{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.35891325071496666, "global_step": 4518, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "learning_rate": 7.936507936507937e-07, "loss": 2.9309, "theoretical_loss": 3.4868973533572363, "tokens_seen": 1650130944 }, { "epoch": 0.0, "learning_rate": 1.5873015873015873e-06, "loss": 3.0573, "theoretical_loss": 3.4868733789973354, "tokens_seen": 1650262016 }, { "epoch": 0.0, "learning_rate": 2.3809523809523808e-06, "loss": 2.9048, "theoretical_loss": 3.4868494070746388, "tokens_seen": 1650393088 }, { "epoch": 0.0, "learning_rate": 3.1746031746031746e-06, "loss": 3.1852, "theoretical_loss": 3.4868254375887053, "tokens_seen": 1650524160 }, { "epoch": 0.0, "learning_rate": 3.968253968253968e-06, "loss": 2.9406, "theoretical_loss": 3.4868014705390937, "tokens_seen": 1650655232 }, { "epoch": 0.0, "learning_rate": 4.7619047619047615e-06, "loss": 2.9301, "theoretical_loss": 3.4867775059253625, "tokens_seen": 1650786304 }, { "epoch": 0.0, "learning_rate": 5.555555555555556e-06, "loss": 2.9694, "theoretical_loss": 3.4867535437470716, "tokens_seen": 1650917376 }, { "epoch": 0.0, "learning_rate": 6.349206349206349e-06, "loss": 2.8537, "theoretical_loss": 3.486729584003779, "tokens_seen": 1651048448 }, { "epoch": 0.0, "learning_rate": 7.142857142857143e-06, "loss": 3.1275, "theoretical_loss": 3.4867056266950454, "tokens_seen": 1651179520 }, { "epoch": 0.0, "learning_rate": 7.936507936507936e-06, "loss": 2.8591, "theoretical_loss": 3.4866816718204294, "tokens_seen": 1651310592 }, { "epoch": 0.0, "learning_rate": 8.73015873015873e-06, "loss": 2.9591, "theoretical_loss": 3.4866577193794903, "tokens_seen": 1651441664 }, { "epoch": 0.0, "learning_rate": 9.523809523809523e-06, "loss": 2.9381, "theoretical_loss": 3.486633769371788, "tokens_seen": 1651572736 }, { "epoch": 0.0, "objective/train/docs_used": 911303, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.985459327697754, "objective/train/theoretical_loss": 3.486621795280263, "objective/train/tokens_used": 22097376, "theoretical_loss": 3.486621795280263, "tokens_seen": 1651638272 }, { "epoch": 0.0, "learning_rate": 1.0317460317460318e-05, "loss": 3.0636, "theoretical_loss": 3.4866098217968826, "tokens_seen": 1651703808 }, { "epoch": 0.0, "learning_rate": 1.1111111111111112e-05, "loss": 2.9173, "theoretical_loss": 3.486585876654333, "tokens_seen": 1651834880 }, { "epoch": 0.0, "learning_rate": 1.1904761904761905e-05, "loss": 2.9235, "theoretical_loss": 3.4865619339437, "tokens_seen": 1651965952 }, { "epoch": 0.0, "learning_rate": 1.2698412698412699e-05, "loss": 2.9036, "theoretical_loss": 3.4865379936645438, "tokens_seen": 1652097024 }, { "epoch": 0.0, "learning_rate": 1.3492063492063492e-05, "loss": 2.8327, "theoretical_loss": 3.486514055816424, "tokens_seen": 1652228096 }, { "epoch": 0.0, "learning_rate": 1.4285714285714285e-05, "loss": 2.8069, "theoretical_loss": 3.4864901203989014, "tokens_seen": 1652359168 }, { "epoch": 0.0, "learning_rate": 1.5079365079365079e-05, "loss": 2.6994, "theoretical_loss": 3.4864661874115357, "tokens_seen": 1652490240 }, { "epoch": 0.0, "learning_rate": 1.5873015873015872e-05, "loss": 2.8277, "theoretical_loss": 3.486442256853888, "tokens_seen": 1652621312 }, { "epoch": 0.0, "learning_rate": 1.6666666666666667e-05, "loss": 2.8779, "theoretical_loss": 3.4864183287255193, "tokens_seen": 1652752384 }, { "epoch": 0.0, "learning_rate": 1.746031746031746e-05, "loss": 2.929, "theoretical_loss": 3.4863944030259897, "tokens_seen": 1652883456 }, { "epoch": 0.0, "learning_rate": 1.8253968253968254e-05, "loss": 2.7098, "theoretical_loss": 3.4863704797548607, "tokens_seen": 1653014528 }, { "epoch": 0.0, "learning_rate": 1.9047619047619046e-05, "loss": 2.7617, "theoretical_loss": 3.4863465589116927, "tokens_seen": 1653145600 }, { "epoch": 0.0, "objective/train/docs_used": 912643, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.1410162448883057, "objective/train/theoretical_loss": 3.486322640496047, "objective/train/tokens_used": 23735776, "theoretical_loss": 3.486322640496047, "tokens_seen": 1653276672 }, { "epoch": 0.0, "learning_rate": 1.984126984126984e-05, "loss": 2.8668, "theoretical_loss": 3.486322640496047, "tokens_seen": 1653276672 }, { "epoch": 0.0, "learning_rate": 2.0634920634920636e-05, "loss": 2.7427, "theoretical_loss": 3.486298724507485, "tokens_seen": 1653407744 }, { "epoch": 0.0, "learning_rate": 2.1428571428571428e-05, "loss": 2.9108, "theoretical_loss": 3.4862748109455675, "tokens_seen": 1653538816 }, { "epoch": 0.0, "learning_rate": 2.2222222222222223e-05, "loss": 2.8479, "theoretical_loss": 3.486250899809857, "tokens_seen": 1653669888 }, { "epoch": 0.0, "learning_rate": 2.3015873015873015e-05, "loss": 2.9489, "theoretical_loss": 3.4862269910999135, "tokens_seen": 1653800960 }, { "epoch": 0.0, "learning_rate": 2.380952380952381e-05, "loss": 2.7904, "theoretical_loss": 3.4862030848153003, "tokens_seen": 1653932032 }, { "epoch": 0.0, "learning_rate": 2.4603174603174602e-05, "loss": 2.6043, "theoretical_loss": 3.4861791809555784, "tokens_seen": 1654063104 }, { "epoch": 0.0, "learning_rate": 2.5396825396825397e-05, "loss": 2.9128, "theoretical_loss": 3.48615527952031, "tokens_seen": 1654194176 }, { "epoch": 0.0, "learning_rate": 2.6190476190476192e-05, "loss": 2.7843, "theoretical_loss": 3.486131380509057, "tokens_seen": 1654325248 }, { "epoch": 0.0, "learning_rate": 2.6984126984126984e-05, "loss": 2.8452, "theoretical_loss": 3.4861074839213813, "tokens_seen": 1654456320 }, { "epoch": 0.0, "learning_rate": 2.777777777777778e-05, "loss": 2.6435, "theoretical_loss": 3.4860835897568454, "tokens_seen": 1654587392 }, { "epoch": 0.0, "learning_rate": 2.857142857142857e-05, "loss": 2.7978, "theoretical_loss": 3.4860596980150116, "tokens_seen": 1654718464 }, { "epoch": 0.0, "learning_rate": 2.9365079365079366e-05, "loss": 2.9673, "theoretical_loss": 3.4860358086954424, "tokens_seen": 1654849536 }, { "epoch": 0.0, "objective/train/docs_used": 913892, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.85264253616333, "objective/train/theoretical_loss": 3.48602386494387, "objective/train/tokens_used": 25374176, "theoretical_loss": 3.48602386494387, "tokens_seen": 1654915072 }, { "epoch": 0.0, "learning_rate": 3.0158730158730158e-05, "loss": 2.9861, "theoretical_loss": 3.4860119217977004, "tokens_seen": 1654980608 }, { "epoch": 0.0, "learning_rate": 3.095238095238095e-05, "loss": 2.9108, "theoretical_loss": 3.485988037321348, "tokens_seen": 1655111680 }, { "epoch": 0.0, "learning_rate": 3.1746031746031745e-05, "loss": 2.7628, "theoretical_loss": 3.4859641552659486, "tokens_seen": 1655242752 }, { "epoch": 0.0, "learning_rate": 3.253968253968254e-05, "loss": 2.6715, "theoretical_loss": 3.485940275631065, "tokens_seen": 1655373824 }, { "epoch": 0.0, "learning_rate": 3.3333333333333335e-05, "loss": 2.9045, "theoretical_loss": 3.4859163984162596, "tokens_seen": 1655504896 }, { "epoch": 0.0, "learning_rate": 3.412698412698413e-05, "loss": 2.7567, "theoretical_loss": 3.4858925236210965, "tokens_seen": 1655635968 }, { "epoch": 0.0, "learning_rate": 3.492063492063492e-05, "loss": 2.787, "theoretical_loss": 3.4858686512451387, "tokens_seen": 1655767040 }, { "epoch": 0.0, "learning_rate": 3.571428571428572e-05, "loss": 2.6928, "theoretical_loss": 3.4858447812879487, "tokens_seen": 1655898112 }, { "epoch": 0.0, "learning_rate": 3.650793650793651e-05, "loss": 2.7383, "theoretical_loss": 3.485820913749091, "tokens_seen": 1656029184 }, { "epoch": 0.0, "learning_rate": 3.730158730158731e-05, "loss": 2.6014, "theoretical_loss": 3.4857970486281293, "tokens_seen": 1656160256 }, { "epoch": 0.0, "learning_rate": 3.809523809523809e-05, "loss": 2.8116, "theoretical_loss": 3.485773185924627, "tokens_seen": 1656291328 }, { "epoch": 0.0, "learning_rate": 3.888888888888889e-05, "loss": 2.6689, "theoretical_loss": 3.4857493256381473, "tokens_seen": 1656422400 }, { "epoch": 0.0, "objective/train/docs_used": 914561, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.774996280670166, "objective/train/theoretical_loss": 3.4857254677682548, "objective/train/tokens_used": 27012576, "theoretical_loss": 3.4857254677682548, "tokens_seen": 1656553472 }, { "epoch": 0.0, "learning_rate": 3.968253968253968e-05, "loss": 2.6872, "theoretical_loss": 3.4857254677682548, "tokens_seen": 1656553472 }, { "epoch": 0.0, "learning_rate": 4.047619047619048e-05, "loss": 2.5722, "theoretical_loss": 3.485701612314514, "tokens_seen": 1656684544 }, { "epoch": 0.0, "learning_rate": 4.126984126984127e-05, "loss": 2.6782, "theoretical_loss": 3.4856777592764883, "tokens_seen": 1656815616 }, { "epoch": 0.0, "learning_rate": 4.2063492063492065e-05, "loss": 2.6618, "theoretical_loss": 3.4856539086537426, "tokens_seen": 1656946688 }, { "epoch": 0.0, "learning_rate": 4.2857142857142856e-05, "loss": 2.6818, "theoretical_loss": 3.485630060445841, "tokens_seen": 1657077760 }, { "epoch": 0.0, "learning_rate": 4.3650793650793655e-05, "loss": 2.4444, "theoretical_loss": 3.485606214652347, "tokens_seen": 1657208832 }, { "epoch": 0.0, "learning_rate": 4.4444444444444447e-05, "loss": 2.5045, "theoretical_loss": 3.4855823712728276, "tokens_seen": 1657339904 }, { "epoch": 0.0, "learning_rate": 4.523809523809524e-05, "loss": 2.5876, "theoretical_loss": 3.4855585303068453, "tokens_seen": 1657470976 }, { "epoch": 0.0, "learning_rate": 4.603174603174603e-05, "loss": 2.5061, "theoretical_loss": 3.4855346917539665, "tokens_seen": 1657602048 }, { "epoch": 0.0, "learning_rate": 4.682539682539683e-05, "loss": 2.6817, "theoretical_loss": 3.4855108556137546, "tokens_seen": 1657733120 }, { "epoch": 0.0, "learning_rate": 4.761904761904762e-05, "loss": 2.6159, "theoretical_loss": 3.4854870218857763, "tokens_seen": 1657864192 }, { "epoch": 0.0, "learning_rate": 4.841269841269841e-05, "loss": 2.5539, "theoretical_loss": 3.485463190569596, "tokens_seen": 1657995264 }, { "epoch": 0.0, "learning_rate": 4.9206349206349204e-05, "loss": 2.6999, "theoretical_loss": 3.485439361664779, "tokens_seen": 1658126336 }, { "epoch": 0.0, "objective/train/docs_used": 915678, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6932287216186523, "objective/train/theoretical_loss": 3.4854274481164964, "objective/train/tokens_used": 28650976, "theoretical_loss": 3.4854274481164964, "tokens_seen": 1658191872 }, { "epoch": 0.01, "learning_rate": 5e-05, "loss": 2.5566, "theoretical_loss": 3.4854155351708913, "tokens_seen": 1658257408 }, { "epoch": 0.01, "learning_rate": 5.0793650793650794e-05, "loss": 2.7198, "theoretical_loss": 3.4853917110874972, "tokens_seen": 1658388480 }, { "epoch": 0.01, "learning_rate": 5.158730158730159e-05, "loss": 2.4993, "theoretical_loss": 3.4853678894141638, "tokens_seen": 1658519552 }, { "epoch": 0.01, "learning_rate": 5.2380952380952384e-05, "loss": 2.5541, "theoretical_loss": 3.4853440701504557, "tokens_seen": 1658650624 }, { "epoch": 0.01, "learning_rate": 5.3174603174603176e-05, "loss": 2.5189, "theoretical_loss": 3.4853202532959395, "tokens_seen": 1658781696 }, { "epoch": 0.01, "learning_rate": 5.396825396825397e-05, "loss": 2.322, "theoretical_loss": 3.4852964388501806, "tokens_seen": 1658912768 }, { "epoch": 0.01, "learning_rate": 5.4761904761904766e-05, "loss": 2.6172, "theoretical_loss": 3.485272626812746, "tokens_seen": 1659043840 }, { "epoch": 0.01, "learning_rate": 5.555555555555556e-05, "loss": 2.5944, "theoretical_loss": 3.4852488171832006, "tokens_seen": 1659174912 }, { "epoch": 0.01, "learning_rate": 5.634920634920635e-05, "loss": 2.5423, "theoretical_loss": 3.4852250099611117, "tokens_seen": 1659305984 }, { "epoch": 0.01, "learning_rate": 5.714285714285714e-05, "loss": 2.5272, "theoretical_loss": 3.485201205146046, "tokens_seen": 1659437056 }, { "epoch": 0.01, "learning_rate": 5.793650793650795e-05, "loss": 2.6078, "theoretical_loss": 3.4851774027375693, "tokens_seen": 1659568128 }, { "epoch": 0.01, "learning_rate": 5.873015873015873e-05, "loss": 2.5119, "theoretical_loss": 3.4851536027352483, "tokens_seen": 1659699200 }, { "epoch": 0.01, "objective/train/docs_used": 916417, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.456707239151001, "objective/train/theoretical_loss": 3.48512980513865, "objective/train/tokens_used": 30289376, "theoretical_loss": 3.48512980513865, "tokens_seen": 1659830272 }, { "epoch": 0.01, "learning_rate": 5.9523809523809524e-05, "loss": 2.6637, "theoretical_loss": 3.48512980513865, "tokens_seen": 1659830272 }, { "epoch": 0.01, "learning_rate": 6.0317460317460316e-05, "loss": 2.3894, "theoretical_loss": 3.4851060099473417, "tokens_seen": 1659961344 }, { "epoch": 0.01, "learning_rate": 6.111111111111112e-05, "loss": 2.5928, "theoretical_loss": 3.4850822171608895, "tokens_seen": 1660092416 }, { "epoch": 0.01, "learning_rate": 6.19047619047619e-05, "loss": 2.4428, "theoretical_loss": 3.4850584267788607, "tokens_seen": 1660223488 }, { "epoch": 0.01, "learning_rate": 6.26984126984127e-05, "loss": 2.7376, "theoretical_loss": 3.4850346388008235, "tokens_seen": 1660354560 }, { "epoch": 0.01, "learning_rate": 6.349206349206349e-05, "loss": 2.4582, "theoretical_loss": 3.4850108532263437, "tokens_seen": 1660485632 }, { "epoch": 0.01, "learning_rate": 6.428571428571429e-05, "loss": 2.409, "theoretical_loss": 3.48498707005499, "tokens_seen": 1660616704 }, { "epoch": 0.01, "learning_rate": 6.507936507936509e-05, "loss": 2.5177, "theoretical_loss": 3.4849632892863296, "tokens_seen": 1660747776 }, { "epoch": 0.01, "learning_rate": 6.587301587301587e-05, "loss": 2.6281, "theoretical_loss": 3.48493951091993, "tokens_seen": 1660878848 }, { "epoch": 0.01, "learning_rate": 6.666666666666667e-05, "loss": 2.5806, "theoretical_loss": 3.4849157349553592, "tokens_seen": 1661009920 }, { "epoch": 0.01, "learning_rate": 6.746031746031747e-05, "loss": 2.5608, "theoretical_loss": 3.4848919613921847, "tokens_seen": 1661140992 }, { "epoch": 0.01, "learning_rate": 6.825396825396825e-05, "loss": 2.5385, "theoretical_loss": 3.484868190229975, "tokens_seen": 1661272064 }, { "epoch": 0.01, "learning_rate": 6.904761904761905e-05, "loss": 2.5632, "theoretical_loss": 3.484844421468298, "tokens_seen": 1661403136 }, { "epoch": 0.01, "objective/train/docs_used": 917689, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.0480103492736816, "objective/train/theoretical_loss": 3.484832537987524, "objective/train/tokens_used": 31927776, "theoretical_loss": 3.484832537987524, "tokens_seen": 1661468672 }, { "epoch": 0.01, "learning_rate": 6.984126984126984e-05, "loss": 2.3842, "theoretical_loss": 3.4848206551067213, "tokens_seen": 1661534208 }, { "epoch": 0.01, "learning_rate": 7.063492063492065e-05, "loss": 2.4595, "theoretical_loss": 3.484796891144814, "tokens_seen": 1661665280 }, { "epoch": 0.01, "learning_rate": 7.142857142857143e-05, "loss": 2.5882, "theoretical_loss": 3.4847731295821447, "tokens_seen": 1661796352 }, { "epoch": 0.01, "learning_rate": 7.222222222222222e-05, "loss": 2.3931, "theoretical_loss": 3.4847493704182817, "tokens_seen": 1661927424 }, { "epoch": 0.01, "learning_rate": 7.301587301587302e-05, "loss": 2.622, "theoretical_loss": 3.4847256136527935, "tokens_seen": 1662058496 }, { "epoch": 0.01, "learning_rate": 7.380952380952382e-05, "loss": 2.5425, "theoretical_loss": 3.4847018592852494, "tokens_seen": 1662189568 }, { "epoch": 0.01, "learning_rate": 7.460317460317461e-05, "loss": 2.601, "theoretical_loss": 3.4846781073152173, "tokens_seen": 1662320640 }, { "epoch": 0.01, "learning_rate": 7.53968253968254e-05, "loss": 2.5712, "theoretical_loss": 3.484654357742267, "tokens_seen": 1662451712 }, { "epoch": 0.01, "learning_rate": 7.619047619047618e-05, "loss": 2.2686, "theoretical_loss": 3.4846306105659677, "tokens_seen": 1662582784 }, { "epoch": 0.01, "learning_rate": 7.6984126984127e-05, "loss": 2.3971, "theoretical_loss": 3.4846068657858877, "tokens_seen": 1662713856 }, { "epoch": 0.01, "learning_rate": 7.777777777777778e-05, "loss": 2.5189, "theoretical_loss": 3.4845831234015976, "tokens_seen": 1662844928 }, { "epoch": 0.01, "learning_rate": 7.857142857142858e-05, "loss": 2.3924, "theoretical_loss": 3.484559383412666, "tokens_seen": 1662976000 }, { "epoch": 0.01, "objective/train/docs_used": 918239, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3232572078704834, "objective/train/theoretical_loss": 3.4845356458186627, "objective/train/tokens_used": 33566176, "theoretical_loss": 3.4845356458186627, "tokens_seen": 1663107072 }, { "epoch": 0.01, "learning_rate": 7.936507936507937e-05, "loss": 2.5743, "theoretical_loss": 3.4845356458186627, "tokens_seen": 1663107072 }, { "epoch": 0.01, "learning_rate": 8.015873015873016e-05, "loss": 2.4116, "theoretical_loss": 3.4845119106191578, "tokens_seen": 1663238144 }, { "epoch": 0.01, "learning_rate": 8.095238095238096e-05, "loss": 2.5243, "theoretical_loss": 3.48448817781372, "tokens_seen": 1663369216 }, { "epoch": 0.01, "learning_rate": 8.174603174603175e-05, "loss": 2.4798, "theoretical_loss": 3.484464447401921, "tokens_seen": 1663500288 }, { "epoch": 0.01, "learning_rate": 8.253968253968255e-05, "loss": 2.3456, "theoretical_loss": 3.4844407193833282, "tokens_seen": 1663631360 }, { "epoch": 0.01, "learning_rate": 8.333333333333334e-05, "loss": 2.4678, "theoretical_loss": 3.484416993757514, "tokens_seen": 1663762432 }, { "epoch": 0.01, "learning_rate": 8.412698412698413e-05, "loss": 2.4106, "theoretical_loss": 3.4843932705240483, "tokens_seen": 1663893504 }, { "epoch": 0.01, "learning_rate": 8.492063492063493e-05, "loss": 2.3669, "theoretical_loss": 3.4843695496825005, "tokens_seen": 1664024576 }, { "epoch": 0.01, "learning_rate": 8.571428571428571e-05, "loss": 2.5092, "theoretical_loss": 3.4843458312324413, "tokens_seen": 1664155648 }, { "epoch": 0.01, "learning_rate": 8.650793650793651e-05, "loss": 2.473, "theoretical_loss": 3.484322115173442, "tokens_seen": 1664286720 }, { "epoch": 0.01, "learning_rate": 8.730158730158731e-05, "loss": 2.5359, "theoretical_loss": 3.4842984015050726, "tokens_seen": 1664417792 }, { "epoch": 0.01, "learning_rate": 8.80952380952381e-05, "loss": 2.6096, "theoretical_loss": 3.484274690226904, "tokens_seen": 1664548864 }, { "epoch": 0.01, "learning_rate": 8.888888888888889e-05, "loss": 2.4428, "theoretical_loss": 3.484250981338507, "tokens_seen": 1664679936 }, { "epoch": 0.01, "objective/train/docs_used": 919306, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.603062629699707, "objective/train/theoretical_loss": 3.484239127790339, "objective/train/tokens_used": 35204576, "theoretical_loss": 3.484239127790339, "tokens_seen": 1664745472 }, { "epoch": 0.01, "learning_rate": 8.968253968253969e-05, "loss": 2.5448, "theoretical_loss": 3.484227274839453, "tokens_seen": 1664811008 }, { "epoch": 0.01, "learning_rate": 9.047619047619048e-05, "loss": 2.5698, "theoretical_loss": 3.484203570729313, "tokens_seen": 1664942080 }, { "epoch": 0.01, "learning_rate": 9.126984126984128e-05, "loss": 2.4996, "theoretical_loss": 3.484179869007658, "tokens_seen": 1665073152 }, { "epoch": 0.01, "learning_rate": 9.206349206349206e-05, "loss": 2.5152, "theoretical_loss": 3.4841561696740597, "tokens_seen": 1665204224 }, { "epoch": 0.01, "learning_rate": 9.285714285714286e-05, "loss": 2.4996, "theoretical_loss": 3.4841324727280893, "tokens_seen": 1665335296 }, { "epoch": 0.01, "learning_rate": 9.365079365079366e-05, "loss": 2.47, "theoretical_loss": 3.484108778169318, "tokens_seen": 1665466368 }, { "epoch": 0.01, "learning_rate": 9.444444444444444e-05, "loss": 2.3922, "theoretical_loss": 3.484085085997318, "tokens_seen": 1665597440 }, { "epoch": 0.01, "learning_rate": 9.523809523809524e-05, "loss": 2.4767, "theoretical_loss": 3.484061396211661, "tokens_seen": 1665728512 }, { "epoch": 0.01, "learning_rate": 9.603174603174604e-05, "loss": 2.5377, "theoretical_loss": 3.4840377088119188, "tokens_seen": 1665859584 }, { "epoch": 0.01, "learning_rate": 9.682539682539682e-05, "loss": 2.61, "theoretical_loss": 3.4840140237976636, "tokens_seen": 1665990656 }, { "epoch": 0.01, "learning_rate": 9.761904761904762e-05, "loss": 2.4506, "theoretical_loss": 3.483990341168467, "tokens_seen": 1666121728 }, { "epoch": 0.01, "learning_rate": 9.841269841269841e-05, "loss": 2.4311, "theoretical_loss": 3.483966660923902, "tokens_seen": 1666252800 }, { "epoch": 0.01, "objective/train/docs_used": 919806, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3898022174835205, "objective/train/theoretical_loss": 3.4839429830635407, "objective/train/tokens_used": 36842976, "theoretical_loss": 3.4839429830635407, "tokens_seen": 1666383872 }, { "epoch": 0.01, "learning_rate": 9.920634920634922e-05, "loss": 2.543, "theoretical_loss": 3.4839429830635407, "tokens_seen": 1666383872 }, { "epoch": 0.01, "learning_rate": 0.0001, "loss": 2.4395, "theoretical_loss": 3.4839193075869543, "tokens_seen": 1666514944 }, { "epoch": 0.01, "learning_rate": 9.999197560584176e-05, "loss": 2.4916, "theoretical_loss": 3.4838956344937175, "tokens_seen": 1666646016 }, { "epoch": 0.01, "learning_rate": 9.998395121168352e-05, "loss": 2.3888, "theoretical_loss": 3.4838719637834012, "tokens_seen": 1666777088 }, { "epoch": 0.01, "learning_rate": 9.997592681752528e-05, "loss": 2.5063, "theoretical_loss": 3.483848295455579, "tokens_seen": 1666908160 }, { "epoch": 0.01, "learning_rate": 9.996790242336704e-05, "loss": 2.4951, "theoretical_loss": 3.483824629509824, "tokens_seen": 1667039232 }, { "epoch": 0.01, "learning_rate": 9.99598780292088e-05, "loss": 2.4226, "theoretical_loss": 3.4838009659457088, "tokens_seen": 1667170304 }, { "epoch": 0.01, "learning_rate": 9.995185363505056e-05, "loss": 2.4431, "theoretical_loss": 3.4837773047628064, "tokens_seen": 1667301376 }, { "epoch": 0.01, "learning_rate": 9.994382924089232e-05, "loss": 2.4434, "theoretical_loss": 3.4837536459606904, "tokens_seen": 1667432448 }, { "epoch": 0.01, "learning_rate": 9.993580484673407e-05, "loss": 2.5021, "theoretical_loss": 3.4837299895389333, "tokens_seen": 1667563520 }, { "epoch": 0.01, "learning_rate": 9.992778045257584e-05, "loss": 2.5622, "theoretical_loss": 3.48370633549711, "tokens_seen": 1667694592 }, { "epoch": 0.01, "learning_rate": 9.991975605841759e-05, "loss": 2.524, "theoretical_loss": 3.4836826838347923, "tokens_seen": 1667825664 }, { "epoch": 0.01, "learning_rate": 9.991173166425936e-05, "loss": 2.5055, "theoretical_loss": 3.483659034551555, "tokens_seen": 1667956736 }, { "epoch": 0.01, "objective/train/docs_used": 921005, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6571173667907715, "objective/train/theoretical_loss": 3.4836472108019585, "objective/train/tokens_used": 38481376, "theoretical_loss": 3.4836472108019585, "tokens_seen": 1668022272 }, { "epoch": 0.01, "learning_rate": 9.990370727010111e-05, "loss": 2.5956, "theoretical_loss": 3.483635387646972, "tokens_seen": 1668087808 }, { "epoch": 0.01, "learning_rate": 9.989568287594288e-05, "loss": 2.5279, "theoretical_loss": 3.4836117431206164, "tokens_seen": 1668218880 }, { "epoch": 0.01, "learning_rate": 9.988765848178463e-05, "loss": 2.45, "theoretical_loss": 3.483588100972063, "tokens_seen": 1668349952 }, { "epoch": 0.01, "learning_rate": 9.987963408762638e-05, "loss": 2.5056, "theoretical_loss": 3.483564461200885, "tokens_seen": 1668481024 }, { "epoch": 0.01, "learning_rate": 9.987160969346815e-05, "loss": 2.56, "theoretical_loss": 3.4835408238066567, "tokens_seen": 1668612096 }, { "epoch": 0.01, "learning_rate": 9.98635852993099e-05, "loss": 2.4673, "theoretical_loss": 3.4835171887889533, "tokens_seen": 1668743168 }, { "epoch": 0.01, "learning_rate": 9.985556090515167e-05, "loss": 2.3887, "theoretical_loss": 3.483493556147349, "tokens_seen": 1668874240 }, { "epoch": 0.01, "learning_rate": 9.984753651099342e-05, "loss": 2.478, "theoretical_loss": 3.4834699258814172, "tokens_seen": 1669005312 }, { "epoch": 0.01, "learning_rate": 9.983951211683519e-05, "loss": 2.6136, "theoretical_loss": 3.483446297990734, "tokens_seen": 1669136384 }, { "epoch": 0.01, "learning_rate": 9.983148772267694e-05, "loss": 2.5876, "theoretical_loss": 3.4834226724748736, "tokens_seen": 1669267456 }, { "epoch": 0.01, "learning_rate": 9.982346332851871e-05, "loss": 2.5149, "theoretical_loss": 3.4833990493334106, "tokens_seen": 1669398528 }, { "epoch": 0.01, "learning_rate": 9.981543893436046e-05, "loss": 2.5755, "theoretical_loss": 3.48337542856592, "tokens_seen": 1669529600 }, { "epoch": 0.01, "objective/train/docs_used": 921680, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.329754590988159, "objective/train/theoretical_loss": 3.4833518101719774, "objective/train/tokens_used": 40119776, "theoretical_loss": 3.4833518101719774, "tokens_seen": 1669660672 }, { "epoch": 0.01, "learning_rate": 9.980741454020222e-05, "loss": 2.4671, "theoretical_loss": 3.4833518101719774, "tokens_seen": 1669660672 }, { "epoch": 0.01, "learning_rate": 9.979939014604398e-05, "loss": 2.3942, "theoretical_loss": 3.4833281941511576, "tokens_seen": 1669791744 }, { "epoch": 0.01, "learning_rate": 9.979136575188574e-05, "loss": 2.4857, "theoretical_loss": 3.4833045805030363, "tokens_seen": 1669922816 }, { "epoch": 0.01, "learning_rate": 9.97833413577275e-05, "loss": 2.4502, "theoretical_loss": 3.483280969227188, "tokens_seen": 1670053888 }, { "epoch": 0.01, "learning_rate": 9.977531696356926e-05, "loss": 2.5831, "theoretical_loss": 3.4832573603231887, "tokens_seen": 1670184960 }, { "epoch": 0.01, "learning_rate": 9.976729256941102e-05, "loss": 2.561, "theoretical_loss": 3.4832337537906146, "tokens_seen": 1670316032 }, { "epoch": 0.01, "learning_rate": 9.975926817525277e-05, "loss": 2.5869, "theoretical_loss": 3.4832101496290404, "tokens_seen": 1670447104 }, { "epoch": 0.01, "learning_rate": 9.975124378109453e-05, "loss": 2.4206, "theoretical_loss": 3.4831865478380433, "tokens_seen": 1670578176 }, { "epoch": 0.01, "learning_rate": 9.97432193869363e-05, "loss": 2.3991, "theoretical_loss": 3.4831629484171978, "tokens_seen": 1670709248 }, { "epoch": 0.01, "learning_rate": 9.973519499277805e-05, "loss": 2.3821, "theoretical_loss": 3.4831393513660807, "tokens_seen": 1670840320 }, { "epoch": 0.01, "learning_rate": 9.972717059861981e-05, "loss": 2.5258, "theoretical_loss": 3.483115756684268, "tokens_seen": 1670971392 }, { "epoch": 0.01, "learning_rate": 9.971914620446157e-05, "loss": 2.4604, "theoretical_loss": 3.4830921643713366, "tokens_seen": 1671102464 }, { "epoch": 0.01, "learning_rate": 9.971112181030333e-05, "loss": 2.5027, "theoretical_loss": 3.483068574426862, "tokens_seen": 1671233536 }, { "epoch": 0.01, "objective/train/docs_used": 922598, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.674171209335327, "objective/train/theoretical_loss": 3.4830567803426637, "objective/train/tokens_used": 41758176, "theoretical_loss": 3.4830567803426637, "tokens_seen": 1671299072 }, { "epoch": 0.01, "learning_rate": 9.970309741614509e-05, "loss": 2.5194, "theoretical_loss": 3.483044986850421, "tokens_seen": 1671364608 }, { "epoch": 0.01, "learning_rate": 9.969507302198684e-05, "loss": 2.4984, "theoretical_loss": 3.48302140164159, "tokens_seen": 1671495680 }, { "epoch": 0.01, "learning_rate": 9.968704862782861e-05, "loss": 2.6113, "theoretical_loss": 3.482997818799947, "tokens_seen": 1671626752 }, { "epoch": 0.01, "learning_rate": 9.967902423367036e-05, "loss": 2.6013, "theoretical_loss": 3.4829742383250673, "tokens_seen": 1671757824 }, { "epoch": 0.01, "learning_rate": 9.967099983951213e-05, "loss": 2.4007, "theoretical_loss": 3.4829506602165283, "tokens_seen": 1671888896 }, { "epoch": 0.01, "learning_rate": 9.966297544535388e-05, "loss": 2.4473, "theoretical_loss": 3.482927084473907, "tokens_seen": 1672019968 }, { "epoch": 0.01, "learning_rate": 9.965495105119565e-05, "loss": 2.3941, "theoretical_loss": 3.482903511096781, "tokens_seen": 1672151040 }, { "epoch": 0.01, "learning_rate": 9.96469266570374e-05, "loss": 2.5564, "theoretical_loss": 3.482879940084727, "tokens_seen": 1672282112 }, { "epoch": 0.01, "learning_rate": 9.963890226287915e-05, "loss": 2.4979, "theoretical_loss": 3.4828563714373226, "tokens_seen": 1672413184 }, { "epoch": 0.01, "learning_rate": 9.963087786872092e-05, "loss": 2.4506, "theoretical_loss": 3.4828328051541453, "tokens_seen": 1672544256 }, { "epoch": 0.01, "learning_rate": 9.962285347456267e-05, "loss": 2.5131, "theoretical_loss": 3.482809241234773, "tokens_seen": 1672675328 }, { "epoch": 0.01, "learning_rate": 9.961482908040444e-05, "loss": 2.4079, "theoretical_loss": 3.482785679678783, "tokens_seen": 1672806400 }, { "epoch": 0.01, "objective/train/docs_used": 923236, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.478933095932007, "objective/train/theoretical_loss": 3.482762120485753, "objective/train/tokens_used": 43396576, "theoretical_loss": 3.482762120485753, "tokens_seen": 1672937472 }, { "epoch": 0.01, "learning_rate": 9.960680468624619e-05, "loss": 2.5005, "theoretical_loss": 3.482762120485753, "tokens_seen": 1672937472 }, { "epoch": 0.01, "learning_rate": 9.959878029208796e-05, "loss": 2.669, "theoretical_loss": 3.482738563655261, "tokens_seen": 1673068544 }, { "epoch": 0.01, "learning_rate": 9.959075589792971e-05, "loss": 2.5706, "theoretical_loss": 3.4827150091868853, "tokens_seen": 1673199616 }, { "epoch": 0.01, "learning_rate": 9.958273150377147e-05, "loss": 2.5295, "theoretical_loss": 3.482691457080204, "tokens_seen": 1673330688 }, { "epoch": 0.01, "learning_rate": 9.957470710961323e-05, "loss": 2.4358, "theoretical_loss": 3.482667907334795, "tokens_seen": 1673461760 }, { "epoch": 0.01, "learning_rate": 9.956668271545499e-05, "loss": 2.4467, "theoretical_loss": 3.482644359950237, "tokens_seen": 1673592832 }, { "epoch": 0.01, "learning_rate": 9.955865832129675e-05, "loss": 2.3988, "theoretical_loss": 3.4826208149261078, "tokens_seen": 1673723904 }, { "epoch": 0.01, "learning_rate": 9.95506339271385e-05, "loss": 2.521, "theoretical_loss": 3.482597272261987, "tokens_seen": 1673854976 }, { "epoch": 0.01, "learning_rate": 9.954260953298027e-05, "loss": 2.409, "theoretical_loss": 3.4825737319574523, "tokens_seen": 1673986048 }, { "epoch": 0.01, "learning_rate": 9.953458513882203e-05, "loss": 2.5548, "theoretical_loss": 3.4825501940120835, "tokens_seen": 1674117120 }, { "epoch": 0.01, "learning_rate": 9.952656074466378e-05, "loss": 2.4541, "theoretical_loss": 3.4825266584254586, "tokens_seen": 1674248192 }, { "epoch": 0.01, "learning_rate": 9.951853635050554e-05, "loss": 2.5549, "theoretical_loss": 3.482503125197157, "tokens_seen": 1674379264 }, { "epoch": 0.01, "learning_rate": 9.95105119563473e-05, "loss": 2.4687, "theoretical_loss": 3.4824795943267577, "tokens_seen": 1674510336 }, { "epoch": 0.01, "objective/train/docs_used": 924425, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.032245397567749, "objective/train/theoretical_loss": 3.48246782977564, "objective/train/tokens_used": 45034976, "theoretical_loss": 3.48246782977564, "tokens_seen": 1674575872 }, { "epoch": 0.01, "learning_rate": 9.950248756218906e-05, "loss": 2.4781, "theoretical_loss": 3.4824560658138397, "tokens_seen": 1674641408 }, { "epoch": 0.02, "learning_rate": 9.949446316803082e-05, "loss": 2.6007, "theoretical_loss": 3.482432539657983, "tokens_seen": 1674772480 }, { "epoch": 0.02, "learning_rate": 9.948643877387258e-05, "loss": 2.4198, "theoretical_loss": 3.4824090158587664, "tokens_seen": 1674903552 }, { "epoch": 0.02, "learning_rate": 9.947841437971434e-05, "loss": 2.49, "theoretical_loss": 3.4823854944157695, "tokens_seen": 1675034624 }, { "epoch": 0.02, "learning_rate": 9.94703899855561e-05, "loss": 2.4733, "theoretical_loss": 3.482361975328572, "tokens_seen": 1675165696 }, { "epoch": 0.02, "learning_rate": 9.946236559139786e-05, "loss": 2.6164, "theoretical_loss": 3.482338458596754, "tokens_seen": 1675296768 }, { "epoch": 0.02, "learning_rate": 9.945434119723961e-05, "loss": 2.4729, "theoretical_loss": 3.482314944219895, "tokens_seen": 1675427840 }, { "epoch": 0.02, "learning_rate": 9.944631680308138e-05, "loss": 2.5106, "theoretical_loss": 3.482291432197575, "tokens_seen": 1675558912 }, { "epoch": 0.02, "learning_rate": 9.943829240892313e-05, "loss": 2.4255, "theoretical_loss": 3.482267922529374, "tokens_seen": 1675689984 }, { "epoch": 0.02, "learning_rate": 9.94302680147649e-05, "loss": 2.4172, "theoretical_loss": 3.482244415214873, "tokens_seen": 1675821056 }, { "epoch": 0.02, "learning_rate": 9.942224362060665e-05, "loss": 2.4485, "theoretical_loss": 3.482220910253651, "tokens_seen": 1675952128 }, { "epoch": 0.02, "learning_rate": 9.941421922644842e-05, "loss": 2.5783, "theoretical_loss": 3.482197407645289, "tokens_seen": 1676083200 }, { "epoch": 0.02, "objective/train/docs_used": 924948, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.1431517601013184, "objective/train/theoretical_loss": 3.4821739073893676, "objective/train/tokens_used": 46673376, "theoretical_loss": 3.4821739073893676, "tokens_seen": 1676214272 }, { "epoch": 0.02, "learning_rate": 9.940619483229017e-05, "loss": 2.4842, "theoretical_loss": 3.4821739073893676, "tokens_seen": 1676214272 }, { "epoch": 0.02, "learning_rate": 9.939817043813192e-05, "loss": 2.5048, "theoretical_loss": 3.482150409485467, "tokens_seen": 1676345344 }, { "epoch": 0.02, "learning_rate": 9.939014604397369e-05, "loss": 2.4242, "theoretical_loss": 3.482126913933169, "tokens_seen": 1676476416 }, { "epoch": 0.02, "learning_rate": 9.938212164981544e-05, "loss": 2.4716, "theoretical_loss": 3.482103420732053, "tokens_seen": 1676607488 }, { "epoch": 0.02, "learning_rate": 9.937409725565721e-05, "loss": 2.4006, "theoretical_loss": 3.482079929881701, "tokens_seen": 1676738560 }, { "epoch": 0.02, "learning_rate": 9.936607286149896e-05, "loss": 2.5438, "theoretical_loss": 3.482056441381694, "tokens_seen": 1676869632 }, { "epoch": 0.02, "learning_rate": 9.935804846734073e-05, "loss": 2.4948, "theoretical_loss": 3.4820329552316123, "tokens_seen": 1677000704 }, { "epoch": 0.02, "learning_rate": 9.935002407318248e-05, "loss": 2.4135, "theoretical_loss": 3.482009471431038, "tokens_seen": 1677131776 }, { "epoch": 0.02, "learning_rate": 9.934199967902424e-05, "loss": 2.4302, "theoretical_loss": 3.4819859899795516, "tokens_seen": 1677262848 }, { "epoch": 0.02, "learning_rate": 9.9333975284866e-05, "loss": 2.4168, "theoretical_loss": 3.481962510876736, "tokens_seen": 1677393920 }, { "epoch": 0.02, "learning_rate": 9.932595089070776e-05, "loss": 2.648, "theoretical_loss": 3.481939034122171, "tokens_seen": 1677524992 }, { "epoch": 0.02, "learning_rate": 9.931792649654952e-05, "loss": 2.4332, "theoretical_loss": 3.4819155597154396, "tokens_seen": 1677656064 }, { "epoch": 0.02, "learning_rate": 9.930990210239128e-05, "loss": 2.4625, "theoretical_loss": 3.4818920876561235, "tokens_seen": 1677787136 }, { "epoch": 0.02, "objective/train/docs_used": 926165, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.494450569152832, "objective/train/theoretical_loss": 3.4818803525066153, "objective/train/tokens_used": 48311776, "theoretical_loss": 3.4818803525066153, "tokens_seen": 1677852672 }, { "epoch": 0.02, "learning_rate": 9.930187770823304e-05, "loss": 2.4562, "theoretical_loss": 3.481868617943804, "tokens_seen": 1677918208 }, { "epoch": 0.02, "learning_rate": 9.92938533140748e-05, "loss": 2.4934, "theoretical_loss": 3.481845150578063, "tokens_seen": 1678049280 }, { "epoch": 0.02, "learning_rate": 9.928582891991655e-05, "loss": 2.4625, "theoretical_loss": 3.481821685558484, "tokens_seen": 1678180352 }, { "epoch": 0.02, "learning_rate": 9.927780452575832e-05, "loss": 2.3664, "theoretical_loss": 3.4817982228846476, "tokens_seen": 1678311424 }, { "epoch": 0.02, "learning_rate": 9.926978013160007e-05, "loss": 2.4554, "theoretical_loss": 3.481774762556137, "tokens_seen": 1678442496 }, { "epoch": 0.02, "learning_rate": 9.926175573744183e-05, "loss": 2.6398, "theoretical_loss": 3.4817513045725343, "tokens_seen": 1678573568 }, { "epoch": 0.02, "learning_rate": 9.925373134328359e-05, "loss": 2.3632, "theoretical_loss": 3.4817278489334225, "tokens_seen": 1678704640 }, { "epoch": 0.02, "learning_rate": 9.924570694912535e-05, "loss": 2.6002, "theoretical_loss": 3.481704395638383, "tokens_seen": 1678835712 }, { "epoch": 0.02, "learning_rate": 9.923768255496711e-05, "loss": 2.5935, "theoretical_loss": 3.4816809446870005, "tokens_seen": 1678966784 }, { "epoch": 0.02, "learning_rate": 9.922965816080886e-05, "loss": 2.6889, "theoretical_loss": 3.4816574960788564, "tokens_seen": 1679097856 }, { "epoch": 0.02, "learning_rate": 9.922163376665063e-05, "loss": 2.2718, "theoretical_loss": 3.4816340498135343, "tokens_seen": 1679228928 }, { "epoch": 0.02, "learning_rate": 9.921360937249238e-05, "loss": 2.4179, "theoretical_loss": 3.4816106058906175, "tokens_seen": 1679360000 }, { "epoch": 0.02, "objective/train/docs_used": 926923, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7133426666259766, "objective/train/theoretical_loss": 3.481587164309688, "objective/train/tokens_used": 49950176, "theoretical_loss": 3.481587164309688, "tokens_seen": 1679491072 }, { "epoch": 0.02, "learning_rate": 9.920558497833415e-05, "loss": 2.4835, "theoretical_loss": 3.481587164309688, "tokens_seen": 1679491072 }, { "epoch": 0.02, "learning_rate": 9.91975605841759e-05, "loss": 2.5025, "theoretical_loss": 3.4815637250703304, "tokens_seen": 1679622144 }, { "epoch": 0.02, "learning_rate": 9.918953619001767e-05, "loss": 2.4519, "theoretical_loss": 3.4815402881721274, "tokens_seen": 1679753216 }, { "epoch": 0.02, "learning_rate": 9.918151179585942e-05, "loss": 2.5021, "theoretical_loss": 3.4815168536146626, "tokens_seen": 1679884288 }, { "epoch": 0.02, "learning_rate": 9.917348740170119e-05, "loss": 2.4785, "theoretical_loss": 3.48149342139752, "tokens_seen": 1680015360 }, { "epoch": 0.02, "learning_rate": 9.916546300754294e-05, "loss": 2.5138, "theoretical_loss": 3.481469991520283, "tokens_seen": 1680146432 }, { "epoch": 0.02, "learning_rate": 9.91574386133847e-05, "loss": 2.4019, "theoretical_loss": 3.4814465639825354, "tokens_seen": 1680277504 }, { "epoch": 0.02, "learning_rate": 9.914941421922646e-05, "loss": 2.5178, "theoretical_loss": 3.481423138783861, "tokens_seen": 1680408576 }, { "epoch": 0.02, "learning_rate": 9.914138982506821e-05, "loss": 2.4615, "theoretical_loss": 3.4813997159238443, "tokens_seen": 1680539648 }, { "epoch": 0.02, "learning_rate": 9.913336543090998e-05, "loss": 2.533, "theoretical_loss": 3.481376295402069, "tokens_seen": 1680670720 }, { "epoch": 0.02, "learning_rate": 9.912534103675173e-05, "loss": 2.5408, "theoretical_loss": 3.4813528772181193, "tokens_seen": 1680801792 }, { "epoch": 0.02, "learning_rate": 9.91173166425935e-05, "loss": 2.4864, "theoretical_loss": 3.4813294613715797, "tokens_seen": 1680932864 }, { "epoch": 0.02, "learning_rate": 9.910929224843525e-05, "loss": 2.567, "theoretical_loss": 3.4813060478620352, "tokens_seen": 1681063936 }, { "epoch": 0.02, "objective/train/docs_used": 927976, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.612856864929199, "objective/train/theoretical_loss": 3.4812943419835065, "objective/train/tokens_used": 51588576, "theoretical_loss": 3.4812943419835065, "tokens_seen": 1681129472 }, { "epoch": 0.02, "learning_rate": 9.9101267854277e-05, "loss": 2.515, "theoretical_loss": 3.48128263668907, "tokens_seen": 1681195008 }, { "epoch": 0.02, "learning_rate": 9.909324346011877e-05, "loss": 2.5095, "theoretical_loss": 3.4812592278522687, "tokens_seen": 1681326080 }, { "epoch": 0.02, "learning_rate": 9.908521906596053e-05, "loss": 2.6043, "theoretical_loss": 3.4812358213512162, "tokens_seen": 1681457152 }, { "epoch": 0.02, "learning_rate": 9.907719467180229e-05, "loss": 2.6276, "theoretical_loss": 3.4812124171854966, "tokens_seen": 1681588224 }, { "epoch": 0.02, "learning_rate": 9.906917027764405e-05, "loss": 2.3657, "theoretical_loss": 3.481189015354696, "tokens_seen": 1681719296 }, { "epoch": 0.02, "learning_rate": 9.906114588348581e-05, "loss": 2.4345, "theoretical_loss": 3.481165615858399, "tokens_seen": 1681850368 }, { "epoch": 0.02, "learning_rate": 9.905312148932757e-05, "loss": 2.6024, "theoretical_loss": 3.481142218696191, "tokens_seen": 1681981440 }, { "epoch": 0.02, "learning_rate": 9.904509709516932e-05, "loss": 2.5844, "theoretical_loss": 3.481118823867657, "tokens_seen": 1682112512 }, { "epoch": 0.02, "learning_rate": 9.903707270101109e-05, "loss": 2.4985, "theoretical_loss": 3.481095431372383, "tokens_seen": 1682243584 }, { "epoch": 0.02, "learning_rate": 9.902904830685284e-05, "loss": 2.4409, "theoretical_loss": 3.481072041209954, "tokens_seen": 1682374656 }, { "epoch": 0.02, "learning_rate": 9.90210239126946e-05, "loss": 2.4786, "theoretical_loss": 3.481048653379955, "tokens_seen": 1682505728 }, { "epoch": 0.02, "learning_rate": 9.901299951853636e-05, "loss": 2.4033, "theoretical_loss": 3.4810252678819733, "tokens_seen": 1682636800 }, { "epoch": 0.02, "objective/train/docs_used": 929273, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6835052967071533, "objective/train/theoretical_loss": 3.481001884715594, "objective/train/tokens_used": 53226976, "theoretical_loss": 3.481001884715594, "tokens_seen": 1682767872 }, { "epoch": 0.02, "learning_rate": 9.900497512437812e-05, "loss": 2.5045, "theoretical_loss": 3.481001884715594, "tokens_seen": 1682767872 }, { "epoch": 0.02, "learning_rate": 9.899695073021988e-05, "loss": 2.4493, "theoretical_loss": 3.4809785038804026, "tokens_seen": 1682898944 }, { "epoch": 0.02, "learning_rate": 9.898892633606163e-05, "loss": 2.5062, "theoretical_loss": 3.480955125375986, "tokens_seen": 1683030016 }, { "epoch": 0.02, "learning_rate": 9.89809019419034e-05, "loss": 2.5045, "theoretical_loss": 3.48093174920193, "tokens_seen": 1683161088 }, { "epoch": 0.02, "learning_rate": 9.897287754774515e-05, "loss": 2.5248, "theoretical_loss": 3.48090837535782, "tokens_seen": 1683292160 }, { "epoch": 0.02, "learning_rate": 9.896485315358692e-05, "loss": 2.5576, "theoretical_loss": 3.4808850038432437, "tokens_seen": 1683423232 }, { "epoch": 0.02, "learning_rate": 9.895682875942867e-05, "loss": 2.398, "theoretical_loss": 3.4808616346577868, "tokens_seen": 1683554304 }, { "epoch": 0.02, "learning_rate": 9.894880436527044e-05, "loss": 2.5147, "theoretical_loss": 3.4808382678010363, "tokens_seen": 1683685376 }, { "epoch": 0.02, "learning_rate": 9.894077997111219e-05, "loss": 2.5106, "theoretical_loss": 3.480814903272579, "tokens_seen": 1683816448 }, { "epoch": 0.02, "learning_rate": 9.893275557695394e-05, "loss": 2.4724, "theoretical_loss": 3.4807915410720005, "tokens_seen": 1683947520 }, { "epoch": 0.02, "learning_rate": 9.892473118279571e-05, "loss": 2.5067, "theoretical_loss": 3.4807681811988895, "tokens_seen": 1684078592 }, { "epoch": 0.02, "learning_rate": 9.891670678863746e-05, "loss": 2.5444, "theoretical_loss": 3.4807448236528318, "tokens_seen": 1684209664 }, { "epoch": 0.02, "learning_rate": 9.890868239447923e-05, "loss": 2.5985, "theoretical_loss": 3.4807214684334147, "tokens_seen": 1684340736 }, { "epoch": 0.02, "objective/train/docs_used": 929850, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.568603992462158, "objective/train/theoretical_loss": 3.480709791696068, "objective/train/tokens_used": 54865376, "theoretical_loss": 3.480709791696068, "tokens_seen": 1684406272 }, { "epoch": 0.02, "learning_rate": 9.890065800032098e-05, "loss": 2.4515, "theoretical_loss": 3.4806981155402257, "tokens_seen": 1684471808 }, { "epoch": 0.02, "learning_rate": 9.889263360616275e-05, "loss": 2.4635, "theoretical_loss": 3.4806747649728518, "tokens_seen": 1684602880 }, { "epoch": 0.02, "learning_rate": 9.88846092120045e-05, "loss": 2.5249, "theoretical_loss": 3.480651416730881, "tokens_seen": 1684733952 }, { "epoch": 0.02, "learning_rate": 9.887658481784626e-05, "loss": 2.5362, "theoretical_loss": 3.4806280708139, "tokens_seen": 1684865024 }, { "epoch": 0.02, "learning_rate": 9.886856042368802e-05, "loss": 2.4881, "theoretical_loss": 3.480604727221497, "tokens_seen": 1684996096 }, { "epoch": 0.02, "learning_rate": 9.886053602952978e-05, "loss": 2.555, "theoretical_loss": 3.4805813859532595, "tokens_seen": 1685127168 }, { "epoch": 0.02, "learning_rate": 9.885251163537154e-05, "loss": 2.6367, "theoretical_loss": 3.480558047008776, "tokens_seen": 1685258240 }, { "epoch": 0.02, "learning_rate": 9.88444872412133e-05, "loss": 2.4256, "theoretical_loss": 3.4805347103876327, "tokens_seen": 1685389312 }, { "epoch": 0.02, "learning_rate": 9.883646284705506e-05, "loss": 2.2893, "theoretical_loss": 3.48051137608942, "tokens_seen": 1685520384 }, { "epoch": 0.02, "learning_rate": 9.882843845289682e-05, "loss": 2.5847, "theoretical_loss": 3.4804880441137245, "tokens_seen": 1685651456 }, { "epoch": 0.02, "learning_rate": 9.882041405873857e-05, "loss": 2.6211, "theoretical_loss": 3.4804647144601346, "tokens_seen": 1685782528 }, { "epoch": 0.02, "learning_rate": 9.881238966458034e-05, "loss": 2.4622, "theoretical_loss": 3.4804413871282396, "tokens_seen": 1685913600 }, { "epoch": 0.02, "objective/train/docs_used": 931063, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.578721523284912, "objective/train/theoretical_loss": 3.480418062117627, "objective/train/tokens_used": 56503776, "theoretical_loss": 3.480418062117627, "tokens_seen": 1686044672 }, { "epoch": 0.02, "learning_rate": 9.880436527042209e-05, "loss": 2.5689, "theoretical_loss": 3.480418062117627, "tokens_seen": 1686044672 }, { "epoch": 0.02, "learning_rate": 9.879634087626386e-05, "loss": 2.4969, "theoretical_loss": 3.4803947394278856, "tokens_seen": 1686175744 }, { "epoch": 0.02, "learning_rate": 9.878831648210561e-05, "loss": 2.5164, "theoretical_loss": 3.4803714190586037, "tokens_seen": 1686306816 }, { "epoch": 0.02, "learning_rate": 9.878029208794736e-05, "loss": 2.4945, "theoretical_loss": 3.4803481010093718, "tokens_seen": 1686437888 }, { "epoch": 0.02, "learning_rate": 9.877226769378913e-05, "loss": 2.399, "theoretical_loss": 3.4803247852797767, "tokens_seen": 1686568960 }, { "epoch": 0.02, "learning_rate": 9.876424329963088e-05, "loss": 2.4799, "theoretical_loss": 3.4803014718694087, "tokens_seen": 1686700032 }, { "epoch": 0.02, "learning_rate": 9.875621890547265e-05, "loss": 2.35, "theoretical_loss": 3.480278160777856, "tokens_seen": 1686831104 }, { "epoch": 0.02, "learning_rate": 9.87481945113144e-05, "loss": 2.4886, "theoretical_loss": 3.480254852004709, "tokens_seen": 1686962176 }, { "epoch": 0.02, "learning_rate": 9.874017011715617e-05, "loss": 2.5754, "theoretical_loss": 3.4802315455495565, "tokens_seen": 1687093248 }, { "epoch": 0.02, "learning_rate": 9.873214572299792e-05, "loss": 2.5741, "theoretical_loss": 3.480208241411987, "tokens_seen": 1687224320 }, { "epoch": 0.02, "learning_rate": 9.872412132883967e-05, "loss": 2.5755, "theoretical_loss": 3.480184939591591, "tokens_seen": 1687355392 }, { "epoch": 0.02, "learning_rate": 9.871609693468144e-05, "loss": 2.4047, "theoretical_loss": 3.4801616400879585, "tokens_seen": 1687486464 }, { "epoch": 0.02, "learning_rate": 9.87080725405232e-05, "loss": 2.4871, "theoretical_loss": 3.480138342900678, "tokens_seen": 1687617536 }, { "epoch": 0.02, "objective/train/docs_used": 931367, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5317652225494385, "objective/train/theoretical_loss": 3.480126695175542, "objective/train/tokens_used": 58142176, "theoretical_loss": 3.480126695175542, "tokens_seen": 1687683072 }, { "epoch": 0.02, "learning_rate": 9.870004814636496e-05, "loss": 2.6203, "theoretical_loss": 3.48011504802934, "tokens_seen": 1687748608 }, { "epoch": 0.02, "learning_rate": 9.869202375220671e-05, "loss": 2.4963, "theoretical_loss": 3.4800917554735347, "tokens_seen": 1687879680 }, { "epoch": 0.02, "learning_rate": 9.868399935804847e-05, "loss": 2.5753, "theoretical_loss": 3.480068465232852, "tokens_seen": 1688010752 }, { "epoch": 0.02, "learning_rate": 9.867597496389023e-05, "loss": 2.546, "theoretical_loss": 3.4800451773068817, "tokens_seen": 1688141824 }, { "epoch": 0.02, "learning_rate": 9.866795056973199e-05, "loss": 2.3913, "theoretical_loss": 3.4800218916952144, "tokens_seen": 1688272896 }, { "epoch": 0.02, "learning_rate": 9.865992617557375e-05, "loss": 2.5884, "theoretical_loss": 3.47999860839744, "tokens_seen": 1688403968 }, { "epoch": 0.02, "learning_rate": 9.86519017814155e-05, "loss": 2.5369, "theoretical_loss": 3.4799753274131495, "tokens_seen": 1688535040 }, { "epoch": 0.02, "learning_rate": 9.864387738725727e-05, "loss": 2.6298, "theoretical_loss": 3.4799520487419335, "tokens_seen": 1688666112 }, { "epoch": 0.02, "learning_rate": 9.863585299309903e-05, "loss": 2.567, "theoretical_loss": 3.4799287723833823, "tokens_seen": 1688797184 }, { "epoch": 0.02, "learning_rate": 9.862782859894078e-05, "loss": 2.4449, "theoretical_loss": 3.4799054983370867, "tokens_seen": 1688928256 }, { "epoch": 0.02, "learning_rate": 9.861980420478255e-05, "loss": 2.5075, "theoretical_loss": 3.4798822266026384, "tokens_seen": 1689059328 }, { "epoch": 0.02, "learning_rate": 9.86117798106243e-05, "loss": 2.5648, "theoretical_loss": 3.4798589571796272, "tokens_seen": 1689190400 }, { "epoch": 0.02, "objective/train/docs_used": 932516, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2863895893096924, "objective/train/theoretical_loss": 3.4798356900676444, "objective/train/tokens_used": 59780576, "theoretical_loss": 3.4798356900676444, "tokens_seen": 1689321472 }, { "epoch": 0.02, "learning_rate": 9.860375541646607e-05, "loss": 2.4437, "theoretical_loss": 3.4798356900676444, "tokens_seen": 1689321472 }, { "epoch": 0.02, "learning_rate": 9.859573102230782e-05, "loss": 2.4915, "theoretical_loss": 3.479812425266282, "tokens_seen": 1689452544 }, { "epoch": 0.02, "learning_rate": 9.858770662814957e-05, "loss": 2.5486, "theoretical_loss": 3.479789162775131, "tokens_seen": 1689583616 }, { "epoch": 0.02, "learning_rate": 9.857968223399134e-05, "loss": 2.4756, "theoretical_loss": 3.479765902593782, "tokens_seen": 1689714688 }, { "epoch": 0.02, "learning_rate": 9.857165783983309e-05, "loss": 2.4487, "theoretical_loss": 3.4797426447218274, "tokens_seen": 1689845760 }, { "epoch": 0.02, "learning_rate": 9.856363344567486e-05, "loss": 2.5253, "theoretical_loss": 3.4797193891588583, "tokens_seen": 1689976832 }, { "epoch": 0.02, "learning_rate": 9.855560905151661e-05, "loss": 2.4716, "theoretical_loss": 3.479696135904467, "tokens_seen": 1690107904 }, { "epoch": 0.02, "learning_rate": 9.854758465735838e-05, "loss": 2.5138, "theoretical_loss": 3.479672884958245, "tokens_seen": 1690238976 }, { "epoch": 0.02, "learning_rate": 9.853956026320013e-05, "loss": 2.4722, "theoretical_loss": 3.4796496363197837, "tokens_seen": 1690370048 }, { "epoch": 0.02, "learning_rate": 9.853153586904188e-05, "loss": 2.4039, "theoretical_loss": 3.4796263899886757, "tokens_seen": 1690501120 }, { "epoch": 0.02, "learning_rate": 9.852351147488365e-05, "loss": 2.2929, "theoretical_loss": 3.4796031459645134, "tokens_seen": 1690632192 }, { "epoch": 0.02, "learning_rate": 9.85154870807254e-05, "loss": 2.5445, "theoretical_loss": 3.4795799042468882, "tokens_seen": 1690763264 }, { "epoch": 0.02, "learning_rate": 9.850746268656717e-05, "loss": 2.4633, "theoretical_loss": 3.4795566648353935, "tokens_seen": 1690894336 }, { "epoch": 0.02, "objective/train/docs_used": 933032, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.596432685852051, "objective/train/theoretical_loss": 3.4795450459943167, "objective/train/tokens_used": 61418976, "theoretical_loss": 3.4795450459943167, "tokens_seen": 1690959872 }, { "epoch": 0.02, "learning_rate": 9.849943829240892e-05, "loss": 2.5766, "theoretical_loss": 3.4795334277296206, "tokens_seen": 1691025408 }, { "epoch": 0.02, "learning_rate": 9.849141389825068e-05, "loss": 2.4845, "theoretical_loss": 3.4795101929291627, "tokens_seen": 1691156480 }, { "epoch": 0.03, "learning_rate": 9.848338950409244e-05, "loss": 2.4419, "theoretical_loss": 3.4794869604336123, "tokens_seen": 1691287552 }, { "epoch": 0.03, "learning_rate": 9.84753651099342e-05, "loss": 2.6624, "theoretical_loss": 3.4794637302425624, "tokens_seen": 1691418624 }, { "epoch": 0.03, "learning_rate": 9.846734071577596e-05, "loss": 2.4082, "theoretical_loss": 3.4794405023556054, "tokens_seen": 1691549696 }, { "epoch": 0.03, "learning_rate": 9.845931632161772e-05, "loss": 2.5021, "theoretical_loss": 3.4794172767723346, "tokens_seen": 1691680768 }, { "epoch": 0.03, "learning_rate": 9.845129192745948e-05, "loss": 2.6212, "theoretical_loss": 3.479394053492343, "tokens_seen": 1691811840 }, { "epoch": 0.03, "learning_rate": 9.844326753330124e-05, "loss": 2.4537, "theoretical_loss": 3.4793708325152237, "tokens_seen": 1691942912 }, { "epoch": 0.03, "learning_rate": 9.843524313914299e-05, "loss": 2.5129, "theoretical_loss": 3.47934761384057, "tokens_seen": 1692073984 }, { "epoch": 0.03, "learning_rate": 9.842721874498476e-05, "loss": 2.5562, "theoretical_loss": 3.4793243974679755, "tokens_seen": 1692205056 }, { "epoch": 0.03, "learning_rate": 9.841919435082651e-05, "loss": 2.409, "theoretical_loss": 3.479301183397033, "tokens_seen": 1692336128 }, { "epoch": 0.03, "learning_rate": 9.841116995666828e-05, "loss": 2.5037, "theoretical_loss": 3.4792779716273365, "tokens_seen": 1692467200 }, { "epoch": 0.03, "objective/train/docs_used": 934026, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.466470241546631, "objective/train/theoretical_loss": 3.4792547621584804, "objective/train/tokens_used": 63057376, "theoretical_loss": 3.4792547621584804, "tokens_seen": 1692598272 }, { "epoch": 0.03, "learning_rate": 9.840314556251003e-05, "loss": 2.4919, "theoretical_loss": 3.4792547621584804, "tokens_seen": 1692598272 }, { "epoch": 0.03, "learning_rate": 9.839512116835178e-05, "loss": 2.5155, "theoretical_loss": 3.479231554990057, "tokens_seen": 1692729344 }, { "epoch": 0.03, "learning_rate": 9.838709677419355e-05, "loss": 2.4349, "theoretical_loss": 3.4792083501216613, "tokens_seen": 1692860416 }, { "epoch": 0.03, "learning_rate": 9.83790723800353e-05, "loss": 2.2746, "theoretical_loss": 3.4791851475528874, "tokens_seen": 1692991488 }, { "epoch": 0.03, "learning_rate": 9.837104798587707e-05, "loss": 2.6368, "theoretical_loss": 3.4791619472833286, "tokens_seen": 1693122560 }, { "epoch": 0.03, "learning_rate": 9.836302359171882e-05, "loss": 2.4889, "theoretical_loss": 3.479138749312579, "tokens_seen": 1693253632 }, { "epoch": 0.03, "learning_rate": 9.835499919756059e-05, "loss": 2.5798, "theoretical_loss": 3.4791155536402343, "tokens_seen": 1693384704 }, { "epoch": 0.03, "learning_rate": 9.834697480340234e-05, "loss": 2.6683, "theoretical_loss": 3.479092360265887, "tokens_seen": 1693515776 }, { "epoch": 0.03, "learning_rate": 9.83389504092441e-05, "loss": 2.4851, "theoretical_loss": 3.479069169189133, "tokens_seen": 1693646848 }, { "epoch": 0.03, "learning_rate": 9.833092601508586e-05, "loss": 2.3859, "theoretical_loss": 3.4790459804095666, "tokens_seen": 1693777920 }, { "epoch": 0.03, "learning_rate": 9.832290162092762e-05, "loss": 2.5253, "theoretical_loss": 3.4790227939267826, "tokens_seen": 1693908992 }, { "epoch": 0.03, "learning_rate": 9.831487722676938e-05, "loss": 2.2977, "theoretical_loss": 3.478999609740375, "tokens_seen": 1694040064 }, { "epoch": 0.03, "learning_rate": 9.830685283261113e-05, "loss": 2.4333, "theoretical_loss": 3.4789764278499398, "tokens_seen": 1694171136 }, { "epoch": 0.03, "objective/train/docs_used": 934482, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2957074642181396, "objective/train/theoretical_loss": 3.478964837765585, "objective/train/tokens_used": 64695776, "theoretical_loss": 3.478964837765585, "tokens_seen": 1694236672 }, { "epoch": 0.03, "learning_rate": 9.82988284384529e-05, "loss": 2.5303, "theoretical_loss": 3.4789532482550714, "tokens_seen": 1694302208 }, { "epoch": 0.03, "learning_rate": 9.829080404429465e-05, "loss": 2.4688, "theoretical_loss": 3.478930070955365, "tokens_seen": 1694433280 }, { "epoch": 0.03, "learning_rate": 9.828277965013641e-05, "loss": 2.443, "theoretical_loss": 3.478906895950416, "tokens_seen": 1694564352 }, { "epoch": 0.03, "learning_rate": 9.827475525597817e-05, "loss": 2.5702, "theoretical_loss": 3.4788837232398198, "tokens_seen": 1694695424 }, { "epoch": 0.03, "learning_rate": 9.826673086181993e-05, "loss": 2.3373, "theoretical_loss": 3.4788605528231713, "tokens_seen": 1694826496 }, { "epoch": 0.03, "learning_rate": 9.82587064676617e-05, "loss": 2.4548, "theoretical_loss": 3.4788373847000664, "tokens_seen": 1694957568 }, { "epoch": 0.03, "learning_rate": 9.825068207350345e-05, "loss": 2.6429, "theoretical_loss": 3.478814218870101, "tokens_seen": 1695088640 }, { "epoch": 0.03, "learning_rate": 9.824265767934521e-05, "loss": 2.3599, "theoretical_loss": 3.4787910553328705, "tokens_seen": 1695219712 }, { "epoch": 0.03, "learning_rate": 9.823463328518697e-05, "loss": 2.3772, "theoretical_loss": 3.4787678940879707, "tokens_seen": 1695350784 }, { "epoch": 0.03, "learning_rate": 9.822660889102873e-05, "loss": 2.5326, "theoretical_loss": 3.478744735134998, "tokens_seen": 1695481856 }, { "epoch": 0.03, "learning_rate": 9.821858449687049e-05, "loss": 2.6442, "theoretical_loss": 3.4787215784735475, "tokens_seen": 1695612928 }, { "epoch": 0.03, "learning_rate": 9.821056010271224e-05, "loss": 2.5572, "theoretical_loss": 3.478698424103216, "tokens_seen": 1695744000 }, { "epoch": 0.03, "objective/train/docs_used": 935631, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5070106983184814, "objective/train/theoretical_loss": 3.4786752720236005, "objective/train/tokens_used": 66334176, "theoretical_loss": 3.4786752720236005, "tokens_seen": 1695875072 }, { "epoch": 0.03, "learning_rate": 9.820253570855401e-05, "loss": 2.6289, "theoretical_loss": 3.4786752720236005, "tokens_seen": 1695875072 }, { "epoch": 0.03, "learning_rate": 9.819451131439576e-05, "loss": 2.659, "theoretical_loss": 3.478652122234296, "tokens_seen": 1696006144 }, { "epoch": 0.03, "learning_rate": 9.818648692023753e-05, "loss": 2.575, "theoretical_loss": 3.4786289747348995, "tokens_seen": 1696137216 }, { "epoch": 0.03, "learning_rate": 9.817846252607928e-05, "loss": 2.637, "theoretical_loss": 3.4786058295250077, "tokens_seen": 1696268288 }, { "epoch": 0.03, "learning_rate": 9.817043813192105e-05, "loss": 2.5304, "theoretical_loss": 3.4785826866042173, "tokens_seen": 1696399360 }, { "epoch": 0.03, "learning_rate": 9.81624137377628e-05, "loss": 2.4708, "theoretical_loss": 3.4785595459721246, "tokens_seen": 1696530432 }, { "epoch": 0.03, "learning_rate": 9.815438934360455e-05, "loss": 2.5821, "theoretical_loss": 3.4785364076283267, "tokens_seen": 1696661504 }, { "epoch": 0.03, "learning_rate": 9.814636494944632e-05, "loss": 2.6878, "theoretical_loss": 3.478513271572421, "tokens_seen": 1696792576 }, { "epoch": 0.03, "learning_rate": 9.813834055528807e-05, "loss": 2.5641, "theoretical_loss": 3.4784901378040036, "tokens_seen": 1696923648 }, { "epoch": 0.03, "learning_rate": 9.813031616112984e-05, "loss": 2.5774, "theoretical_loss": 3.4784670063226732, "tokens_seen": 1697054720 }, { "epoch": 0.03, "learning_rate": 9.812229176697159e-05, "loss": 2.5069, "theoretical_loss": 3.4784438771280257, "tokens_seen": 1697185792 }, { "epoch": 0.03, "learning_rate": 9.811426737281336e-05, "loss": 2.4408, "theoretical_loss": 3.4784207502196587, "tokens_seen": 1697316864 }, { "epoch": 0.03, "learning_rate": 9.810624297865511e-05, "loss": 2.6084, "theoretical_loss": 3.47839762559717, "tokens_seen": 1697447936 }, { "epoch": 0.03, "objective/train/docs_used": 936086, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.701629877090454, "objective/train/theoretical_loss": 3.478386064143004, "objective/train/tokens_used": 67972576, "theoretical_loss": 3.478386064143004, "tokens_seen": 1697513472 }, { "epoch": 0.03, "learning_rate": 9.809821858449687e-05, "loss": 2.5616, "theoretical_loss": 3.4783745032601567, "tokens_seen": 1697579008 }, { "epoch": 0.03, "learning_rate": 9.809019419033863e-05, "loss": 2.5896, "theoretical_loss": 3.4783513832082176, "tokens_seen": 1697710080 }, { "epoch": 0.03, "learning_rate": 9.808216979618039e-05, "loss": 2.4651, "theoretical_loss": 3.4783282654409495, "tokens_seen": 1697841152 }, { "epoch": 0.03, "learning_rate": 9.807414540202215e-05, "loss": 2.5471, "theoretical_loss": 3.4783051499579507, "tokens_seen": 1697972224 }, { "epoch": 0.03, "learning_rate": 9.80661210078639e-05, "loss": 2.5858, "theoretical_loss": 3.4782820367588183, "tokens_seen": 1698103296 }, { "epoch": 0.03, "learning_rate": 9.805809661370567e-05, "loss": 2.4588, "theoretical_loss": 3.4782589258431518, "tokens_seen": 1698234368 }, { "epoch": 0.03, "learning_rate": 9.805007221954742e-05, "loss": 2.4407, "theoretical_loss": 3.4782358172105483, "tokens_seen": 1698365440 }, { "epoch": 0.03, "learning_rate": 9.804204782538918e-05, "loss": 2.6123, "theoretical_loss": 3.478212710860607, "tokens_seen": 1698496512 }, { "epoch": 0.03, "learning_rate": 9.803402343123094e-05, "loss": 2.3523, "theoretical_loss": 3.4781896067929257, "tokens_seen": 1698627584 }, { "epoch": 0.03, "learning_rate": 9.80259990370727e-05, "loss": 2.5561, "theoretical_loss": 3.4781665050071027, "tokens_seen": 1698758656 }, { "epoch": 0.03, "learning_rate": 9.801797464291446e-05, "loss": 2.5973, "theoretical_loss": 3.4781434055027365, "tokens_seen": 1698889728 }, { "epoch": 0.03, "learning_rate": 9.800995024875622e-05, "loss": 2.3148, "theoretical_loss": 3.4781203082794265, "tokens_seen": 1699020800 }, { "epoch": 0.03, "objective/train/docs_used": 937360, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2998149394989014, "objective/train/theoretical_loss": 3.4780972133367714, "objective/train/tokens_used": 69610976, "theoretical_loss": 3.4780972133367714, "tokens_seen": 1699151872 }, { "epoch": 0.03, "learning_rate": 9.800192585459798e-05, "loss": 2.5517, "theoretical_loss": 3.4780972133367714, "tokens_seen": 1699151872 }, { "epoch": 0.03, "learning_rate": 9.799390146043974e-05, "loss": 2.5891, "theoretical_loss": 3.47807412067437, "tokens_seen": 1699282944 }, { "epoch": 0.03, "learning_rate": 9.798587706628149e-05, "loss": 2.5741, "theoretical_loss": 3.478051030291821, "tokens_seen": 1699414016 }, { "epoch": 0.03, "learning_rate": 9.797785267212326e-05, "loss": 2.6164, "theoretical_loss": 3.4780279421887235, "tokens_seen": 1699545088 }, { "epoch": 0.03, "learning_rate": 9.796982827796501e-05, "loss": 2.6742, "theoretical_loss": 3.478004856364677, "tokens_seen": 1699676160 }, { "epoch": 0.03, "learning_rate": 9.796180388380678e-05, "loss": 2.5587, "theoretical_loss": 3.4779817728192803, "tokens_seen": 1699807232 }, { "epoch": 0.03, "learning_rate": 9.795377948964853e-05, "loss": 2.41, "theoretical_loss": 3.477958691552134, "tokens_seen": 1699938304 }, { "epoch": 0.03, "learning_rate": 9.79457550954903e-05, "loss": 2.4044, "theoretical_loss": 3.4779356125628365, "tokens_seen": 1700069376 }, { "epoch": 0.03, "learning_rate": 9.793773070133205e-05, "loss": 2.6259, "theoretical_loss": 3.4779125358509875, "tokens_seen": 1700200448 }, { "epoch": 0.03, "learning_rate": 9.792970630717382e-05, "loss": 2.54, "theoretical_loss": 3.4778894614161873, "tokens_seen": 1700331520 }, { "epoch": 0.03, "learning_rate": 9.792168191301557e-05, "loss": 2.4259, "theoretical_loss": 3.4778663892580353, "tokens_seen": 1700462592 }, { "epoch": 0.03, "learning_rate": 9.791365751885732e-05, "loss": 2.7077, "theoretical_loss": 3.477843319376131, "tokens_seen": 1700593664 }, { "epoch": 0.03, "learning_rate": 9.790563312469909e-05, "loss": 2.6439, "theoretical_loss": 3.4778202517700754, "tokens_seen": 1700724736 }, { "epoch": 0.03, "objective/train/docs_used": 938428, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8109610080718994, "objective/train/theoretical_loss": 3.4778087188203655, "objective/train/tokens_used": 71249376, "theoretical_loss": 3.4778087188203655, "tokens_seen": 1700790272 }, { "epoch": 0.03, "learning_rate": 9.789760873054084e-05, "loss": 2.4903, "theoretical_loss": 3.477797186439468, "tokens_seen": 1700855808 }, { "epoch": 0.03, "learning_rate": 9.788958433638261e-05, "loss": 2.6681, "theoretical_loss": 3.477774123383909, "tokens_seen": 1700986880 }, { "epoch": 0.03, "learning_rate": 9.788155994222436e-05, "loss": 2.4544, "theoretical_loss": 3.477751062602999, "tokens_seen": 1701117952 }, { "epoch": 0.03, "learning_rate": 9.787353554806613e-05, "loss": 2.627, "theoretical_loss": 3.4777280040963383, "tokens_seen": 1701249024 }, { "epoch": 0.03, "learning_rate": 9.786551115390788e-05, "loss": 2.5838, "theoretical_loss": 3.477704947863527, "tokens_seen": 1701380096 }, { "epoch": 0.03, "learning_rate": 9.785748675974964e-05, "loss": 2.5787, "theoretical_loss": 3.4776818939041663, "tokens_seen": 1701511168 }, { "epoch": 0.03, "learning_rate": 9.78494623655914e-05, "loss": 2.6475, "theoretical_loss": 3.477658842217857, "tokens_seen": 1701642240 }, { "epoch": 0.03, "learning_rate": 9.784143797143316e-05, "loss": 2.529, "theoretical_loss": 3.4776357928041994, "tokens_seen": 1701773312 }, { "epoch": 0.03, "learning_rate": 9.783341357727492e-05, "loss": 2.4775, "theoretical_loss": 3.4776127456627948, "tokens_seen": 1701904384 }, { "epoch": 0.03, "learning_rate": 9.782538918311668e-05, "loss": 2.6175, "theoretical_loss": 3.477589700793244, "tokens_seen": 1702035456 }, { "epoch": 0.03, "learning_rate": 9.781736478895844e-05, "loss": 2.5555, "theoretical_loss": 3.477566658195148, "tokens_seen": 1702166528 }, { "epoch": 0.03, "learning_rate": 9.78093403948002e-05, "loss": 2.5131, "theoretical_loss": 3.4775436178681085, "tokens_seen": 1702297600 }, { "epoch": 0.03, "objective/train/docs_used": 939017, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7329940795898438, "objective/train/theoretical_loss": 3.477520579811727, "objective/train/tokens_used": 72887776, "theoretical_loss": 3.477520579811727, "tokens_seen": 1702428672 }, { "epoch": 0.03, "learning_rate": 9.780131600064195e-05, "loss": 2.5521, "theoretical_loss": 3.477520579811727, "tokens_seen": 1702428672 }, { "epoch": 0.03, "learning_rate": 9.779329160648371e-05, "loss": 2.6362, "theoretical_loss": 3.4774975440256037, "tokens_seen": 1702559744 }, { "epoch": 0.03, "learning_rate": 9.778526721232547e-05, "loss": 2.6066, "theoretical_loss": 3.4774745105093414, "tokens_seen": 1702690816 }, { "epoch": 0.03, "learning_rate": 9.777724281816723e-05, "loss": 2.5312, "theoretical_loss": 3.477451479262541, "tokens_seen": 1702821888 }, { "epoch": 0.03, "learning_rate": 9.776921842400899e-05, "loss": 2.5702, "theoretical_loss": 3.477428450284805, "tokens_seen": 1702952960 }, { "epoch": 0.03, "learning_rate": 9.776119402985075e-05, "loss": 2.6024, "theoretical_loss": 3.4774054235757346, "tokens_seen": 1703084032 }, { "epoch": 0.03, "learning_rate": 9.775316963569251e-05, "loss": 2.7145, "theoretical_loss": 3.4773823991349317, "tokens_seen": 1703215104 }, { "epoch": 0.03, "learning_rate": 9.774514524153426e-05, "loss": 2.4479, "theoretical_loss": 3.477359376961999, "tokens_seen": 1703346176 }, { "epoch": 0.03, "learning_rate": 9.773712084737603e-05, "loss": 2.5557, "theoretical_loss": 3.4773363570565374, "tokens_seen": 1703477248 }, { "epoch": 0.03, "learning_rate": 9.772909645321778e-05, "loss": 2.514, "theoretical_loss": 3.47731333941815, "tokens_seen": 1703608320 }, { "epoch": 0.03, "learning_rate": 9.772107205905955e-05, "loss": 2.4775, "theoretical_loss": 3.4772903240464395, "tokens_seen": 1703739392 }, { "epoch": 0.03, "learning_rate": 9.77130476649013e-05, "loss": 2.5532, "theoretical_loss": 3.4772673109410075, "tokens_seen": 1703870464 }, { "epoch": 0.03, "learning_rate": 9.770502327074307e-05, "loss": 2.5049, "theoretical_loss": 3.4772443001014564, "tokens_seen": 1704001536 }, { "epoch": 0.03, "objective/train/docs_used": 939656, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5736076831817627, "objective/train/theoretical_loss": 3.4772327955312625, "objective/train/tokens_used": 74526176, "theoretical_loss": 3.4772327955312625, "tokens_seen": 1704067072 }, { "epoch": 0.03, "learning_rate": 9.769699887658482e-05, "loss": 2.5509, "theoretical_loss": 3.47722129152739, "tokens_seen": 1704132608 }, { "epoch": 0.03, "learning_rate": 9.768897448242657e-05, "loss": 2.5471, "theoretical_loss": 3.4771982852184102, "tokens_seen": 1704263680 }, { "epoch": 0.03, "learning_rate": 9.768095008826834e-05, "loss": 2.4507, "theoretical_loss": 3.4771752811741194, "tokens_seen": 1704394752 }, { "epoch": 0.03, "learning_rate": 9.767292569411009e-05, "loss": 2.5674, "theoretical_loss": 3.477152279394122, "tokens_seen": 1704525824 }, { "epoch": 0.03, "learning_rate": 9.766490129995186e-05, "loss": 2.618, "theoretical_loss": 3.4771292798780196, "tokens_seen": 1704656896 }, { "epoch": 0.03, "learning_rate": 9.765687690579361e-05, "loss": 2.5407, "theoretical_loss": 3.4771062826254155, "tokens_seen": 1704787968 }, { "epoch": 0.03, "learning_rate": 9.764885251163538e-05, "loss": 2.6236, "theoretical_loss": 3.477083287635914, "tokens_seen": 1704919040 }, { "epoch": 0.03, "learning_rate": 9.764082811747713e-05, "loss": 2.7451, "theoretical_loss": 3.4770602949091174, "tokens_seen": 1705050112 }, { "epoch": 0.03, "learning_rate": 9.763280372331889e-05, "loss": 2.5501, "theoretical_loss": 3.4770373044446297, "tokens_seen": 1705181184 }, { "epoch": 0.03, "learning_rate": 9.762477932916065e-05, "loss": 2.6444, "theoretical_loss": 3.4770143162420544, "tokens_seen": 1705312256 }, { "epoch": 0.03, "learning_rate": 9.76167549350024e-05, "loss": 2.6054, "theoretical_loss": 3.4769913303009945, "tokens_seen": 1705443328 }, { "epoch": 0.03, "learning_rate": 9.760873054084417e-05, "loss": 2.6033, "theoretical_loss": 3.4769683466210544, "tokens_seen": 1705574400 }, { "epoch": 0.03, "objective/train/docs_used": 940998, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.642338752746582, "objective/train/theoretical_loss": 3.4769453652018374, "objective/train/tokens_used": 76164576, "theoretical_loss": 3.4769453652018374, "tokens_seen": 1705705472 }, { "epoch": 0.03, "learning_rate": 9.760070614668593e-05, "loss": 2.5956, "theoretical_loss": 3.4769453652018374, "tokens_seen": 1705705472 }, { "epoch": 0.03, "learning_rate": 9.759268175252769e-05, "loss": 2.5025, "theoretical_loss": 3.4769223860429483, "tokens_seen": 1705836544 }, { "epoch": 0.03, "learning_rate": 9.758465735836945e-05, "loss": 2.6917, "theoretical_loss": 3.47689940914399, "tokens_seen": 1705967616 }, { "epoch": 0.03, "learning_rate": 9.757663296421121e-05, "loss": 2.5622, "theoretical_loss": 3.4768764345045677, "tokens_seen": 1706098688 }, { "epoch": 0.03, "learning_rate": 9.756860857005296e-05, "loss": 2.5032, "theoretical_loss": 3.4768534621242853, "tokens_seen": 1706229760 }, { "epoch": 0.03, "learning_rate": 9.756058417589472e-05, "loss": 2.6986, "theoretical_loss": 3.4768304920027466, "tokens_seen": 1706360832 }, { "epoch": 0.03, "learning_rate": 9.755255978173648e-05, "loss": 2.5713, "theoretical_loss": 3.4768075241395566, "tokens_seen": 1706491904 }, { "epoch": 0.03, "learning_rate": 9.754453538757824e-05, "loss": 2.6232, "theoretical_loss": 3.4767845585343196, "tokens_seen": 1706622976 }, { "epoch": 0.03, "learning_rate": 9.753651099342e-05, "loss": 2.5417, "theoretical_loss": 3.47676159518664, "tokens_seen": 1706754048 }, { "epoch": 0.03, "learning_rate": 9.752848659926176e-05, "loss": 2.5311, "theoretical_loss": 3.4767386340961233, "tokens_seen": 1706885120 }, { "epoch": 0.03, "learning_rate": 9.752046220510352e-05, "loss": 2.5849, "theoretical_loss": 3.4767156752623736, "tokens_seen": 1707016192 }, { "epoch": 0.03, "learning_rate": 9.751243781094528e-05, "loss": 2.6568, "theoretical_loss": 3.476692718684996, "tokens_seen": 1707147264 }, { "epoch": 0.03, "learning_rate": 9.750441341678703e-05, "loss": 2.4312, "theoretical_loss": 3.4766697643635958, "tokens_seen": 1707278336 }, { "epoch": 0.03, "objective/train/docs_used": 942016, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.860006809234619, "objective/train/theoretical_loss": 3.4766582880487635, "objective/train/tokens_used": 77802976, "theoretical_loss": 3.4766582880487635, "tokens_seen": 1707343872 }, { "epoch": 0.03, "learning_rate": 9.74963890226288e-05, "loss": 2.6738, "theoretical_loss": 3.4766468122977776, "tokens_seen": 1707409408 }, { "epoch": 0.03, "learning_rate": 9.748836462847055e-05, "loss": 2.507, "theoretical_loss": 3.4766238624871475, "tokens_seen": 1707540480 }, { "epoch": 0.03, "learning_rate": 9.748034023431232e-05, "loss": 2.6324, "theoretical_loss": 3.4766009149313097, "tokens_seen": 1707671552 }, { "epoch": 0.04, "learning_rate": 9.747231584015407e-05, "loss": 2.5417, "theoretical_loss": 3.4765779696298704, "tokens_seen": 1707802624 }, { "epoch": 0.04, "learning_rate": 9.746429144599584e-05, "loss": 2.568, "theoretical_loss": 3.476555026582435, "tokens_seen": 1707933696 }, { "epoch": 0.04, "learning_rate": 9.745626705183759e-05, "loss": 2.4916, "theoretical_loss": 3.4765320857886097, "tokens_seen": 1708064768 }, { "epoch": 0.04, "learning_rate": 9.744824265767934e-05, "loss": 2.4715, "theoretical_loss": 3.476509147247999, "tokens_seen": 1708195840 }, { "epoch": 0.04, "learning_rate": 9.744021826352111e-05, "loss": 2.4469, "theoretical_loss": 3.4764862109602097, "tokens_seen": 1708326912 }, { "epoch": 0.04, "learning_rate": 9.743219386936286e-05, "loss": 2.6352, "theoretical_loss": 3.476463276924847, "tokens_seen": 1708457984 }, { "epoch": 0.04, "learning_rate": 9.742416947520463e-05, "loss": 2.6409, "theoretical_loss": 3.4764403451415173, "tokens_seen": 1708589056 }, { "epoch": 0.04, "learning_rate": 9.741614508104638e-05, "loss": 2.6966, "theoretical_loss": 3.476417415609827, "tokens_seen": 1708720128 }, { "epoch": 0.04, "learning_rate": 9.740812068688815e-05, "loss": 2.5196, "theoretical_loss": 3.476394488329382, "tokens_seen": 1708851200 }, { "epoch": 0.04, "objective/train/docs_used": 942685, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.43986439704895, "objective/train/theoretical_loss": 3.4763715632997885, "objective/train/tokens_used": 79441376, "theoretical_loss": 3.4763715632997885, "tokens_seen": 1708982272 }, { "epoch": 0.04, "learning_rate": 9.74000962927299e-05, "loss": 2.5859, "theoretical_loss": 3.4763715632997885, "tokens_seen": 1708982272 }, { "epoch": 0.04, "learning_rate": 9.739207189857166e-05, "loss": 2.669, "theoretical_loss": 3.476348640520653, "tokens_seen": 1709113344 }, { "epoch": 0.04, "learning_rate": 9.738404750441342e-05, "loss": 2.6132, "theoretical_loss": 3.476325719991582, "tokens_seen": 1709244416 }, { "epoch": 0.04, "learning_rate": 9.737602311025518e-05, "loss": 2.5786, "theoretical_loss": 3.4763028017121824, "tokens_seen": 1709375488 }, { "epoch": 0.04, "learning_rate": 9.736799871609694e-05, "loss": 2.4671, "theoretical_loss": 3.476279885682061, "tokens_seen": 1709506560 }, { "epoch": 0.04, "learning_rate": 9.73599743219387e-05, "loss": 2.8103, "theoretical_loss": 3.476256971900824, "tokens_seen": 1709637632 }, { "epoch": 0.04, "learning_rate": 9.735194992778046e-05, "loss": 2.6216, "theoretical_loss": 3.476234060368079, "tokens_seen": 1709768704 }, { "epoch": 0.04, "learning_rate": 9.734392553362222e-05, "loss": 2.5491, "theoretical_loss": 3.476211151083432, "tokens_seen": 1709899776 }, { "epoch": 0.04, "learning_rate": 9.733590113946397e-05, "loss": 2.5661, "theoretical_loss": 3.476188244046491, "tokens_seen": 1710030848 }, { "epoch": 0.04, "learning_rate": 9.732787674530574e-05, "loss": 2.7132, "theoretical_loss": 3.4761653392568634, "tokens_seen": 1710161920 }, { "epoch": 0.04, "learning_rate": 9.731985235114749e-05, "loss": 2.5686, "theoretical_loss": 3.476142436714156, "tokens_seen": 1710292992 }, { "epoch": 0.04, "learning_rate": 9.731182795698925e-05, "loss": 2.7215, "theoretical_loss": 3.4761195364179764, "tokens_seen": 1710424064 }, { "epoch": 0.04, "learning_rate": 9.730380356283101e-05, "loss": 2.4812, "theoretical_loss": 3.476096638367932, "tokens_seen": 1710555136 }, { "epoch": 0.04, "objective/train/docs_used": 944008, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.545758008956909, "objective/train/theoretical_loss": 3.4760851901850875, "objective/train/tokens_used": 81079776, "theoretical_loss": 3.4760851901850875, "tokens_seen": 1710620672 }, { "epoch": 0.04, "learning_rate": 9.729577916867277e-05, "loss": 2.5449, "theoretical_loss": 3.4760737425636297, "tokens_seen": 1710686208 }, { "epoch": 0.04, "learning_rate": 9.728775477451453e-05, "loss": 2.7649, "theoretical_loss": 3.476050849004679, "tokens_seen": 1710817280 }, { "epoch": 0.04, "learning_rate": 9.72797303803563e-05, "loss": 2.4776, "theoretical_loss": 3.476027957690686, "tokens_seen": 1710948352 }, { "epoch": 0.04, "learning_rate": 9.727170598619805e-05, "loss": 2.4808, "theoretical_loss": 3.4760050686212596, "tokens_seen": 1711079424 }, { "epoch": 0.04, "learning_rate": 9.72636815920398e-05, "loss": 2.5526, "theoretical_loss": 3.475982181796007, "tokens_seen": 1711210496 }, { "epoch": 0.04, "learning_rate": 9.725565719788157e-05, "loss": 2.5859, "theoretical_loss": 3.475959297214537, "tokens_seen": 1711341568 }, { "epoch": 0.04, "learning_rate": 9.724763280372332e-05, "loss": 2.6257, "theoretical_loss": 3.4759364148764575, "tokens_seen": 1711472640 }, { "epoch": 0.04, "learning_rate": 9.723960840956509e-05, "loss": 2.7139, "theoretical_loss": 3.4759135347813768, "tokens_seen": 1711603712 }, { "epoch": 0.04, "learning_rate": 9.723158401540684e-05, "loss": 2.6359, "theoretical_loss": 3.4758906569289034, "tokens_seen": 1711734784 }, { "epoch": 0.04, "learning_rate": 9.722355962124861e-05, "loss": 2.6534, "theoretical_loss": 3.475867781318646, "tokens_seen": 1711865856 }, { "epoch": 0.04, "learning_rate": 9.721553522709036e-05, "loss": 2.6935, "theoretical_loss": 3.4758449079502123, "tokens_seen": 1711996928 }, { "epoch": 0.04, "learning_rate": 9.720751083293211e-05, "loss": 2.6434, "theoretical_loss": 3.475822036823212, "tokens_seen": 1712128000 }, { "epoch": 0.04, "objective/train/docs_used": 944514, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2325708866119385, "objective/train/theoretical_loss": 3.4757991679372533, "objective/train/tokens_used": 82718176, "theoretical_loss": 3.4757991679372533, "tokens_seen": 1712259072 }, { "epoch": 0.04, "learning_rate": 9.719948643877388e-05, "loss": 2.3, "theoretical_loss": 3.4757991679372533, "tokens_seen": 1712259072 }, { "epoch": 0.04, "learning_rate": 9.719146204461563e-05, "loss": 2.7386, "theoretical_loss": 3.4757763012919454, "tokens_seen": 1712390144 }, { "epoch": 0.04, "learning_rate": 9.71834376504574e-05, "loss": 2.6512, "theoretical_loss": 3.4757534368868974, "tokens_seen": 1712521216 }, { "epoch": 0.04, "learning_rate": 9.717541325629915e-05, "loss": 2.5087, "theoretical_loss": 3.4757305747217178, "tokens_seen": 1712652288 }, { "epoch": 0.04, "learning_rate": 9.716738886214092e-05, "loss": 2.5342, "theoretical_loss": 3.475707714796016, "tokens_seen": 1712783360 }, { "epoch": 0.04, "learning_rate": 9.715936446798267e-05, "loss": 2.4806, "theoretical_loss": 3.4756848571094014, "tokens_seen": 1712914432 }, { "epoch": 0.04, "learning_rate": 9.715134007382443e-05, "loss": 2.6971, "theoretical_loss": 3.475662001661483, "tokens_seen": 1713045504 }, { "epoch": 0.04, "learning_rate": 9.714331567966619e-05, "loss": 2.6806, "theoretical_loss": 3.4756391484518714, "tokens_seen": 1713176576 }, { "epoch": 0.04, "learning_rate": 9.713529128550795e-05, "loss": 2.554, "theoretical_loss": 3.475616297480175, "tokens_seen": 1713307648 }, { "epoch": 0.04, "learning_rate": 9.712726689134971e-05, "loss": 2.579, "theoretical_loss": 3.4755934487460047, "tokens_seen": 1713438720 }, { "epoch": 0.04, "learning_rate": 9.711924249719147e-05, "loss": 2.6336, "theoretical_loss": 3.4755706022489687, "tokens_seen": 1713569792 }, { "epoch": 0.04, "learning_rate": 9.711121810303323e-05, "loss": 2.4975, "theoretical_loss": 3.4755477579886778, "tokens_seen": 1713700864 }, { "epoch": 0.04, "learning_rate": 9.710319370887499e-05, "loss": 2.5121, "theoretical_loss": 3.4755249159647414, "tokens_seen": 1713831936 }, { "epoch": 0.04, "objective/train/docs_used": 945657, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.636183261871338, "objective/train/theoretical_loss": 3.4755134957912848, "objective/train/tokens_used": 84356576, "theoretical_loss": 3.4755134957912848, "tokens_seen": 1713897472 }, { "epoch": 0.04, "learning_rate": 9.709516931471674e-05, "loss": 2.5273, "theoretical_loss": 3.4755020761767703, "tokens_seen": 1713963008 }, { "epoch": 0.04, "learning_rate": 9.70871449205585e-05, "loss": 2.4317, "theoretical_loss": 3.4754792386243745, "tokens_seen": 1714094080 }, { "epoch": 0.04, "learning_rate": 9.707912052640026e-05, "loss": 2.4492, "theoretical_loss": 3.475456403307164, "tokens_seen": 1714225152 }, { "epoch": 0.04, "learning_rate": 9.707109613224202e-05, "loss": 2.6308, "theoretical_loss": 3.475433570224749, "tokens_seen": 1714356224 }, { "epoch": 0.04, "learning_rate": 9.706307173808378e-05, "loss": 2.5065, "theoretical_loss": 3.47541073937674, "tokens_seen": 1714487296 }, { "epoch": 0.04, "learning_rate": 9.705504734392554e-05, "loss": 2.4417, "theoretical_loss": 3.4753879107627483, "tokens_seen": 1714618368 }, { "epoch": 0.04, "learning_rate": 9.70470229497673e-05, "loss": 2.5615, "theoretical_loss": 3.4753650843823833, "tokens_seen": 1714749440 }, { "epoch": 0.04, "learning_rate": 9.703899855560905e-05, "loss": 2.5363, "theoretical_loss": 3.475342260235257, "tokens_seen": 1714880512 }, { "epoch": 0.04, "learning_rate": 9.703097416145082e-05, "loss": 2.6402, "theoretical_loss": 3.4753194383209793, "tokens_seen": 1715011584 }, { "epoch": 0.04, "learning_rate": 9.702294976729257e-05, "loss": 2.6409, "theoretical_loss": 3.4752966186391614, "tokens_seen": 1715142656 }, { "epoch": 0.04, "learning_rate": 9.701492537313434e-05, "loss": 2.5318, "theoretical_loss": 3.4752738011894144, "tokens_seen": 1715273728 }, { "epoch": 0.04, "learning_rate": 9.700690097897609e-05, "loss": 2.5512, "theoretical_loss": 3.47525098597135, "tokens_seen": 1715404800 }, { "epoch": 0.04, "objective/train/docs_used": 946326, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3408730030059814, "objective/train/theoretical_loss": 3.4752281729845786, "objective/train/tokens_used": 85994976, "theoretical_loss": 3.4752281729845786, "tokens_seen": 1715535872 }, { "epoch": 0.04, "learning_rate": 9.699887658481786e-05, "loss": 2.4938, "theoretical_loss": 3.4752281729845786, "tokens_seen": 1715535872 }, { "epoch": 0.04, "learning_rate": 9.699085219065961e-05, "loss": 2.5903, "theoretical_loss": 3.475205362228712, "tokens_seen": 1715666944 }, { "epoch": 0.04, "learning_rate": 9.698282779650138e-05, "loss": 2.6315, "theoretical_loss": 3.4751825537033607, "tokens_seen": 1715798016 }, { "epoch": 0.04, "learning_rate": 9.697480340234313e-05, "loss": 2.5496, "theoretical_loss": 3.4751597474081377, "tokens_seen": 1715929088 }, { "epoch": 0.04, "learning_rate": 9.696677900818488e-05, "loss": 2.5582, "theoretical_loss": 3.4751369433426538, "tokens_seen": 1716060160 }, { "epoch": 0.04, "learning_rate": 9.695875461402665e-05, "loss": 2.4528, "theoretical_loss": 3.4751141415065208, "tokens_seen": 1716191232 }, { "epoch": 0.04, "learning_rate": 9.69507302198684e-05, "loss": 2.4352, "theoretical_loss": 3.4750913418993505, "tokens_seen": 1716322304 }, { "epoch": 0.04, "learning_rate": 9.694270582571017e-05, "loss": 2.4735, "theoretical_loss": 3.475068544520755, "tokens_seen": 1716453376 }, { "epoch": 0.04, "learning_rate": 9.693468143155192e-05, "loss": 2.6383, "theoretical_loss": 3.475045749370346, "tokens_seen": 1716584448 }, { "epoch": 0.04, "learning_rate": 9.692665703739369e-05, "loss": 2.6037, "theoretical_loss": 3.475022956447736, "tokens_seen": 1716715520 }, { "epoch": 0.04, "learning_rate": 9.691863264323544e-05, "loss": 2.5211, "theoretical_loss": 3.4750001657525367, "tokens_seen": 1716846592 }, { "epoch": 0.04, "learning_rate": 9.69106082490772e-05, "loss": 2.5279, "theoretical_loss": 3.474977377284361, "tokens_seen": 1716977664 }, { "epoch": 0.04, "learning_rate": 9.690258385491896e-05, "loss": 2.471, "theoretical_loss": 3.4749545910428212, "tokens_seen": 1717108736 }, { "epoch": 0.04, "objective/train/docs_used": 947520, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7320938110351562, "objective/train/theoretical_loss": 3.474943198756918, "objective/train/tokens_used": 87633376, "theoretical_loss": 3.474943198756918, "tokens_seen": 1717174272 }, { "epoch": 0.04, "learning_rate": 9.689455946076072e-05, "loss": 2.5489, "theoretical_loss": 3.474931807027529, "tokens_seen": 1717239808 }, { "epoch": 0.04, "learning_rate": 9.688653506660248e-05, "loss": 2.3712, "theoretical_loss": 3.474909025238098, "tokens_seen": 1717370880 }, { "epoch": 0.04, "learning_rate": 9.687851067244424e-05, "loss": 2.5661, "theoretical_loss": 3.4748862456741403, "tokens_seen": 1717501952 }, { "epoch": 0.04, "learning_rate": 9.6870486278286e-05, "loss": 2.6223, "theoretical_loss": 3.474863468335269, "tokens_seen": 1717633024 }, { "epoch": 0.04, "learning_rate": 9.686246188412776e-05, "loss": 2.6434, "theoretical_loss": 3.474840693221096, "tokens_seen": 1717764096 }, { "epoch": 0.04, "learning_rate": 9.685443748996951e-05, "loss": 2.6146, "theoretical_loss": 3.474817920331236, "tokens_seen": 1717895168 }, { "epoch": 0.04, "learning_rate": 9.684641309581128e-05, "loss": 2.5673, "theoretical_loss": 3.474795149665301, "tokens_seen": 1718026240 }, { "epoch": 0.04, "learning_rate": 9.683838870165303e-05, "loss": 2.5127, "theoretical_loss": 3.4747723812229045, "tokens_seen": 1718157312 }, { "epoch": 0.04, "learning_rate": 9.68303643074948e-05, "loss": 2.5503, "theoretical_loss": 3.4747496150036596, "tokens_seen": 1718288384 }, { "epoch": 0.04, "learning_rate": 9.682233991333655e-05, "loss": 2.6373, "theoretical_loss": 3.47472685100718, "tokens_seen": 1718419456 }, { "epoch": 0.04, "learning_rate": 9.681431551917831e-05, "loss": 2.5278, "theoretical_loss": 3.4747040892330787, "tokens_seen": 1718550528 }, { "epoch": 0.04, "learning_rate": 9.680629112502007e-05, "loss": 2.5697, "theoretical_loss": 3.474681329680969, "tokens_seen": 1718681600 }, { "epoch": 0.04, "objective/train/docs_used": 948221, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.1624414920806885, "objective/train/theoretical_loss": 3.4746585723504655, "objective/train/tokens_used": 89271776, "theoretical_loss": 3.4746585723504655, "tokens_seen": 1718812672 }, { "epoch": 0.04, "learning_rate": 9.679826673086182e-05, "loss": 2.4717, "theoretical_loss": 3.4746585723504655, "tokens_seen": 1718812672 }, { "epoch": 0.04, "learning_rate": 9.679024233670359e-05, "loss": 2.5336, "theoretical_loss": 3.4746358172411815, "tokens_seen": 1718943744 }, { "epoch": 0.04, "learning_rate": 9.678221794254534e-05, "loss": 2.5264, "theoretical_loss": 3.4746130643527304, "tokens_seen": 1719074816 }, { "epoch": 0.04, "learning_rate": 9.677419354838711e-05, "loss": 2.598, "theoretical_loss": 3.474590313684727, "tokens_seen": 1719205888 }, { "epoch": 0.04, "learning_rate": 9.676616915422886e-05, "loss": 2.492, "theoretical_loss": 3.474567565236785, "tokens_seen": 1719336960 }, { "epoch": 0.04, "learning_rate": 9.675814476007063e-05, "loss": 2.4323, "theoretical_loss": 3.4745448190085177, "tokens_seen": 1719468032 }, { "epoch": 0.04, "learning_rate": 9.675012036591238e-05, "loss": 2.5456, "theoretical_loss": 3.4745220749995407, "tokens_seen": 1719599104 }, { "epoch": 0.04, "learning_rate": 9.674209597175413e-05, "loss": 2.4039, "theoretical_loss": 3.474499333209468, "tokens_seen": 1719730176 }, { "epoch": 0.04, "learning_rate": 9.67340715775959e-05, "loss": 2.4198, "theoretical_loss": 3.474476593637913, "tokens_seen": 1719861248 }, { "epoch": 0.04, "learning_rate": 9.672604718343765e-05, "loss": 2.6933, "theoretical_loss": 3.4744538562844913, "tokens_seen": 1719992320 }, { "epoch": 0.04, "learning_rate": 9.671802278927942e-05, "loss": 2.5493, "theoretical_loss": 3.4744311211488172, "tokens_seen": 1720123392 }, { "epoch": 0.04, "learning_rate": 9.670999839512117e-05, "loss": 2.6969, "theoretical_loss": 3.4744083882305055, "tokens_seen": 1720254464 }, { "epoch": 0.04, "learning_rate": 9.670197400096294e-05, "loss": 2.4579, "theoretical_loss": 3.474385657529171, "tokens_seen": 1720385536 }, { "epoch": 0.04, "objective/train/docs_used": 949201, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.1461188793182373, "objective/train/theoretical_loss": 3.47437429300975, "objective/train/tokens_used": 90910176, "theoretical_loss": 3.47437429300975, "tokens_seen": 1720451072 }, { "epoch": 0.04, "learning_rate": 9.669394960680469e-05, "loss": 2.4628, "theoretical_loss": 3.4743629290444282, "tokens_seen": 1720516608 }, { "epoch": 0.04, "learning_rate": 9.668592521264645e-05, "loss": 2.4322, "theoretical_loss": 3.474340202775893, "tokens_seen": 1720647680 }, { "epoch": 0.04, "learning_rate": 9.667790081848821e-05, "loss": 2.5498, "theoretical_loss": 3.47431747872318, "tokens_seen": 1720778752 }, { "epoch": 0.04, "learning_rate": 9.666987642432997e-05, "loss": 2.5226, "theoretical_loss": 3.474294756885904, "tokens_seen": 1720909824 }, { "epoch": 0.04, "learning_rate": 9.666185203017173e-05, "loss": 2.4954, "theoretical_loss": 3.474272037263681, "tokens_seen": 1721040896 }, { "epoch": 0.04, "learning_rate": 9.665382763601349e-05, "loss": 2.6002, "theoretical_loss": 3.474249319856126, "tokens_seen": 1721171968 }, { "epoch": 0.04, "learning_rate": 9.664580324185525e-05, "loss": 2.4917, "theoretical_loss": 3.4742266046628543, "tokens_seen": 1721303040 }, { "epoch": 0.04, "learning_rate": 9.6637778847697e-05, "loss": 2.479, "theoretical_loss": 3.474203891683482, "tokens_seen": 1721434112 }, { "epoch": 0.04, "learning_rate": 9.662975445353877e-05, "loss": 2.479, "theoretical_loss": 3.474181180917624, "tokens_seen": 1721565184 }, { "epoch": 0.04, "learning_rate": 9.662173005938053e-05, "loss": 2.717, "theoretical_loss": 3.4741584723648975, "tokens_seen": 1721696256 }, { "epoch": 0.04, "learning_rate": 9.661370566522228e-05, "loss": 2.6332, "theoretical_loss": 3.474135766024917, "tokens_seen": 1721827328 }, { "epoch": 0.04, "learning_rate": 9.660568127106405e-05, "loss": 2.6836, "theoretical_loss": 3.474113061897299, "tokens_seen": 1721958400 }, { "epoch": 0.04, "objective/train/docs_used": 949909, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.497939348220825, "objective/train/theoretical_loss": 3.474090359981659, "objective/train/tokens_used": 92548576, "theoretical_loss": 3.474090359981659, "tokens_seen": 1722089472 }, { "epoch": 0.04, "learning_rate": 9.65976568769058e-05, "loss": 2.5073, "theoretical_loss": 3.474090359981659, "tokens_seen": 1722089472 }, { "epoch": 0.04, "learning_rate": 9.658963248274757e-05, "loss": 2.4187, "theoretical_loss": 3.4740676602776146, "tokens_seen": 1722220544 }, { "epoch": 0.04, "learning_rate": 9.658160808858932e-05, "loss": 2.5354, "theoretical_loss": 3.4740449627847805, "tokens_seen": 1722351616 }, { "epoch": 0.04, "learning_rate": 9.657358369443108e-05, "loss": 2.428, "theoretical_loss": 3.4740222675027734, "tokens_seen": 1722482688 }, { "epoch": 0.04, "learning_rate": 9.656555930027284e-05, "loss": 2.5802, "theoretical_loss": 3.4739995744312107, "tokens_seen": 1722613760 }, { "epoch": 0.04, "learning_rate": 9.655753490611459e-05, "loss": 2.6047, "theoretical_loss": 3.473976883569708, "tokens_seen": 1722744832 }, { "epoch": 0.04, "learning_rate": 9.654951051195636e-05, "loss": 2.6242, "theoretical_loss": 3.4739541949178827, "tokens_seen": 1722875904 }, { "epoch": 0.04, "learning_rate": 9.654148611779811e-05, "loss": 2.4371, "theoretical_loss": 3.47393150847535, "tokens_seen": 1723006976 }, { "epoch": 0.04, "learning_rate": 9.653346172363988e-05, "loss": 2.3743, "theoretical_loss": 3.473908824241729, "tokens_seen": 1723138048 }, { "epoch": 0.04, "learning_rate": 9.652543732948163e-05, "loss": 2.7518, "theoretical_loss": 3.4738861422166343, "tokens_seen": 1723269120 }, { "epoch": 0.04, "learning_rate": 9.65174129353234e-05, "loss": 2.44, "theoretical_loss": 3.4738634623996845, "tokens_seen": 1723400192 }, { "epoch": 0.04, "learning_rate": 9.650938854116515e-05, "loss": 2.5699, "theoretical_loss": 3.4738407847904966, "tokens_seen": 1723531264 }, { "epoch": 0.04, "learning_rate": 9.65013641470069e-05, "loss": 2.4851, "theoretical_loss": 3.473818109388687, "tokens_seen": 1723662336 }, { "epoch": 0.04, "objective/train/docs_used": 951235, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2583699226379395, "objective/train/theoretical_loss": 3.47380677251543, "objective/train/tokens_used": 94186976, "theoretical_loss": 3.47380677251543, "tokens_seen": 1723727872 }, { "epoch": 0.04, "learning_rate": 9.649333975284867e-05, "loss": 2.5272, "theoretical_loss": 3.473795436193874, "tokens_seen": 1723793408 }, { "epoch": 0.04, "learning_rate": 9.648531535869042e-05, "loss": 2.5135, "theoretical_loss": 3.4737727652056742, "tokens_seen": 1723924480 }, { "epoch": 0.04, "learning_rate": 9.647729096453219e-05, "loss": 2.6295, "theoretical_loss": 3.473750096423705, "tokens_seen": 1724055552 }, { "epoch": 0.04, "learning_rate": 9.646926657037394e-05, "loss": 2.6414, "theoretical_loss": 3.4737274298475844, "tokens_seen": 1724186624 }, { "epoch": 0.05, "learning_rate": 9.646124217621571e-05, "loss": 2.5256, "theoretical_loss": 3.4737047654769304, "tokens_seen": 1724317696 }, { "epoch": 0.05, "learning_rate": 9.645321778205746e-05, "loss": 2.664, "theoretical_loss": 3.4736821033113605, "tokens_seen": 1724448768 }, { "epoch": 0.05, "learning_rate": 9.644519338789922e-05, "loss": 2.5327, "theoretical_loss": 3.473659443350493, "tokens_seen": 1724579840 }, { "epoch": 0.05, "learning_rate": 9.643716899374098e-05, "loss": 2.6298, "theoretical_loss": 3.4736367855939445, "tokens_seen": 1724710912 }, { "epoch": 0.05, "learning_rate": 9.642914459958274e-05, "loss": 2.5402, "theoretical_loss": 3.4736141300413337, "tokens_seen": 1724841984 }, { "epoch": 0.05, "learning_rate": 9.64211202054245e-05, "loss": 2.4692, "theoretical_loss": 3.47359147669228, "tokens_seen": 1724973056 }, { "epoch": 0.05, "learning_rate": 9.641309581126626e-05, "loss": 2.6825, "theoretical_loss": 3.4735688255464003, "tokens_seen": 1725104128 }, { "epoch": 0.05, "learning_rate": 9.640507141710802e-05, "loss": 2.4039, "theoretical_loss": 3.4735461766033136, "tokens_seen": 1725235200 }, { "epoch": 0.05, "objective/train/docs_used": 951881, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6248931884765625, "objective/train/theoretical_loss": 3.4735235298626375, "objective/train/tokens_used": 95825376, "theoretical_loss": 3.4735235298626375, "tokens_seen": 1725366272 }, { "epoch": 0.05, "learning_rate": 9.639704702294978e-05, "loss": 2.55, "theoretical_loss": 3.4735235298626375, "tokens_seen": 1725366272 }, { "epoch": 0.05, "learning_rate": 9.638902262879153e-05, "loss": 2.3494, "theoretical_loss": 3.473500885323992, "tokens_seen": 1725497344 }, { "epoch": 0.05, "learning_rate": 9.63809982346333e-05, "loss": 2.6021, "theoretical_loss": 3.473478242986994, "tokens_seen": 1725628416 }, { "epoch": 0.05, "learning_rate": 9.637297384047505e-05, "loss": 2.4924, "theoretical_loss": 3.473455602851264, "tokens_seen": 1725759488 }, { "epoch": 0.05, "learning_rate": 9.636494944631682e-05, "loss": 2.4347, "theoretical_loss": 3.4734329649164195, "tokens_seen": 1725890560 }, { "epoch": 0.05, "learning_rate": 9.635692505215857e-05, "loss": 2.3637, "theoretical_loss": 3.47341032918208, "tokens_seen": 1726021632 }, { "epoch": 0.05, "learning_rate": 9.634890065800034e-05, "loss": 2.5177, "theoretical_loss": 3.4733876956478644, "tokens_seen": 1726152704 }, { "epoch": 0.05, "learning_rate": 9.634087626384209e-05, "loss": 2.5704, "theoretical_loss": 3.473365064313392, "tokens_seen": 1726283776 }, { "epoch": 0.05, "learning_rate": 9.633285186968385e-05, "loss": 2.6099, "theoretical_loss": 3.4733424351782816, "tokens_seen": 1726414848 }, { "epoch": 0.05, "learning_rate": 9.632482747552561e-05, "loss": 2.6702, "theoretical_loss": 3.4733198082421533, "tokens_seen": 1726545920 }, { "epoch": 0.05, "learning_rate": 9.631680308136736e-05, "loss": 2.5298, "theoretical_loss": 3.4732971835046254, "tokens_seen": 1726676992 }, { "epoch": 0.05, "learning_rate": 9.630877868720913e-05, "loss": 2.4978, "theoretical_loss": 3.4732745609653177, "tokens_seen": 1726808064 }, { "epoch": 0.05, "learning_rate": 9.630075429305088e-05, "loss": 2.4702, "theoretical_loss": 3.4732519406238507, "tokens_seen": 1726939136 }, { "epoch": 0.05, "objective/train/docs_used": 953135, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.060582160949707, "objective/train/theoretical_loss": 3.4732406312771884, "objective/train/tokens_used": 97463776, "theoretical_loss": 3.4732406312771884, "tokens_seen": 1727004672 }, { "epoch": 0.05, "learning_rate": 9.629272989889265e-05, "loss": 2.5876, "theoretical_loss": 3.473229322479843, "tokens_seen": 1727070208 }, { "epoch": 0.05, "learning_rate": 9.62847055047344e-05, "loss": 2.6043, "theoretical_loss": 3.473206706532915, "tokens_seen": 1727201280 }, { "epoch": 0.05, "learning_rate": 9.627668111057617e-05, "loss": 2.5915, "theoretical_loss": 3.4731840927826867, "tokens_seen": 1727332352 }, { "epoch": 0.05, "learning_rate": 9.626865671641792e-05, "loss": 2.4664, "theoretical_loss": 3.4731614812287774, "tokens_seen": 1727463424 }, { "epoch": 0.05, "learning_rate": 9.626063232225967e-05, "loss": 2.5555, "theoretical_loss": 3.4731388718708076, "tokens_seen": 1727594496 }, { "epoch": 0.05, "learning_rate": 9.625260792810144e-05, "loss": 2.5768, "theoretical_loss": 3.4731162647083975, "tokens_seen": 1727725568 }, { "epoch": 0.05, "learning_rate": 9.62445835339432e-05, "loss": 2.6285, "theoretical_loss": 3.473093659741167, "tokens_seen": 1727856640 }, { "epoch": 0.05, "learning_rate": 9.623655913978496e-05, "loss": 2.6888, "theoretical_loss": 3.473071056968737, "tokens_seen": 1727987712 }, { "epoch": 0.05, "learning_rate": 9.622853474562671e-05, "loss": 2.6184, "theoretical_loss": 3.4730484563907273, "tokens_seen": 1728118784 }, { "epoch": 0.05, "learning_rate": 9.622051035146848e-05, "loss": 2.6123, "theoretical_loss": 3.473025858006759, "tokens_seen": 1728249856 }, { "epoch": 0.05, "learning_rate": 9.621248595731023e-05, "loss": 2.4683, "theoretical_loss": 3.4730032618164524, "tokens_seen": 1728380928 }, { "epoch": 0.05, "learning_rate": 9.620446156315199e-05, "loss": 2.4986, "theoretical_loss": 3.472980667819428, "tokens_seen": 1728512000 }, { "epoch": 0.05, "objective/train/docs_used": 953685, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4338219165802, "objective/train/theoretical_loss": 3.472958076015307, "objective/train/tokens_used": 99102176, "theoretical_loss": 3.472958076015307, "tokens_seen": 1728643072 }, { "epoch": 0.05, "learning_rate": 9.619643716899375e-05, "loss": 2.6793, "theoretical_loss": 3.472958076015307, "tokens_seen": 1728643072 }, { "epoch": 0.05, "learning_rate": 9.61884127748355e-05, "loss": 2.3993, "theoretical_loss": 3.4729354864037107, "tokens_seen": 1728774144 }, { "epoch": 0.05, "learning_rate": 9.618038838067727e-05, "loss": 2.5727, "theoretical_loss": 3.4729128989842595, "tokens_seen": 1728905216 }, { "epoch": 0.05, "learning_rate": 9.617236398651903e-05, "loss": 2.5781, "theoretical_loss": 3.4728903137565745, "tokens_seen": 1729036288 }, { "epoch": 0.05, "learning_rate": 9.616433959236079e-05, "loss": 2.5062, "theoretical_loss": 3.472867730720277, "tokens_seen": 1729167360 }, { "epoch": 0.05, "learning_rate": 9.615631519820255e-05, "loss": 2.6331, "theoretical_loss": 3.4728451498749884, "tokens_seen": 1729298432 }, { "epoch": 0.05, "learning_rate": 9.61482908040443e-05, "loss": 2.6771, "theoretical_loss": 3.4728225712203304, "tokens_seen": 1729429504 }, { "epoch": 0.05, "learning_rate": 9.614026640988607e-05, "loss": 2.562, "theoretical_loss": 3.472799994755924, "tokens_seen": 1729560576 }, { "epoch": 0.05, "learning_rate": 9.613224201572782e-05, "loss": 2.5274, "theoretical_loss": 3.4727774204813904, "tokens_seen": 1729691648 }, { "epoch": 0.05, "learning_rate": 9.612421762156959e-05, "loss": 2.4391, "theoretical_loss": 3.472754848396352, "tokens_seen": 1729822720 }, { "epoch": 0.05, "learning_rate": 9.611619322741134e-05, "loss": 2.5883, "theoretical_loss": 3.4727322785004304, "tokens_seen": 1729953792 }, { "epoch": 0.05, "learning_rate": 9.610816883325309e-05, "loss": 2.7006, "theoretical_loss": 3.4727097107932474, "tokens_seen": 1730084864 }, { "epoch": 0.05, "learning_rate": 9.610014443909486e-05, "loss": 2.5706, "theoretical_loss": 3.472687145274425, "tokens_seen": 1730215936 }, { "epoch": 0.05, "objective/train/docs_used": 954893, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3295490741729736, "objective/train/theoretical_loss": 3.472675863335531, "objective/train/tokens_used": 100740576, "theoretical_loss": 3.472675863335531, "tokens_seen": 1730281472 }, { "epoch": 0.05, "learning_rate": 9.609212004493661e-05, "loss": 2.5245, "theoretical_loss": 3.472664581943585, "tokens_seen": 1730347008 }, { "epoch": 0.05, "learning_rate": 9.608409565077838e-05, "loss": 2.5552, "theoretical_loss": 3.4726420208003494, "tokens_seen": 1730478080 }, { "epoch": 0.05, "learning_rate": 9.607607125662013e-05, "loss": 2.5185, "theoretical_loss": 3.4726194618443413, "tokens_seen": 1730609152 }, { "epoch": 0.05, "learning_rate": 9.60680468624619e-05, "loss": 2.53, "theoretical_loss": 3.472596905075182, "tokens_seen": 1730740224 }, { "epoch": 0.05, "learning_rate": 9.606002246830365e-05, "loss": 2.6368, "theoretical_loss": 3.472574350492495, "tokens_seen": 1730871296 }, { "epoch": 0.05, "learning_rate": 9.60519980741454e-05, "loss": 2.4409, "theoretical_loss": 3.4725517980959015, "tokens_seen": 1731002368 }, { "epoch": 0.05, "learning_rate": 9.604397367998717e-05, "loss": 2.6554, "theoretical_loss": 3.472529247885025, "tokens_seen": 1731133440 }, { "epoch": 0.05, "learning_rate": 9.603594928582892e-05, "loss": 2.6305, "theoretical_loss": 3.4725066998594882, "tokens_seen": 1731264512 }, { "epoch": 0.05, "learning_rate": 9.602792489167069e-05, "loss": 2.5356, "theoretical_loss": 3.4724841540189137, "tokens_seen": 1731395584 }, { "epoch": 0.05, "learning_rate": 9.601990049751244e-05, "loss": 2.5259, "theoretical_loss": 3.4724616103629242, "tokens_seen": 1731526656 }, { "epoch": 0.05, "learning_rate": 9.60118761033542e-05, "loss": 2.5755, "theoretical_loss": 3.472439068891143, "tokens_seen": 1731657728 }, { "epoch": 0.05, "learning_rate": 9.600385170919596e-05, "loss": 2.6055, "theoretical_loss": 3.4724165296031924, "tokens_seen": 1731788800 }, { "epoch": 0.05, "objective/train/docs_used": 956184, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5680079460144043, "objective/train/theoretical_loss": 3.472393992498697, "objective/train/tokens_used": 102378976, "theoretical_loss": 3.472393992498697, "tokens_seen": 1731919872 }, { "epoch": 0.05, "learning_rate": 9.599582731503772e-05, "loss": 2.6725, "theoretical_loss": 3.472393992498697, "tokens_seen": 1731919872 }, { "epoch": 0.05, "learning_rate": 9.598780292087948e-05, "loss": 2.5937, "theoretical_loss": 3.472371457577279, "tokens_seen": 1732050944 }, { "epoch": 0.05, "learning_rate": 9.597977852672124e-05, "loss": 2.613, "theoretical_loss": 3.472348924838562, "tokens_seen": 1732182016 }, { "epoch": 0.05, "learning_rate": 9.5971754132563e-05, "loss": 2.5003, "theoretical_loss": 3.4723263942821694, "tokens_seen": 1732313088 }, { "epoch": 0.05, "learning_rate": 9.596372973840476e-05, "loss": 2.5422, "theoretical_loss": 3.4723038659077248, "tokens_seen": 1732444160 }, { "epoch": 0.05, "learning_rate": 9.595570534424651e-05, "loss": 2.5334, "theoretical_loss": 3.4722813397148515, "tokens_seen": 1732575232 }, { "epoch": 0.05, "learning_rate": 9.594768095008828e-05, "loss": 2.4992, "theoretical_loss": 3.4722588157031744, "tokens_seen": 1732706304 }, { "epoch": 0.05, "learning_rate": 9.593965655593003e-05, "loss": 2.5993, "theoretical_loss": 3.4722362938723155, "tokens_seen": 1732837376 }, { "epoch": 0.05, "learning_rate": 9.59316321617718e-05, "loss": 2.5522, "theoretical_loss": 3.472213774221901, "tokens_seen": 1732968448 }, { "epoch": 0.05, "learning_rate": 9.592360776761355e-05, "loss": 2.5903, "theoretical_loss": 3.472191256751552, "tokens_seen": 1733099520 }, { "epoch": 0.05, "learning_rate": 9.59155833734553e-05, "loss": 2.55, "theoretical_loss": 3.4721687414608953, "tokens_seen": 1733230592 }, { "epoch": 0.05, "learning_rate": 9.590755897929707e-05, "loss": 2.6655, "theoretical_loss": 3.472146228349554, "tokens_seen": 1733361664 }, { "epoch": 0.05, "learning_rate": 9.589953458513882e-05, "loss": 2.5992, "theoretical_loss": 3.472123717417152, "tokens_seen": 1733492736 }, { "epoch": 0.05, "objective/train/docs_used": 956719, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.863294839859009, "objective/train/theoretical_loss": 3.472112462767936, "objective/train/tokens_used": 104017376, "theoretical_loss": 3.472112462767936, "tokens_seen": 1733558272 }, { "epoch": 0.05, "learning_rate": 9.589151019098059e-05, "loss": 2.592, "theoretical_loss": 3.4721012086633145, "tokens_seen": 1733623808 }, { "epoch": 0.05, "learning_rate": 9.588348579682234e-05, "loss": 2.6002, "theoretical_loss": 3.472078702087665, "tokens_seen": 1733754880 }, { "epoch": 0.05, "learning_rate": 9.587546140266411e-05, "loss": 2.396, "theoretical_loss": 3.4720561976898288, "tokens_seen": 1733885952 }, { "epoch": 0.05, "learning_rate": 9.586743700850586e-05, "loss": 2.5257, "theoretical_loss": 3.47203369546943, "tokens_seen": 1734017024 }, { "epoch": 0.05, "learning_rate": 9.585941261434761e-05, "loss": 2.2733, "theoretical_loss": 3.472011195426095, "tokens_seen": 1734148096 }, { "epoch": 0.05, "learning_rate": 9.585138822018938e-05, "loss": 2.5829, "theoretical_loss": 3.4719886975594463, "tokens_seen": 1734279168 }, { "epoch": 0.05, "learning_rate": 9.584336382603113e-05, "loss": 2.5333, "theoretical_loss": 3.47196620186911, "tokens_seen": 1734410240 }, { "epoch": 0.05, "learning_rate": 9.58353394318729e-05, "loss": 2.6064, "theoretical_loss": 3.4719437083547113, "tokens_seen": 1734541312 }, { "epoch": 0.05, "learning_rate": 9.582731503771465e-05, "loss": 2.5828, "theoretical_loss": 3.4719212170158755, "tokens_seen": 1734672384 }, { "epoch": 0.05, "learning_rate": 9.581929064355641e-05, "loss": 2.5385, "theoretical_loss": 3.471898727852227, "tokens_seen": 1734803456 }, { "epoch": 0.05, "learning_rate": 9.581126624939817e-05, "loss": 2.5934, "theoretical_loss": 3.4718762408633914, "tokens_seen": 1734934528 }, { "epoch": 0.05, "learning_rate": 9.580324185523993e-05, "loss": 2.6143, "theoretical_loss": 3.471853756048994, "tokens_seen": 1735065600 }, { "epoch": 0.05, "objective/train/docs_used": 957813, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.456376314163208, "objective/train/theoretical_loss": 3.471831273408661, "objective/train/tokens_used": 105655776, "theoretical_loss": 3.471831273408661, "tokens_seen": 1735196672 }, { "epoch": 0.05, "learning_rate": 9.57952174610817e-05, "loss": 2.5648, "theoretical_loss": 3.471831273408661, "tokens_seen": 1735196672 }, { "epoch": 0.05, "learning_rate": 9.578719306692345e-05, "loss": 2.6123, "theoretical_loss": 3.471808792942017, "tokens_seen": 1735327744 }, { "epoch": 0.05, "learning_rate": 9.57791686727652e-05, "loss": 2.5439, "theoretical_loss": 3.4717863146486887, "tokens_seen": 1735458816 }, { "epoch": 0.05, "learning_rate": 9.577114427860697e-05, "loss": 2.5435, "theoretical_loss": 3.471763838528301, "tokens_seen": 1735589888 }, { "epoch": 0.05, "learning_rate": 9.576311988444872e-05, "loss": 2.6144, "theoretical_loss": 3.47174136458048, "tokens_seen": 1735720960 }, { "epoch": 0.05, "learning_rate": 9.575509549029049e-05, "loss": 2.6676, "theoretical_loss": 3.4717188928048524, "tokens_seen": 1735852032 }, { "epoch": 0.05, "learning_rate": 9.574707109613224e-05, "loss": 2.6404, "theoretical_loss": 3.471696423201043, "tokens_seen": 1735983104 }, { "epoch": 0.05, "learning_rate": 9.5739046701974e-05, "loss": 2.42, "theoretical_loss": 3.4716739557686793, "tokens_seen": 1736114176 }, { "epoch": 0.05, "learning_rate": 9.573102230781576e-05, "loss": 2.5927, "theoretical_loss": 3.4716514905073863, "tokens_seen": 1736245248 }, { "epoch": 0.05, "learning_rate": 9.572299791365751e-05, "loss": 2.5552, "theoretical_loss": 3.4716290274167907, "tokens_seen": 1736376320 }, { "epoch": 0.05, "learning_rate": 9.571497351949928e-05, "loss": 2.5069, "theoretical_loss": 3.4716065664965194, "tokens_seen": 1736507392 }, { "epoch": 0.05, "learning_rate": 9.570694912534103e-05, "loss": 2.6358, "theoretical_loss": 3.4715841077461986, "tokens_seen": 1736638464 }, { "epoch": 0.05, "learning_rate": 9.56989247311828e-05, "loss": 2.5779, "theoretical_loss": 3.4715616511654552, "tokens_seen": 1736769536 }, { "epoch": 0.05, "objective/train/docs_used": 958321, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.462451934814453, "objective/train/theoretical_loss": 3.4715504236885577, "objective/train/tokens_used": 107294176, "theoretical_loss": 3.4715504236885577, "tokens_seen": 1736835072 }, { "epoch": 0.05, "learning_rate": 9.569090033702455e-05, "loss": 2.5436, "theoretical_loss": 3.4715391967539153, "tokens_seen": 1736900608 }, { "epoch": 0.05, "learning_rate": 9.568287594286632e-05, "loss": 2.7413, "theoretical_loss": 3.471516744511206, "tokens_seen": 1737031680 }, { "epoch": 0.05, "learning_rate": 9.567485154870807e-05, "loss": 2.4977, "theoretical_loss": 3.471494294436954, "tokens_seen": 1737162752 }, { "epoch": 0.05, "learning_rate": 9.566682715454983e-05, "loss": 2.4846, "theoretical_loss": 3.471471846530787, "tokens_seen": 1737293824 }, { "epoch": 0.05, "learning_rate": 9.565880276039159e-05, "loss": 2.5388, "theoretical_loss": 3.4714494007923307, "tokens_seen": 1737424896 }, { "epoch": 0.05, "learning_rate": 9.565077836623335e-05, "loss": 2.6694, "theoretical_loss": 3.4714269572212135, "tokens_seen": 1737555968 }, { "epoch": 0.05, "learning_rate": 9.564275397207511e-05, "loss": 2.5805, "theoretical_loss": 3.4714045158170626, "tokens_seen": 1737687040 }, { "epoch": 0.05, "learning_rate": 9.563472957791687e-05, "loss": 2.4512, "theoretical_loss": 3.4713820765795047, "tokens_seen": 1737818112 }, { "epoch": 0.05, "learning_rate": 9.562670518375863e-05, "loss": 2.4375, "theoretical_loss": 3.471359639508168, "tokens_seen": 1737949184 }, { "epoch": 0.05, "learning_rate": 9.561868078960038e-05, "loss": 2.5466, "theoretical_loss": 3.471337204602679, "tokens_seen": 1738080256 }, { "epoch": 0.05, "learning_rate": 9.561065639544214e-05, "loss": 2.6156, "theoretical_loss": 3.471314771862666, "tokens_seen": 1738211328 }, { "epoch": 0.05, "learning_rate": 9.56026320012839e-05, "loss": 2.7741, "theoretical_loss": 3.4712923412877563, "tokens_seen": 1738342400 }, { "epoch": 0.05, "objective/train/docs_used": 959398, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5910706520080566, "objective/train/theoretical_loss": 3.4712699128775784, "objective/train/tokens_used": 108932576, "theoretical_loss": 3.4712699128775784, "tokens_seen": 1738473472 }, { "epoch": 0.05, "learning_rate": 9.559460760712566e-05, "loss": 2.5787, "theoretical_loss": 3.4712699128775784, "tokens_seen": 1738473472 }, { "epoch": 0.05, "learning_rate": 9.558658321296742e-05, "loss": 2.7085, "theoretical_loss": 3.4712474866317597, "tokens_seen": 1738604544 }, { "epoch": 0.05, "learning_rate": 9.557855881880918e-05, "loss": 2.5791, "theoretical_loss": 3.4712250625499284, "tokens_seen": 1738735616 }, { "epoch": 0.05, "learning_rate": 9.557053442465094e-05, "loss": 2.5154, "theoretical_loss": 3.471202640631713, "tokens_seen": 1738866688 }, { "epoch": 0.05, "learning_rate": 9.55625100304927e-05, "loss": 2.6328, "theoretical_loss": 3.4711802208767404, "tokens_seen": 1738997760 }, { "epoch": 0.05, "learning_rate": 9.555448563633445e-05, "loss": 2.5779, "theoretical_loss": 3.4711578032846404, "tokens_seen": 1739128832 }, { "epoch": 0.05, "learning_rate": 9.554646124217622e-05, "loss": 2.4909, "theoretical_loss": 3.47113538785504, "tokens_seen": 1739259904 }, { "epoch": 0.05, "learning_rate": 9.553843684801797e-05, "loss": 2.6613, "theoretical_loss": 3.471112974587568, "tokens_seen": 1739390976 }, { "epoch": 0.05, "learning_rate": 9.553041245385974e-05, "loss": 2.647, "theoretical_loss": 3.471090563481854, "tokens_seen": 1739522048 }, { "epoch": 0.05, "learning_rate": 9.552238805970149e-05, "loss": 2.5482, "theoretical_loss": 3.4710681545375253, "tokens_seen": 1739653120 }, { "epoch": 0.05, "learning_rate": 9.551436366554326e-05, "loss": 2.6415, "theoretical_loss": 3.471045747754211, "tokens_seen": 1739784192 }, { "epoch": 0.05, "learning_rate": 9.550633927138501e-05, "loss": 2.5234, "theoretical_loss": 3.4710233431315407, "tokens_seen": 1739915264 }, { "epoch": 0.05, "learning_rate": 9.549831487722676e-05, "loss": 2.4932, "theoretical_loss": 3.471000940669142, "tokens_seen": 1740046336 }, { "epoch": 0.05, "objective/train/docs_used": 960697, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7677383422851562, "objective/train/theoretical_loss": 3.470989740247929, "objective/train/tokens_used": 110570976, "theoretical_loss": 3.470989740247929, "tokens_seen": 1740111872 }, { "epoch": 0.05, "learning_rate": 9.549029048306853e-05, "loss": 2.5452, "theoretical_loss": 3.4709785403666453, "tokens_seen": 1740177408 }, { "epoch": 0.05, "learning_rate": 9.548226608891028e-05, "loss": 2.5708, "theoretical_loss": 3.4709561422236783, "tokens_seen": 1740308480 }, { "epoch": 0.05, "learning_rate": 9.547424169475205e-05, "loss": 2.7149, "theoretical_loss": 3.4709337462398713, "tokens_seen": 1740439552 }, { "epoch": 0.05, "learning_rate": 9.54662173005938e-05, "loss": 2.5015, "theoretical_loss": 3.470911352414853, "tokens_seen": 1740570624 }, { "epoch": 0.05, "learning_rate": 9.545819290643557e-05, "loss": 2.6489, "theoretical_loss": 3.470888960748253, "tokens_seen": 1740701696 }, { "epoch": 0.06, "learning_rate": 9.545016851227732e-05, "loss": 2.6192, "theoretical_loss": 3.4708665712397004, "tokens_seen": 1740832768 }, { "epoch": 0.06, "learning_rate": 9.544214411811908e-05, "loss": 2.6038, "theoretical_loss": 3.4708441838888255, "tokens_seen": 1740963840 }, { "epoch": 0.06, "learning_rate": 9.543411972396084e-05, "loss": 2.6892, "theoretical_loss": 3.4708217986952574, "tokens_seen": 1741094912 }, { "epoch": 0.06, "learning_rate": 9.54260953298026e-05, "loss": 2.5735, "theoretical_loss": 3.4707994156586253, "tokens_seen": 1741225984 }, { "epoch": 0.06, "learning_rate": 9.541807093564436e-05, "loss": 2.6655, "theoretical_loss": 3.47077703477856, "tokens_seen": 1741357056 }, { "epoch": 0.06, "learning_rate": 9.541004654148612e-05, "loss": 2.5846, "theoretical_loss": 3.470754656054691, "tokens_seen": 1741488128 }, { "epoch": 0.06, "learning_rate": 9.540202214732788e-05, "loss": 2.539, "theoretical_loss": 3.470732279486649, "tokens_seen": 1741619200 }, { "epoch": 0.06, "objective/train/docs_used": 961372, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.166609525680542, "objective/train/theoretical_loss": 3.470709905074062, "objective/train/tokens_used": 112209376, "theoretical_loss": 3.470709905074062, "tokens_seen": 1741750272 }, { "epoch": 0.06, "learning_rate": 9.539399775316964e-05, "loss": 2.6481, "theoretical_loss": 3.470709905074062, "tokens_seen": 1741750272 }, { "epoch": 0.06, "learning_rate": 9.53859733590114e-05, "loss": 2.4939, "theoretical_loss": 3.470687532816563, "tokens_seen": 1741881344 }, { "epoch": 0.06, "learning_rate": 9.537794896485315e-05, "loss": 2.488, "theoretical_loss": 3.4706651627137806, "tokens_seen": 1742012416 }, { "epoch": 0.06, "learning_rate": 9.536992457069491e-05, "loss": 2.5908, "theoretical_loss": 3.470642794765345, "tokens_seen": 1742143488 }, { "epoch": 0.06, "learning_rate": 9.536190017653667e-05, "loss": 2.5974, "theoretical_loss": 3.4706204289708875, "tokens_seen": 1742274560 }, { "epoch": 0.06, "learning_rate": 9.535387578237843e-05, "loss": 2.5184, "theoretical_loss": 3.4705980653300386, "tokens_seen": 1742405632 }, { "epoch": 0.06, "learning_rate": 9.53458513882202e-05, "loss": 2.4658, "theoretical_loss": 3.4705757038424285, "tokens_seen": 1742536704 }, { "epoch": 0.06, "learning_rate": 9.533782699406195e-05, "loss": 2.6165, "theoretical_loss": 3.470553344507688, "tokens_seen": 1742667776 }, { "epoch": 0.06, "learning_rate": 9.532980259990371e-05, "loss": 2.5697, "theoretical_loss": 3.470530987325448, "tokens_seen": 1742798848 }, { "epoch": 0.06, "learning_rate": 9.532177820574547e-05, "loss": 2.7383, "theoretical_loss": 3.4705086322953393, "tokens_seen": 1742929920 }, { "epoch": 0.06, "learning_rate": 9.531375381158722e-05, "loss": 2.5499, "theoretical_loss": 3.4704862794169937, "tokens_seen": 1743060992 }, { "epoch": 0.06, "learning_rate": 9.530572941742899e-05, "loss": 2.5394, "theoretical_loss": 3.470463928690041, "tokens_seen": 1743192064 }, { "epoch": 0.06, "learning_rate": 9.529770502327074e-05, "loss": 2.5673, "theoretical_loss": 3.4704415801141133, "tokens_seen": 1743323136 }, { "epoch": 0.06, "objective/train/docs_used": 962529, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.988215923309326, "objective/train/theoretical_loss": 3.4704304066326683, "objective/train/tokens_used": 113847776, "theoretical_loss": 3.4704304066326683, "tokens_seen": 1743388672 }, { "epoch": 0.06, "learning_rate": 9.528968062911251e-05, "loss": 2.7066, "theoretical_loss": 3.4704192336888413, "tokens_seen": 1743454208 }, { "epoch": 0.06, "learning_rate": 9.528165623495426e-05, "loss": 2.6361, "theoretical_loss": 3.4703968894138573, "tokens_seen": 1743585280 }, { "epoch": 0.06, "learning_rate": 9.527363184079603e-05, "loss": 2.6597, "theoretical_loss": 3.470374547288792, "tokens_seen": 1743716352 }, { "epoch": 0.06, "learning_rate": 9.526560744663778e-05, "loss": 2.7227, "theoretical_loss": 3.4703522073132773, "tokens_seen": 1743847424 }, { "epoch": 0.06, "learning_rate": 9.525758305247953e-05, "loss": 2.4906, "theoretical_loss": 3.470329869486944, "tokens_seen": 1743978496 }, { "epoch": 0.06, "learning_rate": 9.52495586583213e-05, "loss": 2.6045, "theoretical_loss": 3.4703075338094256, "tokens_seen": 1744109568 }, { "epoch": 0.06, "learning_rate": 9.524153426416305e-05, "loss": 2.5417, "theoretical_loss": 3.470285200280352, "tokens_seen": 1744240640 }, { "epoch": 0.06, "learning_rate": 9.523350987000482e-05, "loss": 2.6117, "theoretical_loss": 3.470262868899357, "tokens_seen": 1744371712 }, { "epoch": 0.06, "learning_rate": 9.522548547584657e-05, "loss": 2.6588, "theoretical_loss": 3.4702405396660705, "tokens_seen": 1744502784 }, { "epoch": 0.06, "learning_rate": 9.521746108168834e-05, "loss": 2.6923, "theoretical_loss": 3.470218212580126, "tokens_seen": 1744633856 }, { "epoch": 0.06, "learning_rate": 9.520943668753009e-05, "loss": 2.7159, "theoretical_loss": 3.4701958876411556, "tokens_seen": 1744764928 }, { "epoch": 0.06, "learning_rate": 9.520141229337185e-05, "loss": 2.5938, "theoretical_loss": 3.4701735648487912, "tokens_seen": 1744896000 }, { "epoch": 0.06, "objective/train/docs_used": 963086, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.850843667984009, "objective/train/theoretical_loss": 3.4701512442026656, "objective/train/tokens_used": 115486176, "theoretical_loss": 3.4701512442026656, "tokens_seen": 1745027072 }, { "epoch": 0.06, "learning_rate": 9.519338789921361e-05, "loss": 2.575, "theoretical_loss": 3.4701512442026656, "tokens_seen": 1745027072 }, { "epoch": 0.06, "learning_rate": 9.518536350505537e-05, "loss": 2.6056, "theoretical_loss": 3.4701289257024106, "tokens_seen": 1745158144 }, { "epoch": 0.06, "learning_rate": 9.517733911089713e-05, "loss": 2.5319, "theoretical_loss": 3.47010660934766, "tokens_seen": 1745289216 }, { "epoch": 0.06, "learning_rate": 9.516931471673889e-05, "loss": 2.5956, "theoretical_loss": 3.4700842951380446, "tokens_seen": 1745420288 }, { "epoch": 0.06, "learning_rate": 9.516129032258065e-05, "loss": 2.5044, "theoretical_loss": 3.4700619830731982, "tokens_seen": 1745551360 }, { "epoch": 0.06, "learning_rate": 9.51532659284224e-05, "loss": 2.499, "theoretical_loss": 3.4700396731527543, "tokens_seen": 1745682432 }, { "epoch": 0.06, "learning_rate": 9.514524153426416e-05, "loss": 2.6185, "theoretical_loss": 3.4700173653763446, "tokens_seen": 1745813504 }, { "epoch": 0.06, "learning_rate": 9.513721714010593e-05, "loss": 2.5509, "theoretical_loss": 3.469995059743603, "tokens_seen": 1745944576 }, { "epoch": 0.06, "learning_rate": 9.512919274594768e-05, "loss": 2.5919, "theoretical_loss": 3.469972756254162, "tokens_seen": 1746075648 }, { "epoch": 0.06, "learning_rate": 9.512116835178944e-05, "loss": 2.4592, "theoretical_loss": 3.4699504549076545, "tokens_seen": 1746206720 }, { "epoch": 0.06, "learning_rate": 9.51131439576312e-05, "loss": 2.6279, "theoretical_loss": 3.4699281557037147, "tokens_seen": 1746337792 }, { "epoch": 0.06, "learning_rate": 9.510511956347296e-05, "loss": 2.6709, "theoretical_loss": 3.4699058586419755, "tokens_seen": 1746468864 }, { "epoch": 0.06, "learning_rate": 9.509709516931472e-05, "loss": 2.5483, "theoretical_loss": 3.46988356372207, "tokens_seen": 1746599936 }, { "epoch": 0.06, "objective/train/docs_used": 964151, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.168619394302368, "objective/train/theoretical_loss": 3.469872417065191, "objective/train/tokens_used": 117124576, "theoretical_loss": 3.469872417065191, "tokens_seen": 1746665472 }, { "epoch": 0.06, "learning_rate": 9.508907077515648e-05, "loss": 2.4492, "theoretical_loss": 3.4698612709436327, "tokens_seen": 1746731008 }, { "epoch": 0.06, "learning_rate": 9.508104638099824e-05, "loss": 2.6619, "theoretical_loss": 3.469838980306297, "tokens_seen": 1746862080 }, { "epoch": 0.06, "learning_rate": 9.507302198683999e-05, "loss": 2.6721, "theoretical_loss": 3.4698166918096955, "tokens_seen": 1746993152 }, { "epoch": 0.06, "learning_rate": 9.506499759268176e-05, "loss": 2.502, "theoretical_loss": 3.469794405453463, "tokens_seen": 1747124224 }, { "epoch": 0.06, "learning_rate": 9.505697319852351e-05, "loss": 2.599, "theoretical_loss": 3.4697721212372334, "tokens_seen": 1747255296 }, { "epoch": 0.06, "learning_rate": 9.504894880436528e-05, "loss": 2.6716, "theoretical_loss": 3.4697498391606403, "tokens_seen": 1747386368 }, { "epoch": 0.06, "learning_rate": 9.504092441020703e-05, "loss": 2.5568, "theoretical_loss": 3.469727559223318, "tokens_seen": 1747517440 }, { "epoch": 0.06, "learning_rate": 9.50329000160488e-05, "loss": 2.6531, "theoretical_loss": 3.469705281424901, "tokens_seen": 1747648512 }, { "epoch": 0.06, "learning_rate": 9.502487562189055e-05, "loss": 2.653, "theoretical_loss": 3.469683005765023, "tokens_seen": 1747779584 }, { "epoch": 0.06, "learning_rate": 9.50168512277323e-05, "loss": 2.547, "theoretical_loss": 3.469660732243319, "tokens_seen": 1747910656 }, { "epoch": 0.06, "learning_rate": 9.500882683357407e-05, "loss": 2.5207, "theoretical_loss": 3.469638460859423, "tokens_seen": 1748041728 }, { "epoch": 0.06, "learning_rate": 9.500080243941582e-05, "loss": 2.6592, "theoretical_loss": 3.4696161916129693, "tokens_seen": 1748172800 }, { "epoch": 0.06, "objective/train/docs_used": 964837, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 1.9922051429748535, "objective/train/theoretical_loss": 3.469593924503593, "objective/train/tokens_used": 118762976, "theoretical_loss": 3.469593924503593, "tokens_seen": 1748303872 }, { "epoch": 0.06, "learning_rate": 9.499277804525759e-05, "loss": 2.5553, "theoretical_loss": 3.469593924503593, "tokens_seen": 1748303872 }, { "epoch": 0.06, "learning_rate": 9.498475365109934e-05, "loss": 2.5655, "theoretical_loss": 3.4695716595309287, "tokens_seen": 1748434944 }, { "epoch": 0.06, "learning_rate": 9.497672925694111e-05, "loss": 2.6482, "theoretical_loss": 3.469549396694611, "tokens_seen": 1748566016 }, { "epoch": 0.06, "learning_rate": 9.496870486278286e-05, "loss": 2.5774, "theoretical_loss": 3.469527135994275, "tokens_seen": 1748697088 }, { "epoch": 0.06, "learning_rate": 9.496068046862462e-05, "loss": 2.4727, "theoretical_loss": 3.469504877429556, "tokens_seen": 1748828160 }, { "epoch": 0.06, "learning_rate": 9.495265607446638e-05, "loss": 2.5101, "theoretical_loss": 3.4694826210000884, "tokens_seen": 1748959232 }, { "epoch": 0.06, "learning_rate": 9.494463168030814e-05, "loss": 2.6768, "theoretical_loss": 3.4694603667055075, "tokens_seen": 1749090304 }, { "epoch": 0.06, "learning_rate": 9.49366072861499e-05, "loss": 2.553, "theoretical_loss": 3.4694381145454494, "tokens_seen": 1749221376 }, { "epoch": 0.06, "learning_rate": 9.492858289199166e-05, "loss": 2.4986, "theoretical_loss": 3.469415864519548, "tokens_seen": 1749352448 }, { "epoch": 0.06, "learning_rate": 9.492055849783342e-05, "loss": 2.66, "theoretical_loss": 3.469393616627441, "tokens_seen": 1749483520 }, { "epoch": 0.06, "learning_rate": 9.491253410367518e-05, "loss": 2.5936, "theoretical_loss": 3.4693713708687612, "tokens_seen": 1749614592 }, { "epoch": 0.06, "learning_rate": 9.490450970951693e-05, "loss": 2.5735, "theoretical_loss": 3.4693491272431456, "tokens_seen": 1749745664 }, { "epoch": 0.06, "learning_rate": 9.48964853153587e-05, "loss": 2.6236, "theoretical_loss": 3.46932688575023, "tokens_seen": 1749876736 }, { "epoch": 0.06, "objective/train/docs_used": 965967, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4741082191467285, "objective/train/theoretical_loss": 3.4693157658034206, "objective/train/tokens_used": 120401376, "theoretical_loss": 3.4693157658034206, "tokens_seen": 1749942272 }, { "epoch": 0.06, "learning_rate": 9.488846092120045e-05, "loss": 2.6312, "theoretical_loss": 3.46930464638965, "tokens_seen": 1750007808 }, { "epoch": 0.06, "learning_rate": 9.488043652704221e-05, "loss": 2.6914, "theoretical_loss": 3.4692824091610412, "tokens_seen": 1750138880 }, { "epoch": 0.06, "learning_rate": 9.487241213288397e-05, "loss": 2.4969, "theoretical_loss": 3.4692601740640403, "tokens_seen": 1750269952 }, { "epoch": 0.06, "learning_rate": 9.486438773872573e-05, "loss": 2.568, "theoretical_loss": 3.4692379410982825, "tokens_seen": 1750401024 }, { "epoch": 0.06, "learning_rate": 9.485636334456749e-05, "loss": 2.5101, "theoretical_loss": 3.4692157102634047, "tokens_seen": 1750532096 }, { "epoch": 0.06, "learning_rate": 9.484833895040924e-05, "loss": 2.6857, "theoretical_loss": 3.4691934815590426, "tokens_seen": 1750663168 }, { "epoch": 0.06, "learning_rate": 9.484031455625101e-05, "loss": 2.4291, "theoretical_loss": 3.4691712549848326, "tokens_seen": 1750794240 }, { "epoch": 0.06, "learning_rate": 9.483229016209276e-05, "loss": 2.6526, "theoretical_loss": 3.469149030540412, "tokens_seen": 1750925312 }, { "epoch": 0.06, "learning_rate": 9.482426576793453e-05, "loss": 2.5627, "theoretical_loss": 3.4691268082254156, "tokens_seen": 1751056384 }, { "epoch": 0.06, "learning_rate": 9.481624137377628e-05, "loss": 2.5353, "theoretical_loss": 3.4691045880394817, "tokens_seen": 1751187456 }, { "epoch": 0.06, "learning_rate": 9.480821697961805e-05, "loss": 2.5783, "theoretical_loss": 3.4690823699822455, "tokens_seen": 1751318528 }, { "epoch": 0.06, "learning_rate": 9.48001925854598e-05, "loss": 2.6184, "theoretical_loss": 3.469060154053345, "tokens_seen": 1751449600 }, { "epoch": 0.06, "objective/train/docs_used": 966884, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.789921760559082, "objective/train/theoretical_loss": 3.4690379402524165, "objective/train/tokens_used": 122039776, "theoretical_loss": 3.4690379402524165, "tokens_seen": 1751580672 }, { "epoch": 0.06, "learning_rate": 9.479216819130155e-05, "loss": 2.5307, "theoretical_loss": 3.4690379402524165, "tokens_seen": 1751580672 }, { "epoch": 0.06, "learning_rate": 9.478414379714332e-05, "loss": 2.5257, "theoretical_loss": 3.4690157285790972, "tokens_seen": 1751711744 }, { "epoch": 0.06, "learning_rate": 9.477611940298507e-05, "loss": 2.561, "theoretical_loss": 3.468993519033024, "tokens_seen": 1751842816 }, { "epoch": 0.06, "learning_rate": 9.476809500882684e-05, "loss": 2.3812, "theoretical_loss": 3.468971311613834, "tokens_seen": 1751973888 }, { "epoch": 0.06, "learning_rate": 9.476007061466859e-05, "loss": 2.5491, "theoretical_loss": 3.468949106321164, "tokens_seen": 1752104960 }, { "epoch": 0.06, "learning_rate": 9.475204622051036e-05, "loss": 2.5825, "theoretical_loss": 3.468926903154652, "tokens_seen": 1752236032 }, { "epoch": 0.06, "learning_rate": 9.474402182635211e-05, "loss": 2.608, "theoretical_loss": 3.4689047021139356, "tokens_seen": 1752367104 }, { "epoch": 0.06, "learning_rate": 9.473599743219388e-05, "loss": 2.6951, "theoretical_loss": 3.4688825031986514, "tokens_seen": 1752498176 }, { "epoch": 0.06, "learning_rate": 9.472797303803563e-05, "loss": 2.5484, "theoretical_loss": 3.4688603064084376, "tokens_seen": 1752629248 }, { "epoch": 0.06, "learning_rate": 9.471994864387739e-05, "loss": 2.7742, "theoretical_loss": 3.4688381117429317, "tokens_seen": 1752760320 }, { "epoch": 0.06, "learning_rate": 9.471192424971915e-05, "loss": 2.5422, "theoretical_loss": 3.4688159192017713, "tokens_seen": 1752891392 }, { "epoch": 0.06, "learning_rate": 9.47038998555609e-05, "loss": 2.637, "theoretical_loss": 3.4687937287845947, "tokens_seen": 1753022464 }, { "epoch": 0.06, "learning_rate": 9.469587546140267e-05, "loss": 2.5073, "theoretical_loss": 3.4687715404910398, "tokens_seen": 1753153536 }, { "epoch": 0.06, "objective/train/docs_used": 967417, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.34035325050354, "objective/train/theoretical_loss": 3.4687604471405065, "objective/train/tokens_used": 123678176, "theoretical_loss": 3.4687604471405065, "tokens_seen": 1753219072 }, { "epoch": 0.06, "learning_rate": 9.468785106724443e-05, "loss": 2.5376, "theoretical_loss": 3.4687493543207433, "tokens_seen": 1753284608 }, { "epoch": 0.06, "learning_rate": 9.467982667308619e-05, "loss": 2.6251, "theoretical_loss": 3.468727170273345, "tokens_seen": 1753415680 }, { "epoch": 0.06, "learning_rate": 9.467180227892795e-05, "loss": 2.8049, "theoretical_loss": 3.468704988348483, "tokens_seen": 1753546752 }, { "epoch": 0.06, "learning_rate": 9.46637778847697e-05, "loss": 2.5002, "theoretical_loss": 3.468682808545794, "tokens_seen": 1753677824 }, { "epoch": 0.06, "learning_rate": 9.465575349061147e-05, "loss": 2.5396, "theoretical_loss": 3.4686606308649184, "tokens_seen": 1753808896 }, { "epoch": 0.06, "learning_rate": 9.464772909645322e-05, "loss": 2.5852, "theoretical_loss": 3.468638455305493, "tokens_seen": 1753939968 }, { "epoch": 0.06, "learning_rate": 9.463970470229499e-05, "loss": 2.5664, "theoretical_loss": 3.468616281867157, "tokens_seen": 1754071040 }, { "epoch": 0.06, "learning_rate": 9.463168030813674e-05, "loss": 2.6855, "theoretical_loss": 3.46859411054955, "tokens_seen": 1754202112 }, { "epoch": 0.06, "learning_rate": 9.46236559139785e-05, "loss": 2.6069, "theoretical_loss": 3.468571941352309, "tokens_seen": 1754333184 }, { "epoch": 0.06, "learning_rate": 9.461563151982026e-05, "loss": 2.5943, "theoretical_loss": 3.4685497742750737, "tokens_seen": 1754464256 }, { "epoch": 0.06, "learning_rate": 9.460760712566201e-05, "loss": 2.5968, "theoretical_loss": 3.4685276093174835, "tokens_seen": 1754595328 }, { "epoch": 0.06, "learning_rate": 9.459958273150378e-05, "loss": 2.4725, "theoretical_loss": 3.4685054464791767, "tokens_seen": 1754726400 }, { "epoch": 0.06, "objective/train/docs_used": 968410, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6490659713745117, "objective/train/theoretical_loss": 3.4684832857597923, "objective/train/tokens_used": 125316576, "theoretical_loss": 3.4684832857597923, "tokens_seen": 1754857472 }, { "epoch": 0.06, "learning_rate": 9.459155833734553e-05, "loss": 2.5628, "theoretical_loss": 3.4684832857597923, "tokens_seen": 1754857472 }, { "epoch": 0.06, "learning_rate": 9.45835339431873e-05, "loss": 2.5849, "theoretical_loss": 3.46846112715897, "tokens_seen": 1754988544 }, { "epoch": 0.06, "learning_rate": 9.457550954902905e-05, "loss": 2.6675, "theoretical_loss": 3.468438970676348, "tokens_seen": 1755119616 }, { "epoch": 0.06, "learning_rate": 9.456748515487082e-05, "loss": 2.5566, "theoretical_loss": 3.468416816311567, "tokens_seen": 1755250688 }, { "epoch": 0.06, "learning_rate": 9.455946076071257e-05, "loss": 2.4097, "theoretical_loss": 3.468394664064266, "tokens_seen": 1755381760 }, { "epoch": 0.06, "learning_rate": 9.455143636655432e-05, "loss": 2.6195, "theoretical_loss": 3.468372513934084, "tokens_seen": 1755512832 }, { "epoch": 0.06, "learning_rate": 9.454341197239609e-05, "loss": 2.5105, "theoretical_loss": 3.468350365920662, "tokens_seen": 1755643904 }, { "epoch": 0.06, "learning_rate": 9.453538757823784e-05, "loss": 2.5745, "theoretical_loss": 3.468328220023638, "tokens_seen": 1755774976 }, { "epoch": 0.06, "learning_rate": 9.452736318407961e-05, "loss": 2.5794, "theoretical_loss": 3.4683060762426523, "tokens_seen": 1755906048 }, { "epoch": 0.06, "learning_rate": 9.451933878992136e-05, "loss": 2.563, "theoretical_loss": 3.4682839345773457, "tokens_seen": 1756037120 }, { "epoch": 0.06, "learning_rate": 9.451131439576313e-05, "loss": 2.5961, "theoretical_loss": 3.468261795027357, "tokens_seen": 1756168192 }, { "epoch": 0.06, "learning_rate": 9.450329000160488e-05, "loss": 2.6463, "theoretical_loss": 3.4682396575923264, "tokens_seen": 1756299264 }, { "epoch": 0.06, "learning_rate": 9.449526560744664e-05, "loss": 2.5798, "theoretical_loss": 3.468217522271895, "tokens_seen": 1756430336 }, { "epoch": 0.06, "objective/train/docs_used": 968954, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.701629638671875, "objective/train/theoretical_loss": 3.468206455404541, "objective/train/tokens_used": 126954976, "theoretical_loss": 3.468206455404541, "tokens_seen": 1756495872 }, { "epoch": 0.06, "learning_rate": 9.44872412132884e-05, "loss": 2.4543, "theoretical_loss": 3.468195389065702, "tokens_seen": 1756561408 }, { "epoch": 0.06, "learning_rate": 9.447921681913016e-05, "loss": 2.5349, "theoretical_loss": 3.4681732579733886, "tokens_seen": 1756692480 }, { "epoch": 0.06, "learning_rate": 9.447119242497192e-05, "loss": 2.5532, "theoretical_loss": 3.4681511289945943, "tokens_seen": 1756823552 }, { "epoch": 0.06, "learning_rate": 9.446316803081368e-05, "loss": 2.5709, "theoretical_loss": 3.46812900212896, "tokens_seen": 1756954624 }, { "epoch": 0.06, "learning_rate": 9.445514363665544e-05, "loss": 2.5352, "theoretical_loss": 3.468106877376126, "tokens_seen": 1757085696 }, { "epoch": 0.06, "learning_rate": 9.44471192424972e-05, "loss": 2.6435, "theoretical_loss": 3.468084754735734, "tokens_seen": 1757216768 }, { "epoch": 0.07, "learning_rate": 9.443909484833896e-05, "loss": 2.6461, "theoretical_loss": 3.468062634207424, "tokens_seen": 1757347840 }, { "epoch": 0.07, "learning_rate": 9.443107045418072e-05, "loss": 2.655, "theoretical_loss": 3.4680405157908365, "tokens_seen": 1757478912 }, { "epoch": 0.07, "learning_rate": 9.442304606002247e-05, "loss": 2.6677, "theoretical_loss": 3.4680183994856133, "tokens_seen": 1757609984 }, { "epoch": 0.07, "learning_rate": 9.441502166586424e-05, "loss": 2.6228, "theoretical_loss": 3.4679962852913944, "tokens_seen": 1757741056 }, { "epoch": 0.07, "learning_rate": 9.440699727170599e-05, "loss": 2.6249, "theoretical_loss": 3.4679741732078213, "tokens_seen": 1757872128 }, { "epoch": 0.07, "learning_rate": 9.439897287754776e-05, "loss": 2.5379, "theoretical_loss": 3.467952063234536, "tokens_seen": 1758003200 }, { "epoch": 0.07, "objective/train/docs_used": 970319, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6145591735839844, "objective/train/theoretical_loss": 3.4679299553711793, "objective/train/tokens_used": 128593376, "theoretical_loss": 3.4679299553711793, "tokens_seen": 1758134272 }, { "epoch": 0.07, "learning_rate": 9.439094848338951e-05, "loss": 2.6071, "theoretical_loss": 3.4679299553711793, "tokens_seen": 1758134272 }, { "epoch": 0.07, "learning_rate": 9.438292408923127e-05, "loss": 2.4915, "theoretical_loss": 3.467907849617392, "tokens_seen": 1758265344 }, { "epoch": 0.07, "learning_rate": 9.437489969507303e-05, "loss": 2.4659, "theoretical_loss": 3.467885745972816, "tokens_seen": 1758396416 }, { "epoch": 0.07, "learning_rate": 9.436687530091478e-05, "loss": 2.5481, "theoretical_loss": 3.467863644437093, "tokens_seen": 1758527488 }, { "epoch": 0.07, "learning_rate": 9.435885090675655e-05, "loss": 2.6001, "theoretical_loss": 3.4678415450098643, "tokens_seen": 1758658560 }, { "epoch": 0.07, "learning_rate": 9.43508265125983e-05, "loss": 2.5004, "theoretical_loss": 3.467819447690772, "tokens_seen": 1758789632 }, { "epoch": 0.07, "learning_rate": 9.434280211844007e-05, "loss": 2.6864, "theoretical_loss": 3.467797352479458, "tokens_seen": 1758920704 }, { "epoch": 0.07, "learning_rate": 9.433477772428182e-05, "loss": 2.6084, "theoretical_loss": 3.4677752593755633, "tokens_seen": 1759051776 }, { "epoch": 0.07, "learning_rate": 9.432675333012359e-05, "loss": 2.553, "theoretical_loss": 3.4677531683787306, "tokens_seen": 1759182848 }, { "epoch": 0.07, "learning_rate": 9.431872893596534e-05, "loss": 2.6378, "theoretical_loss": 3.4677310794886025, "tokens_seen": 1759313920 }, { "epoch": 0.07, "learning_rate": 9.43107045418071e-05, "loss": 2.38, "theoretical_loss": 3.46770899270482, "tokens_seen": 1759444992 }, { "epoch": 0.07, "learning_rate": 9.430268014764886e-05, "loss": 2.6261, "theoretical_loss": 3.467686908027026, "tokens_seen": 1759576064 }, { "epoch": 0.07, "learning_rate": 9.429465575349061e-05, "loss": 2.6348, "theoretical_loss": 3.4676648254548628, "tokens_seen": 1759707136 }, { "epoch": 0.07, "objective/train/docs_used": 970997, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7837326526641846, "objective/train/theoretical_loss": 3.467653784958281, "objective/train/tokens_used": 130231776, "theoretical_loss": 3.467653784958281, "tokens_seen": 1759772672 }, { "epoch": 0.07, "learning_rate": 9.428663135933238e-05, "loss": 2.6637, "theoretical_loss": 3.467642744987973, "tokens_seen": 1759838208 }, { "epoch": 0.07, "learning_rate": 9.427860696517413e-05, "loss": 2.6002, "theoretical_loss": 3.467620666625999, "tokens_seen": 1759969280 }, { "epoch": 0.07, "learning_rate": 9.42705825710159e-05, "loss": 2.3965, "theoretical_loss": 3.4675985903685826, "tokens_seen": 1760100352 }, { "epoch": 0.07, "learning_rate": 9.426255817685765e-05, "loss": 2.4942, "theoretical_loss": 3.4675765162153676, "tokens_seen": 1760231424 }, { "epoch": 0.07, "learning_rate": 9.42545337826994e-05, "loss": 2.6354, "theoretical_loss": 3.4675544441659962, "tokens_seen": 1760362496 }, { "epoch": 0.07, "learning_rate": 9.424650938854117e-05, "loss": 2.5839, "theoretical_loss": 3.467532374220112, "tokens_seen": 1760493568 }, { "epoch": 0.07, "learning_rate": 9.423848499438293e-05, "loss": 2.5664, "theoretical_loss": 3.4675103063773567, "tokens_seen": 1760624640 }, { "epoch": 0.07, "learning_rate": 9.423046060022469e-05, "loss": 2.6749, "theoretical_loss": 3.4674882406373744, "tokens_seen": 1760755712 }, { "epoch": 0.07, "learning_rate": 9.422243620606645e-05, "loss": 2.5527, "theoretical_loss": 3.4674661769998076, "tokens_seen": 1760886784 }, { "epoch": 0.07, "learning_rate": 9.421441181190821e-05, "loss": 2.4963, "theoretical_loss": 3.4674441154643, "tokens_seen": 1761017856 }, { "epoch": 0.07, "learning_rate": 9.420638741774997e-05, "loss": 2.5026, "theoretical_loss": 3.467422056030494, "tokens_seen": 1761148928 }, { "epoch": 0.07, "learning_rate": 9.419836302359172e-05, "loss": 2.546, "theoretical_loss": 3.467399998698034, "tokens_seen": 1761280000 }, { "epoch": 0.07, "objective/train/docs_used": 972106, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7436439990997314, "objective/train/theoretical_loss": 3.4673779434665635, "objective/train/tokens_used": 131870176, "theoretical_loss": 3.4673779434665635, "tokens_seen": 1761411072 }, { "epoch": 0.07, "learning_rate": 9.419033862943349e-05, "loss": 2.5274, "theoretical_loss": 3.4673779434665635, "tokens_seen": 1761411072 }, { "epoch": 0.07, "learning_rate": 9.418231423527524e-05, "loss": 2.7645, "theoretical_loss": 3.4673558903357247, "tokens_seen": 1761542144 }, { "epoch": 0.07, "learning_rate": 9.4174289841117e-05, "loss": 2.498, "theoretical_loss": 3.467333839305163, "tokens_seen": 1761673216 }, { "epoch": 0.07, "learning_rate": 9.416626544695876e-05, "loss": 2.6431, "theoretical_loss": 3.467311790374521, "tokens_seen": 1761804288 }, { "epoch": 0.07, "learning_rate": 9.415824105280053e-05, "loss": 2.6502, "theoretical_loss": 3.467289743543443, "tokens_seen": 1761935360 }, { "epoch": 0.07, "learning_rate": 9.415021665864228e-05, "loss": 2.5922, "theoretical_loss": 3.4672676988115727, "tokens_seen": 1762066432 }, { "epoch": 0.07, "learning_rate": 9.414219226448404e-05, "loss": 2.5547, "theoretical_loss": 3.4672456561785543, "tokens_seen": 1762197504 }, { "epoch": 0.07, "learning_rate": 9.41341678703258e-05, "loss": 2.5285, "theoretical_loss": 3.4672236156440315, "tokens_seen": 1762328576 }, { "epoch": 0.07, "learning_rate": 9.412614347616755e-05, "loss": 2.7133, "theoretical_loss": 3.467201577207649, "tokens_seen": 1762459648 }, { "epoch": 0.07, "learning_rate": 9.411811908200932e-05, "loss": 2.516, "theoretical_loss": 3.4671795408690507, "tokens_seen": 1762590720 }, { "epoch": 0.07, "learning_rate": 9.411009468785107e-05, "loss": 2.4072, "theoretical_loss": 3.467157506627881, "tokens_seen": 1762721792 }, { "epoch": 0.07, "learning_rate": 9.410207029369284e-05, "loss": 2.3683, "theoretical_loss": 3.4671354744837837, "tokens_seen": 1762852864 }, { "epoch": 0.07, "learning_rate": 9.409404589953459e-05, "loss": 2.6088, "theoretical_loss": 3.4671134444364045, "tokens_seen": 1762983936 }, { "epoch": 0.07, "objective/train/docs_used": 972675, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4672791957855225, "objective/train/theoretical_loss": 3.467102430198873, "objective/train/tokens_used": 133508576, "theoretical_loss": 3.467102430198873, "tokens_seen": 1763049472 }, { "epoch": 0.07, "learning_rate": 9.408602150537636e-05, "loss": 2.4902, "theoretical_loss": 3.4670914164853874, "tokens_seen": 1763115008 }, { "epoch": 0.07, "learning_rate": 9.407799711121811e-05, "loss": 2.5745, "theoretical_loss": 3.4670693906303773, "tokens_seen": 1763246080 }, { "epoch": 0.07, "learning_rate": 9.406997271705986e-05, "loss": 2.6415, "theoretical_loss": 3.4670473668710184, "tokens_seen": 1763377152 }, { "epoch": 0.07, "learning_rate": 9.406194832290163e-05, "loss": 2.6236, "theoretical_loss": 3.467025345206956, "tokens_seen": 1763508224 }, { "epoch": 0.07, "learning_rate": 9.405392392874338e-05, "loss": 2.5835, "theoretical_loss": 3.4670033256378354, "tokens_seen": 1763639296 }, { "epoch": 0.07, "learning_rate": 9.404589953458515e-05, "loss": 2.5896, "theoretical_loss": 3.4669813081633007, "tokens_seen": 1763770368 }, { "epoch": 0.07, "learning_rate": 9.40378751404269e-05, "loss": 2.4005, "theoretical_loss": 3.466959292782998, "tokens_seen": 1763901440 }, { "epoch": 0.07, "learning_rate": 9.402985074626867e-05, "loss": 2.6698, "theoretical_loss": 3.4669372794965723, "tokens_seen": 1764032512 }, { "epoch": 0.07, "learning_rate": 9.402182635211042e-05, "loss": 2.4925, "theoretical_loss": 3.4669152683036684, "tokens_seen": 1764163584 }, { "epoch": 0.07, "learning_rate": 9.401380195795218e-05, "loss": 2.4732, "theoretical_loss": 3.466893259203932, "tokens_seen": 1764294656 }, { "epoch": 0.07, "learning_rate": 9.400577756379394e-05, "loss": 2.4619, "theoretical_loss": 3.4668712521970084, "tokens_seen": 1764425728 }, { "epoch": 0.07, "learning_rate": 9.39977531696357e-05, "loss": 2.4386, "theoretical_loss": 3.466849247282543, "tokens_seen": 1764556800 }, { "epoch": 0.07, "objective/train/docs_used": 973758, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.705296039581299, "objective/train/theoretical_loss": 3.466827244460182, "objective/train/tokens_used": 135146976, "theoretical_loss": 3.466827244460182, "tokens_seen": 1764687872 }, { "epoch": 0.07, "learning_rate": 9.398972877547746e-05, "loss": 2.6076, "theoretical_loss": 3.466827244460182, "tokens_seen": 1764687872 }, { "epoch": 0.07, "learning_rate": 9.398170438131922e-05, "loss": 2.5514, "theoretical_loss": 3.466805243729571, "tokens_seen": 1764818944 }, { "epoch": 0.07, "learning_rate": 9.397367998716098e-05, "loss": 2.4123, "theoretical_loss": 3.4667832450903555, "tokens_seen": 1764950016 }, { "epoch": 0.07, "learning_rate": 9.396565559300274e-05, "loss": 2.5775, "theoretical_loss": 3.4667612485421815, "tokens_seen": 1765081088 }, { "epoch": 0.07, "learning_rate": 9.395763119884449e-05, "loss": 2.443, "theoretical_loss": 3.466739254084695, "tokens_seen": 1765212160 }, { "epoch": 0.07, "learning_rate": 9.394960680468626e-05, "loss": 2.436, "theoretical_loss": 3.4667172617175424, "tokens_seen": 1765343232 }, { "epoch": 0.07, "learning_rate": 9.394158241052801e-05, "loss": 2.5424, "theoretical_loss": 3.4666952714403694, "tokens_seen": 1765474304 }, { "epoch": 0.07, "learning_rate": 9.393355801636978e-05, "loss": 2.4734, "theoretical_loss": 3.466673283252822, "tokens_seen": 1765605376 }, { "epoch": 0.07, "learning_rate": 9.392553362221153e-05, "loss": 2.5327, "theoretical_loss": 3.4666512971545473, "tokens_seen": 1765736448 }, { "epoch": 0.07, "learning_rate": 9.39175092280533e-05, "loss": 2.5953, "theoretical_loss": 3.4666293131451917, "tokens_seen": 1765867520 }, { "epoch": 0.07, "learning_rate": 9.390948483389505e-05, "loss": 2.5874, "theoretical_loss": 3.4666073312244006, "tokens_seen": 1765998592 }, { "epoch": 0.07, "learning_rate": 9.39014604397368e-05, "loss": 2.4698, "theoretical_loss": 3.4665853513918217, "tokens_seen": 1766129664 }, { "epoch": 0.07, "learning_rate": 9.389343604557857e-05, "loss": 2.4093, "theoretical_loss": 3.4665633736471015, "tokens_seen": 1766260736 }, { "epoch": 0.07, "objective/train/docs_used": 974221, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.551424026489258, "objective/train/theoretical_loss": 3.466552385557578, "objective/train/tokens_used": 136785376, "theoretical_loss": 3.466552385557578, "tokens_seen": 1766326272 }, { "epoch": 0.07, "learning_rate": 9.388541165142032e-05, "loss": 2.5791, "theoretical_loss": 3.4665413979898863, "tokens_seen": 1766391808 }, { "epoch": 0.07, "learning_rate": 9.387738725726209e-05, "loss": 2.4563, "theoretical_loss": 3.466519424419823, "tokens_seen": 1766522880 }, { "epoch": 0.07, "learning_rate": 9.386936286310384e-05, "loss": 2.4674, "theoretical_loss": 3.4664974529365593, "tokens_seen": 1766653952 }, { "epoch": 0.07, "learning_rate": 9.386133846894561e-05, "loss": 2.6283, "theoretical_loss": 3.4664754835397416, "tokens_seen": 1766785024 }, { "epoch": 0.07, "learning_rate": 9.385331407478736e-05, "loss": 2.4859, "theoretical_loss": 3.4664535162290164, "tokens_seen": 1766916096 }, { "epoch": 0.07, "learning_rate": 9.384528968062911e-05, "loss": 2.4277, "theoretical_loss": 3.466431551004032, "tokens_seen": 1767047168 }, { "epoch": 0.07, "learning_rate": 9.383726528647088e-05, "loss": 2.4934, "theoretical_loss": 3.466409587864436, "tokens_seen": 1767178240 }, { "epoch": 0.07, "learning_rate": 9.382924089231263e-05, "loss": 2.476, "theoretical_loss": 3.4663876268098734, "tokens_seen": 1767309312 }, { "epoch": 0.07, "learning_rate": 9.38212164981544e-05, "loss": 2.6492, "theoretical_loss": 3.4663656678399946, "tokens_seen": 1767440384 }, { "epoch": 0.07, "learning_rate": 9.381319210399615e-05, "loss": 2.4337, "theoretical_loss": 3.4663437109544453, "tokens_seen": 1767571456 }, { "epoch": 0.07, "learning_rate": 9.380516770983792e-05, "loss": 2.5528, "theoretical_loss": 3.4663217561528734, "tokens_seen": 1767702528 }, { "epoch": 0.07, "learning_rate": 9.379714331567967e-05, "loss": 2.5765, "theoretical_loss": 3.466299803434927, "tokens_seen": 1767833600 }, { "epoch": 0.07, "objective/train/docs_used": 975253, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.134099006652832, "objective/train/theoretical_loss": 3.4662778528002534, "objective/train/tokens_used": 138423776, "theoretical_loss": 3.4662778528002534, "tokens_seen": 1767964672 }, { "epoch": 0.07, "learning_rate": 9.378911892152144e-05, "loss": 2.5428, "theoretical_loss": 3.4662778528002534, "tokens_seen": 1767964672 }, { "epoch": 0.07, "learning_rate": 9.37810945273632e-05, "loss": 2.4819, "theoretical_loss": 3.4662559042485013, "tokens_seen": 1768095744 }, { "epoch": 0.07, "learning_rate": 9.377307013320495e-05, "loss": 2.6061, "theoretical_loss": 3.466233957779318, "tokens_seen": 1768226816 }, { "epoch": 0.07, "learning_rate": 9.376504573904671e-05, "loss": 2.6773, "theoretical_loss": 3.466212013392351, "tokens_seen": 1768357888 }, { "epoch": 0.07, "learning_rate": 9.375702134488847e-05, "loss": 2.5121, "theoretical_loss": 3.46619007108725, "tokens_seen": 1768488960 }, { "epoch": 0.07, "learning_rate": 9.374899695073023e-05, "loss": 2.5764, "theoretical_loss": 3.466168130863662, "tokens_seen": 1768620032 }, { "epoch": 0.07, "learning_rate": 9.374097255657199e-05, "loss": 2.594, "theoretical_loss": 3.466146192721235, "tokens_seen": 1768751104 }, { "epoch": 0.07, "learning_rate": 9.373294816241375e-05, "loss": 2.6474, "theoretical_loss": 3.4661242566596187, "tokens_seen": 1768882176 }, { "epoch": 0.07, "learning_rate": 9.37249237682555e-05, "loss": 2.6483, "theoretical_loss": 3.46610232267846, "tokens_seen": 1769013248 }, { "epoch": 0.07, "learning_rate": 9.371689937409726e-05, "loss": 2.4411, "theoretical_loss": 3.4660803907774094, "tokens_seen": 1769144320 }, { "epoch": 0.07, "learning_rate": 9.370887497993903e-05, "loss": 2.5386, "theoretical_loss": 3.4660584609561145, "tokens_seen": 1769275392 }, { "epoch": 0.07, "learning_rate": 9.370085058578078e-05, "loss": 2.427, "theoretical_loss": 3.4660365332142233, "tokens_seen": 1769406464 }, { "epoch": 0.07, "learning_rate": 9.369282619162255e-05, "loss": 2.6418, "theoretical_loss": 3.4660146075513856, "tokens_seen": 1769537536 }, { "epoch": 0.07, "objective/train/docs_used": 975853, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6023433208465576, "objective/train/theoretical_loss": 3.466003645499502, "objective/train/tokens_used": 140062176, "theoretical_loss": 3.466003645499502, "tokens_seen": 1769603072 }, { "epoch": 0.07, "learning_rate": 9.36848017974643e-05, "loss": 2.4855, "theoretical_loss": 3.46599268396725, "tokens_seen": 1769668608 }, { "epoch": 0.07, "learning_rate": 9.367677740330607e-05, "loss": 2.5534, "theoretical_loss": 3.465970762461466, "tokens_seen": 1769799680 }, { "epoch": 0.07, "learning_rate": 9.366875300914782e-05, "loss": 2.7021, "theoretical_loss": 3.4659488430336816, "tokens_seen": 1769930752 }, { "epoch": 0.07, "learning_rate": 9.366072861498957e-05, "loss": 2.5897, "theoretical_loss": 3.4659269256835468, "tokens_seen": 1770061824 }, { "epoch": 0.07, "learning_rate": 9.365270422083134e-05, "loss": 2.5387, "theoretical_loss": 3.4659050104107103, "tokens_seen": 1770192896 }, { "epoch": 0.07, "learning_rate": 9.364467982667309e-05, "loss": 2.4854, "theoretical_loss": 3.465883097214822, "tokens_seen": 1770323968 }, { "epoch": 0.07, "learning_rate": 9.363665543251486e-05, "loss": 2.6062, "theoretical_loss": 3.465861186095531, "tokens_seen": 1770455040 }, { "epoch": 0.07, "learning_rate": 9.362863103835661e-05, "loss": 2.6616, "theoretical_loss": 3.4658392770524866, "tokens_seen": 1770586112 }, { "epoch": 0.07, "learning_rate": 9.362060664419838e-05, "loss": 2.54, "theoretical_loss": 3.4658173700853387, "tokens_seen": 1770717184 }, { "epoch": 0.07, "learning_rate": 9.361258225004013e-05, "loss": 2.4839, "theoretical_loss": 3.465795465193737, "tokens_seen": 1770848256 }, { "epoch": 0.07, "learning_rate": 9.360455785588188e-05, "loss": 2.5301, "theoretical_loss": 3.465773562377331, "tokens_seen": 1770979328 }, { "epoch": 0.07, "learning_rate": 9.359653346172365e-05, "loss": 2.4655, "theoretical_loss": 3.4657516616357706, "tokens_seen": 1771110400 }, { "epoch": 0.07, "objective/train/docs_used": 976931, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 1.991319179534912, "objective/train/theoretical_loss": 3.4657297629687056, "objective/train/tokens_used": 141700576, "theoretical_loss": 3.4657297629687056, "tokens_seen": 1771241472 }, { "epoch": 0.07, "learning_rate": 9.35885090675654e-05, "loss": 2.6133, "theoretical_loss": 3.4657297629687056, "tokens_seen": 1771241472 }, { "epoch": 0.07, "learning_rate": 9.358048467340717e-05, "loss": 2.5932, "theoretical_loss": 3.4657078663757868, "tokens_seen": 1771372544 }, { "epoch": 0.07, "learning_rate": 9.357246027924892e-05, "loss": 2.5936, "theoretical_loss": 3.4656859718566633, "tokens_seen": 1771503616 }, { "epoch": 0.07, "learning_rate": 9.356443588509069e-05, "loss": 2.5908, "theoretical_loss": 3.4656640794109856, "tokens_seen": 1771634688 }, { "epoch": 0.07, "learning_rate": 9.355641149093244e-05, "loss": 2.5438, "theoretical_loss": 3.4656421890384044, "tokens_seen": 1771765760 }, { "epoch": 0.07, "learning_rate": 9.35483870967742e-05, "loss": 2.4126, "theoretical_loss": 3.465620300738569, "tokens_seen": 1771896832 }, { "epoch": 0.07, "learning_rate": 9.354036270261596e-05, "loss": 2.6821, "theoretical_loss": 3.465598414511131, "tokens_seen": 1772027904 }, { "epoch": 0.07, "learning_rate": 9.353233830845772e-05, "loss": 2.5196, "theoretical_loss": 3.46557653035574, "tokens_seen": 1772158976 }, { "epoch": 0.07, "learning_rate": 9.352431391429948e-05, "loss": 2.6704, "theoretical_loss": 3.4655546482720476, "tokens_seen": 1772290048 }, { "epoch": 0.07, "learning_rate": 9.351628952014124e-05, "loss": 2.6822, "theoretical_loss": 3.4655327682597035, "tokens_seen": 1772421120 }, { "epoch": 0.07, "learning_rate": 9.3508265125983e-05, "loss": 2.4813, "theoretical_loss": 3.465510890318359, "tokens_seen": 1772552192 }, { "epoch": 0.07, "learning_rate": 9.350024073182476e-05, "loss": 2.5966, "theoretical_loss": 3.465489014447665, "tokens_seen": 1772683264 }, { "epoch": 0.07, "learning_rate": 9.349221633766652e-05, "loss": 2.6041, "theoretical_loss": 3.465467140647272, "tokens_seen": 1772814336 }, { "epoch": 0.07, "objective/train/docs_used": 977871, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.325096368789673, "objective/train/theoretical_loss": 3.4654562045233295, "objective/train/tokens_used": 143338976, "theoretical_loss": 3.4654562045233295, "tokens_seen": 1772879872 }, { "epoch": 0.07, "learning_rate": 9.348419194350828e-05, "loss": 2.5723, "theoretical_loss": 3.465445268916832, "tokens_seen": 1772945408 }, { "epoch": 0.07, "learning_rate": 9.347616754935003e-05, "loss": 2.5643, "theoretical_loss": 3.4654233992559944, "tokens_seen": 1773076480 }, { "epoch": 0.07, "learning_rate": 9.34681431551918e-05, "loss": 2.7092, "theoretical_loss": 3.465401531664412, "tokens_seen": 1773207552 }, { "epoch": 0.07, "learning_rate": 9.346011876103355e-05, "loss": 2.5261, "theoretical_loss": 3.465379666141735, "tokens_seen": 1773338624 }, { "epoch": 0.07, "learning_rate": 9.345209436687532e-05, "loss": 2.6486, "theoretical_loss": 3.4653578026876155, "tokens_seen": 1773469696 }, { "epoch": 0.07, "learning_rate": 9.344406997271707e-05, "loss": 2.4702, "theoretical_loss": 3.4653359413017046, "tokens_seen": 1773600768 }, { "epoch": 0.07, "learning_rate": 9.343604557855882e-05, "loss": 2.6074, "theoretical_loss": 3.465314081983654, "tokens_seen": 1773731840 }, { "epoch": 0.08, "learning_rate": 9.342802118440059e-05, "loss": 2.7033, "theoretical_loss": 3.4652922247331155, "tokens_seen": 1773862912 }, { "epoch": 0.08, "learning_rate": 9.341999679024234e-05, "loss": 2.6171, "theoretical_loss": 3.4652703695497404, "tokens_seen": 1773993984 }, { "epoch": 0.08, "learning_rate": 9.341197239608411e-05, "loss": 2.6683, "theoretical_loss": 3.4652485164331805, "tokens_seen": 1774125056 }, { "epoch": 0.08, "learning_rate": 9.340394800192586e-05, "loss": 2.5473, "theoretical_loss": 3.4652266653830877, "tokens_seen": 1774256128 }, { "epoch": 0.08, "learning_rate": 9.339592360776763e-05, "loss": 2.503, "theoretical_loss": 3.465204816399114, "tokens_seen": 1774387200 }, { "epoch": 0.08, "objective/train/docs_used": 978597, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.45396089553833, "objective/train/theoretical_loss": 3.4651829694809115, "objective/train/tokens_used": 144977376, "theoretical_loss": 3.4651829694809115, "tokens_seen": 1774518272 }, { "epoch": 0.08, "learning_rate": 9.338789921360938e-05, "loss": 2.3952, "theoretical_loss": 3.4651829694809115, "tokens_seen": 1774518272 }, { "epoch": 0.08, "learning_rate": 9.337987481945113e-05, "loss": 2.6399, "theoretical_loss": 3.4651611246281324, "tokens_seen": 1774649344 }, { "epoch": 0.08, "learning_rate": 9.33718504252929e-05, "loss": 2.53, "theoretical_loss": 3.465139281840429, "tokens_seen": 1774780416 }, { "epoch": 0.08, "learning_rate": 9.336382603113465e-05, "loss": 2.4813, "theoretical_loss": 3.465117441117453, "tokens_seen": 1774911488 }, { "epoch": 0.08, "learning_rate": 9.335580163697642e-05, "loss": 2.6267, "theoretical_loss": 3.4650956024588577, "tokens_seen": 1775042560 }, { "epoch": 0.08, "learning_rate": 9.334777724281817e-05, "loss": 2.4972, "theoretical_loss": 3.465073765864295, "tokens_seen": 1775173632 }, { "epoch": 0.08, "learning_rate": 9.333975284865993e-05, "loss": 2.6277, "theoretical_loss": 3.4650519313334174, "tokens_seen": 1775304704 }, { "epoch": 0.08, "learning_rate": 9.33317284545017e-05, "loss": 2.5438, "theoretical_loss": 3.465030098865877, "tokens_seen": 1775435776 }, { "epoch": 0.08, "learning_rate": 9.332370406034345e-05, "loss": 2.4372, "theoretical_loss": 3.465008268461328, "tokens_seen": 1775566848 }, { "epoch": 0.08, "learning_rate": 9.331567966618521e-05, "loss": 2.3739, "theoretical_loss": 3.4649864401194215, "tokens_seen": 1775697920 }, { "epoch": 0.08, "learning_rate": 9.330765527202697e-05, "loss": 2.4286, "theoretical_loss": 3.464964613839812, "tokens_seen": 1775828992 }, { "epoch": 0.08, "learning_rate": 9.329963087786872e-05, "loss": 2.6974, "theoretical_loss": 3.464942789622151, "tokens_seen": 1775960064 }, { "epoch": 0.08, "learning_rate": 9.329160648371049e-05, "loss": 2.4616, "theoretical_loss": 3.4649209674660923, "tokens_seen": 1776091136 }, { "epoch": 0.08, "objective/train/docs_used": 979583, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8104419708251953, "objective/train/theoretical_loss": 3.464910057161056, "objective/train/tokens_used": 146615776, "theoretical_loss": 3.464910057161056, "tokens_seen": 1776156672 }, { "epoch": 0.08, "learning_rate": 9.328358208955224e-05, "loss": 2.6414, "theoretical_loss": 3.4648991473712893, "tokens_seen": 1776222208 }, { "epoch": 0.08, "learning_rate": 9.3275557695394e-05, "loss": 2.524, "theoretical_loss": 3.464877329337395, "tokens_seen": 1776353280 }, { "epoch": 0.08, "learning_rate": 9.326753330123576e-05, "loss": 2.7874, "theoretical_loss": 3.4648555133640624, "tokens_seen": 1776484352 }, { "epoch": 0.08, "learning_rate": 9.325950890707753e-05, "loss": 2.5007, "theoretical_loss": 3.4648336994509448, "tokens_seen": 1776615424 }, { "epoch": 0.08, "learning_rate": 9.325148451291928e-05, "loss": 2.6567, "theoretical_loss": 3.464811887597696, "tokens_seen": 1776746496 }, { "epoch": 0.08, "learning_rate": 9.324346011876103e-05, "loss": 2.5699, "theoretical_loss": 3.4647900778039693, "tokens_seen": 1776877568 }, { "epoch": 0.08, "learning_rate": 9.32354357246028e-05, "loss": 2.5771, "theoretical_loss": 3.4647682700694187, "tokens_seen": 1777008640 }, { "epoch": 0.08, "learning_rate": 9.322741133044455e-05, "loss": 2.6564, "theoretical_loss": 3.4647464643936976, "tokens_seen": 1777139712 }, { "epoch": 0.08, "learning_rate": 9.321938693628632e-05, "loss": 2.5478, "theoretical_loss": 3.4647246607764606, "tokens_seen": 1777270784 }, { "epoch": 0.08, "learning_rate": 9.321136254212807e-05, "loss": 2.5323, "theoretical_loss": 3.4647028592173603, "tokens_seen": 1777401856 }, { "epoch": 0.08, "learning_rate": 9.320333814796983e-05, "loss": 2.6142, "theoretical_loss": 3.4646810597160513, "tokens_seen": 1777532928 }, { "epoch": 0.08, "learning_rate": 9.319531375381159e-05, "loss": 2.6264, "theoretical_loss": 3.4646592622721872, "tokens_seen": 1777664000 }, { "epoch": 0.08, "objective/train/docs_used": 979848, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.569380044937134, "objective/train/theoretical_loss": 3.4646374668854234, "objective/train/tokens_used": 148254176, "theoretical_loss": 3.4646374668854234, "tokens_seen": 1777795072 }, { "epoch": 0.08, "learning_rate": 9.318728935965335e-05, "loss": 2.6984, "theoretical_loss": 3.4646374668854234, "tokens_seen": 1777795072 }, { "epoch": 0.08, "learning_rate": 9.317926496549511e-05, "loss": 2.4607, "theoretical_loss": 3.464615673555413, "tokens_seen": 1777926144 }, { "epoch": 0.08, "learning_rate": 9.317124057133686e-05, "loss": 2.5602, "theoretical_loss": 3.4645938822818105, "tokens_seen": 1778057216 }, { "epoch": 0.08, "learning_rate": 9.316321617717863e-05, "loss": 2.5766, "theoretical_loss": 3.4645720930642705, "tokens_seen": 1778188288 }, { "epoch": 0.08, "learning_rate": 9.315519178302038e-05, "loss": 2.5064, "theoretical_loss": 3.4645503059024474, "tokens_seen": 1778319360 }, { "epoch": 0.08, "learning_rate": 9.314716738886214e-05, "loss": 2.6483, "theoretical_loss": 3.4645285207959953, "tokens_seen": 1778450432 }, { "epoch": 0.08, "learning_rate": 9.31391429947039e-05, "loss": 2.437, "theoretical_loss": 3.4645067377445695, "tokens_seen": 1778581504 }, { "epoch": 0.08, "learning_rate": 9.313111860054566e-05, "loss": 2.7707, "theoretical_loss": 3.4644849567478246, "tokens_seen": 1778712576 }, { "epoch": 0.08, "learning_rate": 9.312309420638742e-05, "loss": 2.6932, "theoretical_loss": 3.464463177805415, "tokens_seen": 1778843648 }, { "epoch": 0.08, "learning_rate": 9.311506981222918e-05, "loss": 2.5334, "theoretical_loss": 3.4644414009169964, "tokens_seen": 1778974720 }, { "epoch": 0.08, "learning_rate": 9.310704541807093e-05, "loss": 2.6573, "theoretical_loss": 3.464419626082223, "tokens_seen": 1779105792 }, { "epoch": 0.08, "learning_rate": 9.30990210239127e-05, "loss": 2.5421, "theoretical_loss": 3.46439785330075, "tokens_seen": 1779236864 }, { "epoch": 0.08, "learning_rate": 9.309099662975445e-05, "loss": 2.6231, "theoretical_loss": 3.464376082572233, "tokens_seen": 1779367936 }, { "epoch": 0.08, "objective/train/docs_used": 981185, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7941625118255615, "objective/train/theoretical_loss": 3.464365197977725, "objective/train/tokens_used": 149892576, "theoretical_loss": 3.464365197977725, "tokens_seen": 1779433472 }, { "epoch": 0.08, "learning_rate": 9.308297223559622e-05, "loss": 2.5164, "theoretical_loss": 3.464354313896327, "tokens_seen": 1779499008 }, { "epoch": 0.08, "learning_rate": 9.307494784143797e-05, "loss": 2.5201, "theoretical_loss": 3.4643325472726865, "tokens_seen": 1779630080 }, { "epoch": 0.08, "learning_rate": 9.306692344727974e-05, "loss": 2.4844, "theoretical_loss": 3.4643107827009674, "tokens_seen": 1779761152 }, { "epoch": 0.08, "learning_rate": 9.305889905312149e-05, "loss": 2.6485, "theoretical_loss": 3.464289020180826, "tokens_seen": 1779892224 }, { "epoch": 0.08, "learning_rate": 9.305087465896324e-05, "loss": 2.5563, "theoretical_loss": 3.4642672597119173, "tokens_seen": 1780023296 }, { "epoch": 0.08, "learning_rate": 9.304285026480501e-05, "loss": 2.4439, "theoretical_loss": 3.464245501293896, "tokens_seen": 1780154368 }, { "epoch": 0.08, "learning_rate": 9.303482587064676e-05, "loss": 2.5935, "theoretical_loss": 3.4642237449264193, "tokens_seen": 1780285440 }, { "epoch": 0.08, "learning_rate": 9.302680147648853e-05, "loss": 2.6076, "theoretical_loss": 3.4642019906091424, "tokens_seen": 1780416512 }, { "epoch": 0.08, "learning_rate": 9.301877708233028e-05, "loss": 2.6269, "theoretical_loss": 3.464180238341721, "tokens_seen": 1780547584 }, { "epoch": 0.08, "learning_rate": 9.301075268817204e-05, "loss": 2.5585, "theoretical_loss": 3.464158488123811, "tokens_seen": 1780678656 }, { "epoch": 0.08, "learning_rate": 9.30027282940138e-05, "loss": 2.6408, "theoretical_loss": 3.464136739955069, "tokens_seen": 1780809728 }, { "epoch": 0.08, "learning_rate": 9.299470389985556e-05, "loss": 2.7265, "theoretical_loss": 3.4641149938351505, "tokens_seen": 1780940800 }, { "epoch": 0.08, "objective/train/docs_used": 981883, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.705595016479492, "objective/train/theoretical_loss": 3.4640932497637125, "objective/train/tokens_used": 151530976, "theoretical_loss": 3.4640932497637125, "tokens_seen": 1781071872 }, { "epoch": 0.08, "learning_rate": 9.298667950569732e-05, "loss": 2.6347, "theoretical_loss": 3.4640932497637125, "tokens_seen": 1781071872 }, { "epoch": 0.08, "learning_rate": 9.297865511153908e-05, "loss": 2.4974, "theoretical_loss": 3.4640715077404103, "tokens_seen": 1781202944 }, { "epoch": 0.08, "learning_rate": 9.297063071738084e-05, "loss": 2.5395, "theoretical_loss": 3.464049767764901, "tokens_seen": 1781334016 }, { "epoch": 0.08, "learning_rate": 9.29626063232226e-05, "loss": 2.5233, "theoretical_loss": 3.4640280298368413, "tokens_seen": 1781465088 }, { "epoch": 0.08, "learning_rate": 9.295458192906435e-05, "loss": 2.5728, "theoretical_loss": 3.4640062939558867, "tokens_seen": 1781596160 }, { "epoch": 0.08, "learning_rate": 9.294655753490612e-05, "loss": 2.7589, "theoretical_loss": 3.4639845601216948, "tokens_seen": 1781727232 }, { "epoch": 0.08, "learning_rate": 9.293853314074787e-05, "loss": 2.6934, "theoretical_loss": 3.4639628283339223, "tokens_seen": 1781858304 }, { "epoch": 0.08, "learning_rate": 9.293050874658963e-05, "loss": 2.5922, "theoretical_loss": 3.463941098592225, "tokens_seen": 1781989376 }, { "epoch": 0.08, "learning_rate": 9.292248435243139e-05, "loss": 2.5203, "theoretical_loss": 3.4639193708962615, "tokens_seen": 1782120448 }, { "epoch": 0.08, "learning_rate": 9.291445995827315e-05, "loss": 2.6153, "theoretical_loss": 3.4638976452456873, "tokens_seen": 1782251520 }, { "epoch": 0.08, "learning_rate": 9.290643556411491e-05, "loss": 2.5865, "theoretical_loss": 3.4638759216401596, "tokens_seen": 1782382592 }, { "epoch": 0.08, "learning_rate": 9.289841116995666e-05, "loss": 2.4681, "theoretical_loss": 3.4638542000793358, "tokens_seen": 1782513664 }, { "epoch": 0.08, "learning_rate": 9.289038677579843e-05, "loss": 2.6832, "theoretical_loss": 3.463832480562873, "tokens_seen": 1782644736 }, { "epoch": 0.08, "objective/train/docs_used": 982960, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.1652934551239014, "objective/train/theoretical_loss": 3.463821621571171, "objective/train/tokens_used": 153169376, "theoretical_loss": 3.463821621571171, "tokens_seen": 1782710272 }, { "epoch": 0.08, "learning_rate": 9.288236238164018e-05, "loss": 2.4893, "theoretical_loss": 3.4638107630904296, "tokens_seen": 1782775808 }, { "epoch": 0.08, "learning_rate": 9.287433798748195e-05, "loss": 2.5598, "theoretical_loss": 3.4637890476616615, "tokens_seen": 1782906880 }, { "epoch": 0.08, "learning_rate": 9.28663135933237e-05, "loss": 2.5903, "theoretical_loss": 3.4637673342762265, "tokens_seen": 1783037952 }, { "epoch": 0.08, "learning_rate": 9.285828919916547e-05, "loss": 2.4675, "theoretical_loss": 3.463745622933782, "tokens_seen": 1783169024 }, { "epoch": 0.08, "learning_rate": 9.285026480500722e-05, "loss": 2.426, "theoretical_loss": 3.463723913633987, "tokens_seen": 1783300096 }, { "epoch": 0.08, "learning_rate": 9.284224041084899e-05, "loss": 2.5681, "theoretical_loss": 3.4637022063764977, "tokens_seen": 1783431168 }, { "epoch": 0.08, "learning_rate": 9.283421601669074e-05, "loss": 2.6727, "theoretical_loss": 3.4636805011609724, "tokens_seen": 1783562240 }, { "epoch": 0.08, "learning_rate": 9.28261916225325e-05, "loss": 2.682, "theoretical_loss": 3.4636587979870685, "tokens_seen": 1783693312 }, { "epoch": 0.08, "learning_rate": 9.281816722837426e-05, "loss": 2.4822, "theoretical_loss": 3.4636370968544448, "tokens_seen": 1783824384 }, { "epoch": 0.08, "learning_rate": 9.281014283421601e-05, "loss": 2.5931, "theoretical_loss": 3.4636153977627586, "tokens_seen": 1783955456 }, { "epoch": 0.08, "learning_rate": 9.280211844005778e-05, "loss": 2.5113, "theoretical_loss": 3.4635937007116686, "tokens_seen": 1784086528 }, { "epoch": 0.08, "learning_rate": 9.279409404589953e-05, "loss": 2.4104, "theoretical_loss": 3.4635720057008332, "tokens_seen": 1784217600 }, { "epoch": 0.08, "objective/train/docs_used": 984043, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.182506799697876, "objective/train/theoretical_loss": 3.4635503127299097, "objective/train/tokens_used": 154807776, "theoretical_loss": 3.4635503127299097, "tokens_seen": 1784348672 }, { "epoch": 0.08, "learning_rate": 9.27860696517413e-05, "loss": 2.5613, "theoretical_loss": 3.4635503127299097, "tokens_seen": 1784348672 }, { "epoch": 0.08, "learning_rate": 9.277804525758305e-05, "loss": 2.5171, "theoretical_loss": 3.4635286217985572, "tokens_seen": 1784479744 }, { "epoch": 0.08, "learning_rate": 9.27700208634248e-05, "loss": 2.5051, "theoretical_loss": 3.463506932906434, "tokens_seen": 1784610816 }, { "epoch": 0.08, "learning_rate": 9.276199646926657e-05, "loss": 2.6653, "theoretical_loss": 3.4634852460531986, "tokens_seen": 1784741888 }, { "epoch": 0.08, "learning_rate": 9.275397207510833e-05, "loss": 2.6551, "theoretical_loss": 3.46346356123851, "tokens_seen": 1784872960 }, { "epoch": 0.08, "learning_rate": 9.274594768095009e-05, "loss": 2.5599, "theoretical_loss": 3.463441878462026, "tokens_seen": 1785004032 }, { "epoch": 0.08, "learning_rate": 9.273792328679185e-05, "loss": 2.4816, "theoretical_loss": 3.463420197723406, "tokens_seen": 1785135104 }, { "epoch": 0.08, "learning_rate": 9.272989889263361e-05, "loss": 2.5629, "theoretical_loss": 3.4633985190223084, "tokens_seen": 1785266176 }, { "epoch": 0.08, "learning_rate": 9.272187449847537e-05, "loss": 2.5634, "theoretical_loss": 3.463376842358393, "tokens_seen": 1785397248 }, { "epoch": 0.08, "learning_rate": 9.271385010431712e-05, "loss": 2.6212, "theoretical_loss": 3.4633551677313186, "tokens_seen": 1785528320 }, { "epoch": 0.08, "learning_rate": 9.270582571015889e-05, "loss": 2.5097, "theoretical_loss": 3.463333495140744, "tokens_seen": 1785659392 }, { "epoch": 0.08, "learning_rate": 9.269780131600064e-05, "loss": 2.5634, "theoretical_loss": 3.463311824586328, "tokens_seen": 1785790464 }, { "epoch": 0.08, "learning_rate": 9.26897769218424e-05, "loss": 2.545, "theoretical_loss": 3.463290156067731, "tokens_seen": 1785921536 }, { "epoch": 0.08, "objective/train/docs_used": 984688, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8167052268981934, "objective/train/theoretical_loss": 3.4632793225717573, "objective/train/tokens_used": 156446176, "theoretical_loss": 3.4632793225717573, "tokens_seen": 1785987072 }, { "epoch": 0.08, "learning_rate": 9.268175252768416e-05, "loss": 2.7015, "theoretical_loss": 3.463268489584611, "tokens_seen": 1786052608 }, { "epoch": 0.08, "learning_rate": 9.267372813352592e-05, "loss": 2.6884, "theoretical_loss": 3.463246825136628, "tokens_seen": 1786183680 }, { "epoch": 0.08, "learning_rate": 9.266570373936768e-05, "loss": 2.5157, "theoretical_loss": 3.4632251627234423, "tokens_seen": 1786314752 }, { "epoch": 0.08, "learning_rate": 9.265767934520943e-05, "loss": 2.5907, "theoretical_loss": 3.4632035023447125, "tokens_seen": 1786445824 }, { "epoch": 0.08, "learning_rate": 9.26496549510512e-05, "loss": 2.6504, "theoretical_loss": 3.4631818440000988, "tokens_seen": 1786576896 }, { "epoch": 0.08, "learning_rate": 9.264163055689295e-05, "loss": 2.5668, "theoretical_loss": 3.4631601876892604, "tokens_seen": 1786707968 }, { "epoch": 0.08, "learning_rate": 9.263360616273472e-05, "loss": 2.6424, "theoretical_loss": 3.4631385334118576, "tokens_seen": 1786839040 }, { "epoch": 0.08, "learning_rate": 9.262558176857647e-05, "loss": 2.6048, "theoretical_loss": 3.463116881167551, "tokens_seen": 1786970112 }, { "epoch": 0.08, "learning_rate": 9.261755737441824e-05, "loss": 2.4931, "theoretical_loss": 3.463095230955999, "tokens_seen": 1787101184 }, { "epoch": 0.08, "learning_rate": 9.260953298025999e-05, "loss": 2.5467, "theoretical_loss": 3.463073582776863, "tokens_seen": 1787232256 }, { "epoch": 0.08, "learning_rate": 9.260150858610174e-05, "loss": 2.3647, "theoretical_loss": 3.4630519366298023, "tokens_seen": 1787363328 }, { "epoch": 0.08, "learning_rate": 9.259348419194351e-05, "loss": 2.4978, "theoretical_loss": 3.4630302925144782, "tokens_seen": 1787494400 }, { "epoch": 0.08, "objective/train/docs_used": 985933, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 1.6754649877548218, "objective/train/theoretical_loss": 3.4630086504305497, "objective/train/tokens_used": 158084576, "theoretical_loss": 3.4630086504305497, "tokens_seen": 1787625472 }, { "epoch": 0.08, "learning_rate": 9.258545979778526e-05, "loss": 2.4287, "theoretical_loss": 3.4630086504305497, "tokens_seen": 1787625472 }, { "epoch": 0.08, "learning_rate": 9.257743540362703e-05, "loss": 2.4368, "theoretical_loss": 3.4629870103776783, "tokens_seen": 1787756544 }, { "epoch": 0.08, "learning_rate": 9.256941100946878e-05, "loss": 2.5856, "theoretical_loss": 3.4629653723555247, "tokens_seen": 1787887616 }, { "epoch": 0.08, "learning_rate": 9.256138661531055e-05, "loss": 2.4903, "theoretical_loss": 3.4629437363637483, "tokens_seen": 1788018688 }, { "epoch": 0.08, "learning_rate": 9.25533622211523e-05, "loss": 2.4972, "theoretical_loss": 3.4629221024020107, "tokens_seen": 1788149760 }, { "epoch": 0.08, "learning_rate": 9.254533782699407e-05, "loss": 2.4449, "theoretical_loss": 3.4629004704699726, "tokens_seen": 1788280832 }, { "epoch": 0.08, "learning_rate": 9.253731343283582e-05, "loss": 2.5606, "theoretical_loss": 3.462878840567294, "tokens_seen": 1788411904 }, { "epoch": 0.08, "learning_rate": 9.252928903867758e-05, "loss": 2.577, "theoretical_loss": 3.4628572126936366, "tokens_seen": 1788542976 }, { "epoch": 0.08, "learning_rate": 9.252126464451934e-05, "loss": 2.5612, "theoretical_loss": 3.4628355868486613, "tokens_seen": 1788674048 }, { "epoch": 0.08, "learning_rate": 9.25132402503611e-05, "loss": 2.4463, "theoretical_loss": 3.462813963032029, "tokens_seen": 1788805120 }, { "epoch": 0.08, "learning_rate": 9.250521585620286e-05, "loss": 2.64, "theoretical_loss": 3.462792341243401, "tokens_seen": 1788936192 }, { "epoch": 0.08, "learning_rate": 9.249719146204462e-05, "loss": 2.543, "theoretical_loss": 3.462770721482438, "tokens_seen": 1789067264 }, { "epoch": 0.08, "learning_rate": 9.248916706788638e-05, "loss": 2.6057, "theoretical_loss": 3.462749103748802, "tokens_seen": 1789198336 }, { "epoch": 0.08, "objective/train/docs_used": 986505, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2175519466400146, "objective/train/theoretical_loss": 3.4627382956421258, "objective/train/tokens_used": 159722976, "theoretical_loss": 3.4627382956421258, "tokens_seen": 1789263872 }, { "epoch": 0.08, "learning_rate": 9.248114267372814e-05, "loss": 2.491, "theoretical_loss": 3.4627274880421544, "tokens_seen": 1789329408 }, { "epoch": 0.08, "learning_rate": 9.247311827956989e-05, "loss": 2.6415, "theoretical_loss": 3.462705874362156, "tokens_seen": 1789460480 }, { "epoch": 0.08, "learning_rate": 9.246509388541166e-05, "loss": 2.3435, "theoretical_loss": 3.4626842627084695, "tokens_seen": 1789591552 }, { "epoch": 0.08, "learning_rate": 9.245706949125341e-05, "loss": 2.4189, "theoretical_loss": 3.4626626530807547, "tokens_seen": 1789722624 }, { "epoch": 0.08, "learning_rate": 9.244904509709518e-05, "loss": 2.4021, "theoretical_loss": 3.4626410454786756, "tokens_seen": 1789853696 }, { "epoch": 0.08, "learning_rate": 9.244102070293693e-05, "loss": 2.4551, "theoretical_loss": 3.462619439901892, "tokens_seen": 1789984768 }, { "epoch": 0.08, "learning_rate": 9.24329963087787e-05, "loss": 2.6548, "theoretical_loss": 3.462597836350067, "tokens_seen": 1790115840 }, { "epoch": 0.09, "learning_rate": 9.242497191462045e-05, "loss": 2.5905, "theoretical_loss": 3.462576234822863, "tokens_seen": 1790246912 }, { "epoch": 0.09, "learning_rate": 9.24169475204622e-05, "loss": 2.6165, "theoretical_loss": 3.46255463531994, "tokens_seen": 1790377984 }, { "epoch": 0.09, "learning_rate": 9.240892312630397e-05, "loss": 2.4857, "theoretical_loss": 3.462533037840962, "tokens_seen": 1790509056 }, { "epoch": 0.09, "learning_rate": 9.240089873214572e-05, "loss": 2.6329, "theoretical_loss": 3.4625114423855905, "tokens_seen": 1790640128 }, { "epoch": 0.09, "learning_rate": 9.239287433798749e-05, "loss": 2.4891, "theoretical_loss": 3.4624898489534877, "tokens_seen": 1790771200 }, { "epoch": 0.09, "objective/train/docs_used": 987700, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4643940925598145, "objective/train/theoretical_loss": 3.462468257544317, "objective/train/tokens_used": 161361376, "theoretical_loss": 3.462468257544317, "tokens_seen": 1790902272 }, { "epoch": 0.09, "learning_rate": 9.238484994382924e-05, "loss": 2.3413, "theoretical_loss": 3.462468257544317, "tokens_seen": 1790902272 }, { "epoch": 0.09, "learning_rate": 9.237682554967101e-05, "loss": 2.4009, "theoretical_loss": 3.462446668157739, "tokens_seen": 1791033344 }, { "epoch": 0.09, "learning_rate": 9.236880115551276e-05, "loss": 2.665, "theoretical_loss": 3.462425080793418, "tokens_seen": 1791164416 }, { "epoch": 0.09, "learning_rate": 9.236077676135451e-05, "loss": 2.5981, "theoretical_loss": 3.4624034954510154, "tokens_seen": 1791295488 }, { "epoch": 0.09, "learning_rate": 9.235275236719628e-05, "loss": 2.6992, "theoretical_loss": 3.462381912130194, "tokens_seen": 1791426560 }, { "epoch": 0.09, "learning_rate": 9.234472797303803e-05, "loss": 2.6392, "theoretical_loss": 3.462360330830618, "tokens_seen": 1791557632 }, { "epoch": 0.09, "learning_rate": 9.23367035788798e-05, "loss": 2.5319, "theoretical_loss": 3.4623387515519486, "tokens_seen": 1791688704 }, { "epoch": 0.09, "learning_rate": 9.232867918472155e-05, "loss": 2.487, "theoretical_loss": 3.462317174293849, "tokens_seen": 1791819776 }, { "epoch": 0.09, "learning_rate": 9.232065479056332e-05, "loss": 2.4134, "theoretical_loss": 3.462295599055983, "tokens_seen": 1791950848 }, { "epoch": 0.09, "learning_rate": 9.231263039640507e-05, "loss": 2.387, "theoretical_loss": 3.4622740258380134, "tokens_seen": 1792081920 }, { "epoch": 0.09, "learning_rate": 9.230460600224683e-05, "loss": 2.5494, "theoretical_loss": 3.4622524546396027, "tokens_seen": 1792212992 }, { "epoch": 0.09, "learning_rate": 9.229658160808859e-05, "loss": 2.4921, "theoretical_loss": 3.4622308854604156, "tokens_seen": 1792344064 }, { "epoch": 0.09, "learning_rate": 9.228855721393035e-05, "loss": 2.4341, "theoretical_loss": 3.462209318300114, "tokens_seen": 1792475136 }, { "epoch": 0.09, "objective/train/docs_used": 988349, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3251655101776123, "objective/train/theoretical_loss": 3.46219853547694, "objective/train/tokens_used": 162999776, "theoretical_loss": 3.46219853547694, "tokens_seen": 1792540672 }, { "epoch": 0.09, "learning_rate": 9.228053281977211e-05, "loss": 2.5972, "theoretical_loss": 3.4621877531583616, "tokens_seen": 1792606208 }, { "epoch": 0.09, "learning_rate": 9.227250842561387e-05, "loss": 2.5254, "theoretical_loss": 3.4621661900348224, "tokens_seen": 1792737280 }, { "epoch": 0.09, "learning_rate": 9.226448403145563e-05, "loss": 2.4832, "theoretical_loss": 3.46214462892916, "tokens_seen": 1792868352 }, { "epoch": 0.09, "learning_rate": 9.225645963729739e-05, "loss": 2.4044, "theoretical_loss": 3.462123069841038, "tokens_seen": 1792999424 }, { "epoch": 0.09, "learning_rate": 9.224843524313914e-05, "loss": 2.6646, "theoretical_loss": 3.4621015127701193, "tokens_seen": 1793130496 }, { "epoch": 0.09, "learning_rate": 9.22404108489809e-05, "loss": 2.4458, "theoretical_loss": 3.462079957716069, "tokens_seen": 1793261568 }, { "epoch": 0.09, "learning_rate": 9.223238645482266e-05, "loss": 2.4208, "theoretical_loss": 3.46205840467855, "tokens_seen": 1793392640 }, { "epoch": 0.09, "learning_rate": 9.222436206066443e-05, "loss": 2.5817, "theoretical_loss": 3.462036853657227, "tokens_seen": 1793523712 }, { "epoch": 0.09, "learning_rate": 9.221633766650618e-05, "loss": 2.5792, "theoretical_loss": 3.4620153046517634, "tokens_seen": 1793654784 }, { "epoch": 0.09, "learning_rate": 9.220831327234795e-05, "loss": 2.5938, "theoretical_loss": 3.461993757661824, "tokens_seen": 1793785856 }, { "epoch": 0.09, "learning_rate": 9.22002888781897e-05, "loss": 2.5438, "theoretical_loss": 3.4619722126870727, "tokens_seen": 1793916928 }, { "epoch": 0.09, "learning_rate": 9.219226448403146e-05, "loss": 2.5115, "theoretical_loss": 3.461950669727174, "tokens_seen": 1794048000 }, { "epoch": 0.09, "objective/train/docs_used": 989683, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.0829696655273438, "objective/train/theoretical_loss": 3.461929128781792, "objective/train/tokens_used": 164638176, "theoretical_loss": 3.461929128781792, "tokens_seen": 1794179072 }, { "epoch": 0.09, "learning_rate": 9.218424008987322e-05, "loss": 2.343, "theoretical_loss": 3.461929128781792, "tokens_seen": 1794179072 }, { "epoch": 0.09, "learning_rate": 9.217621569571497e-05, "loss": 2.5634, "theoretical_loss": 3.4619075898505915, "tokens_seen": 1794310144 }, { "epoch": 0.09, "learning_rate": 9.216819130155674e-05, "loss": 2.5899, "theoretical_loss": 3.4618860529332363, "tokens_seen": 1794441216 }, { "epoch": 0.09, "learning_rate": 9.216016690739849e-05, "loss": 2.4196, "theoretical_loss": 3.461864518029392, "tokens_seen": 1794572288 }, { "epoch": 0.09, "learning_rate": 9.215214251324026e-05, "loss": 2.5227, "theoretical_loss": 3.4618429851387233, "tokens_seen": 1794703360 }, { "epoch": 0.09, "learning_rate": 9.214411811908201e-05, "loss": 2.442, "theoretical_loss": 3.4618214542608943, "tokens_seen": 1794834432 }, { "epoch": 0.09, "learning_rate": 9.213609372492378e-05, "loss": 2.5729, "theoretical_loss": 3.46179992539557, "tokens_seen": 1794965504 }, { "epoch": 0.09, "learning_rate": 9.212806933076553e-05, "loss": 2.5655, "theoretical_loss": 3.4617783985424158, "tokens_seen": 1795096576 }, { "epoch": 0.09, "learning_rate": 9.212004493660728e-05, "loss": 2.4973, "theoretical_loss": 3.461756873701096, "tokens_seen": 1795227648 }, { "epoch": 0.09, "learning_rate": 9.211202054244905e-05, "loss": 2.7353, "theoretical_loss": 3.461735350871277, "tokens_seen": 1795358720 }, { "epoch": 0.09, "learning_rate": 9.21039961482908e-05, "loss": 2.5919, "theoretical_loss": 3.4617138300526222, "tokens_seen": 1795489792 }, { "epoch": 0.09, "learning_rate": 9.209597175413257e-05, "loss": 2.4516, "theoretical_loss": 3.461692311244798, "tokens_seen": 1795620864 }, { "epoch": 0.09, "learning_rate": 9.208794735997432e-05, "loss": 2.4701, "theoretical_loss": 3.4616707944474703, "tokens_seen": 1795751936 }, { "epoch": 0.09, "objective/train/docs_used": 990185, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.836434841156006, "objective/train/theoretical_loss": 3.4616600368026376, "objective/train/tokens_used": 166276576, "theoretical_loss": 3.4616600368026376, "tokens_seen": 1795817472 }, { "epoch": 0.09, "learning_rate": 9.207992296581609e-05, "loss": 2.5007, "theoretical_loss": 3.4616492796603033, "tokens_seen": 1795883008 }, { "epoch": 0.09, "learning_rate": 9.207189857165784e-05, "loss": 2.6287, "theoretical_loss": 3.4616277668829634, "tokens_seen": 1796014080 }, { "epoch": 0.09, "learning_rate": 9.20638741774996e-05, "loss": 2.6319, "theoretical_loss": 3.461606256115116, "tokens_seen": 1796145152 }, { "epoch": 0.09, "learning_rate": 9.205584978334136e-05, "loss": 2.3697, "theoretical_loss": 3.4615847473564263, "tokens_seen": 1796276224 }, { "epoch": 0.09, "learning_rate": 9.204782538918312e-05, "loss": 2.5783, "theoretical_loss": 3.4615632406065604, "tokens_seen": 1796407296 }, { "epoch": 0.09, "learning_rate": 9.203980099502488e-05, "loss": 2.4935, "theoretical_loss": 3.4615417358651843, "tokens_seen": 1796538368 }, { "epoch": 0.09, "learning_rate": 9.203177660086664e-05, "loss": 2.6099, "theoretical_loss": 3.4615202331319637, "tokens_seen": 1796669440 }, { "epoch": 0.09, "learning_rate": 9.20237522067084e-05, "loss": 2.6371, "theoretical_loss": 3.4614987324065645, "tokens_seen": 1796800512 }, { "epoch": 0.09, "learning_rate": 9.201572781255016e-05, "loss": 2.464, "theoretical_loss": 3.461477233688653, "tokens_seen": 1796931584 }, { "epoch": 0.09, "learning_rate": 9.200770341839191e-05, "loss": 2.609, "theoretical_loss": 3.4614557369778947, "tokens_seen": 1797062656 }, { "epoch": 0.09, "learning_rate": 9.199967902423368e-05, "loss": 2.5767, "theoretical_loss": 3.461434242273957, "tokens_seen": 1797193728 }, { "epoch": 0.09, "learning_rate": 9.199165463007543e-05, "loss": 2.641, "theoretical_loss": 3.461412749576505, "tokens_seen": 1797324800 }, { "epoch": 0.09, "objective/train/docs_used": 991515, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3866214752197266, "objective/train/theoretical_loss": 3.4613912588852065, "objective/train/tokens_used": 167914976, "theoretical_loss": 3.4613912588852065, "tokens_seen": 1797455872 }, { "epoch": 0.09, "learning_rate": 9.19836302359172e-05, "loss": 2.6222, "theoretical_loss": 3.4613912588852065, "tokens_seen": 1797455872 }, { "epoch": 0.09, "learning_rate": 9.197560584175895e-05, "loss": 2.6715, "theoretical_loss": 3.4613697701997266, "tokens_seen": 1797586944 }, { "epoch": 0.09, "learning_rate": 9.196758144760072e-05, "loss": 2.5473, "theoretical_loss": 3.4613482835197322, "tokens_seen": 1797718016 }, { "epoch": 0.09, "learning_rate": 9.195955705344247e-05, "loss": 2.5753, "theoretical_loss": 3.4613267988448913, "tokens_seen": 1797849088 }, { "epoch": 0.09, "learning_rate": 9.195153265928422e-05, "loss": 2.6014, "theoretical_loss": 3.4613053161748684, "tokens_seen": 1797980160 }, { "epoch": 0.09, "learning_rate": 9.194350826512599e-05, "loss": 2.5159, "theoretical_loss": 3.461283835509332, "tokens_seen": 1798111232 }, { "epoch": 0.09, "learning_rate": 9.193548387096774e-05, "loss": 2.5353, "theoretical_loss": 3.4612623568479477, "tokens_seen": 1798242304 }, { "epoch": 0.09, "learning_rate": 9.192745947680951e-05, "loss": 2.5962, "theoretical_loss": 3.4612408801903833, "tokens_seen": 1798373376 }, { "epoch": 0.09, "learning_rate": 9.191943508265126e-05, "loss": 2.5236, "theoretical_loss": 3.4612194055363057, "tokens_seen": 1798504448 }, { "epoch": 0.09, "learning_rate": 9.191141068849303e-05, "loss": 2.5268, "theoretical_loss": 3.461197932885382, "tokens_seen": 1798635520 }, { "epoch": 0.09, "learning_rate": 9.190338629433478e-05, "loss": 2.5341, "theoretical_loss": 3.4611764622372796, "tokens_seen": 1798766592 }, { "epoch": 0.09, "learning_rate": 9.189536190017655e-05, "loss": 2.538, "theoretical_loss": 3.4611549935916646, "tokens_seen": 1798897664 }, { "epoch": 0.09, "learning_rate": 9.18873375060183e-05, "loss": 2.6186, "theoretical_loss": 3.461133526948206, "tokens_seen": 1799028736 }, { "epoch": 0.09, "objective/train/docs_used": 992149, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.483299732208252, "objective/train/theoretical_loss": 3.4611227943771814, "objective/train/tokens_used": 169553376, "theoretical_loss": 3.4611227943771814, "tokens_seen": 1799094272 }, { "epoch": 0.09, "learning_rate": 9.187931311186005e-05, "loss": 2.565, "theoretical_loss": 3.4611120623065705, "tokens_seen": 1799159808 }, { "epoch": 0.09, "learning_rate": 9.187128871770182e-05, "loss": 2.5304, "theoretical_loss": 3.4610905996664254, "tokens_seen": 1799290880 }, { "epoch": 0.09, "learning_rate": 9.186326432354357e-05, "loss": 2.5325, "theoretical_loss": 3.461069139027438, "tokens_seen": 1799421952 }, { "epoch": 0.09, "learning_rate": 9.185523992938534e-05, "loss": 2.6046, "theoretical_loss": 3.461047680389277, "tokens_seen": 1799553024 }, { "epoch": 0.09, "learning_rate": 9.18472155352271e-05, "loss": 2.5433, "theoretical_loss": 3.4610262237516096, "tokens_seen": 1799684096 }, { "epoch": 0.09, "learning_rate": 9.183919114106886e-05, "loss": 2.5845, "theoretical_loss": 3.461004769114103, "tokens_seen": 1799815168 }, { "epoch": 0.09, "learning_rate": 9.183116674691061e-05, "loss": 2.5856, "theoretical_loss": 3.460983316476426, "tokens_seen": 1799946240 }, { "epoch": 0.09, "learning_rate": 9.182314235275237e-05, "loss": 2.6222, "theoretical_loss": 3.4609618658382466, "tokens_seen": 1800077312 }, { "epoch": 0.09, "learning_rate": 9.181511795859413e-05, "loss": 2.5011, "theoretical_loss": 3.4609404171992324, "tokens_seen": 1800208384 }, { "epoch": 0.09, "learning_rate": 9.180709356443589e-05, "loss": 2.5571, "theoretical_loss": 3.4609189705590513, "tokens_seen": 1800339456 }, { "epoch": 0.09, "learning_rate": 9.179906917027765e-05, "loss": 2.3876, "theoretical_loss": 3.460897525917373, "tokens_seen": 1800470528 }, { "epoch": 0.09, "learning_rate": 9.17910447761194e-05, "loss": 2.5827, "theoretical_loss": 3.4608760832738636, "tokens_seen": 1800601600 }, { "epoch": 0.09, "objective/train/docs_used": 993279, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.028233766555786, "objective/train/theoretical_loss": 3.4608546426281928, "objective/train/tokens_used": 171191776, "theoretical_loss": 3.4608546426281928, "tokens_seen": 1800732672 }, { "epoch": 0.09, "learning_rate": 9.178302038196117e-05, "loss": 2.5501, "theoretical_loss": 3.4608546426281928, "tokens_seen": 1800732672 }, { "epoch": 0.09, "learning_rate": 9.177499598780293e-05, "loss": 2.4478, "theoretical_loss": 3.4608332039800294, "tokens_seen": 1800863744 }, { "epoch": 0.09, "learning_rate": 9.176697159364468e-05, "loss": 2.4669, "theoretical_loss": 3.460811767329041, "tokens_seen": 1800994816 }, { "epoch": 0.09, "learning_rate": 9.175894719948645e-05, "loss": 2.6295, "theoretical_loss": 3.4607903326748968, "tokens_seen": 1801125888 }, { "epoch": 0.09, "learning_rate": 9.17509228053282e-05, "loss": 2.6089, "theoretical_loss": 3.460768900017265, "tokens_seen": 1801256960 }, { "epoch": 0.09, "learning_rate": 9.174289841116997e-05, "loss": 2.6017, "theoretical_loss": 3.460747469355815, "tokens_seen": 1801388032 }, { "epoch": 0.09, "learning_rate": 9.173487401701172e-05, "loss": 2.6287, "theoretical_loss": 3.460726040690215, "tokens_seen": 1801519104 }, { "epoch": 0.09, "learning_rate": 9.172684962285349e-05, "loss": 2.5425, "theoretical_loss": 3.460704614020135, "tokens_seen": 1801650176 }, { "epoch": 0.09, "learning_rate": 9.171882522869524e-05, "loss": 2.4997, "theoretical_loss": 3.4606831893452425, "tokens_seen": 1801781248 }, { "epoch": 0.09, "learning_rate": 9.171080083453699e-05, "loss": 2.5585, "theoretical_loss": 3.4606617666652078, "tokens_seen": 1801912320 }, { "epoch": 0.09, "learning_rate": 9.170277644037876e-05, "loss": 2.5326, "theoretical_loss": 3.4606403459796997, "tokens_seen": 1802043392 }, { "epoch": 0.09, "learning_rate": 9.169475204622051e-05, "loss": 2.6015, "theoretical_loss": 3.4606189272883876, "tokens_seen": 1802174464 }, { "epoch": 0.09, "learning_rate": 9.168672765206228e-05, "loss": 2.5776, "theoretical_loss": 3.46059751059094, "tokens_seen": 1802305536 }, { "epoch": 0.09, "objective/train/docs_used": 993741, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3329663276672363, "objective/train/theoretical_loss": 3.4605868029898126, "objective/train/tokens_used": 172830176, "theoretical_loss": 3.4605868029898126, "tokens_seen": 1802371072 }, { "epoch": 0.09, "learning_rate": 9.167870325790403e-05, "loss": 2.5359, "theoretical_loss": 3.4605760958870277, "tokens_seen": 1802436608 }, { "epoch": 0.09, "learning_rate": 9.16706788637458e-05, "loss": 2.6917, "theoretical_loss": 3.4605546831763188, "tokens_seen": 1802567680 }, { "epoch": 0.09, "learning_rate": 9.166265446958755e-05, "loss": 2.5287, "theoretical_loss": 3.460533272458484, "tokens_seen": 1802698752 }, { "epoch": 0.09, "learning_rate": 9.16546300754293e-05, "loss": 2.6202, "theoretical_loss": 3.460511863733192, "tokens_seen": 1802829824 }, { "epoch": 0.09, "learning_rate": 9.164660568127107e-05, "loss": 2.6136, "theoretical_loss": 3.4604904570001134, "tokens_seen": 1802960896 }, { "epoch": 0.09, "learning_rate": 9.163858128711282e-05, "loss": 2.7156, "theoretical_loss": 3.460469052258917, "tokens_seen": 1803091968 }, { "epoch": 0.09, "learning_rate": 9.163055689295459e-05, "loss": 2.6449, "theoretical_loss": 3.460447649509274, "tokens_seen": 1803223040 }, { "epoch": 0.09, "learning_rate": 9.162253249879634e-05, "loss": 2.6126, "theoretical_loss": 3.4604262487508533, "tokens_seen": 1803354112 }, { "epoch": 0.09, "learning_rate": 9.161450810463811e-05, "loss": 2.6197, "theoretical_loss": 3.460404849983325, "tokens_seen": 1803485184 }, { "epoch": 0.09, "learning_rate": 9.160648371047986e-05, "loss": 2.4627, "theoretical_loss": 3.4603834532063598, "tokens_seen": 1803616256 }, { "epoch": 0.09, "learning_rate": 9.159845931632163e-05, "loss": 2.5741, "theoretical_loss": 3.4603620584196273, "tokens_seen": 1803747328 }, { "epoch": 0.09, "learning_rate": 9.159043492216338e-05, "loss": 2.4163, "theoretical_loss": 3.460340665622798, "tokens_seen": 1803878400 }, { "epoch": 0.09, "objective/train/docs_used": 994982, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4623970985412598, "objective/train/theoretical_loss": 3.460319274815543, "objective/train/tokens_used": 174468576, "theoretical_loss": 3.460319274815543, "tokens_seen": 1804009472 }, { "epoch": 0.09, "learning_rate": 9.158241052800514e-05, "loss": 2.5558, "theoretical_loss": 3.460319274815543, "tokens_seen": 1804009472 }, { "epoch": 0.09, "learning_rate": 9.15743861338469e-05, "loss": 2.6833, "theoretical_loss": 3.460297885997531, "tokens_seen": 1804140544 }, { "epoch": 0.09, "learning_rate": 9.156636173968866e-05, "loss": 2.5265, "theoretical_loss": 3.4602764991684345, "tokens_seen": 1804271616 }, { "epoch": 0.09, "learning_rate": 9.155833734553042e-05, "loss": 2.3257, "theoretical_loss": 3.4602551143279228, "tokens_seen": 1804402688 }, { "epoch": 0.09, "learning_rate": 9.155031295137218e-05, "loss": 2.5691, "theoretical_loss": 3.4602337314756664, "tokens_seen": 1804533760 }, { "epoch": 0.09, "learning_rate": 9.154228855721394e-05, "loss": 2.5498, "theoretical_loss": 3.4602123506113376, "tokens_seen": 1804664832 }, { "epoch": 0.09, "learning_rate": 9.15342641630557e-05, "loss": 2.545, "theoretical_loss": 3.460190971734605, "tokens_seen": 1804795904 }, { "epoch": 0.09, "learning_rate": 9.152623976889745e-05, "loss": 2.6298, "theoretical_loss": 3.4601695948451416, "tokens_seen": 1804926976 }, { "epoch": 0.09, "learning_rate": 9.151821537473922e-05, "loss": 2.6275, "theoretical_loss": 3.460148219942617, "tokens_seen": 1805058048 }, { "epoch": 0.09, "learning_rate": 9.151019098058097e-05, "loss": 2.5375, "theoretical_loss": 3.4601268470267024, "tokens_seen": 1805189120 }, { "epoch": 0.09, "learning_rate": 9.150216658642274e-05, "loss": 2.4901, "theoretical_loss": 3.4601054760970698, "tokens_seen": 1805320192 }, { "epoch": 0.09, "learning_rate": 9.149414219226449e-05, "loss": 2.6574, "theoretical_loss": 3.4600841071533894, "tokens_seen": 1805451264 }, { "epoch": 0.09, "learning_rate": 9.148611779810626e-05, "loss": 2.508, "theoretical_loss": 3.460062740195333, "tokens_seen": 1805582336 }, { "epoch": 0.09, "objective/train/docs_used": 995914, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.526226758956909, "objective/train/theoretical_loss": 3.4600520574608113, "objective/train/tokens_used": 176106976, "theoretical_loss": 3.4600520574608113, "tokens_seen": 1805647872 }, { "epoch": 0.09, "learning_rate": 9.147809340394801e-05, "loss": 2.4625, "theoretical_loss": 3.4600413752225725, "tokens_seen": 1805713408 }, { "epoch": 0.09, "learning_rate": 9.147006900978976e-05, "loss": 2.4556, "theoretical_loss": 3.460020012234778, "tokens_seen": 1805844480 }, { "epoch": 0.09, "learning_rate": 9.146204461563153e-05, "loss": 2.56, "theoretical_loss": 3.459998651231622, "tokens_seen": 1805975552 }, { "epoch": 0.09, "learning_rate": 9.145402022147328e-05, "loss": 2.5144, "theoretical_loss": 3.4599772922127765, "tokens_seen": 1806106624 }, { "epoch": 0.09, "learning_rate": 9.144599582731505e-05, "loss": 2.5989, "theoretical_loss": 3.459955935177912, "tokens_seen": 1806237696 }, { "epoch": 0.09, "learning_rate": 9.14379714331568e-05, "loss": 2.6055, "theoretical_loss": 3.459934580126701, "tokens_seen": 1806368768 }, { "epoch": 0.09, "learning_rate": 9.142994703899857e-05, "loss": 2.5955, "theoretical_loss": 3.4599132270588155, "tokens_seen": 1806499840 }, { "epoch": 0.09, "learning_rate": 9.142192264484032e-05, "loss": 2.6991, "theoretical_loss": 3.459891875973927, "tokens_seen": 1806630912 }, { "epoch": 0.1, "learning_rate": 9.141389825068207e-05, "loss": 2.6054, "theoretical_loss": 3.4598705268717076, "tokens_seen": 1806761984 }, { "epoch": 0.1, "learning_rate": 9.140587385652384e-05, "loss": 2.6119, "theoretical_loss": 3.4598491797518287, "tokens_seen": 1806893056 }, { "epoch": 0.1, "learning_rate": 9.13978494623656e-05, "loss": 2.6178, "theoretical_loss": 3.459827834613964, "tokens_seen": 1807024128 }, { "epoch": 0.1, "learning_rate": 9.138982506820736e-05, "loss": 2.7865, "theoretical_loss": 3.4598064914577846, "tokens_seen": 1807155200 }, { "epoch": 0.1, "objective/train/docs_used": 996601, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5091497898101807, "objective/train/theoretical_loss": 3.459785150282963, "objective/train/tokens_used": 177745376, "theoretical_loss": 3.459785150282963, "tokens_seen": 1807286272 }, { "epoch": 0.1, "learning_rate": 9.138180067404911e-05, "loss": 2.6171, "theoretical_loss": 3.459785150282963, "tokens_seen": 1807286272 }, { "epoch": 0.1, "learning_rate": 9.137377627989088e-05, "loss": 2.4956, "theoretical_loss": 3.4597638110891715, "tokens_seen": 1807417344 }, { "epoch": 0.1, "learning_rate": 9.136575188573263e-05, "loss": 2.5369, "theoretical_loss": 3.459742473876083, "tokens_seen": 1807548416 }, { "epoch": 0.1, "learning_rate": 9.135772749157439e-05, "loss": 2.5218, "theoretical_loss": 3.4597211386433697, "tokens_seen": 1807679488 }, { "epoch": 0.1, "learning_rate": 9.134970309741615e-05, "loss": 2.54, "theoretical_loss": 3.459699805390705, "tokens_seen": 1807810560 }, { "epoch": 0.1, "learning_rate": 9.13416787032579e-05, "loss": 2.5169, "theoretical_loss": 3.45967847411776, "tokens_seen": 1807941632 }, { "epoch": 0.1, "learning_rate": 9.133365430909967e-05, "loss": 2.6301, "theoretical_loss": 3.4596571448242086, "tokens_seen": 1808072704 }, { "epoch": 0.1, "learning_rate": 9.132562991494143e-05, "loss": 2.6115, "theoretical_loss": 3.4596358175097235, "tokens_seen": 1808203776 }, { "epoch": 0.1, "learning_rate": 9.131760552078319e-05, "loss": 2.5861, "theoretical_loss": 3.4596144921739773, "tokens_seen": 1808334848 }, { "epoch": 0.1, "learning_rate": 9.130958112662495e-05, "loss": 2.5045, "theoretical_loss": 3.459593168816644, "tokens_seen": 1808465920 }, { "epoch": 0.1, "learning_rate": 9.13015567324667e-05, "loss": 2.571, "theoretical_loss": 3.4595718474373953, "tokens_seen": 1808596992 }, { "epoch": 0.1, "learning_rate": 9.129353233830847e-05, "loss": 2.6187, "theoretical_loss": 3.459550528035905, "tokens_seen": 1808728064 }, { "epoch": 0.1, "learning_rate": 9.128550794415022e-05, "loss": 2.5396, "theoretical_loss": 3.459529210611847, "tokens_seen": 1808859136 }, { "epoch": 0.1, "objective/train/docs_used": 997711, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7451651096343994, "objective/train/theoretical_loss": 3.4595185526412524, "objective/train/tokens_used": 179383776, "theoretical_loss": 3.4595185526412524, "tokens_seen": 1808924672 }, { "epoch": 0.1, "learning_rate": 9.127748354999199e-05, "loss": 2.5304, "theoretical_loss": 3.4595078951648937, "tokens_seen": 1808990208 }, { "epoch": 0.1, "learning_rate": 9.126945915583374e-05, "loss": 2.5649, "theoretical_loss": 3.4594865816947182, "tokens_seen": 1809121280 }, { "epoch": 0.1, "learning_rate": 9.12614347616755e-05, "loss": 2.6362, "theoretical_loss": 3.4594652702009956, "tokens_seen": 1809252352 }, { "epoch": 0.1, "learning_rate": 9.125341036751726e-05, "loss": 2.6765, "theoretical_loss": 3.4594439606833975, "tokens_seen": 1809383424 }, { "epoch": 0.1, "learning_rate": 9.124538597335903e-05, "loss": 2.5887, "theoretical_loss": 3.4594226531415986, "tokens_seen": 1809514496 }, { "epoch": 0.1, "learning_rate": 9.123736157920078e-05, "loss": 2.5145, "theoretical_loss": 3.459401347575273, "tokens_seen": 1809645568 }, { "epoch": 0.1, "learning_rate": 9.122933718504253e-05, "loss": 2.6369, "theoretical_loss": 3.4593800439840936, "tokens_seen": 1809776640 }, { "epoch": 0.1, "learning_rate": 9.12213127908843e-05, "loss": 2.6258, "theoretical_loss": 3.4593587423677348, "tokens_seen": 1809907712 }, { "epoch": 0.1, "learning_rate": 9.121328839672605e-05, "loss": 2.3457, "theoretical_loss": 3.45933744272587, "tokens_seen": 1810038784 }, { "epoch": 0.1, "learning_rate": 9.120526400256782e-05, "loss": 2.4792, "theoretical_loss": 3.4593161450581738, "tokens_seen": 1810169856 }, { "epoch": 0.1, "learning_rate": 9.119723960840957e-05, "loss": 2.5877, "theoretical_loss": 3.45929484936432, "tokens_seen": 1810300928 }, { "epoch": 0.1, "learning_rate": 9.118921521425134e-05, "loss": 2.4965, "theoretical_loss": 3.459273555643983, "tokens_seen": 1810432000 }, { "epoch": 0.1, "objective/train/docs_used": 998245, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.329852342605591, "objective/train/theoretical_loss": 3.459252263896837, "objective/train/tokens_used": 181022176, "theoretical_loss": 3.459252263896837, "tokens_seen": 1810563072 }, { "epoch": 0.1, "learning_rate": 9.118119082009309e-05, "loss": 2.5769, "theoretical_loss": 3.459252263896837, "tokens_seen": 1810563072 }, { "epoch": 0.1, "learning_rate": 9.117316642593484e-05, "loss": 2.6172, "theoretical_loss": 3.4592309741225558, "tokens_seen": 1810694144 }, { "epoch": 0.1, "learning_rate": 9.116514203177661e-05, "loss": 2.6355, "theoretical_loss": 3.4592096863208144, "tokens_seen": 1810825216 }, { "epoch": 0.1, "learning_rate": 9.115711763761836e-05, "loss": 2.6663, "theoretical_loss": 3.4591884004912874, "tokens_seen": 1810956288 }, { "epoch": 0.1, "learning_rate": 9.114909324346013e-05, "loss": 2.4647, "theoretical_loss": 3.4591671166336484, "tokens_seen": 1811087360 }, { "epoch": 0.1, "learning_rate": 9.114106884930188e-05, "loss": 2.5586, "theoretical_loss": 3.4591458347475736, "tokens_seen": 1811218432 }, { "epoch": 0.1, "learning_rate": 9.113304445514365e-05, "loss": 2.6566, "theoretical_loss": 3.4591245548327363, "tokens_seen": 1811349504 }, { "epoch": 0.1, "learning_rate": 9.11250200609854e-05, "loss": 2.4653, "theoretical_loss": 3.4591032768888117, "tokens_seen": 1811480576 }, { "epoch": 0.1, "learning_rate": 9.111699566682716e-05, "loss": 2.5683, "theoretical_loss": 3.459082000915475, "tokens_seen": 1811611648 }, { "epoch": 0.1, "learning_rate": 9.110897127266892e-05, "loss": 2.5607, "theoretical_loss": 3.4590607269124005, "tokens_seen": 1811742720 }, { "epoch": 0.1, "learning_rate": 9.110094687851068e-05, "loss": 2.577, "theoretical_loss": 3.4590394548792642, "tokens_seen": 1811873792 }, { "epoch": 0.1, "learning_rate": 9.109292248435244e-05, "loss": 2.5945, "theoretical_loss": 3.459018184815741, "tokens_seen": 1812004864 }, { "epoch": 0.1, "learning_rate": 9.10848980901942e-05, "loss": 2.5688, "theoretical_loss": 3.458996916721505, "tokens_seen": 1812135936 }, { "epoch": 0.1, "objective/train/docs_used": 999249, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.812991142272949, "objective/train/theoretical_loss": 3.4589862834127683, "objective/train/tokens_used": 182660576, "theoretical_loss": 3.4589862834127683, "tokens_seen": 1812201472 }, { "epoch": 0.1, "learning_rate": 9.107687369603596e-05, "loss": 2.5249, "theoretical_loss": 3.458975650596232, "tokens_seen": 1812267008 }, { "epoch": 0.1, "learning_rate": 9.106884930187772e-05, "loss": 2.6084, "theoretical_loss": 3.458954386439598, "tokens_seen": 1812398080 }, { "epoch": 0.1, "learning_rate": 9.106082490771947e-05, "loss": 2.5619, "theoretical_loss": 3.4589331242512777, "tokens_seen": 1812529152 }, { "epoch": 0.1, "learning_rate": 9.105280051356124e-05, "loss": 2.4964, "theoretical_loss": 3.458911864030947, "tokens_seen": 1812660224 }, { "epoch": 0.1, "learning_rate": 9.104477611940299e-05, "loss": 2.6568, "theoretical_loss": 3.458890605778281, "tokens_seen": 1812791296 }, { "epoch": 0.1, "learning_rate": 9.103675172524476e-05, "loss": 2.5938, "theoretical_loss": 3.4588693494929563, "tokens_seen": 1812922368 }, { "epoch": 0.1, "learning_rate": 9.102872733108651e-05, "loss": 2.7681, "theoretical_loss": 3.4588480951746474, "tokens_seen": 1813053440 }, { "epoch": 0.1, "learning_rate": 9.102070293692828e-05, "loss": 2.5224, "theoretical_loss": 3.4588268428230307, "tokens_seen": 1813184512 }, { "epoch": 0.1, "learning_rate": 9.101267854277003e-05, "loss": 2.6445, "theoretical_loss": 3.458805592437782, "tokens_seen": 1813315584 }, { "epoch": 0.1, "learning_rate": 9.100465414861178e-05, "loss": 2.6035, "theoretical_loss": 3.458784344018577, "tokens_seen": 1813446656 }, { "epoch": 0.1, "learning_rate": 9.099662975445355e-05, "loss": 2.6548, "theoretical_loss": 3.4587630975650923, "tokens_seen": 1813577728 }, { "epoch": 0.1, "learning_rate": 9.09886053602953e-05, "loss": 2.4354, "theoretical_loss": 3.4587418530770035, "tokens_seen": 1813708800 }, { "debugging/Self-BLEU-5": 0.3156689391415616, "debugging/distinct-1-grams": 0.796493726160272, "debugging/distinct-2-grams": 0.9767311252314004, "debugging/entropy-1-grams": 5.133892638116764, "debugging/entropy-2-grams": 5.762233051450052, "debugging/length": 475.0, "debugging/num_segments": 5, "debugging/score": 0.006195986574498804, "debugging/score_std": 0.00502469901082792, "epoch": 0.1, "objective/train/docs_used": 999799, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.311352252960205, "objective/train/theoretical_loss": 3.458720610553987, "objective/train/tokens_used": 184298976, "theoretical_loss": 3.458720610553987, "tokens_seen": 1813839872 }, { "epoch": 0.1, "learning_rate": 9.098058096613707e-05, "loss": 2.6938, "theoretical_loss": 3.458720610553987, "tokens_seen": 1813839872 }, { "epoch": 0.1, "learning_rate": 9.097255657197882e-05, "loss": 2.2563, "theoretical_loss": 3.4586993699957187, "tokens_seen": 1813970944 }, { "epoch": 0.1, "learning_rate": 9.096453217782059e-05, "loss": 2.4608, "theoretical_loss": 3.458678131401876, "tokens_seen": 1814102016 }, { "epoch": 0.1, "learning_rate": 9.095650778366234e-05, "loss": 2.5433, "theoretical_loss": 3.4586568947721337, "tokens_seen": 1814233088 }, { "epoch": 0.1, "learning_rate": 9.094848338950411e-05, "loss": 2.5471, "theoretical_loss": 3.4586356601061694, "tokens_seen": 1814364160 }, { "epoch": 0.1, "learning_rate": 9.094045899534586e-05, "loss": 2.4761, "theoretical_loss": 3.4586144274036594, "tokens_seen": 1814495232 }, { "epoch": 0.1, "learning_rate": 9.093243460118761e-05, "loss": 2.4453, "theoretical_loss": 3.4585931966642804, "tokens_seen": 1814626304 }, { "epoch": 0.1, "learning_rate": 9.092441020702938e-05, "loss": 2.5264, "theoretical_loss": 3.4585719678877087, "tokens_seen": 1814757376 }, { "epoch": 0.1, "learning_rate": 9.091638581287113e-05, "loss": 2.4763, "theoretical_loss": 3.458550741073622, "tokens_seen": 1814888448 }, { "epoch": 0.1, "learning_rate": 9.09083614187129e-05, "loss": 2.5676, "theoretical_loss": 3.458529516221696, "tokens_seen": 1815019520 }, { "epoch": 0.1, "learning_rate": 9.090033702455465e-05, "loss": 2.6084, "theoretical_loss": 3.458508293331609, "tokens_seen": 1815150592 }, { "epoch": 0.1, "learning_rate": 9.089231263039642e-05, "loss": 2.3971, "theoretical_loss": 3.4584870724030363, "tokens_seen": 1815281664 }, { "epoch": 0.1, "learning_rate": 9.088428823623817e-05, "loss": 2.4878, "theoretical_loss": 3.4584658534356567, "tokens_seen": 1815412736 }, { "epoch": 0.1, "objective/train/docs_used": 1000512, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2753448486328125, "objective/train/theoretical_loss": 3.458455244687313, "objective/train/tokens_used": 185937376, "theoretical_loss": 3.458455244687313, "tokens_seen": 1815478272 }, { "epoch": 0.1, "learning_rate": 9.087626384207993e-05, "loss": 2.5719, "theoretical_loss": 3.458444636429146, "tokens_seen": 1815543808 }, { "epoch": 0.1, "learning_rate": 9.08682394479217e-05, "loss": 2.5104, "theoretical_loss": 3.4584234213831824, "tokens_seen": 1815674880 }, { "epoch": 0.1, "learning_rate": 9.086021505376345e-05, "loss": 2.5449, "theoretical_loss": 3.458402208297443, "tokens_seen": 1815805952 }, { "epoch": 0.1, "learning_rate": 9.085219065960521e-05, "loss": 2.3611, "theoretical_loss": 3.458380997171605, "tokens_seen": 1815937024 }, { "epoch": 0.1, "learning_rate": 9.084416626544697e-05, "loss": 2.706, "theoretical_loss": 3.458359788005346, "tokens_seen": 1816068096 }, { "epoch": 0.1, "learning_rate": 9.083614187128873e-05, "loss": 2.609, "theoretical_loss": 3.458338580798343, "tokens_seen": 1816199168 }, { "epoch": 0.1, "learning_rate": 9.082811747713049e-05, "loss": 2.4692, "theoretical_loss": 3.458317375550275, "tokens_seen": 1816330240 }, { "epoch": 0.1, "learning_rate": 9.082009308297224e-05, "loss": 2.5132, "theoretical_loss": 3.4582961722608188, "tokens_seen": 1816461312 }, { "epoch": 0.1, "learning_rate": 9.0812068688814e-05, "loss": 2.6601, "theoretical_loss": 3.4582749709296516, "tokens_seen": 1816592384 }, { "epoch": 0.1, "learning_rate": 9.080404429465576e-05, "loss": 2.4721, "theoretical_loss": 3.4582537715564525, "tokens_seen": 1816723456 }, { "epoch": 0.1, "learning_rate": 9.079601990049753e-05, "loss": 2.5989, "theoretical_loss": 3.458232574140899, "tokens_seen": 1816854528 }, { "epoch": 0.1, "learning_rate": 9.078799550633928e-05, "loss": 2.6828, "theoretical_loss": 3.4582113786826683, "tokens_seen": 1816985600 }, { "epoch": 0.1, "objective/train/docs_used": 1001919, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5981926918029785, "objective/train/theoretical_loss": 3.4581901851814396, "objective/train/tokens_used": 187575776, "theoretical_loss": 3.4581901851814396, "tokens_seen": 1817116672 }, { "epoch": 0.1, "learning_rate": 9.077997111218105e-05, "loss": 2.5979, "theoretical_loss": 3.4581901851814396, "tokens_seen": 1817116672 }, { "epoch": 0.1, "learning_rate": 9.07719467180228e-05, "loss": 2.5523, "theoretical_loss": 3.4581689936368907, "tokens_seen": 1817247744 }, { "epoch": 0.1, "learning_rate": 9.076392232386455e-05, "loss": 2.5556, "theoretical_loss": 3.458147804048699, "tokens_seen": 1817378816 }, { "epoch": 0.1, "learning_rate": 9.075589792970632e-05, "loss": 2.611, "theoretical_loss": 3.4581266164165445, "tokens_seen": 1817509888 }, { "epoch": 0.1, "learning_rate": 9.074787353554807e-05, "loss": 2.4397, "theoretical_loss": 3.4581054307401047, "tokens_seen": 1817640960 }, { "epoch": 0.1, "learning_rate": 9.073984914138984e-05, "loss": 2.5673, "theoretical_loss": 3.458084247019057, "tokens_seen": 1817772032 }, { "epoch": 0.1, "learning_rate": 9.073182474723159e-05, "loss": 2.7918, "theoretical_loss": 3.458063065253082, "tokens_seen": 1817903104 }, { "epoch": 0.1, "learning_rate": 9.072380035307334e-05, "loss": 2.456, "theoretical_loss": 3.4580418854418573, "tokens_seen": 1818034176 }, { "epoch": 0.1, "learning_rate": 9.071577595891511e-05, "loss": 2.5651, "theoretical_loss": 3.458020707585061, "tokens_seen": 1818165248 }, { "epoch": 0.1, "learning_rate": 9.070775156475686e-05, "loss": 2.4939, "theoretical_loss": 3.457999531682373, "tokens_seen": 1818296320 }, { "epoch": 0.1, "learning_rate": 9.069972717059863e-05, "loss": 2.6449, "theoretical_loss": 3.457978357733471, "tokens_seen": 1818427392 }, { "epoch": 0.1, "learning_rate": 9.069170277644038e-05, "loss": 2.6295, "theoretical_loss": 3.457957185738035, "tokens_seen": 1818558464 }, { "epoch": 0.1, "learning_rate": 9.068367838228215e-05, "loss": 2.6178, "theoretical_loss": 3.4579360156957435, "tokens_seen": 1818689536 }, { "epoch": 0.1, "objective/train/docs_used": 1002350, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.173239231109619, "objective/train/theoretical_loss": 3.4579254314069265, "objective/train/tokens_used": 189214176, "theoretical_loss": 3.4579254314069265, "tokens_seen": 1818755072 }, { "epoch": 0.1, "learning_rate": 9.06756539881239e-05, "loss": 2.5297, "theoretical_loss": 3.4579148476062755, "tokens_seen": 1818820608 }, { "epoch": 0.1, "learning_rate": 9.066762959396566e-05, "loss": 2.7194, "theoretical_loss": 3.4578936814693106, "tokens_seen": 1818951680 }, { "epoch": 0.1, "learning_rate": 9.065960519980742e-05, "loss": 2.5981, "theoretical_loss": 3.4578725172845273, "tokens_seen": 1819082752 }, { "epoch": 0.1, "learning_rate": 9.065158080564918e-05, "loss": 2.6458, "theoretical_loss": 3.4578513550516057, "tokens_seen": 1819213824 }, { "epoch": 0.1, "learning_rate": 9.064355641149094e-05, "loss": 2.7472, "theoretical_loss": 3.457830194770225, "tokens_seen": 1819344896 }, { "epoch": 0.1, "learning_rate": 9.06355320173327e-05, "loss": 2.4691, "theoretical_loss": 3.457809036440064, "tokens_seen": 1819475968 }, { "epoch": 0.1, "learning_rate": 9.062750762317445e-05, "loss": 2.6148, "theoretical_loss": 3.4577878800608026, "tokens_seen": 1819607040 }, { "epoch": 0.1, "learning_rate": 9.061948322901622e-05, "loss": 2.5812, "theoretical_loss": 3.457766725632121, "tokens_seen": 1819738112 }, { "epoch": 0.1, "learning_rate": 9.061145883485797e-05, "loss": 2.7097, "theoretical_loss": 3.457745573153698, "tokens_seen": 1819869184 }, { "epoch": 0.1, "learning_rate": 9.060343444069974e-05, "loss": 2.554, "theoretical_loss": 3.4577244226252137, "tokens_seen": 1820000256 }, { "epoch": 0.1, "learning_rate": 9.059541004654149e-05, "loss": 2.6685, "theoretical_loss": 3.4577032740463483, "tokens_seen": 1820131328 }, { "epoch": 0.1, "learning_rate": 9.058738565238326e-05, "loss": 2.4642, "theoretical_loss": 3.4576821274167813, "tokens_seen": 1820262400 }, { "epoch": 0.1, "objective/train/docs_used": 1003363, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7602131366729736, "objective/train/theoretical_loss": 3.4576609827361926, "objective/train/tokens_used": 190852576, "theoretical_loss": 3.4576609827361926, "tokens_seen": 1820393472 }, { "epoch": 0.1, "learning_rate": 9.057936125822501e-05, "loss": 2.6968, "theoretical_loss": 3.4576609827361926, "tokens_seen": 1820393472 }, { "epoch": 0.1, "learning_rate": 9.057133686406676e-05, "loss": 2.5316, "theoretical_loss": 3.457639840004262, "tokens_seen": 1820524544 }, { "epoch": 0.1, "learning_rate": 9.056331246990853e-05, "loss": 2.6249, "theoretical_loss": 3.4576186992206708, "tokens_seen": 1820655616 }, { "epoch": 0.1, "learning_rate": 9.055528807575028e-05, "loss": 2.7948, "theoretical_loss": 3.457597560385098, "tokens_seen": 1820786688 }, { "epoch": 0.1, "learning_rate": 9.054726368159205e-05, "loss": 2.6169, "theoretical_loss": 3.4575764234972253, "tokens_seen": 1820917760 }, { "epoch": 0.1, "learning_rate": 9.05392392874338e-05, "loss": 2.6168, "theoretical_loss": 3.457555288556731, "tokens_seen": 1821048832 }, { "epoch": 0.1, "learning_rate": 9.053121489327556e-05, "loss": 2.6352, "theoretical_loss": 3.4575341555632972, "tokens_seen": 1821179904 }, { "epoch": 0.1, "learning_rate": 9.052319049911732e-05, "loss": 2.5961, "theoretical_loss": 3.457513024516604, "tokens_seen": 1821310976 }, { "epoch": 0.1, "learning_rate": 9.051516610495908e-05, "loss": 2.6443, "theoretical_loss": 3.4574918954163314, "tokens_seen": 1821442048 }, { "epoch": 0.1, "learning_rate": 9.050714171080084e-05, "loss": 2.6054, "theoretical_loss": 3.4574707682621613, "tokens_seen": 1821573120 }, { "epoch": 0.1, "learning_rate": 9.04991173166426e-05, "loss": 2.5453, "theoretical_loss": 3.4574496430537733, "tokens_seen": 1821704192 }, { "epoch": 0.1, "learning_rate": 9.049109292248436e-05, "loss": 2.583, "theoretical_loss": 3.4574285197908483, "tokens_seen": 1821835264 }, { "epoch": 0.1, "learning_rate": 9.048306852832611e-05, "loss": 2.6039, "theoretical_loss": 3.4574073984730678, "tokens_seen": 1821966336 }, { "epoch": 0.1, "objective/train/docs_used": 1003939, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.497610569000244, "objective/train/theoretical_loss": 3.4573968385435068, "objective/train/tokens_used": 192490976, "theoretical_loss": 3.4573968385435068, "tokens_seen": 1822031872 }, { "epoch": 0.1, "learning_rate": 9.047504413416787e-05, "loss": 2.4516, "theoretical_loss": 3.4573862791001124, "tokens_seen": 1822097408 }, { "epoch": 0.1, "learning_rate": 9.046701974000963e-05, "loss": 2.6655, "theoretical_loss": 3.457365161671663, "tokens_seen": 1822228480 }, { "epoch": 0.1, "learning_rate": 9.045899534585139e-05, "loss": 2.5194, "theoretical_loss": 3.457344046187401, "tokens_seen": 1822359552 }, { "epoch": 0.1, "learning_rate": 9.045097095169315e-05, "loss": 2.7839, "theoretical_loss": 3.4573229326470076, "tokens_seen": 1822490624 }, { "epoch": 0.1, "learning_rate": 9.044294655753491e-05, "loss": 2.6435, "theoretical_loss": 3.457301821050164, "tokens_seen": 1822621696 }, { "epoch": 0.1, "learning_rate": 9.043492216337666e-05, "loss": 2.49, "theoretical_loss": 3.4572807113965514, "tokens_seen": 1822752768 }, { "epoch": 0.1, "learning_rate": 9.042689776921843e-05, "loss": 2.6145, "theoretical_loss": 3.457259603685851, "tokens_seen": 1822883840 }, { "epoch": 0.1, "learning_rate": 9.041887337506018e-05, "loss": 2.575, "theoretical_loss": 3.4572384979177455, "tokens_seen": 1823014912 }, { "epoch": 0.1, "learning_rate": 9.041084898090195e-05, "loss": 2.5775, "theoretical_loss": 3.457217394091915, "tokens_seen": 1823145984 }, { "epoch": 0.11, "learning_rate": 9.04028245867437e-05, "loss": 2.7332, "theoretical_loss": 3.457196292208042, "tokens_seen": 1823277056 }, { "epoch": 0.11, "learning_rate": 9.039480019258547e-05, "loss": 2.5805, "theoretical_loss": 3.457175192265808, "tokens_seen": 1823408128 }, { "epoch": 0.11, "learning_rate": 9.038677579842722e-05, "loss": 2.6114, "theoretical_loss": 3.4571540942648946, "tokens_seen": 1823539200 }, { "epoch": 0.11, "objective/train/docs_used": 1005294, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7300758361816406, "objective/train/theoretical_loss": 3.457132998204984, "objective/train/tokens_used": 194129376, "theoretical_loss": 3.457132998204984, "tokens_seen": 1823670272 }, { "epoch": 0.11, "learning_rate": 9.037875140426897e-05, "loss": 2.4603, "theoretical_loss": 3.457132998204984, "tokens_seen": 1823670272 }, { "epoch": 0.11, "learning_rate": 9.037072701011074e-05, "loss": 2.6913, "theoretical_loss": 3.4571119040857576, "tokens_seen": 1823801344 }, { "epoch": 0.11, "learning_rate": 9.036270261595249e-05, "loss": 2.5986, "theoretical_loss": 3.457090811906898, "tokens_seen": 1823932416 }, { "epoch": 0.11, "learning_rate": 9.035467822179426e-05, "loss": 2.6846, "theoretical_loss": 3.457069721668087, "tokens_seen": 1824063488 }, { "epoch": 0.11, "learning_rate": 9.034665382763601e-05, "loss": 2.4226, "theoretical_loss": 3.4570486333690065, "tokens_seen": 1824194560 }, { "epoch": 0.11, "learning_rate": 9.033862943347777e-05, "loss": 2.5405, "theoretical_loss": 3.4570275470093397, "tokens_seen": 1824325632 }, { "epoch": 0.11, "learning_rate": 9.033060503931953e-05, "loss": 2.5503, "theoretical_loss": 3.457006462588768, "tokens_seen": 1824456704 }, { "epoch": 0.11, "learning_rate": 9.032258064516129e-05, "loss": 2.6584, "theoretical_loss": 3.4569853801069743, "tokens_seen": 1824587776 }, { "epoch": 0.11, "learning_rate": 9.031455625100305e-05, "loss": 2.5576, "theoretical_loss": 3.4569642995636407, "tokens_seen": 1824718848 }, { "epoch": 0.11, "learning_rate": 9.03065318568448e-05, "loss": 2.5872, "theoretical_loss": 3.4569432209584496, "tokens_seen": 1824849920 }, { "epoch": 0.11, "learning_rate": 9.029850746268657e-05, "loss": 2.5797, "theoretical_loss": 3.4569221442910845, "tokens_seen": 1824980992 }, { "epoch": 0.11, "learning_rate": 9.029048306852833e-05, "loss": 2.764, "theoretical_loss": 3.4569010695612272, "tokens_seen": 1825112064 }, { "epoch": 0.11, "learning_rate": 9.028245867437008e-05, "loss": 2.614, "theoretical_loss": 3.456879996768561, "tokens_seen": 1825243136 }, { "epoch": 0.11, "objective/train/docs_used": 1005912, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 1.9645118713378906, "objective/train/theoretical_loss": 3.4568694610985755, "objective/train/tokens_used": 195767776, "theoretical_loss": 3.4568694610985755, "tokens_seen": 1825308672 }, { "epoch": 0.11, "learning_rate": 9.027443428021185e-05, "loss": 2.4857, "theoretical_loss": 3.456858925912768, "tokens_seen": 1825374208 }, { "epoch": 0.11, "learning_rate": 9.02664098860536e-05, "loss": 2.4868, "theoretical_loss": 3.4568378569935323, "tokens_seen": 1825505280 }, { "epoch": 0.11, "learning_rate": 9.025838549189537e-05, "loss": 2.5564, "theoretical_loss": 3.456816790010536, "tokens_seen": 1825636352 }, { "epoch": 0.11, "learning_rate": 9.025036109773712e-05, "loss": 2.5284, "theoretical_loss": 3.456795724963462, "tokens_seen": 1825767424 }, { "epoch": 0.11, "learning_rate": 9.024233670357888e-05, "loss": 2.6118, "theoretical_loss": 3.456774661851995, "tokens_seen": 1825898496 }, { "epoch": 0.11, "learning_rate": 9.023431230942064e-05, "loss": 2.5575, "theoretical_loss": 3.4567536006758157, "tokens_seen": 1826029568 }, { "epoch": 0.11, "learning_rate": 9.022628791526239e-05, "loss": 2.5875, "theoretical_loss": 3.4567325414346093, "tokens_seen": 1826160640 }, { "epoch": 0.11, "learning_rate": 9.021826352110416e-05, "loss": 2.5273, "theoretical_loss": 3.456711484128059, "tokens_seen": 1826291712 }, { "epoch": 0.11, "learning_rate": 9.021023912694591e-05, "loss": 2.6429, "theoretical_loss": 3.456690428755847, "tokens_seen": 1826422784 }, { "epoch": 0.11, "learning_rate": 9.020221473278768e-05, "loss": 2.6974, "theoretical_loss": 3.4566693753176576, "tokens_seen": 1826553856 }, { "epoch": 0.11, "learning_rate": 9.019419033862943e-05, "loss": 2.4875, "theoretical_loss": 3.4566483238131753, "tokens_seen": 1826684928 }, { "epoch": 0.11, "learning_rate": 9.01861659444712e-05, "loss": 2.562, "theoretical_loss": 3.4566272742420825, "tokens_seen": 1826816000 }, { "epoch": 0.11, "objective/train/docs_used": 1007268, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4686145782470703, "objective/train/theoretical_loss": 3.4566062266040634, "objective/train/tokens_used": 197406176, "theoretical_loss": 3.4566062266040634, "tokens_seen": 1826947072 }, { "epoch": 0.11, "learning_rate": 9.017814155031295e-05, "loss": 2.4638, "theoretical_loss": 3.4566062266040634, "tokens_seen": 1826947072 }, { "epoch": 0.11, "learning_rate": 9.01701171561547e-05, "loss": 2.4429, "theoretical_loss": 3.4565851808988013, "tokens_seen": 1827078144 }, { "epoch": 0.11, "learning_rate": 9.016209276199647e-05, "loss": 2.7966, "theoretical_loss": 3.456564137125981, "tokens_seen": 1827209216 }, { "epoch": 0.11, "learning_rate": 9.015406836783822e-05, "loss": 2.4905, "theoretical_loss": 3.4565430952852862, "tokens_seen": 1827340288 }, { "epoch": 0.11, "learning_rate": 9.014604397367999e-05, "loss": 2.517, "theoretical_loss": 3.4565220553764004, "tokens_seen": 1827471360 }, { "epoch": 0.11, "learning_rate": 9.013801957952174e-05, "loss": 2.5002, "theoretical_loss": 3.456501017399008, "tokens_seen": 1827602432 }, { "epoch": 0.11, "learning_rate": 9.012999518536351e-05, "loss": 2.5377, "theoretical_loss": 3.4564799813527936, "tokens_seen": 1827733504 }, { "epoch": 0.11, "learning_rate": 9.012197079120526e-05, "loss": 2.5164, "theoretical_loss": 3.4564589472374405, "tokens_seen": 1827864576 }, { "epoch": 0.11, "learning_rate": 9.011394639704702e-05, "loss": 2.6276, "theoretical_loss": 3.456437915052634, "tokens_seen": 1827995648 }, { "epoch": 0.11, "learning_rate": 9.010592200288878e-05, "loss": 2.5136, "theoretical_loss": 3.456416884798058, "tokens_seen": 1828126720 }, { "epoch": 0.11, "learning_rate": 9.009789760873054e-05, "loss": 2.4933, "theoretical_loss": 3.456395856473397, "tokens_seen": 1828257792 }, { "epoch": 0.11, "learning_rate": 9.00898732145723e-05, "loss": 2.5778, "theoretical_loss": 3.456374830078336, "tokens_seen": 1828388864 }, { "epoch": 0.11, "learning_rate": 9.008184882041406e-05, "loss": 2.4973, "theoretical_loss": 3.456353805612559, "tokens_seen": 1828519936 }, { "epoch": 0.11, "objective/train/docs_used": 1007941, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.821141481399536, "objective/train/theoretical_loss": 3.4563432941030534, "objective/train/tokens_used": 199044576, "theoretical_loss": 3.4563432941030534, "tokens_seen": 1828585472 }, { "epoch": 0.11, "learning_rate": 9.007382442625582e-05, "loss": 2.5797, "theoretical_loss": 3.4563327830757506, "tokens_seen": 1828651008 }, { "epoch": 0.11, "learning_rate": 9.006580003209758e-05, "loss": 2.6534, "theoretical_loss": 3.4563117624675965, "tokens_seen": 1828782080 }, { "epoch": 0.11, "learning_rate": 9.005777563793933e-05, "loss": 2.4451, "theoretical_loss": 3.456290743787781, "tokens_seen": 1828913152 }, { "epoch": 0.11, "learning_rate": 9.00497512437811e-05, "loss": 2.5534, "theoretical_loss": 3.4562697270359886, "tokens_seen": 1829044224 }, { "epoch": 0.11, "learning_rate": 9.004172684962285e-05, "loss": 2.6204, "theoretical_loss": 3.456248712211905, "tokens_seen": 1829175296 }, { "epoch": 0.11, "learning_rate": 9.003370245546462e-05, "loss": 2.5441, "theoretical_loss": 3.456227699315215, "tokens_seen": 1829306368 }, { "epoch": 0.11, "learning_rate": 9.002567806130637e-05, "loss": 2.5198, "theoretical_loss": 3.4562066883456035, "tokens_seen": 1829437440 }, { "epoch": 0.11, "learning_rate": 9.001765366714814e-05, "loss": 2.3917, "theoretical_loss": 3.4561856793027568, "tokens_seen": 1829568512 }, { "epoch": 0.11, "learning_rate": 9.000962927298989e-05, "loss": 2.5708, "theoretical_loss": 3.4561646721863584, "tokens_seen": 1829699584 }, { "epoch": 0.11, "learning_rate": 9.000160487883166e-05, "loss": 2.6934, "theoretical_loss": 3.4561436669960948, "tokens_seen": 1829830656 }, { "epoch": 0.11, "learning_rate": 8.999358048467341e-05, "loss": 2.5966, "theoretical_loss": 3.4561226637316516, "tokens_seen": 1829961728 }, { "epoch": 0.11, "learning_rate": 8.998555609051516e-05, "loss": 2.5411, "theoretical_loss": 3.456101662392714, "tokens_seen": 1830092800 }, { "epoch": 0.11, "objective/train/docs_used": 1009247, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.360668659210205, "objective/train/theoretical_loss": 3.456080662978967, "objective/train/tokens_used": 200682976, "theoretical_loss": 3.456080662978967, "tokens_seen": 1830223872 }, { "epoch": 0.11, "learning_rate": 8.997753169635693e-05, "loss": 2.4355, "theoretical_loss": 3.456080662978967, "tokens_seen": 1830223872 }, { "epoch": 0.11, "learning_rate": 8.996950730219868e-05, "loss": 2.5936, "theoretical_loss": 3.4560596654900975, "tokens_seen": 1830354944 }, { "epoch": 0.11, "learning_rate": 8.996148290804045e-05, "loss": 2.5363, "theoretical_loss": 3.45603866992579, "tokens_seen": 1830486016 }, { "epoch": 0.11, "learning_rate": 8.99534585138822e-05, "loss": 2.514, "theoretical_loss": 3.4560176762857315, "tokens_seen": 1830617088 }, { "epoch": 0.11, "learning_rate": 8.994543411972397e-05, "loss": 2.4908, "theoretical_loss": 3.455996684569607, "tokens_seen": 1830748160 }, { "epoch": 0.11, "learning_rate": 8.993740972556572e-05, "loss": 2.5477, "theoretical_loss": 3.4559756947771025, "tokens_seen": 1830879232 }, { "epoch": 0.11, "learning_rate": 8.992938533140747e-05, "loss": 2.4969, "theoretical_loss": 3.4559547069079044, "tokens_seen": 1831010304 }, { "epoch": 0.11, "learning_rate": 8.992136093724924e-05, "loss": 2.5748, "theoretical_loss": 3.4559337209616987, "tokens_seen": 1831141376 }, { "epoch": 0.11, "learning_rate": 8.9913336543091e-05, "loss": 2.5423, "theoretical_loss": 3.4559127369381715, "tokens_seen": 1831272448 }, { "epoch": 0.11, "learning_rate": 8.990531214893276e-05, "loss": 2.7193, "theoretical_loss": 3.455891754837009, "tokens_seen": 1831403520 }, { "epoch": 0.11, "learning_rate": 8.989728775477451e-05, "loss": 2.4593, "theoretical_loss": 3.455870774657898, "tokens_seen": 1831534592 }, { "epoch": 0.11, "learning_rate": 8.988926336061628e-05, "loss": 2.6618, "theoretical_loss": 3.4558497964005244, "tokens_seen": 1831665664 }, { "epoch": 0.11, "learning_rate": 8.988123896645803e-05, "loss": 2.5454, "theoretical_loss": 3.455828820064575, "tokens_seen": 1831796736 }, { "epoch": 0.11, "objective/train/docs_used": 1009780, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6938259601593018, "objective/train/theoretical_loss": 3.455818332617036, "objective/train/tokens_used": 202321376, "theoretical_loss": 3.455818332617036, "tokens_seen": 1831862272 }, { "epoch": 0.11, "learning_rate": 8.987321457229979e-05, "loss": 2.6755, "theoretical_loss": 3.455807845649735, "tokens_seen": 1831927808 }, { "epoch": 0.11, "learning_rate": 8.986519017814155e-05, "loss": 2.6631, "theoretical_loss": 3.4557868731556933, "tokens_seen": 1832058880 }, { "epoch": 0.11, "learning_rate": 8.98571657839833e-05, "loss": 2.7067, "theoretical_loss": 3.4557659025821352, "tokens_seen": 1832189952 }, { "epoch": 0.11, "learning_rate": 8.984914138982507e-05, "loss": 2.4591, "theoretical_loss": 3.455744933928748, "tokens_seen": 1832321024 }, { "epoch": 0.11, "learning_rate": 8.984111699566683e-05, "loss": 2.3704, "theoretical_loss": 3.455723967195218, "tokens_seen": 1832452096 }, { "epoch": 0.11, "learning_rate": 8.983309260150859e-05, "loss": 2.5915, "theoretical_loss": 3.4557030023812323, "tokens_seen": 1832583168 }, { "epoch": 0.11, "learning_rate": 8.982506820735035e-05, "loss": 2.576, "theoretical_loss": 3.455682039486478, "tokens_seen": 1832714240 }, { "epoch": 0.11, "learning_rate": 8.98170438131921e-05, "loss": 2.5794, "theoretical_loss": 3.4556610785106425, "tokens_seen": 1832845312 }, { "epoch": 0.11, "learning_rate": 8.980901941903387e-05, "loss": 2.6461, "theoretical_loss": 3.4556401194534123, "tokens_seen": 1832976384 }, { "epoch": 0.11, "learning_rate": 8.980099502487562e-05, "loss": 2.632, "theoretical_loss": 3.4556191623144747, "tokens_seen": 1833107456 }, { "epoch": 0.11, "learning_rate": 8.979297063071739e-05, "loss": 2.6446, "theoretical_loss": 3.4555982070935176, "tokens_seen": 1833238528 }, { "epoch": 0.11, "learning_rate": 8.978494623655914e-05, "loss": 2.5226, "theoretical_loss": 3.455577253790228, "tokens_seen": 1833369600 }, { "epoch": 0.11, "objective/train/docs_used": 1010758, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.585409164428711, "objective/train/theoretical_loss": 3.455556302404293, "objective/train/tokens_used": 203959776, "theoretical_loss": 3.455556302404293, "tokens_seen": 1833500672 }, { "epoch": 0.11, "learning_rate": 8.97769218424009e-05, "loss": 2.5823, "theoretical_loss": 3.455556302404293, "tokens_seen": 1833500672 }, { "epoch": 0.11, "learning_rate": 8.976889744824266e-05, "loss": 2.422, "theoretical_loss": 3.4555353529354003, "tokens_seen": 1833631744 }, { "epoch": 0.11, "learning_rate": 8.976087305408441e-05, "loss": 2.6723, "theoretical_loss": 3.455514405383237, "tokens_seen": 1833762816 }, { "epoch": 0.11, "learning_rate": 8.975284865992618e-05, "loss": 2.5495, "theoretical_loss": 3.4554934597474922, "tokens_seen": 1833893888 }, { "epoch": 0.11, "learning_rate": 8.974482426576793e-05, "loss": 2.5599, "theoretical_loss": 3.4554725160278528, "tokens_seen": 1834024960 }, { "epoch": 0.11, "learning_rate": 8.97367998716097e-05, "loss": 2.482, "theoretical_loss": 3.4554515742240057, "tokens_seen": 1834156032 }, { "epoch": 0.11, "learning_rate": 8.972877547745145e-05, "loss": 2.6911, "theoretical_loss": 3.45543063433564, "tokens_seen": 1834287104 }, { "epoch": 0.11, "learning_rate": 8.972075108329322e-05, "loss": 2.6872, "theoretical_loss": 3.455409696362443, "tokens_seen": 1834418176 }, { "epoch": 0.11, "learning_rate": 8.971272668913497e-05, "loss": 2.5098, "theoretical_loss": 3.4553887603041034, "tokens_seen": 1834549248 }, { "epoch": 0.11, "learning_rate": 8.970470229497674e-05, "loss": 2.7034, "theoretical_loss": 3.4553678261603085, "tokens_seen": 1834680320 }, { "epoch": 0.11, "learning_rate": 8.969667790081849e-05, "loss": 2.482, "theoretical_loss": 3.455346893930747, "tokens_seen": 1834811392 }, { "epoch": 0.11, "learning_rate": 8.968865350666024e-05, "loss": 2.5007, "theoretical_loss": 3.4553259636151066, "tokens_seen": 1834942464 }, { "epoch": 0.11, "learning_rate": 8.968062911250201e-05, "loss": 2.5495, "theoretical_loss": 3.455305035213076, "tokens_seen": 1835073536 }, { "epoch": 0.11, "objective/train/docs_used": 1011050, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.749027729034424, "objective/train/theoretical_loss": 3.455294571729567, "objective/train/tokens_used": 205598176, "theoretical_loss": 3.455294571729567, "tokens_seen": 1835139072 }, { "epoch": 0.11, "learning_rate": 8.967260471834376e-05, "loss": 2.7461, "theoretical_loss": 3.455284108724344, "tokens_seen": 1835204608 }, { "epoch": 0.11, "learning_rate": 8.966458032418553e-05, "loss": 2.8043, "theoretical_loss": 3.455263184148598, "tokens_seen": 1835335680 }, { "epoch": 0.11, "learning_rate": 8.965655593002728e-05, "loss": 2.5974, "theoretical_loss": 3.455242261485527, "tokens_seen": 1835466752 }, { "epoch": 0.11, "learning_rate": 8.964853153586905e-05, "loss": 2.4865, "theoretical_loss": 3.45522134073482, "tokens_seen": 1835597824 }, { "epoch": 0.11, "learning_rate": 8.96405071417108e-05, "loss": 2.6225, "theoretical_loss": 3.455200421896165, "tokens_seen": 1835728896 }, { "epoch": 0.11, "learning_rate": 8.963248274755256e-05, "loss": 2.6943, "theoretical_loss": 3.4551795049692515, "tokens_seen": 1835859968 }, { "epoch": 0.11, "learning_rate": 8.962445835339432e-05, "loss": 2.5456, "theoretical_loss": 3.455158589953768, "tokens_seen": 1835991040 }, { "epoch": 0.11, "learning_rate": 8.961643395923608e-05, "loss": 2.5191, "theoretical_loss": 3.455137676849403, "tokens_seen": 1836122112 }, { "epoch": 0.11, "learning_rate": 8.960840956507784e-05, "loss": 2.6647, "theoretical_loss": 3.4551167656558457, "tokens_seen": 1836253184 }, { "epoch": 0.11, "learning_rate": 8.96003851709196e-05, "loss": 2.5848, "theoretical_loss": 3.4550958563727856, "tokens_seen": 1836384256 }, { "epoch": 0.11, "learning_rate": 8.959236077676136e-05, "loss": 2.5178, "theoretical_loss": 3.4550749489999113, "tokens_seen": 1836515328 }, { "epoch": 0.11, "learning_rate": 8.958433638260312e-05, "loss": 2.5931, "theoretical_loss": 3.455054043536912, "tokens_seen": 1836646400 }, { "epoch": 0.11, "objective/train/docs_used": 1012285, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6171936988830566, "objective/train/theoretical_loss": 3.4550331399834766, "objective/train/tokens_used": 207236576, "theoretical_loss": 3.4550331399834766, "tokens_seen": 1836777472 }, { "epoch": 0.11, "learning_rate": 8.957631198844487e-05, "loss": 2.5262, "theoretical_loss": 3.4550331399834766, "tokens_seen": 1836777472 }, { "epoch": 0.11, "learning_rate": 8.956828759428664e-05, "loss": 2.6416, "theoretical_loss": 3.455012238339296, "tokens_seen": 1836908544 }, { "epoch": 0.11, "learning_rate": 8.956026320012839e-05, "loss": 2.5346, "theoretical_loss": 3.4549913386040574, "tokens_seen": 1837039616 }, { "epoch": 0.11, "learning_rate": 8.955223880597016e-05, "loss": 2.6475, "theoretical_loss": 3.4549704407774517, "tokens_seen": 1837170688 }, { "epoch": 0.11, "learning_rate": 8.954421441181191e-05, "loss": 2.7004, "theoretical_loss": 3.4549495448591685, "tokens_seen": 1837301760 }, { "epoch": 0.11, "learning_rate": 8.953619001765368e-05, "loss": 2.6106, "theoretical_loss": 3.4549286508488963, "tokens_seen": 1837432832 }, { "epoch": 0.11, "learning_rate": 8.952816562349543e-05, "loss": 2.5996, "theoretical_loss": 3.4549077587463257, "tokens_seen": 1837563904 }, { "epoch": 0.11, "learning_rate": 8.952014122933718e-05, "loss": 2.8679, "theoretical_loss": 3.4548868685511467, "tokens_seen": 1837694976 }, { "epoch": 0.11, "learning_rate": 8.951211683517895e-05, "loss": 2.6206, "theoretical_loss": 3.4548659802630484, "tokens_seen": 1837826048 }, { "epoch": 0.11, "learning_rate": 8.95040924410207e-05, "loss": 2.4902, "theoretical_loss": 3.4548450938817212, "tokens_seen": 1837957120 }, { "epoch": 0.11, "learning_rate": 8.949606804686247e-05, "loss": 2.5685, "theoretical_loss": 3.454824209406855, "tokens_seen": 1838088192 }, { "epoch": 0.11, "learning_rate": 8.948804365270422e-05, "loss": 2.6072, "theoretical_loss": 3.4548033268381397, "tokens_seen": 1838219264 }, { "epoch": 0.11, "learning_rate": 8.948001925854599e-05, "loss": 2.5037, "theoretical_loss": 3.4547824461752654, "tokens_seen": 1838350336 }, { "epoch": 0.11, "objective/train/docs_used": 1012947, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6320648193359375, "objective/train/theoretical_loss": 3.4547720065584215, "objective/train/tokens_used": 208874976, "theoretical_loss": 3.4547720065584215, "tokens_seen": 1838415872 }, { "epoch": 0.11, "learning_rate": 8.947199486438774e-05, "loss": 2.6496, "theoretical_loss": 3.454761567417922, "tokens_seen": 1838481408 }, { "epoch": 0.11, "learning_rate": 8.94639704702295e-05, "loss": 2.6935, "theoretical_loss": 3.4547406905658002, "tokens_seen": 1838612480 }, { "epoch": 0.11, "learning_rate": 8.945594607607126e-05, "loss": 2.5356, "theoretical_loss": 3.454719815618591, "tokens_seen": 1838743552 }, { "epoch": 0.11, "learning_rate": 8.944792168191301e-05, "loss": 2.5054, "theoretical_loss": 3.4546989425759835, "tokens_seen": 1838874624 }, { "epoch": 0.11, "learning_rate": 8.943989728775478e-05, "loss": 2.663, "theoretical_loss": 3.4546780714376695, "tokens_seen": 1839005696 }, { "epoch": 0.11, "learning_rate": 8.943187289359653e-05, "loss": 2.5547, "theoretical_loss": 3.4546572022033377, "tokens_seen": 1839136768 }, { "epoch": 0.11, "learning_rate": 8.94238484994383e-05, "loss": 2.3293, "theoretical_loss": 3.4546363348726805, "tokens_seen": 1839267840 }, { "epoch": 0.11, "learning_rate": 8.941582410528005e-05, "loss": 2.6209, "theoretical_loss": 3.4546154694453883, "tokens_seen": 1839398912 }, { "epoch": 0.11, "learning_rate": 8.94077997111218e-05, "loss": 2.5797, "theoretical_loss": 3.454594605921151, "tokens_seen": 1839529984 }, { "epoch": 0.11, "learning_rate": 8.939977531696357e-05, "loss": 2.6429, "theoretical_loss": 3.454573744299661, "tokens_seen": 1839661056 }, { "epoch": 0.12, "learning_rate": 8.939175092280533e-05, "loss": 2.6145, "theoretical_loss": 3.454552884580607, "tokens_seen": 1839792128 }, { "epoch": 0.12, "learning_rate": 8.93837265286471e-05, "loss": 2.5151, "theoretical_loss": 3.454532026763682, "tokens_seen": 1839923200 }, { "epoch": 0.12, "objective/train/docs_used": 1014005, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.342764139175415, "objective/train/theoretical_loss": 3.4545111708485763, "objective/train/tokens_used": 210513376, "theoretical_loss": 3.4545111708485763, "tokens_seen": 1840054272 }, { "epoch": 0.12, "learning_rate": 8.937570213448885e-05, "loss": 2.5643, "theoretical_loss": 3.4545111708485763, "tokens_seen": 1840054272 }, { "epoch": 0.12, "learning_rate": 8.936767774033061e-05, "loss": 2.7472, "theoretical_loss": 3.4544903168349808, "tokens_seen": 1840185344 }, { "epoch": 0.12, "learning_rate": 8.935965334617237e-05, "loss": 2.6095, "theoretical_loss": 3.454469464722587, "tokens_seen": 1840316416 }, { "epoch": 0.12, "learning_rate": 8.935162895201413e-05, "loss": 2.5621, "theoretical_loss": 3.4544486145110858, "tokens_seen": 1840447488 }, { "epoch": 0.12, "learning_rate": 8.934360455785589e-05, "loss": 2.4646, "theoretical_loss": 3.4544277662001694, "tokens_seen": 1840578560 }, { "epoch": 0.12, "learning_rate": 8.933558016369764e-05, "loss": 2.5613, "theoretical_loss": 3.4544069197895286, "tokens_seen": 1840709632 }, { "epoch": 0.12, "learning_rate": 8.93275557695394e-05, "loss": 2.6725, "theoretical_loss": 3.454386075278855, "tokens_seen": 1840840704 }, { "epoch": 0.12, "learning_rate": 8.931953137538116e-05, "loss": 2.6032, "theoretical_loss": 3.4543652326678402, "tokens_seen": 1840971776 }, { "epoch": 0.12, "learning_rate": 8.931150698122293e-05, "loss": 2.5783, "theoretical_loss": 3.454344391956176, "tokens_seen": 1841102848 }, { "epoch": 0.12, "learning_rate": 8.930348258706468e-05, "loss": 2.541, "theoretical_loss": 3.4543235531435537, "tokens_seen": 1841233920 }, { "epoch": 0.12, "learning_rate": 8.929545819290645e-05, "loss": 2.5714, "theoretical_loss": 3.454302716229665, "tokens_seen": 1841364992 }, { "epoch": 0.12, "learning_rate": 8.92874337987482e-05, "loss": 2.6278, "theoretical_loss": 3.454281881214203, "tokens_seen": 1841496064 }, { "epoch": 0.12, "learning_rate": 8.927940940458995e-05, "loss": 2.5964, "theoretical_loss": 3.454261048096858, "tokens_seen": 1841627136 }, { "epoch": 0.12, "objective/train/docs_used": 1015263, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.241819381713867, "objective/train/theoretical_loss": 3.4542506322498836, "objective/train/tokens_used": 212151776, "theoretical_loss": 3.4542506322498836, "tokens_seen": 1841692672 }, { "epoch": 0.12, "learning_rate": 8.927138501043172e-05, "loss": 2.6414, "theoretical_loss": 3.454240216877323, "tokens_seen": 1841758208 }, { "epoch": 0.12, "learning_rate": 8.926336061627347e-05, "loss": 2.5861, "theoretical_loss": 3.45421938755529, "tokens_seen": 1841889280 }, { "epoch": 0.12, "learning_rate": 8.925533622211524e-05, "loss": 2.6131, "theoretical_loss": 3.4541985601304503, "tokens_seen": 1842020352 }, { "epoch": 0.12, "learning_rate": 8.924731182795699e-05, "loss": 2.6827, "theoretical_loss": 3.4541777346024976, "tokens_seen": 1842151424 }, { "epoch": 0.12, "learning_rate": 8.923928743379876e-05, "loss": 2.583, "theoretical_loss": 3.4541569109711228, "tokens_seen": 1842282496 }, { "epoch": 0.12, "learning_rate": 8.923126303964051e-05, "loss": 2.5736, "theoretical_loss": 3.4541360892360187, "tokens_seen": 1842413568 }, { "epoch": 0.12, "learning_rate": 8.922323864548226e-05, "loss": 2.6615, "theoretical_loss": 3.4541152693968784, "tokens_seen": 1842544640 }, { "epoch": 0.12, "learning_rate": 8.921521425132403e-05, "loss": 2.6997, "theoretical_loss": 3.4540944514533933, "tokens_seen": 1842675712 }, { "epoch": 0.12, "learning_rate": 8.920718985716578e-05, "loss": 2.5494, "theoretical_loss": 3.4540736354052575, "tokens_seen": 1842806784 }, { "epoch": 0.12, "learning_rate": 8.919916546300755e-05, "loss": 2.6537, "theoretical_loss": 3.4540528212521617, "tokens_seen": 1842937856 }, { "epoch": 0.12, "learning_rate": 8.91911410688493e-05, "loss": 2.5037, "theoretical_loss": 3.4540320089938, "tokens_seen": 1843068928 }, { "epoch": 0.12, "learning_rate": 8.918311667469107e-05, "loss": 2.5411, "theoretical_loss": 3.454011198629865, "tokens_seen": 1843200000 }, { "epoch": 0.12, "objective/train/docs_used": 1015948, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7754783630371094, "objective/train/theoretical_loss": 3.453990390160049, "objective/train/tokens_used": 213790176, "theoretical_loss": 3.453990390160049, "tokens_seen": 1843331072 }, { "epoch": 0.12, "learning_rate": 8.917509228053282e-05, "loss": 2.6023, "theoretical_loss": 3.453990390160049, "tokens_seen": 1843331072 }, { "epoch": 0.12, "learning_rate": 8.916706788637458e-05, "loss": 2.6114, "theoretical_loss": 3.453969583584045, "tokens_seen": 1843462144 }, { "epoch": 0.12, "learning_rate": 8.915904349221634e-05, "loss": 2.6495, "theoretical_loss": 3.453948778901547, "tokens_seen": 1843593216 }, { "epoch": 0.12, "learning_rate": 8.91510190980581e-05, "loss": 2.5029, "theoretical_loss": 3.4539279761122472, "tokens_seen": 1843724288 }, { "epoch": 0.12, "learning_rate": 8.914299470389986e-05, "loss": 2.4838, "theoretical_loss": 3.453907175215839, "tokens_seen": 1843855360 }, { "epoch": 0.12, "learning_rate": 8.913497030974162e-05, "loss": 2.468, "theoretical_loss": 3.453886376212015, "tokens_seen": 1843986432 }, { "epoch": 0.12, "learning_rate": 8.912694591558338e-05, "loss": 2.4687, "theoretical_loss": 3.4538655791004693, "tokens_seen": 1844117504 }, { "epoch": 0.12, "learning_rate": 8.911892152142514e-05, "loss": 2.4653, "theoretical_loss": 3.4538447838808954, "tokens_seen": 1844248576 }, { "epoch": 0.12, "learning_rate": 8.911089712726689e-05, "loss": 2.7463, "theoretical_loss": 3.453823990552986, "tokens_seen": 1844379648 }, { "epoch": 0.12, "learning_rate": 8.910287273310866e-05, "loss": 2.459, "theoretical_loss": 3.453803199116435, "tokens_seen": 1844510720 }, { "epoch": 0.12, "learning_rate": 8.909484833895041e-05, "loss": 2.4994, "theoretical_loss": 3.4537824095709366, "tokens_seen": 1844641792 }, { "epoch": 0.12, "learning_rate": 8.908682394479218e-05, "loss": 2.4943, "theoretical_loss": 3.4537616219161835, "tokens_seen": 1844772864 }, { "epoch": 0.12, "learning_rate": 8.907879955063393e-05, "loss": 2.5477, "theoretical_loss": 3.4537408361518693, "tokens_seen": 1844903936 }, { "epoch": 0.12, "objective/train/docs_used": 1017203, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5314950942993164, "objective/train/theoretical_loss": 3.4537304439785315, "objective/train/tokens_used": 215428576, "theoretical_loss": 3.4537304439785315, "tokens_seen": 1844969472 }, { "epoch": 0.12, "learning_rate": 8.90707751564757e-05, "loss": 2.4796, "theoretical_loss": 3.4537200522776885, "tokens_seen": 1845035008 }, { "epoch": 0.12, "learning_rate": 8.906275076231745e-05, "loss": 2.7601, "theoretical_loss": 3.4536992702933347, "tokens_seen": 1845166080 }, { "epoch": 0.12, "learning_rate": 8.905472636815922e-05, "loss": 2.4913, "theoretical_loss": 3.4536784901985014, "tokens_seen": 1845297152 }, { "epoch": 0.12, "learning_rate": 8.904670197400097e-05, "loss": 2.5899, "theoretical_loss": 3.4536577119928835, "tokens_seen": 1845428224 }, { "epoch": 0.12, "learning_rate": 8.903867757984272e-05, "loss": 2.3805, "theoretical_loss": 3.4536369356761742, "tokens_seen": 1845559296 }, { "epoch": 0.12, "learning_rate": 8.903065318568449e-05, "loss": 2.6224, "theoretical_loss": 3.4536161612480685, "tokens_seen": 1845690368 }, { "epoch": 0.12, "learning_rate": 8.902262879152624e-05, "loss": 2.5781, "theoretical_loss": 3.4535953887082598, "tokens_seen": 1845821440 }, { "epoch": 0.12, "learning_rate": 8.901460439736801e-05, "loss": 2.5903, "theoretical_loss": 3.453574618056443, "tokens_seen": 1845952512 }, { "epoch": 0.12, "learning_rate": 8.900658000320976e-05, "loss": 2.5429, "theoretical_loss": 3.4535538492923123, "tokens_seen": 1846083584 }, { "epoch": 0.12, "learning_rate": 8.899855560905153e-05, "loss": 2.5974, "theoretical_loss": 3.4535330824155617, "tokens_seen": 1846214656 }, { "epoch": 0.12, "learning_rate": 8.899053121489328e-05, "loss": 2.464, "theoretical_loss": 3.4535123174258864, "tokens_seen": 1846345728 }, { "epoch": 0.12, "learning_rate": 8.898250682073503e-05, "loss": 2.6167, "theoretical_loss": 3.4534915543229805, "tokens_seen": 1846476800 }, { "epoch": 0.12, "objective/train/docs_used": 1017618, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7826731204986572, "objective/train/theoretical_loss": 3.453470793106539, "objective/train/tokens_used": 217066976, "theoretical_loss": 3.453470793106539, "tokens_seen": 1846607872 }, { "epoch": 0.12, "learning_rate": 8.89744824265768e-05, "loss": 2.5319, "theoretical_loss": 3.453470793106539, "tokens_seen": 1846607872 }, { "epoch": 0.12, "learning_rate": 8.896645803241855e-05, "loss": 2.6329, "theoretical_loss": 3.453450033776256, "tokens_seen": 1846738944 }, { "epoch": 0.12, "learning_rate": 8.895843363826032e-05, "loss": 2.4089, "theoretical_loss": 3.4534292763318275, "tokens_seen": 1846870016 }, { "epoch": 0.12, "learning_rate": 8.895040924410207e-05, "loss": 2.5541, "theoretical_loss": 3.453408520772947, "tokens_seen": 1847001088 }, { "epoch": 0.12, "learning_rate": 8.894238484994384e-05, "loss": 2.5275, "theoretical_loss": 3.4533877670993096, "tokens_seen": 1847132160 }, { "epoch": 0.12, "learning_rate": 8.89343604557856e-05, "loss": 2.4709, "theoretical_loss": 3.4533670153106115, "tokens_seen": 1847263232 }, { "epoch": 0.12, "learning_rate": 8.892633606162735e-05, "loss": 2.4936, "theoretical_loss": 3.4533462654065463, "tokens_seen": 1847394304 }, { "epoch": 0.12, "learning_rate": 8.891831166746911e-05, "loss": 2.6244, "theoretical_loss": 3.4533255173868103, "tokens_seen": 1847525376 }, { "epoch": 0.12, "learning_rate": 8.891028727331087e-05, "loss": 2.5891, "theoretical_loss": 3.4533047712510982, "tokens_seen": 1847656448 }, { "epoch": 0.12, "learning_rate": 8.890226287915263e-05, "loss": 2.4386, "theoretical_loss": 3.453284026999105, "tokens_seen": 1847787520 }, { "epoch": 0.12, "learning_rate": 8.889423848499439e-05, "loss": 2.5154, "theoretical_loss": 3.453263284630527, "tokens_seen": 1847918592 }, { "epoch": 0.12, "learning_rate": 8.888621409083615e-05, "loss": 2.5096, "theoretical_loss": 3.4532425441450587, "tokens_seen": 1848049664 }, { "epoch": 0.12, "learning_rate": 8.88781896966779e-05, "loss": 2.5441, "theoretical_loss": 3.4532218055423956, "tokens_seen": 1848180736 }, { "epoch": 0.12, "objective/train/docs_used": 1018803, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3100316524505615, "objective/train/theoretical_loss": 3.4532114369470213, "objective/train/tokens_used": 218705376, "theoretical_loss": 3.4532114369470213, "tokens_seen": 1848246272 }, { "epoch": 0.12, "learning_rate": 8.887016530251966e-05, "loss": 2.3515, "theoretical_loss": 3.453201068822234, "tokens_seen": 1848311808 }, { "epoch": 0.12, "learning_rate": 8.886214090836143e-05, "loss": 2.4272, "theoretical_loss": 3.4531803339842693, "tokens_seen": 1848442880 }, { "epoch": 0.12, "learning_rate": 8.885411651420318e-05, "loss": 2.4624, "theoretical_loss": 3.4531596010281964, "tokens_seen": 1848573952 }, { "epoch": 0.12, "learning_rate": 8.884609212004495e-05, "loss": 2.5151, "theoretical_loss": 3.453138869953712, "tokens_seen": 1848705024 }, { "epoch": 0.12, "learning_rate": 8.88380677258867e-05, "loss": 2.622, "theoretical_loss": 3.4531181407605116, "tokens_seen": 1848836096 }, { "epoch": 0.12, "learning_rate": 8.883004333172847e-05, "loss": 2.6775, "theoretical_loss": 3.4530974134482917, "tokens_seen": 1848967168 }, { "epoch": 0.12, "learning_rate": 8.882201893757022e-05, "loss": 2.564, "theoretical_loss": 3.4530766880167474, "tokens_seen": 1849098240 }, { "epoch": 0.12, "learning_rate": 8.881399454341197e-05, "loss": 2.6534, "theoretical_loss": 3.4530559644655754, "tokens_seen": 1849229312 }, { "epoch": 0.12, "learning_rate": 8.880597014925374e-05, "loss": 2.5453, "theoretical_loss": 3.4530352427944715, "tokens_seen": 1849360384 }, { "epoch": 0.12, "learning_rate": 8.879794575509549e-05, "loss": 2.6126, "theoretical_loss": 3.453014523003132, "tokens_seen": 1849491456 }, { "epoch": 0.12, "learning_rate": 8.878992136093726e-05, "loss": 2.5405, "theoretical_loss": 3.452993805091253, "tokens_seen": 1849622528 }, { "epoch": 0.12, "learning_rate": 8.878189696677901e-05, "loss": 2.6833, "theoretical_loss": 3.4529730890585313, "tokens_seen": 1849753600 }, { "epoch": 0.12, "objective/train/docs_used": 1019322, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4879612922668457, "objective/train/theoretical_loss": 3.4529523749046627, "objective/train/tokens_used": 220343776, "theoretical_loss": 3.4529523749046627, "tokens_seen": 1849884672 }, { "epoch": 0.12, "learning_rate": 8.877387257262078e-05, "loss": 2.5839, "theoretical_loss": 3.4529523749046627, "tokens_seen": 1849884672 }, { "epoch": 0.12, "learning_rate": 8.876584817846253e-05, "loss": 2.5506, "theoretical_loss": 3.4529316626293443, "tokens_seen": 1850015744 }, { "epoch": 0.12, "learning_rate": 8.87578237843043e-05, "loss": 2.5309, "theoretical_loss": 3.4529109522322727, "tokens_seen": 1850146816 }, { "epoch": 0.12, "learning_rate": 8.874979939014605e-05, "loss": 2.5407, "theoretical_loss": 3.452890243713144, "tokens_seen": 1850277888 }, { "epoch": 0.12, "learning_rate": 8.87417749959878e-05, "loss": 2.4533, "theoretical_loss": 3.452869537071655, "tokens_seen": 1850408960 }, { "epoch": 0.12, "learning_rate": 8.873375060182957e-05, "loss": 2.5144, "theoretical_loss": 3.4528488323075024, "tokens_seen": 1850540032 }, { "epoch": 0.12, "learning_rate": 8.872572620767132e-05, "loss": 2.4951, "theoretical_loss": 3.452828129420384, "tokens_seen": 1850671104 }, { "epoch": 0.12, "learning_rate": 8.871770181351309e-05, "loss": 2.5968, "theoretical_loss": 3.4528074284099954, "tokens_seen": 1850802176 }, { "epoch": 0.12, "learning_rate": 8.870967741935484e-05, "loss": 2.7107, "theoretical_loss": 3.452786729276034, "tokens_seen": 1850933248 }, { "epoch": 0.12, "learning_rate": 8.870165302519661e-05, "loss": 2.4595, "theoretical_loss": 3.4527660320181974, "tokens_seen": 1851064320 }, { "epoch": 0.12, "learning_rate": 8.869362863103836e-05, "loss": 2.5775, "theoretical_loss": 3.452745336636182, "tokens_seen": 1851195392 }, { "epoch": 0.12, "learning_rate": 8.868560423688012e-05, "loss": 2.627, "theoretical_loss": 3.4527246431296854, "tokens_seen": 1851326464 }, { "epoch": 0.12, "learning_rate": 8.867757984272188e-05, "loss": 2.5698, "theoretical_loss": 3.452703951498405, "tokens_seen": 1851457536 }, { "epoch": 0.12, "objective/train/docs_used": 1019593, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.545363426208496, "objective/train/theoretical_loss": 3.452693606385876, "objective/train/tokens_used": 221982176, "theoretical_loss": 3.452693606385876, "tokens_seen": 1851523072 }, { "epoch": 0.12, "learning_rate": 8.866955544856364e-05, "loss": 2.6297, "theoretical_loss": 3.452683261742038, "tokens_seen": 1851588608 }, { "epoch": 0.12, "learning_rate": 8.86615310544054e-05, "loss": 2.664, "theoretical_loss": 3.452662573860281, "tokens_seen": 1851719680 }, { "epoch": 0.12, "learning_rate": 8.865350666024716e-05, "loss": 2.5098, "theoretical_loss": 3.4526418878528333, "tokens_seen": 1851850752 }, { "epoch": 0.12, "learning_rate": 8.864548226608892e-05, "loss": 2.5979, "theoretical_loss": 3.452621203719391, "tokens_seen": 1851981824 }, { "epoch": 0.12, "learning_rate": 8.863745787193068e-05, "loss": 2.6701, "theoretical_loss": 3.452600521459652, "tokens_seen": 1852112896 }, { "epoch": 0.12, "learning_rate": 8.862943347777243e-05, "loss": 2.5514, "theoretical_loss": 3.4525798410733133, "tokens_seen": 1852243968 }, { "epoch": 0.12, "learning_rate": 8.86214090836142e-05, "loss": 2.615, "theoretical_loss": 3.4525591625600742, "tokens_seen": 1852375040 }, { "epoch": 0.12, "learning_rate": 8.861338468945595e-05, "loss": 2.4747, "theoretical_loss": 3.4525384859196313, "tokens_seen": 1852506112 }, { "epoch": 0.12, "learning_rate": 8.860536029529772e-05, "loss": 2.5322, "theoretical_loss": 3.4525178111516834, "tokens_seen": 1852637184 }, { "epoch": 0.12, "learning_rate": 8.859733590113947e-05, "loss": 2.6186, "theoretical_loss": 3.4524971382559277, "tokens_seen": 1852768256 }, { "epoch": 0.12, "learning_rate": 8.858931150698124e-05, "loss": 2.5473, "theoretical_loss": 3.4524764672320627, "tokens_seen": 1852899328 }, { "epoch": 0.12, "learning_rate": 8.858128711282299e-05, "loss": 2.5329, "theoretical_loss": 3.452455798079786, "tokens_seen": 1853030400 }, { "epoch": 0.12, "objective/train/docs_used": 1020710, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7822999954223633, "objective/train/theoretical_loss": 3.4524351307987966, "objective/train/tokens_used": 223620576, "theoretical_loss": 3.4524351307987966, "tokens_seen": 1853161472 }, { "epoch": 0.12, "learning_rate": 8.857326271866474e-05, "loss": 2.7127, "theoretical_loss": 3.4524351307987966, "tokens_seen": 1853161472 }, { "epoch": 0.12, "learning_rate": 8.856523832450651e-05, "loss": 2.6761, "theoretical_loss": 3.452414465388792, "tokens_seen": 1853292544 }, { "epoch": 0.12, "learning_rate": 8.855721393034826e-05, "loss": 2.5513, "theoretical_loss": 3.4523938018494706, "tokens_seen": 1853423616 }, { "epoch": 0.12, "learning_rate": 8.854918953619003e-05, "loss": 2.5548, "theoretical_loss": 3.4523731401805313, "tokens_seen": 1853554688 }, { "epoch": 0.12, "learning_rate": 8.854116514203178e-05, "loss": 2.7371, "theoretical_loss": 3.4523524803816716, "tokens_seen": 1853685760 }, { "epoch": 0.12, "learning_rate": 8.853314074787355e-05, "loss": 2.5652, "theoretical_loss": 3.4523318224525914, "tokens_seen": 1853816832 }, { "epoch": 0.12, "learning_rate": 8.85251163537153e-05, "loss": 2.6217, "theoretical_loss": 3.452311166392988, "tokens_seen": 1853947904 }, { "epoch": 0.12, "learning_rate": 8.851709195955705e-05, "loss": 2.6013, "theoretical_loss": 3.452290512202561, "tokens_seen": 1854078976 }, { "epoch": 0.12, "learning_rate": 8.850906756539882e-05, "loss": 2.4873, "theoretical_loss": 3.4522698598810084, "tokens_seen": 1854210048 }, { "epoch": 0.12, "learning_rate": 8.850104317124057e-05, "loss": 2.5747, "theoretical_loss": 3.4522492094280297, "tokens_seen": 1854341120 }, { "epoch": 0.12, "learning_rate": 8.849301877708234e-05, "loss": 2.4675, "theoretical_loss": 3.4522285608433236, "tokens_seen": 1854472192 }, { "epoch": 0.12, "learning_rate": 8.84849943829241e-05, "loss": 2.6827, "theoretical_loss": 3.4522079141265882, "tokens_seen": 1854603264 }, { "epoch": 0.12, "learning_rate": 8.847696998876586e-05, "loss": 2.6778, "theoretical_loss": 3.4521872692775237, "tokens_seen": 1854734336 }, { "epoch": 0.12, "objective/train/docs_used": 1021316, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.845196485519409, "objective/train/theoretical_loss": 3.4521769475532738, "objective/train/tokens_used": 225258976, "theoretical_loss": 3.4521769475532738, "tokens_seen": 1854799872 }, { "epoch": 0.12, "learning_rate": 8.846894559460761e-05, "loss": 2.6202, "theoretical_loss": 3.4521666262958286, "tokens_seen": 1854865408 }, { "epoch": 0.12, "learning_rate": 8.846092120044937e-05, "loss": 2.4954, "theoretical_loss": 3.452145985181202, "tokens_seen": 1854996480 }, { "epoch": 0.12, "learning_rate": 8.845289680629113e-05, "loss": 2.6478, "theoretical_loss": 3.452125345933343, "tokens_seen": 1855127552 }, { "epoch": 0.12, "learning_rate": 8.844487241213289e-05, "loss": 2.4197, "theoretical_loss": 3.4521047085519516, "tokens_seen": 1855258624 }, { "epoch": 0.12, "learning_rate": 8.843684801797465e-05, "loss": 2.4694, "theoretical_loss": 3.4520840730367266, "tokens_seen": 1855389696 }, { "epoch": 0.12, "learning_rate": 8.842882362381641e-05, "loss": 2.6215, "theoretical_loss": 3.4520634393873673, "tokens_seen": 1855520768 }, { "epoch": 0.12, "learning_rate": 8.842079922965817e-05, "loss": 2.4711, "theoretical_loss": 3.452042807603574, "tokens_seen": 1855651840 }, { "epoch": 0.12, "learning_rate": 8.841277483549993e-05, "loss": 2.5541, "theoretical_loss": 3.452022177685045, "tokens_seen": 1855782912 }, { "epoch": 0.12, "learning_rate": 8.84047504413417e-05, "loss": 2.691, "theoretical_loss": 3.452001549631481, "tokens_seen": 1855913984 }, { "epoch": 0.12, "learning_rate": 8.839672604718345e-05, "loss": 2.7246, "theoretical_loss": 3.4519809234425813, "tokens_seen": 1856045056 }, { "epoch": 0.12, "learning_rate": 8.83887016530252e-05, "loss": 2.3307, "theoretical_loss": 3.451960299118046, "tokens_seen": 1856176128 }, { "epoch": 0.13, "learning_rate": 8.838067725886697e-05, "loss": 2.498, "theoretical_loss": 3.4519396766575747, "tokens_seen": 1856307200 }, { "epoch": 0.13, "objective/train/docs_used": 1022409, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.1355769634246826, "objective/train/theoretical_loss": 3.4519190560608672, "objective/train/tokens_used": 226897376, "theoretical_loss": 3.4519190560608672, "tokens_seen": 1856438272 }, { "epoch": 0.13, "learning_rate": 8.837265286470872e-05, "loss": 2.5153, "theoretical_loss": 3.4519190560608672, "tokens_seen": 1856438272 }, { "epoch": 0.13, "learning_rate": 8.836462847055049e-05, "loss": 2.4026, "theoretical_loss": 3.4518984373276234, "tokens_seen": 1856569344 }, { "epoch": 0.13, "learning_rate": 8.835660407639224e-05, "loss": 2.5265, "theoretical_loss": 3.4518778204575433, "tokens_seen": 1856700416 }, { "epoch": 0.13, "learning_rate": 8.8348579682234e-05, "loss": 2.4881, "theoretical_loss": 3.4518572054503283, "tokens_seen": 1856831488 }, { "epoch": 0.13, "learning_rate": 8.834055528807576e-05, "loss": 2.5827, "theoretical_loss": 3.4518365923056766, "tokens_seen": 1856962560 }, { "epoch": 0.13, "learning_rate": 8.833253089391751e-05, "loss": 2.6521, "theoretical_loss": 3.45181598102329, "tokens_seen": 1857093632 }, { "epoch": 0.13, "learning_rate": 8.832450649975928e-05, "loss": 2.5918, "theoretical_loss": 3.451795371602868, "tokens_seen": 1857224704 }, { "epoch": 0.13, "learning_rate": 8.831648210560103e-05, "loss": 2.5057, "theoretical_loss": 3.451774764044112, "tokens_seen": 1857355776 }, { "epoch": 0.13, "learning_rate": 8.83084577114428e-05, "loss": 2.551, "theoretical_loss": 3.4517541583467213, "tokens_seen": 1857486848 }, { "epoch": 0.13, "learning_rate": 8.830043331728455e-05, "loss": 2.5057, "theoretical_loss": 3.4517335545103967, "tokens_seen": 1857617920 }, { "epoch": 0.13, "learning_rate": 8.829240892312632e-05, "loss": 2.4116, "theoretical_loss": 3.45171295253484, "tokens_seen": 1857748992 }, { "epoch": 0.13, "learning_rate": 8.828438452896807e-05, "loss": 2.5689, "theoretical_loss": 3.45169235241975, "tokens_seen": 1857880064 }, { "epoch": 0.13, "learning_rate": 8.827636013480982e-05, "loss": 2.4917, "theoretical_loss": 3.4516717541648285, "tokens_seen": 1858011136 }, { "epoch": 0.13, "objective/train/docs_used": 1023672, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6751821041107178, "objective/train/theoretical_loss": 3.4516614557348375, "objective/train/tokens_used": 228535776, "theoretical_loss": 3.4516614557348375, "tokens_seen": 1858076672 }, { "epoch": 0.13, "learning_rate": 8.826833574065159e-05, "loss": 2.6894, "theoretical_loss": 3.451651157769777, "tokens_seen": 1858142208 }, { "epoch": 0.13, "learning_rate": 8.826031134649334e-05, "loss": 2.6551, "theoretical_loss": 3.4516305632342945, "tokens_seen": 1858273280 }, { "epoch": 0.13, "learning_rate": 8.825228695233511e-05, "loss": 2.5355, "theoretical_loss": 3.451609970558083, "tokens_seen": 1858404352 }, { "epoch": 0.13, "learning_rate": 8.824426255817686e-05, "loss": 2.499, "theoretical_loss": 3.4515893797408443, "tokens_seen": 1858535424 }, { "epoch": 0.13, "learning_rate": 8.823623816401863e-05, "loss": 2.6607, "theoretical_loss": 3.4515687907822787, "tokens_seen": 1858666496 }, { "epoch": 0.13, "learning_rate": 8.822821376986038e-05, "loss": 2.6997, "theoretical_loss": 3.451548203682088, "tokens_seen": 1858797568 }, { "epoch": 0.13, "learning_rate": 8.822018937570214e-05, "loss": 2.5268, "theoretical_loss": 3.4515276184399717, "tokens_seen": 1858928640 }, { "epoch": 0.13, "learning_rate": 8.82121649815439e-05, "loss": 2.7007, "theoretical_loss": 3.4515070350556334, "tokens_seen": 1859059712 }, { "epoch": 0.13, "learning_rate": 8.820414058738566e-05, "loss": 2.4948, "theoretical_loss": 3.451486453528773, "tokens_seen": 1859190784 }, { "epoch": 0.13, "learning_rate": 8.819611619322742e-05, "loss": 2.494, "theoretical_loss": 3.4514658738590915, "tokens_seen": 1859321856 }, { "epoch": 0.13, "learning_rate": 8.818809179906918e-05, "loss": 2.7342, "theoretical_loss": 3.451445296046292, "tokens_seen": 1859452928 }, { "epoch": 0.13, "learning_rate": 8.818006740491094e-05, "loss": 2.7209, "theoretical_loss": 3.4514247200900754, "tokens_seen": 1859584000 }, { "epoch": 0.13, "objective/train/docs_used": 1024297, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4555611610412598, "objective/train/theoretical_loss": 3.4514041459901432, "objective/train/tokens_used": 230174176, "theoretical_loss": 3.4514041459901432, "tokens_seen": 1859715072 }, { "epoch": 0.13, "learning_rate": 8.81720430107527e-05, "loss": 2.6482, "theoretical_loss": 3.4514041459901432, "tokens_seen": 1859715072 }, { "epoch": 0.13, "learning_rate": 8.816401861659445e-05, "loss": 2.6735, "theoretical_loss": 3.451383573746197, "tokens_seen": 1859846144 }, { "epoch": 0.13, "learning_rate": 8.815599422243622e-05, "loss": 2.5834, "theoretical_loss": 3.451363003357939, "tokens_seen": 1859977216 }, { "epoch": 0.13, "learning_rate": 8.814796982827797e-05, "loss": 2.5174, "theoretical_loss": 3.4513424348250705, "tokens_seen": 1860108288 }, { "epoch": 0.13, "learning_rate": 8.813994543411974e-05, "loss": 2.6091, "theoretical_loss": 3.451321868147294, "tokens_seen": 1860239360 }, { "epoch": 0.13, "learning_rate": 8.813192103996149e-05, "loss": 2.6376, "theoretical_loss": 3.4513013033243114, "tokens_seen": 1860370432 }, { "epoch": 0.13, "learning_rate": 8.812389664580326e-05, "loss": 2.616, "theoretical_loss": 3.4512807403558243, "tokens_seen": 1860501504 }, { "epoch": 0.13, "learning_rate": 8.811587225164501e-05, "loss": 2.5169, "theoretical_loss": 3.451260179241536, "tokens_seen": 1860632576 }, { "epoch": 0.13, "learning_rate": 8.810784785748678e-05, "loss": 2.7286, "theoretical_loss": 3.4512396199811466, "tokens_seen": 1860763648 }, { "epoch": 0.13, "learning_rate": 8.809982346332853e-05, "loss": 2.5962, "theoretical_loss": 3.4512190625743604, "tokens_seen": 1860894720 }, { "epoch": 0.13, "learning_rate": 8.809179906917028e-05, "loss": 2.6989, "theoretical_loss": 3.4511985070208793, "tokens_seen": 1861025792 }, { "epoch": 0.13, "learning_rate": 8.808377467501205e-05, "loss": 2.6519, "theoretical_loss": 3.4511779533204043, "tokens_seen": 1861156864 }, { "epoch": 0.13, "learning_rate": 8.80757502808538e-05, "loss": 2.6076, "theoretical_loss": 3.4511574014726403, "tokens_seen": 1861287936 }, { "epoch": 0.13, "objective/train/docs_used": 1025545, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7613306045532227, "objective/train/theoretical_loss": 3.4511471262434306, "objective/train/tokens_used": 231812576, "theoretical_loss": 3.4511471262434306, "tokens_seen": 1861353472 }, { "epoch": 0.13, "learning_rate": 8.806772588669557e-05, "loss": 2.6141, "theoretical_loss": 3.4511368514772878, "tokens_seen": 1861419008 }, { "epoch": 0.13, "learning_rate": 8.805970149253732e-05, "loss": 2.6818, "theoretical_loss": 3.4511163033340497, "tokens_seen": 1861550080 }, { "epoch": 0.13, "learning_rate": 8.805167709837908e-05, "loss": 2.695, "theoretical_loss": 3.45109575704263, "tokens_seen": 1861681152 }, { "epoch": 0.13, "learning_rate": 8.804365270422084e-05, "loss": 2.4897, "theoretical_loss": 3.4510752126027295, "tokens_seen": 1861812224 }, { "epoch": 0.13, "learning_rate": 8.80356283100626e-05, "loss": 2.526, "theoretical_loss": 3.451054670014053, "tokens_seen": 1861943296 }, { "epoch": 0.13, "learning_rate": 8.802760391590436e-05, "loss": 2.564, "theoretical_loss": 3.451034129276302, "tokens_seen": 1862074368 }, { "epoch": 0.13, "learning_rate": 8.801957952174611e-05, "loss": 2.4509, "theoretical_loss": 3.45101359038918, "tokens_seen": 1862205440 }, { "epoch": 0.13, "learning_rate": 8.801155512758788e-05, "loss": 2.6459, "theoretical_loss": 3.4509930533523905, "tokens_seen": 1862336512 }, { "epoch": 0.13, "learning_rate": 8.800353073342963e-05, "loss": 2.5574, "theoretical_loss": 3.4509725181656354, "tokens_seen": 1862467584 }, { "epoch": 0.13, "learning_rate": 8.799550633927139e-05, "loss": 2.6837, "theoretical_loss": 3.4509519848286185, "tokens_seen": 1862598656 }, { "epoch": 0.13, "learning_rate": 8.798748194511315e-05, "loss": 2.6577, "theoretical_loss": 3.4509314533410436, "tokens_seen": 1862729728 }, { "epoch": 0.13, "learning_rate": 8.797945755095491e-05, "loss": 2.6322, "theoretical_loss": 3.450910923702613, "tokens_seen": 1862860800 }, { "epoch": 0.13, "objective/train/docs_used": 1026014, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.46268630027771, "objective/train/theoretical_loss": 3.4508903959130306, "objective/train/tokens_used": 233450976, "theoretical_loss": 3.4508903959130306, "tokens_seen": 1862991872 }, { "epoch": 0.13, "learning_rate": 8.797143315679667e-05, "loss": 2.6023, "theoretical_loss": 3.4508903959130306, "tokens_seen": 1862991872 }, { "epoch": 0.13, "learning_rate": 8.796340876263843e-05, "loss": 2.5201, "theoretical_loss": 3.450869869972, "tokens_seen": 1863122944 }, { "epoch": 0.13, "learning_rate": 8.795538436848018e-05, "loss": 2.762, "theoretical_loss": 3.4508493458792246, "tokens_seen": 1863254016 }, { "epoch": 0.13, "learning_rate": 8.794735997432195e-05, "loss": 2.7066, "theoretical_loss": 3.4508288236344073, "tokens_seen": 1863385088 }, { "epoch": 0.13, "learning_rate": 8.79393355801637e-05, "loss": 2.5886, "theoretical_loss": 3.4508083032372525, "tokens_seen": 1863516160 }, { "epoch": 0.13, "learning_rate": 8.793131118600547e-05, "loss": 2.6404, "theoretical_loss": 3.4507877846874635, "tokens_seen": 1863647232 }, { "epoch": 0.13, "learning_rate": 8.792328679184722e-05, "loss": 2.7333, "theoretical_loss": 3.4507672679847445, "tokens_seen": 1863778304 }, { "epoch": 0.13, "learning_rate": 8.791526239768899e-05, "loss": 2.4031, "theoretical_loss": 3.450746753128799, "tokens_seen": 1863909376 }, { "epoch": 0.13, "learning_rate": 8.790723800353074e-05, "loss": 2.5246, "theoretical_loss": 3.4507262401193315, "tokens_seen": 1864040448 }, { "epoch": 0.13, "learning_rate": 8.789921360937249e-05, "loss": 2.3183, "theoretical_loss": 3.450705728956045, "tokens_seen": 1864171520 }, { "epoch": 0.13, "learning_rate": 8.789118921521426e-05, "loss": 2.6952, "theoretical_loss": 3.450685219638644, "tokens_seen": 1864302592 }, { "epoch": 0.13, "learning_rate": 8.788316482105601e-05, "loss": 2.5442, "theoretical_loss": 3.4506647121668332, "tokens_seen": 1864433664 }, { "epoch": 0.13, "learning_rate": 8.787514042689778e-05, "loss": 2.6715, "theoretical_loss": 3.450644206540316, "tokens_seen": 1864564736 }, { "epoch": 0.13, "objective/train/docs_used": 1027424, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5948421955108643, "objective/train/theoretical_loss": 3.4506339544189504, "objective/train/tokens_used": 235089376, "theoretical_loss": 3.4506339544189504, "tokens_seen": 1864630272 }, { "epoch": 0.13, "learning_rate": 8.786711603273953e-05, "loss": 2.6985, "theoretical_loss": 3.450623702758797, "tokens_seen": 1864695808 }, { "epoch": 0.13, "learning_rate": 8.785909163858129e-05, "loss": 2.6472, "theoretical_loss": 3.450603200821981, "tokens_seen": 1864826880 }, { "epoch": 0.13, "learning_rate": 8.785106724442305e-05, "loss": 2.7063, "theoretical_loss": 3.450582700729571, "tokens_seen": 1864957952 }, { "epoch": 0.13, "learning_rate": 8.78430428502648e-05, "loss": 2.7536, "theoretical_loss": 3.4505622024812728, "tokens_seen": 1865089024 }, { "epoch": 0.13, "learning_rate": 8.783501845610657e-05, "loss": 2.6416, "theoretical_loss": 3.4505417060767902, "tokens_seen": 1865220096 }, { "epoch": 0.13, "learning_rate": 8.782699406194833e-05, "loss": 2.7129, "theoretical_loss": 3.450521211515828, "tokens_seen": 1865351168 }, { "epoch": 0.13, "learning_rate": 8.781896966779009e-05, "loss": 2.5542, "theoretical_loss": 3.4505007187980916, "tokens_seen": 1865482240 }, { "epoch": 0.13, "learning_rate": 8.781094527363185e-05, "loss": 2.6308, "theoretical_loss": 3.4504802279232845, "tokens_seen": 1865613312 }, { "epoch": 0.13, "learning_rate": 8.78029208794736e-05, "loss": 2.5675, "theoretical_loss": 3.4504597388911122, "tokens_seen": 1865744384 }, { "epoch": 0.13, "learning_rate": 8.779489648531536e-05, "loss": 2.5833, "theoretical_loss": 3.4504392517012796, "tokens_seen": 1865875456 }, { "epoch": 0.13, "learning_rate": 8.778687209115712e-05, "loss": 2.6185, "theoretical_loss": 3.450418766353491, "tokens_seen": 1866006528 }, { "epoch": 0.13, "learning_rate": 8.777884769699888e-05, "loss": 2.7469, "theoretical_loss": 3.450398282847452, "tokens_seen": 1866137600 }, { "epoch": 0.13, "objective/train/docs_used": 1028516, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7388877868652344, "objective/train/theoretical_loss": 3.450377801182868, "objective/train/tokens_used": 236727776, "theoretical_loss": 3.450377801182868, "tokens_seen": 1866268672 }, { "epoch": 0.13, "learning_rate": 8.777082330284064e-05, "loss": 2.6726, "theoretical_loss": 3.450377801182868, "tokens_seen": 1866268672 }, { "epoch": 0.13, "learning_rate": 8.776279890868239e-05, "loss": 2.642, "theoretical_loss": 3.4503573213594434, "tokens_seen": 1866399744 }, { "epoch": 0.13, "learning_rate": 8.775477451452416e-05, "loss": 2.6033, "theoretical_loss": 3.4503368433768835, "tokens_seen": 1866530816 }, { "epoch": 0.13, "learning_rate": 8.774675012036591e-05, "loss": 2.4442, "theoretical_loss": 3.450316367234894, "tokens_seen": 1866661888 }, { "epoch": 0.13, "learning_rate": 8.773872572620768e-05, "loss": 2.513, "theoretical_loss": 3.45029589293318, "tokens_seen": 1866792960 }, { "epoch": 0.13, "learning_rate": 8.773070133204943e-05, "loss": 2.5866, "theoretical_loss": 3.450275420471447, "tokens_seen": 1866924032 }, { "epoch": 0.13, "learning_rate": 8.77226769378912e-05, "loss": 2.7162, "theoretical_loss": 3.4502549498494, "tokens_seen": 1867055104 }, { "epoch": 0.13, "learning_rate": 8.771465254373295e-05, "loss": 2.518, "theoretical_loss": 3.4502344810667456, "tokens_seen": 1867186176 }, { "epoch": 0.13, "learning_rate": 8.77066281495747e-05, "loss": 2.6779, "theoretical_loss": 3.4502140141231887, "tokens_seen": 1867317248 }, { "epoch": 0.13, "learning_rate": 8.769860375541647e-05, "loss": 2.6632, "theoretical_loss": 3.450193549018435, "tokens_seen": 1867448320 }, { "epoch": 0.13, "learning_rate": 8.769057936125822e-05, "loss": 2.4661, "theoretical_loss": 3.4501730857521906, "tokens_seen": 1867579392 }, { "epoch": 0.13, "learning_rate": 8.768255496709999e-05, "loss": 2.5559, "theoretical_loss": 3.4501526243241605, "tokens_seen": 1867710464 }, { "epoch": 0.13, "learning_rate": 8.767453057294174e-05, "loss": 2.7043, "theoretical_loss": 3.4501321647340513, "tokens_seen": 1867841536 }, { "epoch": 0.13, "objective/train/docs_used": 1029117, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6434731483459473, "objective/train/theoretical_loss": 3.4501219356281254, "objective/train/tokens_used": 238366176, "theoretical_loss": 3.4501219356281254, "tokens_seen": 1867907072 }, { "epoch": 0.13, "learning_rate": 8.76665061787835e-05, "loss": 2.7618, "theoretical_loss": 3.450111706981569, "tokens_seen": 1867972608 }, { "epoch": 0.13, "learning_rate": 8.765848178462526e-05, "loss": 2.5639, "theoretical_loss": 3.4500912510664192, "tokens_seen": 1868103680 }, { "epoch": 0.13, "learning_rate": 8.765045739046702e-05, "loss": 2.6544, "theoretical_loss": 3.4500707969883084, "tokens_seen": 1868234752 }, { "epoch": 0.13, "learning_rate": 8.764243299630878e-05, "loss": 2.5797, "theoretical_loss": 3.450050344746942, "tokens_seen": 1868365824 }, { "epoch": 0.13, "learning_rate": 8.763440860215054e-05, "loss": 2.5622, "theoretical_loss": 3.450029894342028, "tokens_seen": 1868496896 }, { "epoch": 0.13, "learning_rate": 8.76263842079923e-05, "loss": 2.6618, "theoretical_loss": 3.4500094457732704, "tokens_seen": 1868627968 }, { "epoch": 0.13, "learning_rate": 8.761835981383406e-05, "loss": 2.7035, "theoretical_loss": 3.449988999040377, "tokens_seen": 1868759040 }, { "epoch": 0.13, "learning_rate": 8.761033541967581e-05, "loss": 2.6978, "theoretical_loss": 3.4499685541430543, "tokens_seen": 1868890112 }, { "epoch": 0.13, "learning_rate": 8.760231102551758e-05, "loss": 2.5531, "theoretical_loss": 3.4499481110810075, "tokens_seen": 1869021184 }, { "epoch": 0.13, "learning_rate": 8.759428663135933e-05, "loss": 2.5434, "theoretical_loss": 3.449927669853945, "tokens_seen": 1869152256 }, { "epoch": 0.13, "learning_rate": 8.75862622372011e-05, "loss": 2.5474, "theoretical_loss": 3.449907230461572, "tokens_seen": 1869283328 }, { "epoch": 0.13, "learning_rate": 8.757823784304285e-05, "loss": 2.6269, "theoretical_loss": 3.4498867929035955, "tokens_seen": 1869414400 }, { "epoch": 0.13, "objective/train/docs_used": 1029470, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6646249294281006, "objective/train/theoretical_loss": 3.449866357179723, "objective/train/tokens_used": 240004576, "theoretical_loss": 3.449866357179723, "tokens_seen": 1869545472 }, { "epoch": 0.13, "learning_rate": 8.75702134488846e-05, "loss": 2.7099, "theoretical_loss": 3.449866357179723, "tokens_seen": 1869545472 }, { "epoch": 0.13, "learning_rate": 8.756218905472637e-05, "loss": 2.635, "theoretical_loss": 3.44984592328966, "tokens_seen": 1869676544 }, { "epoch": 0.13, "learning_rate": 8.755416466056812e-05, "loss": 2.4857, "theoretical_loss": 3.4498254912331148, "tokens_seen": 1869807616 }, { "epoch": 0.13, "learning_rate": 8.754614026640989e-05, "loss": 2.7694, "theoretical_loss": 3.4498050610097932, "tokens_seen": 1869938688 }, { "epoch": 0.13, "learning_rate": 8.753811587225164e-05, "loss": 2.489, "theoretical_loss": 3.4497846326194033, "tokens_seen": 1870069760 }, { "epoch": 0.13, "learning_rate": 8.753009147809341e-05, "loss": 2.479, "theoretical_loss": 3.449764206061652, "tokens_seen": 1870200832 }, { "epoch": 0.13, "learning_rate": 8.752206708393516e-05, "loss": 2.6056, "theoretical_loss": 3.4497437813362453, "tokens_seen": 1870331904 }, { "epoch": 0.13, "learning_rate": 8.751404268977691e-05, "loss": 2.7141, "theoretical_loss": 3.449723358442892, "tokens_seen": 1870462976 }, { "epoch": 0.13, "learning_rate": 8.750601829561868e-05, "loss": 2.6245, "theoretical_loss": 3.449702937381298, "tokens_seen": 1870594048 }, { "epoch": 0.13, "learning_rate": 8.749799390146043e-05, "loss": 2.6135, "theoretical_loss": 3.449682518151172, "tokens_seen": 1870725120 }, { "epoch": 0.13, "learning_rate": 8.74899695073022e-05, "loss": 2.6038, "theoretical_loss": 3.449662100752221, "tokens_seen": 1870856192 }, { "epoch": 0.13, "learning_rate": 8.748194511314395e-05, "loss": 2.5741, "theoretical_loss": 3.449641685184152, "tokens_seen": 1870987264 }, { "epoch": 0.13, "learning_rate": 8.747392071898572e-05, "loss": 2.5113, "theoretical_loss": 3.4496212714466727, "tokens_seen": 1871118336 }, { "epoch": 0.13, "objective/train/docs_used": 1030778, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5349972248077393, "objective/train/theoretical_loss": 3.449611065264313, "objective/train/tokens_used": 241642976, "theoretical_loss": 3.449611065264313, "tokens_seen": 1871183872 }, { "epoch": 0.13, "learning_rate": 8.746589632482747e-05, "loss": 2.6365, "theoretical_loss": 3.449600859539491, "tokens_seen": 1871249408 }, { "epoch": 0.13, "learning_rate": 8.745787193066924e-05, "loss": 2.6784, "theoretical_loss": 3.4495804494623146, "tokens_seen": 1871380480 }, { "epoch": 0.13, "learning_rate": 8.7449847536511e-05, "loss": 2.6674, "theoretical_loss": 3.449560041214851, "tokens_seen": 1871511552 }, { "epoch": 0.13, "learning_rate": 8.744182314235275e-05, "loss": 2.6919, "theoretical_loss": 3.4495396347968086, "tokens_seen": 1871642624 }, { "epoch": 0.13, "learning_rate": 8.743379874819451e-05, "loss": 2.5136, "theoretical_loss": 3.449519230207895, "tokens_seen": 1871773696 }, { "epoch": 0.13, "learning_rate": 8.742577435403627e-05, "loss": 2.4687, "theoretical_loss": 3.4494988274478184, "tokens_seen": 1871904768 }, { "epoch": 0.13, "learning_rate": 8.741774995987803e-05, "loss": 2.5254, "theoretical_loss": 3.4494784265162863, "tokens_seen": 1872035840 }, { "epoch": 0.13, "learning_rate": 8.740972556571979e-05, "loss": 2.5739, "theoretical_loss": 3.449458027413007, "tokens_seen": 1872166912 }, { "epoch": 0.13, "learning_rate": 8.740170117156155e-05, "loss": 2.6662, "theoretical_loss": 3.4494376301376892, "tokens_seen": 1872297984 }, { "epoch": 0.13, "learning_rate": 8.73936767774033e-05, "loss": 2.3869, "theoretical_loss": 3.4494172346900402, "tokens_seen": 1872429056 }, { "epoch": 0.13, "learning_rate": 8.738565238324506e-05, "loss": 2.5522, "theoretical_loss": 3.4493968410697695, "tokens_seen": 1872560128 }, { "epoch": 0.13, "learning_rate": 8.737762798908683e-05, "loss": 2.5246, "theoretical_loss": 3.449376449276584, "tokens_seen": 1872691200 }, { "epoch": 0.13, "objective/train/docs_used": 1031383, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6211163997650146, "objective/train/theoretical_loss": 3.449356059310194, "objective/train/tokens_used": 243281376, "theoretical_loss": 3.449356059310194, "tokens_seen": 1872822272 }, { "epoch": 0.14, "learning_rate": 8.736960359492858e-05, "loss": 2.53, "theoretical_loss": 3.449356059310194, "tokens_seen": 1872822272 }, { "epoch": 0.14, "learning_rate": 8.736157920077035e-05, "loss": 2.6602, "theoretical_loss": 3.4493356711703065, "tokens_seen": 1872953344 }, { "epoch": 0.14, "learning_rate": 8.73535548066121e-05, "loss": 2.4317, "theoretical_loss": 3.44931528485663, "tokens_seen": 1873084416 }, { "epoch": 0.14, "learning_rate": 8.734553041245387e-05, "loss": 2.6668, "theoretical_loss": 3.4492949003688747, "tokens_seen": 1873215488 }, { "epoch": 0.14, "learning_rate": 8.733750601829562e-05, "loss": 2.6375, "theoretical_loss": 3.4492745177067476, "tokens_seen": 1873346560 }, { "epoch": 0.14, "learning_rate": 8.732948162413737e-05, "loss": 2.6174, "theoretical_loss": 3.4492541368699587, "tokens_seen": 1873477632 }, { "epoch": 0.14, "learning_rate": 8.732145722997914e-05, "loss": 2.4123, "theoretical_loss": 3.4492337578582157, "tokens_seen": 1873608704 }, { "epoch": 0.14, "learning_rate": 8.731343283582089e-05, "loss": 2.5342, "theoretical_loss": 3.449213380671229, "tokens_seen": 1873739776 }, { "epoch": 0.14, "learning_rate": 8.730540844166266e-05, "loss": 2.5533, "theoretical_loss": 3.4491930053087065, "tokens_seen": 1873870848 }, { "epoch": 0.14, "learning_rate": 8.729738404750441e-05, "loss": 2.5133, "theoretical_loss": 3.449172631770357, "tokens_seen": 1874001920 }, { "epoch": 0.14, "learning_rate": 8.728935965334618e-05, "loss": 2.657, "theoretical_loss": 3.4491522600558904, "tokens_seen": 1874132992 }, { "epoch": 0.14, "learning_rate": 8.728133525918793e-05, "loss": 2.5688, "theoretical_loss": 3.449131890165016, "tokens_seen": 1874264064 }, { "epoch": 0.14, "learning_rate": 8.727331086502968e-05, "loss": 2.6587, "theoretical_loss": 3.4491115220974424, "tokens_seen": 1874395136 }, { "epoch": 0.14, "objective/train/docs_used": 1032518, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.44793438911438, "objective/train/theoretical_loss": 3.449101338747303, "objective/train/tokens_used": 244919776, "theoretical_loss": 3.449101338747303, "tokens_seen": 1874460672 }, { "epoch": 0.14, "learning_rate": 8.726528647087145e-05, "loss": 2.5988, "theoretical_loss": 3.449091155852879, "tokens_seen": 1874526208 }, { "epoch": 0.14, "learning_rate": 8.72572620767132e-05, "loss": 2.6024, "theoretical_loss": 3.4490707914310352, "tokens_seen": 1874657280 }, { "epoch": 0.14, "learning_rate": 8.724923768255497e-05, "loss": 2.5618, "theoretical_loss": 3.4490504288316215, "tokens_seen": 1874788352 }, { "epoch": 0.14, "learning_rate": 8.724121328839672e-05, "loss": 2.5314, "theoretical_loss": 3.4490300680543458, "tokens_seen": 1874919424 }, { "epoch": 0.14, "learning_rate": 8.723318889423849e-05, "loss": 2.5469, "theoretical_loss": 3.4490097090989185, "tokens_seen": 1875050496 }, { "epoch": 0.14, "learning_rate": 8.722516450008024e-05, "loss": 2.4655, "theoretical_loss": 3.448989351965049, "tokens_seen": 1875181568 }, { "epoch": 0.14, "learning_rate": 8.7217140105922e-05, "loss": 2.6071, "theoretical_loss": 3.4489689966524475, "tokens_seen": 1875312640 }, { "epoch": 0.14, "learning_rate": 8.720911571176376e-05, "loss": 2.5736, "theoretical_loss": 3.448948643160823, "tokens_seen": 1875443712 }, { "epoch": 0.14, "learning_rate": 8.720109131760552e-05, "loss": 2.5253, "theoretical_loss": 3.448928291489886, "tokens_seen": 1875574784 }, { "epoch": 0.14, "learning_rate": 8.719306692344728e-05, "loss": 2.6315, "theoretical_loss": 3.448907941639346, "tokens_seen": 1875705856 }, { "epoch": 0.14, "learning_rate": 8.718504252928904e-05, "loss": 2.6466, "theoretical_loss": 3.448887593608914, "tokens_seen": 1875836928 }, { "epoch": 0.14, "learning_rate": 8.71770181351308e-05, "loss": 2.6256, "theoretical_loss": 3.448867247398298, "tokens_seen": 1875968000 }, { "epoch": 0.14, "objective/train/docs_used": 1033022, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8692405223846436, "objective/train/theoretical_loss": 3.448846903007211, "objective/train/tokens_used": 246558176, "theoretical_loss": 3.448846903007211, "tokens_seen": 1876099072 }, { "epoch": 0.14, "learning_rate": 8.716899374097256e-05, "loss": 2.6994, "theoretical_loss": 3.448846903007211, "tokens_seen": 1876099072 }, { "epoch": 0.14, "learning_rate": 8.716096934681432e-05, "loss": 2.6321, "theoretical_loss": 3.44882656043536, "tokens_seen": 1876230144 }, { "epoch": 0.14, "learning_rate": 8.715294495265608e-05, "loss": 2.5969, "theoretical_loss": 3.4488062196824574, "tokens_seen": 1876361216 }, { "epoch": 0.14, "learning_rate": 8.714492055849783e-05, "loss": 2.6608, "theoretical_loss": 3.448785880748213, "tokens_seen": 1876492288 }, { "epoch": 0.14, "learning_rate": 8.71368961643396e-05, "loss": 2.6761, "theoretical_loss": 3.4487655436323363, "tokens_seen": 1876623360 }, { "epoch": 0.14, "learning_rate": 8.712887177018135e-05, "loss": 2.585, "theoretical_loss": 3.4487452083345396, "tokens_seen": 1876754432 }, { "epoch": 0.14, "learning_rate": 8.712084737602312e-05, "loss": 2.4546, "theoretical_loss": 3.4487248748545314, "tokens_seen": 1876885504 }, { "epoch": 0.14, "learning_rate": 8.711282298186487e-05, "loss": 2.7406, "theoretical_loss": 3.4487045431920236, "tokens_seen": 1877016576 }, { "epoch": 0.14, "learning_rate": 8.710479858770664e-05, "loss": 2.6284, "theoretical_loss": 3.4486842133467266, "tokens_seen": 1877147648 }, { "epoch": 0.14, "learning_rate": 8.709677419354839e-05, "loss": 2.6591, "theoretical_loss": 3.448663885318351, "tokens_seen": 1877278720 }, { "epoch": 0.14, "learning_rate": 8.708874979939014e-05, "loss": 2.6345, "theoretical_loss": 3.448643559106607, "tokens_seen": 1877409792 }, { "epoch": 0.14, "learning_rate": 8.708072540523191e-05, "loss": 2.6332, "theoretical_loss": 3.4486232347112065, "tokens_seen": 1877540864 }, { "epoch": 0.14, "learning_rate": 8.707270101107366e-05, "loss": 2.6165, "theoretical_loss": 3.4486029121318595, "tokens_seen": 1877671936 }, { "epoch": 0.14, "objective/train/docs_used": 1034169, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.514636993408203, "objective/train/theoretical_loss": 3.448592751523116, "objective/train/tokens_used": 248196576, "theoretical_loss": 3.448592751523116, "tokens_seen": 1877737472 }, { "epoch": 0.14, "learning_rate": 8.706467661691543e-05, "loss": 2.5063, "theoretical_loss": 3.4485825913682775, "tokens_seen": 1877803008 }, { "epoch": 0.14, "learning_rate": 8.705665222275718e-05, "loss": 2.5341, "theoretical_loss": 3.448562272420171, "tokens_seen": 1877934080 }, { "epoch": 0.14, "learning_rate": 8.704862782859895e-05, "loss": 2.6616, "theoretical_loss": 3.4485419552872525, "tokens_seen": 1878065152 }, { "epoch": 0.14, "learning_rate": 8.70406034344407e-05, "loss": 2.5614, "theoretical_loss": 3.4485216399692313, "tokens_seen": 1878196224 }, { "epoch": 0.14, "learning_rate": 8.703257904028245e-05, "loss": 2.5634, "theoretical_loss": 3.44850132646582, "tokens_seen": 1878327296 }, { "epoch": 0.14, "learning_rate": 8.702455464612422e-05, "loss": 2.6708, "theoretical_loss": 3.4484810147767293, "tokens_seen": 1878458368 }, { "epoch": 0.14, "learning_rate": 8.701653025196597e-05, "loss": 2.7177, "theoretical_loss": 3.4484607049016702, "tokens_seen": 1878589440 }, { "epoch": 0.14, "learning_rate": 8.700850585780774e-05, "loss": 2.6269, "theoretical_loss": 3.4484403968403554, "tokens_seen": 1878720512 }, { "epoch": 0.14, "learning_rate": 8.70004814636495e-05, "loss": 2.6973, "theoretical_loss": 3.4484200905924953, "tokens_seen": 1878851584 }, { "epoch": 0.14, "learning_rate": 8.699245706949126e-05, "loss": 2.6511, "theoretical_loss": 3.448399786157802, "tokens_seen": 1878982656 }, { "epoch": 0.14, "learning_rate": 8.698443267533301e-05, "loss": 2.6892, "theoretical_loss": 3.448379483535987, "tokens_seen": 1879113728 }, { "epoch": 0.14, "learning_rate": 8.697640828117477e-05, "loss": 2.6398, "theoretical_loss": 3.4483591827267617, "tokens_seen": 1879244800 }, { "epoch": 0.14, "objective/train/docs_used": 1034880, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.418511390686035, "objective/train/theoretical_loss": 3.4483388837298383, "objective/train/tokens_used": 249834976, "theoretical_loss": 3.4483388837298383, "tokens_seen": 1879375872 }, { "epoch": 0.14, "learning_rate": 8.696838388701653e-05, "loss": 2.6351, "theoretical_loss": 3.4483388837298383, "tokens_seen": 1879375872 }, { "epoch": 0.14, "learning_rate": 8.696035949285829e-05, "loss": 2.7967, "theoretical_loss": 3.448318586544928, "tokens_seen": 1879506944 }, { "epoch": 0.14, "learning_rate": 8.695233509870005e-05, "loss": 2.7456, "theoretical_loss": 3.4482982911717435, "tokens_seen": 1879638016 }, { "epoch": 0.14, "learning_rate": 8.69443107045418e-05, "loss": 2.6476, "theoretical_loss": 3.4482779976099964, "tokens_seen": 1879769088 }, { "epoch": 0.14, "learning_rate": 8.693628631038357e-05, "loss": 2.56, "theoretical_loss": 3.448257705859399, "tokens_seen": 1879900160 }, { "epoch": 0.14, "learning_rate": 8.692826191622533e-05, "loss": 2.6287, "theoretical_loss": 3.448237415919663, "tokens_seen": 1880031232 }, { "epoch": 0.14, "learning_rate": 8.692023752206708e-05, "loss": 2.6142, "theoretical_loss": 3.4482171277905005, "tokens_seen": 1880162304 }, { "epoch": 0.14, "learning_rate": 8.691221312790885e-05, "loss": 2.5558, "theoretical_loss": 3.448196841471624, "tokens_seen": 1880293376 }, { "epoch": 0.14, "learning_rate": 8.69041887337506e-05, "loss": 2.4915, "theoretical_loss": 3.4481765569627463, "tokens_seen": 1880424448 }, { "epoch": 0.14, "learning_rate": 8.689616433959237e-05, "loss": 2.5497, "theoretical_loss": 3.448156274263578, "tokens_seen": 1880555520 }, { "epoch": 0.14, "learning_rate": 8.688813994543412e-05, "loss": 2.507, "theoretical_loss": 3.4481359933738336, "tokens_seen": 1880686592 }, { "epoch": 0.14, "learning_rate": 8.688011555127589e-05, "loss": 2.6085, "theoretical_loss": 3.448115714293225, "tokens_seen": 1880817664 }, { "epoch": 0.14, "learning_rate": 8.687209115711764e-05, "loss": 2.612, "theoretical_loss": 3.448095437021464, "tokens_seen": 1880948736 }, { "epoch": 0.14, "objective/train/docs_used": 1035982, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5726146697998047, "objective/train/theoretical_loss": 3.448085299063812, "objective/train/tokens_used": 251473376, "theoretical_loss": 3.448085299063812, "tokens_seen": 1881014272 }, { "epoch": 0.14, "learning_rate": 8.68640667629594e-05, "loss": 2.4888, "theoretical_loss": 3.448075161558264, "tokens_seen": 1881079808 }, { "epoch": 0.14, "learning_rate": 8.685604236880116e-05, "loss": 2.6203, "theoretical_loss": 3.448054887903337, "tokens_seen": 1881210880 }, { "epoch": 0.14, "learning_rate": 8.684801797464291e-05, "loss": 2.797, "theoretical_loss": 3.448034616056396, "tokens_seen": 1881341952 }, { "epoch": 0.14, "learning_rate": 8.683999358048468e-05, "loss": 2.6205, "theoretical_loss": 3.4480143460171546, "tokens_seen": 1881473024 }, { "epoch": 0.14, "learning_rate": 8.683196918632643e-05, "loss": 2.4245, "theoretical_loss": 3.447994077785325, "tokens_seen": 1881604096 }, { "epoch": 0.14, "learning_rate": 8.68239447921682e-05, "loss": 2.6329, "theoretical_loss": 3.4479738113606198, "tokens_seen": 1881735168 }, { "epoch": 0.14, "learning_rate": 8.681592039800995e-05, "loss": 2.5879, "theoretical_loss": 3.4479535467427525, "tokens_seen": 1881866240 }, { "epoch": 0.14, "learning_rate": 8.680789600385172e-05, "loss": 2.6507, "theoretical_loss": 3.4479332839314365, "tokens_seen": 1881997312 }, { "epoch": 0.14, "learning_rate": 8.679987160969347e-05, "loss": 2.6719, "theoretical_loss": 3.447913022926384, "tokens_seen": 1882128384 }, { "epoch": 0.14, "learning_rate": 8.679184721553522e-05, "loss": 2.528, "theoretical_loss": 3.4478927637273094, "tokens_seen": 1882259456 }, { "epoch": 0.14, "learning_rate": 8.678382282137699e-05, "loss": 2.6766, "theoretical_loss": 3.4478725063339244, "tokens_seen": 1882390528 }, { "epoch": 0.14, "learning_rate": 8.677579842721874e-05, "loss": 2.5525, "theoretical_loss": 3.447852250745944, "tokens_seen": 1882521600 }, { "epoch": 0.14, "objective/train/docs_used": 1036495, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.561272144317627, "objective/train/theoretical_loss": 3.447831996963081, "objective/train/tokens_used": 253111776, "theoretical_loss": 3.447831996963081, "tokens_seen": 1882652672 }, { "epoch": 0.14, "learning_rate": 8.676777403306051e-05, "loss": 2.7183, "theoretical_loss": 3.447831996963081, "tokens_seen": 1882652672 }, { "epoch": 0.14, "learning_rate": 8.675974963890226e-05, "loss": 2.7235, "theoretical_loss": 3.4478117449850485, "tokens_seen": 1882783744 }, { "epoch": 0.14, "learning_rate": 8.675172524474403e-05, "loss": 2.5957, "theoretical_loss": 3.44779149481156, "tokens_seen": 1882914816 }, { "epoch": 0.14, "learning_rate": 8.674370085058578e-05, "loss": 2.5508, "theoretical_loss": 3.44777124644233, "tokens_seen": 1883045888 }, { "epoch": 0.14, "learning_rate": 8.673567645642754e-05, "loss": 2.5098, "theoretical_loss": 3.447750999877072, "tokens_seen": 1883176960 }, { "epoch": 0.14, "learning_rate": 8.67276520622693e-05, "loss": 2.605, "theoretical_loss": 3.4477307551154985, "tokens_seen": 1883308032 }, { "epoch": 0.14, "learning_rate": 8.671962766811106e-05, "loss": 2.7175, "theoretical_loss": 3.4477105121573244, "tokens_seen": 1883439104 }, { "epoch": 0.14, "learning_rate": 8.671160327395282e-05, "loss": 2.5376, "theoretical_loss": 3.4476902710022634, "tokens_seen": 1883570176 }, { "epoch": 0.14, "learning_rate": 8.670357887979458e-05, "loss": 2.5942, "theoretical_loss": 3.4476700316500293, "tokens_seen": 1883701248 }, { "epoch": 0.14, "learning_rate": 8.669555448563634e-05, "loss": 2.6486, "theoretical_loss": 3.447649794100336, "tokens_seen": 1883832320 }, { "epoch": 0.14, "learning_rate": 8.66875300914781e-05, "loss": 2.6782, "theoretical_loss": 3.447629558352898, "tokens_seen": 1883963392 }, { "epoch": 0.14, "learning_rate": 8.667950569731985e-05, "loss": 2.6249, "theoretical_loss": 3.447609324407429, "tokens_seen": 1884094464 }, { "epoch": 0.14, "learning_rate": 8.667148130316162e-05, "loss": 2.5889, "theoretical_loss": 3.4475890922636436, "tokens_seen": 1884225536 }, { "epoch": 0.14, "objective/train/docs_used": 1037807, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6240508556365967, "objective/train/theoretical_loss": 3.447578976867293, "objective/train/tokens_used": 254750176, "theoretical_loss": 3.447578976867293, "tokens_seen": 1884291072 }, { "epoch": 0.14, "learning_rate": 8.666345690900337e-05, "loss": 2.5063, "theoretical_loss": 3.4475688619212557, "tokens_seen": 1884356608 }, { "epoch": 0.14, "learning_rate": 8.665543251484514e-05, "loss": 2.8423, "theoretical_loss": 3.4475486333799794, "tokens_seen": 1884487680 }, { "epoch": 0.14, "learning_rate": 8.664740812068689e-05, "loss": 2.6742, "theoretical_loss": 3.4475284066395293, "tokens_seen": 1884618752 }, { "epoch": 0.14, "learning_rate": 8.663938372652866e-05, "loss": 2.7974, "theoretical_loss": 3.4475081816996207, "tokens_seen": 1884749824 }, { "epoch": 0.14, "learning_rate": 8.663135933237041e-05, "loss": 2.5465, "theoretical_loss": 3.4474879585599667, "tokens_seen": 1884880896 }, { "epoch": 0.14, "learning_rate": 8.662333493821216e-05, "loss": 2.6038, "theoretical_loss": 3.4474677372202827, "tokens_seen": 1885011968 }, { "epoch": 0.14, "learning_rate": 8.661531054405393e-05, "loss": 2.6861, "theoretical_loss": 3.4474475176802835, "tokens_seen": 1885143040 }, { "epoch": 0.14, "learning_rate": 8.660728614989568e-05, "loss": 2.4443, "theoretical_loss": 3.447427299939683, "tokens_seen": 1885274112 }, { "epoch": 0.14, "learning_rate": 8.659926175573745e-05, "loss": 2.4831, "theoretical_loss": 3.4474070839981974, "tokens_seen": 1885405184 }, { "epoch": 0.14, "learning_rate": 8.65912373615792e-05, "loss": 2.6502, "theoretical_loss": 3.4473868698555403, "tokens_seen": 1885536256 }, { "epoch": 0.14, "learning_rate": 8.658321296742097e-05, "loss": 2.6768, "theoretical_loss": 3.4473666575114263, "tokens_seen": 1885667328 }, { "epoch": 0.14, "learning_rate": 8.657518857326272e-05, "loss": 2.5946, "theoretical_loss": 3.4473464469655717, "tokens_seen": 1885798400 }, { "epoch": 0.14, "objective/train/docs_used": 1038424, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4946327209472656, "objective/train/theoretical_loss": 3.4473262382176912, "objective/train/tokens_used": 256388576, "theoretical_loss": 3.4473262382176912, "tokens_seen": 1885929472 }, { "epoch": 0.14, "learning_rate": 8.656716417910447e-05, "loss": 2.6475, "theoretical_loss": 3.4473262382176912, "tokens_seen": 1885929472 }, { "epoch": 0.14, "learning_rate": 8.655913978494624e-05, "loss": 2.5066, "theoretical_loss": 3.447306031267499, "tokens_seen": 1886060544 }, { "epoch": 0.14, "learning_rate": 8.6551115390788e-05, "loss": 2.7395, "theoretical_loss": 3.4472858261147112, "tokens_seen": 1886191616 }, { "epoch": 0.14, "learning_rate": 8.654309099662976e-05, "loss": 2.6795, "theoretical_loss": 3.4472656227590424, "tokens_seen": 1886322688 }, { "epoch": 0.14, "learning_rate": 8.653506660247151e-05, "loss": 2.8047, "theoretical_loss": 3.4472454212002086, "tokens_seen": 1886453760 }, { "epoch": 0.14, "learning_rate": 8.652704220831328e-05, "loss": 2.4961, "theoretical_loss": 3.4472252214379244, "tokens_seen": 1886584832 }, { "epoch": 0.14, "learning_rate": 8.651901781415503e-05, "loss": 2.5358, "theoretical_loss": 3.447205023471906, "tokens_seen": 1886715904 }, { "epoch": 0.14, "learning_rate": 8.65109934199968e-05, "loss": 2.6046, "theoretical_loss": 3.4471848273018684, "tokens_seen": 1886846976 }, { "epoch": 0.14, "learning_rate": 8.650296902583855e-05, "loss": 2.5741, "theoretical_loss": 3.447164632927527, "tokens_seen": 1886978048 }, { "epoch": 0.14, "learning_rate": 8.649494463168031e-05, "loss": 2.6877, "theoretical_loss": 3.4471444403485982, "tokens_seen": 1887109120 }, { "epoch": 0.14, "learning_rate": 8.648692023752207e-05, "loss": 2.652, "theoretical_loss": 3.447124249564797, "tokens_seen": 1887240192 }, { "epoch": 0.14, "learning_rate": 8.647889584336383e-05, "loss": 2.5121, "theoretical_loss": 3.447104060575839, "tokens_seen": 1887371264 }, { "epoch": 0.14, "learning_rate": 8.64708714492056e-05, "loss": 2.5847, "theoretical_loss": 3.44708387338144, "tokens_seen": 1887502336 }, { "epoch": 0.14, "objective/train/docs_used": 1039538, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6100270748138428, "objective/train/theoretical_loss": 3.447073780457112, "objective/train/tokens_used": 258026976, "theoretical_loss": 3.447073780457112, "tokens_seen": 1887567872 }, { "epoch": 0.14, "learning_rate": 8.646284705504735e-05, "loss": 2.7654, "theoretical_loss": 3.4470636879813172, "tokens_seen": 1887633408 }, { "epoch": 0.14, "learning_rate": 8.645482266088911e-05, "loss": 2.5572, "theoretical_loss": 3.447043504375185, "tokens_seen": 1887764480 }, { "epoch": 0.14, "learning_rate": 8.644679826673087e-05, "loss": 2.5704, "theoretical_loss": 3.4470233225627602, "tokens_seen": 1887895552 }, { "epoch": 0.14, "learning_rate": 8.643877387257262e-05, "loss": 2.6712, "theoretical_loss": 3.4470031425437586, "tokens_seen": 1888026624 }, { "epoch": 0.14, "learning_rate": 8.643074947841439e-05, "loss": 2.435, "theoretical_loss": 3.4469829643178964, "tokens_seen": 1888157696 }, { "epoch": 0.14, "learning_rate": 8.642272508425614e-05, "loss": 2.7374, "theoretical_loss": 3.44696278788489, "tokens_seen": 1888288768 }, { "epoch": 0.14, "learning_rate": 8.64147006900979e-05, "loss": 2.6685, "theoretical_loss": 3.446942613244455, "tokens_seen": 1888419840 }, { "epoch": 0.14, "learning_rate": 8.640667629593966e-05, "loss": 2.6827, "theoretical_loss": 3.4469224403963086, "tokens_seen": 1888550912 }, { "epoch": 0.14, "learning_rate": 8.639865190178143e-05, "loss": 2.5129, "theoretical_loss": 3.4469022693401667, "tokens_seen": 1888681984 }, { "epoch": 0.14, "learning_rate": 8.639062750762318e-05, "loss": 2.7268, "theoretical_loss": 3.4468821000757464, "tokens_seen": 1888813056 }, { "epoch": 0.14, "learning_rate": 8.638260311346493e-05, "loss": 2.5836, "theoretical_loss": 3.446861932602763, "tokens_seen": 1888944128 }, { "epoch": 0.14, "learning_rate": 8.63745787193067e-05, "loss": 2.5723, "theoretical_loss": 3.446841766920934, "tokens_seen": 1889075200 }, { "epoch": 0.14, "objective/train/docs_used": 1040979, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.302525758743286, "objective/train/theoretical_loss": 3.446821603029976, "objective/train/tokens_used": 259665376, "theoretical_loss": 3.446821603029976, "tokens_seen": 1889206272 }, { "epoch": 0.14, "learning_rate": 8.636655432514845e-05, "loss": 2.5104, "theoretical_loss": 3.446821603029976, "tokens_seen": 1889206272 }, { "epoch": 0.15, "learning_rate": 8.635852993099022e-05, "loss": 2.7808, "theoretical_loss": 3.446801440929606, "tokens_seen": 1889337344 }, { "epoch": 0.15, "learning_rate": 8.635050553683197e-05, "loss": 2.6051, "theoretical_loss": 3.4467812806195397, "tokens_seen": 1889468416 }, { "epoch": 0.15, "learning_rate": 8.634248114267374e-05, "loss": 2.4754, "theoretical_loss": 3.4467611220994954, "tokens_seen": 1889599488 }, { "epoch": 0.15, "learning_rate": 8.633445674851549e-05, "loss": 2.5734, "theoretical_loss": 3.4467409653691887, "tokens_seen": 1889730560 }, { "epoch": 0.15, "learning_rate": 8.632643235435724e-05, "loss": 2.636, "theoretical_loss": 3.4467208104283373, "tokens_seen": 1889861632 }, { "epoch": 0.15, "learning_rate": 8.631840796019901e-05, "loss": 2.5367, "theoretical_loss": 3.446700657276658, "tokens_seen": 1889992704 }, { "epoch": 0.15, "learning_rate": 8.631038356604076e-05, "loss": 2.6323, "theoretical_loss": 3.446680505913868, "tokens_seen": 1890123776 }, { "epoch": 0.15, "learning_rate": 8.630235917188253e-05, "loss": 2.6413, "theoretical_loss": 3.4466603563396845, "tokens_seen": 1890254848 }, { "epoch": 0.15, "learning_rate": 8.629433477772428e-05, "loss": 2.6791, "theoretical_loss": 3.4466402085538252, "tokens_seen": 1890385920 }, { "epoch": 0.15, "learning_rate": 8.628631038356605e-05, "loss": 2.63, "theoretical_loss": 3.4466200625560064, "tokens_seen": 1890516992 }, { "epoch": 0.15, "learning_rate": 8.62782859894078e-05, "loss": 2.6905, "theoretical_loss": 3.446599918345946, "tokens_seen": 1890648064 }, { "epoch": 0.15, "learning_rate": 8.627026159524956e-05, "loss": 2.6284, "theoretical_loss": 3.446579775923362, "tokens_seen": 1890779136 }, { "epoch": 0.15, "objective/train/docs_used": 1041612, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.944444417953491, "objective/train/theoretical_loss": 3.4465697053822844, "objective/train/tokens_used": 261303776, "theoretical_loss": 3.4465697053822844, "tokens_seen": 1890844672 }, { "epoch": 0.15, "learning_rate": 8.626223720109132e-05, "loss": 2.6467, "theoretical_loss": 3.4465596352879704, "tokens_seen": 1890910208 }, { "epoch": 0.15, "learning_rate": 8.625421280693308e-05, "loss": 2.4845, "theoretical_loss": 3.44653949643949, "tokens_seen": 1891041280 }, { "epoch": 0.15, "learning_rate": 8.624618841277484e-05, "loss": 2.5807, "theoretical_loss": 3.446519359377638, "tokens_seen": 1891172352 }, { "epoch": 0.15, "learning_rate": 8.62381640186166e-05, "loss": 2.6668, "theoretical_loss": 3.4464992241021326, "tokens_seen": 1891303424 }, { "epoch": 0.15, "learning_rate": 8.623013962445836e-05, "loss": 2.4901, "theoretical_loss": 3.4464790906126908, "tokens_seen": 1891434496 }, { "epoch": 0.15, "learning_rate": 8.622211523030012e-05, "loss": 2.7128, "theoretical_loss": 3.446458958909031, "tokens_seen": 1891565568 }, { "epoch": 0.15, "learning_rate": 8.621409083614188e-05, "loss": 2.6439, "theoretical_loss": 3.4464388289908703, "tokens_seen": 1891696640 }, { "epoch": 0.15, "learning_rate": 8.620606644198364e-05, "loss": 2.6116, "theoretical_loss": 3.4464187008579277, "tokens_seen": 1891827712 }, { "epoch": 0.15, "learning_rate": 8.619804204782539e-05, "loss": 2.5137, "theoretical_loss": 3.446398574509921, "tokens_seen": 1891958784 }, { "epoch": 0.15, "learning_rate": 8.619001765366716e-05, "loss": 2.5921, "theoretical_loss": 3.4463784499465673, "tokens_seen": 1892089856 }, { "epoch": 0.15, "learning_rate": 8.618199325950891e-05, "loss": 2.5911, "theoretical_loss": 3.4463583271675855, "tokens_seen": 1892220928 }, { "epoch": 0.15, "learning_rate": 8.617396886535068e-05, "loss": 2.5032, "theoretical_loss": 3.446338206172694, "tokens_seen": 1892352000 }, { "epoch": 0.15, "objective/train/docs_used": 1042686, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5551538467407227, "objective/train/theoretical_loss": 3.44631808696161, "objective/train/tokens_used": 262942176, "theoretical_loss": 3.44631808696161, "tokens_seen": 1892483072 }, { "epoch": 0.15, "learning_rate": 8.616594447119243e-05, "loss": 2.7494, "theoretical_loss": 3.44631808696161, "tokens_seen": 1892483072 }, { "epoch": 0.15, "learning_rate": 8.61579200770342e-05, "loss": 2.5324, "theoretical_loss": 3.4462979695340534, "tokens_seen": 1892614144 }, { "epoch": 0.15, "learning_rate": 8.614989568287595e-05, "loss": 2.5766, "theoretical_loss": 3.446277853889742, "tokens_seen": 1892745216 }, { "epoch": 0.15, "learning_rate": 8.61418712887177e-05, "loss": 2.5912, "theoretical_loss": 3.4462577400283934, "tokens_seen": 1892876288 }, { "epoch": 0.15, "learning_rate": 8.613384689455947e-05, "loss": 2.5151, "theoretical_loss": 3.446237627949727, "tokens_seen": 1893007360 }, { "epoch": 0.15, "learning_rate": 8.612582250040122e-05, "loss": 2.7351, "theoretical_loss": 3.4462175176534613, "tokens_seen": 1893138432 }, { "epoch": 0.15, "learning_rate": 8.611779810624299e-05, "loss": 2.4694, "theoretical_loss": 3.446197409139314, "tokens_seen": 1893269504 }, { "epoch": 0.15, "learning_rate": 8.610977371208474e-05, "loss": 2.7383, "theoretical_loss": 3.4461773024070057, "tokens_seen": 1893400576 }, { "epoch": 0.15, "learning_rate": 8.610174931792651e-05, "loss": 2.2325, "theoretical_loss": 3.4461571974562535, "tokens_seen": 1893531648 }, { "epoch": 0.15, "learning_rate": 8.609372492376826e-05, "loss": 2.5378, "theoretical_loss": 3.446137094286777, "tokens_seen": 1893662720 }, { "epoch": 0.15, "learning_rate": 8.608570052961001e-05, "loss": 2.6805, "theoretical_loss": 3.446116992898295, "tokens_seen": 1893793792 }, { "epoch": 0.15, "learning_rate": 8.607767613545178e-05, "loss": 2.5466, "theoretical_loss": 3.4460968932905263, "tokens_seen": 1893924864 }, { "epoch": 0.15, "learning_rate": 8.606965174129353e-05, "loss": 2.6135, "theoretical_loss": 3.44607679546319, "tokens_seen": 1894055936 }, { "epoch": 0.15, "objective/train/docs_used": 1043318, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7247536182403564, "objective/train/theoretical_loss": 3.4460667472170963, "objective/train/tokens_used": 264580576, "theoretical_loss": 3.4460667472170963, "tokens_seen": 1894121472 }, { "epoch": 0.15, "learning_rate": 8.60616273471353e-05, "loss": 2.5337, "theoretical_loss": 3.446056699416005, "tokens_seen": 1894187008 }, { "epoch": 0.15, "learning_rate": 8.605360295297705e-05, "loss": 2.5264, "theoretical_loss": 3.4460366051486915, "tokens_seen": 1894318080 }, { "epoch": 0.15, "learning_rate": 8.604557855881882e-05, "loss": 2.5163, "theoretical_loss": 3.446016512660967, "tokens_seen": 1894449152 }, { "epoch": 0.15, "learning_rate": 8.603755416466057e-05, "loss": 2.7092, "theoretical_loss": 3.445996421952552, "tokens_seen": 1894580224 }, { "epoch": 0.15, "learning_rate": 8.602952977050233e-05, "loss": 2.5607, "theoretical_loss": 3.445976333023166, "tokens_seen": 1894711296 }, { "epoch": 0.15, "learning_rate": 8.60215053763441e-05, "loss": 2.5803, "theoretical_loss": 3.4459562458725275, "tokens_seen": 1894842368 }, { "epoch": 0.15, "learning_rate": 8.601348098218585e-05, "loss": 2.594, "theoretical_loss": 3.4459361605003567, "tokens_seen": 1894973440 }, { "epoch": 0.15, "learning_rate": 8.600545658802761e-05, "loss": 2.8107, "theoretical_loss": 3.4459160769063724, "tokens_seen": 1895104512 }, { "epoch": 0.15, "learning_rate": 8.599743219386937e-05, "loss": 2.5098, "theoretical_loss": 3.4458959950902948, "tokens_seen": 1895235584 }, { "epoch": 0.15, "learning_rate": 8.598940779971113e-05, "loss": 2.6863, "theoretical_loss": 3.445875915051844, "tokens_seen": 1895366656 }, { "epoch": 0.15, "learning_rate": 8.598138340555289e-05, "loss": 2.6153, "theoretical_loss": 3.445855836790739, "tokens_seen": 1895497728 }, { "epoch": 0.15, "learning_rate": 8.597335901139464e-05, "loss": 2.2966, "theoretical_loss": 3.4458357603066996, "tokens_seen": 1895628800 }, { "epoch": 0.15, "objective/train/docs_used": 1044449, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.796210289001465, "objective/train/theoretical_loss": 3.4458156855994457, "objective/train/tokens_used": 266218976, "theoretical_loss": 3.4458156855994457, "tokens_seen": 1895759872 }, { "epoch": 0.15, "learning_rate": 8.59653346172364e-05, "loss": 2.6243, "theoretical_loss": 3.4458156855994457, "tokens_seen": 1895759872 }, { "epoch": 0.15, "learning_rate": 8.595731022307816e-05, "loss": 2.5802, "theoretical_loss": 3.4457956126686975, "tokens_seen": 1895890944 }, { "epoch": 0.15, "learning_rate": 8.594928582891993e-05, "loss": 2.5255, "theoretical_loss": 3.445775541514175, "tokens_seen": 1896022016 }, { "epoch": 0.15, "learning_rate": 8.594126143476168e-05, "loss": 2.5969, "theoretical_loss": 3.4457554721355983, "tokens_seen": 1896153088 }, { "epoch": 0.15, "learning_rate": 8.593323704060345e-05, "loss": 2.5436, "theoretical_loss": 3.4457354045326873, "tokens_seen": 1896284160 }, { "epoch": 0.15, "learning_rate": 8.59252126464452e-05, "loss": 2.5276, "theoretical_loss": 3.4457153387051616, "tokens_seen": 1896415232 }, { "epoch": 0.15, "learning_rate": 8.591718825228697e-05, "loss": 2.4972, "theoretical_loss": 3.4456952746527425, "tokens_seen": 1896546304 }, { "epoch": 0.15, "learning_rate": 8.590916385812872e-05, "loss": 2.7186, "theoretical_loss": 3.44567521237515, "tokens_seen": 1896677376 }, { "epoch": 0.15, "learning_rate": 8.590113946397047e-05, "loss": 2.5623, "theoretical_loss": 3.4456551518721037, "tokens_seen": 1896808448 }, { "epoch": 0.15, "learning_rate": 8.589311506981224e-05, "loss": 2.7065, "theoretical_loss": 3.4456350931433253, "tokens_seen": 1896939520 }, { "epoch": 0.15, "learning_rate": 8.588509067565399e-05, "loss": 2.5867, "theoretical_loss": 3.445615036188534, "tokens_seen": 1897070592 }, { "epoch": 0.15, "learning_rate": 8.587706628149576e-05, "loss": 2.5036, "theoretical_loss": 3.4455949810074515, "tokens_seen": 1897201664 }, { "epoch": 0.15, "learning_rate": 8.586904188733751e-05, "loss": 2.444, "theoretical_loss": 3.4455749275997976, "tokens_seen": 1897332736 }, { "epoch": 0.15, "objective/train/docs_used": 1045116, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.431305408477783, "objective/train/theoretical_loss": 3.4455649015609193, "objective/train/tokens_used": 267857376, "theoretical_loss": 3.4455649015609193, "tokens_seen": 1897398272 }, { "epoch": 0.15, "learning_rate": 8.586101749317928e-05, "loss": 2.4496, "theoretical_loss": 3.4455548759652936, "tokens_seen": 1897463808 }, { "epoch": 0.15, "learning_rate": 8.585299309902103e-05, "loss": 2.6165, "theoretical_loss": 3.4455348261036596, "tokens_seen": 1897594880 }, { "epoch": 0.15, "learning_rate": 8.584496870486278e-05, "loss": 2.5878, "theoretical_loss": 3.4455147780146174, "tokens_seen": 1897725952 }, { "epoch": 0.15, "learning_rate": 8.583694431070455e-05, "loss": 2.5, "theoretical_loss": 3.4454947316978863, "tokens_seen": 1897857024 }, { "epoch": 0.15, "learning_rate": 8.58289199165463e-05, "loss": 2.6347, "theoretical_loss": 3.445474687153189, "tokens_seen": 1897988096 }, { "epoch": 0.15, "learning_rate": 8.582089552238807e-05, "loss": 2.5567, "theoretical_loss": 3.4454546443802454, "tokens_seen": 1898119168 }, { "epoch": 0.15, "learning_rate": 8.581287112822982e-05, "loss": 2.5346, "theoretical_loss": 3.445434603378777, "tokens_seen": 1898250240 }, { "epoch": 0.15, "learning_rate": 8.580484673407159e-05, "loss": 2.698, "theoretical_loss": 3.4454145641485043, "tokens_seen": 1898381312 }, { "epoch": 0.15, "learning_rate": 8.579682233991334e-05, "loss": 2.5593, "theoretical_loss": 3.4453945266891486, "tokens_seen": 1898512384 }, { "epoch": 0.15, "learning_rate": 8.57887979457551e-05, "loss": 2.5686, "theoretical_loss": 3.4453744910004325, "tokens_seen": 1898643456 }, { "epoch": 0.15, "learning_rate": 8.578077355159686e-05, "loss": 2.5992, "theoretical_loss": 3.4453544570820758, "tokens_seen": 1898774528 }, { "epoch": 0.15, "learning_rate": 8.577274915743862e-05, "loss": 2.5495, "theoretical_loss": 3.4453344249338005, "tokens_seen": 1898905600 }, { "epoch": 0.15, "objective/train/docs_used": 1046217, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.568467855453491, "objective/train/theoretical_loss": 3.4453143945553277, "objective/train/tokens_used": 269495776, "theoretical_loss": 3.4453143945553277, "tokens_seen": 1899036672 }, { "epoch": 0.15, "learning_rate": 8.576472476328038e-05, "loss": 2.6635, "theoretical_loss": 3.4453143945553277, "tokens_seen": 1899036672 }, { "epoch": 0.15, "learning_rate": 8.575670036912214e-05, "loss": 2.591, "theoretical_loss": 3.4452943659463795, "tokens_seen": 1899167744 }, { "epoch": 0.15, "learning_rate": 8.57486759749639e-05, "loss": 2.6946, "theoretical_loss": 3.445274339106677, "tokens_seen": 1899298816 }, { "epoch": 0.15, "learning_rate": 8.574065158080566e-05, "loss": 2.5766, "theoretical_loss": 3.4452543140359415, "tokens_seen": 1899429888 }, { "epoch": 0.15, "learning_rate": 8.573262718664741e-05, "loss": 2.6189, "theoretical_loss": 3.4452342907338958, "tokens_seen": 1899560960 }, { "epoch": 0.15, "learning_rate": 8.572460279248918e-05, "loss": 2.556, "theoretical_loss": 3.4452142692002603, "tokens_seen": 1899692032 }, { "epoch": 0.15, "learning_rate": 8.571657839833093e-05, "loss": 2.5013, "theoretical_loss": 3.445194249434758, "tokens_seen": 1899823104 }, { "epoch": 0.15, "learning_rate": 8.57085540041727e-05, "loss": 2.8092, "theoretical_loss": 3.4451742314371105, "tokens_seen": 1899954176 }, { "epoch": 0.15, "learning_rate": 8.570052961001445e-05, "loss": 2.6131, "theoretical_loss": 3.4451542152070393, "tokens_seen": 1900085248 }, { "epoch": 0.15, "learning_rate": 8.569250521585622e-05, "loss": 2.5599, "theoretical_loss": 3.445134200744267, "tokens_seen": 1900216320 }, { "epoch": 0.15, "learning_rate": 8.568448082169797e-05, "loss": 2.4855, "theoretical_loss": 3.4451141880485148, "tokens_seen": 1900347392 }, { "epoch": 0.15, "learning_rate": 8.567645642753972e-05, "loss": 2.5664, "theoretical_loss": 3.445094177119505, "tokens_seen": 1900478464 }, { "epoch": 0.15, "learning_rate": 8.566843203338149e-05, "loss": 2.6017, "theoretical_loss": 3.4450741679569608, "tokens_seen": 1900609536 }, { "epoch": 0.15, "objective/train/docs_used": 1047224, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.664011001586914, "objective/train/theoretical_loss": 3.445064164038026, "objective/train/tokens_used": 271134176, "theoretical_loss": 3.445064164038026, "tokens_seen": 1900675072 }, { "epoch": 0.15, "learning_rate": 8.566040763922324e-05, "loss": 2.6229, "theoretical_loss": 3.4450541605606038, "tokens_seen": 1900740608 }, { "epoch": 0.15, "learning_rate": 8.565238324506501e-05, "loss": 2.5607, "theoretical_loss": 3.4450341549301564, "tokens_seen": 1900871680 }, { "epoch": 0.15, "learning_rate": 8.564435885090676e-05, "loss": 2.4586, "theoretical_loss": 3.4450141510653403, "tokens_seen": 1901002752 }, { "epoch": 0.15, "learning_rate": 8.563633445674853e-05, "loss": 2.4271, "theoretical_loss": 3.444994148965879, "tokens_seen": 1901133824 }, { "epoch": 0.15, "learning_rate": 8.562831006259028e-05, "loss": 2.6382, "theoretical_loss": 3.444974148631495, "tokens_seen": 1901264896 }, { "epoch": 0.15, "learning_rate": 8.562028566843204e-05, "loss": 2.6207, "theoretical_loss": 3.4449541500619096, "tokens_seen": 1901395968 }, { "epoch": 0.15, "learning_rate": 8.56122612742738e-05, "loss": 2.6704, "theoretical_loss": 3.4449341532568467, "tokens_seen": 1901527040 }, { "epoch": 0.15, "learning_rate": 8.560423688011555e-05, "loss": 2.7467, "theoretical_loss": 3.4449141582160285, "tokens_seen": 1901658112 }, { "epoch": 0.15, "learning_rate": 8.559621248595732e-05, "loss": 2.6002, "theoretical_loss": 3.4448941649391784, "tokens_seen": 1901789184 }, { "epoch": 0.15, "learning_rate": 8.558818809179907e-05, "loss": 2.6505, "theoretical_loss": 3.444874173426018, "tokens_seen": 1901920256 }, { "epoch": 0.15, "learning_rate": 8.558016369764084e-05, "loss": 2.5797, "theoretical_loss": 3.444854183676271, "tokens_seen": 1902051328 }, { "epoch": 0.15, "learning_rate": 8.55721393034826e-05, "loss": 2.5589, "theoretical_loss": 3.44483419568966, "tokens_seen": 1902182400 }, { "epoch": 0.15, "objective/train/docs_used": 1047772, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.691408395767212, "objective/train/theoretical_loss": 3.444814209465909, "objective/train/tokens_used": 272772576, "theoretical_loss": 3.444814209465909, "tokens_seen": 1902313472 }, { "epoch": 0.15, "learning_rate": 8.556411490932436e-05, "loss": 2.6686, "theoretical_loss": 3.444814209465909, "tokens_seen": 1902313472 }, { "epoch": 0.15, "learning_rate": 8.555609051516611e-05, "loss": 2.7427, "theoretical_loss": 3.44479422500474, "tokens_seen": 1902444544 }, { "epoch": 0.15, "learning_rate": 8.554806612100787e-05, "loss": 2.5318, "theoretical_loss": 3.4447742423058765, "tokens_seen": 1902575616 }, { "epoch": 0.15, "learning_rate": 8.554004172684963e-05, "loss": 2.5627, "theoretical_loss": 3.4447542613690416, "tokens_seen": 1902706688 }, { "epoch": 0.15, "learning_rate": 8.553201733269139e-05, "loss": 2.6165, "theoretical_loss": 3.4447342821939584, "tokens_seen": 1902837760 }, { "epoch": 0.15, "learning_rate": 8.552399293853315e-05, "loss": 2.651, "theoretical_loss": 3.4447143047803506, "tokens_seen": 1902968832 }, { "epoch": 0.15, "learning_rate": 8.551596854437491e-05, "loss": 2.5404, "theoretical_loss": 3.444694329127941, "tokens_seen": 1903099904 }, { "epoch": 0.15, "learning_rate": 8.550794415021667e-05, "loss": 2.6022, "theoretical_loss": 3.4446743552364545, "tokens_seen": 1903230976 }, { "epoch": 0.15, "learning_rate": 8.549991975605843e-05, "loss": 2.5531, "theoretical_loss": 3.444654383105613, "tokens_seen": 1903362048 }, { "epoch": 0.15, "learning_rate": 8.549189536190018e-05, "loss": 2.6862, "theoretical_loss": 3.44463441273514, "tokens_seen": 1903493120 }, { "epoch": 0.15, "learning_rate": 8.548387096774195e-05, "loss": 2.7709, "theoretical_loss": 3.4446144441247606, "tokens_seen": 1903624192 }, { "epoch": 0.15, "learning_rate": 8.54758465735837e-05, "loss": 2.5729, "theoretical_loss": 3.4445944772741974, "tokens_seen": 1903755264 }, { "epoch": 0.15, "learning_rate": 8.546782217942547e-05, "loss": 2.6571, "theoretical_loss": 3.444574512183175, "tokens_seen": 1903886336 }, { "epoch": 0.15, "objective/train/docs_used": 1048883, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7661850452423096, "objective/train/theoretical_loss": 3.444564530297405, "objective/train/tokens_used": 274410976, "theoretical_loss": 3.444564530297405, "tokens_seen": 1903951872 }, { "epoch": 0.15, "learning_rate": 8.545979778526722e-05, "loss": 2.6316, "theoretical_loss": 3.444554548851416, "tokens_seen": 1904017408 }, { "epoch": 0.15, "learning_rate": 8.545177339110899e-05, "loss": 2.3924, "theoretical_loss": 3.4445345872786453, "tokens_seen": 1904148480 }, { "epoch": 0.15, "learning_rate": 8.544374899695074e-05, "loss": 2.6117, "theoretical_loss": 3.444514627464587, "tokens_seen": 1904279552 }, { "epoch": 0.15, "learning_rate": 8.543572460279249e-05, "loss": 2.5, "theoretical_loss": 3.4444946694089635, "tokens_seen": 1904410624 }, { "epoch": 0.15, "learning_rate": 8.542770020863426e-05, "loss": 2.6789, "theoretical_loss": 3.444474713111501, "tokens_seen": 1904541696 }, { "epoch": 0.15, "learning_rate": 8.541967581447601e-05, "loss": 2.5173, "theoretical_loss": 3.444454758571922, "tokens_seen": 1904672768 }, { "epoch": 0.15, "learning_rate": 8.541165142031778e-05, "loss": 2.6405, "theoretical_loss": 3.444434805789952, "tokens_seen": 1904803840 }, { "epoch": 0.15, "learning_rate": 8.540362702615953e-05, "loss": 2.5492, "theoretical_loss": 3.444414854765314, "tokens_seen": 1904934912 }, { "epoch": 0.15, "learning_rate": 8.53956026320013e-05, "loss": 2.4519, "theoretical_loss": 3.4443949054977336, "tokens_seen": 1905065984 }, { "epoch": 0.15, "learning_rate": 8.538757823784305e-05, "loss": 2.4236, "theoretical_loss": 3.444374957986934, "tokens_seen": 1905197056 }, { "epoch": 0.15, "learning_rate": 8.53795538436848e-05, "loss": 2.5186, "theoretical_loss": 3.4443550122326405, "tokens_seen": 1905328128 }, { "epoch": 0.15, "learning_rate": 8.537152944952657e-05, "loss": 2.5571, "theoretical_loss": 3.444335068234577, "tokens_seen": 1905459200 }, { "epoch": 0.15, "objective/train/docs_used": 1049481, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8444793224334717, "objective/train/theoretical_loss": 3.4443151259924685, "objective/train/tokens_used": 276049376, "theoretical_loss": 3.4443151259924685, "tokens_seen": 1905590272 }, { "epoch": 0.15, "learning_rate": 8.536350505536833e-05, "loss": 2.7023, "theoretical_loss": 3.4443151259924685, "tokens_seen": 1905590272 }, { "epoch": 0.15, "learning_rate": 8.535548066121009e-05, "loss": 2.5393, "theoretical_loss": 3.4442951855060393, "tokens_seen": 1905721344 }, { "epoch": 0.16, "learning_rate": 8.534745626705184e-05, "loss": 2.5215, "theoretical_loss": 3.444275246775015, "tokens_seen": 1905852416 }, { "epoch": 0.16, "learning_rate": 8.533943187289361e-05, "loss": 2.6303, "theoretical_loss": 3.444255309799119, "tokens_seen": 1905983488 }, { "epoch": 0.16, "learning_rate": 8.533140747873536e-05, "loss": 2.6693, "theoretical_loss": 3.4442353745780765, "tokens_seen": 1906114560 }, { "epoch": 0.16, "learning_rate": 8.532338308457712e-05, "loss": 2.6018, "theoretical_loss": 3.444215441111613, "tokens_seen": 1906245632 }, { "epoch": 0.16, "learning_rate": 8.531535869041888e-05, "loss": 2.6154, "theoretical_loss": 3.444195509399453, "tokens_seen": 1906376704 }, { "epoch": 0.16, "learning_rate": 8.530733429626064e-05, "loss": 2.6642, "theoretical_loss": 3.444175579441322, "tokens_seen": 1906507776 }, { "epoch": 0.16, "learning_rate": 8.52993099021024e-05, "loss": 2.5281, "theoretical_loss": 3.4441556512369442, "tokens_seen": 1906638848 }, { "epoch": 0.16, "learning_rate": 8.529128550794416e-05, "loss": 2.5748, "theoretical_loss": 3.4441357247860456, "tokens_seen": 1906769920 }, { "epoch": 0.16, "learning_rate": 8.528326111378591e-05, "loss": 2.4962, "theoretical_loss": 3.4441158000883507, "tokens_seen": 1906900992 }, { "epoch": 0.16, "learning_rate": 8.527523671962768e-05, "loss": 2.6538, "theoretical_loss": 3.4440958771435852, "tokens_seen": 1907032064 }, { "epoch": 0.16, "learning_rate": 8.526721232546943e-05, "loss": 2.7252, "theoretical_loss": 3.444075955951474, "tokens_seen": 1907163136 }, { "epoch": 0.16, "objective/train/docs_used": 1050597, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.0899901390075684, "objective/train/theoretical_loss": 3.4440659960125783, "objective/train/tokens_used": 277687776, "theoretical_loss": 3.4440659960125783, "tokens_seen": 1907228672 }, { "epoch": 0.16, "learning_rate": 8.52591879313112e-05, "loss": 2.5514, "theoretical_loss": 3.4440560365117427, "tokens_seen": 1907294208 }, { "epoch": 0.16, "learning_rate": 8.525116353715295e-05, "loss": 2.5941, "theoretical_loss": 3.4440361188241173, "tokens_seen": 1907425280 }, { "epoch": 0.16, "learning_rate": 8.524313914299472e-05, "loss": 2.58, "theoretical_loss": 3.4440162028883226, "tokens_seen": 1907556352 }, { "epoch": 0.16, "learning_rate": 8.523511474883647e-05, "loss": 2.3977, "theoretical_loss": 3.443996288704084, "tokens_seen": 1907687424 }, { "epoch": 0.16, "learning_rate": 8.522709035467822e-05, "loss": 2.5847, "theoretical_loss": 3.443976376271128, "tokens_seen": 1907818496 }, { "epoch": 0.16, "learning_rate": 8.521906596051999e-05, "loss": 2.6012, "theoretical_loss": 3.4439564655891797, "tokens_seen": 1907949568 }, { "epoch": 0.16, "learning_rate": 8.521104156636174e-05, "loss": 2.6382, "theoretical_loss": 3.4439365566579645, "tokens_seen": 1908080640 }, { "epoch": 0.16, "learning_rate": 8.520301717220351e-05, "loss": 2.4888, "theoretical_loss": 3.4439166494772087, "tokens_seen": 1908211712 }, { "epoch": 0.16, "learning_rate": 8.519499277804526e-05, "loss": 2.7036, "theoretical_loss": 3.4438967440466386, "tokens_seen": 1908342784 }, { "epoch": 0.16, "learning_rate": 8.518696838388702e-05, "loss": 2.6248, "theoretical_loss": 3.443876840365979, "tokens_seen": 1908473856 }, { "epoch": 0.16, "learning_rate": 8.517894398972878e-05, "loss": 2.6627, "theoretical_loss": 3.443856938434957, "tokens_seen": 1908604928 }, { "epoch": 0.16, "learning_rate": 8.517091959557054e-05, "loss": 2.5851, "theoretical_loss": 3.443837038253298, "tokens_seen": 1908736000 }, { "epoch": 0.16, "objective/train/docs_used": 1051040, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.524304151535034, "objective/train/theoretical_loss": 3.4438171398207285, "objective/train/tokens_used": 279326176, "theoretical_loss": 3.4438171398207285, "tokens_seen": 1908867072 }, { "epoch": 0.16, "learning_rate": 8.51628952014123e-05, "loss": 2.5677, "theoretical_loss": 3.4438171398207285, "tokens_seen": 1908867072 }, { "epoch": 0.16, "learning_rate": 8.515487080725406e-05, "loss": 2.5599, "theoretical_loss": 3.4437972431369746, "tokens_seen": 1908998144 }, { "epoch": 0.16, "learning_rate": 8.514684641309582e-05, "loss": 2.7516, "theoretical_loss": 3.443777348201762, "tokens_seen": 1909129216 }, { "epoch": 0.16, "learning_rate": 8.513882201893758e-05, "loss": 2.4558, "theoretical_loss": 3.4437574550148176, "tokens_seen": 1909260288 }, { "epoch": 0.16, "learning_rate": 8.513079762477933e-05, "loss": 2.6508, "theoretical_loss": 3.4437375635758674, "tokens_seen": 1909391360 }, { "epoch": 0.16, "learning_rate": 8.51227732306211e-05, "loss": 2.6752, "theoretical_loss": 3.4437176738846382, "tokens_seen": 1909522432 }, { "epoch": 0.16, "learning_rate": 8.511474883646285e-05, "loss": 2.6269, "theoretical_loss": 3.4436977859408566, "tokens_seen": 1909653504 }, { "epoch": 0.16, "learning_rate": 8.510672444230461e-05, "loss": 2.6323, "theoretical_loss": 3.443677899744249, "tokens_seen": 1909784576 }, { "epoch": 0.16, "learning_rate": 8.509870004814637e-05, "loss": 2.6092, "theoretical_loss": 3.4436580152945413, "tokens_seen": 1909915648 }, { "epoch": 0.16, "learning_rate": 8.509067565398812e-05, "loss": 2.6991, "theoretical_loss": 3.4436381325914613, "tokens_seen": 1910046720 }, { "epoch": 0.16, "learning_rate": 8.508265125982989e-05, "loss": 2.7202, "theoretical_loss": 3.443618251634735, "tokens_seen": 1910177792 }, { "epoch": 0.16, "learning_rate": 8.507462686567164e-05, "loss": 2.397, "theoretical_loss": 3.443598372424089, "tokens_seen": 1910308864 }, { "epoch": 0.16, "learning_rate": 8.506660247151341e-05, "loss": 2.4573, "theoretical_loss": 3.443578494959251, "tokens_seen": 1910439936 }, { "epoch": 0.16, "objective/train/docs_used": 1051849, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3590188026428223, "objective/train/theoretical_loss": 3.4435685568814245, "objective/train/tokens_used": 280964576, "theoretical_loss": 3.4435685568814245, "tokens_seen": 1910505472 }, { "epoch": 0.16, "learning_rate": 8.505857807735516e-05, "loss": 2.6063, "theoretical_loss": 3.4435586192399477, "tokens_seen": 1910571008 }, { "epoch": 0.16, "learning_rate": 8.505055368319693e-05, "loss": 2.6925, "theoretical_loss": 3.443538745265905, "tokens_seen": 1910702080 }, { "epoch": 0.16, "learning_rate": 8.504252928903868e-05, "loss": 2.6257, "theoretical_loss": 3.443518873036852, "tokens_seen": 1910833152 }, { "epoch": 0.16, "learning_rate": 8.503450489488043e-05, "loss": 2.5778, "theoretical_loss": 3.443499002552514, "tokens_seen": 1910964224 }, { "epoch": 0.16, "learning_rate": 8.50264805007222e-05, "loss": 2.5841, "theoretical_loss": 3.4434791338126187, "tokens_seen": 1911095296 }, { "epoch": 0.16, "learning_rate": 8.501845610656395e-05, "loss": 2.5606, "theoretical_loss": 3.4434592668168933, "tokens_seen": 1911226368 }, { "epoch": 0.16, "learning_rate": 8.501043171240572e-05, "loss": 2.7309, "theoretical_loss": 3.4434394015650653, "tokens_seen": 1911357440 }, { "epoch": 0.16, "learning_rate": 8.500240731824747e-05, "loss": 2.6035, "theoretical_loss": 3.443419538056862, "tokens_seen": 1911488512 }, { "epoch": 0.16, "learning_rate": 8.499438292408923e-05, "loss": 2.5485, "theoretical_loss": 3.4433996762920107, "tokens_seen": 1911619584 }, { "epoch": 0.16, "learning_rate": 8.498635852993099e-05, "loss": 2.6219, "theoretical_loss": 3.443379816270239, "tokens_seen": 1911750656 }, { "epoch": 0.16, "learning_rate": 8.497833413577275e-05, "loss": 2.6035, "theoretical_loss": 3.443359957991275, "tokens_seen": 1911881728 }, { "epoch": 0.16, "learning_rate": 8.497030974161451e-05, "loss": 2.4346, "theoretical_loss": 3.4433401014548446, "tokens_seen": 1912012800 }, { "epoch": 0.16, "objective/train/docs_used": 1052490, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.299638032913208, "objective/train/theoretical_loss": 3.4433202466606767, "objective/train/tokens_used": 282602976, "theoretical_loss": 3.4433202466606767, "tokens_seen": 1912143872 }, { "epoch": 0.16, "learning_rate": 8.496228534745627e-05, "loss": 2.6254, "theoretical_loss": 3.4433202466606767, "tokens_seen": 1912143872 }, { "epoch": 0.16, "learning_rate": 8.495426095329802e-05, "loss": 2.4929, "theoretical_loss": 3.4433003936084994, "tokens_seen": 1912274944 }, { "epoch": 0.16, "learning_rate": 8.494623655913979e-05, "loss": 2.5459, "theoretical_loss": 3.44328054229804, "tokens_seen": 1912406016 }, { "epoch": 0.16, "learning_rate": 8.493821216498154e-05, "loss": 2.4689, "theoretical_loss": 3.4432606927290257, "tokens_seen": 1912537088 }, { "epoch": 0.16, "learning_rate": 8.49301877708233e-05, "loss": 2.6596, "theoretical_loss": 3.443240844901185, "tokens_seen": 1912668160 }, { "epoch": 0.16, "learning_rate": 8.492216337666506e-05, "loss": 2.5501, "theoretical_loss": 3.4432209988142457, "tokens_seen": 1912799232 }, { "epoch": 0.16, "learning_rate": 8.491413898250683e-05, "loss": 2.5588, "theoretical_loss": 3.4432011544679364, "tokens_seen": 1912930304 }, { "epoch": 0.16, "learning_rate": 8.490611458834858e-05, "loss": 2.631, "theoretical_loss": 3.4431813118619847, "tokens_seen": 1913061376 }, { "epoch": 0.16, "learning_rate": 8.489809019419033e-05, "loss": 2.6742, "theoretical_loss": 3.4431614709961185, "tokens_seen": 1913192448 }, { "epoch": 0.16, "learning_rate": 8.48900658000321e-05, "loss": 2.4847, "theoretical_loss": 3.4431416318700663, "tokens_seen": 1913323520 }, { "epoch": 0.16, "learning_rate": 8.488204140587385e-05, "loss": 2.4341, "theoretical_loss": 3.443121794483557, "tokens_seen": 1913454592 }, { "epoch": 0.16, "learning_rate": 8.487401701171562e-05, "loss": 2.7514, "theoretical_loss": 3.4431019588363174, "tokens_seen": 1913585664 }, { "epoch": 0.16, "learning_rate": 8.486599261755737e-05, "loss": 2.5306, "theoretical_loss": 3.443082124928077, "tokens_seen": 1913716736 }, { "epoch": 0.16, "objective/train/docs_used": 1053877, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7644686698913574, "objective/train/theoretical_loss": 3.443072208625997, "objective/train/tokens_used": 284241376, "theoretical_loss": 3.443072208625997, "tokens_seen": 1913782272 }, { "epoch": 0.16, "learning_rate": 8.485796822339914e-05, "loss": 2.6458, "theoretical_loss": 3.4430622927585643, "tokens_seen": 1913847808 }, { "epoch": 0.16, "learning_rate": 8.484994382924089e-05, "loss": 2.5744, "theoretical_loss": 3.4430424623275075, "tokens_seen": 1913978880 }, { "epoch": 0.16, "learning_rate": 8.484191943508264e-05, "loss": 2.5332, "theoretical_loss": 3.443022633634635, "tokens_seen": 1914109952 }, { "epoch": 0.16, "learning_rate": 8.483389504092441e-05, "loss": 2.6024, "theoretical_loss": 3.443002806679676, "tokens_seen": 1914241024 }, { "epoch": 0.16, "learning_rate": 8.482587064676616e-05, "loss": 2.6787, "theoretical_loss": 3.4429829814623583, "tokens_seen": 1914372096 }, { "epoch": 0.16, "learning_rate": 8.481784625260793e-05, "loss": 2.4463, "theoretical_loss": 3.442963157982412, "tokens_seen": 1914503168 }, { "epoch": 0.16, "learning_rate": 8.480982185844968e-05, "loss": 2.6267, "theoretical_loss": 3.4429433362395643, "tokens_seen": 1914634240 }, { "epoch": 0.16, "learning_rate": 8.480179746429145e-05, "loss": 2.3903, "theoretical_loss": 3.442923516233545, "tokens_seen": 1914765312 }, { "epoch": 0.16, "learning_rate": 8.47937730701332e-05, "loss": 2.4938, "theoretical_loss": 3.4429036979640832, "tokens_seen": 1914896384 }, { "epoch": 0.16, "learning_rate": 8.478574867597496e-05, "loss": 2.5209, "theoretical_loss": 3.442883881430908, "tokens_seen": 1915027456 }, { "epoch": 0.16, "learning_rate": 8.477772428181672e-05, "loss": 2.6642, "theoretical_loss": 3.442864066633747, "tokens_seen": 1915158528 }, { "epoch": 0.16, "learning_rate": 8.476969988765848e-05, "loss": 2.6275, "theoretical_loss": 3.442844253572331, "tokens_seen": 1915289600 }, { "epoch": 0.16, "objective/train/docs_used": 1054366, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.9697377681732178, "objective/train/theoretical_loss": 3.4428244422463887, "objective/train/tokens_used": 285879776, "theoretical_loss": 3.4428244422463887, "tokens_seen": 1915420672 }, { "epoch": 0.16, "learning_rate": 8.476167549350024e-05, "loss": 2.6293, "theoretical_loss": 3.4428244422463887, "tokens_seen": 1915420672 }, { "epoch": 0.16, "learning_rate": 8.4753651099342e-05, "loss": 2.4817, "theoretical_loss": 3.4428046326556494, "tokens_seen": 1915551744 }, { "epoch": 0.16, "learning_rate": 8.474562670518376e-05, "loss": 2.4483, "theoretical_loss": 3.4427848247998423, "tokens_seen": 1915682816 }, { "epoch": 0.16, "learning_rate": 8.473760231102552e-05, "loss": 2.6185, "theoretical_loss": 3.4427650186786964, "tokens_seen": 1915813888 }, { "epoch": 0.16, "learning_rate": 8.472957791686727e-05, "loss": 2.5448, "theoretical_loss": 3.4427452142919415, "tokens_seen": 1915944960 }, { "epoch": 0.16, "learning_rate": 8.472155352270904e-05, "loss": 2.5756, "theoretical_loss": 3.442725411639307, "tokens_seen": 1916076032 }, { "epoch": 0.16, "learning_rate": 8.471352912855079e-05, "loss": 2.6482, "theoretical_loss": 3.4427056107205223, "tokens_seen": 1916207104 }, { "epoch": 0.16, "learning_rate": 8.470550473439256e-05, "loss": 2.5003, "theoretical_loss": 3.442685811535317, "tokens_seen": 1916338176 }, { "epoch": 0.16, "learning_rate": 8.469748034023431e-05, "loss": 2.7789, "theoretical_loss": 3.4426660140834215, "tokens_seen": 1916469248 }, { "epoch": 0.16, "learning_rate": 8.468945594607608e-05, "loss": 2.5984, "theoretical_loss": 3.442646218364565, "tokens_seen": 1916600320 }, { "epoch": 0.16, "learning_rate": 8.468143155191783e-05, "loss": 2.5712, "theoretical_loss": 3.442626424378477, "tokens_seen": 1916731392 }, { "epoch": 0.16, "learning_rate": 8.467340715775958e-05, "loss": 2.5202, "theoretical_loss": 3.4426066321248876, "tokens_seen": 1916862464 }, { "epoch": 0.16, "learning_rate": 8.466538276360135e-05, "loss": 2.5165, "theoretical_loss": 3.442586841603527, "tokens_seen": 1916993536 }, { "epoch": 0.16, "objective/train/docs_used": 1055597, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.398259162902832, "objective/train/theoretical_loss": 3.4425769469923475, "objective/train/tokens_used": 287518176, "theoretical_loss": 3.4425769469923475, "tokens_seen": 1917059072 }, { "epoch": 0.16, "learning_rate": 8.46573583694431e-05, "loss": 2.6192, "theoretical_loss": 3.442567052814124, "tokens_seen": 1917124608 }, { "epoch": 0.16, "learning_rate": 8.464933397528487e-05, "loss": 2.4034, "theoretical_loss": 3.442547265756411, "tokens_seen": 1917255680 }, { "epoch": 0.16, "learning_rate": 8.464130958112662e-05, "loss": 2.4076, "theoretical_loss": 3.4425274804301154, "tokens_seen": 1917386752 }, { "epoch": 0.16, "learning_rate": 8.463328518696839e-05, "loss": 2.6285, "theoretical_loss": 3.442507696834969, "tokens_seen": 1917517824 }, { "epoch": 0.16, "learning_rate": 8.462526079281014e-05, "loss": 2.4717, "theoretical_loss": 3.4424879149707017, "tokens_seen": 1917648896 }, { "epoch": 0.16, "learning_rate": 8.461723639865191e-05, "loss": 2.649, "theoretical_loss": 3.442468134837043, "tokens_seen": 1917779968 }, { "epoch": 0.16, "learning_rate": 8.460921200449366e-05, "loss": 2.5644, "theoretical_loss": 3.4424483564337245, "tokens_seen": 1917911040 }, { "epoch": 0.16, "learning_rate": 8.460118761033541e-05, "loss": 2.5549, "theoretical_loss": 3.4424285797604757, "tokens_seen": 1918042112 }, { "epoch": 0.16, "learning_rate": 8.459316321617718e-05, "loss": 2.5357, "theoretical_loss": 3.4424088048170276, "tokens_seen": 1918173184 }, { "epoch": 0.16, "learning_rate": 8.458513882201893e-05, "loss": 2.4481, "theoretical_loss": 3.4423890316031107, "tokens_seen": 1918304256 }, { "epoch": 0.16, "learning_rate": 8.45771144278607e-05, "loss": 2.3806, "theoretical_loss": 3.442369260118455, "tokens_seen": 1918435328 }, { "epoch": 0.16, "learning_rate": 8.456909003370245e-05, "loss": 2.5601, "theoretical_loss": 3.4423494903627914, "tokens_seen": 1918566400 }, { "epoch": 0.16, "objective/train/docs_used": 1056906, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.411397695541382, "objective/train/theoretical_loss": 3.4423297223358507, "objective/train/tokens_used": 289156576, "theoretical_loss": 3.4423297223358507, "tokens_seen": 1918697472 }, { "epoch": 0.16, "learning_rate": 8.456106563954422e-05, "loss": 2.6341, "theoretical_loss": 3.4423297223358507, "tokens_seen": 1918697472 }, { "epoch": 0.16, "learning_rate": 8.455304124538597e-05, "loss": 2.4965, "theoretical_loss": 3.4423099560373633, "tokens_seen": 1918828544 }, { "epoch": 0.16, "learning_rate": 8.454501685122773e-05, "loss": 2.4888, "theoretical_loss": 3.4422901914670607, "tokens_seen": 1918959616 }, { "epoch": 0.16, "learning_rate": 8.45369924570695e-05, "loss": 2.5232, "theoretical_loss": 3.4422704286246733, "tokens_seen": 1919090688 }, { "epoch": 0.16, "learning_rate": 8.452896806291125e-05, "loss": 2.6028, "theoretical_loss": 3.4422506675099322, "tokens_seen": 1919221760 }, { "epoch": 0.16, "learning_rate": 8.452094366875301e-05, "loss": 2.4306, "theoretical_loss": 3.4422309081225686, "tokens_seen": 1919352832 }, { "epoch": 0.16, "learning_rate": 8.451291927459477e-05, "loss": 2.6496, "theoretical_loss": 3.442211150462313, "tokens_seen": 1919483904 }, { "epoch": 0.16, "learning_rate": 8.450489488043653e-05, "loss": 2.539, "theoretical_loss": 3.442191394528897, "tokens_seen": 1919614976 }, { "epoch": 0.16, "learning_rate": 8.449687048627829e-05, "loss": 2.485, "theoretical_loss": 3.4421716403220515, "tokens_seen": 1919746048 }, { "epoch": 0.16, "learning_rate": 8.448884609212004e-05, "loss": 2.6431, "theoretical_loss": 3.442151887841508, "tokens_seen": 1919877120 }, { "epoch": 0.16, "learning_rate": 8.44808216979618e-05, "loss": 2.7387, "theoretical_loss": 3.442132137086997, "tokens_seen": 1920008192 }, { "epoch": 0.16, "learning_rate": 8.447279730380356e-05, "loss": 2.5542, "theoretical_loss": 3.442112388058251, "tokens_seen": 1920139264 }, { "epoch": 0.16, "learning_rate": 8.446477290964533e-05, "loss": 2.557, "theoretical_loss": 3.4420926407550008, "tokens_seen": 1920270336 }, { "epoch": 0.16, "objective/train/docs_used": 1057603, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.503713607788086, "objective/train/theoretical_loss": 3.442082767750353, "objective/train/tokens_used": 290794976, "theoretical_loss": 3.442082767750353, "tokens_seen": 1920335872 }, { "epoch": 0.16, "learning_rate": 8.445674851548708e-05, "loss": 2.6189, "theoretical_loss": 3.4420728951769783, "tokens_seen": 1920401408 }, { "epoch": 0.16, "learning_rate": 8.444872412132885e-05, "loss": 2.5225, "theoretical_loss": 3.4420531513239148, "tokens_seen": 1920532480 }, { "epoch": 0.16, "learning_rate": 8.44406997271706e-05, "loss": 2.5151, "theoretical_loss": 3.4420334091955413, "tokens_seen": 1920663552 }, { "epoch": 0.16, "learning_rate": 8.443267533301235e-05, "loss": 2.6353, "theoretical_loss": 3.4420136687915903, "tokens_seen": 1920794624 }, { "epoch": 0.16, "learning_rate": 8.442465093885412e-05, "loss": 2.4835, "theoretical_loss": 3.441993930111793, "tokens_seen": 1920925696 }, { "epoch": 0.16, "learning_rate": 8.441662654469587e-05, "loss": 2.5794, "theoretical_loss": 3.4419741931558816, "tokens_seen": 1921056768 }, { "epoch": 0.16, "learning_rate": 8.440860215053764e-05, "loss": 2.6229, "theoretical_loss": 3.441954457923588, "tokens_seen": 1921187840 }, { "epoch": 0.16, "learning_rate": 8.440057775637939e-05, "loss": 2.4523, "theoretical_loss": 3.4419347244146437, "tokens_seen": 1921318912 }, { "epoch": 0.16, "learning_rate": 8.439255336222116e-05, "loss": 2.6327, "theoretical_loss": 3.4419149926287806, "tokens_seen": 1921449984 }, { "epoch": 0.16, "learning_rate": 8.438452896806291e-05, "loss": 2.6949, "theoretical_loss": 3.4418952625657306, "tokens_seen": 1921581056 }, { "epoch": 0.16, "learning_rate": 8.437650457390466e-05, "loss": 2.5928, "theoretical_loss": 3.4418755342252267, "tokens_seen": 1921712128 }, { "epoch": 0.16, "learning_rate": 8.436848017974643e-05, "loss": 2.5388, "theoretical_loss": 3.4418558076070003, "tokens_seen": 1921843200 }, { "epoch": 0.16, "objective/train/docs_used": 1058944, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.295034885406494, "objective/train/theoretical_loss": 3.4418360827107835, "objective/train/tokens_used": 292433376, "theoretical_loss": 3.4418360827107835, "tokens_seen": 1921974272 }, { "epoch": 0.16, "learning_rate": 8.436045578558818e-05, "loss": 2.5935, "theoretical_loss": 3.4418360827107835, "tokens_seen": 1921974272 }, { "epoch": 0.16, "learning_rate": 8.435243139142995e-05, "loss": 2.6164, "theoretical_loss": 3.4418163595363094, "tokens_seen": 1922105344 }, { "epoch": 0.16, "learning_rate": 8.43444069972717e-05, "loss": 2.6415, "theoretical_loss": 3.441796638083309, "tokens_seen": 1922236416 }, { "epoch": 0.17, "learning_rate": 8.433638260311347e-05, "loss": 2.6415, "theoretical_loss": 3.4417769183515157, "tokens_seen": 1922367488 }, { "epoch": 0.17, "learning_rate": 8.432835820895522e-05, "loss": 2.5213, "theoretical_loss": 3.441757200340662, "tokens_seen": 1922498560 }, { "epoch": 0.17, "learning_rate": 8.432033381479699e-05, "loss": 2.7047, "theoretical_loss": 3.4417374840504795, "tokens_seen": 1922629632 }, { "epoch": 0.17, "learning_rate": 8.431230942063874e-05, "loss": 2.324, "theoretical_loss": 3.441717769480701, "tokens_seen": 1922760704 }, { "epoch": 0.17, "learning_rate": 8.43042850264805e-05, "loss": 2.7517, "theoretical_loss": 3.4416980566310604, "tokens_seen": 1922891776 }, { "epoch": 0.17, "learning_rate": 8.429626063232226e-05, "loss": 2.5888, "theoretical_loss": 3.4416783455012885, "tokens_seen": 1923022848 }, { "epoch": 0.17, "learning_rate": 8.428823623816402e-05, "loss": 2.452, "theoretical_loss": 3.4416586360911197, "tokens_seen": 1923153920 }, { "epoch": 0.17, "learning_rate": 8.428021184400578e-05, "loss": 2.6019, "theoretical_loss": 3.4416389284002853, "tokens_seen": 1923284992 }, { "epoch": 0.17, "learning_rate": 8.427218744984754e-05, "loss": 2.622, "theoretical_loss": 3.4416192224285194, "tokens_seen": 1923416064 }, { "epoch": 0.17, "learning_rate": 8.42641630556893e-05, "loss": 2.642, "theoretical_loss": 3.441599518175554, "tokens_seen": 1923547136 }, { "epoch": 0.17, "objective/train/docs_used": 1059301, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.181396245956421, "objective/train/theoretical_loss": 3.441589666693538, "objective/train/tokens_used": 294071776, "theoretical_loss": 3.441589666693538, "tokens_seen": 1923612672 }, { "epoch": 0.17, "learning_rate": 8.425613866153106e-05, "loss": 2.6953, "theoretical_loss": 3.4415798156411226, "tokens_seen": 1923678208 }, { "epoch": 0.17, "learning_rate": 8.424811426737281e-05, "loss": 2.5897, "theoretical_loss": 3.4415601148249584, "tokens_seen": 1923809280 }, { "epoch": 0.17, "learning_rate": 8.424008987321458e-05, "loss": 2.5453, "theoretical_loss": 3.441540415726794, "tokens_seen": 1923940352 }, { "epoch": 0.17, "learning_rate": 8.423206547905633e-05, "loss": 2.4628, "theoretical_loss": 3.4415207183463625, "tokens_seen": 1924071424 }, { "epoch": 0.17, "learning_rate": 8.42240410848981e-05, "loss": 2.5583, "theoretical_loss": 3.441501022683398, "tokens_seen": 1924202496 }, { "epoch": 0.17, "learning_rate": 8.421601669073985e-05, "loss": 2.5335, "theoretical_loss": 3.4414813287376327, "tokens_seen": 1924333568 }, { "epoch": 0.17, "learning_rate": 8.420799229658162e-05, "loss": 2.5424, "theoretical_loss": 3.4414616365088007, "tokens_seen": 1924464640 }, { "epoch": 0.17, "learning_rate": 8.419996790242337e-05, "loss": 2.5134, "theoretical_loss": 3.441441945996635, "tokens_seen": 1924595712 }, { "epoch": 0.17, "learning_rate": 8.419194350826512e-05, "loss": 2.5916, "theoretical_loss": 3.441422257200869, "tokens_seen": 1924726784 }, { "epoch": 0.17, "learning_rate": 8.418391911410689e-05, "loss": 2.4067, "theoretical_loss": 3.4414025701212365, "tokens_seen": 1924857856 }, { "epoch": 0.17, "learning_rate": 8.417589471994864e-05, "loss": 2.5449, "theoretical_loss": 3.441382884757471, "tokens_seen": 1924988928 }, { "epoch": 0.17, "learning_rate": 8.416787032579041e-05, "loss": 2.6262, "theoretical_loss": 3.441363201109305, "tokens_seen": 1925120000 }, { "epoch": 0.17, "objective/train/docs_used": 1060240, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.727017641067505, "objective/train/theoretical_loss": 3.4413435191764745, "objective/train/tokens_used": 295710176, "theoretical_loss": 3.4413435191764745, "tokens_seen": 1925251072 }, { "epoch": 0.17, "learning_rate": 8.415984593163216e-05, "loss": 2.5072, "theoretical_loss": 3.4413435191764745, "tokens_seen": 1925251072 }, { "epoch": 0.17, "learning_rate": 8.415182153747393e-05, "loss": 2.5916, "theoretical_loss": 3.441323838958712, "tokens_seen": 1925382144 }, { "epoch": 0.17, "learning_rate": 8.414379714331568e-05, "loss": 2.6457, "theoretical_loss": 3.4413041604557506, "tokens_seen": 1925513216 }, { "epoch": 0.17, "learning_rate": 8.413577274915743e-05, "loss": 2.5979, "theoretical_loss": 3.4412844836673253, "tokens_seen": 1925644288 }, { "epoch": 0.17, "learning_rate": 8.41277483549992e-05, "loss": 2.5639, "theoretical_loss": 3.4412648085931696, "tokens_seen": 1925775360 }, { "epoch": 0.17, "learning_rate": 8.411972396084095e-05, "loss": 2.5317, "theoretical_loss": 3.4412451352330176, "tokens_seen": 1925906432 }, { "epoch": 0.17, "learning_rate": 8.411169956668272e-05, "loss": 2.553, "theoretical_loss": 3.441225463586603, "tokens_seen": 1926037504 }, { "epoch": 0.17, "learning_rate": 8.410367517252447e-05, "loss": 2.5041, "theoretical_loss": 3.44120579365366, "tokens_seen": 1926168576 }, { "epoch": 0.17, "learning_rate": 8.409565077836624e-05, "loss": 2.4359, "theoretical_loss": 3.4411861254339238, "tokens_seen": 1926299648 }, { "epoch": 0.17, "learning_rate": 8.4087626384208e-05, "loss": 2.5611, "theoretical_loss": 3.441166458927127, "tokens_seen": 1926430720 }, { "epoch": 0.17, "learning_rate": 8.407960199004975e-05, "loss": 2.4728, "theoretical_loss": 3.441146794133005, "tokens_seen": 1926561792 }, { "epoch": 0.17, "learning_rate": 8.407157759589151e-05, "loss": 2.4456, "theoretical_loss": 3.4411271310512914, "tokens_seen": 1926692864 }, { "epoch": 0.17, "learning_rate": 8.406355320173327e-05, "loss": 2.5713, "theoretical_loss": 3.441107469681721, "tokens_seen": 1926823936 }, { "epoch": 0.17, "objective/train/docs_used": 1060815, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.9482412338256836, "objective/train/theoretical_loss": 3.441097639638907, "objective/train/tokens_used": 297348576, "theoretical_loss": 3.441097639638907, "tokens_seen": 1926889472 }, { "epoch": 0.17, "learning_rate": 8.405552880757503e-05, "loss": 2.6823, "theoretical_loss": 3.4410878100240287, "tokens_seen": 1926955008 }, { "epoch": 0.17, "learning_rate": 8.404750441341679e-05, "loss": 2.6304, "theoretical_loss": 3.441068152077948, "tokens_seen": 1927086080 }, { "epoch": 0.17, "learning_rate": 8.403948001925855e-05, "loss": 2.7154, "theoretical_loss": 3.441048495843215, "tokens_seen": 1927217152 }, { "epoch": 0.17, "learning_rate": 8.403145562510031e-05, "loss": 2.6143, "theoretical_loss": 3.4410288413195627, "tokens_seen": 1927348224 }, { "epoch": 0.17, "learning_rate": 8.402343123094206e-05, "loss": 2.5897, "theoretical_loss": 3.4410091885067264, "tokens_seen": 1927479296 }, { "epoch": 0.17, "learning_rate": 8.401540683678383e-05, "loss": 2.7404, "theoretical_loss": 3.440989537404441, "tokens_seen": 1927610368 }, { "epoch": 0.17, "learning_rate": 8.400738244262558e-05, "loss": 2.4568, "theoretical_loss": 3.4409698880124413, "tokens_seen": 1927741440 }, { "epoch": 0.17, "learning_rate": 8.399935804846735e-05, "loss": 2.6438, "theoretical_loss": 3.4409502403304626, "tokens_seen": 1927872512 }, { "epoch": 0.17, "learning_rate": 8.39913336543091e-05, "loss": 2.471, "theoretical_loss": 3.440930594358239, "tokens_seen": 1928003584 }, { "epoch": 0.17, "learning_rate": 8.398330926015087e-05, "loss": 2.4765, "theoretical_loss": 3.440910950095506, "tokens_seen": 1928134656 }, { "epoch": 0.17, "learning_rate": 8.397528486599262e-05, "loss": 2.6399, "theoretical_loss": 3.4408913075419987, "tokens_seen": 1928265728 }, { "epoch": 0.17, "learning_rate": 8.396726047183439e-05, "loss": 2.6799, "theoretical_loss": 3.440871666697452, "tokens_seen": 1928396800 }, { "epoch": 0.17, "objective/train/docs_used": 1062086, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.882824659347534, "objective/train/theoretical_loss": 3.4408520275616006, "objective/train/tokens_used": 298986976, "theoretical_loss": 3.4408520275616006, "tokens_seen": 1928527872 }, { "epoch": 0.17, "learning_rate": 8.395923607767614e-05, "loss": 2.5886, "theoretical_loss": 3.4408520275616006, "tokens_seen": 1928527872 }, { "epoch": 0.17, "learning_rate": 8.395121168351789e-05, "loss": 2.5707, "theoretical_loss": 3.4408323901341813, "tokens_seen": 1928658944 }, { "epoch": 0.17, "learning_rate": 8.394318728935966e-05, "loss": 2.5705, "theoretical_loss": 3.440812754414927, "tokens_seen": 1928790016 }, { "epoch": 0.17, "learning_rate": 8.393516289520141e-05, "loss": 2.5066, "theoretical_loss": 3.4407931204035753, "tokens_seen": 1928921088 }, { "epoch": 0.17, "learning_rate": 8.392713850104318e-05, "loss": 2.6055, "theoretical_loss": 3.440773488099861, "tokens_seen": 1929052160 }, { "epoch": 0.17, "learning_rate": 8.391911410688493e-05, "loss": 2.5717, "theoretical_loss": 3.440753857503519, "tokens_seen": 1929183232 }, { "epoch": 0.17, "learning_rate": 8.39110897127267e-05, "loss": 2.5541, "theoretical_loss": 3.440734228614285, "tokens_seen": 1929314304 }, { "epoch": 0.17, "learning_rate": 8.390306531856845e-05, "loss": 2.587, "theoretical_loss": 3.4407146014318943, "tokens_seen": 1929445376 }, { "epoch": 0.17, "learning_rate": 8.38950409244102e-05, "loss": 2.4675, "theoretical_loss": 3.4406949759560836, "tokens_seen": 1929576448 }, { "epoch": 0.17, "learning_rate": 8.388701653025197e-05, "loss": 2.6033, "theoretical_loss": 3.440675352186587, "tokens_seen": 1929707520 }, { "epoch": 0.17, "learning_rate": 8.387899213609372e-05, "loss": 2.6421, "theoretical_loss": 3.4406557301231424, "tokens_seen": 1929838592 }, { "epoch": 0.17, "learning_rate": 8.387096774193549e-05, "loss": 2.5802, "theoretical_loss": 3.440636109765484, "tokens_seen": 1929969664 }, { "epoch": 0.17, "learning_rate": 8.386294334777724e-05, "loss": 2.5423, "theoretical_loss": 3.440616491113348, "tokens_seen": 1930100736 }, { "epoch": 0.17, "objective/train/docs_used": 1062749, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.9142069816589355, "objective/train/theoretical_loss": 3.4406066824267683, "objective/train/tokens_used": 300625376, "theoretical_loss": 3.4406066824267683, "tokens_seen": 1930166272 }, { "epoch": 0.17, "learning_rate": 8.385491895361901e-05, "loss": 2.5888, "theoretical_loss": 3.44059687416647, "tokens_seen": 1930231808 }, { "epoch": 0.17, "learning_rate": 8.384689455946076e-05, "loss": 2.6012, "theoretical_loss": 3.4405772589245873, "tokens_seen": 1930362880 }, { "epoch": 0.17, "learning_rate": 8.383887016530252e-05, "loss": 2.6279, "theoretical_loss": 3.440557645387435, "tokens_seen": 1930493952 }, { "epoch": 0.17, "learning_rate": 8.383084577114428e-05, "loss": 2.4806, "theoretical_loss": 3.4405380335547493, "tokens_seen": 1930625024 }, { "epoch": 0.17, "learning_rate": 8.382282137698604e-05, "loss": 2.426, "theoretical_loss": 3.4405184234262656, "tokens_seen": 1930756096 }, { "epoch": 0.17, "learning_rate": 8.38147969828278e-05, "loss": 2.6313, "theoretical_loss": 3.440498815001722, "tokens_seen": 1930887168 }, { "epoch": 0.17, "learning_rate": 8.380677258866956e-05, "loss": 2.521, "theoretical_loss": 3.440479208280853, "tokens_seen": 1931018240 }, { "epoch": 0.17, "learning_rate": 8.379874819451132e-05, "loss": 2.6727, "theoretical_loss": 3.4404596032633963, "tokens_seen": 1931149312 }, { "epoch": 0.17, "learning_rate": 8.379072380035308e-05, "loss": 2.4125, "theoretical_loss": 3.440439999949087, "tokens_seen": 1931280384 }, { "epoch": 0.17, "learning_rate": 8.378269940619483e-05, "loss": 2.5713, "theoretical_loss": 3.440420398337663, "tokens_seen": 1931411456 }, { "epoch": 0.17, "learning_rate": 8.37746750120366e-05, "loss": 2.5418, "theoretical_loss": 3.4404007984288594, "tokens_seen": 1931542528 }, { "epoch": 0.17, "learning_rate": 8.376665061787835e-05, "loss": 2.4928, "theoretical_loss": 3.4403812002224137, "tokens_seen": 1931673600 }, { "epoch": 0.17, "objective/train/docs_used": 1064305, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5660760402679443, "objective/train/theoretical_loss": 3.4403616037180624, "objective/train/tokens_used": 302263776, "theoretical_loss": 3.4403616037180624, "tokens_seen": 1931804672 }, { "epoch": 0.17, "learning_rate": 8.375862622372012e-05, "loss": 2.5156, "theoretical_loss": 3.4403616037180624, "tokens_seen": 1931804672 }, { "epoch": 0.17, "learning_rate": 8.375060182956187e-05, "loss": 2.6033, "theoretical_loss": 3.440342008915542, "tokens_seen": 1931935744 }, { "epoch": 0.17, "learning_rate": 8.374257743540364e-05, "loss": 2.5074, "theoretical_loss": 3.440322415814589, "tokens_seen": 1932066816 }, { "epoch": 0.17, "learning_rate": 8.373455304124539e-05, "loss": 2.4945, "theoretical_loss": 3.4403028244149416, "tokens_seen": 1932197888 }, { "epoch": 0.17, "learning_rate": 8.372652864708714e-05, "loss": 2.5083, "theoretical_loss": 3.440283234716335, "tokens_seen": 1932328960 }, { "epoch": 0.17, "learning_rate": 8.371850425292891e-05, "loss": 2.6815, "theoretical_loss": 3.440263646718507, "tokens_seen": 1932460032 }, { "epoch": 0.17, "learning_rate": 8.371047985877066e-05, "loss": 2.4038, "theoretical_loss": 3.440244060421194, "tokens_seen": 1932591104 }, { "epoch": 0.17, "learning_rate": 8.370245546461243e-05, "loss": 2.7489, "theoretical_loss": 3.4402244758241336, "tokens_seen": 1932722176 }, { "epoch": 0.17, "learning_rate": 8.369443107045418e-05, "loss": 2.545, "theoretical_loss": 3.4402048929270634, "tokens_seen": 1932853248 }, { "epoch": 0.17, "learning_rate": 8.368640667629595e-05, "loss": 2.616, "theoretical_loss": 3.4401853117297194, "tokens_seen": 1932984320 }, { "epoch": 0.17, "learning_rate": 8.36783822821377e-05, "loss": 2.4904, "theoretical_loss": 3.440165732231839, "tokens_seen": 1933115392 }, { "epoch": 0.17, "learning_rate": 8.367035788797947e-05, "loss": 2.5622, "theoretical_loss": 3.440146154433161, "tokens_seen": 1933246464 }, { "epoch": 0.17, "learning_rate": 8.366233349382122e-05, "loss": 2.7073, "theoretical_loss": 3.4401265783334205, "tokens_seen": 1933377536 }, { "epoch": 0.17, "objective/train/docs_used": 1064860, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6494178771972656, "objective/train/theoretical_loss": 3.4401167909205705, "objective/train/tokens_used": 303902176, "theoretical_loss": 3.4401167909205705, "tokens_seen": 1933443072 }, { "epoch": 0.17, "learning_rate": 8.365430909966297e-05, "loss": 2.6403, "theoretical_loss": 3.440107003932357, "tokens_seen": 1933508608 }, { "epoch": 0.17, "learning_rate": 8.364628470550474e-05, "loss": 2.6015, "theoretical_loss": 3.4400874312297063, "tokens_seen": 1933639680 }, { "epoch": 0.17, "learning_rate": 8.36382603113465e-05, "loss": 2.5343, "theoretical_loss": 3.4400678602252066, "tokens_seen": 1933770752 }, { "epoch": 0.17, "learning_rate": 8.363023591718826e-05, "loss": 2.6448, "theoretical_loss": 3.440048290918596, "tokens_seen": 1933901824 }, { "epoch": 0.17, "learning_rate": 8.362221152303001e-05, "loss": 2.5473, "theoretical_loss": 3.4400287233096107, "tokens_seen": 1934032896 }, { "epoch": 0.17, "learning_rate": 8.361418712887178e-05, "loss": 2.6655, "theoretical_loss": 3.4400091573979896, "tokens_seen": 1934163968 }, { "epoch": 0.17, "learning_rate": 8.360616273471353e-05, "loss": 2.5688, "theoretical_loss": 3.439989593183471, "tokens_seen": 1934295040 }, { "epoch": 0.17, "learning_rate": 8.359813834055529e-05, "loss": 2.6169, "theoretical_loss": 3.439970030665791, "tokens_seen": 1934426112 }, { "epoch": 0.17, "learning_rate": 8.359011394639705e-05, "loss": 2.6115, "theoretical_loss": 3.4399504698446886, "tokens_seen": 1934557184 }, { "epoch": 0.17, "learning_rate": 8.358208955223881e-05, "loss": 2.4359, "theoretical_loss": 3.4399309107199016, "tokens_seen": 1934688256 }, { "epoch": 0.17, "learning_rate": 8.357406515808057e-05, "loss": 2.5228, "theoretical_loss": 3.439911353291168, "tokens_seen": 1934819328 }, { "epoch": 0.17, "learning_rate": 8.356604076392233e-05, "loss": 2.558, "theoretical_loss": 3.439891797558225, "tokens_seen": 1934950400 }, { "epoch": 0.17, "objective/train/docs_used": 1065725, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4864485263824463, "objective/train/theoretical_loss": 3.439872243520812, "objective/train/tokens_used": 305540576, "theoretical_loss": 3.439872243520812, "tokens_seen": 1935081472 }, { "epoch": 0.17, "learning_rate": 8.35580163697641e-05, "loss": 2.5461, "theoretical_loss": 3.439872243520812, "tokens_seen": 1935081472 }, { "epoch": 0.17, "learning_rate": 8.354999197560585e-05, "loss": 2.4294, "theoretical_loss": 3.4398526911786664, "tokens_seen": 1935212544 }, { "epoch": 0.17, "learning_rate": 8.35419675814476e-05, "loss": 2.48, "theoretical_loss": 3.439833140531527, "tokens_seen": 1935343616 }, { "epoch": 0.17, "learning_rate": 8.353394318728937e-05, "loss": 2.5107, "theoretical_loss": 3.4398135915791315, "tokens_seen": 1935474688 }, { "epoch": 0.17, "learning_rate": 8.352591879313112e-05, "loss": 2.4663, "theoretical_loss": 3.439794044321218, "tokens_seen": 1935605760 }, { "epoch": 0.17, "learning_rate": 8.351789439897289e-05, "loss": 2.6062, "theoretical_loss": 3.439774498757526, "tokens_seen": 1935736832 }, { "epoch": 0.17, "learning_rate": 8.350987000481464e-05, "loss": 2.6143, "theoretical_loss": 3.4397549548877926, "tokens_seen": 1935867904 }, { "epoch": 0.17, "learning_rate": 8.35018456106564e-05, "loss": 2.699, "theoretical_loss": 3.439735412711757, "tokens_seen": 1935998976 }, { "epoch": 0.17, "learning_rate": 8.349382121649816e-05, "loss": 2.5616, "theoretical_loss": 3.4397158722291583, "tokens_seen": 1936130048 }, { "epoch": 0.17, "learning_rate": 8.348579682233991e-05, "loss": 2.5023, "theoretical_loss": 3.439696333439734, "tokens_seen": 1936261120 }, { "epoch": 0.17, "learning_rate": 8.347777242818168e-05, "loss": 2.5506, "theoretical_loss": 3.4396767963432238, "tokens_seen": 1936392192 }, { "epoch": 0.17, "learning_rate": 8.346974803402343e-05, "loss": 2.5643, "theoretical_loss": 3.4396572609393656, "tokens_seen": 1936523264 }, { "epoch": 0.17, "learning_rate": 8.34617236398652e-05, "loss": 2.4963, "theoretical_loss": 3.439637727227899, "tokens_seen": 1936654336 }, { "epoch": 0.17, "objective/train/docs_used": 1066923, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5231666564941406, "objective/train/theoretical_loss": 3.439627961006731, "objective/train/tokens_used": 307178976, "theoretical_loss": 3.439627961006731, "tokens_seen": 1936719872 }, { "epoch": 0.17, "learning_rate": 8.345369924570695e-05, "loss": 2.5335, "theoretical_loss": 3.4396181952085625, "tokens_seen": 1936785408 }, { "epoch": 0.17, "learning_rate": 8.344567485154872e-05, "loss": 2.4004, "theoretical_loss": 3.4395986648810943, "tokens_seen": 1936916480 }, { "epoch": 0.17, "learning_rate": 8.343765045739047e-05, "loss": 2.6551, "theoretical_loss": 3.439579136245235, "tokens_seen": 1937047552 }, { "epoch": 0.17, "learning_rate": 8.342962606323223e-05, "loss": 2.5923, "theoretical_loss": 3.439559609300722, "tokens_seen": 1937178624 }, { "epoch": 0.17, "learning_rate": 8.342160166907399e-05, "loss": 2.5727, "theoretical_loss": 3.4395400840472954, "tokens_seen": 1937309696 }, { "epoch": 0.17, "learning_rate": 8.341357727491575e-05, "loss": 2.6667, "theoretical_loss": 3.439520560484694, "tokens_seen": 1937440768 }, { "epoch": 0.17, "learning_rate": 8.340555288075751e-05, "loss": 2.6872, "theoretical_loss": 3.439501038612657, "tokens_seen": 1937571840 }, { "epoch": 0.17, "learning_rate": 8.339752848659926e-05, "loss": 2.6223, "theoretical_loss": 3.4394815184309238, "tokens_seen": 1937702912 }, { "epoch": 0.17, "learning_rate": 8.338950409244103e-05, "loss": 2.5338, "theoretical_loss": 3.4394619999392333, "tokens_seen": 1937833984 }, { "epoch": 0.17, "learning_rate": 8.338147969828278e-05, "loss": 2.6202, "theoretical_loss": 3.4394424831373254, "tokens_seen": 1937965056 }, { "epoch": 0.17, "learning_rate": 8.337345530412455e-05, "loss": 2.556, "theoretical_loss": 3.439422968024939, "tokens_seen": 1938096128 }, { "epoch": 0.17, "learning_rate": 8.33654309099663e-05, "loss": 2.4879, "theoretical_loss": 3.4394034546018144, "tokens_seen": 1938227200 }, { "epoch": 0.17, "objective/train/docs_used": 1067414, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6843504905700684, "objective/train/theoretical_loss": 3.439383942867691, "objective/train/tokens_used": 308817376, "theoretical_loss": 3.439383942867691, "tokens_seen": 1938358272 }, { "epoch": 0.17, "learning_rate": 8.335740651580806e-05, "loss": 2.6718, "theoretical_loss": 3.439383942867691, "tokens_seen": 1938358272 }, { "epoch": 0.17, "learning_rate": 8.334938212164982e-05, "loss": 2.4463, "theoretical_loss": 3.439364432822307, "tokens_seen": 1938489344 }, { "epoch": 0.17, "learning_rate": 8.334135772749158e-05, "loss": 2.5893, "theoretical_loss": 3.439344924465404, "tokens_seen": 1938620416 }, { "epoch": 0.18, "learning_rate": 8.333333333333334e-05, "loss": 2.442, "theoretical_loss": 3.439325417796721, "tokens_seen": 1938751488 }, { "epoch": 0.18, "learning_rate": 8.33253089391751e-05, "loss": 2.5347, "theoretical_loss": 3.439305912815997, "tokens_seen": 1938882560 }, { "epoch": 0.18, "learning_rate": 8.331728454501686e-05, "loss": 2.6075, "theoretical_loss": 3.439286409522973, "tokens_seen": 1939013632 }, { "epoch": 0.18, "learning_rate": 8.330926015085862e-05, "loss": 2.453, "theoretical_loss": 3.4392669079173883, "tokens_seen": 1939144704 }, { "epoch": 0.18, "learning_rate": 8.330123575670037e-05, "loss": 2.5933, "theoretical_loss": 3.439247407998983, "tokens_seen": 1939275776 }, { "epoch": 0.18, "learning_rate": 8.329321136254214e-05, "loss": 2.4927, "theoretical_loss": 3.439227909767497, "tokens_seen": 1939406848 }, { "epoch": 0.18, "learning_rate": 8.328518696838389e-05, "loss": 2.5154, "theoretical_loss": 3.439208413222671, "tokens_seen": 1939537920 }, { "epoch": 0.18, "learning_rate": 8.327716257422566e-05, "loss": 2.6075, "theoretical_loss": 3.439188918364244, "tokens_seen": 1939668992 }, { "epoch": 0.18, "learning_rate": 8.326913818006741e-05, "loss": 2.5968, "theoretical_loss": 3.439169425191957, "tokens_seen": 1939800064 }, { "epoch": 0.18, "learning_rate": 8.326111378590918e-05, "loss": 2.4055, "theoretical_loss": 3.4391499337055498, "tokens_seen": 1939931136 }, { "epoch": 0.18, "objective/train/docs_used": 1068508, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.342050313949585, "objective/train/theoretical_loss": 3.4391401885944703, "objective/train/tokens_used": 310455776, "theoretical_loss": 3.4391401885944703, "tokens_seen": 1939996672 }, { "epoch": 0.18, "learning_rate": 8.325308939175093e-05, "loss": 2.4388, "theoretical_loss": 3.439130443904763, "tokens_seen": 1940062208 }, { "epoch": 0.18, "learning_rate": 8.324506499759268e-05, "loss": 2.5146, "theoretical_loss": 3.439110955789338, "tokens_seen": 1940193280 }, { "epoch": 0.18, "learning_rate": 8.323704060343445e-05, "loss": 2.6226, "theoretical_loss": 3.4390914693590124, "tokens_seen": 1940324352 }, { "epoch": 0.18, "learning_rate": 8.32290162092762e-05, "loss": 2.5135, "theoretical_loss": 3.4390719846135296, "tokens_seen": 1940455424 }, { "epoch": 0.18, "learning_rate": 8.322099181511797e-05, "loss": 2.5178, "theoretical_loss": 3.439052501552628, "tokens_seen": 1940586496 }, { "epoch": 0.18, "learning_rate": 8.321296742095972e-05, "loss": 2.5175, "theoretical_loss": 3.4390330201760495, "tokens_seen": 1940717568 }, { "epoch": 0.18, "learning_rate": 8.320494302680149e-05, "loss": 2.4956, "theoretical_loss": 3.439013540483535, "tokens_seen": 1940848640 }, { "epoch": 0.18, "learning_rate": 8.319691863264324e-05, "loss": 2.4474, "theoretical_loss": 3.4389940624748236, "tokens_seen": 1940979712 }, { "epoch": 0.18, "learning_rate": 8.3188894238485e-05, "loss": 2.4981, "theoretical_loss": 3.4389745861496577, "tokens_seen": 1941110784 }, { "epoch": 0.18, "learning_rate": 8.318086984432676e-05, "loss": 2.4449, "theoretical_loss": 3.438955111507777, "tokens_seen": 1941241856 }, { "epoch": 0.18, "learning_rate": 8.317284545016852e-05, "loss": 2.5271, "theoretical_loss": 3.4389356385489234, "tokens_seen": 1941372928 }, { "epoch": 0.18, "learning_rate": 8.316482105601028e-05, "loss": 2.5505, "theoretical_loss": 3.4389161672728368, "tokens_seen": 1941504000 }, { "epoch": 0.18, "objective/train/docs_used": 1069184, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.391522169113159, "objective/train/theoretical_loss": 3.438896697679259, "objective/train/tokens_used": 312094176, "theoretical_loss": 3.438896697679259, "tokens_seen": 1941635072 }, { "epoch": 0.18, "learning_rate": 8.315679666185203e-05, "loss": 2.5596, "theoretical_loss": 3.438896697679259, "tokens_seen": 1941635072 }, { "epoch": 0.18, "learning_rate": 8.31487722676938e-05, "loss": 2.4975, "theoretical_loss": 3.4388772297679306, "tokens_seen": 1941766144 }, { "epoch": 0.18, "learning_rate": 8.314074787353555e-05, "loss": 2.5002, "theoretical_loss": 3.438857763538593, "tokens_seen": 1941897216 }, { "epoch": 0.18, "learning_rate": 8.313272347937731e-05, "loss": 2.5132, "theoretical_loss": 3.4388382989909863, "tokens_seen": 1942028288 }, { "epoch": 0.18, "learning_rate": 8.312469908521907e-05, "loss": 2.5543, "theoretical_loss": 3.4388188361248533, "tokens_seen": 1942159360 }, { "epoch": 0.18, "learning_rate": 8.311667469106083e-05, "loss": 2.381, "theoretical_loss": 3.4387993749399346, "tokens_seen": 1942290432 }, { "epoch": 0.18, "learning_rate": 8.31086502969026e-05, "loss": 2.3192, "theoretical_loss": 3.4387799154359717, "tokens_seen": 1942421504 }, { "epoch": 0.18, "learning_rate": 8.310062590274435e-05, "loss": 2.5764, "theoretical_loss": 3.4387604576127053, "tokens_seen": 1942552576 }, { "epoch": 0.18, "learning_rate": 8.309260150858611e-05, "loss": 2.3756, "theoretical_loss": 3.438741001469878, "tokens_seen": 1942683648 }, { "epoch": 0.18, "learning_rate": 8.308457711442787e-05, "loss": 2.4454, "theoretical_loss": 3.43872154700723, "tokens_seen": 1942814720 }, { "epoch": 0.18, "learning_rate": 8.307655272026962e-05, "loss": 2.5725, "theoretical_loss": 3.438702094224504, "tokens_seen": 1942945792 }, { "epoch": 0.18, "learning_rate": 8.306852832611139e-05, "loss": 2.4494, "theoretical_loss": 3.438682643121441, "tokens_seen": 1943076864 }, { "epoch": 0.18, "learning_rate": 8.306050393195314e-05, "loss": 2.3641, "theoretical_loss": 3.4386631936977827, "tokens_seen": 1943207936 }, { "epoch": 0.18, "objective/train/docs_used": 1070317, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5918591022491455, "objective/train/theoretical_loss": 3.43865346961565, "objective/train/tokens_used": 313732576, "theoretical_loss": 3.43865346961565, "tokens_seen": 1943273472 }, { "epoch": 0.18, "learning_rate": 8.305247953779491e-05, "loss": 2.6276, "theoretical_loss": 3.4386437459532715, "tokens_seen": 1943339008 }, { "epoch": 0.18, "learning_rate": 8.304445514363666e-05, "loss": 2.5815, "theoretical_loss": 3.438624299887648, "tokens_seen": 1943470080 }, { "epoch": 0.18, "learning_rate": 8.303643074947843e-05, "loss": 2.3985, "theoretical_loss": 3.438604855500655, "tokens_seen": 1943601152 }, { "epoch": 0.18, "learning_rate": 8.302840635532018e-05, "loss": 2.5067, "theoretical_loss": 3.438585412792034, "tokens_seen": 1943732224 }, { "epoch": 0.18, "learning_rate": 8.302038196116195e-05, "loss": 2.605, "theoretical_loss": 3.438565971761528, "tokens_seen": 1943863296 }, { "epoch": 0.18, "learning_rate": 8.30123575670037e-05, "loss": 2.4195, "theoretical_loss": 3.4385465324088766, "tokens_seen": 1943994368 }, { "epoch": 0.18, "learning_rate": 8.300433317284545e-05, "loss": 2.5353, "theoretical_loss": 3.4385270947338245, "tokens_seen": 1944125440 }, { "epoch": 0.18, "learning_rate": 8.299630877868722e-05, "loss": 2.5632, "theoretical_loss": 3.4385076587361123, "tokens_seen": 1944256512 }, { "epoch": 0.18, "learning_rate": 8.298828438452897e-05, "loss": 2.4589, "theoretical_loss": 3.438488224415483, "tokens_seen": 1944387584 }, { "epoch": 0.18, "learning_rate": 8.298025999037074e-05, "loss": 2.4504, "theoretical_loss": 3.438468791771678, "tokens_seen": 1944518656 }, { "epoch": 0.18, "learning_rate": 8.297223559621249e-05, "loss": 2.7096, "theoretical_loss": 3.4384493608044404, "tokens_seen": 1944649728 }, { "epoch": 0.18, "learning_rate": 8.296421120205426e-05, "loss": 2.494, "theoretical_loss": 3.4384299315135123, "tokens_seen": 1944780800 }, { "epoch": 0.18, "objective/train/docs_used": 1070782, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.948735237121582, "objective/train/theoretical_loss": 3.438410503898636, "objective/train/tokens_used": 315370976, "theoretical_loss": 3.438410503898636, "tokens_seen": 1944911872 }, { "epoch": 0.18, "learning_rate": 8.295618680789601e-05, "loss": 2.5633, "theoretical_loss": 3.438410503898636, "tokens_seen": 1944911872 }, { "epoch": 0.18, "learning_rate": 8.294816241373777e-05, "loss": 2.4825, "theoretical_loss": 3.4383910779595537, "tokens_seen": 1945042944 }, { "epoch": 0.18, "learning_rate": 8.294013801957953e-05, "loss": 2.593, "theoretical_loss": 3.438371653696009, "tokens_seen": 1945174016 }, { "epoch": 0.18, "learning_rate": 8.293211362542129e-05, "loss": 2.4795, "theoretical_loss": 3.4383522311077432, "tokens_seen": 1945305088 }, { "epoch": 0.18, "learning_rate": 8.292408923126305e-05, "loss": 2.6767, "theoretical_loss": 3.4383328101945, "tokens_seen": 1945436160 }, { "epoch": 0.18, "learning_rate": 8.29160648371048e-05, "loss": 2.5267, "theoretical_loss": 3.4383133909560213, "tokens_seen": 1945567232 }, { "epoch": 0.18, "learning_rate": 8.290804044294657e-05, "loss": 2.5432, "theoretical_loss": 3.4382939733920503, "tokens_seen": 1945698304 }, { "epoch": 0.18, "learning_rate": 8.290001604878832e-05, "loss": 2.5006, "theoretical_loss": 3.4382745575023295, "tokens_seen": 1945829376 }, { "epoch": 0.18, "learning_rate": 8.289199165463008e-05, "loss": 2.5887, "theoretical_loss": 3.4382551432866024, "tokens_seen": 1945960448 }, { "epoch": 0.18, "learning_rate": 8.288396726047184e-05, "loss": 2.4243, "theoretical_loss": 3.438235730744611, "tokens_seen": 1946091520 }, { "epoch": 0.18, "learning_rate": 8.28759428663136e-05, "loss": 2.5427, "theoretical_loss": 3.4382163198760995, "tokens_seen": 1946222592 }, { "epoch": 0.18, "learning_rate": 8.286791847215536e-05, "loss": 2.6092, "theoretical_loss": 3.43819691068081, "tokens_seen": 1946353664 }, { "epoch": 0.18, "learning_rate": 8.285989407799712e-05, "loss": 2.4251, "theoretical_loss": 3.4381775031584856, "tokens_seen": 1946484736 }, { "epoch": 0.18, "objective/train/docs_used": 1072031, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5232903957366943, "objective/train/theoretical_loss": 3.4381678000246056, "objective/train/tokens_used": 317009376, "theoretical_loss": 3.4381678000246056, "tokens_seen": 1946550272 }, { "epoch": 0.18, "learning_rate": 8.285186968383888e-05, "loss": 2.6291, "theoretical_loss": 3.43815809730887, "tokens_seen": 1946615808 }, { "epoch": 0.18, "learning_rate": 8.284384528968064e-05, "loss": 2.5281, "theoretical_loss": 3.438138693131706, "tokens_seen": 1946746880 }, { "epoch": 0.18, "learning_rate": 8.283582089552239e-05, "loss": 2.4736, "theoretical_loss": 3.4381192906267373, "tokens_seen": 1946877952 }, { "epoch": 0.18, "learning_rate": 8.282779650136416e-05, "loss": 2.538, "theoretical_loss": 3.4380998897937065, "tokens_seen": 1947009024 }, { "epoch": 0.18, "learning_rate": 8.281977210720591e-05, "loss": 2.4369, "theoretical_loss": 3.438080490632358, "tokens_seen": 1947140096 }, { "epoch": 0.18, "learning_rate": 8.281174771304768e-05, "loss": 2.7668, "theoretical_loss": 3.4380610931424345, "tokens_seen": 1947271168 }, { "epoch": 0.18, "learning_rate": 8.280372331888943e-05, "loss": 2.5402, "theoretical_loss": 3.438041697323679, "tokens_seen": 1947402240 }, { "epoch": 0.18, "learning_rate": 8.27956989247312e-05, "loss": 2.499, "theoretical_loss": 3.4380223031758366, "tokens_seen": 1947533312 }, { "epoch": 0.18, "learning_rate": 8.278767453057295e-05, "loss": 2.3226, "theoretical_loss": 3.43800291069865, "tokens_seen": 1947664384 }, { "epoch": 0.18, "learning_rate": 8.27796501364147e-05, "loss": 2.5558, "theoretical_loss": 3.4379835198918625, "tokens_seen": 1947795456 }, { "epoch": 0.18, "learning_rate": 8.277162574225647e-05, "loss": 2.5019, "theoretical_loss": 3.4379641307552182, "tokens_seen": 1947926528 }, { "epoch": 0.18, "learning_rate": 8.276360134809822e-05, "loss": 2.51, "theoretical_loss": 3.4379447432884613, "tokens_seen": 1948057600 }, { "epoch": 0.18, "objective/train/docs_used": 1072626, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2674527168273926, "objective/train/theoretical_loss": 3.437925357491335, "objective/train/tokens_used": 318647776, "theoretical_loss": 3.437925357491335, "tokens_seen": 1948188672 }, { "epoch": 0.18, "learning_rate": 8.275557695393999e-05, "loss": 2.5014, "theoretical_loss": 3.437925357491335, "tokens_seen": 1948188672 }, { "epoch": 0.18, "learning_rate": 8.274755255978174e-05, "loss": 2.5272, "theoretical_loss": 3.437905973363584, "tokens_seen": 1948319744 }, { "epoch": 0.18, "learning_rate": 8.273952816562351e-05, "loss": 2.659, "theoretical_loss": 3.4378865909049514, "tokens_seen": 1948450816 }, { "epoch": 0.18, "learning_rate": 8.273150377146526e-05, "loss": 2.5092, "theoretical_loss": 3.437867210115181, "tokens_seen": 1948581888 }, { "epoch": 0.18, "learning_rate": 8.272347937730703e-05, "loss": 2.5668, "theoretical_loss": 3.437847830994018, "tokens_seen": 1948712960 }, { "epoch": 0.18, "learning_rate": 8.271545498314878e-05, "loss": 2.5462, "theoretical_loss": 3.437828453541206, "tokens_seen": 1948844032 }, { "epoch": 0.18, "learning_rate": 8.270743058899054e-05, "loss": 2.7522, "theoretical_loss": 3.437809077756489, "tokens_seen": 1948975104 }, { "epoch": 0.18, "learning_rate": 8.26994061948323e-05, "loss": 2.6027, "theoretical_loss": 3.437789703639611, "tokens_seen": 1949106176 }, { "epoch": 0.18, "learning_rate": 8.269138180067406e-05, "loss": 2.5426, "theoretical_loss": 3.4377703311903174, "tokens_seen": 1949237248 }, { "epoch": 0.18, "learning_rate": 8.268335740651582e-05, "loss": 2.6411, "theoretical_loss": 3.437750960408351, "tokens_seen": 1949368320 }, { "epoch": 0.18, "learning_rate": 8.267533301235758e-05, "loss": 2.4644, "theoretical_loss": 3.437731591293457, "tokens_seen": 1949499392 }, { "epoch": 0.18, "learning_rate": 8.266730861819934e-05, "loss": 2.5176, "theoretical_loss": 3.4377122238453803, "tokens_seen": 1949630464 }, { "epoch": 0.18, "learning_rate": 8.26592842240411e-05, "loss": 2.6194, "theoretical_loss": 3.4376928580638646, "tokens_seen": 1949761536 }, { "epoch": 0.18, "objective/train/docs_used": 1073976, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5130059719085693, "objective/train/theoretical_loss": 3.4376831757979875, "objective/train/tokens_used": 320286176, "theoretical_loss": 3.4376831757979875, "tokens_seen": 1949827072 }, { "epoch": 0.18, "learning_rate": 8.265125982988285e-05, "loss": 2.536, "theoretical_loss": 3.437673493948655, "tokens_seen": 1949892608 }, { "epoch": 0.18, "learning_rate": 8.264323543572461e-05, "loss": 2.6313, "theoretical_loss": 3.437654131499496, "tokens_seen": 1950023680 }, { "epoch": 0.18, "learning_rate": 8.263521104156637e-05, "loss": 2.5475, "theoretical_loss": 3.437634770716132, "tokens_seen": 1950154752 }, { "epoch": 0.18, "learning_rate": 8.262718664740813e-05, "loss": 2.6455, "theoretical_loss": 3.4376154115983084, "tokens_seen": 1950285824 }, { "epoch": 0.18, "learning_rate": 8.261916225324989e-05, "loss": 2.3373, "theoretical_loss": 3.4375960541457697, "tokens_seen": 1950416896 }, { "epoch": 0.18, "learning_rate": 8.261113785909164e-05, "loss": 2.4703, "theoretical_loss": 3.43757669835826, "tokens_seen": 1950547968 }, { "epoch": 0.18, "learning_rate": 8.260311346493341e-05, "loss": 2.5127, "theoretical_loss": 3.4375573442355254, "tokens_seen": 1950679040 }, { "epoch": 0.18, "learning_rate": 8.259508907077516e-05, "loss": 2.3909, "theoretical_loss": 3.43753799177731, "tokens_seen": 1950810112 }, { "epoch": 0.18, "learning_rate": 8.258706467661693e-05, "loss": 2.4477, "theoretical_loss": 3.437518640983359, "tokens_seen": 1950941184 }, { "epoch": 0.18, "learning_rate": 8.257904028245868e-05, "loss": 2.3993, "theoretical_loss": 3.4374992918534177, "tokens_seen": 1951072256 }, { "epoch": 0.18, "learning_rate": 8.257101588830045e-05, "loss": 2.5379, "theoretical_loss": 3.4374799443872317, "tokens_seen": 1951203328 }, { "epoch": 0.18, "learning_rate": 8.25629914941422e-05, "loss": 2.5257, "theoretical_loss": 3.4374605985845452, "tokens_seen": 1951334400 }, { "epoch": 0.18, "objective/train/docs_used": 1074351, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.568186044692993, "objective/train/theoretical_loss": 3.437441254445104, "objective/train/tokens_used": 321924576, "theoretical_loss": 3.437441254445104, "tokens_seen": 1951465472 }, { "epoch": 0.18, "learning_rate": 8.255496709998395e-05, "loss": 2.6717, "theoretical_loss": 3.437441254445104, "tokens_seen": 1951465472 }, { "epoch": 0.18, "learning_rate": 8.254694270582572e-05, "loss": 2.6209, "theoretical_loss": 3.437421911968653, "tokens_seen": 1951596544 }, { "epoch": 0.18, "learning_rate": 8.253891831166747e-05, "loss": 2.4163, "theoretical_loss": 3.4374025711549385, "tokens_seen": 1951727616 }, { "epoch": 0.18, "learning_rate": 8.253089391750924e-05, "loss": 2.5627, "theoretical_loss": 3.437383232003705, "tokens_seen": 1951858688 }, { "epoch": 0.18, "learning_rate": 8.252286952335099e-05, "loss": 2.4633, "theoretical_loss": 3.4373638945146983, "tokens_seen": 1951989760 }, { "epoch": 0.18, "learning_rate": 8.251484512919275e-05, "loss": 2.5232, "theoretical_loss": 3.4373445586876636, "tokens_seen": 1952120832 }, { "epoch": 0.18, "learning_rate": 8.250682073503451e-05, "loss": 2.5546, "theoretical_loss": 3.4373252245223473, "tokens_seen": 1952251904 }, { "epoch": 0.18, "learning_rate": 8.249879634087627e-05, "loss": 2.5288, "theoretical_loss": 3.437305892018494, "tokens_seen": 1952382976 }, { "epoch": 0.18, "learning_rate": 8.249077194671803e-05, "loss": 2.6947, "theoretical_loss": 3.43728656117585, "tokens_seen": 1952514048 }, { "epoch": 0.18, "learning_rate": 8.248274755255979e-05, "loss": 2.6388, "theoretical_loss": 3.437267231994161, "tokens_seen": 1952645120 }, { "epoch": 0.18, "learning_rate": 8.247472315840154e-05, "loss": 2.5857, "theoretical_loss": 3.437247904473173, "tokens_seen": 1952776192 }, { "epoch": 0.18, "learning_rate": 8.24666987642433e-05, "loss": 2.5235, "theoretical_loss": 3.4372285786126318, "tokens_seen": 1952907264 }, { "epoch": 0.18, "learning_rate": 8.245867437008506e-05, "loss": 2.5955, "theoretical_loss": 3.4372092544122825, "tokens_seen": 1953038336 }, { "epoch": 0.18, "objective/train/docs_used": 1075675, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2708775997161865, "objective/train/theoretical_loss": 3.4371995929346006, "objective/train/tokens_used": 323562976, "theoretical_loss": 3.4371995929346006, "tokens_seen": 1953103872 }, { "epoch": 0.18, "learning_rate": 8.245064997592683e-05, "loss": 2.5938, "theoretical_loss": 3.437189931871872, "tokens_seen": 1953169408 }, { "epoch": 0.18, "learning_rate": 8.244262558176858e-05, "loss": 2.6016, "theoretical_loss": 3.437170610991146, "tokens_seen": 1953300480 }, { "epoch": 0.18, "learning_rate": 8.243460118761035e-05, "loss": 2.3854, "theoretical_loss": 3.437151291769851, "tokens_seen": 1953431552 }, { "epoch": 0.18, "learning_rate": 8.24265767934521e-05, "loss": 2.59, "theoretical_loss": 3.4371319742077326, "tokens_seen": 1953562624 }, { "epoch": 0.18, "learning_rate": 8.241855239929385e-05, "loss": 2.4944, "theoretical_loss": 3.437112658304537, "tokens_seen": 1953693696 }, { "epoch": 0.18, "learning_rate": 8.241052800513562e-05, "loss": 2.6604, "theoretical_loss": 3.437093344060011, "tokens_seen": 1953824768 }, { "epoch": 0.18, "learning_rate": 8.240250361097737e-05, "loss": 2.4924, "theoretical_loss": 3.437074031473901, "tokens_seen": 1953955840 }, { "epoch": 0.18, "learning_rate": 8.239447921681914e-05, "loss": 2.5443, "theoretical_loss": 3.4370547205459516, "tokens_seen": 1954086912 }, { "epoch": 0.18, "learning_rate": 8.238645482266089e-05, "loss": 2.5714, "theoretical_loss": 3.4370354112759114, "tokens_seen": 1954217984 }, { "epoch": 0.18, "learning_rate": 8.237843042850264e-05, "loss": 2.5103, "theoretical_loss": 3.4370161036635256, "tokens_seen": 1954349056 }, { "epoch": 0.18, "learning_rate": 8.237040603434441e-05, "loss": 2.5072, "theoretical_loss": 3.4369967977085416, "tokens_seen": 1954480128 }, { "epoch": 0.18, "learning_rate": 8.236238164018616e-05, "loss": 2.4739, "theoretical_loss": 3.436977493410706, "tokens_seen": 1954611200 }, { "epoch": 0.18, "objective/train/docs_used": 1076263, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2574844360351562, "objective/train/theoretical_loss": 3.436958190769764, "objective/train/tokens_used": 325201376, "theoretical_loss": 3.436958190769764, "tokens_seen": 1954742272 }, { "epoch": 0.18, "learning_rate": 8.235435724602793e-05, "loss": 2.5272, "theoretical_loss": 3.436958190769764, "tokens_seen": 1954742272 }, { "epoch": 0.18, "learning_rate": 8.234633285186968e-05, "loss": 2.5859, "theoretical_loss": 3.4369388897854636, "tokens_seen": 1954873344 }, { "epoch": 0.18, "learning_rate": 8.233830845771145e-05, "loss": 2.4542, "theoretical_loss": 3.436919590457552, "tokens_seen": 1955004416 }, { "epoch": 0.18, "learning_rate": 8.23302840635532e-05, "loss": 2.6658, "theoretical_loss": 3.4369002927857744, "tokens_seen": 1955135488 }, { "epoch": 0.19, "learning_rate": 8.232225966939496e-05, "loss": 2.5128, "theoretical_loss": 3.436880996769879, "tokens_seen": 1955266560 }, { "epoch": 0.19, "learning_rate": 8.231423527523672e-05, "loss": 2.4474, "theoretical_loss": 3.436861702409612, "tokens_seen": 1955397632 }, { "epoch": 0.19, "learning_rate": 8.230621088107848e-05, "loss": 2.6982, "theoretical_loss": 3.4368424097047208, "tokens_seen": 1955528704 }, { "epoch": 0.19, "learning_rate": 8.229818648692024e-05, "loss": 2.4142, "theoretical_loss": 3.4368231186549525, "tokens_seen": 1955659776 }, { "epoch": 0.19, "learning_rate": 8.2290162092762e-05, "loss": 2.4942, "theoretical_loss": 3.436803829260054, "tokens_seen": 1955790848 }, { "epoch": 0.19, "learning_rate": 8.228213769860375e-05, "loss": 2.5114, "theoretical_loss": 3.4367845415197715, "tokens_seen": 1955921920 }, { "epoch": 0.19, "learning_rate": 8.227411330444552e-05, "loss": 2.4439, "theoretical_loss": 3.436765255433854, "tokens_seen": 1956052992 }, { "epoch": 0.19, "learning_rate": 8.226608891028727e-05, "loss": 2.5427, "theoretical_loss": 3.436745971002048, "tokens_seen": 1956184064 }, { "epoch": 0.19, "learning_rate": 8.225806451612904e-05, "loss": 2.5254, "theoretical_loss": 3.4367266882241, "tokens_seen": 1956315136 }, { "epoch": 0.19, "objective/train/docs_used": 1077404, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.399874210357666, "objective/train/theoretical_loss": 3.436717047455245, "objective/train/tokens_used": 326839776, "theoretical_loss": 3.436717047455245, "tokens_seen": 1956380672 }, { "epoch": 0.19, "learning_rate": 8.225004012197079e-05, "loss": 2.5416, "theoretical_loss": 3.436707407099759, "tokens_seen": 1956446208 }, { "epoch": 0.19, "learning_rate": 8.224201572781256e-05, "loss": 2.7036, "theoretical_loss": 3.436688127628771, "tokens_seen": 1956577280 }, { "epoch": 0.19, "learning_rate": 8.223399133365431e-05, "loss": 2.6167, "theoretical_loss": 3.436668849810884, "tokens_seen": 1956708352 }, { "epoch": 0.19, "learning_rate": 8.222596693949606e-05, "loss": 2.5307, "theoretical_loss": 3.4366495736458456, "tokens_seen": 1956839424 }, { "epoch": 0.19, "learning_rate": 8.221794254533783e-05, "loss": 2.6293, "theoretical_loss": 3.436630299133403, "tokens_seen": 1956970496 }, { "epoch": 0.19, "learning_rate": 8.220991815117958e-05, "loss": 2.5141, "theoretical_loss": 3.4366110262733045, "tokens_seen": 1957101568 }, { "epoch": 0.19, "learning_rate": 8.220189375702135e-05, "loss": 2.4892, "theoretical_loss": 3.4365917550652973, "tokens_seen": 1957232640 }, { "epoch": 0.19, "learning_rate": 8.21938693628631e-05, "loss": 2.6137, "theoretical_loss": 3.4365724855091293, "tokens_seen": 1957363712 }, { "epoch": 0.19, "learning_rate": 8.218584496870485e-05, "loss": 2.4596, "theoretical_loss": 3.4365532176045486, "tokens_seen": 1957494784 }, { "epoch": 0.19, "learning_rate": 8.217782057454662e-05, "loss": 2.5889, "theoretical_loss": 3.4365339513513025, "tokens_seen": 1957625856 }, { "epoch": 0.19, "learning_rate": 8.216979618038837e-05, "loss": 2.5195, "theoretical_loss": 3.436514686749139, "tokens_seen": 1957756928 }, { "epoch": 0.19, "learning_rate": 8.216177178623014e-05, "loss": 2.48, "theoretical_loss": 3.4364954237978065, "tokens_seen": 1957888000 }, { "epoch": 0.19, "objective/train/docs_used": 1077881, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7305192947387695, "objective/train/theoretical_loss": 3.436476162497053, "objective/train/tokens_used": 328478176, "theoretical_loss": 3.436476162497053, "tokens_seen": 1958019072 }, { "epoch": 0.19, "learning_rate": 8.21537473920719e-05, "loss": 2.5058, "theoretical_loss": 3.436476162497053, "tokens_seen": 1958019072 }, { "epoch": 0.19, "learning_rate": 8.214572299791366e-05, "loss": 2.514, "theoretical_loss": 3.4364569028466265, "tokens_seen": 1958150144 }, { "epoch": 0.19, "learning_rate": 8.213769860375541e-05, "loss": 2.4948, "theoretical_loss": 3.436437644846275, "tokens_seen": 1958281216 }, { "epoch": 0.19, "learning_rate": 8.212967420959717e-05, "loss": 2.4609, "theoretical_loss": 3.4364183884957464, "tokens_seen": 1958412288 }, { "epoch": 0.19, "learning_rate": 8.212164981543893e-05, "loss": 2.6224, "theoretical_loss": 3.43639913379479, "tokens_seen": 1958543360 }, { "epoch": 0.19, "learning_rate": 8.211362542128069e-05, "loss": 2.5121, "theoretical_loss": 3.4363798807431527, "tokens_seen": 1958674432 }, { "epoch": 0.19, "learning_rate": 8.210560102712245e-05, "loss": 2.5246, "theoretical_loss": 3.4363606293405837, "tokens_seen": 1958805504 }, { "epoch": 0.19, "learning_rate": 8.209757663296421e-05, "loss": 2.598, "theoretical_loss": 3.436341379586832, "tokens_seen": 1958936576 }, { "epoch": 0.19, "learning_rate": 8.208955223880597e-05, "loss": 2.6481, "theoretical_loss": 3.436322131481645, "tokens_seen": 1959067648 }, { "epoch": 0.19, "learning_rate": 8.208152784464773e-05, "loss": 2.3418, "theoretical_loss": 3.4363028850247717, "tokens_seen": 1959198720 }, { "epoch": 0.19, "learning_rate": 8.20735034504895e-05, "loss": 2.378, "theoretical_loss": 3.4362836402159607, "tokens_seen": 1959329792 }, { "epoch": 0.19, "learning_rate": 8.206547905633125e-05, "loss": 2.6566, "theoretical_loss": 3.43626439705496, "tokens_seen": 1959460864 }, { "epoch": 0.19, "learning_rate": 8.2057454662173e-05, "loss": 2.5246, "theoretical_loss": 3.4362451555415197, "tokens_seen": 1959591936 }, { "epoch": 0.19, "objective/train/docs_used": 1079297, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.692366600036621, "objective/train/theoretical_loss": 3.436235535402555, "objective/train/tokens_used": 330116576, "theoretical_loss": 3.436235535402555, "tokens_seen": 1959657472 }, { "epoch": 0.19, "learning_rate": 8.204943026801477e-05, "loss": 2.4392, "theoretical_loss": 3.4362259156753874, "tokens_seen": 1959723008 }, { "epoch": 0.19, "learning_rate": 8.204140587385652e-05, "loss": 2.5496, "theoretical_loss": 3.436206677456312, "tokens_seen": 1959854080 }, { "epoch": 0.19, "learning_rate": 8.203338147969829e-05, "loss": 2.4338, "theoretical_loss": 3.4361874408840425, "tokens_seen": 1959985152 }, { "epoch": 0.19, "learning_rate": 8.202535708554004e-05, "loss": 2.618, "theoretical_loss": 3.436168205958328, "tokens_seen": 1960116224 }, { "epoch": 0.19, "learning_rate": 8.20173326913818e-05, "loss": 2.4057, "theoretical_loss": 3.436148972678917, "tokens_seen": 1960247296 }, { "epoch": 0.19, "learning_rate": 8.200930829722356e-05, "loss": 2.6132, "theoretical_loss": 3.43612974104556, "tokens_seen": 1960378368 }, { "epoch": 0.19, "learning_rate": 8.200128390306531e-05, "loss": 2.3999, "theoretical_loss": 3.436110511058004, "tokens_seen": 1960509440 }, { "epoch": 0.19, "learning_rate": 8.199325950890708e-05, "loss": 2.5032, "theoretical_loss": 3.4360912827159997, "tokens_seen": 1960640512 }, { "epoch": 0.19, "learning_rate": 8.198523511474883e-05, "loss": 2.4567, "theoretical_loss": 3.436072056019295, "tokens_seen": 1960771584 }, { "epoch": 0.19, "learning_rate": 8.19772107205906e-05, "loss": 2.5605, "theoretical_loss": 3.4360528309676406, "tokens_seen": 1960902656 }, { "epoch": 0.19, "learning_rate": 8.196918632643235e-05, "loss": 2.5381, "theoretical_loss": 3.436033607560785, "tokens_seen": 1961033728 }, { "epoch": 0.19, "learning_rate": 8.196116193227412e-05, "loss": 2.5904, "theoretical_loss": 3.4360143857984777, "tokens_seen": 1961164800 }, { "epoch": 0.19, "objective/train/docs_used": 1080387, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7845823764801025, "objective/train/theoretical_loss": 3.4359951656804677, "objective/train/tokens_used": 331754976, "theoretical_loss": 3.4359951656804677, "tokens_seen": 1961295872 }, { "epoch": 0.19, "learning_rate": 8.195313753811587e-05, "loss": 2.6118, "theoretical_loss": 3.4359951656804677, "tokens_seen": 1961295872 }, { "epoch": 0.19, "learning_rate": 8.194511314395762e-05, "loss": 2.5112, "theoretical_loss": 3.435975947206505, "tokens_seen": 1961426944 }, { "epoch": 0.19, "learning_rate": 8.193708874979939e-05, "loss": 2.5616, "theoretical_loss": 3.4359567303763385, "tokens_seen": 1961558016 }, { "epoch": 0.19, "learning_rate": 8.192906435564114e-05, "loss": 2.472, "theoretical_loss": 3.4359375151897185, "tokens_seen": 1961689088 }, { "epoch": 0.19, "learning_rate": 8.192103996148291e-05, "loss": 2.5179, "theoretical_loss": 3.4359183016463946, "tokens_seen": 1961820160 }, { "epoch": 0.19, "learning_rate": 8.191301556732466e-05, "loss": 2.4378, "theoretical_loss": 3.435899089746116, "tokens_seen": 1961951232 }, { "epoch": 0.19, "learning_rate": 8.190499117316643e-05, "loss": 2.3945, "theoretical_loss": 3.435879879488633, "tokens_seen": 1962082304 }, { "epoch": 0.19, "learning_rate": 8.189696677900818e-05, "loss": 2.592, "theoretical_loss": 3.4358606708736943, "tokens_seen": 1962213376 }, { "epoch": 0.19, "learning_rate": 8.188894238484994e-05, "loss": 2.5364, "theoretical_loss": 3.435841463901051, "tokens_seen": 1962344448 }, { "epoch": 0.19, "learning_rate": 8.18809179906917e-05, "loss": 2.5239, "theoretical_loss": 3.4358222585704525, "tokens_seen": 1962475520 }, { "epoch": 0.19, "learning_rate": 8.187289359653346e-05, "loss": 2.5953, "theoretical_loss": 3.4358030548816485, "tokens_seen": 1962606592 }, { "epoch": 0.19, "learning_rate": 8.186486920237522e-05, "loss": 2.4442, "theoretical_loss": 3.4357838528343896, "tokens_seen": 1962737664 }, { "epoch": 0.19, "learning_rate": 8.185684480821698e-05, "loss": 2.5722, "theoretical_loss": 3.4357646524284258, "tokens_seen": 1962868736 }, { "epoch": 0.19, "objective/train/docs_used": 1080811, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.384798049926758, "objective/train/theoretical_loss": 3.435755052840851, "objective/train/tokens_used": 333393376, "theoretical_loss": 3.435755052840851, "tokens_seen": 1962934272 }, { "epoch": 0.19, "learning_rate": 8.184882041405874e-05, "loss": 2.4926, "theoretical_loss": 3.4357454536635066, "tokens_seen": 1962999808 }, { "epoch": 0.19, "learning_rate": 8.18407960199005e-05, "loss": 2.5586, "theoretical_loss": 3.4357262565393825, "tokens_seen": 1963130880 }, { "epoch": 0.19, "learning_rate": 8.183277162574225e-05, "loss": 2.5849, "theoretical_loss": 3.4357070610558043, "tokens_seen": 1963261952 }, { "epoch": 0.19, "learning_rate": 8.182474723158402e-05, "loss": 2.5331, "theoretical_loss": 3.4356878672125215, "tokens_seen": 1963393024 }, { "epoch": 0.19, "learning_rate": 8.181672283742577e-05, "loss": 2.433, "theoretical_loss": 3.4356686750092846, "tokens_seen": 1963524096 }, { "epoch": 0.19, "learning_rate": 8.180869844326754e-05, "loss": 2.6144, "theoretical_loss": 3.4356494844458445, "tokens_seen": 1963655168 }, { "epoch": 0.19, "learning_rate": 8.180067404910929e-05, "loss": 2.5068, "theoretical_loss": 3.435630295521951, "tokens_seen": 1963786240 }, { "epoch": 0.19, "learning_rate": 8.179264965495106e-05, "loss": 2.4976, "theoretical_loss": 3.4356111082373553, "tokens_seen": 1963917312 }, { "epoch": 0.19, "learning_rate": 8.178462526079281e-05, "loss": 2.5668, "theoretical_loss": 3.435591922591807, "tokens_seen": 1964048384 }, { "epoch": 0.19, "learning_rate": 8.177660086663458e-05, "loss": 2.5539, "theoretical_loss": 3.4355727385850576, "tokens_seen": 1964179456 }, { "epoch": 0.19, "learning_rate": 8.176857647247633e-05, "loss": 2.4739, "theoretical_loss": 3.435553556216858, "tokens_seen": 1964310528 }, { "epoch": 0.19, "learning_rate": 8.176055207831808e-05, "loss": 2.4657, "theoretical_loss": 3.4355343754869576, "tokens_seen": 1964441600 }, { "epoch": 0.19, "objective/train/docs_used": 1082085, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.1325109004974365, "objective/train/theoretical_loss": 3.4355151963951083, "objective/train/tokens_used": 335031776, "theoretical_loss": 3.4355151963951083, "tokens_seen": 1964572672 }, { "epoch": 0.19, "learning_rate": 8.175252768415985e-05, "loss": 2.5154, "theoretical_loss": 3.4355151963951083, "tokens_seen": 1964572672 }, { "epoch": 0.19, "learning_rate": 8.17445032900016e-05, "loss": 2.586, "theoretical_loss": 3.4354960189410604, "tokens_seen": 1964703744 }, { "epoch": 0.19, "learning_rate": 8.173647889584337e-05, "loss": 2.6418, "theoretical_loss": 3.4354768431245652, "tokens_seen": 1964834816 }, { "epoch": 0.19, "learning_rate": 8.172845450168512e-05, "loss": 2.566, "theoretical_loss": 3.4354576689453733, "tokens_seen": 1964965888 }, { "epoch": 0.19, "learning_rate": 8.172043010752689e-05, "loss": 2.5053, "theoretical_loss": 3.4354384964032363, "tokens_seen": 1965096960 }, { "epoch": 0.19, "learning_rate": 8.171240571336864e-05, "loss": 2.4657, "theoretical_loss": 3.435419325497904, "tokens_seen": 1965228032 }, { "epoch": 0.19, "learning_rate": 8.17043813192104e-05, "loss": 2.5324, "theoretical_loss": 3.4354001562291288, "tokens_seen": 1965359104 }, { "epoch": 0.19, "learning_rate": 8.169635692505216e-05, "loss": 2.468, "theoretical_loss": 3.4353809885966613, "tokens_seen": 1965490176 }, { "epoch": 0.19, "learning_rate": 8.168833253089391e-05, "loss": 2.5399, "theoretical_loss": 3.4353618226002527, "tokens_seen": 1965621248 }, { "epoch": 0.19, "learning_rate": 8.168030813673568e-05, "loss": 2.5918, "theoretical_loss": 3.4353426582396542, "tokens_seen": 1965752320 }, { "epoch": 0.19, "learning_rate": 8.167228374257743e-05, "loss": 2.5126, "theoretical_loss": 3.4353234955146177, "tokens_seen": 1965883392 }, { "epoch": 0.19, "learning_rate": 8.16642593484192e-05, "loss": 2.4399, "theoretical_loss": 3.4353043344248944, "tokens_seen": 1966014464 }, { "epoch": 0.19, "learning_rate": 8.165623495426095e-05, "loss": 2.465, "theoretical_loss": 3.435285174970235, "tokens_seen": 1966145536 }, { "epoch": 0.19, "objective/train/docs_used": 1082685, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4286997318267822, "objective/train/theoretical_loss": 3.435275595855977, "objective/train/tokens_used": 336670176, "theoretical_loss": 3.435275595855977, "tokens_seen": 1966211072 }, { "epoch": 0.19, "learning_rate": 8.164821056010271e-05, "loss": 2.4441, "theoretical_loss": 3.4352660171503913, "tokens_seen": 1966276608 }, { "epoch": 0.19, "learning_rate": 8.164018616594447e-05, "loss": 2.586, "theoretical_loss": 3.4352468609651154, "tokens_seen": 1966407680 }, { "epoch": 0.19, "learning_rate": 8.163216177178623e-05, "loss": 2.5347, "theoretical_loss": 3.4352277064141585, "tokens_seen": 1966538752 }, { "epoch": 0.19, "learning_rate": 8.1624137377628e-05, "loss": 2.4911, "theoretical_loss": 3.4352085534972723, "tokens_seen": 1966669824 }, { "epoch": 0.19, "learning_rate": 8.161611298346975e-05, "loss": 2.6297, "theoretical_loss": 3.435189402214208, "tokens_seen": 1966800896 }, { "epoch": 0.19, "learning_rate": 8.160808858931151e-05, "loss": 2.4255, "theoretical_loss": 3.4351702525647183, "tokens_seen": 1966931968 }, { "epoch": 0.19, "learning_rate": 8.160006419515327e-05, "loss": 2.5653, "theoretical_loss": 3.4351511045485545, "tokens_seen": 1967063040 }, { "epoch": 0.19, "learning_rate": 8.159203980099502e-05, "loss": 2.5436, "theoretical_loss": 3.4351319581654685, "tokens_seen": 1967194112 }, { "epoch": 0.19, "learning_rate": 8.158401540683679e-05, "loss": 2.6022, "theoretical_loss": 3.4351128134152122, "tokens_seen": 1967325184 }, { "epoch": 0.19, "learning_rate": 8.157599101267854e-05, "loss": 2.4349, "theoretical_loss": 3.4350936702975376, "tokens_seen": 1967456256 }, { "epoch": 0.19, "learning_rate": 8.15679666185203e-05, "loss": 2.4711, "theoretical_loss": 3.435074528812197, "tokens_seen": 1967587328 }, { "epoch": 0.19, "learning_rate": 8.155994222436206e-05, "loss": 2.5145, "theoretical_loss": 3.435055388958942, "tokens_seen": 1967718400 }, { "epoch": 0.19, "objective/train/docs_used": 1083643, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4880809783935547, "objective/train/theoretical_loss": 3.435036250737525, "objective/train/tokens_used": 338308576, "theoretical_loss": 3.435036250737525, "tokens_seen": 1967849472 }, { "epoch": 0.19, "learning_rate": 8.155191783020383e-05, "loss": 2.5831, "theoretical_loss": 3.435036250737525, "tokens_seen": 1967849472 }, { "epoch": 0.19, "learning_rate": 8.154389343604558e-05, "loss": 2.5233, "theoretical_loss": 3.4350171141476986, "tokens_seen": 1967980544 }, { "epoch": 0.19, "learning_rate": 8.153586904188733e-05, "loss": 2.4237, "theoretical_loss": 3.4349979791892142, "tokens_seen": 1968111616 }, { "epoch": 0.19, "learning_rate": 8.15278446477291e-05, "loss": 2.5631, "theoretical_loss": 3.4349788458618247, "tokens_seen": 1968242688 }, { "epoch": 0.19, "learning_rate": 8.151982025357085e-05, "loss": 2.4931, "theoretical_loss": 3.434959714165282, "tokens_seen": 1968373760 }, { "epoch": 0.19, "learning_rate": 8.151179585941262e-05, "loss": 2.7059, "theoretical_loss": 3.4349405840993392, "tokens_seen": 1968504832 }, { "epoch": 0.19, "learning_rate": 8.150377146525437e-05, "loss": 2.4426, "theoretical_loss": 3.434921455663748, "tokens_seen": 1968635904 }, { "epoch": 0.19, "learning_rate": 8.149574707109614e-05, "loss": 2.5333, "theoretical_loss": 3.4349023288582616, "tokens_seen": 1968766976 }, { "epoch": 0.19, "learning_rate": 8.148772267693789e-05, "loss": 2.5284, "theoretical_loss": 3.4348832036826322, "tokens_seen": 1968898048 }, { "epoch": 0.19, "learning_rate": 8.147969828277966e-05, "loss": 2.5927, "theoretical_loss": 3.434864080136612, "tokens_seen": 1969029120 }, { "epoch": 0.19, "learning_rate": 8.147167388862141e-05, "loss": 2.7439, "theoretical_loss": 3.434844958219955, "tokens_seen": 1969160192 }, { "epoch": 0.19, "learning_rate": 8.146364949446317e-05, "loss": 2.5252, "theoretical_loss": 3.434825837932413, "tokens_seen": 1969291264 }, { "epoch": 0.19, "learning_rate": 8.145562510030493e-05, "loss": 2.674, "theoretical_loss": 3.4348067192737384, "tokens_seen": 1969422336 }, { "epoch": 0.19, "objective/train/docs_used": 1084001, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6455793380737305, "objective/train/theoretical_loss": 3.4347971605551493, "objective/train/tokens_used": 339946976, "theoretical_loss": 3.4347971605551493, "tokens_seen": 1969487872 }, { "epoch": 0.19, "learning_rate": 8.144760070614668e-05, "loss": 2.4479, "theoretical_loss": 3.4347876022436843, "tokens_seen": 1969553408 }, { "epoch": 0.19, "learning_rate": 8.143957631198845e-05, "loss": 2.5325, "theoretical_loss": 3.434768486842005, "tokens_seen": 1969684480 }, { "epoch": 0.19, "learning_rate": 8.14315519178302e-05, "loss": 2.6488, "theoretical_loss": 3.434749373068451, "tokens_seen": 1969815552 }, { "epoch": 0.19, "learning_rate": 8.142352752367197e-05, "loss": 2.4881, "theoretical_loss": 3.4347302609227772, "tokens_seen": 1969946624 }, { "epoch": 0.19, "learning_rate": 8.141550312951372e-05, "loss": 2.4335, "theoretical_loss": 3.434711150404736, "tokens_seen": 1970077696 }, { "epoch": 0.19, "learning_rate": 8.140747873535548e-05, "loss": 2.48, "theoretical_loss": 3.4346920415140803, "tokens_seen": 1970208768 }, { "epoch": 0.19, "learning_rate": 8.139945434119724e-05, "loss": 2.6898, "theoretical_loss": 3.434672934250564, "tokens_seen": 1970339840 }, { "epoch": 0.19, "learning_rate": 8.1391429947039e-05, "loss": 2.5244, "theoretical_loss": 3.434653828613939, "tokens_seen": 1970470912 }, { "epoch": 0.19, "learning_rate": 8.138340555288076e-05, "loss": 2.4301, "theoretical_loss": 3.434634724603961, "tokens_seen": 1970601984 }, { "epoch": 0.19, "learning_rate": 8.137538115872252e-05, "loss": 2.5058, "theoretical_loss": 3.43461562222038, "tokens_seen": 1970733056 }, { "epoch": 0.19, "learning_rate": 8.136735676456428e-05, "loss": 2.5456, "theoretical_loss": 3.434596521462952, "tokens_seen": 1970864128 }, { "epoch": 0.19, "learning_rate": 8.135933237040604e-05, "loss": 2.4504, "theoretical_loss": 3.4345774223314294, "tokens_seen": 1970995200 }, { "epoch": 0.19, "objective/train/docs_used": 1085329, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.782360076904297, "objective/train/theoretical_loss": 3.4345583248255656, "objective/train/tokens_used": 341585376, "theoretical_loss": 3.4345583248255656, "tokens_seen": 1971126272 }, { "epoch": 0.19, "learning_rate": 8.135130797624779e-05, "loss": 2.4819, "theoretical_loss": 3.4345583248255656, "tokens_seen": 1971126272 }, { "epoch": 0.19, "learning_rate": 8.134328358208956e-05, "loss": 2.6041, "theoretical_loss": 3.434539228945115, "tokens_seen": 1971257344 }, { "epoch": 0.19, "learning_rate": 8.133525918793131e-05, "loss": 2.5261, "theoretical_loss": 3.43452013468983, "tokens_seen": 1971388416 }, { "epoch": 0.19, "learning_rate": 8.132723479377308e-05, "loss": 2.4584, "theoretical_loss": 3.4345010420594653, "tokens_seen": 1971519488 }, { "epoch": 0.19, "learning_rate": 8.131921039961483e-05, "loss": 2.3719, "theoretical_loss": 3.434481951053774, "tokens_seen": 1971650560 }, { "epoch": 0.2, "learning_rate": 8.13111860054566e-05, "loss": 2.5287, "theoretical_loss": 3.4344628616725097, "tokens_seen": 1971781632 }, { "epoch": 0.2, "learning_rate": 8.130316161129835e-05, "loss": 2.7084, "theoretical_loss": 3.4344437739154268, "tokens_seen": 1971912704 }, { "epoch": 0.2, "learning_rate": 8.12951372171401e-05, "loss": 2.4448, "theoretical_loss": 3.434424687782279, "tokens_seen": 1972043776 }, { "epoch": 0.2, "learning_rate": 8.128711282298187e-05, "loss": 2.3985, "theoretical_loss": 3.4344056032728196, "tokens_seen": 1972174848 }, { "epoch": 0.2, "learning_rate": 8.127908842882362e-05, "loss": 2.6247, "theoretical_loss": 3.4343865203868034, "tokens_seen": 1972305920 }, { "epoch": 0.2, "learning_rate": 8.127106403466539e-05, "loss": 2.6056, "theoretical_loss": 3.4343674391239842, "tokens_seen": 1972436992 }, { "epoch": 0.2, "learning_rate": 8.126303964050714e-05, "loss": 2.6542, "theoretical_loss": 3.4343483594841158, "tokens_seen": 1972568064 }, { "epoch": 0.2, "learning_rate": 8.125501524634891e-05, "loss": 2.48, "theoretical_loss": 3.4343292814669524, "tokens_seen": 1972699136 }, { "epoch": 0.2, "objective/train/docs_used": 1086429, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2790753841400146, "objective/train/theoretical_loss": 3.4343197430668084, "objective/train/tokens_used": 343223776, "theoretical_loss": 3.4343197430668084, "tokens_seen": 1972764672 }, { "epoch": 0.2, "learning_rate": 8.124699085219066e-05, "loss": 2.6051, "theoretical_loss": 3.4343102050722485, "tokens_seen": 1972830208 }, { "epoch": 0.2, "learning_rate": 8.123896645803242e-05, "loss": 2.5512, "theoretical_loss": 3.434291130299758, "tokens_seen": 1972961280 }, { "epoch": 0.2, "learning_rate": 8.123094206387418e-05, "loss": 2.3245, "theoretical_loss": 3.4342720571492356, "tokens_seen": 1973092352 }, { "epoch": 0.2, "learning_rate": 8.122291766971594e-05, "loss": 2.538, "theoretical_loss": 3.434252985620435, "tokens_seen": 1973223424 }, { "epoch": 0.2, "learning_rate": 8.12148932755577e-05, "loss": 2.6185, "theoretical_loss": 3.4342339157131114, "tokens_seen": 1973354496 }, { "epoch": 0.2, "learning_rate": 8.120686888139945e-05, "loss": 2.5536, "theoretical_loss": 3.434214847427018, "tokens_seen": 1973485568 }, { "epoch": 0.2, "learning_rate": 8.119884448724122e-05, "loss": 2.3628, "theoretical_loss": 3.434195780761911, "tokens_seen": 1973616640 }, { "epoch": 0.2, "learning_rate": 8.119082009308297e-05, "loss": 2.4919, "theoretical_loss": 3.434176715717544, "tokens_seen": 1973747712 }, { "epoch": 0.2, "learning_rate": 8.118279569892473e-05, "loss": 2.5678, "theoretical_loss": 3.4341576522936714, "tokens_seen": 1973878784 }, { "epoch": 0.2, "learning_rate": 8.11747713047665e-05, "loss": 2.6441, "theoretical_loss": 3.4341385904900483, "tokens_seen": 1974009856 }, { "epoch": 0.2, "learning_rate": 8.116674691060825e-05, "loss": 2.5687, "theoretical_loss": 3.4341195303064294, "tokens_seen": 1974140928 }, { "epoch": 0.2, "learning_rate": 8.115872251645001e-05, "loss": 2.5559, "theoretical_loss": 3.4341004717425694, "tokens_seen": 1974272000 }, { "epoch": 0.2, "objective/train/docs_used": 1087105, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.320133924484253, "objective/train/theoretical_loss": 3.4340814147982233, "objective/train/tokens_used": 344862176, "theoretical_loss": 3.4340814147982233, "tokens_seen": 1974403072 }, { "epoch": 0.2, "learning_rate": 8.115069812229177e-05, "loss": 2.4467, "theoretical_loss": 3.4340814147982233, "tokens_seen": 1974403072 }, { "epoch": 0.2, "learning_rate": 8.114267372813353e-05, "loss": 2.4401, "theoretical_loss": 3.4340623594731454, "tokens_seen": 1974534144 }, { "epoch": 0.2, "learning_rate": 8.113464933397529e-05, "loss": 2.6396, "theoretical_loss": 3.4340433057670907, "tokens_seen": 1974665216 }, { "epoch": 0.2, "learning_rate": 8.112662493981705e-05, "loss": 2.6161, "theoretical_loss": 3.434024253679815, "tokens_seen": 1974796288 }, { "epoch": 0.2, "learning_rate": 8.111860054565881e-05, "loss": 2.8056, "theoretical_loss": 3.434005203211073, "tokens_seen": 1974927360 }, { "epoch": 0.2, "learning_rate": 8.111057615150056e-05, "loss": 2.5804, "theoretical_loss": 3.4339861543606194, "tokens_seen": 1975058432 }, { "epoch": 0.2, "learning_rate": 8.110255175734233e-05, "loss": 2.4813, "theoretical_loss": 3.4339671071282094, "tokens_seen": 1975189504 }, { "epoch": 0.2, "learning_rate": 8.109452736318408e-05, "loss": 2.5453, "theoretical_loss": 3.4339480615135987, "tokens_seen": 1975320576 }, { "epoch": 0.2, "learning_rate": 8.108650296902585e-05, "loss": 2.618, "theoretical_loss": 3.4339290175165424, "tokens_seen": 1975451648 }, { "epoch": 0.2, "learning_rate": 8.10784785748676e-05, "loss": 2.5069, "theoretical_loss": 3.4339099751367956, "tokens_seen": 1975582720 }, { "epoch": 0.2, "learning_rate": 8.107045418070937e-05, "loss": 2.498, "theoretical_loss": 3.433890934374114, "tokens_seen": 1975713792 }, { "epoch": 0.2, "learning_rate": 8.106242978655112e-05, "loss": 2.5357, "theoretical_loss": 3.4338718952282523, "tokens_seen": 1975844864 }, { "epoch": 0.2, "learning_rate": 8.105440539239287e-05, "loss": 2.4488, "theoretical_loss": 3.4338528576989664, "tokens_seen": 1975975936 }, { "epoch": 0.2, "objective/train/docs_used": 1088406, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.515301465988159, "objective/train/theoretical_loss": 3.433843339540463, "objective/train/tokens_used": 346500576, "theoretical_loss": 3.433843339540463, "tokens_seen": 1976041472 }, { "epoch": 0.2, "learning_rate": 8.104638099823464e-05, "loss": 2.4573, "theoretical_loss": 3.4338338217860116, "tokens_seen": 1976107008 }, { "epoch": 0.2, "learning_rate": 8.103835660407639e-05, "loss": 2.5943, "theoretical_loss": 3.433814787489144, "tokens_seen": 1976238080 }, { "epoch": 0.2, "learning_rate": 8.103033220991816e-05, "loss": 2.6775, "theoretical_loss": 3.433795754808119, "tokens_seen": 1976369152 }, { "epoch": 0.2, "learning_rate": 8.102230781575991e-05, "loss": 2.5248, "theoretical_loss": 3.4337767237426924, "tokens_seen": 1976500224 }, { "epoch": 0.2, "learning_rate": 8.101428342160168e-05, "loss": 2.5347, "theoretical_loss": 3.43375769429262, "tokens_seen": 1976631296 }, { "epoch": 0.2, "learning_rate": 8.100625902744343e-05, "loss": 2.4765, "theoretical_loss": 3.4337386664576566, "tokens_seen": 1976762368 }, { "epoch": 0.2, "learning_rate": 8.099823463328519e-05, "loss": 2.4933, "theoretical_loss": 3.433719640237559, "tokens_seen": 1976893440 }, { "epoch": 0.2, "learning_rate": 8.099021023912695e-05, "loss": 2.6944, "theoretical_loss": 3.433700615632083, "tokens_seen": 1977024512 }, { "epoch": 0.2, "learning_rate": 8.09821858449687e-05, "loss": 2.5308, "theoretical_loss": 3.4336815926409843, "tokens_seen": 1977155584 }, { "epoch": 0.2, "learning_rate": 8.097416145081047e-05, "loss": 2.6257, "theoretical_loss": 3.433662571264019, "tokens_seen": 1977286656 }, { "epoch": 0.2, "learning_rate": 8.096613705665222e-05, "loss": 2.4688, "theoretical_loss": 3.4336435515009436, "tokens_seen": 1977417728 }, { "epoch": 0.2, "learning_rate": 8.095811266249399e-05, "loss": 2.4563, "theoretical_loss": 3.433624533351513, "tokens_seen": 1977548800 }, { "debugging/Self-BLEU-5": 0.5383387038400951, "debugging/distinct-1-grams": 0.7574669480332016, "debugging/distinct-2-grams": 0.9486429063642564, "debugging/entropy-1-grams": 5.78993480235564, "debugging/entropy-2-grams": 6.620072215652616, "debugging/length": 591.3, "debugging/num_segments": 10, "debugging/score": 0.004437156266881241, "debugging/score_std": 0.0035169708188302635, "epoch": 0.2, "objective/train/docs_used": 1088993, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6108832359313965, "objective/train/theoretical_loss": 3.4336055168154846, "objective/train/tokens_used": 348138976, "theoretical_loss": 3.4336055168154846, "tokens_seen": 1977679872 }, { "epoch": 0.2, "learning_rate": 8.095008826833574e-05, "loss": 2.5314, "theoretical_loss": 3.4336055168154846, "tokens_seen": 1977679872 }, { "epoch": 0.2, "learning_rate": 8.09420638741775e-05, "loss": 2.5313, "theoretical_loss": 3.4335865018926146, "tokens_seen": 1977810944 }, { "epoch": 0.2, "learning_rate": 8.093403948001926e-05, "loss": 2.3723, "theoretical_loss": 3.433567488582658, "tokens_seen": 1977942016 }, { "epoch": 0.2, "learning_rate": 8.092601508586102e-05, "loss": 2.4287, "theoretical_loss": 3.433548476885372, "tokens_seen": 1978073088 }, { "epoch": 0.2, "learning_rate": 8.091799069170278e-05, "loss": 2.49, "theoretical_loss": 3.433529466800514, "tokens_seen": 1978204160 }, { "epoch": 0.2, "learning_rate": 8.090996629754454e-05, "loss": 2.5583, "theoretical_loss": 3.433510458327838, "tokens_seen": 1978335232 }, { "epoch": 0.2, "learning_rate": 8.09019419033863e-05, "loss": 2.4996, "theoretical_loss": 3.4334914514671024, "tokens_seen": 1978466304 }, { "epoch": 0.2, "learning_rate": 8.089391750922806e-05, "loss": 2.5626, "theoretical_loss": 3.433472446218063, "tokens_seen": 1978597376 }, { "epoch": 0.2, "learning_rate": 8.088589311506981e-05, "loss": 2.5652, "theoretical_loss": 3.4334534425804764, "tokens_seen": 1978728448 }, { "epoch": 0.2, "learning_rate": 8.087786872091158e-05, "loss": 2.4335, "theoretical_loss": 3.4334344405541, "tokens_seen": 1978859520 }, { "epoch": 0.2, "learning_rate": 8.086984432675333e-05, "loss": 2.495, "theoretical_loss": 3.433415440138689, "tokens_seen": 1978990592 }, { "epoch": 0.2, "learning_rate": 8.08618199325951e-05, "loss": 2.4582, "theoretical_loss": 3.4333964413340015, "tokens_seen": 1979121664 }, { "epoch": 0.2, "learning_rate": 8.085379553843685e-05, "loss": 2.5634, "theoretical_loss": 3.4333774441397935, "tokens_seen": 1979252736 }, { "epoch": 0.2, "objective/train/docs_used": 1090226, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6920125484466553, "objective/train/theoretical_loss": 3.4333679461465434, "objective/train/tokens_used": 349777376, "theoretical_loss": 3.4333679461465434, "tokens_seen": 1979318272 }, { "epoch": 0.2, "learning_rate": 8.084577114427862e-05, "loss": 2.6059, "theoretical_loss": 3.433358448555822, "tokens_seen": 1979383808 }, { "epoch": 0.2, "learning_rate": 8.083774675012037e-05, "loss": 2.5343, "theoretical_loss": 3.4333394545818443, "tokens_seen": 1979514880 }, { "epoch": 0.2, "learning_rate": 8.082972235596214e-05, "loss": 2.588, "theoretical_loss": 3.433320462217617, "tokens_seen": 1979645952 }, { "epoch": 0.2, "learning_rate": 8.082169796180389e-05, "loss": 2.5355, "theoretical_loss": 3.4333014714628973, "tokens_seen": 1979777024 }, { "epoch": 0.2, "learning_rate": 8.081367356764564e-05, "loss": 2.7333, "theoretical_loss": 3.433282482317442, "tokens_seen": 1979908096 }, { "epoch": 0.2, "learning_rate": 8.080564917348741e-05, "loss": 2.4832, "theoretical_loss": 3.4332634947810083, "tokens_seen": 1980039168 }, { "epoch": 0.2, "learning_rate": 8.079762477932916e-05, "loss": 2.5933, "theoretical_loss": 3.4332445088533534, "tokens_seen": 1980170240 }, { "epoch": 0.2, "learning_rate": 8.078960038517093e-05, "loss": 2.5591, "theoretical_loss": 3.433225524534235, "tokens_seen": 1980301312 }, { "epoch": 0.2, "learning_rate": 8.078157599101268e-05, "loss": 2.6487, "theoretical_loss": 3.433206541823409, "tokens_seen": 1980432384 }, { "epoch": 0.2, "learning_rate": 8.077355159685445e-05, "loss": 2.4959, "theoretical_loss": 3.433187560720634, "tokens_seen": 1980563456 }, { "epoch": 0.2, "learning_rate": 8.07655272026962e-05, "loss": 2.6386, "theoretical_loss": 3.433168581225667, "tokens_seen": 1980694528 }, { "epoch": 0.2, "learning_rate": 8.075750280853796e-05, "loss": 2.5598, "theoretical_loss": 3.433149603338266, "tokens_seen": 1980825600 }, { "epoch": 0.2, "objective/train/docs_used": 1091410, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.233128309249878, "objective/train/theoretical_loss": 3.433130627058187, "objective/train/tokens_used": 351415776, "theoretical_loss": 3.433130627058187, "tokens_seen": 1980956672 }, { "epoch": 0.2, "learning_rate": 8.074947841437972e-05, "loss": 2.4389, "theoretical_loss": 3.433130627058187, "tokens_seen": 1980956672 }, { "epoch": 0.2, "learning_rate": 8.074145402022148e-05, "loss": 2.6606, "theoretical_loss": 3.433111652385189, "tokens_seen": 1981087744 }, { "epoch": 0.2, "learning_rate": 8.073342962606324e-05, "loss": 2.614, "theoretical_loss": 3.433092679319029, "tokens_seen": 1981218816 }, { "epoch": 0.2, "learning_rate": 8.0725405231905e-05, "loss": 2.5507, "theoretical_loss": 3.4330737078594638, "tokens_seen": 1981349888 }, { "epoch": 0.2, "learning_rate": 8.071738083774676e-05, "loss": 2.7506, "theoretical_loss": 3.433054738006253, "tokens_seen": 1981480960 }, { "epoch": 0.2, "learning_rate": 8.070935644358851e-05, "loss": 2.5217, "theoretical_loss": 3.4330357697591527, "tokens_seen": 1981612032 }, { "epoch": 0.2, "learning_rate": 8.070133204943027e-05, "loss": 2.5975, "theoretical_loss": 3.4330168031179213, "tokens_seen": 1981743104 }, { "epoch": 0.2, "learning_rate": 8.069330765527203e-05, "loss": 2.5798, "theoretical_loss": 3.4329978380823167, "tokens_seen": 1981874176 }, { "epoch": 0.2, "learning_rate": 8.068528326111379e-05, "loss": 2.5297, "theoretical_loss": 3.432978874652097, "tokens_seen": 1982005248 }, { "epoch": 0.2, "learning_rate": 8.067725886695555e-05, "loss": 2.5863, "theoretical_loss": 3.4329599128270196, "tokens_seen": 1982136320 }, { "epoch": 0.2, "learning_rate": 8.066923447279731e-05, "loss": 2.4973, "theoretical_loss": 3.4329409526068426, "tokens_seen": 1982267392 }, { "epoch": 0.2, "learning_rate": 8.066121007863907e-05, "loss": 2.4644, "theoretical_loss": 3.4329219939913247, "tokens_seen": 1982398464 }, { "epoch": 0.2, "learning_rate": 8.065318568448083e-05, "loss": 2.5549, "theoretical_loss": 3.432903036980223, "tokens_seen": 1982529536 }, { "epoch": 0.2, "objective/train/docs_used": 1091797, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8559112548828125, "objective/train/theoretical_loss": 3.4328935590762533, "objective/train/tokens_used": 353054176, "theoretical_loss": 3.4328935590762533, "tokens_seen": 1982595072 }, { "epoch": 0.2, "learning_rate": 8.064516129032258e-05, "loss": 2.5303, "theoretical_loss": 3.4328840815732966, "tokens_seen": 1982660608 }, { "epoch": 0.2, "learning_rate": 8.063713689616435e-05, "loss": 2.521, "theoretical_loss": 3.4328651277703033, "tokens_seen": 1982791680 }, { "epoch": 0.2, "learning_rate": 8.06291125020061e-05, "loss": 2.5473, "theoretical_loss": 3.432846175571001, "tokens_seen": 1982922752 }, { "epoch": 0.2, "learning_rate": 8.062108810784787e-05, "loss": 2.6932, "theoretical_loss": 3.4328272249751492, "tokens_seen": 1983053824 }, { "epoch": 0.2, "learning_rate": 8.061306371368962e-05, "loss": 2.6695, "theoretical_loss": 3.432808275982505, "tokens_seen": 1983184896 }, { "epoch": 0.2, "learning_rate": 8.060503931953139e-05, "loss": 2.5086, "theoretical_loss": 3.4327893285928273, "tokens_seen": 1983315968 }, { "epoch": 0.2, "learning_rate": 8.059701492537314e-05, "loss": 2.5303, "theoretical_loss": 3.432770382805874, "tokens_seen": 1983447040 }, { "epoch": 0.2, "learning_rate": 8.058899053121489e-05, "loss": 2.5718, "theoretical_loss": 3.432751438621405, "tokens_seen": 1983578112 }, { "epoch": 0.2, "learning_rate": 8.058096613705666e-05, "loss": 2.6883, "theoretical_loss": 3.4327324960391783, "tokens_seen": 1983709184 }, { "epoch": 0.2, "learning_rate": 8.057294174289841e-05, "loss": 2.6956, "theoretical_loss": 3.4327135550589514, "tokens_seen": 1983840256 }, { "epoch": 0.2, "learning_rate": 8.056491734874018e-05, "loss": 2.6438, "theoretical_loss": 3.432694615680485, "tokens_seen": 1983971328 }, { "epoch": 0.2, "learning_rate": 8.055689295458193e-05, "loss": 2.5278, "theoretical_loss": 3.4326756779035357, "tokens_seen": 1984102400 }, { "epoch": 0.2, "objective/train/docs_used": 1092754, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2613365650177, "objective/train/theoretical_loss": 3.4326567417278637, "objective/train/tokens_used": 354692576, "theoretical_loss": 3.4326567417278637, "tokens_seen": 1984233472 }, { "epoch": 0.2, "learning_rate": 8.05488685604237e-05, "loss": 2.54, "theoretical_loss": 3.4326567417278637, "tokens_seen": 1984233472 }, { "epoch": 0.2, "learning_rate": 8.054084416626545e-05, "loss": 2.4222, "theoretical_loss": 3.432637807153228, "tokens_seen": 1984364544 }, { "epoch": 0.2, "learning_rate": 8.053281977210722e-05, "loss": 2.6055, "theoretical_loss": 3.4326188741793864, "tokens_seen": 1984495616 }, { "epoch": 0.2, "learning_rate": 8.052479537794897e-05, "loss": 2.5514, "theoretical_loss": 3.4325999428060987, "tokens_seen": 1984626688 }, { "epoch": 0.2, "learning_rate": 8.051677098379073e-05, "loss": 2.7175, "theoretical_loss": 3.432581013033124, "tokens_seen": 1984757760 }, { "epoch": 0.2, "learning_rate": 8.050874658963249e-05, "loss": 2.4612, "theoretical_loss": 3.43256208486022, "tokens_seen": 1984888832 }, { "epoch": 0.2, "learning_rate": 8.050072219547425e-05, "loss": 2.5532, "theoretical_loss": 3.4325431582871473, "tokens_seen": 1985019904 }, { "epoch": 0.2, "learning_rate": 8.049269780131601e-05, "loss": 2.4733, "theoretical_loss": 3.4325242333136643, "tokens_seen": 1985150976 }, { "epoch": 0.2, "learning_rate": 8.048467340715777e-05, "loss": 2.6379, "theoretical_loss": 3.432505309939531, "tokens_seen": 1985282048 }, { "epoch": 0.2, "learning_rate": 8.047664901299953e-05, "loss": 2.6498, "theoretical_loss": 3.432486388164506, "tokens_seen": 1985413120 }, { "epoch": 0.2, "learning_rate": 8.046862461884128e-05, "loss": 2.5589, "theoretical_loss": 3.432467467988348, "tokens_seen": 1985544192 }, { "epoch": 0.2, "learning_rate": 8.046060022468304e-05, "loss": 2.6433, "theoretical_loss": 3.4324485494108177, "tokens_seen": 1985675264 }, { "epoch": 0.2, "learning_rate": 8.04525758305248e-05, "loss": 2.59, "theoretical_loss": 3.4324296324316736, "tokens_seen": 1985806336 }, { "epoch": 0.2, "objective/train/docs_used": 1093159, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2847602367401123, "objective/train/theoretical_loss": 3.432420174541422, "objective/train/tokens_used": 356330976, "theoretical_loss": 3.432420174541422, "tokens_seen": 1985871872 }, { "epoch": 0.2, "learning_rate": 8.044455143636656e-05, "loss": 2.6555, "theoretical_loss": 3.4324107170506757, "tokens_seen": 1985937408 }, { "epoch": 0.2, "learning_rate": 8.043652704220832e-05, "loss": 2.5619, "theoretical_loss": 3.432391803267583, "tokens_seen": 1986068480 }, { "epoch": 0.2, "learning_rate": 8.042850264805008e-05, "loss": 2.7432, "theoretical_loss": 3.432372891082156, "tokens_seen": 1986199552 }, { "epoch": 0.2, "learning_rate": 8.042047825389184e-05, "loss": 2.6317, "theoretical_loss": 3.4323539804941534, "tokens_seen": 1986330624 }, { "epoch": 0.2, "learning_rate": 8.04124538597336e-05, "loss": 2.5587, "theoretical_loss": 3.432335071503335, "tokens_seen": 1986461696 }, { "epoch": 0.2, "learning_rate": 8.040442946557535e-05, "loss": 2.7005, "theoretical_loss": 3.4323161641094604, "tokens_seen": 1986592768 }, { "epoch": 0.2, "learning_rate": 8.039640507141712e-05, "loss": 2.6133, "theoretical_loss": 3.4322972583122904, "tokens_seen": 1986723840 }, { "epoch": 0.2, "learning_rate": 8.038838067725887e-05, "loss": 2.6584, "theoretical_loss": 3.4322783541115838, "tokens_seen": 1986854912 }, { "epoch": 0.2, "learning_rate": 8.038035628310064e-05, "loss": 2.6685, "theoretical_loss": 3.4322594515071003, "tokens_seen": 1986985984 }, { "epoch": 0.2, "learning_rate": 8.037233188894239e-05, "loss": 2.6374, "theoretical_loss": 3.4322405504986007, "tokens_seen": 1987117056 }, { "epoch": 0.2, "learning_rate": 8.036430749478416e-05, "loss": 2.6503, "theoretical_loss": 3.432221651085845, "tokens_seen": 1987248128 }, { "epoch": 0.2, "learning_rate": 8.035628310062591e-05, "loss": 2.637, "theoretical_loss": 3.4322027532685926, "tokens_seen": 1987379200 }, { "epoch": 0.2, "objective/train/docs_used": 1094053, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.824427366256714, "objective/train/theoretical_loss": 3.4321838570466037, "objective/train/tokens_used": 357969376, "theoretical_loss": 3.4321838570466037, "tokens_seen": 1987510272 }, { "epoch": 0.2, "learning_rate": 8.034825870646766e-05, "loss": 2.658, "theoretical_loss": 3.4321838570466037, "tokens_seen": 1987510272 }, { "epoch": 0.2, "learning_rate": 8.034023431230943e-05, "loss": 2.568, "theoretical_loss": 3.4321649624196384, "tokens_seen": 1987641344 }, { "epoch": 0.2, "learning_rate": 8.033220991815118e-05, "loss": 2.4641, "theoretical_loss": 3.432146069387458, "tokens_seen": 1987772416 }, { "epoch": 0.2, "learning_rate": 8.032418552399295e-05, "loss": 2.5967, "theoretical_loss": 3.432127177949821, "tokens_seen": 1987903488 }, { "epoch": 0.2, "learning_rate": 8.03161611298347e-05, "loss": 2.6815, "theoretical_loss": 3.4321082881064884, "tokens_seen": 1988034560 }, { "epoch": 0.2, "learning_rate": 8.030813673567647e-05, "loss": 2.6253, "theoretical_loss": 3.4320893998572215, "tokens_seen": 1988165632 }, { "epoch": 0.21, "learning_rate": 8.030011234151822e-05, "loss": 2.5624, "theoretical_loss": 3.4320705132017792, "tokens_seen": 1988296704 }, { "epoch": 0.21, "learning_rate": 8.029208794735998e-05, "loss": 2.6203, "theoretical_loss": 3.4320516281399227, "tokens_seen": 1988427776 }, { "epoch": 0.21, "learning_rate": 8.028406355320174e-05, "loss": 2.6069, "theoretical_loss": 3.432032744671413, "tokens_seen": 1988558848 }, { "epoch": 0.21, "learning_rate": 8.02760391590435e-05, "loss": 2.6759, "theoretical_loss": 3.4320138627960097, "tokens_seen": 1988689920 }, { "epoch": 0.21, "learning_rate": 8.026801476488526e-05, "loss": 2.5438, "theoretical_loss": 3.4319949825134737, "tokens_seen": 1988820992 }, { "epoch": 0.21, "learning_rate": 8.025999037072702e-05, "loss": 2.6477, "theoretical_loss": 3.4319761038235663, "tokens_seen": 1988952064 }, { "epoch": 0.21, "learning_rate": 8.025196597656878e-05, "loss": 2.5959, "theoretical_loss": 3.4319572267260474, "tokens_seen": 1989083136 }, { "epoch": 0.21, "objective/train/docs_used": 1094680, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8489184379577637, "objective/train/theoretical_loss": 3.431947788774359, "objective/train/tokens_used": 359607776, "theoretical_loss": 3.431947788774359, "tokens_seen": 1989148672 }, { "epoch": 0.21, "learning_rate": 8.024394158241054e-05, "loss": 2.6793, "theoretical_loss": 3.431938351220678, "tokens_seen": 1989214208 }, { "epoch": 0.21, "learning_rate": 8.023591718825229e-05, "loss": 2.6479, "theoretical_loss": 3.4319194773072192, "tokens_seen": 1989345280 }, { "epoch": 0.21, "learning_rate": 8.022789279409406e-05, "loss": 2.5235, "theoretical_loss": 3.4319006049854313, "tokens_seen": 1989476352 }, { "epoch": 0.21, "learning_rate": 8.021986839993581e-05, "loss": 2.6275, "theoretical_loss": 3.431881734255076, "tokens_seen": 1989607424 }, { "epoch": 0.21, "learning_rate": 8.021184400577757e-05, "loss": 2.6886, "theoretical_loss": 3.4318628651159138, "tokens_seen": 1989738496 }, { "epoch": 0.21, "learning_rate": 8.020381961161933e-05, "loss": 2.5579, "theoretical_loss": 3.4318439975677055, "tokens_seen": 1989869568 }, { "epoch": 0.21, "learning_rate": 8.01957952174611e-05, "loss": 2.5609, "theoretical_loss": 3.4318251316102124, "tokens_seen": 1990000640 }, { "epoch": 0.21, "learning_rate": 8.018777082330285e-05, "loss": 2.5518, "theoretical_loss": 3.4318062672431964, "tokens_seen": 1990131712 }, { "epoch": 0.21, "learning_rate": 8.017974642914461e-05, "loss": 2.7237, "theoretical_loss": 3.431787404466417, "tokens_seen": 1990262784 }, { "epoch": 0.21, "learning_rate": 8.017172203498637e-05, "loss": 2.5871, "theoretical_loss": 3.431768543279637, "tokens_seen": 1990393856 }, { "epoch": 0.21, "learning_rate": 8.016369764082812e-05, "loss": 2.4901, "theoretical_loss": 3.431749683682617, "tokens_seen": 1990524928 }, { "epoch": 0.21, "learning_rate": 8.015567324666989e-05, "loss": 2.6152, "theoretical_loss": 3.4317308256751184, "tokens_seen": 1990656000 }, { "epoch": 0.21, "objective/train/docs_used": 1095652, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7679972648620605, "objective/train/theoretical_loss": 3.4317119692569023, "objective/train/tokens_used": 361246176, "theoretical_loss": 3.4317119692569023, "tokens_seen": 1990787072 }, { "epoch": 0.21, "learning_rate": 8.014764885251164e-05, "loss": 2.5604, "theoretical_loss": 3.4317119692569023, "tokens_seen": 1990787072 }, { "epoch": 0.21, "learning_rate": 8.013962445835341e-05, "loss": 2.7062, "theoretical_loss": 3.431693114427731, "tokens_seen": 1990918144 }, { "epoch": 0.21, "learning_rate": 8.013160006419516e-05, "loss": 2.6594, "theoretical_loss": 3.431674261187365, "tokens_seen": 1991049216 }, { "epoch": 0.21, "learning_rate": 8.012357567003693e-05, "loss": 2.5803, "theoretical_loss": 3.4316554095355665, "tokens_seen": 1991180288 }, { "epoch": 0.21, "learning_rate": 8.011555127587868e-05, "loss": 2.6, "theoretical_loss": 3.4316365594720963, "tokens_seen": 1991311360 }, { "epoch": 0.21, "learning_rate": 8.010752688172043e-05, "loss": 2.5181, "theoretical_loss": 3.431617710996717, "tokens_seen": 1991442432 }, { "epoch": 0.21, "learning_rate": 8.00995024875622e-05, "loss": 2.4289, "theoretical_loss": 3.4315988641091906, "tokens_seen": 1991573504 }, { "epoch": 0.21, "learning_rate": 8.009147809340395e-05, "loss": 2.6387, "theoretical_loss": 3.431580018809277, "tokens_seen": 1991704576 }, { "epoch": 0.21, "learning_rate": 8.008345369924572e-05, "loss": 2.6242, "theoretical_loss": 3.4315611750967396, "tokens_seen": 1991835648 }, { "epoch": 0.21, "learning_rate": 8.007542930508747e-05, "loss": 2.5817, "theoretical_loss": 3.4315423329713397, "tokens_seen": 1991966720 }, { "epoch": 0.21, "learning_rate": 8.006740491092924e-05, "loss": 2.5962, "theoretical_loss": 3.4315234924328393, "tokens_seen": 1992097792 }, { "epoch": 0.21, "learning_rate": 8.005938051677099e-05, "loss": 2.5673, "theoretical_loss": 3.4315046534810003, "tokens_seen": 1992228864 }, { "epoch": 0.21, "learning_rate": 8.005135612261275e-05, "loss": 2.4096, "theoretical_loss": 3.4314858161155852, "tokens_seen": 1992359936 }, { "epoch": 0.21, "objective/train/docs_used": 1096842, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.55733060836792, "objective/train/theoretical_loss": 3.431476398027712, "objective/train/tokens_used": 362884576, "theoretical_loss": 3.431476398027712, "tokens_seen": 1992425472 }, { "epoch": 0.21, "learning_rate": 8.004333172845451e-05, "loss": 2.5752, "theoretical_loss": 3.431466980336355, "tokens_seen": 1992491008 }, { "epoch": 0.21, "learning_rate": 8.003530733429627e-05, "loss": 2.6069, "theoretical_loss": 3.4314481461430724, "tokens_seen": 1992622080 }, { "epoch": 0.21, "learning_rate": 8.002728294013803e-05, "loss": 2.6405, "theoretical_loss": 3.4314293135355003, "tokens_seen": 1992753152 }, { "epoch": 0.21, "learning_rate": 8.001925854597979e-05, "loss": 2.6135, "theoretical_loss": 3.4314104825133995, "tokens_seen": 1992884224 }, { "epoch": 0.21, "learning_rate": 8.001123415182155e-05, "loss": 2.6832, "theoretical_loss": 3.431391653076533, "tokens_seen": 1993015296 }, { "epoch": 0.21, "learning_rate": 8.00032097576633e-05, "loss": 2.5489, "theoretical_loss": 3.4313728252246634, "tokens_seen": 1993146368 }, { "epoch": 0.21, "learning_rate": 7.999518536350506e-05, "loss": 2.6522, "theoretical_loss": 3.431353998957552, "tokens_seen": 1993277440 }, { "epoch": 0.21, "learning_rate": 7.998716096934683e-05, "loss": 2.7246, "theoretical_loss": 3.4313351742749623, "tokens_seen": 1993408512 }, { "epoch": 0.21, "learning_rate": 7.997913657518858e-05, "loss": 2.7097, "theoretical_loss": 3.4313163511766565, "tokens_seen": 1993539584 }, { "epoch": 0.21, "learning_rate": 7.997111218103034e-05, "loss": 2.7062, "theoretical_loss": 3.431297529662397, "tokens_seen": 1993670656 }, { "epoch": 0.21, "learning_rate": 7.99630877868721e-05, "loss": 2.7169, "theoretical_loss": 3.431278709731946, "tokens_seen": 1993801728 }, { "epoch": 0.21, "learning_rate": 7.995506339271386e-05, "loss": 2.5566, "theoretical_loss": 3.431259891385067, "tokens_seen": 1993932800 }, { "epoch": 0.21, "objective/train/docs_used": 1097514, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8202970027923584, "objective/train/theoretical_loss": 3.4312410746215214, "objective/train/tokens_used": 364522976, "theoretical_loss": 3.4312410746215214, "tokens_seen": 1994063872 }, { "epoch": 0.21, "learning_rate": 7.994703899855562e-05, "loss": 2.5633, "theoretical_loss": 3.4312410746215214, "tokens_seen": 1994063872 }, { "epoch": 0.21, "learning_rate": 7.993901460439737e-05, "loss": 2.7012, "theoretical_loss": 3.431222259441073, "tokens_seen": 1994194944 }, { "epoch": 0.21, "learning_rate": 7.993099021023914e-05, "loss": 2.6065, "theoretical_loss": 3.431203445843485, "tokens_seen": 1994326016 }, { "epoch": 0.21, "learning_rate": 7.992296581608089e-05, "loss": 2.6068, "theoretical_loss": 3.4311846338285186, "tokens_seen": 1994457088 }, { "epoch": 0.21, "learning_rate": 7.991494142192266e-05, "loss": 2.527, "theoretical_loss": 3.4311658233959372, "tokens_seen": 1994588160 }, { "epoch": 0.21, "learning_rate": 7.990691702776441e-05, "loss": 2.5969, "theoretical_loss": 3.431147014545505, "tokens_seen": 1994719232 }, { "epoch": 0.21, "learning_rate": 7.989889263360616e-05, "loss": 2.7501, "theoretical_loss": 3.4311282072769833, "tokens_seen": 1994850304 }, { "epoch": 0.21, "learning_rate": 7.989086823944793e-05, "loss": 2.5058, "theoretical_loss": 3.4311094015901364, "tokens_seen": 1994981376 }, { "epoch": 0.21, "learning_rate": 7.988284384528968e-05, "loss": 2.6537, "theoretical_loss": 3.431090597484727, "tokens_seen": 1995112448 }, { "epoch": 0.21, "learning_rate": 7.987481945113145e-05, "loss": 2.6582, "theoretical_loss": 3.431071794960517, "tokens_seen": 1995243520 }, { "epoch": 0.21, "learning_rate": 7.98667950569732e-05, "loss": 2.6121, "theoretical_loss": 3.431052994017272, "tokens_seen": 1995374592 }, { "epoch": 0.21, "learning_rate": 7.985877066281497e-05, "loss": 2.6305, "theoretical_loss": 3.431034194654753, "tokens_seen": 1995505664 }, { "epoch": 0.21, "learning_rate": 7.985074626865672e-05, "loss": 2.5408, "theoretical_loss": 3.4310153968727244, "tokens_seen": 1995636736 }, { "epoch": 0.21, "objective/train/docs_used": 1098702, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.530543804168701, "objective/train/theoretical_loss": 3.43100599857432, "objective/train/tokens_used": 366161376, "theoretical_loss": 3.43100599857432, "tokens_seen": 1995702272 }, { "epoch": 0.21, "learning_rate": 7.984272187449848e-05, "loss": 2.6301, "theoretical_loss": 3.4309966006709494, "tokens_seen": 1995767808 }, { "epoch": 0.21, "learning_rate": 7.983469748034024e-05, "loss": 2.6202, "theoretical_loss": 3.4309778060491913, "tokens_seen": 1995898880 }, { "epoch": 0.21, "learning_rate": 7.9826673086182e-05, "loss": 2.5104, "theoretical_loss": 3.430959013007213, "tokens_seen": 1996029952 }, { "epoch": 0.21, "learning_rate": 7.981864869202376e-05, "loss": 2.7801, "theoretical_loss": 3.4309402215447795, "tokens_seen": 1996161024 }, { "epoch": 0.21, "learning_rate": 7.981062429786552e-05, "loss": 2.6228, "theoretical_loss": 3.4309214316616528, "tokens_seen": 1996292096 }, { "epoch": 0.21, "learning_rate": 7.980259990370727e-05, "loss": 2.6714, "theoretical_loss": 3.4309026433575966, "tokens_seen": 1996423168 }, { "epoch": 0.21, "learning_rate": 7.979457550954904e-05, "loss": 2.7458, "theoretical_loss": 3.4308838566323754, "tokens_seen": 1996554240 }, { "epoch": 0.21, "learning_rate": 7.978655111539079e-05, "loss": 2.6498, "theoretical_loss": 3.4308650714857523, "tokens_seen": 1996685312 }, { "epoch": 0.21, "learning_rate": 7.977852672123256e-05, "loss": 2.5197, "theoretical_loss": 3.4308462879174915, "tokens_seen": 1996816384 }, { "epoch": 0.21, "learning_rate": 7.977050232707431e-05, "loss": 2.5544, "theoretical_loss": 3.4308275059273563, "tokens_seen": 1996947456 }, { "epoch": 0.21, "learning_rate": 7.976247793291608e-05, "loss": 2.6762, "theoretical_loss": 3.430808725515111, "tokens_seen": 1997078528 }, { "epoch": 0.21, "learning_rate": 7.975445353875783e-05, "loss": 2.636, "theoretical_loss": 3.4307899466805187, "tokens_seen": 1997209600 }, { "epoch": 0.21, "objective/train/docs_used": 1099661, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.868697166442871, "objective/train/theoretical_loss": 3.4307711694233443, "objective/train/tokens_used": 367799776, "theoretical_loss": 3.4307711694233443, "tokens_seen": 1997340672 }, { "epoch": 0.21, "learning_rate": 7.974642914459958e-05, "loss": 2.6351, "theoretical_loss": 3.4307711694233443, "tokens_seen": 1997340672 }, { "epoch": 0.21, "learning_rate": 7.973840475044135e-05, "loss": 2.5566, "theoretical_loss": 3.430752393743351, "tokens_seen": 1997471744 }, { "epoch": 0.21, "learning_rate": 7.97303803562831e-05, "loss": 2.6381, "theoretical_loss": 3.4307336196403035, "tokens_seen": 1997602816 }, { "epoch": 0.21, "learning_rate": 7.972235596212487e-05, "loss": 2.7568, "theoretical_loss": 3.4307148471139657, "tokens_seen": 1997733888 }, { "epoch": 0.21, "learning_rate": 7.971433156796662e-05, "loss": 2.5882, "theoretical_loss": 3.4306960761641014, "tokens_seen": 1997864960 }, { "epoch": 0.21, "learning_rate": 7.970630717380837e-05, "loss": 2.578, "theoretical_loss": 3.430677306790475, "tokens_seen": 1997996032 }, { "epoch": 0.21, "learning_rate": 7.969828277965014e-05, "loss": 2.5211, "theoretical_loss": 3.430658538992851, "tokens_seen": 1998127104 }, { "epoch": 0.21, "learning_rate": 7.96902583854919e-05, "loss": 2.6673, "theoretical_loss": 3.4306397727709936, "tokens_seen": 1998258176 }, { "epoch": 0.21, "learning_rate": 7.968223399133366e-05, "loss": 2.5972, "theoretical_loss": 3.4306210081246666, "tokens_seen": 1998389248 }, { "epoch": 0.21, "learning_rate": 7.967420959717541e-05, "loss": 2.7591, "theoretical_loss": 3.4306022450536355, "tokens_seen": 1998520320 }, { "epoch": 0.21, "learning_rate": 7.966618520301718e-05, "loss": 2.6882, "theoretical_loss": 3.4305834835576636, "tokens_seen": 1998651392 }, { "epoch": 0.21, "learning_rate": 7.965816080885893e-05, "loss": 2.7569, "theoretical_loss": 3.4305647236365164, "tokens_seen": 1998782464 }, { "epoch": 0.21, "learning_rate": 7.965013641470069e-05, "loss": 2.6635, "theoretical_loss": 3.4305459652899577, "tokens_seen": 1998913536 }, { "epoch": 0.21, "objective/train/docs_used": 1100646, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.641911745071411, "objective/train/theoretical_loss": 3.4305365867070754, "objective/train/tokens_used": 369438176, "theoretical_loss": 3.4305365867070754, "tokens_seen": 1998979072 }, { "epoch": 0.21, "learning_rate": 7.964211202054245e-05, "loss": 2.6049, "theoretical_loss": 3.430527208517752, "tokens_seen": 1999044608 }, { "epoch": 0.21, "learning_rate": 7.96340876263842e-05, "loss": 2.576, "theoretical_loss": 3.430508453319665, "tokens_seen": 1999175680 }, { "epoch": 0.21, "learning_rate": 7.962606323222597e-05, "loss": 2.6991, "theoretical_loss": 3.4304896996954604, "tokens_seen": 1999306752 }, { "epoch": 0.21, "learning_rate": 7.961803883806773e-05, "loss": 2.5727, "theoretical_loss": 3.430470947644903, "tokens_seen": 1999437824 }, { "epoch": 0.21, "learning_rate": 7.961001444390948e-05, "loss": 2.7093, "theoretical_loss": 3.4304521971677584, "tokens_seen": 1999568896 }, { "epoch": 0.21, "learning_rate": 7.960199004975125e-05, "loss": 2.5668, "theoretical_loss": 3.4304334482637904, "tokens_seen": 1999699968 }, { "epoch": 0.21, "learning_rate": 7.9593965655593e-05, "loss": 2.6501, "theoretical_loss": 3.430414700932765, "tokens_seen": 1999831040 }, { "epoch": 0.21, "learning_rate": 7.958594126143477e-05, "loss": 2.5152, "theoretical_loss": 3.430395955174446, "tokens_seen": 1999962112 }, { "epoch": 0.21, "learning_rate": 7.957791686727652e-05, "loss": 2.6144, "theoretical_loss": 3.4303772109885995, "tokens_seen": 2000093184 }, { "epoch": 0.21, "learning_rate": 7.956989247311829e-05, "loss": 2.5636, "theoretical_loss": 3.43035846837499, "tokens_seen": 2000224256 }, { "epoch": 0.21, "learning_rate": 7.956186807896004e-05, "loss": 2.5419, "theoretical_loss": 3.430339727333383, "tokens_seen": 2000355328 }, { "epoch": 0.21, "learning_rate": 7.955384368480179e-05, "loss": 2.6236, "theoretical_loss": 3.4303209878635426, "tokens_seen": 2000486400 }, { "epoch": 0.21, "objective/train/docs_used": 1101283, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.551567792892456, "objective/train/theoretical_loss": 3.4303022499652354, "objective/train/tokens_used": 371076576, "theoretical_loss": 3.4303022499652354, "tokens_seen": 2000617472 }, { "epoch": 0.21, "learning_rate": 7.954581929064356e-05, "loss": 2.6409, "theoretical_loss": 3.4303022499652354, "tokens_seen": 2000617472 }, { "epoch": 0.21, "learning_rate": 7.953779489648531e-05, "loss": 2.5011, "theoretical_loss": 3.4302835136382255, "tokens_seen": 2000748544 }, { "epoch": 0.21, "learning_rate": 7.952977050232708e-05, "loss": 2.5693, "theoretical_loss": 3.430264778882279, "tokens_seen": 2000879616 }, { "epoch": 0.21, "learning_rate": 7.952174610816883e-05, "loss": 2.6041, "theoretical_loss": 3.430246045697161, "tokens_seen": 2001010688 }, { "epoch": 0.21, "learning_rate": 7.951372171401059e-05, "loss": 2.501, "theoretical_loss": 3.4302273140826367, "tokens_seen": 2001141760 }, { "epoch": 0.21, "learning_rate": 7.950569731985235e-05, "loss": 2.7269, "theoretical_loss": 3.430208584038472, "tokens_seen": 2001272832 }, { "epoch": 0.21, "learning_rate": 7.94976729256941e-05, "loss": 2.5952, "theoretical_loss": 3.430189855564432, "tokens_seen": 2001403904 }, { "epoch": 0.21, "learning_rate": 7.948964853153587e-05, "loss": 2.5573, "theoretical_loss": 3.430171128660283, "tokens_seen": 2001534976 }, { "epoch": 0.21, "learning_rate": 7.948162413737762e-05, "loss": 2.6459, "theoretical_loss": 3.4301524033257893, "tokens_seen": 2001666048 }, { "epoch": 0.21, "learning_rate": 7.947359974321939e-05, "loss": 2.7094, "theoretical_loss": 3.4301336795607176, "tokens_seen": 2001797120 }, { "epoch": 0.21, "learning_rate": 7.946557534906114e-05, "loss": 2.5831, "theoretical_loss": 3.4301149573648333, "tokens_seen": 2001928192 }, { "epoch": 0.21, "learning_rate": 7.94575509549029e-05, "loss": 2.5399, "theoretical_loss": 3.430096236737902, "tokens_seen": 2002059264 }, { "epoch": 0.21, "learning_rate": 7.944952656074466e-05, "loss": 2.5772, "theoretical_loss": 3.43007751767969, "tokens_seen": 2002190336 }, { "epoch": 0.21, "objective/train/docs_used": 1102695, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5117669105529785, "objective/train/theoretical_loss": 3.4300681587387807, "objective/train/tokens_used": 372714976, "theoretical_loss": 3.4300681587387807, "tokens_seen": 2002255872 }, { "epoch": 0.21, "learning_rate": 7.944150216658642e-05, "loss": 2.5223, "theoretical_loss": 3.430058800189963, "tokens_seen": 2002321408 }, { "epoch": 0.21, "learning_rate": 7.943347777242818e-05, "loss": 2.5086, "theoretical_loss": 3.4300400842684864, "tokens_seen": 2002452480 }, { "epoch": 0.21, "learning_rate": 7.942545337826994e-05, "loss": 2.5602, "theoretical_loss": 3.430021369915027, "tokens_seen": 2002583552 }, { "epoch": 0.21, "learning_rate": 7.94174289841117e-05, "loss": 2.6225, "theoretical_loss": 3.43000265712935, "tokens_seen": 2002714624 }, { "epoch": 0.21, "learning_rate": 7.940940458995346e-05, "loss": 2.5564, "theoretical_loss": 3.4299839459112222, "tokens_seen": 2002845696 }, { "epoch": 0.21, "learning_rate": 7.940138019579521e-05, "loss": 2.5846, "theoretical_loss": 3.4299652362604087, "tokens_seen": 2002976768 }, { "epoch": 0.21, "learning_rate": 7.939335580163698e-05, "loss": 2.578, "theoretical_loss": 3.429946528176677, "tokens_seen": 2003107840 }, { "epoch": 0.21, "learning_rate": 7.938533140747873e-05, "loss": 2.5654, "theoretical_loss": 3.429927821659793, "tokens_seen": 2003238912 }, { "epoch": 0.21, "learning_rate": 7.93773070133205e-05, "loss": 2.5886, "theoretical_loss": 3.4299091167095215, "tokens_seen": 2003369984 }, { "epoch": 0.21, "learning_rate": 7.936928261916225e-05, "loss": 2.5983, "theoretical_loss": 3.42989041332563, "tokens_seen": 2003501056 }, { "epoch": 0.21, "learning_rate": 7.936125822500402e-05, "loss": 2.6702, "theoretical_loss": 3.4298717115078854, "tokens_seen": 2003632128 }, { "epoch": 0.21, "learning_rate": 7.935323383084577e-05, "loss": 2.6316, "theoretical_loss": 3.4298530112560535, "tokens_seen": 2003763200 }, { "epoch": 0.21, "objective/train/docs_used": 1103208, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2795019149780273, "objective/train/theoretical_loss": 3.4298343125699002, "objective/train/tokens_used": 374353376, "theoretical_loss": 3.4298343125699002, "tokens_seen": 2003894272 }, { "epoch": 0.21, "learning_rate": 7.934520943668752e-05, "loss": 2.4894, "theoretical_loss": 3.4298343125699002, "tokens_seen": 2003894272 }, { "epoch": 0.21, "learning_rate": 7.933718504252929e-05, "loss": 2.4966, "theoretical_loss": 3.4298156154491934, "tokens_seen": 2004025344 }, { "epoch": 0.21, "learning_rate": 7.932916064837104e-05, "loss": 2.4743, "theoretical_loss": 3.429796919893698, "tokens_seen": 2004156416 }, { "epoch": 0.21, "learning_rate": 7.932113625421281e-05, "loss": 2.5501, "theoretical_loss": 3.429778225903182, "tokens_seen": 2004287488 }, { "epoch": 0.21, "learning_rate": 7.931311186005456e-05, "loss": 2.5934, "theoretical_loss": 3.4297595334774114, "tokens_seen": 2004418560 }, { "epoch": 0.21, "learning_rate": 7.930508746589633e-05, "loss": 2.5649, "theoretical_loss": 3.429740842616153, "tokens_seen": 2004549632 }, { "epoch": 0.21, "learning_rate": 7.929706307173808e-05, "loss": 2.5895, "theoretical_loss": 3.4297221533191737, "tokens_seen": 2004680704 }, { "epoch": 0.22, "learning_rate": 7.928903867757984e-05, "loss": 2.699, "theoretical_loss": 3.4297034655862406, "tokens_seen": 2004811776 }, { "epoch": 0.22, "learning_rate": 7.92810142834216e-05, "loss": 2.6369, "theoretical_loss": 3.4296847794171197, "tokens_seen": 2004942848 }, { "epoch": 0.22, "learning_rate": 7.927298988926336e-05, "loss": 2.5172, "theoretical_loss": 3.4296660948115782, "tokens_seen": 2005073920 }, { "epoch": 0.22, "learning_rate": 7.926496549510512e-05, "loss": 2.4895, "theoretical_loss": 3.429647411769384, "tokens_seen": 2005204992 }, { "epoch": 0.22, "learning_rate": 7.925694110094687e-05, "loss": 2.5339, "theoretical_loss": 3.429628730290303, "tokens_seen": 2005336064 }, { "epoch": 0.22, "learning_rate": 7.924891670678864e-05, "loss": 2.7384, "theoretical_loss": 3.429610050374103, "tokens_seen": 2005467136 }, { "epoch": 0.22, "objective/train/docs_used": 1104357, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6726927757263184, "objective/train/theoretical_loss": 3.4296007110020104, "objective/train/tokens_used": 375991776, "theoretical_loss": 3.4296007110020104, "tokens_seen": 2005532672 }, { "epoch": 0.22, "learning_rate": 7.92408923126304e-05, "loss": 2.5657, "theoretical_loss": 3.4295913720205506, "tokens_seen": 2005598208 }, { "epoch": 0.22, "learning_rate": 7.923286791847216e-05, "loss": 2.6706, "theoretical_loss": 3.429572695229413, "tokens_seen": 2005729280 }, { "epoch": 0.22, "learning_rate": 7.922484352431391e-05, "loss": 2.65, "theoretical_loss": 3.4295540200004577, "tokens_seen": 2005860352 }, { "epoch": 0.22, "learning_rate": 7.921681913015567e-05, "loss": 2.5002, "theoretical_loss": 3.429535346333452, "tokens_seen": 2005991424 }, { "epoch": 0.22, "learning_rate": 7.920879473599743e-05, "loss": 2.581, "theoretical_loss": 3.4295166742281626, "tokens_seen": 2006122496 }, { "epoch": 0.22, "learning_rate": 7.920077034183919e-05, "loss": 2.6358, "theoretical_loss": 3.4294980036843583, "tokens_seen": 2006253568 }, { "epoch": 0.22, "learning_rate": 7.919274594768095e-05, "loss": 2.5882, "theoretical_loss": 3.4294793347018047, "tokens_seen": 2006384640 }, { "epoch": 0.22, "learning_rate": 7.918472155352271e-05, "loss": 2.6364, "theoretical_loss": 3.4294606672802703, "tokens_seen": 2006515712 }, { "epoch": 0.22, "learning_rate": 7.917669715936447e-05, "loss": 2.4089, "theoretical_loss": 3.4294420014195226, "tokens_seen": 2006646784 }, { "epoch": 0.22, "learning_rate": 7.916867276520623e-05, "loss": 2.6042, "theoretical_loss": 3.429423337119329, "tokens_seen": 2006777856 }, { "epoch": 0.22, "learning_rate": 7.916064837104798e-05, "loss": 2.5733, "theoretical_loss": 3.429404674379457, "tokens_seen": 2006908928 }, { "epoch": 0.22, "learning_rate": 7.915262397688975e-05, "loss": 2.5246, "theoretical_loss": 3.4293860131996743, "tokens_seen": 2007040000 }, { "epoch": 0.22, "objective/train/docs_used": 1104987, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.601886749267578, "objective/train/theoretical_loss": 3.4293673535797486, "objective/train/tokens_used": 377630176, "theoretical_loss": 3.4293673535797486, "tokens_seen": 2007171072 }, { "epoch": 0.22, "learning_rate": 7.91445995827315e-05, "loss": 2.6193, "theoretical_loss": 3.4293673535797486, "tokens_seen": 2007171072 }, { "epoch": 0.22, "learning_rate": 7.913657518857327e-05, "loss": 2.4668, "theoretical_loss": 3.4293486955194483, "tokens_seen": 2007302144 }, { "epoch": 0.22, "learning_rate": 7.912855079441502e-05, "loss": 2.5915, "theoretical_loss": 3.42933003901854, "tokens_seen": 2007433216 }, { "epoch": 0.22, "learning_rate": 7.912052640025679e-05, "loss": 2.5267, "theoretical_loss": 3.429311384076793, "tokens_seen": 2007564288 }, { "epoch": 0.22, "learning_rate": 7.911250200609854e-05, "loss": 2.5647, "theoretical_loss": 3.429292730693974, "tokens_seen": 2007695360 }, { "epoch": 0.22, "learning_rate": 7.910447761194029e-05, "loss": 2.6728, "theoretical_loss": 3.4292740788698515, "tokens_seen": 2007826432 }, { "epoch": 0.22, "learning_rate": 7.909645321778206e-05, "loss": 2.4985, "theoretical_loss": 3.4292554286041934, "tokens_seen": 2007957504 }, { "epoch": 0.22, "learning_rate": 7.908842882362381e-05, "loss": 2.5317, "theoretical_loss": 3.4292367798967676, "tokens_seen": 2008088576 }, { "epoch": 0.22, "learning_rate": 7.908040442946558e-05, "loss": 2.5953, "theoretical_loss": 3.4292181327473426, "tokens_seen": 2008219648 }, { "epoch": 0.22, "learning_rate": 7.907238003530733e-05, "loss": 2.493, "theoretical_loss": 3.429199487155686, "tokens_seen": 2008350720 }, { "epoch": 0.22, "learning_rate": 7.90643556411491e-05, "loss": 2.5788, "theoretical_loss": 3.429180843121567, "tokens_seen": 2008481792 }, { "epoch": 0.22, "learning_rate": 7.905633124699085e-05, "loss": 2.4941, "theoretical_loss": 3.429162200644753, "tokens_seen": 2008612864 }, { "epoch": 0.22, "learning_rate": 7.90483068528326e-05, "loss": 2.5688, "theoretical_loss": 3.429143559725013, "tokens_seen": 2008743936 }, { "epoch": 0.22, "objective/train/docs_used": 1106102, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6869187355041504, "objective/train/theoretical_loss": 3.429134239848973, "objective/train/tokens_used": 379268576, "theoretical_loss": 3.429134239848973, "tokens_seen": 2008809472 }, { "epoch": 0.22, "learning_rate": 7.904028245867437e-05, "loss": 2.5055, "theoretical_loss": 3.4291249203621144, "tokens_seen": 2008875008 }, { "epoch": 0.22, "learning_rate": 7.903225806451613e-05, "loss": 2.5025, "theoretical_loss": 3.429106282555826, "tokens_seen": 2009006080 }, { "epoch": 0.22, "learning_rate": 7.902423367035789e-05, "loss": 2.3999, "theoretical_loss": 3.429087646305917, "tokens_seen": 2009137152 }, { "epoch": 0.22, "learning_rate": 7.901620927619964e-05, "loss": 2.5506, "theoretical_loss": 3.429069011612155, "tokens_seen": 2009268224 }, { "epoch": 0.22, "learning_rate": 7.900818488204141e-05, "loss": 2.51, "theoretical_loss": 3.4290503784743085, "tokens_seen": 2009399296 }, { "epoch": 0.22, "learning_rate": 7.900016048788316e-05, "loss": 2.5117, "theoretical_loss": 3.429031746892147, "tokens_seen": 2009530368 }, { "epoch": 0.22, "learning_rate": 7.899213609372492e-05, "loss": 2.6431, "theoretical_loss": 3.4290131168654385, "tokens_seen": 2009661440 }, { "epoch": 0.22, "learning_rate": 7.898411169956668e-05, "loss": 2.5644, "theoretical_loss": 3.4289944883939514, "tokens_seen": 2009792512 }, { "epoch": 0.22, "learning_rate": 7.897608730540844e-05, "loss": 2.5707, "theoretical_loss": 3.428975861477455, "tokens_seen": 2009923584 }, { "epoch": 0.22, "learning_rate": 7.89680629112502e-05, "loss": 2.5151, "theoretical_loss": 3.4289572361157186, "tokens_seen": 2010054656 }, { "epoch": 0.22, "learning_rate": 7.896003851709196e-05, "loss": 2.5642, "theoretical_loss": 3.42893861230851, "tokens_seen": 2010185728 }, { "epoch": 0.22, "learning_rate": 7.895201412293372e-05, "loss": 2.6082, "theoretical_loss": 3.428919990055599, "tokens_seen": 2010316800 }, { "epoch": 0.22, "objective/train/docs_used": 1107415, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5254108905792236, "objective/train/theoretical_loss": 3.428901369356754, "objective/train/tokens_used": 380906976, "theoretical_loss": 3.428901369356754, "tokens_seen": 2010447872 }, { "epoch": 0.22, "learning_rate": 7.894398972877548e-05, "loss": 2.5814, "theoretical_loss": 3.428901369356754, "tokens_seen": 2010447872 }, { "epoch": 0.22, "learning_rate": 7.893596533461724e-05, "loss": 2.6239, "theoretical_loss": 3.4288827502117436, "tokens_seen": 2010578944 }, { "epoch": 0.22, "learning_rate": 7.8927940940459e-05, "loss": 2.671, "theoretical_loss": 3.4288641326203377, "tokens_seen": 2010710016 }, { "epoch": 0.22, "learning_rate": 7.891991654630075e-05, "loss": 2.6009, "theoretical_loss": 3.428845516582305, "tokens_seen": 2010841088 }, { "epoch": 0.22, "learning_rate": 7.891189215214252e-05, "loss": 2.6173, "theoretical_loss": 3.428826902097415, "tokens_seen": 2010972160 }, { "epoch": 0.22, "learning_rate": 7.890386775798427e-05, "loss": 2.5199, "theoretical_loss": 3.4288082891654366, "tokens_seen": 2011103232 }, { "epoch": 0.22, "learning_rate": 7.889584336382604e-05, "loss": 2.5374, "theoretical_loss": 3.4287896777861384, "tokens_seen": 2011234304 }, { "epoch": 0.22, "learning_rate": 7.888781896966779e-05, "loss": 2.5707, "theoretical_loss": 3.4287710679592913, "tokens_seen": 2011365376 }, { "epoch": 0.22, "learning_rate": 7.887979457550956e-05, "loss": 2.4322, "theoretical_loss": 3.4287524596846635, "tokens_seen": 2011496448 }, { "epoch": 0.22, "learning_rate": 7.887177018135131e-05, "loss": 2.4975, "theoretical_loss": 3.4287338529620244, "tokens_seen": 2011627520 }, { "epoch": 0.22, "learning_rate": 7.886374578719306e-05, "loss": 2.6787, "theoretical_loss": 3.4287152477911436, "tokens_seen": 2011758592 }, { "epoch": 0.22, "learning_rate": 7.885572139303483e-05, "loss": 2.4772, "theoretical_loss": 3.428696644171791, "tokens_seen": 2011889664 }, { "epoch": 0.22, "learning_rate": 7.884769699887658e-05, "loss": 2.6306, "theoretical_loss": 3.428678042103736, "tokens_seen": 2012020736 }, { "epoch": 0.22, "objective/train/docs_used": 1108037, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8532683849334717, "objective/train/theoretical_loss": 3.428668741651373, "objective/train/tokens_used": 382545376, "theoretical_loss": 3.428668741651373, "tokens_seen": 2012086272 }, { "epoch": 0.22, "learning_rate": 7.883967260471835e-05, "loss": 2.5644, "theoretical_loss": 3.4286594415867477, "tokens_seen": 2012151808 }, { "epoch": 0.22, "learning_rate": 7.88316482105601e-05, "loss": 2.5605, "theoretical_loss": 3.428640842620596, "tokens_seen": 2012282880 }, { "epoch": 0.22, "learning_rate": 7.882362381640187e-05, "loss": 2.5797, "theoretical_loss": 3.4286222452050508, "tokens_seen": 2012413952 }, { "epoch": 0.22, "learning_rate": 7.881559942224362e-05, "loss": 2.5207, "theoretical_loss": 3.428603649339882, "tokens_seen": 2012545024 }, { "epoch": 0.22, "learning_rate": 7.880757502808538e-05, "loss": 2.5635, "theoretical_loss": 3.4285850550248584, "tokens_seen": 2012676096 }, { "epoch": 0.22, "learning_rate": 7.879955063392714e-05, "loss": 2.7104, "theoretical_loss": 3.4285664622597514, "tokens_seen": 2012807168 }, { "epoch": 0.22, "learning_rate": 7.87915262397689e-05, "loss": 2.5008, "theoretical_loss": 3.4285478710443296, "tokens_seen": 2012938240 }, { "epoch": 0.22, "learning_rate": 7.878350184561066e-05, "loss": 2.5471, "theoretical_loss": 3.4285292813783634, "tokens_seen": 2013069312 }, { "epoch": 0.22, "learning_rate": 7.877547745145242e-05, "loss": 2.4691, "theoretical_loss": 3.4285106932616225, "tokens_seen": 2013200384 }, { "epoch": 0.22, "learning_rate": 7.876745305729418e-05, "loss": 2.6136, "theoretical_loss": 3.428492106693878, "tokens_seen": 2013331456 }, { "epoch": 0.22, "learning_rate": 7.875942866313593e-05, "loss": 2.4363, "theoretical_loss": 3.4284735216748983, "tokens_seen": 2013462528 }, { "epoch": 0.22, "learning_rate": 7.875140426897769e-05, "loss": 2.4772, "theoretical_loss": 3.428454938204455, "tokens_seen": 2013593600 }, { "epoch": 0.22, "objective/train/docs_used": 1109293, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.913437604904175, "objective/train/theoretical_loss": 3.4284363562823175, "objective/train/tokens_used": 384183776, "theoretical_loss": 3.4284363562823175, "tokens_seen": 2013724672 }, { "epoch": 0.22, "learning_rate": 7.874337987481945e-05, "loss": 2.6037, "theoretical_loss": 3.4284363562823175, "tokens_seen": 2013724672 }, { "epoch": 0.22, "learning_rate": 7.873535548066121e-05, "loss": 2.6943, "theoretical_loss": 3.4284177759082564, "tokens_seen": 2013855744 }, { "epoch": 0.22, "learning_rate": 7.872733108650297e-05, "loss": 2.5348, "theoretical_loss": 3.428399197082042, "tokens_seen": 2013986816 }, { "epoch": 0.22, "learning_rate": 7.871930669234473e-05, "loss": 2.5262, "theoretical_loss": 3.428380619803444, "tokens_seen": 2014117888 }, { "epoch": 0.22, "learning_rate": 7.87112822981865e-05, "loss": 2.6206, "theoretical_loss": 3.4283620440722333, "tokens_seen": 2014248960 }, { "epoch": 0.22, "learning_rate": 7.870325790402825e-05, "loss": 2.5586, "theoretical_loss": 3.4283434698881807, "tokens_seen": 2014380032 }, { "epoch": 0.22, "learning_rate": 7.869523350987e-05, "loss": 2.6104, "theoretical_loss": 3.4283248972510556, "tokens_seen": 2014511104 }, { "epoch": 0.22, "learning_rate": 7.868720911571177e-05, "loss": 2.5206, "theoretical_loss": 3.42830632616063, "tokens_seen": 2014642176 }, { "epoch": 0.22, "learning_rate": 7.867918472155352e-05, "loss": 2.5558, "theoretical_loss": 3.4282877566166734, "tokens_seen": 2014773248 }, { "epoch": 0.22, "learning_rate": 7.867116032739529e-05, "loss": 2.5839, "theoretical_loss": 3.4282691886189562, "tokens_seen": 2014904320 }, { "epoch": 0.22, "learning_rate": 7.866313593323704e-05, "loss": 2.5297, "theoretical_loss": 3.4282506221672504, "tokens_seen": 2015035392 }, { "epoch": 0.22, "learning_rate": 7.865511153907881e-05, "loss": 2.3923, "theoretical_loss": 3.4282320572613254, "tokens_seen": 2015166464 }, { "epoch": 0.22, "learning_rate": 7.864708714492056e-05, "loss": 2.5233, "theoretical_loss": 3.4282134939009525, "tokens_seen": 2015297536 }, { "epoch": 0.22, "objective/train/docs_used": 1109931, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3523659706115723, "objective/train/theoretical_loss": 3.4282042128002765, "objective/train/tokens_used": 385822176, "theoretical_loss": 3.4282042128002765, "tokens_seen": 2015363072 }, { "epoch": 0.22, "learning_rate": 7.863906275076233e-05, "loss": 2.4561, "theoretical_loss": 3.4281949320859026, "tokens_seen": 2015428608 }, { "epoch": 0.22, "learning_rate": 7.863103835660408e-05, "loss": 2.6495, "theoretical_loss": 3.4281763718159466, "tokens_seen": 2015559680 }, { "epoch": 0.22, "learning_rate": 7.862301396244583e-05, "loss": 2.6334, "theoretical_loss": 3.4281578130908548, "tokens_seen": 2015690752 }, { "epoch": 0.22, "learning_rate": 7.86149895682876e-05, "loss": 2.5526, "theoretical_loss": 3.4281392559103994, "tokens_seen": 2015821824 }, { "epoch": 0.22, "learning_rate": 7.860696517412935e-05, "loss": 2.5854, "theoretical_loss": 3.4281207002743503, "tokens_seen": 2015952896 }, { "epoch": 0.22, "learning_rate": 7.859894077997112e-05, "loss": 2.5802, "theoretical_loss": 3.428102146182479, "tokens_seen": 2016083968 }, { "epoch": 0.22, "learning_rate": 7.859091638581287e-05, "loss": 2.4863, "theoretical_loss": 3.4280835936345566, "tokens_seen": 2016215040 }, { "epoch": 0.22, "learning_rate": 7.858289199165464e-05, "loss": 2.5008, "theoretical_loss": 3.4280650426303536, "tokens_seen": 2016346112 }, { "epoch": 0.22, "learning_rate": 7.857486759749639e-05, "loss": 2.5014, "theoretical_loss": 3.428046493169643, "tokens_seen": 2016477184 }, { "epoch": 0.22, "learning_rate": 7.856684320333815e-05, "loss": 2.6355, "theoretical_loss": 3.428027945252194, "tokens_seen": 2016608256 }, { "epoch": 0.22, "learning_rate": 7.855881880917991e-05, "loss": 2.4802, "theoretical_loss": 3.428009398877779, "tokens_seen": 2016739328 }, { "epoch": 0.22, "learning_rate": 7.855079441502167e-05, "loss": 2.4685, "theoretical_loss": 3.427990854046169, "tokens_seen": 2016870400 }, { "epoch": 0.22, "objective/train/docs_used": 1110878, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.68257212638855, "objective/train/theoretical_loss": 3.427972310757136, "objective/train/tokens_used": 387460576, "theoretical_loss": 3.427972310757136, "tokens_seen": 2017001472 }, { "epoch": 0.22, "learning_rate": 7.854277002086343e-05, "loss": 2.5736, "theoretical_loss": 3.427972310757136, "tokens_seen": 2017001472 }, { "epoch": 0.22, "learning_rate": 7.853474562670519e-05, "loss": 2.4837, "theoretical_loss": 3.4279537690104513, "tokens_seen": 2017132544 }, { "epoch": 0.22, "learning_rate": 7.852672123254695e-05, "loss": 2.5344, "theoretical_loss": 3.427935228805886, "tokens_seen": 2017263616 }, { "epoch": 0.22, "learning_rate": 7.85186968383887e-05, "loss": 2.5714, "theoretical_loss": 3.4279166901432117, "tokens_seen": 2017394688 }, { "epoch": 0.22, "learning_rate": 7.851067244423046e-05, "loss": 2.5783, "theoretical_loss": 3.4278981530221997, "tokens_seen": 2017525760 }, { "epoch": 0.22, "learning_rate": 7.850264805007222e-05, "loss": 2.4, "theoretical_loss": 3.4278796174426223, "tokens_seen": 2017656832 }, { "epoch": 0.22, "learning_rate": 7.849462365591398e-05, "loss": 2.5503, "theoretical_loss": 3.427861083404251, "tokens_seen": 2017787904 }, { "epoch": 0.22, "learning_rate": 7.848659926175574e-05, "loss": 2.5524, "theoretical_loss": 3.4278425509068575, "tokens_seen": 2017918976 }, { "epoch": 0.22, "learning_rate": 7.84785748675975e-05, "loss": 2.5015, "theoretical_loss": 3.4278240199502137, "tokens_seen": 2018050048 }, { "epoch": 0.22, "learning_rate": 7.847055047343926e-05, "loss": 2.567, "theoretical_loss": 3.4278054905340913, "tokens_seen": 2018181120 }, { "epoch": 0.22, "learning_rate": 7.846252607928102e-05, "loss": 2.4222, "theoretical_loss": 3.427786962658262, "tokens_seen": 2018312192 }, { "epoch": 0.22, "learning_rate": 7.845450168512277e-05, "loss": 2.6226, "theoretical_loss": 3.4277684363224985, "tokens_seen": 2018443264 }, { "epoch": 0.22, "learning_rate": 7.844647729096454e-05, "loss": 2.3817, "theoretical_loss": 3.427749911526572, "tokens_seen": 2018574336 }, { "epoch": 0.22, "objective/train/docs_used": 1111506, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.24393630027771, "objective/train/theoretical_loss": 3.427740649705976, "objective/train/tokens_used": 389098976, "theoretical_loss": 3.427740649705976, "tokens_seen": 2018639872 }, { "epoch": 0.22, "learning_rate": 7.843845289680629e-05, "loss": 2.5344, "theoretical_loss": 3.4277313882702547, "tokens_seen": 2018705408 }, { "epoch": 0.22, "learning_rate": 7.843042850264806e-05, "loss": 2.5531, "theoretical_loss": 3.4277128665533194, "tokens_seen": 2018836480 }, { "epoch": 0.22, "learning_rate": 7.842240410848981e-05, "loss": 2.4688, "theoretical_loss": 3.427694346375537, "tokens_seen": 2018967552 }, { "epoch": 0.22, "learning_rate": 7.841437971433158e-05, "loss": 2.4856, "theoretical_loss": 3.4276758277366803, "tokens_seen": 2019098624 }, { "epoch": 0.22, "learning_rate": 7.840635532017333e-05, "loss": 2.441, "theoretical_loss": 3.427657310636522, "tokens_seen": 2019229696 }, { "epoch": 0.22, "learning_rate": 7.839833092601508e-05, "loss": 2.566, "theoretical_loss": 3.4276387950748335, "tokens_seen": 2019360768 }, { "epoch": 0.22, "learning_rate": 7.839030653185685e-05, "loss": 2.4114, "theoretical_loss": 3.427620281051388, "tokens_seen": 2019491840 }, { "epoch": 0.22, "learning_rate": 7.83822821376986e-05, "loss": 2.4363, "theoretical_loss": 3.4276017685659577, "tokens_seen": 2019622912 }, { "epoch": 0.22, "learning_rate": 7.837425774354037e-05, "loss": 2.4333, "theoretical_loss": 3.4275832576183145, "tokens_seen": 2019753984 }, { "epoch": 0.22, "learning_rate": 7.836623334938212e-05, "loss": 2.5378, "theoretical_loss": 3.4275647482082308, "tokens_seen": 2019885056 }, { "epoch": 0.22, "learning_rate": 7.835820895522389e-05, "loss": 2.5506, "theoretical_loss": 3.42754624033548, "tokens_seen": 2020016128 }, { "epoch": 0.22, "learning_rate": 7.835018456106564e-05, "loss": 2.5029, "theoretical_loss": 3.427527733999834, "tokens_seen": 2020147200 }, { "epoch": 0.22, "objective/train/docs_used": 1112541, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.220839262008667, "objective/train/theoretical_loss": 3.427509229201066, "objective/train/tokens_used": 390737376, "theoretical_loss": 3.427509229201066, "tokens_seen": 2020278272 }, { "epoch": 0.22, "learning_rate": 7.83421601669074e-05, "loss": 2.5173, "theoretical_loss": 3.427509229201066, "tokens_seen": 2020278272 }, { "epoch": 0.22, "learning_rate": 7.833413577274916e-05, "loss": 2.5369, "theoretical_loss": 3.4274907259389478, "tokens_seen": 2020409344 }, { "epoch": 0.22, "learning_rate": 7.832611137859092e-05, "loss": 2.5579, "theoretical_loss": 3.4274722242132523, "tokens_seen": 2020540416 }, { "epoch": 0.22, "learning_rate": 7.831808698443268e-05, "loss": 2.5063, "theoretical_loss": 3.427453724023753, "tokens_seen": 2020671488 }, { "epoch": 0.22, "learning_rate": 7.831006259027444e-05, "loss": 2.6443, "theoretical_loss": 3.427435225370222, "tokens_seen": 2020802560 }, { "epoch": 0.22, "learning_rate": 7.83020381961162e-05, "loss": 2.6026, "theoretical_loss": 3.427416728252433, "tokens_seen": 2020933632 }, { "epoch": 0.22, "learning_rate": 7.829401380195796e-05, "loss": 2.4388, "theoretical_loss": 3.427398232670158, "tokens_seen": 2021064704 }, { "epoch": 0.22, "learning_rate": 7.828598940779972e-05, "loss": 2.6021, "theoretical_loss": 3.4273797386231704, "tokens_seen": 2021195776 }, { "epoch": 0.23, "learning_rate": 7.827796501364148e-05, "loss": 2.5579, "theoretical_loss": 3.427361246111243, "tokens_seen": 2021326848 }, { "epoch": 0.23, "learning_rate": 7.826994061948323e-05, "loss": 2.4902, "theoretical_loss": 3.4273427551341493, "tokens_seen": 2021457920 }, { "epoch": 0.23, "learning_rate": 7.8261916225325e-05, "loss": 2.552, "theoretical_loss": 3.4273242656916616, "tokens_seen": 2021588992 }, { "epoch": 0.23, "learning_rate": 7.825389183116675e-05, "loss": 2.5983, "theoretical_loss": 3.4273057777835536, "tokens_seen": 2021720064 }, { "epoch": 0.23, "learning_rate": 7.824586743700851e-05, "loss": 2.4921, "theoretical_loss": 3.427287291409599, "tokens_seen": 2021851136 }, { "epoch": 0.23, "objective/train/docs_used": 1113127, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.679295539855957, "objective/train/theoretical_loss": 3.4272780487978576, "objective/train/tokens_used": 392375776, "theoretical_loss": 3.4272780487978576, "tokens_seen": 2021916672 }, { "epoch": 0.23, "learning_rate": 7.823784304285027e-05, "loss": 2.5579, "theoretical_loss": 3.42726880656957, "tokens_seen": 2021982208 }, { "epoch": 0.23, "learning_rate": 7.822981864869203e-05, "loss": 2.3652, "theoretical_loss": 3.42725032326324, "tokens_seen": 2022113280 }, { "epoch": 0.23, "learning_rate": 7.822179425453379e-05, "loss": 2.7038, "theoretical_loss": 3.427231841490384, "tokens_seen": 2022244352 }, { "epoch": 0.23, "learning_rate": 7.821376986037554e-05, "loss": 2.4067, "theoretical_loss": 3.4272133612507734, "tokens_seen": 2022375424 }, { "epoch": 0.23, "learning_rate": 7.820574546621731e-05, "loss": 2.5086, "theoretical_loss": 3.427194882544182, "tokens_seen": 2022506496 }, { "epoch": 0.23, "learning_rate": 7.819772107205906e-05, "loss": 2.5402, "theoretical_loss": 3.4271764053703837, "tokens_seen": 2022637568 }, { "epoch": 0.23, "learning_rate": 7.818969667790083e-05, "loss": 2.4729, "theoretical_loss": 3.4271579297291526, "tokens_seen": 2022768640 }, { "epoch": 0.23, "learning_rate": 7.818167228374258e-05, "loss": 2.4496, "theoretical_loss": 3.4271394556202615, "tokens_seen": 2022899712 }, { "epoch": 0.23, "learning_rate": 7.817364788958435e-05, "loss": 2.3272, "theoretical_loss": 3.4271209830434843, "tokens_seen": 2023030784 }, { "epoch": 0.23, "learning_rate": 7.81656234954261e-05, "loss": 2.6453, "theoretical_loss": 3.4271025119985943, "tokens_seen": 2023161856 }, { "epoch": 0.23, "learning_rate": 7.815759910126785e-05, "loss": 2.5107, "theoretical_loss": 3.4270840424853657, "tokens_seen": 2023292928 }, { "epoch": 0.23, "learning_rate": 7.814957470710962e-05, "loss": 2.5706, "theoretical_loss": 3.427065574503572, "tokens_seen": 2023424000 }, { "epoch": 0.23, "objective/train/docs_used": 1114473, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.07741641998291, "objective/train/theoretical_loss": 3.427047108052988, "objective/train/tokens_used": 394014176, "theoretical_loss": 3.427047108052988, "tokens_seen": 2023555072 }, { "epoch": 0.23, "learning_rate": 7.814155031295137e-05, "loss": 2.5202, "theoretical_loss": 3.427047108052988, "tokens_seen": 2023555072 }, { "epoch": 0.23, "learning_rate": 7.813352591879314e-05, "loss": 2.4878, "theoretical_loss": 3.4270286431333856, "tokens_seen": 2023686144 }, { "epoch": 0.23, "learning_rate": 7.812550152463489e-05, "loss": 2.4707, "theoretical_loss": 3.4270101797445403, "tokens_seen": 2023817216 }, { "epoch": 0.23, "learning_rate": 7.811747713047666e-05, "loss": 2.4816, "theoretical_loss": 3.4269917178862253, "tokens_seen": 2023948288 }, { "epoch": 0.23, "learning_rate": 7.810945273631841e-05, "loss": 2.5005, "theoretical_loss": 3.426973257558216, "tokens_seen": 2024079360 }, { "epoch": 0.23, "learning_rate": 7.810142834216017e-05, "loss": 2.58, "theoretical_loss": 3.4269547987602844, "tokens_seen": 2024210432 }, { "epoch": 0.23, "learning_rate": 7.809340394800193e-05, "loss": 2.5284, "theoretical_loss": 3.4269363414922056, "tokens_seen": 2024341504 }, { "epoch": 0.23, "learning_rate": 7.808537955384369e-05, "loss": 2.5707, "theoretical_loss": 3.4269178857537543, "tokens_seen": 2024472576 }, { "epoch": 0.23, "learning_rate": 7.807735515968545e-05, "loss": 2.5637, "theoretical_loss": 3.426899431544704, "tokens_seen": 2024603648 }, { "epoch": 0.23, "learning_rate": 7.80693307655272e-05, "loss": 2.5071, "theoretical_loss": 3.4268809788648293, "tokens_seen": 2024734720 }, { "epoch": 0.23, "learning_rate": 7.806130637136897e-05, "loss": 2.4704, "theoretical_loss": 3.4268625277139044, "tokens_seen": 2024865792 }, { "epoch": 0.23, "learning_rate": 7.805328197721073e-05, "loss": 2.5377, "theoretical_loss": 3.4268440780917038, "tokens_seen": 2024996864 }, { "epoch": 0.23, "learning_rate": 7.804525758305248e-05, "loss": 2.5348, "theoretical_loss": 3.4268256299980013, "tokens_seen": 2025127936 }, { "epoch": 0.23, "objective/train/docs_used": 1115722, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8050100803375244, "objective/train/theoretical_loss": 3.4268164065242663, "objective/train/tokens_used": 395652576, "theoretical_loss": 3.4268164065242663, "tokens_seen": 2025193472 }, { "epoch": 0.23, "learning_rate": 7.803723318889425e-05, "loss": 2.5304, "theoretical_loss": 3.4268071834325715, "tokens_seen": 2025259008 }, { "epoch": 0.23, "learning_rate": 7.8029208794736e-05, "loss": 2.4983, "theoretical_loss": 3.42678873839519, "tokens_seen": 2025390080 }, { "epoch": 0.23, "learning_rate": 7.802118440057776e-05, "loss": 2.4642, "theoretical_loss": 3.42677029488563, "tokens_seen": 2025521152 }, { "epoch": 0.23, "learning_rate": 7.801316000641952e-05, "loss": 2.5015, "theoretical_loss": 3.426751852903667, "tokens_seen": 2025652224 }, { "epoch": 0.23, "learning_rate": 7.800513561226128e-05, "loss": 2.3507, "theoretical_loss": 3.4267334124490754, "tokens_seen": 2025783296 }, { "epoch": 0.23, "learning_rate": 7.799711121810304e-05, "loss": 2.6905, "theoretical_loss": 3.426714973521629, "tokens_seen": 2025914368 }, { "epoch": 0.23, "learning_rate": 7.79890868239448e-05, "loss": 2.538, "theoretical_loss": 3.426696536121104, "tokens_seen": 2026045440 }, { "epoch": 0.23, "learning_rate": 7.798106242978656e-05, "loss": 2.598, "theoretical_loss": 3.4266781002472744, "tokens_seen": 2026176512 }, { "epoch": 0.23, "learning_rate": 7.797303803562831e-05, "loss": 2.4248, "theoretical_loss": 3.4266596658999147, "tokens_seen": 2026307584 }, { "epoch": 0.23, "learning_rate": 7.796501364147008e-05, "loss": 2.5141, "theoretical_loss": 3.4266412330788008, "tokens_seen": 2026438656 }, { "epoch": 0.23, "learning_rate": 7.795698924731183e-05, "loss": 2.3992, "theoretical_loss": 3.426622801783707, "tokens_seen": 2026569728 }, { "epoch": 0.23, "learning_rate": 7.79489648531536e-05, "loss": 2.6258, "theoretical_loss": 3.426604372014408, "tokens_seen": 2026700800 }, { "epoch": 0.23, "objective/train/docs_used": 1116448, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.421637535095215, "objective/train/theoretical_loss": 3.426585943770679, "objective/train/tokens_used": 397290976, "theoretical_loss": 3.426585943770679, "tokens_seen": 2026831872 }, { "epoch": 0.23, "learning_rate": 7.794094045899535e-05, "loss": 2.5952, "theoretical_loss": 3.426585943770679, "tokens_seen": 2026831872 }, { "epoch": 0.23, "learning_rate": 7.793291606483712e-05, "loss": 2.3854, "theoretical_loss": 3.4265675170522956, "tokens_seen": 2026962944 }, { "epoch": 0.23, "learning_rate": 7.792489167067887e-05, "loss": 2.4204, "theoretical_loss": 3.426549091859032, "tokens_seen": 2027094016 }, { "epoch": 0.23, "learning_rate": 7.791686727652062e-05, "loss": 2.4967, "theoretical_loss": 3.4265306681906647, "tokens_seen": 2027225088 }, { "epoch": 0.23, "learning_rate": 7.790884288236239e-05, "loss": 2.5815, "theoretical_loss": 3.4265122460469675, "tokens_seen": 2027356160 }, { "epoch": 0.23, "learning_rate": 7.790081848820414e-05, "loss": 2.5191, "theoretical_loss": 3.4264938254277166, "tokens_seen": 2027487232 }, { "epoch": 0.23, "learning_rate": 7.789279409404591e-05, "loss": 2.5302, "theoretical_loss": 3.4264754063326865, "tokens_seen": 2027618304 }, { "epoch": 0.23, "learning_rate": 7.788476969988766e-05, "loss": 2.4264, "theoretical_loss": 3.426456988761654, "tokens_seen": 2027749376 }, { "epoch": 0.23, "learning_rate": 7.787674530572943e-05, "loss": 2.5857, "theoretical_loss": 3.426438572714393, "tokens_seen": 2027880448 }, { "epoch": 0.23, "learning_rate": 7.786872091157118e-05, "loss": 2.2381, "theoretical_loss": 3.4264201581906795, "tokens_seen": 2028011520 }, { "epoch": 0.23, "learning_rate": 7.786069651741294e-05, "loss": 2.4836, "theoretical_loss": 3.4264017451902893, "tokens_seen": 2028142592 }, { "epoch": 0.23, "learning_rate": 7.78526721232547e-05, "loss": 2.5449, "theoretical_loss": 3.426383333712997, "tokens_seen": 2028273664 }, { "epoch": 0.23, "learning_rate": 7.784464772909646e-05, "loss": 2.5556, "theoretical_loss": 3.4263649237585794, "tokens_seen": 2028404736 }, { "epoch": 0.23, "objective/train/docs_used": 1117705, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.295840263366699, "objective/train/theoretical_loss": 3.426355719352378, "objective/train/tokens_used": 398929376, "theoretical_loss": 3.426355719352378, "tokens_seen": 2028470272 }, { "epoch": 0.23, "learning_rate": 7.783662333493822e-05, "loss": 2.5011, "theoretical_loss": 3.4263465153268116, "tokens_seen": 2028535808 }, { "epoch": 0.23, "learning_rate": 7.782859894077998e-05, "loss": 2.512, "theoretical_loss": 3.4263281084174695, "tokens_seen": 2028666880 }, { "epoch": 0.23, "learning_rate": 7.782057454662174e-05, "loss": 2.5264, "theoretical_loss": 3.426309703030328, "tokens_seen": 2028797952 }, { "epoch": 0.23, "learning_rate": 7.78125501524635e-05, "loss": 2.483, "theoretical_loss": 3.4262912991651637, "tokens_seen": 2028929024 }, { "epoch": 0.23, "learning_rate": 7.780452575830525e-05, "loss": 2.5553, "theoretical_loss": 3.426272896821753, "tokens_seen": 2029060096 }, { "epoch": 0.23, "learning_rate": 7.779650136414702e-05, "loss": 2.4207, "theoretical_loss": 3.42625449599987, "tokens_seen": 2029191168 }, { "epoch": 0.23, "learning_rate": 7.778847696998877e-05, "loss": 2.5201, "theoretical_loss": 3.4262360966992924, "tokens_seen": 2029322240 }, { "epoch": 0.23, "learning_rate": 7.778045257583053e-05, "loss": 2.5239, "theoretical_loss": 3.426217698919795, "tokens_seen": 2029453312 }, { "epoch": 0.23, "learning_rate": 7.777242818167229e-05, "loss": 2.6701, "theoretical_loss": 3.426199302661155, "tokens_seen": 2029584384 }, { "epoch": 0.23, "learning_rate": 7.776440378751405e-05, "loss": 2.6428, "theoretical_loss": 3.426180907923147, "tokens_seen": 2029715456 }, { "epoch": 0.23, "learning_rate": 7.775637939335581e-05, "loss": 2.6025, "theoretical_loss": 3.426162514705548, "tokens_seen": 2029846528 }, { "epoch": 0.23, "learning_rate": 7.774835499919756e-05, "loss": 2.5519, "theoretical_loss": 3.426144123008134, "tokens_seen": 2029977600 }, { "epoch": 0.23, "objective/train/docs_used": 1118254, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5424864292144775, "objective/train/theoretical_loss": 3.426125732830682, "objective/train/tokens_used": 400567776, "theoretical_loss": 3.426125732830682, "tokens_seen": 2030108672 }, { "epoch": 0.23, "learning_rate": 7.774033060503933e-05, "loss": 2.4409, "theoretical_loss": 3.426125732830682, "tokens_seen": 2030108672 }, { "epoch": 0.23, "learning_rate": 7.773230621088108e-05, "loss": 2.5376, "theoretical_loss": 3.4261073441729666, "tokens_seen": 2030239744 }, { "epoch": 0.23, "learning_rate": 7.772428181672285e-05, "loss": 2.584, "theoretical_loss": 3.4260889570347652, "tokens_seen": 2030370816 }, { "epoch": 0.23, "learning_rate": 7.77162574225646e-05, "loss": 2.712, "theoretical_loss": 3.4260705714158544, "tokens_seen": 2030501888 }, { "epoch": 0.23, "learning_rate": 7.770823302840637e-05, "loss": 2.4428, "theoretical_loss": 3.42605218731601, "tokens_seen": 2030632960 }, { "epoch": 0.23, "learning_rate": 7.770020863424812e-05, "loss": 2.7793, "theoretical_loss": 3.4260338047350087, "tokens_seen": 2030764032 }, { "epoch": 0.23, "learning_rate": 7.769218424008987e-05, "loss": 2.4869, "theoretical_loss": 3.426015423672627, "tokens_seen": 2030895104 }, { "epoch": 0.23, "learning_rate": 7.768415984593164e-05, "loss": 2.64, "theoretical_loss": 3.425997044128641, "tokens_seen": 2031026176 }, { "epoch": 0.23, "learning_rate": 7.76761354517734e-05, "loss": 2.5178, "theoretical_loss": 3.4259786661028278, "tokens_seen": 2031157248 }, { "epoch": 0.23, "learning_rate": 7.766811105761516e-05, "loss": 2.6034, "theoretical_loss": 3.425960289594964, "tokens_seen": 2031288320 }, { "epoch": 0.23, "learning_rate": 7.766008666345691e-05, "loss": 2.4537, "theoretical_loss": 3.4259419146048264, "tokens_seen": 2031419392 }, { "epoch": 0.23, "learning_rate": 7.765206226929868e-05, "loss": 2.5037, "theoretical_loss": 3.425923541132191, "tokens_seen": 2031550464 }, { "epoch": 0.23, "learning_rate": 7.764403787514043e-05, "loss": 2.4198, "theoretical_loss": 3.4259051691768354, "tokens_seen": 2031681536 }, { "epoch": 0.23, "objective/train/docs_used": 1119492, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.9781367778778076, "objective/train/theoretical_loss": 3.425895983768068, "objective/train/tokens_used": 402206176, "theoretical_loss": 3.425895983768068, "tokens_seen": 2031747072 }, { "epoch": 0.23, "learning_rate": 7.76360134809822e-05, "loss": 2.6343, "theoretical_loss": 3.425886798738537, "tokens_seen": 2031812608 }, { "epoch": 0.23, "learning_rate": 7.762798908682395e-05, "loss": 2.4851, "theoretical_loss": 3.4258684298170703, "tokens_seen": 2031943680 }, { "epoch": 0.23, "learning_rate": 7.76199646926657e-05, "loss": 2.4681, "theoretical_loss": 3.4258500624122146, "tokens_seen": 2032074752 }, { "epoch": 0.23, "learning_rate": 7.761194029850747e-05, "loss": 2.6024, "theoretical_loss": 3.425831696523746, "tokens_seen": 2032205824 }, { "epoch": 0.23, "learning_rate": 7.760391590434923e-05, "loss": 2.4884, "theoretical_loss": 3.4258133321514412, "tokens_seen": 2032336896 }, { "epoch": 0.23, "learning_rate": 7.759589151019099e-05, "loss": 2.526, "theoretical_loss": 3.425794969295078, "tokens_seen": 2032467968 }, { "epoch": 0.23, "learning_rate": 7.758786711603275e-05, "loss": 2.591, "theoretical_loss": 3.4257766079544334, "tokens_seen": 2032599040 }, { "epoch": 0.23, "learning_rate": 7.757984272187451e-05, "loss": 2.6383, "theoretical_loss": 3.4257582481292834, "tokens_seen": 2032730112 }, { "epoch": 0.23, "learning_rate": 7.757181832771627e-05, "loss": 2.5295, "theoretical_loss": 3.4257398898194067, "tokens_seen": 2032861184 }, { "epoch": 0.23, "learning_rate": 7.756379393355802e-05, "loss": 2.5044, "theoretical_loss": 3.4257215330245794, "tokens_seen": 2032992256 }, { "epoch": 0.23, "learning_rate": 7.755576953939979e-05, "loss": 2.4814, "theoretical_loss": 3.4257031777445803, "tokens_seen": 2033123328 }, { "epoch": 0.23, "learning_rate": 7.754774514524154e-05, "loss": 2.6389, "theoretical_loss": 3.425684823979185, "tokens_seen": 2033254400 }, { "epoch": 0.23, "objective/train/docs_used": 1120104, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5154623985290527, "objective/train/theoretical_loss": 3.4256664717281717, "objective/train/tokens_used": 403844576, "theoretical_loss": 3.4256664717281717, "tokens_seen": 2033385472 }, { "epoch": 0.23, "learning_rate": 7.75397207510833e-05, "loss": 2.6288, "theoretical_loss": 3.4256664717281717, "tokens_seen": 2033385472 }, { "epoch": 0.23, "learning_rate": 7.753169635692506e-05, "loss": 2.3721, "theoretical_loss": 3.425648120991318, "tokens_seen": 2033516544 }, { "epoch": 0.23, "learning_rate": 7.752367196276682e-05, "loss": 2.4796, "theoretical_loss": 3.4256297717684014, "tokens_seen": 2033647616 }, { "epoch": 0.23, "learning_rate": 7.751564756860858e-05, "loss": 2.4372, "theoretical_loss": 3.4256114240591993, "tokens_seen": 2033778688 }, { "epoch": 0.23, "learning_rate": 7.750762317445033e-05, "loss": 2.5127, "theoretical_loss": 3.425593077863489, "tokens_seen": 2033909760 }, { "epoch": 0.23, "learning_rate": 7.74995987802921e-05, "loss": 2.485, "theoretical_loss": 3.4255747331810484, "tokens_seen": 2034040832 }, { "epoch": 0.23, "learning_rate": 7.749157438613385e-05, "loss": 2.4735, "theoretical_loss": 3.4255563900116552, "tokens_seen": 2034171904 }, { "epoch": 0.23, "learning_rate": 7.748354999197562e-05, "loss": 2.678, "theoretical_loss": 3.4255380483550866, "tokens_seen": 2034302976 }, { "epoch": 0.23, "learning_rate": 7.747552559781737e-05, "loss": 2.4402, "theoretical_loss": 3.4255197082111213, "tokens_seen": 2034434048 }, { "epoch": 0.23, "learning_rate": 7.746750120365914e-05, "loss": 2.557, "theoretical_loss": 3.4255013695795364, "tokens_seen": 2034565120 }, { "epoch": 0.23, "learning_rate": 7.745947680950089e-05, "loss": 2.5052, "theoretical_loss": 3.4254830324601104, "tokens_seen": 2034696192 }, { "epoch": 0.23, "learning_rate": 7.745145241534264e-05, "loss": 2.4159, "theoretical_loss": 3.425464696852621, "tokens_seen": 2034827264 }, { "epoch": 0.23, "learning_rate": 7.744342802118441e-05, "loss": 2.4283, "theoretical_loss": 3.425446362756845, "tokens_seen": 2034958336 }, { "epoch": 0.23, "objective/train/docs_used": 1121212, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.1749119758605957, "objective/train/theoretical_loss": 3.425437196275781, "objective/train/tokens_used": 405482976, "theoretical_loss": 3.425437196275781, "tokens_seen": 2035023872 }, { "epoch": 0.23, "learning_rate": 7.743540362702616e-05, "loss": 2.4383, "theoretical_loss": 3.4254280301725624, "tokens_seen": 2035089408 }, { "epoch": 0.23, "learning_rate": 7.742737923286793e-05, "loss": 2.679, "theoretical_loss": 3.42540969909955, "tokens_seen": 2035220480 }, { "epoch": 0.23, "learning_rate": 7.741935483870968e-05, "loss": 2.5305, "theoretical_loss": 3.4253913695375857, "tokens_seen": 2035351552 }, { "epoch": 0.23, "learning_rate": 7.741133044455145e-05, "loss": 2.6681, "theoretical_loss": 3.4253730414864485, "tokens_seen": 2035482624 }, { "epoch": 0.23, "learning_rate": 7.74033060503932e-05, "loss": 2.4899, "theoretical_loss": 3.4253547149459163, "tokens_seen": 2035613696 }, { "epoch": 0.23, "learning_rate": 7.739528165623496e-05, "loss": 2.7024, "theoretical_loss": 3.425336389915767, "tokens_seen": 2035744768 }, { "epoch": 0.23, "learning_rate": 7.738725726207672e-05, "loss": 2.5845, "theoretical_loss": 3.4253180663957794, "tokens_seen": 2035875840 }, { "epoch": 0.23, "learning_rate": 7.737923286791848e-05, "loss": 2.458, "theoretical_loss": 3.425299744385731, "tokens_seen": 2036006912 }, { "epoch": 0.23, "learning_rate": 7.737120847376024e-05, "loss": 2.6014, "theoretical_loss": 3.425281423885401, "tokens_seen": 2036137984 }, { "epoch": 0.23, "learning_rate": 7.7363184079602e-05, "loss": 2.533, "theoretical_loss": 3.425263104894568, "tokens_seen": 2036269056 }, { "epoch": 0.23, "learning_rate": 7.735515968544376e-05, "loss": 2.5045, "theoretical_loss": 3.4252447874130096, "tokens_seen": 2036400128 }, { "epoch": 0.23, "learning_rate": 7.734713529128552e-05, "loss": 2.6825, "theoretical_loss": 3.4252264714405047, "tokens_seen": 2036531200 }, { "epoch": 0.23, "objective/train/docs_used": 1122341, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8938703536987305, "objective/train/theoretical_loss": 3.425208156976832, "objective/train/tokens_used": 407121376, "theoretical_loss": 3.425208156976832, "tokens_seen": 2036662272 }, { "epoch": 0.23, "learning_rate": 7.733911089712728e-05, "loss": 2.7145, "theoretical_loss": 3.425208156976832, "tokens_seen": 2036662272 }, { "epoch": 0.23, "learning_rate": 7.733108650296904e-05, "loss": 2.5617, "theoretical_loss": 3.4251898440217703, "tokens_seen": 2036793344 }, { "epoch": 0.23, "learning_rate": 7.732306210881079e-05, "loss": 2.5561, "theoretical_loss": 3.4251715325750975, "tokens_seen": 2036924416 }, { "epoch": 0.23, "learning_rate": 7.731503771465256e-05, "loss": 2.4023, "theoretical_loss": 3.4251532226365935, "tokens_seen": 2037055488 }, { "epoch": 0.23, "learning_rate": 7.730701332049431e-05, "loss": 2.6135, "theoretical_loss": 3.425134914206036, "tokens_seen": 2037186560 }, { "epoch": 0.23, "learning_rate": 7.729898892633608e-05, "loss": 2.6687, "theoretical_loss": 3.425116607283204, "tokens_seen": 2037317632 }, { "epoch": 0.23, "learning_rate": 7.729096453217783e-05, "loss": 2.655, "theoretical_loss": 3.4250983018678767, "tokens_seen": 2037448704 }, { "epoch": 0.23, "learning_rate": 7.72829401380196e-05, "loss": 2.4404, "theoretical_loss": 3.4250799979598328, "tokens_seen": 2037579776 }, { "epoch": 0.23, "learning_rate": 7.727491574386135e-05, "loss": 2.5581, "theoretical_loss": 3.425061695558851, "tokens_seen": 2037710848 }, { "epoch": 0.24, "learning_rate": 7.72668913497031e-05, "loss": 2.5581, "theoretical_loss": 3.4250433946647107, "tokens_seen": 2037841920 }, { "epoch": 0.24, "learning_rate": 7.725886695554487e-05, "loss": 2.5608, "theoretical_loss": 3.4250250952771912, "tokens_seen": 2037972992 }, { "epoch": 0.24, "learning_rate": 7.725084256138662e-05, "loss": 2.624, "theoretical_loss": 3.4250067973960707, "tokens_seen": 2038104064 }, { "epoch": 0.24, "learning_rate": 7.724281816722839e-05, "loss": 2.5705, "theoretical_loss": 3.4249885010211285, "tokens_seen": 2038235136 }, { "epoch": 0.24, "objective/train/docs_used": 1123098, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.203213691711426, "objective/train/theoretical_loss": 3.4249793533984056, "objective/train/tokens_used": 408759776, "theoretical_loss": 3.4249793533984056, "tokens_seen": 2038300672 }, { "epoch": 0.24, "learning_rate": 7.723479377307014e-05, "loss": 2.4835, "theoretical_loss": 3.4249702061521443, "tokens_seen": 2038366208 }, { "epoch": 0.24, "learning_rate": 7.72267693789119e-05, "loss": 2.6758, "theoretical_loss": 3.4249519127888974, "tokens_seen": 2038497280 }, { "epoch": 0.24, "learning_rate": 7.721874498475366e-05, "loss": 2.4282, "theoretical_loss": 3.4249336209311667, "tokens_seen": 2038628352 }, { "epoch": 0.24, "learning_rate": 7.721072059059541e-05, "loss": 2.6284, "theoretical_loss": 3.4249153305787314, "tokens_seen": 2038759424 }, { "epoch": 0.24, "learning_rate": 7.720269619643718e-05, "loss": 2.627, "theoretical_loss": 3.424897041731371, "tokens_seen": 2038890496 }, { "epoch": 0.24, "learning_rate": 7.719467180227893e-05, "loss": 2.6302, "theoretical_loss": 3.4248787543888652, "tokens_seen": 2039021568 }, { "epoch": 0.24, "learning_rate": 7.71866474081207e-05, "loss": 2.4353, "theoretical_loss": 3.424860468550993, "tokens_seen": 2039152640 }, { "epoch": 0.24, "learning_rate": 7.717862301396245e-05, "loss": 2.3627, "theoretical_loss": 3.4248421842175336, "tokens_seen": 2039283712 }, { "epoch": 0.24, "learning_rate": 7.71705986198042e-05, "loss": 2.5219, "theoretical_loss": 3.424823901388268, "tokens_seen": 2039414784 }, { "epoch": 0.24, "learning_rate": 7.716257422564597e-05, "loss": 2.5145, "theoretical_loss": 3.4248056200629744, "tokens_seen": 2039545856 }, { "epoch": 0.24, "learning_rate": 7.715454983148773e-05, "loss": 2.599, "theoretical_loss": 3.4247873402414326, "tokens_seen": 2039676928 }, { "epoch": 0.24, "learning_rate": 7.714652543732949e-05, "loss": 2.5982, "theoretical_loss": 3.4247690619234223, "tokens_seen": 2039808000 }, { "epoch": 0.24, "objective/train/docs_used": 1124234, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5336902141571045, "objective/train/theoretical_loss": 3.424750785108724, "objective/train/tokens_used": 410398176, "theoretical_loss": 3.424750785108724, "tokens_seen": 2039939072 }, { "epoch": 0.24, "learning_rate": 7.713850104317125e-05, "loss": 2.4938, "theoretical_loss": 3.424750785108724, "tokens_seen": 2039939072 }, { "epoch": 0.24, "learning_rate": 7.7130476649013e-05, "loss": 2.2913, "theoretical_loss": 3.424732509797117, "tokens_seen": 2040070144 }, { "epoch": 0.24, "learning_rate": 7.712245225485477e-05, "loss": 2.5377, "theoretical_loss": 3.4247142359883807, "tokens_seen": 2040201216 }, { "epoch": 0.24, "learning_rate": 7.711442786069652e-05, "loss": 2.4873, "theoretical_loss": 3.4246959636822956, "tokens_seen": 2040332288 }, { "epoch": 0.24, "learning_rate": 7.710640346653829e-05, "loss": 2.3766, "theoretical_loss": 3.4246776928786415, "tokens_seen": 2040463360 }, { "epoch": 0.24, "learning_rate": 7.709837907238004e-05, "loss": 2.6614, "theoretical_loss": 3.4246594235771983, "tokens_seen": 2040594432 }, { "epoch": 0.24, "learning_rate": 7.70903546782218e-05, "loss": 2.5043, "theoretical_loss": 3.4246411557777456, "tokens_seen": 2040725504 }, { "epoch": 0.24, "learning_rate": 7.708233028406356e-05, "loss": 2.4483, "theoretical_loss": 3.424622889480064, "tokens_seen": 2040856576 }, { "epoch": 0.24, "learning_rate": 7.707430588990531e-05, "loss": 2.4879, "theoretical_loss": 3.4246046246839335, "tokens_seen": 2040987648 }, { "epoch": 0.24, "learning_rate": 7.706628149574708e-05, "loss": 2.6033, "theoretical_loss": 3.4245863613891343, "tokens_seen": 2041118720 }, { "epoch": 0.24, "learning_rate": 7.705825710158883e-05, "loss": 2.4924, "theoretical_loss": 3.424568099595446, "tokens_seen": 2041249792 }, { "epoch": 0.24, "learning_rate": 7.70502327074306e-05, "loss": 2.6618, "theoretical_loss": 3.4245498393026494, "tokens_seen": 2041380864 }, { "epoch": 0.24, "learning_rate": 7.704220831327235e-05, "loss": 2.5485, "theoretical_loss": 3.424531580510525, "tokens_seen": 2041511936 }, { "epoch": 0.24, "objective/train/docs_used": 1124915, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.3148484230041504, "objective/train/theoretical_loss": 3.424522451677146, "objective/train/tokens_used": 412036576, "theoretical_loss": 3.424522451677146, "tokens_seen": 2041577472 }, { "epoch": 0.24, "learning_rate": 7.70341839191141e-05, "loss": 2.4715, "theoretical_loss": 3.4245133232188527, "tokens_seen": 2041643008 }, { "epoch": 0.24, "learning_rate": 7.702615952495587e-05, "loss": 2.4311, "theoretical_loss": 3.424495067427413, "tokens_seen": 2041774080 }, { "epoch": 0.24, "learning_rate": 7.701813513079762e-05, "loss": 2.5216, "theoretical_loss": 3.424476813135986, "tokens_seen": 2041905152 }, { "epoch": 0.24, "learning_rate": 7.701011073663939e-05, "loss": 2.6073, "theoretical_loss": 3.4244585603443527, "tokens_seen": 2042036224 }, { "epoch": 0.24, "learning_rate": 7.700208634248114e-05, "loss": 2.609, "theoretical_loss": 3.4244403090522937, "tokens_seen": 2042167296 }, { "epoch": 0.24, "learning_rate": 7.699406194832291e-05, "loss": 2.5124, "theoretical_loss": 3.424422059259589, "tokens_seen": 2042298368 }, { "epoch": 0.24, "learning_rate": 7.698603755416466e-05, "loss": 2.4612, "theoretical_loss": 3.4244038109660195, "tokens_seen": 2042429440 }, { "epoch": 0.24, "learning_rate": 7.697801316000642e-05, "loss": 2.4941, "theoretical_loss": 3.4243855641713665, "tokens_seen": 2042560512 }, { "epoch": 0.24, "learning_rate": 7.696998876584818e-05, "loss": 2.7861, "theoretical_loss": 3.4243673188754093, "tokens_seen": 2042691584 }, { "epoch": 0.24, "learning_rate": 7.696196437168994e-05, "loss": 2.5488, "theoretical_loss": 3.42434907507793, "tokens_seen": 2042822656 }, { "epoch": 0.24, "learning_rate": 7.69539399775317e-05, "loss": 2.5045, "theoretical_loss": 3.424330832778708, "tokens_seen": 2042953728 }, { "epoch": 0.24, "learning_rate": 7.694591558337346e-05, "loss": 2.592, "theoretical_loss": 3.4243125919775257, "tokens_seen": 2043084800 }, { "epoch": 0.24, "objective/train/docs_used": 1125983, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.432866096496582, "objective/train/theoretical_loss": 3.424294352674163, "objective/train/tokens_used": 413674976, "theoretical_loss": 3.424294352674163, "tokens_seen": 2043215872 }, { "epoch": 0.24, "learning_rate": 7.693789118921521e-05, "loss": 2.4698, "theoretical_loss": 3.424294352674163, "tokens_seen": 2043215872 }, { "epoch": 0.24, "learning_rate": 7.692986679505698e-05, "loss": 2.5398, "theoretical_loss": 3.424276114868401, "tokens_seen": 2043346944 }, { "epoch": 0.24, "learning_rate": 7.692184240089873e-05, "loss": 2.5617, "theoretical_loss": 3.424257878560021, "tokens_seen": 2043478016 }, { "epoch": 0.24, "learning_rate": 7.69138180067405e-05, "loss": 2.4823, "theoretical_loss": 3.4242396437488036, "tokens_seen": 2043609088 }, { "epoch": 0.24, "learning_rate": 7.690579361258225e-05, "loss": 2.4788, "theoretical_loss": 3.42422141043453, "tokens_seen": 2043740160 }, { "epoch": 0.24, "learning_rate": 7.689776921842402e-05, "loss": 2.6387, "theoretical_loss": 3.4242031786169815, "tokens_seen": 2043871232 }, { "epoch": 0.24, "learning_rate": 7.688974482426577e-05, "loss": 2.5571, "theoretical_loss": 3.424184948295939, "tokens_seen": 2044002304 }, { "epoch": 0.24, "learning_rate": 7.688172043010752e-05, "loss": 2.6261, "theoretical_loss": 3.4241667194711845, "tokens_seen": 2044133376 }, { "epoch": 0.24, "learning_rate": 7.687369603594929e-05, "loss": 2.4709, "theoretical_loss": 3.424148492142498, "tokens_seen": 2044264448 }, { "epoch": 0.24, "learning_rate": 7.686567164179104e-05, "loss": 2.561, "theoretical_loss": 3.4241302663096613, "tokens_seen": 2044395520 }, { "epoch": 0.24, "learning_rate": 7.685764724763281e-05, "loss": 2.473, "theoretical_loss": 3.4241120419724567, "tokens_seen": 2044526592 }, { "epoch": 0.24, "learning_rate": 7.684962285347456e-05, "loss": 2.5965, "theoretical_loss": 3.424093819130664, "tokens_seen": 2044657664 }, { "epoch": 0.24, "learning_rate": 7.684159845931632e-05, "loss": 2.4805, "theoretical_loss": 3.4240755977840656, "tokens_seen": 2044788736 }, { "epoch": 0.24, "objective/train/docs_used": 1126576, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.342308759689331, "objective/train/theoretical_loss": 3.4240664876713955, "objective/train/tokens_used": 415313376, "theoretical_loss": 3.4240664876713955, "tokens_seen": 2044854272 }, { "epoch": 0.24, "learning_rate": 7.683357406515808e-05, "loss": 2.5274, "theoretical_loss": 3.424057377932443, "tokens_seen": 2044919808 }, { "epoch": 0.24, "learning_rate": 7.682554967099984e-05, "loss": 2.5764, "theoretical_loss": 3.4240391595755773, "tokens_seen": 2045050880 }, { "epoch": 0.24, "learning_rate": 7.68175252768416e-05, "loss": 2.4526, "theoretical_loss": 3.42402094271325, "tokens_seen": 2045181952 }, { "epoch": 0.24, "learning_rate": 7.680950088268335e-05, "loss": 2.5081, "theoretical_loss": 3.424002727345244, "tokens_seen": 2045313024 }, { "epoch": 0.24, "learning_rate": 7.680147648852512e-05, "loss": 2.4939, "theoretical_loss": 3.423984513471339, "tokens_seen": 2045444096 }, { "epoch": 0.24, "learning_rate": 7.679345209436687e-05, "loss": 2.3648, "theoretical_loss": 3.423966301091318, "tokens_seen": 2045575168 }, { "epoch": 0.24, "learning_rate": 7.678542770020863e-05, "loss": 2.5418, "theoretical_loss": 3.423948090204963, "tokens_seen": 2045706240 }, { "epoch": 0.24, "learning_rate": 7.67774033060504e-05, "loss": 2.4757, "theoretical_loss": 3.4239298808120546, "tokens_seen": 2045837312 }, { "epoch": 0.24, "learning_rate": 7.676937891189215e-05, "loss": 2.4403, "theoretical_loss": 3.423911672912376, "tokens_seen": 2045968384 }, { "epoch": 0.24, "learning_rate": 7.676135451773391e-05, "loss": 2.442, "theoretical_loss": 3.4238934665057084, "tokens_seen": 2046099456 }, { "epoch": 0.24, "learning_rate": 7.675333012357567e-05, "loss": 2.3285, "theoretical_loss": 3.4238752615918333, "tokens_seen": 2046230528 }, { "epoch": 0.24, "learning_rate": 7.674530572941742e-05, "loss": 2.6098, "theoretical_loss": 3.4238570581705337, "tokens_seen": 2046361600 }, { "epoch": 0.24, "objective/train/docs_used": 1127632, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.210137128829956, "objective/train/theoretical_loss": 3.423838856241591, "objective/train/tokens_used": 416951776, "theoretical_loss": 3.423838856241591, "tokens_seen": 2046492672 }, { "epoch": 0.24, "learning_rate": 7.673728133525919e-05, "loss": 2.4872, "theoretical_loss": 3.423838856241591, "tokens_seen": 2046492672 }, { "epoch": 0.24, "learning_rate": 7.672925694110094e-05, "loss": 2.629, "theoretical_loss": 3.4238206558047874, "tokens_seen": 2046623744 }, { "epoch": 0.24, "learning_rate": 7.672123254694271e-05, "loss": 2.4547, "theoretical_loss": 3.423802456859905, "tokens_seen": 2046754816 }, { "epoch": 0.24, "learning_rate": 7.671320815278446e-05, "loss": 2.4999, "theoretical_loss": 3.423784259406726, "tokens_seen": 2046885888 }, { "epoch": 0.24, "learning_rate": 7.670518375862623e-05, "loss": 2.4866, "theoretical_loss": 3.4237660634450324, "tokens_seen": 2047016960 }, { "epoch": 0.24, "learning_rate": 7.669715936446798e-05, "loss": 2.4262, "theoretical_loss": 3.4237478689746066, "tokens_seen": 2047148032 }, { "epoch": 0.24, "learning_rate": 7.668913497030975e-05, "loss": 2.3548, "theoretical_loss": 3.4237296759952316, "tokens_seen": 2047279104 }, { "epoch": 0.24, "learning_rate": 7.66811105761515e-05, "loss": 2.546, "theoretical_loss": 3.423711484506689, "tokens_seen": 2047410176 }, { "epoch": 0.24, "learning_rate": 7.667308618199325e-05, "loss": 2.5654, "theoretical_loss": 3.4236932945087615, "tokens_seen": 2047541248 }, { "epoch": 0.24, "learning_rate": 7.666506178783502e-05, "loss": 2.4401, "theoretical_loss": 3.423675106001231, "tokens_seen": 2047672320 }, { "epoch": 0.24, "learning_rate": 7.665703739367677e-05, "loss": 2.4974, "theoretical_loss": 3.423656918983881, "tokens_seen": 2047803392 }, { "epoch": 0.24, "learning_rate": 7.664901299951854e-05, "loss": 2.4328, "theoretical_loss": 3.4236387334564933, "tokens_seen": 2047934464 }, { "epoch": 0.24, "learning_rate": 7.664098860536029e-05, "loss": 2.392, "theoretical_loss": 3.42362054941885, "tokens_seen": 2048065536 }, { "epoch": 0.24, "objective/train/docs_used": 1128264, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3604929447174072, "objective/train/theoretical_loss": 3.423611457958615, "objective/train/tokens_used": 418590176, "theoretical_loss": 3.423611457958615, "tokens_seen": 2048131072 }, { "epoch": 0.24, "learning_rate": 7.663296421120206e-05, "loss": 2.6137, "theoretical_loss": 3.4236023668707354, "tokens_seen": 2048196608 }, { "epoch": 0.24, "learning_rate": 7.662493981704381e-05, "loss": 2.468, "theoretical_loss": 3.4235841858119302, "tokens_seen": 2048327680 }, { "epoch": 0.24, "learning_rate": 7.661691542288557e-05, "loss": 2.5171, "theoretical_loss": 3.4235660062422184, "tokens_seen": 2048458752 }, { "epoch": 0.24, "learning_rate": 7.660889102872733e-05, "loss": 2.5218, "theoretical_loss": 3.423547828161383, "tokens_seen": 2048589824 }, { "epoch": 0.24, "learning_rate": 7.660086663456909e-05, "loss": 2.4335, "theoretical_loss": 3.423529651569206, "tokens_seen": 2048720896 }, { "epoch": 0.24, "learning_rate": 7.659284224041085e-05, "loss": 2.4221, "theoretical_loss": 3.42351147646547, "tokens_seen": 2048851968 }, { "epoch": 0.24, "learning_rate": 7.65848178462526e-05, "loss": 2.413, "theoretical_loss": 3.423493302849959, "tokens_seen": 2048983040 }, { "epoch": 0.24, "learning_rate": 7.657679345209437e-05, "loss": 2.5214, "theoretical_loss": 3.4234751307224554, "tokens_seen": 2049114112 }, { "epoch": 0.24, "learning_rate": 7.656876905793612e-05, "loss": 2.4643, "theoretical_loss": 3.423456960082742, "tokens_seen": 2049245184 }, { "epoch": 0.24, "learning_rate": 7.656074466377788e-05, "loss": 2.4014, "theoretical_loss": 3.423438790930602, "tokens_seen": 2049376256 }, { "epoch": 0.24, "learning_rate": 7.655272026961964e-05, "loss": 2.6017, "theoretical_loss": 3.4234206232658186, "tokens_seen": 2049507328 }, { "epoch": 0.24, "learning_rate": 7.65446958754614e-05, "loss": 2.4636, "theoretical_loss": 3.4234024570881747, "tokens_seen": 2049638400 }, { "epoch": 0.24, "objective/train/docs_used": 1129400, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.676476001739502, "objective/train/theoretical_loss": 3.4233842923974542, "objective/train/tokens_used": 420228576, "theoretical_loss": 3.4233842923974542, "tokens_seen": 2049769472 }, { "epoch": 0.24, "learning_rate": 7.653667148130316e-05, "loss": 2.6725, "theoretical_loss": 3.4233842923974542, "tokens_seen": 2049769472 }, { "epoch": 0.24, "learning_rate": 7.652864708714492e-05, "loss": 2.5167, "theoretical_loss": 3.423366129193439, "tokens_seen": 2049900544 }, { "epoch": 0.24, "learning_rate": 7.652062269298668e-05, "loss": 2.4925, "theoretical_loss": 3.4233479674759137, "tokens_seen": 2050031616 }, { "epoch": 0.24, "learning_rate": 7.651259829882844e-05, "loss": 2.3167, "theoretical_loss": 3.4233298072446607, "tokens_seen": 2050162688 }, { "epoch": 0.24, "learning_rate": 7.650457390467019e-05, "loss": 2.3845, "theoretical_loss": 3.423311648499464, "tokens_seen": 2050293760 }, { "epoch": 0.24, "learning_rate": 7.649654951051196e-05, "loss": 2.5006, "theoretical_loss": 3.4232934912401065, "tokens_seen": 2050424832 }, { "epoch": 0.24, "learning_rate": 7.648852511635371e-05, "loss": 2.4615, "theoretical_loss": 3.4232753354663714, "tokens_seen": 2050555904 }, { "epoch": 0.24, "learning_rate": 7.648050072219548e-05, "loss": 2.4585, "theoretical_loss": 3.4232571811780437, "tokens_seen": 2050686976 }, { "epoch": 0.24, "learning_rate": 7.647247632803723e-05, "loss": 2.5186, "theoretical_loss": 3.423239028374905, "tokens_seen": 2050818048 }, { "epoch": 0.24, "learning_rate": 7.6464451933879e-05, "loss": 2.4754, "theoretical_loss": 3.4232208770567403, "tokens_seen": 2050949120 }, { "epoch": 0.24, "learning_rate": 7.645642753972075e-05, "loss": 2.5151, "theoretical_loss": 3.4232027272233325, "tokens_seen": 2051080192 }, { "epoch": 0.24, "learning_rate": 7.64484031455625e-05, "loss": 2.7089, "theoretical_loss": 3.4231845788744657, "tokens_seen": 2051211264 }, { "epoch": 0.24, "learning_rate": 7.644037875140427e-05, "loss": 2.5746, "theoretical_loss": 3.423166432009923, "tokens_seen": 2051342336 }, { "epoch": 0.24, "objective/train/docs_used": 1129773, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.70103120803833, "objective/train/theoretical_loss": 3.4231573591342057, "objective/train/tokens_used": 421866976, "theoretical_loss": 3.4231573591342057, "tokens_seen": 2051407872 }, { "epoch": 0.24, "learning_rate": 7.643235435724602e-05, "loss": 2.5127, "theoretical_loss": 3.4231482866294884, "tokens_seen": 2051473408 }, { "epoch": 0.24, "learning_rate": 7.642432996308779e-05, "loss": 2.4836, "theoretical_loss": 3.423130142732947, "tokens_seen": 2051604480 }, { "epoch": 0.24, "learning_rate": 7.641630556892954e-05, "loss": 2.4251, "theoretical_loss": 3.4231120003200806, "tokens_seen": 2051735552 }, { "epoch": 0.24, "learning_rate": 7.640828117477131e-05, "loss": 2.48, "theoretical_loss": 3.4230938593906743, "tokens_seen": 2051866624 }, { "epoch": 0.24, "learning_rate": 7.640025678061306e-05, "loss": 2.5438, "theoretical_loss": 3.423075719944512, "tokens_seen": 2051997696 }, { "epoch": 0.24, "learning_rate": 7.639223238645483e-05, "loss": 2.416, "theoretical_loss": 3.4230575819813778, "tokens_seen": 2052128768 }, { "epoch": 0.24, "learning_rate": 7.638420799229658e-05, "loss": 2.6944, "theoretical_loss": 3.423039445501055, "tokens_seen": 2052259840 }, { "epoch": 0.24, "learning_rate": 7.637618359813834e-05, "loss": 2.3774, "theoretical_loss": 3.423021310503328, "tokens_seen": 2052390912 }, { "epoch": 0.24, "learning_rate": 7.63681592039801e-05, "loss": 2.3808, "theoretical_loss": 3.423003176987981, "tokens_seen": 2052521984 }, { "epoch": 0.24, "learning_rate": 7.636013480982186e-05, "loss": 2.4532, "theoretical_loss": 3.4229850449547987, "tokens_seen": 2052653056 }, { "epoch": 0.24, "learning_rate": 7.635211041566362e-05, "loss": 2.5069, "theoretical_loss": 3.422966914403565, "tokens_seen": 2052784128 }, { "epoch": 0.24, "learning_rate": 7.634408602150538e-05, "loss": 2.4513, "theoretical_loss": 3.4229487853340634, "tokens_seen": 2052915200 }, { "epoch": 0.24, "objective/train/docs_used": 1130910, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.115882396697998, "objective/train/theoretical_loss": 3.422930657746079, "objective/train/tokens_used": 423505376, "theoretical_loss": 3.422930657746079, "tokens_seen": 2053046272 }, { "epoch": 0.24, "learning_rate": 7.633606162734714e-05, "loss": 2.514, "theoretical_loss": 3.422930657746079, "tokens_seen": 2053046272 }, { "epoch": 0.24, "learning_rate": 7.63280372331889e-05, "loss": 2.5533, "theoretical_loss": 3.4229125316393967, "tokens_seen": 2053177344 }, { "epoch": 0.24, "learning_rate": 7.632001283903065e-05, "loss": 2.4303, "theoretical_loss": 3.4228944070137994, "tokens_seen": 2053308416 }, { "epoch": 0.24, "learning_rate": 7.631198844487241e-05, "loss": 2.5642, "theoretical_loss": 3.4228762838690727, "tokens_seen": 2053439488 }, { "epoch": 0.24, "learning_rate": 7.630396405071417e-05, "loss": 2.5084, "theoretical_loss": 3.4228581622050007, "tokens_seen": 2053570560 }, { "epoch": 0.24, "learning_rate": 7.629593965655593e-05, "loss": 2.5071, "theoretical_loss": 3.422840042021368, "tokens_seen": 2053701632 }, { "epoch": 0.24, "learning_rate": 7.628791526239769e-05, "loss": 2.5386, "theoretical_loss": 3.422821923317959, "tokens_seen": 2053832704 }, { "epoch": 0.24, "learning_rate": 7.627989086823945e-05, "loss": 2.5596, "theoretical_loss": 3.422803806094559, "tokens_seen": 2053963776 }, { "epoch": 0.24, "learning_rate": 7.627186647408121e-05, "loss": 2.6299, "theoretical_loss": 3.4227856903509517, "tokens_seen": 2054094848 }, { "epoch": 0.24, "learning_rate": 7.626384207992296e-05, "loss": 2.2966, "theoretical_loss": 3.4227675760869225, "tokens_seen": 2054225920 }, { "epoch": 0.25, "learning_rate": 7.625581768576473e-05, "loss": 2.5026, "theoretical_loss": 3.4227494633022557, "tokens_seen": 2054356992 }, { "epoch": 0.25, "learning_rate": 7.624779329160648e-05, "loss": 2.4105, "theoretical_loss": 3.4227313519967364, "tokens_seen": 2054488064 }, { "epoch": 0.25, "learning_rate": 7.623976889744825e-05, "loss": 2.467, "theoretical_loss": 3.4227132421701496, "tokens_seen": 2054619136 }, { "epoch": 0.25, "objective/train/docs_used": 1131479, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.358306884765625, "objective/train/theoretical_loss": 3.422704187811389, "objective/train/tokens_used": 425143776, "theoretical_loss": 3.422704187811389, "tokens_seen": 2054684672 }, { "epoch": 0.25, "learning_rate": 7.623174450329e-05, "loss": 2.499, "theoretical_loss": 3.4226951338222804, "tokens_seen": 2054750208 }, { "epoch": 0.25, "learning_rate": 7.622372010913177e-05, "loss": 2.6132, "theoretical_loss": 3.422677026952913, "tokens_seen": 2054881280 }, { "epoch": 0.25, "learning_rate": 7.621569571497352e-05, "loss": 2.6517, "theoretical_loss": 3.4226589215618324, "tokens_seen": 2055012352 }, { "epoch": 0.25, "learning_rate": 7.620767132081527e-05, "loss": 2.5089, "theoretical_loss": 3.4226408176488246, "tokens_seen": 2055143424 }, { "epoch": 0.25, "learning_rate": 7.619964692665704e-05, "loss": 2.5091, "theoretical_loss": 3.4226227152136737, "tokens_seen": 2055274496 }, { "epoch": 0.25, "learning_rate": 7.619162253249879e-05, "loss": 2.397, "theoretical_loss": 3.422604614256165, "tokens_seen": 2055405568 }, { "epoch": 0.25, "learning_rate": 7.618359813834056e-05, "loss": 2.5668, "theoretical_loss": 3.422586514776084, "tokens_seen": 2055536640 }, { "epoch": 0.25, "learning_rate": 7.617557374418231e-05, "loss": 2.6935, "theoretical_loss": 3.422568416773216, "tokens_seen": 2055667712 }, { "epoch": 0.25, "learning_rate": 7.616754935002408e-05, "loss": 2.3887, "theoretical_loss": 3.4225503202473453, "tokens_seen": 2055798784 }, { "epoch": 0.25, "learning_rate": 7.615952495586583e-05, "loss": 2.48, "theoretical_loss": 3.4225322251982586, "tokens_seen": 2055929856 }, { "epoch": 0.25, "learning_rate": 7.615150056170759e-05, "loss": 2.3379, "theoretical_loss": 3.42251413162574, "tokens_seen": 2056060928 }, { "epoch": 0.25, "learning_rate": 7.614347616754935e-05, "loss": 2.5482, "theoretical_loss": 3.4224960395295763, "tokens_seen": 2056192000 }, { "epoch": 0.25, "objective/train/docs_used": 1132437, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8114593029022217, "objective/train/theoretical_loss": 3.4224779489095516, "objective/train/tokens_used": 426782176, "theoretical_loss": 3.4224779489095516, "tokens_seen": 2056323072 }, { "epoch": 0.25, "learning_rate": 7.61354517733911e-05, "loss": 2.4689, "theoretical_loss": 3.4224779489095516, "tokens_seen": 2056323072 }, { "epoch": 0.25, "learning_rate": 7.612742737923287e-05, "loss": 2.5355, "theoretical_loss": 3.422459859765452, "tokens_seen": 2056454144 }, { "epoch": 0.25, "learning_rate": 7.611940298507463e-05, "loss": 2.4067, "theoretical_loss": 3.422441772097063, "tokens_seen": 2056585216 }, { "epoch": 0.25, "learning_rate": 7.611137859091639e-05, "loss": 2.5967, "theoretical_loss": 3.42242368590417, "tokens_seen": 2056716288 }, { "epoch": 0.25, "learning_rate": 7.610335419675815e-05, "loss": 2.5954, "theoretical_loss": 3.4224056011865587, "tokens_seen": 2056847360 }, { "epoch": 0.25, "learning_rate": 7.609532980259991e-05, "loss": 2.5875, "theoretical_loss": 3.4223875179440144, "tokens_seen": 2056978432 }, { "epoch": 0.25, "learning_rate": 7.608730540844167e-05, "loss": 2.7053, "theoretical_loss": 3.4223694361763233, "tokens_seen": 2057109504 }, { "epoch": 0.25, "learning_rate": 7.607928101428342e-05, "loss": 2.5196, "theoretical_loss": 3.422351355883271, "tokens_seen": 2057240576 }, { "epoch": 0.25, "learning_rate": 7.607125662012518e-05, "loss": 2.5351, "theoretical_loss": 3.4223332770646437, "tokens_seen": 2057371648 }, { "epoch": 0.25, "learning_rate": 7.606323222596694e-05, "loss": 2.3518, "theoretical_loss": 3.422315199720227, "tokens_seen": 2057502720 }, { "epoch": 0.25, "learning_rate": 7.60552078318087e-05, "loss": 2.4935, "theoretical_loss": 3.422297123849806, "tokens_seen": 2057633792 }, { "epoch": 0.25, "learning_rate": 7.604718343765046e-05, "loss": 2.5397, "theoretical_loss": 3.4222790494531674, "tokens_seen": 2057764864 }, { "epoch": 0.25, "learning_rate": 7.603915904349222e-05, "loss": 2.4483, "theoretical_loss": 3.4222609765300973, "tokens_seen": 2057895936 }, { "epoch": 0.25, "objective/train/docs_used": 1133547, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.867115020751953, "objective/train/theoretical_loss": 3.4222519406210834, "objective/train/tokens_used": 428420576, "theoretical_loss": 3.4222519406210834, "tokens_seen": 2057961472 }, { "epoch": 0.25, "learning_rate": 7.603113464933398e-05, "loss": 2.597, "theoretical_loss": 3.422242905080381, "tokens_seen": 2058027008 }, { "epoch": 0.25, "learning_rate": 7.602311025517573e-05, "loss": 2.6583, "theoretical_loss": 3.4222248351038056, "tokens_seen": 2058158080 }, { "epoch": 0.25, "learning_rate": 7.60150858610175e-05, "loss": 2.541, "theoretical_loss": 3.4222067666001563, "tokens_seen": 2058289152 }, { "epoch": 0.25, "learning_rate": 7.600706146685925e-05, "loss": 2.3219, "theoretical_loss": 3.4221886995692197, "tokens_seen": 2058420224 }, { "epoch": 0.25, "learning_rate": 7.599903707270102e-05, "loss": 2.3508, "theoretical_loss": 3.422170634010782, "tokens_seen": 2058551296 }, { "epoch": 0.25, "learning_rate": 7.599101267854277e-05, "loss": 2.6189, "theoretical_loss": 3.422152569924629, "tokens_seen": 2058682368 }, { "epoch": 0.25, "learning_rate": 7.598298828438454e-05, "loss": 2.4071, "theoretical_loss": 3.422134507310548, "tokens_seen": 2058813440 }, { "epoch": 0.25, "learning_rate": 7.597496389022629e-05, "loss": 2.3938, "theoretical_loss": 3.4221164461683236, "tokens_seen": 2058944512 }, { "epoch": 0.25, "learning_rate": 7.596693949606804e-05, "loss": 2.4802, "theoretical_loss": 3.422098386497744, "tokens_seen": 2059075584 }, { "epoch": 0.25, "learning_rate": 7.595891510190981e-05, "loss": 2.5144, "theoretical_loss": 3.4220803282985948, "tokens_seen": 2059206656 }, { "epoch": 0.25, "learning_rate": 7.595089070775156e-05, "loss": 2.7101, "theoretical_loss": 3.4220622715706623, "tokens_seen": 2059337728 }, { "epoch": 0.25, "learning_rate": 7.594286631359333e-05, "loss": 2.3732, "theoretical_loss": 3.4220442163137332, "tokens_seen": 2059468800 }, { "epoch": 0.25, "objective/train/docs_used": 1134157, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.846436023712158, "objective/train/theoretical_loss": 3.4220261625275943, "objective/train/tokens_used": 430058976, "theoretical_loss": 3.4220261625275943, "tokens_seen": 2059599872 }, { "epoch": 0.25, "learning_rate": 7.593484191943508e-05, "loss": 2.6596, "theoretical_loss": 3.4220261625275943, "tokens_seen": 2059599872 }, { "epoch": 0.25, "learning_rate": 7.592681752527685e-05, "loss": 2.5337, "theoretical_loss": 3.422008110212032, "tokens_seen": 2059730944 }, { "epoch": 0.25, "learning_rate": 7.59187931311186e-05, "loss": 2.5602, "theoretical_loss": 3.4219900593668324, "tokens_seen": 2059862016 }, { "epoch": 0.25, "learning_rate": 7.591076873696036e-05, "loss": 2.5941, "theoretical_loss": 3.421972009991783, "tokens_seen": 2059993088 }, { "epoch": 0.25, "learning_rate": 7.590274434280212e-05, "loss": 2.5179, "theoretical_loss": 3.4219539620866706, "tokens_seen": 2060124160 }, { "epoch": 0.25, "learning_rate": 7.589471994864388e-05, "loss": 2.6585, "theoretical_loss": 3.4219359156512814, "tokens_seen": 2060255232 }, { "epoch": 0.25, "learning_rate": 7.588669555448564e-05, "loss": 2.6361, "theoretical_loss": 3.4219178706854025, "tokens_seen": 2060386304 }, { "epoch": 0.25, "learning_rate": 7.58786711603274e-05, "loss": 2.5926, "theoretical_loss": 3.4218998271888204, "tokens_seen": 2060517376 }, { "epoch": 0.25, "learning_rate": 7.587064676616916e-05, "loss": 2.5422, "theoretical_loss": 3.4218817851613226, "tokens_seen": 2060648448 }, { "epoch": 0.25, "learning_rate": 7.586262237201092e-05, "loss": 2.7374, "theoretical_loss": 3.4218637446026956, "tokens_seen": 2060779520 }, { "epoch": 0.25, "learning_rate": 7.585459797785267e-05, "loss": 2.4504, "theoretical_loss": 3.4218457055127267, "tokens_seen": 2060910592 }, { "epoch": 0.25, "learning_rate": 7.584657358369444e-05, "loss": 2.5051, "theoretical_loss": 3.421827667891203, "tokens_seen": 2061041664 }, { "epoch": 0.25, "learning_rate": 7.583854918953619e-05, "loss": 2.5216, "theoretical_loss": 3.4218096317379114, "tokens_seen": 2061172736 }, { "epoch": 0.25, "objective/train/docs_used": 1135026, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5525903701782227, "objective/train/theoretical_loss": 3.4218006142117856, "objective/train/tokens_used": 431697376, "theoretical_loss": 3.4218006142117856, "tokens_seen": 2061238272 }, { "epoch": 0.25, "learning_rate": 7.583052479537795e-05, "loss": 2.5612, "theoretical_loss": 3.4217915970526387, "tokens_seen": 2061303808 }, { "epoch": 0.25, "learning_rate": 7.582250040121971e-05, "loss": 2.5427, "theoretical_loss": 3.4217735638351727, "tokens_seen": 2061434880 }, { "epoch": 0.25, "learning_rate": 7.581447600706147e-05, "loss": 2.5848, "theoretical_loss": 3.4217555320853004, "tokens_seen": 2061565952 }, { "epoch": 0.25, "learning_rate": 7.580645161290323e-05, "loss": 2.6088, "theoretical_loss": 3.4217375018028084, "tokens_seen": 2061697024 }, { "epoch": 0.25, "learning_rate": 7.579842721874498e-05, "loss": 2.4747, "theoretical_loss": 3.4217194729874856, "tokens_seen": 2061828096 }, { "epoch": 0.25, "learning_rate": 7.579040282458675e-05, "loss": 2.652, "theoretical_loss": 3.4217014456391177, "tokens_seen": 2061959168 }, { "epoch": 0.25, "learning_rate": 7.57823784304285e-05, "loss": 2.4892, "theoretical_loss": 3.421683419757493, "tokens_seen": 2062090240 }, { "epoch": 0.25, "learning_rate": 7.577435403627027e-05, "loss": 2.5593, "theoretical_loss": 3.4216653953423988, "tokens_seen": 2062221312 }, { "epoch": 0.25, "learning_rate": 7.576632964211202e-05, "loss": 2.5257, "theoretical_loss": 3.4216473723936223, "tokens_seen": 2062352384 }, { "epoch": 0.25, "learning_rate": 7.575830524795379e-05, "loss": 2.7893, "theoretical_loss": 3.421629350910951, "tokens_seen": 2062483456 }, { "epoch": 0.25, "learning_rate": 7.575028085379554e-05, "loss": 2.51, "theoretical_loss": 3.421611330894174, "tokens_seen": 2062614528 }, { "epoch": 0.25, "learning_rate": 7.574225645963731e-05, "loss": 2.6444, "theoretical_loss": 3.4215933123430764, "tokens_seen": 2062745600 }, { "epoch": 0.25, "objective/train/docs_used": 1136115, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5874297618865967, "objective/train/theoretical_loss": 3.421575295257447, "objective/train/tokens_used": 433335776, "theoretical_loss": 3.421575295257447, "tokens_seen": 2062876672 }, { "epoch": 0.25, "learning_rate": 7.573423206547906e-05, "loss": 2.6763, "theoretical_loss": 3.421575295257447, "tokens_seen": 2062876672 }, { "epoch": 0.25, "learning_rate": 7.572620767132081e-05, "loss": 2.6042, "theoretical_loss": 3.4215572796370743, "tokens_seen": 2063007744 }, { "epoch": 0.25, "learning_rate": 7.571818327716258e-05, "loss": 2.534, "theoretical_loss": 3.421539265481745, "tokens_seen": 2063138816 }, { "epoch": 0.25, "learning_rate": 7.571015888300433e-05, "loss": 2.4402, "theoretical_loss": 3.4215212527912477, "tokens_seen": 2063269888 }, { "epoch": 0.25, "learning_rate": 7.57021344888461e-05, "loss": 2.5752, "theoretical_loss": 3.4215032415653694, "tokens_seen": 2063400960 }, { "epoch": 0.25, "learning_rate": 7.569411009468785e-05, "loss": 2.5612, "theoretical_loss": 3.4214852318038984, "tokens_seen": 2063532032 }, { "epoch": 0.25, "learning_rate": 7.568608570052962e-05, "loss": 2.5878, "theoretical_loss": 3.421467223506623, "tokens_seen": 2063663104 }, { "epoch": 0.25, "learning_rate": 7.567806130637137e-05, "loss": 2.6003, "theoretical_loss": 3.4214492166733304, "tokens_seen": 2063794176 }, { "epoch": 0.25, "learning_rate": 7.567003691221313e-05, "loss": 2.5608, "theoretical_loss": 3.4214312113038092, "tokens_seen": 2063925248 }, { "epoch": 0.25, "learning_rate": 7.566201251805489e-05, "loss": 2.5979, "theoretical_loss": 3.421413207397847, "tokens_seen": 2064056320 }, { "epoch": 0.25, "learning_rate": 7.565398812389665e-05, "loss": 2.4841, "theoretical_loss": 3.4213952049552327, "tokens_seen": 2064187392 }, { "epoch": 0.25, "learning_rate": 7.564596372973841e-05, "loss": 2.6268, "theoretical_loss": 3.4213772039757537, "tokens_seen": 2064318464 }, { "epoch": 0.25, "learning_rate": 7.563793933558017e-05, "loss": 2.5995, "theoretical_loss": 3.4213592044591987, "tokens_seen": 2064449536 }, { "epoch": 0.25, "objective/train/docs_used": 1136653, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.423219919204712, "objective/train/theoretical_loss": 3.421350205249451, "objective/train/tokens_used": 434974176, "theoretical_loss": 3.421350205249451, "tokens_seen": 2064515072 }, { "epoch": 0.25, "learning_rate": 7.562991494142193e-05, "loss": 2.6165, "theoretical_loss": 3.421341206405355, "tokens_seen": 2064580608 }, { "epoch": 0.25, "learning_rate": 7.562189054726369e-05, "loss": 2.5891, "theoretical_loss": 3.421323209814012, "tokens_seen": 2064711680 }, { "epoch": 0.25, "learning_rate": 7.561386615310544e-05, "loss": 2.7013, "theoretical_loss": 3.421305214684958, "tokens_seen": 2064842752 }, { "epoch": 0.25, "learning_rate": 7.56058417589472e-05, "loss": 2.492, "theoretical_loss": 3.4212872210179803, "tokens_seen": 2064973824 }, { "epoch": 0.25, "learning_rate": 7.559781736478896e-05, "loss": 2.4796, "theoretical_loss": 3.4212692288128683, "tokens_seen": 2065104896 }, { "epoch": 0.25, "learning_rate": 7.558979297063073e-05, "loss": 2.5791, "theoretical_loss": 3.4212512380694102, "tokens_seen": 2065235968 }, { "epoch": 0.25, "learning_rate": 7.558176857647248e-05, "loss": 2.5923, "theoretical_loss": 3.421233248787394, "tokens_seen": 2065367040 }, { "epoch": 0.25, "learning_rate": 7.557374418231424e-05, "loss": 2.5457, "theoretical_loss": 3.4212152609666093, "tokens_seen": 2065498112 }, { "epoch": 0.25, "learning_rate": 7.5565719788156e-05, "loss": 2.4568, "theoretical_loss": 3.4211972746068438, "tokens_seen": 2065629184 }, { "epoch": 0.25, "learning_rate": 7.555769539399775e-05, "loss": 2.6654, "theoretical_loss": 3.4211792897078865, "tokens_seen": 2065760256 }, { "epoch": 0.25, "learning_rate": 7.554967099983952e-05, "loss": 2.5968, "theoretical_loss": 3.421161306269526, "tokens_seen": 2065891328 }, { "epoch": 0.25, "learning_rate": 7.554164660568127e-05, "loss": 2.6777, "theoretical_loss": 3.4211433242915508, "tokens_seen": 2066022400 }, { "epoch": 0.25, "objective/train/docs_used": 1137651, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5705416202545166, "objective/train/theoretical_loss": 3.42112534377375, "objective/train/tokens_used": 436612576, "theoretical_loss": 3.42112534377375, "tokens_seen": 2066153472 }, { "epoch": 0.25, "learning_rate": 7.553362221152304e-05, "loss": 2.4683, "theoretical_loss": 3.42112534377375, "tokens_seen": 2066153472 }, { "epoch": 0.25, "learning_rate": 7.552559781736479e-05, "loss": 2.6254, "theoretical_loss": 3.4211073647159123, "tokens_seen": 2066284544 }, { "epoch": 0.25, "learning_rate": 7.551757342320656e-05, "loss": 2.5953, "theoretical_loss": 3.4210893871178265, "tokens_seen": 2066415616 }, { "epoch": 0.25, "learning_rate": 7.550954902904831e-05, "loss": 2.5553, "theoretical_loss": 3.4210714109792817, "tokens_seen": 2066546688 }, { "epoch": 0.25, "learning_rate": 7.550152463489006e-05, "loss": 2.7196, "theoretical_loss": 3.421053436300067, "tokens_seen": 2066677760 }, { "epoch": 0.25, "learning_rate": 7.549350024073183e-05, "loss": 2.4659, "theoretical_loss": 3.4210354630799706, "tokens_seen": 2066808832 }, { "epoch": 0.25, "learning_rate": 7.548547584657358e-05, "loss": 2.5927, "theoretical_loss": 3.421017491318782, "tokens_seen": 2066939904 }, { "epoch": 0.25, "learning_rate": 7.547745145241535e-05, "loss": 2.6497, "theoretical_loss": 3.4209995210162907, "tokens_seen": 2067070976 }, { "epoch": 0.25, "learning_rate": 7.54694270582571e-05, "loss": 2.6517, "theoretical_loss": 3.420981552172285, "tokens_seen": 2067202048 }, { "epoch": 0.25, "learning_rate": 7.546140266409887e-05, "loss": 2.595, "theoretical_loss": 3.4209635847865547, "tokens_seen": 2067333120 }, { "epoch": 0.25, "learning_rate": 7.545337826994062e-05, "loss": 2.6581, "theoretical_loss": 3.420945618858889, "tokens_seen": 2067464192 }, { "epoch": 0.25, "learning_rate": 7.544535387578239e-05, "loss": 2.6037, "theoretical_loss": 3.420927654389077, "tokens_seen": 2067595264 }, { "epoch": 0.25, "learning_rate": 7.543732948162414e-05, "loss": 2.6536, "theoretical_loss": 3.4209096913769073, "tokens_seen": 2067726336 }, { "epoch": 0.25, "objective/train/docs_used": 1138992, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7702243328094482, "objective/train/theoretical_loss": 3.420900710417373, "objective/train/tokens_used": 438250976, "theoretical_loss": 3.420900710417373, "tokens_seen": 2067791872 }, { "epoch": 0.25, "learning_rate": 7.54293050874659e-05, "loss": 2.6081, "theoretical_loss": 3.4208917298221704, "tokens_seen": 2067857408 }, { "epoch": 0.25, "learning_rate": 7.542128069330766e-05, "loss": 2.7191, "theoretical_loss": 3.420873769724655, "tokens_seen": 2067988480 }, { "epoch": 0.25, "learning_rate": 7.541325629914942e-05, "loss": 2.6877, "theoretical_loss": 3.4208558110841505, "tokens_seen": 2068119552 }, { "epoch": 0.25, "learning_rate": 7.540523190499118e-05, "loss": 2.628, "theoretical_loss": 3.420837853900447, "tokens_seen": 2068250624 }, { "epoch": 0.25, "learning_rate": 7.539720751083294e-05, "loss": 2.6661, "theoretical_loss": 3.420819898173333, "tokens_seen": 2068381696 }, { "epoch": 0.25, "learning_rate": 7.53891831166747e-05, "loss": 2.7085, "theoretical_loss": 3.4208019439025987, "tokens_seen": 2068512768 }, { "epoch": 0.25, "learning_rate": 7.538115872251646e-05, "loss": 2.6818, "theoretical_loss": 3.420783991088034, "tokens_seen": 2068643840 }, { "epoch": 0.25, "learning_rate": 7.537313432835821e-05, "loss": 2.6039, "theoretical_loss": 3.420766039729428, "tokens_seen": 2068774912 }, { "epoch": 0.25, "learning_rate": 7.536510993419998e-05, "loss": 2.6936, "theoretical_loss": 3.4207480898265707, "tokens_seen": 2068905984 }, { "epoch": 0.25, "learning_rate": 7.535708554004173e-05, "loss": 2.5812, "theoretical_loss": 3.4207301413792512, "tokens_seen": 2069037056 }, { "epoch": 0.25, "learning_rate": 7.53490611458835e-05, "loss": 2.5736, "theoretical_loss": 3.4207121943872605, "tokens_seen": 2069168128 }, { "epoch": 0.25, "learning_rate": 7.534103675172525e-05, "loss": 2.5398, "theoretical_loss": 3.4206942488503875, "tokens_seen": 2069299200 }, { "epoch": 0.25, "objective/train/docs_used": 1139620, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4647276401519775, "objective/train/theoretical_loss": 3.420676304768422, "objective/train/tokens_used": 439889376, "theoretical_loss": 3.420676304768422, "tokens_seen": 2069430272 }, { "epoch": 0.25, "learning_rate": 7.533301235756701e-05, "loss": 2.4302, "theoretical_loss": 3.420676304768422, "tokens_seen": 2069430272 }, { "epoch": 0.25, "learning_rate": 7.532498796340877e-05, "loss": 2.4488, "theoretical_loss": 3.420658362141154, "tokens_seen": 2069561344 }, { "epoch": 0.25, "learning_rate": 7.531696356925052e-05, "loss": 2.548, "theoretical_loss": 3.4206404209683736, "tokens_seen": 2069692416 }, { "epoch": 0.25, "learning_rate": 7.530893917509229e-05, "loss": 2.6237, "theoretical_loss": 3.4206224812498713, "tokens_seen": 2069823488 }, { "epoch": 0.25, "learning_rate": 7.530091478093404e-05, "loss": 2.6025, "theoretical_loss": 3.420604542985436, "tokens_seen": 2069954560 }, { "epoch": 0.25, "learning_rate": 7.529289038677581e-05, "loss": 2.6769, "theoretical_loss": 3.420586606174859, "tokens_seen": 2070085632 }, { "epoch": 0.25, "learning_rate": 7.528486599261756e-05, "loss": 2.6906, "theoretical_loss": 3.4205686708179295, "tokens_seen": 2070216704 }, { "epoch": 0.25, "learning_rate": 7.527684159845933e-05, "loss": 2.6362, "theoretical_loss": 3.4205507369144383, "tokens_seen": 2070347776 }, { "epoch": 0.25, "learning_rate": 7.526881720430108e-05, "loss": 2.5741, "theoretical_loss": 3.420532804464175, "tokens_seen": 2070478848 }, { "epoch": 0.25, "learning_rate": 7.526079281014283e-05, "loss": 2.5953, "theoretical_loss": 3.4205148734669306, "tokens_seen": 2070609920 }, { "epoch": 0.26, "learning_rate": 7.52527684159846e-05, "loss": 2.5617, "theoretical_loss": 3.420496943922495, "tokens_seen": 2070740992 }, { "epoch": 0.26, "learning_rate": 7.524474402182635e-05, "loss": 2.6418, "theoretical_loss": 3.420479015830658, "tokens_seen": 2070872064 }, { "epoch": 0.26, "learning_rate": 7.523671962766812e-05, "loss": 2.5491, "theoretical_loss": 3.4204610891912104, "tokens_seen": 2071003136 }, { "epoch": 0.26, "objective/train/docs_used": 1140715, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.805562973022461, "objective/train/theoretical_loss": 3.4204521264160674, "objective/train/tokens_used": 441527776, "theoretical_loss": 3.4204521264160674, "tokens_seen": 2071068672 }, { "epoch": 0.26, "learning_rate": 7.522869523350987e-05, "loss": 2.6671, "theoretical_loss": 3.4204431640039434, "tokens_seen": 2071134208 }, { "epoch": 0.26, "learning_rate": 7.522067083935164e-05, "loss": 2.6024, "theoretical_loss": 3.4204252402686466, "tokens_seen": 2071265280 }, { "epoch": 0.26, "learning_rate": 7.521264644519339e-05, "loss": 2.6613, "theoretical_loss": 3.42040731798511, "tokens_seen": 2071396352 }, { "epoch": 0.26, "learning_rate": 7.520462205103515e-05, "loss": 2.5674, "theoretical_loss": 3.4203893971531256, "tokens_seen": 2071527424 }, { "epoch": 0.26, "learning_rate": 7.519659765687691e-05, "loss": 2.6061, "theoretical_loss": 3.420371477772483, "tokens_seen": 2071658496 }, { "epoch": 0.26, "learning_rate": 7.518857326271867e-05, "loss": 2.6192, "theoretical_loss": 3.4203535598429737, "tokens_seen": 2071789568 }, { "epoch": 0.26, "learning_rate": 7.518054886856043e-05, "loss": 2.5439, "theoretical_loss": 3.4203356433643877, "tokens_seen": 2071920640 }, { "epoch": 0.26, "learning_rate": 7.517252447440219e-05, "loss": 2.5913, "theoretical_loss": 3.4203177283365154, "tokens_seen": 2072051712 }, { "epoch": 0.26, "learning_rate": 7.516450008024395e-05, "loss": 2.5747, "theoretical_loss": 3.420299814759148, "tokens_seen": 2072182784 }, { "epoch": 0.26, "learning_rate": 7.51564756860857e-05, "loss": 2.4535, "theoretical_loss": 3.4202819026320768, "tokens_seen": 2072313856 }, { "epoch": 0.26, "learning_rate": 7.514845129192747e-05, "loss": 2.5588, "theoretical_loss": 3.4202639919550926, "tokens_seen": 2072444928 }, { "epoch": 0.26, "learning_rate": 7.514042689776923e-05, "loss": 2.5327, "theoretical_loss": 3.420246082727985, "tokens_seen": 2072576000 }, { "epoch": 0.26, "objective/train/docs_used": 1141285, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8016202449798584, "objective/train/theoretical_loss": 3.4202281749505463, "objective/train/tokens_used": 443166176, "theoretical_loss": 3.4202281749505463, "tokens_seen": 2072707072 }, { "epoch": 0.26, "learning_rate": 7.513240250361098e-05, "loss": 2.6386, "theoretical_loss": 3.4202281749505463, "tokens_seen": 2072707072 }, { "epoch": 0.26, "learning_rate": 7.512437810945275e-05, "loss": 2.6437, "theoretical_loss": 3.420210268622567, "tokens_seen": 2072838144 }, { "epoch": 0.26, "learning_rate": 7.51163537152945e-05, "loss": 2.5769, "theoretical_loss": 3.4201923637438383, "tokens_seen": 2072969216 }, { "epoch": 0.26, "learning_rate": 7.510832932113627e-05, "loss": 2.5697, "theoretical_loss": 3.4201744603141515, "tokens_seen": 2073100288 }, { "epoch": 0.26, "learning_rate": 7.510030492697802e-05, "loss": 2.6594, "theoretical_loss": 3.420156558333297, "tokens_seen": 2073231360 }, { "epoch": 0.26, "learning_rate": 7.509228053281979e-05, "loss": 2.5883, "theoretical_loss": 3.420138657801066, "tokens_seen": 2073362432 }, { "epoch": 0.26, "learning_rate": 7.508425613866154e-05, "loss": 2.6873, "theoretical_loss": 3.4201207587172506, "tokens_seen": 2073493504 }, { "epoch": 0.26, "learning_rate": 7.507623174450329e-05, "loss": 2.4776, "theoretical_loss": 3.4201028610816415, "tokens_seen": 2073624576 }, { "epoch": 0.26, "learning_rate": 7.506820735034506e-05, "loss": 2.6836, "theoretical_loss": 3.4200849648940306, "tokens_seen": 2073755648 }, { "epoch": 0.26, "learning_rate": 7.506018295618681e-05, "loss": 2.4895, "theoretical_loss": 3.420067070154208, "tokens_seen": 2073886720 }, { "epoch": 0.26, "learning_rate": 7.505215856202858e-05, "loss": 2.6642, "theoretical_loss": 3.420049176861966, "tokens_seen": 2074017792 }, { "epoch": 0.26, "learning_rate": 7.504413416787033e-05, "loss": 2.6899, "theoretical_loss": 3.420031285017096, "tokens_seen": 2074148864 }, { "epoch": 0.26, "learning_rate": 7.50361097737121e-05, "loss": 2.5657, "theoretical_loss": 3.4200133946193887, "tokens_seen": 2074279936 }, { "epoch": 0.26, "objective/train/docs_used": 1142545, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.676924228668213, "objective/train/theoretical_loss": 3.420004449963156, "objective/train/tokens_used": 444804576, "theoretical_loss": 3.420004449963156, "tokens_seen": 2074345472 }, { "epoch": 0.26, "learning_rate": 7.502808537955385e-05, "loss": 2.5201, "theoretical_loss": 3.419995505668637, "tokens_seen": 2074411008 }, { "epoch": 0.26, "learning_rate": 7.50200609853956e-05, "loss": 2.8034, "theoretical_loss": 3.419977618164631, "tokens_seen": 2074542080 }, { "epoch": 0.26, "learning_rate": 7.501203659123737e-05, "loss": 2.6983, "theoretical_loss": 3.419959732107163, "tokens_seen": 2074673152 }, { "epoch": 0.26, "learning_rate": 7.500401219707912e-05, "loss": 2.6096, "theoretical_loss": 3.419941847496025, "tokens_seen": 2074804224 }, { "epoch": 0.26, "learning_rate": 7.499598780292089e-05, "loss": 2.5876, "theoretical_loss": 3.4199239643310078, "tokens_seen": 2074935296 }, { "epoch": 0.26, "learning_rate": 7.498796340876264e-05, "loss": 2.69, "theoretical_loss": 3.4199060826119037, "tokens_seen": 2075066368 }, { "epoch": 0.26, "learning_rate": 7.497993901460441e-05, "loss": 2.7312, "theoretical_loss": 3.4198882023385044, "tokens_seen": 2075197440 }, { "epoch": 0.26, "learning_rate": 7.497191462044616e-05, "loss": 2.5642, "theoretical_loss": 3.419870323510602, "tokens_seen": 2075328512 }, { "epoch": 0.26, "learning_rate": 7.496389022628792e-05, "loss": 2.6408, "theoretical_loss": 3.4198524461279876, "tokens_seen": 2075459584 }, { "epoch": 0.26, "learning_rate": 7.495586583212968e-05, "loss": 2.5564, "theoretical_loss": 3.419834570190454, "tokens_seen": 2075590656 }, { "epoch": 0.26, "learning_rate": 7.494784143797144e-05, "loss": 2.6452, "theoretical_loss": 3.4198166956977927, "tokens_seen": 2075721728 }, { "epoch": 0.26, "learning_rate": 7.49398170438132e-05, "loss": 2.7101, "theoretical_loss": 3.4197988226497955, "tokens_seen": 2075852800 }, { "epoch": 0.26, "objective/train/docs_used": 1143739, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 1.8349756002426147, "objective/train/theoretical_loss": 3.4197809510462545, "objective/train/tokens_used": 446442976, "theoretical_loss": 3.4197809510462545, "tokens_seen": 2075983872 }, { "epoch": 0.26, "learning_rate": 7.493179264965496e-05, "loss": 2.4385, "theoretical_loss": 3.4197809510462545, "tokens_seen": 2075983872 }, { "epoch": 0.26, "learning_rate": 7.492376825549672e-05, "loss": 2.5877, "theoretical_loss": 3.4197630808869617, "tokens_seen": 2076114944 }, { "epoch": 0.26, "learning_rate": 7.491574386133848e-05, "loss": 2.6309, "theoretical_loss": 3.41974521217171, "tokens_seen": 2076246016 }, { "epoch": 0.26, "learning_rate": 7.490771946718023e-05, "loss": 2.5745, "theoretical_loss": 3.4197273449002905, "tokens_seen": 2076377088 }, { "epoch": 0.26, "learning_rate": 7.4899695073022e-05, "loss": 2.5886, "theoretical_loss": 3.419709479072496, "tokens_seen": 2076508160 }, { "epoch": 0.26, "learning_rate": 7.489167067886375e-05, "loss": 2.5897, "theoretical_loss": 3.4196916146881184, "tokens_seen": 2076639232 }, { "epoch": 0.26, "learning_rate": 7.488364628470552e-05, "loss": 2.5993, "theoretical_loss": 3.419673751746951, "tokens_seen": 2076770304 }, { "epoch": 0.26, "learning_rate": 7.487562189054727e-05, "loss": 2.5253, "theoretical_loss": 3.4196558902487846, "tokens_seen": 2076901376 }, { "epoch": 0.26, "learning_rate": 7.486759749638904e-05, "loss": 2.5155, "theoretical_loss": 3.419638030193413, "tokens_seen": 2077032448 }, { "epoch": 0.26, "learning_rate": 7.485957310223079e-05, "loss": 2.5878, "theoretical_loss": 3.419620171580627, "tokens_seen": 2077163520 }, { "epoch": 0.26, "learning_rate": 7.485154870807254e-05, "loss": 2.5929, "theoretical_loss": 3.419602314410221, "tokens_seen": 2077294592 }, { "epoch": 0.26, "learning_rate": 7.484352431391431e-05, "loss": 2.5011, "theoretical_loss": 3.4195844586819857, "tokens_seen": 2077425664 }, { "epoch": 0.26, "learning_rate": 7.483549991975606e-05, "loss": 2.5385, "theoretical_loss": 3.419566604395715, "tokens_seen": 2077556736 }, { "epoch": 0.26, "objective/train/docs_used": 1144416, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.198617935180664, "objective/train/theoretical_loss": 3.419557677793251, "objective/train/tokens_used": 448081376, "theoretical_loss": 3.419557677793251, "tokens_seen": 2077622272 }, { "epoch": 0.26, "learning_rate": 7.482747552559783e-05, "loss": 2.745, "theoretical_loss": 3.4195487515512006, "tokens_seen": 2077687808 }, { "epoch": 0.26, "learning_rate": 7.481945113143958e-05, "loss": 2.7385, "theoretical_loss": 3.4195309001482355, "tokens_seen": 2077818880 }, { "epoch": 0.26, "learning_rate": 7.481142673728135e-05, "loss": 2.5999, "theoretical_loss": 3.4195130501866124, "tokens_seen": 2077949952 }, { "epoch": 0.26, "learning_rate": 7.48034023431231e-05, "loss": 2.5005, "theoretical_loss": 3.419495201666124, "tokens_seen": 2078081024 }, { "epoch": 0.26, "learning_rate": 7.479537794896487e-05, "loss": 2.6459, "theoretical_loss": 3.4194773545865633, "tokens_seen": 2078212096 }, { "epoch": 0.26, "learning_rate": 7.478735355480662e-05, "loss": 2.5881, "theoretical_loss": 3.419459508947723, "tokens_seen": 2078343168 }, { "epoch": 0.26, "learning_rate": 7.477932916064837e-05, "loss": 2.5664, "theoretical_loss": 3.419441664749395, "tokens_seen": 2078474240 }, { "epoch": 0.26, "learning_rate": 7.477130476649014e-05, "loss": 2.6161, "theoretical_loss": 3.4194238219913737, "tokens_seen": 2078605312 }, { "epoch": 0.26, "learning_rate": 7.47632803723319e-05, "loss": 2.4891, "theoretical_loss": 3.4194059806734507, "tokens_seen": 2078736384 }, { "epoch": 0.26, "learning_rate": 7.475525597817366e-05, "loss": 2.4439, "theoretical_loss": 3.41938814079542, "tokens_seen": 2078867456 }, { "epoch": 0.26, "learning_rate": 7.474723158401541e-05, "loss": 2.6349, "theoretical_loss": 3.4193703023570743, "tokens_seen": 2078998528 }, { "epoch": 0.26, "learning_rate": 7.473920718985718e-05, "loss": 2.5836, "theoretical_loss": 3.419352465358206, "tokens_seen": 2079129600 }, { "epoch": 0.26, "objective/train/docs_used": 1145721, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.877591133117676, "objective/train/theoretical_loss": 3.4193346297986094, "objective/train/tokens_used": 449719776, "theoretical_loss": 3.4193346297986094, "tokens_seen": 2079260672 }, { "epoch": 0.26, "learning_rate": 7.473118279569893e-05, "loss": 2.5798, "theoretical_loss": 3.4193346297986094, "tokens_seen": 2079260672 }, { "epoch": 0.26, "learning_rate": 7.472315840154069e-05, "loss": 2.6641, "theoretical_loss": 3.4193167956780766, "tokens_seen": 2079391744 }, { "epoch": 0.26, "learning_rate": 7.471513400738245e-05, "loss": 2.5415, "theoretical_loss": 3.4192989629964012, "tokens_seen": 2079522816 }, { "epoch": 0.26, "learning_rate": 7.47071096132242e-05, "loss": 2.5896, "theoretical_loss": 3.419281131753377, "tokens_seen": 2079653888 }, { "epoch": 0.26, "learning_rate": 7.469908521906597e-05, "loss": 2.5036, "theoretical_loss": 3.4192633019487966, "tokens_seen": 2079784960 }, { "epoch": 0.26, "learning_rate": 7.469106082490773e-05, "loss": 2.5731, "theoretical_loss": 3.419245473582453, "tokens_seen": 2079916032 }, { "epoch": 0.26, "learning_rate": 7.468303643074949e-05, "loss": 2.6267, "theoretical_loss": 3.4192276466541403, "tokens_seen": 2080047104 }, { "epoch": 0.26, "learning_rate": 7.467501203659125e-05, "loss": 2.4308, "theoretical_loss": 3.4192098211636512, "tokens_seen": 2080178176 }, { "epoch": 0.26, "learning_rate": 7.4666987642433e-05, "loss": 2.3999, "theoretical_loss": 3.41919199711078, "tokens_seen": 2080309248 }, { "epoch": 0.26, "learning_rate": 7.465896324827477e-05, "loss": 2.392, "theoretical_loss": 3.4191741744953195, "tokens_seen": 2080440320 }, { "epoch": 0.26, "learning_rate": 7.465093885411652e-05, "loss": 2.5587, "theoretical_loss": 3.4191563533170637, "tokens_seen": 2080571392 }, { "epoch": 0.26, "learning_rate": 7.464291445995829e-05, "loss": 2.6061, "theoretical_loss": 3.4191385335758055, "tokens_seen": 2080702464 }, { "epoch": 0.26, "learning_rate": 7.463489006580004e-05, "loss": 2.5783, "theoretical_loss": 3.4191207152713394, "tokens_seen": 2080833536 }, { "epoch": 0.26, "objective/train/docs_used": 1146352, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6832168102264404, "objective/train/theoretical_loss": 3.4191118066578383, "objective/train/tokens_used": 451358176, "theoretical_loss": 3.4191118066578383, "tokens_seen": 2080899072 }, { "epoch": 0.26, "learning_rate": 7.46268656716418e-05, "loss": 2.6432, "theoretical_loss": 3.419102898403459, "tokens_seen": 2080964608 }, { "epoch": 0.26, "learning_rate": 7.461884127748356e-05, "loss": 2.6323, "theoretical_loss": 3.4190850829719563, "tokens_seen": 2081095680 }, { "epoch": 0.26, "learning_rate": 7.461081688332531e-05, "loss": 2.6782, "theoretical_loss": 3.4190672689766277, "tokens_seen": 2081226752 }, { "epoch": 0.26, "learning_rate": 7.460279248916708e-05, "loss": 2.674, "theoretical_loss": 3.419049456417265, "tokens_seen": 2081357824 }, { "epoch": 0.26, "learning_rate": 7.459476809500883e-05, "loss": 2.6025, "theoretical_loss": 3.419031645293663, "tokens_seen": 2081488896 }, { "epoch": 0.26, "learning_rate": 7.45867437008506e-05, "loss": 2.5709, "theoretical_loss": 3.419013835605615, "tokens_seen": 2081619968 }, { "epoch": 0.26, "learning_rate": 7.457871930669235e-05, "loss": 2.6332, "theoretical_loss": 3.4189960273529154, "tokens_seen": 2081751040 }, { "epoch": 0.26, "learning_rate": 7.457069491253412e-05, "loss": 2.4575, "theoretical_loss": 3.4189782205353576, "tokens_seen": 2081882112 }, { "epoch": 0.26, "learning_rate": 7.456267051837587e-05, "loss": 2.4975, "theoretical_loss": 3.4189604151527364, "tokens_seen": 2082013184 }, { "epoch": 0.26, "learning_rate": 7.455464612421762e-05, "loss": 2.4342, "theoretical_loss": 3.418942611204845, "tokens_seen": 2082144256 }, { "epoch": 0.26, "learning_rate": 7.454662173005939e-05, "loss": 2.6555, "theoretical_loss": 3.418924808691478, "tokens_seen": 2082275328 }, { "epoch": 0.26, "learning_rate": 7.453859733590114e-05, "loss": 2.6492, "theoretical_loss": 3.4189070076124293, "tokens_seen": 2082406400 }, { "epoch": 0.26, "objective/train/docs_used": 1147344, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.602205753326416, "objective/train/theoretical_loss": 3.418889207967493, "objective/train/tokens_used": 452996576, "theoretical_loss": 3.418889207967493, "tokens_seen": 2082537472 }, { "epoch": 0.26, "learning_rate": 7.453057294174291e-05, "loss": 2.6904, "theoretical_loss": 3.418889207967493, "tokens_seen": 2082537472 }, { "epoch": 0.26, "learning_rate": 7.452254854758466e-05, "loss": 2.7331, "theoretical_loss": 3.4188714097564636, "tokens_seen": 2082668544 }, { "epoch": 0.26, "learning_rate": 7.451452415342643e-05, "loss": 2.4997, "theoretical_loss": 3.4188536129791354, "tokens_seen": 2082799616 }, { "epoch": 0.26, "learning_rate": 7.450649975926818e-05, "loss": 2.6038, "theoretical_loss": 3.418835817635303, "tokens_seen": 2082930688 }, { "epoch": 0.26, "learning_rate": 7.449847536510994e-05, "loss": 2.5311, "theoretical_loss": 3.41881802372476, "tokens_seen": 2083061760 }, { "epoch": 0.26, "learning_rate": 7.44904509709517e-05, "loss": 2.5913, "theoretical_loss": 3.4188002312473005, "tokens_seen": 2083192832 }, { "epoch": 0.26, "learning_rate": 7.448242657679346e-05, "loss": 2.5041, "theoretical_loss": 3.41878244020272, "tokens_seen": 2083323904 }, { "epoch": 0.26, "learning_rate": 7.447440218263522e-05, "loss": 2.5274, "theoretical_loss": 3.418764650590812, "tokens_seen": 2083454976 }, { "epoch": 0.26, "learning_rate": 7.446637778847698e-05, "loss": 2.6918, "theoretical_loss": 3.4187468624113717, "tokens_seen": 2083586048 }, { "epoch": 0.26, "learning_rate": 7.445835339431873e-05, "loss": 2.6158, "theoretical_loss": 3.4187290756641935, "tokens_seen": 2083717120 }, { "epoch": 0.26, "learning_rate": 7.44503290001605e-05, "loss": 2.7229, "theoretical_loss": 3.418711290349072, "tokens_seen": 2083848192 }, { "epoch": 0.26, "learning_rate": 7.444230460600225e-05, "loss": 2.6351, "theoretical_loss": 3.418693506465801, "tokens_seen": 2083979264 }, { "epoch": 0.26, "learning_rate": 7.443428021184402e-05, "loss": 2.7173, "theoretical_loss": 3.418675724014177, "tokens_seen": 2084110336 }, { "epoch": 0.26, "objective/train/docs_used": 1148002, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8767805099487305, "objective/train/theoretical_loss": 3.418666833325168, "objective/train/tokens_used": 454634976, "theoretical_loss": 3.418666833325168, "tokens_seen": 2084175872 }, { "epoch": 0.26, "learning_rate": 7.442625581768577e-05, "loss": 2.6401, "theoretical_loss": 3.4186579429939927, "tokens_seen": 2084241408 }, { "epoch": 0.26, "learning_rate": 7.441823142352754e-05, "loss": 2.4244, "theoretical_loss": 3.4186401634050445, "tokens_seen": 2084372480 }, { "epoch": 0.26, "learning_rate": 7.441020702936929e-05, "loss": 2.5651, "theoretical_loss": 3.4186223852471262, "tokens_seen": 2084503552 }, { "epoch": 0.26, "learning_rate": 7.440218263521104e-05, "loss": 2.6412, "theoretical_loss": 3.4186046085200332, "tokens_seen": 2084634624 }, { "epoch": 0.26, "learning_rate": 7.439415824105281e-05, "loss": 2.3889, "theoretical_loss": 3.41858683322356, "tokens_seen": 2084765696 }, { "epoch": 0.26, "learning_rate": 7.438613384689456e-05, "loss": 2.5168, "theoretical_loss": 3.4185690593575018, "tokens_seen": 2084896768 }, { "epoch": 0.26, "learning_rate": 7.437810945273633e-05, "loss": 2.6211, "theoretical_loss": 3.418551286921653, "tokens_seen": 2085027840 }, { "epoch": 0.26, "learning_rate": 7.437008505857808e-05, "loss": 2.6815, "theoretical_loss": 3.4185335159158097, "tokens_seen": 2085158912 }, { "epoch": 0.26, "learning_rate": 7.436206066441983e-05, "loss": 2.4658, "theoretical_loss": 3.418515746339766, "tokens_seen": 2085289984 }, { "epoch": 0.26, "learning_rate": 7.43540362702616e-05, "loss": 2.5338, "theoretical_loss": 3.4184979781933174, "tokens_seen": 2085421056 }, { "epoch": 0.26, "learning_rate": 7.434601187610335e-05, "loss": 2.5857, "theoretical_loss": 3.418480211476259, "tokens_seen": 2085552128 }, { "epoch": 0.26, "learning_rate": 7.433798748194512e-05, "loss": 2.5221, "theoretical_loss": 3.4184624461883857, "tokens_seen": 2085683200 }, { "epoch": 0.26, "objective/train/docs_used": 1149018, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5401086807250977, "objective/train/theoretical_loss": 3.4184446823294934, "objective/train/tokens_used": 456273376, "theoretical_loss": 3.4184446823294934, "tokens_seen": 2085814272 }, { "epoch": 0.26, "learning_rate": 7.432996308778687e-05, "loss": 2.5125, "theoretical_loss": 3.4184446823294934, "tokens_seen": 2085814272 }, { "epoch": 0.26, "learning_rate": 7.432193869362864e-05, "loss": 2.5801, "theoretical_loss": 3.4184269198993773, "tokens_seen": 2085945344 }, { "epoch": 0.26, "learning_rate": 7.43139142994704e-05, "loss": 2.4876, "theoretical_loss": 3.4184091588978314, "tokens_seen": 2086076416 }, { "epoch": 0.26, "learning_rate": 7.430588990531215e-05, "loss": 2.8154, "theoretical_loss": 3.418391399324653, "tokens_seen": 2086207488 }, { "epoch": 0.26, "learning_rate": 7.429786551115391e-05, "loss": 2.5371, "theoretical_loss": 3.4183736411796364, "tokens_seen": 2086338560 }, { "epoch": 0.26, "learning_rate": 7.428984111699567e-05, "loss": 2.6052, "theoretical_loss": 3.4183558844625765, "tokens_seen": 2086469632 }, { "epoch": 0.26, "learning_rate": 7.428181672283743e-05, "loss": 2.4728, "theoretical_loss": 3.41833812917327, "tokens_seen": 2086600704 }, { "epoch": 0.26, "learning_rate": 7.427379232867919e-05, "loss": 2.4988, "theoretical_loss": 3.418320375311512, "tokens_seen": 2086731776 }, { "epoch": 0.26, "learning_rate": 7.426576793452094e-05, "loss": 2.7325, "theoretical_loss": 3.4183026228770976, "tokens_seen": 2086862848 }, { "epoch": 0.26, "learning_rate": 7.425774354036271e-05, "loss": 2.6953, "theoretical_loss": 3.4182848718698233, "tokens_seen": 2086993920 }, { "epoch": 0.26, "learning_rate": 7.424971914620446e-05, "loss": 2.5597, "theoretical_loss": 3.4182671222894836, "tokens_seen": 2087124992 }, { "epoch": 0.27, "learning_rate": 7.424169475204623e-05, "loss": 2.5705, "theoretical_loss": 3.4182493741358755, "tokens_seen": 2087256064 }, { "epoch": 0.27, "learning_rate": 7.423367035788798e-05, "loss": 2.5799, "theoretical_loss": 3.4182316274087934, "tokens_seen": 2087387136 }, { "epoch": 0.27, "objective/train/docs_used": 1149534, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.625910997390747, "objective/train/theoretical_loss": 3.4182227545801362, "objective/train/tokens_used": 457911776, "theoretical_loss": 3.4182227545801362, "tokens_seen": 2087452672 }, { "epoch": 0.27, "learning_rate": 7.422564596372975e-05, "loss": 2.5105, "theoretical_loss": 3.418213882108034, "tokens_seen": 2087518208 }, { "epoch": 0.27, "learning_rate": 7.42176215695715e-05, "loss": 2.5787, "theoretical_loss": 3.4181961382333927, "tokens_seen": 2087649280 }, { "epoch": 0.27, "learning_rate": 7.420959717541325e-05, "loss": 2.8012, "theoretical_loss": 3.4181783957846656, "tokens_seen": 2087780352 }, { "epoch": 0.27, "learning_rate": 7.420157278125502e-05, "loss": 2.6249, "theoretical_loss": 3.4181606547616488, "tokens_seen": 2087911424 }, { "epoch": 0.27, "learning_rate": 7.419354838709677e-05, "loss": 2.4211, "theoretical_loss": 3.418142915164138, "tokens_seen": 2088042496 }, { "epoch": 0.27, "learning_rate": 7.418552399293854e-05, "loss": 2.5906, "theoretical_loss": 3.4181251769919285, "tokens_seen": 2088173568 }, { "epoch": 0.27, "learning_rate": 7.417749959878029e-05, "loss": 2.5411, "theoretical_loss": 3.4181074402448175, "tokens_seen": 2088304640 }, { "epoch": 0.27, "learning_rate": 7.416947520462205e-05, "loss": 2.4751, "theoretical_loss": 3.4180897049226004, "tokens_seen": 2088435712 }, { "epoch": 0.27, "learning_rate": 7.416145081046381e-05, "loss": 2.6218, "theoretical_loss": 3.418071971025073, "tokens_seen": 2088566784 }, { "epoch": 0.27, "learning_rate": 7.415342641630557e-05, "loss": 2.6089, "theoretical_loss": 3.4180542385520325, "tokens_seen": 2088697856 }, { "epoch": 0.27, "learning_rate": 7.414540202214733e-05, "loss": 2.607, "theoretical_loss": 3.4180365075032744, "tokens_seen": 2088828928 }, { "epoch": 0.27, "learning_rate": 7.413737762798909e-05, "loss": 2.5198, "theoretical_loss": 3.4180187778785953, "tokens_seen": 2088960000 }, { "epoch": 0.27, "objective/train/docs_used": 1150483, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.777698040008545, "objective/train/theoretical_loss": 3.4180010496777906, "objective/train/tokens_used": 459550176, "theoretical_loss": 3.4180010496777906, "tokens_seen": 2089091072 }, { "epoch": 0.27, "learning_rate": 7.412935323383084e-05, "loss": 2.5472, "theoretical_loss": 3.4180010496777906, "tokens_seen": 2089091072 }, { "epoch": 0.27, "learning_rate": 7.41213288396726e-05, "loss": 2.4747, "theoretical_loss": 3.4179833229006578, "tokens_seen": 2089222144 }, { "epoch": 0.27, "learning_rate": 7.411330444551436e-05, "loss": 2.588, "theoretical_loss": 3.417965597546992, "tokens_seen": 2089353216 }, { "epoch": 0.27, "learning_rate": 7.410528005135612e-05, "loss": 2.6024, "theoretical_loss": 3.4179478736165914, "tokens_seen": 2089484288 }, { "epoch": 0.27, "learning_rate": 7.409725565719788e-05, "loss": 2.6249, "theoretical_loss": 3.4179301511092506, "tokens_seen": 2089615360 }, { "epoch": 0.27, "learning_rate": 7.408923126303964e-05, "loss": 2.5506, "theoretical_loss": 3.417912430024767, "tokens_seen": 2089746432 }, { "epoch": 0.27, "learning_rate": 7.40812068688814e-05, "loss": 2.6243, "theoretical_loss": 3.417894710362937, "tokens_seen": 2089877504 }, { "epoch": 0.27, "learning_rate": 7.407318247472315e-05, "loss": 2.5144, "theoretical_loss": 3.4178769921235572, "tokens_seen": 2090008576 }, { "epoch": 0.27, "learning_rate": 7.406515808056492e-05, "loss": 2.5705, "theoretical_loss": 3.4178592753064243, "tokens_seen": 2090139648 }, { "epoch": 0.27, "learning_rate": 7.405713368640667e-05, "loss": 2.5887, "theoretical_loss": 3.417841559911335, "tokens_seen": 2090270720 }, { "epoch": 0.27, "learning_rate": 7.404910929224844e-05, "loss": 2.55, "theoretical_loss": 3.4178238459380856, "tokens_seen": 2090401792 }, { "epoch": 0.27, "learning_rate": 7.404108489809019e-05, "loss": 2.6119, "theoretical_loss": 3.417806133386473, "tokens_seen": 2090532864 }, { "epoch": 0.27, "learning_rate": 7.403306050393196e-05, "loss": 2.5615, "theoretical_loss": 3.417788422256294, "tokens_seen": 2090663936 }, { "epoch": 0.27, "objective/train/docs_used": 1151030, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4571421146392822, "objective/train/theoretical_loss": 3.417779567224179, "objective/train/tokens_used": 461188576, "theoretical_loss": 3.417779567224179, "tokens_seen": 2090729472 }, { "epoch": 0.27, "learning_rate": 7.402503610977371e-05, "loss": 2.5956, "theoretical_loss": 3.417770712547346, "tokens_seen": 2090795008 }, { "epoch": 0.27, "learning_rate": 7.401701171561546e-05, "loss": 2.6705, "theoretical_loss": 3.417753004259425, "tokens_seen": 2090926080 }, { "epoch": 0.27, "learning_rate": 7.400898732145723e-05, "loss": 2.5801, "theoretical_loss": 3.417735297392328, "tokens_seen": 2091057152 }, { "epoch": 0.27, "learning_rate": 7.400096292729898e-05, "loss": 2.3938, "theoretical_loss": 3.417717591945853, "tokens_seen": 2091188224 }, { "epoch": 0.27, "learning_rate": 7.399293853314075e-05, "loss": 2.5603, "theoretical_loss": 3.4176998879197957, "tokens_seen": 2091319296 }, { "epoch": 0.27, "learning_rate": 7.39849141389825e-05, "loss": 2.653, "theoretical_loss": 3.4176821853139536, "tokens_seen": 2091450368 }, { "epoch": 0.27, "learning_rate": 7.397688974482427e-05, "loss": 2.4816, "theoretical_loss": 3.417664484128124, "tokens_seen": 2091581440 }, { "epoch": 0.27, "learning_rate": 7.396886535066602e-05, "loss": 2.5028, "theoretical_loss": 3.4176467843621037, "tokens_seen": 2091712512 }, { "epoch": 0.27, "learning_rate": 7.396084095650778e-05, "loss": 2.6164, "theoretical_loss": 3.4176290860156904, "tokens_seen": 2091843584 }, { "epoch": 0.27, "learning_rate": 7.395281656234954e-05, "loss": 2.5514, "theoretical_loss": 3.4176113890886803, "tokens_seen": 2091974656 }, { "epoch": 0.27, "learning_rate": 7.39447921681913e-05, "loss": 2.4333, "theoretical_loss": 3.417593693580872, "tokens_seen": 2092105728 }, { "epoch": 0.27, "learning_rate": 7.393676777403306e-05, "loss": 2.5275, "theoretical_loss": 3.4175759994920614, "tokens_seen": 2092236800 }, { "epoch": 0.27, "objective/train/docs_used": 1152201, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.786438226699829, "objective/train/theoretical_loss": 3.4175583068220465, "objective/train/tokens_used": 462826976, "theoretical_loss": 3.4175583068220465, "tokens_seen": 2092367872 }, { "epoch": 0.27, "learning_rate": 7.392874337987482e-05, "loss": 2.5778, "theoretical_loss": 3.4175583068220465, "tokens_seen": 2092367872 }, { "epoch": 0.27, "learning_rate": 7.392071898571658e-05, "loss": 2.4786, "theoretical_loss": 3.4175406155706245, "tokens_seen": 2092498944 }, { "epoch": 0.27, "learning_rate": 7.391269459155834e-05, "loss": 2.4362, "theoretical_loss": 3.4175229257375936, "tokens_seen": 2092630016 }, { "epoch": 0.27, "learning_rate": 7.390467019740009e-05, "loss": 2.5344, "theoretical_loss": 3.41750523732275, "tokens_seen": 2092761088 }, { "epoch": 0.27, "learning_rate": 7.389664580324186e-05, "loss": 2.4331, "theoretical_loss": 3.417487550325892, "tokens_seen": 2092892160 }, { "epoch": 0.27, "learning_rate": 7.388862140908361e-05, "loss": 2.5797, "theoretical_loss": 3.417469864746817, "tokens_seen": 2093023232 }, { "epoch": 0.27, "learning_rate": 7.388059701492537e-05, "loss": 2.7274, "theoretical_loss": 3.417452180585322, "tokens_seen": 2093154304 }, { "epoch": 0.27, "learning_rate": 7.387257262076713e-05, "loss": 2.4391, "theoretical_loss": 3.4174344978412057, "tokens_seen": 2093285376 }, { "epoch": 0.27, "learning_rate": 7.38645482266089e-05, "loss": 2.5441, "theoretical_loss": 3.417416816514265, "tokens_seen": 2093416448 }, { "epoch": 0.27, "learning_rate": 7.385652383245065e-05, "loss": 2.6726, "theoretical_loss": 3.4173991366042973, "tokens_seen": 2093547520 }, { "epoch": 0.27, "learning_rate": 7.384849943829241e-05, "loss": 2.7121, "theoretical_loss": 3.4173814581111017, "tokens_seen": 2093678592 }, { "epoch": 0.27, "learning_rate": 7.384047504413417e-05, "loss": 2.4686, "theoretical_loss": 3.4173637810344744, "tokens_seen": 2093809664 }, { "epoch": 0.27, "learning_rate": 7.383245064997592e-05, "loss": 2.4801, "theoretical_loss": 3.417346105374214, "tokens_seen": 2093940736 }, { "epoch": 0.27, "objective/train/docs_used": 1153513, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.452009439468384, "objective/train/theoretical_loss": 3.4173372680751584, "objective/train/tokens_used": 464465376, "theoretical_loss": 3.4173372680751584, "tokens_seen": 2094006272 }, { "epoch": 0.27, "learning_rate": 7.382442625581769e-05, "loss": 2.5239, "theoretical_loss": 3.4173284311301186, "tokens_seen": 2094071808 }, { "epoch": 0.27, "learning_rate": 7.381640186165944e-05, "loss": 2.5636, "theoretical_loss": 3.4173107583019853, "tokens_seen": 2094202880 }, { "epoch": 0.27, "learning_rate": 7.380837746750121e-05, "loss": 2.6055, "theoretical_loss": 3.4172930868896128, "tokens_seen": 2094333952 }, { "epoch": 0.27, "learning_rate": 7.380035307334296e-05, "loss": 2.4273, "theoretical_loss": 3.4172754168927986, "tokens_seen": 2094465024 }, { "epoch": 0.27, "learning_rate": 7.379232867918473e-05, "loss": 2.5082, "theoretical_loss": 3.417257748311341, "tokens_seen": 2094596096 }, { "epoch": 0.27, "learning_rate": 7.378430428502648e-05, "loss": 2.5065, "theoretical_loss": 3.4172400811450387, "tokens_seen": 2094727168 }, { "epoch": 0.27, "learning_rate": 7.377627989086823e-05, "loss": 2.4834, "theoretical_loss": 3.417222415393688, "tokens_seen": 2094858240 }, { "epoch": 0.27, "learning_rate": 7.376825549671e-05, "loss": 2.5748, "theoretical_loss": 3.4172047510570893, "tokens_seen": 2094989312 }, { "epoch": 0.27, "learning_rate": 7.376023110255175e-05, "loss": 2.5213, "theoretical_loss": 3.417187088135039, "tokens_seen": 2095120384 }, { "epoch": 0.27, "learning_rate": 7.375220670839352e-05, "loss": 2.5495, "theoretical_loss": 3.4171694266273365, "tokens_seen": 2095251456 }, { "epoch": 0.27, "learning_rate": 7.374418231423527e-05, "loss": 2.5869, "theoretical_loss": 3.417151766533779, "tokens_seen": 2095382528 }, { "epoch": 0.27, "learning_rate": 7.373615792007704e-05, "loss": 2.5185, "theoretical_loss": 3.4171341078541664, "tokens_seen": 2095513600 }, { "epoch": 0.27, "objective/train/docs_used": 1154141, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.867384433746338, "objective/train/theoretical_loss": 3.4171164505882956, "objective/train/tokens_used": 466103776, "theoretical_loss": 3.4171164505882956, "tokens_seen": 2095644672 }, { "epoch": 0.27, "learning_rate": 7.372813352591879e-05, "loss": 2.6868, "theoretical_loss": 3.4171164505882956, "tokens_seen": 2095644672 }, { "epoch": 0.27, "learning_rate": 7.372010913176055e-05, "loss": 2.5989, "theoretical_loss": 3.4170987947359652, "tokens_seen": 2095775744 }, { "epoch": 0.27, "learning_rate": 7.371208473760231e-05, "loss": 2.413, "theoretical_loss": 3.4170811402969745, "tokens_seen": 2095906816 }, { "epoch": 0.27, "learning_rate": 7.370406034344407e-05, "loss": 2.5521, "theoretical_loss": 3.417063487271121, "tokens_seen": 2096037888 }, { "epoch": 0.27, "learning_rate": 7.369603594928583e-05, "loss": 2.6069, "theoretical_loss": 3.417045835658204, "tokens_seen": 2096168960 }, { "epoch": 0.27, "learning_rate": 7.368801155512759e-05, "loss": 2.5313, "theoretical_loss": 3.417028185458021, "tokens_seen": 2096300032 }, { "epoch": 0.27, "learning_rate": 7.367998716096935e-05, "loss": 2.5615, "theoretical_loss": 3.4170105366703725, "tokens_seen": 2096431104 }, { "epoch": 0.27, "learning_rate": 7.36719627668111e-05, "loss": 2.5193, "theoretical_loss": 3.416992889295055, "tokens_seen": 2096562176 }, { "epoch": 0.27, "learning_rate": 7.366393837265286e-05, "loss": 2.6729, "theoretical_loss": 3.4169752433318688, "tokens_seen": 2096693248 }, { "epoch": 0.27, "learning_rate": 7.365591397849463e-05, "loss": 2.569, "theoretical_loss": 3.416957598780612, "tokens_seen": 2096824320 }, { "epoch": 0.27, "learning_rate": 7.364788958433638e-05, "loss": 2.5365, "theoretical_loss": 3.416939955641083, "tokens_seen": 2096955392 }, { "epoch": 0.27, "learning_rate": 7.363986519017815e-05, "loss": 2.612, "theoretical_loss": 3.4169223139130813, "tokens_seen": 2097086464 }, { "epoch": 0.27, "learning_rate": 7.36318407960199e-05, "loss": 2.6286, "theoretical_loss": 3.416904673596406, "tokens_seen": 2097217536 }, { "epoch": 0.27, "objective/train/docs_used": 1155185, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.522921085357666, "objective/train/theoretical_loss": 3.416895853967252, "objective/train/tokens_used": 467742176, "theoretical_loss": 3.416895853967252, "tokens_seen": 2097283072 }, { "epoch": 0.27, "learning_rate": 7.362381640186166e-05, "loss": 2.5816, "theoretical_loss": 3.4168870346908546, "tokens_seen": 2097348608 }, { "epoch": 0.27, "learning_rate": 7.361579200770342e-05, "loss": 2.6585, "theoretical_loss": 3.416869397196227, "tokens_seen": 2097479680 }, { "epoch": 0.27, "learning_rate": 7.360776761354517e-05, "loss": 2.5419, "theoretical_loss": 3.416851761112323, "tokens_seen": 2097610752 }, { "epoch": 0.27, "learning_rate": 7.359974321938694e-05, "loss": 2.5072, "theoretical_loss": 3.41683412643894, "tokens_seen": 2097741824 }, { "epoch": 0.27, "learning_rate": 7.359171882522869e-05, "loss": 2.6212, "theoretical_loss": 3.4168164931758778, "tokens_seen": 2097872896 }, { "epoch": 0.27, "learning_rate": 7.358369443107046e-05, "loss": 2.5326, "theoretical_loss": 3.416798861322936, "tokens_seen": 2098003968 }, { "epoch": 0.27, "learning_rate": 7.357567003691221e-05, "loss": 2.4969, "theoretical_loss": 3.416781230879913, "tokens_seen": 2098135040 }, { "epoch": 0.27, "learning_rate": 7.356764564275398e-05, "loss": 2.6687, "theoretical_loss": 3.416763601846608, "tokens_seen": 2098266112 }, { "epoch": 0.27, "learning_rate": 7.355962124859573e-05, "loss": 2.6071, "theoretical_loss": 3.416745974222821, "tokens_seen": 2098397184 }, { "epoch": 0.27, "learning_rate": 7.35515968544375e-05, "loss": 2.3758, "theoretical_loss": 3.41672834800835, "tokens_seen": 2098528256 }, { "epoch": 0.27, "learning_rate": 7.354357246027925e-05, "loss": 2.5369, "theoretical_loss": 3.416710723202996, "tokens_seen": 2098659328 }, { "epoch": 0.27, "learning_rate": 7.3535548066121e-05, "loss": 2.5719, "theoretical_loss": 3.4166930998065568, "tokens_seen": 2098790400 }, { "epoch": 0.27, "objective/train/docs_used": 1155992, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.662437677383423, "objective/train/theoretical_loss": 3.416675477818832, "objective/train/tokens_used": 469380576, "theoretical_loss": 3.416675477818832, "tokens_seen": 2098921472 }, { "epoch": 0.27, "learning_rate": 7.352752367196277e-05, "loss": 2.5635, "theoretical_loss": 3.416675477818832, "tokens_seen": 2098921472 }, { "epoch": 0.27, "learning_rate": 7.351949927780452e-05, "loss": 2.5932, "theoretical_loss": 3.4166578572396222, "tokens_seen": 2099052544 }, { "epoch": 0.27, "learning_rate": 7.351147488364629e-05, "loss": 2.5664, "theoretical_loss": 3.416640238068726, "tokens_seen": 2099183616 }, { "epoch": 0.27, "learning_rate": 7.350345048948804e-05, "loss": 2.681, "theoretical_loss": 3.416622620305943, "tokens_seen": 2099314688 }, { "epoch": 0.27, "learning_rate": 7.349542609532981e-05, "loss": 2.4963, "theoretical_loss": 3.416605003951073, "tokens_seen": 2099445760 }, { "epoch": 0.27, "learning_rate": 7.348740170117156e-05, "loss": 2.552, "theoretical_loss": 3.416587389003915, "tokens_seen": 2099576832 }, { "epoch": 0.27, "learning_rate": 7.347937730701332e-05, "loss": 2.5934, "theoretical_loss": 3.416569775464269, "tokens_seen": 2099707904 }, { "epoch": 0.27, "learning_rate": 7.347135291285508e-05, "loss": 2.6231, "theoretical_loss": 3.416552163331935, "tokens_seen": 2099838976 }, { "epoch": 0.27, "learning_rate": 7.346332851869684e-05, "loss": 2.7223, "theoretical_loss": 3.4165345526067123, "tokens_seen": 2099970048 }, { "epoch": 0.27, "learning_rate": 7.34553041245386e-05, "loss": 2.4176, "theoretical_loss": 3.416516943288401, "tokens_seen": 2100101120 }, { "epoch": 0.27, "learning_rate": 7.344727973038036e-05, "loss": 2.5539, "theoretical_loss": 3.4164993353768005, "tokens_seen": 2100232192 }, { "epoch": 0.27, "learning_rate": 7.343925533622212e-05, "loss": 2.4916, "theoretical_loss": 3.4164817288717106, "tokens_seen": 2100363264 }, { "epoch": 0.27, "learning_rate": 7.343123094206388e-05, "loss": 2.5582, "theoretical_loss": 3.416464123772932, "tokens_seen": 2100494336 }, { "epoch": 0.27, "objective/train/docs_used": 1157390, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.862727642059326, "objective/train/theoretical_loss": 3.4164553217508464, "objective/train/tokens_used": 471018976, "theoretical_loss": 3.4164553217508464, "tokens_seen": 2100559872 }, { "epoch": 0.27, "learning_rate": 7.342320654790563e-05, "loss": 2.6738, "theoretical_loss": 3.416446520080264, "tokens_seen": 2100625408 }, { "epoch": 0.27, "learning_rate": 7.34151821537474e-05, "loss": 2.4995, "theoretical_loss": 3.4164289177935063, "tokens_seen": 2100756480 }, { "epoch": 0.27, "learning_rate": 7.340715775958915e-05, "loss": 2.5321, "theoretical_loss": 3.4164113169124595, "tokens_seen": 2100887552 }, { "epoch": 0.27, "learning_rate": 7.339913336543092e-05, "loss": 2.392, "theoretical_loss": 3.4163937174369234, "tokens_seen": 2101018624 }, { "epoch": 0.27, "learning_rate": 7.339110897127267e-05, "loss": 2.5493, "theoretical_loss": 3.4163761193666975, "tokens_seen": 2101149696 }, { "epoch": 0.27, "learning_rate": 7.338308457711443e-05, "loss": 2.6421, "theoretical_loss": 3.416358522701583, "tokens_seen": 2101280768 }, { "epoch": 0.27, "learning_rate": 7.337506018295619e-05, "loss": 2.4459, "theoretical_loss": 3.4163409274413796, "tokens_seen": 2101411840 }, { "epoch": 0.27, "learning_rate": 7.336703578879794e-05, "loss": 2.5705, "theoretical_loss": 3.416323333585888, "tokens_seen": 2101542912 }, { "epoch": 0.27, "learning_rate": 7.335901139463971e-05, "loss": 2.6294, "theoretical_loss": 3.416305741134907, "tokens_seen": 2101673984 }, { "epoch": 0.27, "learning_rate": 7.335098700048146e-05, "loss": 2.4931, "theoretical_loss": 3.4162881500882385, "tokens_seen": 2101805056 }, { "epoch": 0.27, "learning_rate": 7.334296260632323e-05, "loss": 2.6335, "theoretical_loss": 3.416270560445682, "tokens_seen": 2101936128 }, { "epoch": 0.27, "learning_rate": 7.333493821216498e-05, "loss": 2.42, "theoretical_loss": 3.4162529722070385, "tokens_seen": 2102067200 }, { "epoch": 0.27, "objective/train/docs_used": 1158091, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.529240369796753, "objective/train/theoretical_loss": 3.416235385372107, "objective/train/tokens_used": 472657376, "theoretical_loss": 3.416235385372107, "tokens_seen": 2102198272 }, { "epoch": 0.27, "learning_rate": 7.332691381800675e-05, "loss": 2.4835, "theoretical_loss": 3.416235385372107, "tokens_seen": 2102198272 }, { "epoch": 0.27, "learning_rate": 7.33188894238485e-05, "loss": 2.739, "theoretical_loss": 3.41621779994069, "tokens_seen": 2102329344 }, { "epoch": 0.27, "learning_rate": 7.331086502969025e-05, "loss": 2.4934, "theoretical_loss": 3.4162002159125864, "tokens_seen": 2102460416 }, { "epoch": 0.27, "learning_rate": 7.330284063553202e-05, "loss": 2.552, "theoretical_loss": 3.4161826332875975, "tokens_seen": 2102591488 }, { "epoch": 0.27, "learning_rate": 7.329481624137377e-05, "loss": 2.4815, "theoretical_loss": 3.4161650520655233, "tokens_seen": 2102722560 }, { "epoch": 0.27, "learning_rate": 7.328679184721554e-05, "loss": 2.5107, "theoretical_loss": 3.4161474722461653, "tokens_seen": 2102853632 }, { "epoch": 0.27, "learning_rate": 7.32787674530573e-05, "loss": 2.6167, "theoretical_loss": 3.416129893829323, "tokens_seen": 2102984704 }, { "epoch": 0.27, "learning_rate": 7.327074305889906e-05, "loss": 2.5377, "theoretical_loss": 3.4161123168147984, "tokens_seen": 2103115776 }, { "epoch": 0.27, "learning_rate": 7.326271866474081e-05, "loss": 2.4288, "theoretical_loss": 3.416094741202391, "tokens_seen": 2103246848 }, { "epoch": 0.27, "learning_rate": 7.325469427058258e-05, "loss": 2.3633, "theoretical_loss": 3.416077166991903, "tokens_seen": 2103377920 }, { "epoch": 0.27, "learning_rate": 7.324666987642433e-05, "loss": 2.5235, "theoretical_loss": 3.416059594183134, "tokens_seen": 2103508992 }, { "epoch": 0.27, "learning_rate": 7.323864548226609e-05, "loss": 2.5012, "theoretical_loss": 3.416042022775885, "tokens_seen": 2103640064 }, { "epoch": 0.28, "learning_rate": 7.323062108810785e-05, "loss": 2.6528, "theoretical_loss": 3.4160244527699577, "tokens_seen": 2103771136 }, { "epoch": 0.28, "objective/train/docs_used": 1159419, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5441555976867676, "objective/train/theoretical_loss": 3.4160156682924274, "objective/train/tokens_used": 474295776, "theoretical_loss": 3.4160156682924274, "tokens_seen": 2103836672 }, { "epoch": 0.28, "learning_rate": 7.32225966939496e-05, "loss": 2.5143, "theoretical_loss": 3.416006884165152, "tokens_seen": 2103902208 }, { "epoch": 0.28, "learning_rate": 7.321457229979137e-05, "loss": 2.5838, "theoretical_loss": 3.4159893169612703, "tokens_seen": 2104033280 }, { "epoch": 0.28, "learning_rate": 7.320654790563313e-05, "loss": 2.4385, "theoretical_loss": 3.4159717511581116, "tokens_seen": 2104164352 }, { "epoch": 0.28, "learning_rate": 7.319852351147489e-05, "loss": 2.4926, "theoretical_loss": 3.415954186755479, "tokens_seen": 2104295424 }, { "epoch": 0.28, "learning_rate": 7.319049911731665e-05, "loss": 2.4845, "theoretical_loss": 3.4159366237531725, "tokens_seen": 2104426496 }, { "epoch": 0.28, "learning_rate": 7.31824747231584e-05, "loss": 2.5013, "theoretical_loss": 3.4159190621509934, "tokens_seen": 2104557568 }, { "epoch": 0.28, "learning_rate": 7.317445032900017e-05, "loss": 2.5026, "theoretical_loss": 3.415901501948743, "tokens_seen": 2104688640 }, { "epoch": 0.28, "learning_rate": 7.316642593484192e-05, "loss": 2.6106, "theoretical_loss": 3.4158839431462225, "tokens_seen": 2104819712 }, { "epoch": 0.28, "learning_rate": 7.315840154068369e-05, "loss": 2.3583, "theoretical_loss": 3.415866385743233, "tokens_seen": 2104950784 }, { "epoch": 0.28, "learning_rate": 7.315037714652544e-05, "loss": 2.5919, "theoretical_loss": 3.415848829739576, "tokens_seen": 2105081856 }, { "epoch": 0.28, "learning_rate": 7.31423527523672e-05, "loss": 2.559, "theoretical_loss": 3.4158312751350532, "tokens_seen": 2105212928 }, { "epoch": 0.28, "learning_rate": 7.313432835820896e-05, "loss": 2.5438, "theoretical_loss": 3.4158137219294655, "tokens_seen": 2105344000 }, { "epoch": 0.28, "objective/train/docs_used": 1160366, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4362361431121826, "objective/train/theoretical_loss": 3.415796170122615, "objective/train/tokens_used": 475934176, "theoretical_loss": 3.415796170122615, "tokens_seen": 2105475072 }, { "epoch": 0.28, "learning_rate": 7.312630396405071e-05, "loss": 2.5706, "theoretical_loss": 3.415796170122615, "tokens_seen": 2105475072 }, { "epoch": 0.28, "learning_rate": 7.311827956989248e-05, "loss": 2.4484, "theoretical_loss": 3.415778619714302, "tokens_seen": 2105606144 }, { "epoch": 0.28, "learning_rate": 7.311025517573423e-05, "loss": 2.4628, "theoretical_loss": 3.4157610707043284, "tokens_seen": 2105737216 }, { "epoch": 0.28, "learning_rate": 7.3102230781576e-05, "loss": 2.5434, "theoretical_loss": 3.415743523092497, "tokens_seen": 2105868288 }, { "epoch": 0.28, "learning_rate": 7.309420638741775e-05, "loss": 2.5093, "theoretical_loss": 3.4157259768786075, "tokens_seen": 2105999360 }, { "epoch": 0.28, "learning_rate": 7.308618199325952e-05, "loss": 2.443, "theoretical_loss": 3.415708432062463, "tokens_seen": 2106130432 }, { "epoch": 0.28, "learning_rate": 7.307815759910127e-05, "loss": 2.6816, "theoretical_loss": 3.415690888643865, "tokens_seen": 2106261504 }, { "epoch": 0.28, "learning_rate": 7.307013320494302e-05, "loss": 2.6213, "theoretical_loss": 3.415673346622614, "tokens_seen": 2106392576 }, { "epoch": 0.28, "learning_rate": 7.306210881078479e-05, "loss": 2.6613, "theoretical_loss": 3.415655805998513, "tokens_seen": 2106523648 }, { "epoch": 0.28, "learning_rate": 7.305408441662654e-05, "loss": 2.52, "theoretical_loss": 3.415638266771363, "tokens_seen": 2106654720 }, { "epoch": 0.28, "learning_rate": 7.304606002246831e-05, "loss": 2.6228, "theoretical_loss": 3.415620728940967, "tokens_seen": 2106785792 }, { "epoch": 0.28, "learning_rate": 7.303803562831006e-05, "loss": 2.4382, "theoretical_loss": 3.4156031925071257, "tokens_seen": 2106916864 }, { "epoch": 0.28, "learning_rate": 7.303001123415183e-05, "loss": 2.595, "theoretical_loss": 3.4155856574696415, "tokens_seen": 2107047936 }, { "epoch": 0.28, "objective/train/docs_used": 1161079, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.0612099170684814, "objective/train/theoretical_loss": 3.4155768904744717, "objective/train/tokens_used": 477572576, "theoretical_loss": 3.4155768904744717, "tokens_seen": 2107113472 }, { "epoch": 0.28, "learning_rate": 7.302198683999358e-05, "loss": 2.6113, "theoretical_loss": 3.415568123828317, "tokens_seen": 2107179008 }, { "epoch": 0.28, "learning_rate": 7.301396244583534e-05, "loss": 2.6254, "theoretical_loss": 3.4155505915829525, "tokens_seen": 2107310080 }, { "epoch": 0.28, "learning_rate": 7.30059380516771e-05, "loss": 2.5267, "theoretical_loss": 3.4155330607333516, "tokens_seen": 2107441152 }, { "epoch": 0.28, "learning_rate": 7.299791365751886e-05, "loss": 2.4114, "theoretical_loss": 3.4155155312793157, "tokens_seen": 2107572224 }, { "epoch": 0.28, "learning_rate": 7.298988926336062e-05, "loss": 2.5213, "theoretical_loss": 3.4154980032206472, "tokens_seen": 2107703296 }, { "epoch": 0.28, "learning_rate": 7.298186486920238e-05, "loss": 2.6768, "theoretical_loss": 3.415480476557148, "tokens_seen": 2107834368 }, { "epoch": 0.28, "learning_rate": 7.297384047504414e-05, "loss": 2.4979, "theoretical_loss": 3.4154629512886205, "tokens_seen": 2107965440 }, { "epoch": 0.28, "learning_rate": 7.29658160808859e-05, "loss": 2.5837, "theoretical_loss": 3.415445427414867, "tokens_seen": 2108096512 }, { "epoch": 0.28, "learning_rate": 7.295779168672765e-05, "loss": 2.5184, "theoretical_loss": 3.4154279049356897, "tokens_seen": 2108227584 }, { "epoch": 0.28, "learning_rate": 7.294976729256942e-05, "loss": 2.487, "theoretical_loss": 3.4154103838508907, "tokens_seen": 2108358656 }, { "epoch": 0.28, "learning_rate": 7.294174289841117e-05, "loss": 2.5156, "theoretical_loss": 3.415392864160273, "tokens_seen": 2108489728 }, { "epoch": 0.28, "learning_rate": 7.293371850425294e-05, "loss": 2.6961, "theoretical_loss": 3.4153753458636382, "tokens_seen": 2108620800 }, { "epoch": 0.28, "objective/train/docs_used": 1162446, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.430760622024536, "objective/train/theoretical_loss": 3.415357828960789, "objective/train/tokens_used": 479210976, "theoretical_loss": 3.415357828960789, "tokens_seen": 2108751872 }, { "epoch": 0.28, "learning_rate": 7.292569411009469e-05, "loss": 2.5458, "theoretical_loss": 3.415357828960789, "tokens_seen": 2108751872 }, { "epoch": 0.28, "learning_rate": 7.291766971593646e-05, "loss": 2.6747, "theoretical_loss": 3.4153403134515283, "tokens_seen": 2108882944 }, { "epoch": 0.28, "learning_rate": 7.290964532177821e-05, "loss": 2.5972, "theoretical_loss": 3.415322799335658, "tokens_seen": 2109014016 }, { "epoch": 0.28, "learning_rate": 7.290162092761998e-05, "loss": 2.4214, "theoretical_loss": 3.4153052866129814, "tokens_seen": 2109145088 }, { "epoch": 0.28, "learning_rate": 7.289359653346173e-05, "loss": 2.5883, "theoretical_loss": 3.4152877752833, "tokens_seen": 2109276160 }, { "epoch": 0.28, "learning_rate": 7.288557213930348e-05, "loss": 2.6023, "theoretical_loss": 3.4152702653464178, "tokens_seen": 2109407232 }, { "epoch": 0.28, "learning_rate": 7.287754774514525e-05, "loss": 2.5545, "theoretical_loss": 3.4152527568021362, "tokens_seen": 2109538304 }, { "epoch": 0.28, "learning_rate": 7.2869523350987e-05, "loss": 2.5484, "theoretical_loss": 3.4152352496502596, "tokens_seen": 2109669376 }, { "epoch": 0.28, "learning_rate": 7.286149895682877e-05, "loss": 2.5416, "theoretical_loss": 3.4152177438905893, "tokens_seen": 2109800448 }, { "epoch": 0.28, "learning_rate": 7.285347456267052e-05, "loss": 2.501, "theoretical_loss": 3.415200239522928, "tokens_seen": 2109931520 }, { "epoch": 0.28, "learning_rate": 7.284545016851229e-05, "loss": 2.663, "theoretical_loss": 3.4151827365470795, "tokens_seen": 2110062592 }, { "epoch": 0.28, "learning_rate": 7.283742577435404e-05, "loss": 2.5081, "theoretical_loss": 3.4151652349628465, "tokens_seen": 2110193664 }, { "epoch": 0.28, "learning_rate": 7.28294013801958e-05, "loss": 2.4971, "theoretical_loss": 3.415147734770031, "tokens_seen": 2110324736 }, { "epoch": 0.28, "objective/train/docs_used": 1163114, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.412344217300415, "objective/train/theoretical_loss": 3.4151389851953438, "objective/train/tokens_used": 480849376, "theoretical_loss": 3.4151389851953438, "tokens_seen": 2110390272 }, { "epoch": 0.28, "learning_rate": 7.282137698603756e-05, "loss": 2.6189, "theoretical_loss": 3.4151302359684372, "tokens_seen": 2110455808 }, { "epoch": 0.28, "learning_rate": 7.281335259187931e-05, "loss": 2.4918, "theoretical_loss": 3.4151127385578675, "tokens_seen": 2110586880 }, { "epoch": 0.28, "learning_rate": 7.280532819772108e-05, "loss": 2.634, "theoretical_loss": 3.415095242538125, "tokens_seen": 2110717952 }, { "epoch": 0.28, "learning_rate": 7.279730380356283e-05, "loss": 2.5236, "theoretical_loss": 3.415077747909013, "tokens_seen": 2110849024 }, { "epoch": 0.28, "learning_rate": 7.27892794094046e-05, "loss": 2.4392, "theoretical_loss": 3.415060254670334, "tokens_seen": 2110980096 }, { "epoch": 0.28, "learning_rate": 7.278125501524635e-05, "loss": 2.4037, "theoretical_loss": 3.415042762821892, "tokens_seen": 2111111168 }, { "epoch": 0.28, "learning_rate": 7.27732306210881e-05, "loss": 2.5942, "theoretical_loss": 3.41502527236349, "tokens_seen": 2111242240 }, { "epoch": 0.28, "learning_rate": 7.276520622692987e-05, "loss": 2.5675, "theoretical_loss": 3.4150077832949304, "tokens_seen": 2111373312 }, { "epoch": 0.28, "learning_rate": 7.275718183277163e-05, "loss": 2.6567, "theoretical_loss": 3.4149902956160174, "tokens_seen": 2111504384 }, { "epoch": 0.28, "learning_rate": 7.274915743861339e-05, "loss": 2.4766, "theoretical_loss": 3.4149728093265543, "tokens_seen": 2111635456 }, { "epoch": 0.28, "learning_rate": 7.274113304445515e-05, "loss": 2.5684, "theoretical_loss": 3.4149553244263444, "tokens_seen": 2111766528 }, { "epoch": 0.28, "learning_rate": 7.273310865029691e-05, "loss": 2.4782, "theoretical_loss": 3.4149378409151905, "tokens_seen": 2111897600 }, { "epoch": 0.28, "objective/train/docs_used": 1164441, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.561600923538208, "objective/train/theoretical_loss": 3.414920358792897, "objective/train/tokens_used": 482487776, "theoretical_loss": 3.414920358792897, "tokens_seen": 2112028672 }, { "epoch": 0.28, "learning_rate": 7.272508425613867e-05, "loss": 2.6107, "theoretical_loss": 3.414920358792897, "tokens_seen": 2112028672 }, { "epoch": 0.28, "learning_rate": 7.271705986198042e-05, "loss": 2.3715, "theoretical_loss": 3.4149028780592667, "tokens_seen": 2112159744 }, { "epoch": 0.28, "learning_rate": 7.270903546782219e-05, "loss": 2.5158, "theoretical_loss": 3.414885398714103, "tokens_seen": 2112290816 }, { "epoch": 0.28, "learning_rate": 7.270101107366394e-05, "loss": 2.5778, "theoretical_loss": 3.41486792075721, "tokens_seen": 2112421888 }, { "epoch": 0.28, "learning_rate": 7.26929866795057e-05, "loss": 2.5923, "theoretical_loss": 3.4148504441883913, "tokens_seen": 2112552960 }, { "epoch": 0.28, "learning_rate": 7.268496228534746e-05, "loss": 2.5593, "theoretical_loss": 3.414832969007451, "tokens_seen": 2112684032 }, { "epoch": 0.28, "learning_rate": 7.267693789118923e-05, "loss": 2.5111, "theoretical_loss": 3.414815495214191, "tokens_seen": 2112815104 }, { "epoch": 0.28, "learning_rate": 7.266891349703098e-05, "loss": 2.6016, "theoretical_loss": 3.4147980228084167, "tokens_seen": 2112946176 }, { "epoch": 0.28, "learning_rate": 7.266088910287273e-05, "loss": 2.488, "theoretical_loss": 3.414780551789931, "tokens_seen": 2113077248 }, { "epoch": 0.28, "learning_rate": 7.26528647087145e-05, "loss": 2.5142, "theoretical_loss": 3.4147630821585384, "tokens_seen": 2113208320 }, { "epoch": 0.28, "learning_rate": 7.264484031455625e-05, "loss": 2.5584, "theoretical_loss": 3.414745613914042, "tokens_seen": 2113339392 }, { "epoch": 0.28, "learning_rate": 7.263681592039802e-05, "loss": 2.5915, "theoretical_loss": 3.4147281470562465, "tokens_seen": 2113470464 }, { "epoch": 0.28, "learning_rate": 7.262879152623977e-05, "loss": 2.5965, "theoretical_loss": 3.414710681584956, "tokens_seen": 2113601536 }, { "epoch": 0.28, "objective/train/docs_used": 1165048, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.748392105102539, "objective/train/theoretical_loss": 3.4147019493691877, "objective/train/tokens_used": 484126176, "theoretical_loss": 3.4147019493691877, "tokens_seen": 2113667072 }, { "epoch": 0.28, "learning_rate": 7.262076713208154e-05, "loss": 2.5164, "theoretical_loss": 3.414693217499973, "tokens_seen": 2113732608 }, { "epoch": 0.28, "learning_rate": 7.261274273792329e-05, "loss": 2.5338, "theoretical_loss": 3.4146757548011024, "tokens_seen": 2113863680 }, { "epoch": 0.28, "learning_rate": 7.260471834376506e-05, "loss": 2.5395, "theoretical_loss": 3.4146582934881486, "tokens_seen": 2113994752 }, { "epoch": 0.28, "learning_rate": 7.259669394960681e-05, "loss": 2.6317, "theoretical_loss": 3.4146408335609157, "tokens_seen": 2114125824 }, { "epoch": 0.28, "learning_rate": 7.258866955544856e-05, "loss": 2.4815, "theoretical_loss": 3.4146233750192065, "tokens_seen": 2114256896 }, { "epoch": 0.28, "learning_rate": 7.258064516129033e-05, "loss": 2.6, "theoretical_loss": 3.4146059178628265, "tokens_seen": 2114387968 }, { "epoch": 0.28, "learning_rate": 7.257262076713208e-05, "loss": 2.5097, "theoretical_loss": 3.4145884620915803, "tokens_seen": 2114519040 }, { "epoch": 0.28, "learning_rate": 7.256459637297385e-05, "loss": 2.5873, "theoretical_loss": 3.4145710077052707, "tokens_seen": 2114650112 }, { "epoch": 0.28, "learning_rate": 7.25565719788156e-05, "loss": 2.4999, "theoretical_loss": 3.4145535547037023, "tokens_seen": 2114781184 }, { "epoch": 0.28, "learning_rate": 7.254854758465737e-05, "loss": 2.7295, "theoretical_loss": 3.4145361030866805, "tokens_seen": 2114912256 }, { "epoch": 0.28, "learning_rate": 7.254052319049912e-05, "loss": 2.6493, "theoretical_loss": 3.414518652854009, "tokens_seen": 2115043328 }, { "epoch": 0.28, "learning_rate": 7.253249879634088e-05, "loss": 2.5968, "theoretical_loss": 3.414501204005492, "tokens_seen": 2115174400 }, { "epoch": 0.28, "objective/train/docs_used": 1166321, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.903796672821045, "objective/train/theoretical_loss": 3.4144837565409336, "objective/train/tokens_used": 485764576, "theoretical_loss": 3.4144837565409336, "tokens_seen": 2115305472 }, { "epoch": 0.28, "learning_rate": 7.252447440218264e-05, "loss": 2.6134, "theoretical_loss": 3.4144837565409336, "tokens_seen": 2115305472 }, { "epoch": 0.28, "learning_rate": 7.25164500080244e-05, "loss": 2.4382, "theoretical_loss": 3.4144663104601394, "tokens_seen": 2115436544 }, { "epoch": 0.28, "learning_rate": 7.250842561386616e-05, "loss": 2.5556, "theoretical_loss": 3.414448865762913, "tokens_seen": 2115567616 }, { "epoch": 0.28, "learning_rate": 7.250040121970792e-05, "loss": 2.5734, "theoretical_loss": 3.4144314224490597, "tokens_seen": 2115698688 }, { "epoch": 0.28, "learning_rate": 7.249237682554968e-05, "loss": 2.4823, "theoretical_loss": 3.4144139805183835, "tokens_seen": 2115829760 }, { "epoch": 0.28, "learning_rate": 7.248435243139144e-05, "loss": 2.659, "theoretical_loss": 3.4143965399706895, "tokens_seen": 2115960832 }, { "epoch": 0.28, "learning_rate": 7.247632803723319e-05, "loss": 2.3849, "theoretical_loss": 3.4143791008057818, "tokens_seen": 2116091904 }, { "epoch": 0.28, "learning_rate": 7.246830364307496e-05, "loss": 2.5886, "theoretical_loss": 3.4143616630234654, "tokens_seen": 2116222976 }, { "epoch": 0.28, "learning_rate": 7.246027924891671e-05, "loss": 2.5187, "theoretical_loss": 3.4143442266235455, "tokens_seen": 2116354048 }, { "epoch": 0.28, "learning_rate": 7.245225485475848e-05, "loss": 2.4991, "theoretical_loss": 3.414326791605826, "tokens_seen": 2116485120 }, { "epoch": 0.28, "learning_rate": 7.244423046060023e-05, "loss": 2.6432, "theoretical_loss": 3.414309357970113, "tokens_seen": 2116616192 }, { "epoch": 0.28, "learning_rate": 7.2436206066442e-05, "loss": 2.5072, "theoretical_loss": 3.41429192571621, "tokens_seen": 2116747264 }, { "epoch": 0.28, "learning_rate": 7.242818167228375e-05, "loss": 2.4161, "theoretical_loss": 3.414274494843923, "tokens_seen": 2116878336 }, { "epoch": 0.28, "objective/train/docs_used": 1166896, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.635124921798706, "objective/train/theoretical_loss": 3.4142657799258243, "objective/train/tokens_used": 487402976, "theoretical_loss": 3.4142657799258243, "tokens_seen": 2116943872 }, { "epoch": 0.28, "learning_rate": 7.24201572781255e-05, "loss": 2.4925, "theoretical_loss": 3.4142570653530564, "tokens_seen": 2117009408 }, { "epoch": 0.28, "learning_rate": 7.241213288396727e-05, "loss": 2.5845, "theoretical_loss": 3.414239637243415, "tokens_seen": 2117140480 }, { "epoch": 0.28, "learning_rate": 7.240410848980902e-05, "loss": 2.4822, "theoretical_loss": 3.414222210514805, "tokens_seen": 2117271552 }, { "epoch": 0.28, "learning_rate": 7.239608409565079e-05, "loss": 2.6126, "theoretical_loss": 3.41420478516703, "tokens_seen": 2117402624 }, { "epoch": 0.28, "learning_rate": 7.238805970149254e-05, "loss": 2.4967, "theoretical_loss": 3.414187361199896, "tokens_seen": 2117533696 }, { "epoch": 0.28, "learning_rate": 7.238003530733431e-05, "loss": 2.5921, "theoretical_loss": 3.414169938613208, "tokens_seen": 2117664768 }, { "epoch": 0.28, "learning_rate": 7.237201091317606e-05, "loss": 2.5172, "theoretical_loss": 3.4141525174067704, "tokens_seen": 2117795840 }, { "epoch": 0.28, "learning_rate": 7.236398651901781e-05, "loss": 2.5096, "theoretical_loss": 3.41413509758039, "tokens_seen": 2117926912 }, { "epoch": 0.28, "learning_rate": 7.235596212485958e-05, "loss": 2.3977, "theoretical_loss": 3.414117679133871, "tokens_seen": 2118057984 }, { "epoch": 0.28, "learning_rate": 7.234793773070133e-05, "loss": 2.5047, "theoretical_loss": 3.414100262067019, "tokens_seen": 2118189056 }, { "epoch": 0.28, "learning_rate": 7.23399133365431e-05, "loss": 2.5251, "theoretical_loss": 3.4140828463796398, "tokens_seen": 2118320128 }, { "epoch": 0.28, "learning_rate": 7.233188894238485e-05, "loss": 2.5052, "theoretical_loss": 3.4140654320715376, "tokens_seen": 2118451200 }, { "epoch": 0.28, "objective/train/docs_used": 1168203, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 1.8257986307144165, "objective/train/theoretical_loss": 3.414048019142519, "objective/train/tokens_used": 489041376, "theoretical_loss": 3.414048019142519, "tokens_seen": 2118582272 }, { "epoch": 0.28, "learning_rate": 7.232386454822662e-05, "loss": 2.3003, "theoretical_loss": 3.414048019142519, "tokens_seen": 2118582272 }, { "epoch": 0.28, "learning_rate": 7.231584015406837e-05, "loss": 2.4695, "theoretical_loss": 3.414030607592389, "tokens_seen": 2118713344 }, { "epoch": 0.28, "learning_rate": 7.230781575991014e-05, "loss": 2.5806, "theoretical_loss": 3.414013197420953, "tokens_seen": 2118844416 }, { "epoch": 0.28, "learning_rate": 7.22997913657519e-05, "loss": 2.6054, "theoretical_loss": 3.413995788628017, "tokens_seen": 2118975488 }, { "epoch": 0.28, "learning_rate": 7.229176697159365e-05, "loss": 2.5005, "theoretical_loss": 3.4139783812133864, "tokens_seen": 2119106560 }, { "epoch": 0.28, "learning_rate": 7.228374257743541e-05, "loss": 2.6139, "theoretical_loss": 3.4139609751768663, "tokens_seen": 2119237632 }, { "epoch": 0.28, "learning_rate": 7.227571818327717e-05, "loss": 2.5732, "theoretical_loss": 3.413943570518263, "tokens_seen": 2119368704 }, { "epoch": 0.28, "learning_rate": 7.226769378911893e-05, "loss": 2.5102, "theoretical_loss": 3.413926167237382, "tokens_seen": 2119499776 }, { "epoch": 0.28, "learning_rate": 7.225966939496069e-05, "loss": 2.4827, "theoretical_loss": 3.413908765334029, "tokens_seen": 2119630848 }, { "epoch": 0.28, "learning_rate": 7.225164500080245e-05, "loss": 2.5705, "theoretical_loss": 3.41389136480801, "tokens_seen": 2119761920 }, { "epoch": 0.28, "learning_rate": 7.22436206066442e-05, "loss": 2.5917, "theoretical_loss": 3.413873965659131, "tokens_seen": 2119892992 }, { "epoch": 0.28, "learning_rate": 7.223559621248596e-05, "loss": 2.3827, "theoretical_loss": 3.4138565678871973, "tokens_seen": 2120024064 }, { "epoch": 0.28, "learning_rate": 7.222757181832773e-05, "loss": 2.4065, "theoretical_loss": 3.4138391714920147, "tokens_seen": 2120155136 }, { "epoch": 0.28, "objective/train/docs_used": 1168773, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.275192975997925, "objective/train/theoretical_loss": 3.413830473810645, "objective/train/tokens_used": 490679776, "theoretical_loss": 3.413830473810645, "tokens_seen": 2120220672 }, { "epoch": 0.29, "learning_rate": 7.221954742416948e-05, "loss": 2.4053, "theoretical_loss": 3.41382177647339, "tokens_seen": 2120286208 }, { "epoch": 0.29, "learning_rate": 7.221152303001125e-05, "loss": 2.4353, "theoretical_loss": 3.413804382831129, "tokens_seen": 2120417280 }, { "epoch": 0.29, "learning_rate": 7.2203498635853e-05, "loss": 2.467, "theoretical_loss": 3.413786990565037, "tokens_seen": 2120548352 }, { "epoch": 0.29, "learning_rate": 7.219547424169477e-05, "loss": 2.5341, "theoretical_loss": 3.4137695996749207, "tokens_seen": 2120679424 }, { "epoch": 0.29, "learning_rate": 7.218744984753652e-05, "loss": 2.3319, "theoretical_loss": 3.413752210160586, "tokens_seen": 2120810496 }, { "epoch": 0.29, "learning_rate": 7.217942545337827e-05, "loss": 2.5447, "theoretical_loss": 3.413734822021839, "tokens_seen": 2120941568 }, { "epoch": 0.29, "learning_rate": 7.217140105922004e-05, "loss": 2.7, "theoretical_loss": 3.413717435258486, "tokens_seen": 2121072640 }, { "epoch": 0.29, "learning_rate": 7.216337666506179e-05, "loss": 2.5168, "theoretical_loss": 3.413700049870333, "tokens_seen": 2121203712 }, { "epoch": 0.29, "learning_rate": 7.215535227090356e-05, "loss": 2.4817, "theoretical_loss": 3.413682665857187, "tokens_seen": 2121334784 }, { "epoch": 0.29, "learning_rate": 7.214732787674531e-05, "loss": 2.5098, "theoretical_loss": 3.4136652832188528, "tokens_seen": 2121465856 }, { "epoch": 0.29, "learning_rate": 7.213930348258708e-05, "loss": 2.5541, "theoretical_loss": 3.4136479019551382, "tokens_seen": 2121596928 }, { "epoch": 0.29, "learning_rate": 7.213127908842883e-05, "loss": 2.3622, "theoretical_loss": 3.4136305220658496, "tokens_seen": 2121728000 }, { "epoch": 0.29, "objective/train/docs_used": 1169526, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4141392707824707, "objective/train/theoretical_loss": 3.4136131435507924, "objective/train/tokens_used": 492318176, "theoretical_loss": 3.4136131435507924, "tokens_seen": 2121859072 }, { "epoch": 0.29, "learning_rate": 7.212325469427058e-05, "loss": 2.4221, "theoretical_loss": 3.4136131435507924, "tokens_seen": 2121859072 }, { "epoch": 0.29, "learning_rate": 7.211523030011235e-05, "loss": 2.4602, "theoretical_loss": 3.413595766409774, "tokens_seen": 2121990144 }, { "epoch": 0.29, "learning_rate": 7.21072059059541e-05, "loss": 2.3975, "theoretical_loss": 3.4135783906425994, "tokens_seen": 2122121216 }, { "epoch": 0.29, "learning_rate": 7.209918151179587e-05, "loss": 2.443, "theoretical_loss": 3.4135610162490773, "tokens_seen": 2122252288 }, { "epoch": 0.29, "learning_rate": 7.209115711763762e-05, "loss": 2.387, "theoretical_loss": 3.4135436432290125, "tokens_seen": 2122383360 }, { "epoch": 0.29, "learning_rate": 7.208313272347939e-05, "loss": 2.696, "theoretical_loss": 3.4135262715822123, "tokens_seen": 2122514432 }, { "epoch": 0.29, "learning_rate": 7.207510832932114e-05, "loss": 2.6137, "theoretical_loss": 3.4135089013084836, "tokens_seen": 2122645504 }, { "epoch": 0.29, "learning_rate": 7.20670839351629e-05, "loss": 2.4994, "theoretical_loss": 3.413491532407633, "tokens_seen": 2122776576 }, { "epoch": 0.29, "learning_rate": 7.205905954100466e-05, "loss": 2.6495, "theoretical_loss": 3.4134741648794664, "tokens_seen": 2122907648 }, { "epoch": 0.29, "learning_rate": 7.205103514684642e-05, "loss": 2.5628, "theoretical_loss": 3.413456798723792, "tokens_seen": 2123038720 }, { "epoch": 0.29, "learning_rate": 7.204301075268818e-05, "loss": 2.5866, "theoretical_loss": 3.4134394339404155, "tokens_seen": 2123169792 }, { "epoch": 0.29, "learning_rate": 7.203498635852994e-05, "loss": 2.5491, "theoretical_loss": 3.413422070529144, "tokens_seen": 2123300864 }, { "epoch": 0.29, "learning_rate": 7.20269619643717e-05, "loss": 2.5522, "theoretical_loss": 3.413404708489785, "tokens_seen": 2123431936 }, { "epoch": 0.29, "objective/train/docs_used": 1170253, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3250832557678223, "objective/train/theoretical_loss": 3.4133960279845117, "objective/train/tokens_used": 493956576, "theoretical_loss": 3.4133960279845117, "tokens_seen": 2123497472 }, { "epoch": 0.29, "learning_rate": 7.201893757021346e-05, "loss": 2.3746, "theoretical_loss": 3.4133873478221446, "tokens_seen": 2123563008 }, { "epoch": 0.29, "learning_rate": 7.201091317605521e-05, "loss": 2.4241, "theoretical_loss": 3.41336998852603, "tokens_seen": 2123694080 }, { "epoch": 0.29, "learning_rate": 7.200288878189698e-05, "loss": 2.6194, "theoretical_loss": 3.4133526306012483, "tokens_seen": 2123825152 }, { "epoch": 0.29, "learning_rate": 7.199486438773873e-05, "loss": 2.455, "theoretical_loss": 3.4133352740476064, "tokens_seen": 2123956224 }, { "epoch": 0.29, "learning_rate": 7.19868399935805e-05, "loss": 2.6033, "theoretical_loss": 3.413317918864912, "tokens_seen": 2124087296 }, { "epoch": 0.29, "learning_rate": 7.197881559942225e-05, "loss": 2.5824, "theoretical_loss": 3.4133005650529715, "tokens_seen": 2124218368 }, { "epoch": 0.29, "learning_rate": 7.197079120526402e-05, "loss": 2.3911, "theoretical_loss": 3.413283212611592, "tokens_seen": 2124349440 }, { "epoch": 0.29, "learning_rate": 7.196276681110577e-05, "loss": 2.5731, "theoretical_loss": 3.413265861540582, "tokens_seen": 2124480512 }, { "epoch": 0.29, "learning_rate": 7.195474241694754e-05, "loss": 2.4229, "theoretical_loss": 3.413248511839747, "tokens_seen": 2124611584 }, { "epoch": 0.29, "learning_rate": 7.194671802278929e-05, "loss": 2.4563, "theoretical_loss": 3.413231163508895, "tokens_seen": 2124742656 }, { "epoch": 0.29, "learning_rate": 7.193869362863104e-05, "loss": 2.3896, "theoretical_loss": 3.4132138165478336, "tokens_seen": 2124873728 }, { "epoch": 0.29, "learning_rate": 7.193066923447281e-05, "loss": 2.5088, "theoretical_loss": 3.41319647095637, "tokens_seen": 2125004800 }, { "epoch": 0.29, "objective/train/docs_used": 1171472, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.178572177886963, "objective/train/theoretical_loss": 3.4131791267343115, "objective/train/tokens_used": 495594976, "theoretical_loss": 3.4131791267343115, "tokens_seen": 2125135872 }, { "epoch": 0.29, "learning_rate": 7.192264484031456e-05, "loss": 2.6308, "theoretical_loss": 3.4131791267343115, "tokens_seen": 2125135872 }, { "epoch": 0.29, "learning_rate": 7.191462044615633e-05, "loss": 2.7299, "theoretical_loss": 3.413161783881465, "tokens_seen": 2125266944 }, { "epoch": 0.29, "learning_rate": 7.190659605199808e-05, "loss": 2.5909, "theoretical_loss": 3.4131444423976394, "tokens_seen": 2125398016 }, { "epoch": 0.29, "learning_rate": 7.189857165783985e-05, "loss": 2.5526, "theoretical_loss": 3.4131271022826413, "tokens_seen": 2125529088 }, { "epoch": 0.29, "learning_rate": 7.18905472636816e-05, "loss": 2.5848, "theoretical_loss": 3.4131097635362777, "tokens_seen": 2125660160 }, { "epoch": 0.29, "learning_rate": 7.188252286952335e-05, "loss": 2.4472, "theoretical_loss": 3.4130924261583573, "tokens_seen": 2125791232 }, { "epoch": 0.29, "learning_rate": 7.187449847536512e-05, "loss": 2.3733, "theoretical_loss": 3.413075090148687, "tokens_seen": 2125922304 }, { "epoch": 0.29, "learning_rate": 7.186647408120687e-05, "loss": 2.59, "theoretical_loss": 3.413057755507075, "tokens_seen": 2126053376 }, { "epoch": 0.29, "learning_rate": 7.185844968704864e-05, "loss": 2.6016, "theoretical_loss": 3.413040422233328, "tokens_seen": 2126184448 }, { "epoch": 0.29, "learning_rate": 7.18504252928904e-05, "loss": 2.4636, "theoretical_loss": 3.413023090327255, "tokens_seen": 2126315520 }, { "epoch": 0.29, "learning_rate": 7.184240089873216e-05, "loss": 2.5986, "theoretical_loss": 3.413005759788663, "tokens_seen": 2126446592 }, { "epoch": 0.29, "learning_rate": 7.183437650457391e-05, "loss": 2.439, "theoretical_loss": 3.412988430617361, "tokens_seen": 2126577664 }, { "epoch": 0.29, "learning_rate": 7.182635211041567e-05, "loss": 2.4675, "theoretical_loss": 3.4129711028131546, "tokens_seen": 2126708736 }, { "epoch": 0.29, "objective/train/docs_used": 1172068, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.215710163116455, "objective/train/theoretical_loss": 3.4129624394236533, "objective/train/tokens_used": 497233376, "theoretical_loss": 3.4129624394236533, "tokens_seen": 2126774272 }, { "epoch": 0.29, "learning_rate": 7.181832771625743e-05, "loss": 2.4494, "theoretical_loss": 3.4129537763758537, "tokens_seen": 2126839808 }, { "epoch": 0.29, "learning_rate": 7.181030332209919e-05, "loss": 2.4569, "theoretical_loss": 3.4129364513052654, "tokens_seen": 2126970880 }, { "epoch": 0.29, "learning_rate": 7.180227892794095e-05, "loss": 2.6965, "theoretical_loss": 3.4129191276011985, "tokens_seen": 2127101952 }, { "epoch": 0.29, "learning_rate": 7.17942545337827e-05, "loss": 2.5618, "theoretical_loss": 3.4129018052634597, "tokens_seen": 2127233024 }, { "epoch": 0.29, "learning_rate": 7.178623013962446e-05, "loss": 2.5163, "theoretical_loss": 3.412884484291858, "tokens_seen": 2127364096 }, { "epoch": 0.29, "learning_rate": 7.177820574546623e-05, "loss": 2.3182, "theoretical_loss": 3.412867164686201, "tokens_seen": 2127495168 }, { "epoch": 0.29, "learning_rate": 7.177018135130798e-05, "loss": 2.6376, "theoretical_loss": 3.4128498464462975, "tokens_seen": 2127626240 }, { "epoch": 0.29, "learning_rate": 7.176215695714975e-05, "loss": 2.5248, "theoretical_loss": 3.412832529571955, "tokens_seen": 2127757312 }, { "epoch": 0.29, "learning_rate": 7.17541325629915e-05, "loss": 2.6077, "theoretical_loss": 3.4128152140629817, "tokens_seen": 2127888384 }, { "epoch": 0.29, "learning_rate": 7.174610816883327e-05, "loss": 2.599, "theoretical_loss": 3.4127978999191866, "tokens_seen": 2128019456 }, { "epoch": 0.29, "learning_rate": 7.173808377467502e-05, "loss": 2.4737, "theoretical_loss": 3.412780587140378, "tokens_seen": 2128150528 }, { "epoch": 0.29, "learning_rate": 7.173005938051677e-05, "loss": 2.668, "theoretical_loss": 3.4127632757263626, "tokens_seen": 2128281600 }, { "epoch": 0.29, "objective/train/docs_used": 1173293, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4033796787261963, "objective/train/theoretical_loss": 3.4127459656769505, "objective/train/tokens_used": 498871776, "theoretical_loss": 3.4127459656769505, "tokens_seen": 2128412672 }, { "epoch": 0.29, "learning_rate": 7.172203498635854e-05, "loss": 2.4953, "theoretical_loss": 3.4127459656769505, "tokens_seen": 2128412672 }, { "epoch": 0.29, "learning_rate": 7.171401059220029e-05, "loss": 2.5535, "theoretical_loss": 3.4127286569919493, "tokens_seen": 2128543744 }, { "epoch": 0.29, "learning_rate": 7.170598619804206e-05, "loss": 2.6874, "theoretical_loss": 3.412711349671168, "tokens_seen": 2128674816 }, { "epoch": 0.29, "learning_rate": 7.169796180388381e-05, "loss": 2.7642, "theoretical_loss": 3.4126940437144144, "tokens_seen": 2128805888 }, { "epoch": 0.29, "learning_rate": 7.168993740972557e-05, "loss": 2.5645, "theoretical_loss": 3.4126767391214976, "tokens_seen": 2128936960 }, { "epoch": 0.29, "learning_rate": 7.168191301556733e-05, "loss": 2.432, "theoretical_loss": 3.412659435892226, "tokens_seen": 2129068032 }, { "epoch": 0.29, "learning_rate": 7.167388862140908e-05, "loss": 2.6999, "theoretical_loss": 3.412642134026408, "tokens_seen": 2129199104 }, { "epoch": 0.29, "learning_rate": 7.166586422725085e-05, "loss": 2.7949, "theoretical_loss": 3.4126248335238527, "tokens_seen": 2129330176 }, { "epoch": 0.29, "learning_rate": 7.16578398330926e-05, "loss": 2.5417, "theoretical_loss": 3.412607534384368, "tokens_seen": 2129461248 }, { "epoch": 0.29, "learning_rate": 7.164981543893436e-05, "loss": 2.44, "theoretical_loss": 3.4125902366077634, "tokens_seen": 2129592320 }, { "epoch": 0.29, "learning_rate": 7.164179104477612e-05, "loss": 2.6147, "theoretical_loss": 3.412572940193847, "tokens_seen": 2129723392 }, { "epoch": 0.29, "learning_rate": 7.163376665061788e-05, "loss": 2.5853, "theoretical_loss": 3.412555645142428, "tokens_seen": 2129854464 }, { "epoch": 0.29, "learning_rate": 7.162574225645964e-05, "loss": 2.5282, "theoretical_loss": 3.4125383514533154, "tokens_seen": 2129985536 }, { "epoch": 0.29, "objective/train/docs_used": 1173914, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.899099111557007, "objective/train/theoretical_loss": 3.412529705119564, "objective/train/tokens_used": 500510176, "theoretical_loss": 3.412529705119564, "tokens_seen": 2130051072 }, { "epoch": 0.29, "learning_rate": 7.16177178623014e-05, "loss": 2.7238, "theoretical_loss": 3.4125210591263175, "tokens_seen": 2130116608 }, { "epoch": 0.29, "learning_rate": 7.160969346814316e-05, "loss": 2.476, "theoretical_loss": 3.412503768161244, "tokens_seen": 2130247680 }, { "epoch": 0.29, "learning_rate": 7.160166907398492e-05, "loss": 2.6226, "theoretical_loss": 3.412486478557903, "tokens_seen": 2130378752 }, { "epoch": 0.29, "learning_rate": 7.159364467982667e-05, "loss": 2.5487, "theoretical_loss": 3.412469190316104, "tokens_seen": 2130509824 }, { "epoch": 0.29, "learning_rate": 7.158562028566844e-05, "loss": 2.613, "theoretical_loss": 3.412451903435656, "tokens_seen": 2130640896 }, { "epoch": 0.29, "learning_rate": 7.157759589151019e-05, "loss": 2.5264, "theoretical_loss": 3.412434617916368, "tokens_seen": 2130771968 }, { "epoch": 0.29, "learning_rate": 7.156957149735196e-05, "loss": 2.5246, "theoretical_loss": 3.4124173337580492, "tokens_seen": 2130903040 }, { "epoch": 0.29, "learning_rate": 7.156154710319371e-05, "loss": 2.5873, "theoretical_loss": 3.412400050960508, "tokens_seen": 2131034112 }, { "epoch": 0.29, "learning_rate": 7.155352270903546e-05, "loss": 2.4861, "theoretical_loss": 3.4123827695235542, "tokens_seen": 2131165184 }, { "epoch": 0.29, "learning_rate": 7.154549831487723e-05, "loss": 2.4371, "theoretical_loss": 3.412365489446998, "tokens_seen": 2131296256 }, { "epoch": 0.29, "learning_rate": 7.153747392071898e-05, "loss": 2.6413, "theoretical_loss": 3.412348210730647, "tokens_seen": 2131427328 }, { "epoch": 0.29, "learning_rate": 7.152944952656075e-05, "loss": 2.63, "theoretical_loss": 3.412330933374311, "tokens_seen": 2131558400 }, { "epoch": 0.29, "objective/train/docs_used": 1174844, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.1257615089416504, "objective/train/theoretical_loss": 3.4123136573777995, "objective/train/tokens_used": 502148576, "theoretical_loss": 3.4123136573777995, "tokens_seen": 2131689472 }, { "epoch": 0.29, "learning_rate": 7.15214251324025e-05, "loss": 2.4612, "theoretical_loss": 3.4123136573777995, "tokens_seen": 2131689472 }, { "epoch": 0.29, "learning_rate": 7.151340073824427e-05, "loss": 2.5941, "theoretical_loss": 3.412296382740922, "tokens_seen": 2131820544 }, { "epoch": 0.29, "learning_rate": 7.150537634408602e-05, "loss": 2.627, "theoretical_loss": 3.412279109463488, "tokens_seen": 2131951616 }, { "epoch": 0.29, "learning_rate": 7.149735194992778e-05, "loss": 2.6469, "theoretical_loss": 3.412261837545307, "tokens_seen": 2132082688 }, { "epoch": 0.29, "learning_rate": 7.148932755576954e-05, "loss": 2.5935, "theoretical_loss": 3.412244566986187, "tokens_seen": 2132213760 }, { "epoch": 0.29, "learning_rate": 7.14813031616113e-05, "loss": 2.6418, "theoretical_loss": 3.4122272977859396, "tokens_seen": 2132344832 }, { "epoch": 0.29, "learning_rate": 7.147327876745306e-05, "loss": 2.3957, "theoretical_loss": 3.412210029944374, "tokens_seen": 2132475904 }, { "epoch": 0.29, "learning_rate": 7.146525437329482e-05, "loss": 2.485, "theoretical_loss": 3.4121927634612987, "tokens_seen": 2132606976 }, { "epoch": 0.29, "learning_rate": 7.145722997913657e-05, "loss": 2.6645, "theoretical_loss": 3.4121754983365236, "tokens_seen": 2132738048 }, { "epoch": 0.29, "learning_rate": 7.144920558497834e-05, "loss": 2.559, "theoretical_loss": 3.4121582345698593, "tokens_seen": 2132869120 }, { "epoch": 0.29, "learning_rate": 7.144118119082009e-05, "loss": 2.4919, "theoretical_loss": 3.4121409721611147, "tokens_seen": 2133000192 }, { "epoch": 0.29, "learning_rate": 7.143315679666185e-05, "loss": 2.6485, "theoretical_loss": 3.4121237111101, "tokens_seen": 2133131264 }, { "epoch": 0.29, "learning_rate": 7.142513240250361e-05, "loss": 2.671, "theoretical_loss": 3.4121064514166246, "tokens_seen": 2133262336 }, { "epoch": 0.29, "objective/train/docs_used": 1175941, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.085177183151245, "objective/train/theoretical_loss": 3.4120978220789047, "objective/train/tokens_used": 503786976, "theoretical_loss": 3.4120978220789047, "tokens_seen": 2133327872 }, { "epoch": 0.29, "learning_rate": 7.141710800834537e-05, "loss": 2.5607, "theoretical_loss": 3.412089193080498, "tokens_seen": 2133393408 }, { "epoch": 0.29, "learning_rate": 7.140908361418713e-05, "loss": 2.4615, "theoretical_loss": 3.4120719361015315, "tokens_seen": 2133524480 }, { "epoch": 0.29, "learning_rate": 7.140105922002888e-05, "loss": 2.628, "theoretical_loss": 3.4120546804795335, "tokens_seen": 2133655552 }, { "epoch": 0.29, "learning_rate": 7.139303482587065e-05, "loss": 2.5597, "theoretical_loss": 3.412037426214315, "tokens_seen": 2133786624 }, { "epoch": 0.29, "learning_rate": 7.13850104317124e-05, "loss": 2.605, "theoretical_loss": 3.4120201733056854, "tokens_seen": 2133917696 }, { "epoch": 0.29, "learning_rate": 7.137698603755417e-05, "loss": 2.7064, "theoretical_loss": 3.412002921753455, "tokens_seen": 2134048768 }, { "epoch": 0.29, "learning_rate": 7.136896164339592e-05, "loss": 2.4825, "theoretical_loss": 3.4119856715574333, "tokens_seen": 2134179840 }, { "epoch": 0.29, "learning_rate": 7.136093724923767e-05, "loss": 2.6736, "theoretical_loss": 3.411968422717431, "tokens_seen": 2134310912 }, { "epoch": 0.29, "learning_rate": 7.135291285507944e-05, "loss": 2.6355, "theoretical_loss": 3.4119511752332583, "tokens_seen": 2134441984 }, { "epoch": 0.29, "learning_rate": 7.13448884609212e-05, "loss": 2.6273, "theoretical_loss": 3.411933929104725, "tokens_seen": 2134573056 }, { "epoch": 0.29, "learning_rate": 7.133686406676296e-05, "loss": 2.4734, "theoretical_loss": 3.4119166843316413, "tokens_seen": 2134704128 }, { "epoch": 0.29, "learning_rate": 7.132883967260471e-05, "loss": 2.5344, "theoretical_loss": 3.411899440913818, "tokens_seen": 2134835200 }, { "epoch": 0.29, "objective/train/docs_used": 1176552, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7227025032043457, "objective/train/theoretical_loss": 3.411882198851065, "objective/train/tokens_used": 505425376, "theoretical_loss": 3.411882198851065, "tokens_seen": 2134966272 }, { "epoch": 0.29, "learning_rate": 7.132081527844648e-05, "loss": 2.5944, "theoretical_loss": 3.411882198851065, "tokens_seen": 2134966272 }, { "epoch": 0.29, "learning_rate": 7.131279088428823e-05, "loss": 2.5057, "theoretical_loss": 3.411864958143192, "tokens_seen": 2135097344 }, { "epoch": 0.29, "learning_rate": 7.130476649013e-05, "loss": 2.6437, "theoretical_loss": 3.4118477187900105, "tokens_seen": 2135228416 }, { "epoch": 0.29, "learning_rate": 7.129674209597175e-05, "loss": 2.5914, "theoretical_loss": 3.41183048079133, "tokens_seen": 2135359488 }, { "epoch": 0.29, "learning_rate": 7.12887177018135e-05, "loss": 2.6825, "theoretical_loss": 3.4118132441469617, "tokens_seen": 2135490560 }, { "epoch": 0.29, "learning_rate": 7.128069330765527e-05, "loss": 2.5187, "theoretical_loss": 3.4117960088567156, "tokens_seen": 2135621632 }, { "epoch": 0.29, "learning_rate": 7.127266891349703e-05, "loss": 2.599, "theoretical_loss": 3.4117787749204025, "tokens_seen": 2135752704 }, { "epoch": 0.29, "learning_rate": 7.126464451933879e-05, "loss": 2.5976, "theoretical_loss": 3.4117615423378327, "tokens_seen": 2135883776 }, { "epoch": 0.29, "learning_rate": 7.125662012518055e-05, "loss": 2.543, "theoretical_loss": 3.4117443111088175, "tokens_seen": 2136014848 }, { "epoch": 0.29, "learning_rate": 7.124859573102231e-05, "loss": 2.6293, "theoretical_loss": 3.411727081233166, "tokens_seen": 2136145920 }, { "epoch": 0.29, "learning_rate": 7.124057133686407e-05, "loss": 2.4895, "theoretical_loss": 3.41170985271069, "tokens_seen": 2136276992 }, { "epoch": 0.29, "learning_rate": 7.123254694270582e-05, "loss": 2.4808, "theoretical_loss": 3.4116926255412006, "tokens_seen": 2136408064 }, { "epoch": 0.29, "learning_rate": 7.122452254854759e-05, "loss": 2.6091, "theoretical_loss": 3.4116753997245075, "tokens_seen": 2136539136 }, { "epoch": 0.29, "objective/train/docs_used": 1177438, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7546355724334717, "objective/train/theoretical_loss": 3.4116667873234006, "objective/train/tokens_used": 507063776, "theoretical_loss": 3.4116667873234006, "tokens_seen": 2136604672 }, { "epoch": 0.29, "learning_rate": 7.121649815438934e-05, "loss": 2.6309, "theoretical_loss": 3.4116581752604223, "tokens_seen": 2136670208 }, { "epoch": 0.3, "learning_rate": 7.12084737602311e-05, "loss": 2.6768, "theoretical_loss": 3.411640952148755, "tokens_seen": 2136801280 }, { "epoch": 0.3, "learning_rate": 7.120044936607286e-05, "loss": 2.6746, "theoretical_loss": 3.4116237303893175, "tokens_seen": 2136932352 }, { "epoch": 0.3, "learning_rate": 7.119242497191462e-05, "loss": 2.5923, "theoretical_loss": 3.4116065099819197, "tokens_seen": 2137063424 }, { "epoch": 0.3, "learning_rate": 7.118440057775638e-05, "loss": 2.6047, "theoretical_loss": 3.411589290926374, "tokens_seen": 2137194496 }, { "epoch": 0.3, "learning_rate": 7.117637618359813e-05, "loss": 2.61, "theoretical_loss": 3.411572073222489, "tokens_seen": 2137325568 }, { "epoch": 0.3, "learning_rate": 7.11683517894399e-05, "loss": 2.686, "theoretical_loss": 3.4115548568700778, "tokens_seen": 2137456640 }, { "epoch": 0.3, "learning_rate": 7.116032739528165e-05, "loss": 2.5895, "theoretical_loss": 3.411537641868951, "tokens_seen": 2137587712 }, { "epoch": 0.3, "learning_rate": 7.115230300112342e-05, "loss": 2.5471, "theoretical_loss": 3.411520428218919, "tokens_seen": 2137718784 }, { "epoch": 0.3, "learning_rate": 7.114427860696517e-05, "loss": 2.399, "theoretical_loss": 3.4115032159197938, "tokens_seen": 2137849856 }, { "epoch": 0.3, "learning_rate": 7.113625421280694e-05, "loss": 2.5753, "theoretical_loss": 3.4114860049713855, "tokens_seen": 2137980928 }, { "epoch": 0.3, "learning_rate": 7.112822981864869e-05, "loss": 2.725, "theoretical_loss": 3.4114687953735063, "tokens_seen": 2138112000 }, { "epoch": 0.3, "objective/train/docs_used": 1177662, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.830423593521118, "objective/train/theoretical_loss": 3.411451587125967, "objective/train/tokens_used": 508702176, "theoretical_loss": 3.411451587125967, "tokens_seen": 2138243072 }, { "epoch": 0.3, "learning_rate": 7.112020542449044e-05, "loss": 2.6082, "theoretical_loss": 3.411451587125967, "tokens_seen": 2138243072 }, { "epoch": 0.3, "learning_rate": 7.111218103033221e-05, "loss": 2.5201, "theoretical_loss": 3.411434380228579, "tokens_seen": 2138374144 }, { "epoch": 0.3, "learning_rate": 7.110415663617396e-05, "loss": 2.5066, "theoretical_loss": 3.4114171746811537, "tokens_seen": 2138505216 }, { "epoch": 0.3, "learning_rate": 7.109613224201573e-05, "loss": 2.6115, "theoretical_loss": 3.4113999704835023, "tokens_seen": 2138636288 }, { "epoch": 0.3, "learning_rate": 7.108810784785748e-05, "loss": 2.6089, "theoretical_loss": 3.4113827676354362, "tokens_seen": 2138767360 }, { "epoch": 0.3, "learning_rate": 7.108008345369925e-05, "loss": 2.7135, "theoretical_loss": 3.411365566136767, "tokens_seen": 2138898432 }, { "epoch": 0.3, "learning_rate": 7.1072059059541e-05, "loss": 2.6082, "theoretical_loss": 3.4113483659873056, "tokens_seen": 2139029504 }, { "epoch": 0.3, "learning_rate": 7.106403466538276e-05, "loss": 2.597, "theoretical_loss": 3.4113311671868645, "tokens_seen": 2139160576 }, { "epoch": 0.3, "learning_rate": 7.105601027122452e-05, "loss": 2.7267, "theoretical_loss": 3.4113139697352546, "tokens_seen": 2139291648 }, { "epoch": 0.3, "learning_rate": 7.104798587706628e-05, "loss": 2.533, "theoretical_loss": 3.4112967736322872, "tokens_seen": 2139422720 }, { "epoch": 0.3, "learning_rate": 7.103996148290804e-05, "loss": 2.5938, "theoretical_loss": 3.4112795788777746, "tokens_seen": 2139553792 }, { "epoch": 0.3, "learning_rate": 7.10319370887498e-05, "loss": 2.6091, "theoretical_loss": 3.4112623854715283, "tokens_seen": 2139684864 }, { "epoch": 0.3, "learning_rate": 7.102391269459156e-05, "loss": 2.6935, "theoretical_loss": 3.41124519341336, "tokens_seen": 2139815936 }, { "epoch": 0.3, "objective/train/docs_used": 1178610, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.592621088027954, "objective/train/theoretical_loss": 3.4112365978897463, "objective/train/tokens_used": 510340576, "theoretical_loss": 3.4112365978897463, "tokens_seen": 2139881472 }, { "epoch": 0.3, "learning_rate": 7.101588830043332e-05, "loss": 2.612, "theoretical_loss": 3.411228002703081, "tokens_seen": 2139947008 }, { "epoch": 0.3, "learning_rate": 7.100786390627508e-05, "loss": 2.6124, "theoretical_loss": 3.4112108133405035, "tokens_seen": 2140078080 }, { "epoch": 0.3, "learning_rate": 7.099983951211684e-05, "loss": 2.5644, "theoretical_loss": 3.4111936253254393, "tokens_seen": 2140209152 }, { "epoch": 0.3, "learning_rate": 7.099181511795859e-05, "loss": 2.6921, "theoretical_loss": 3.4111764386576997, "tokens_seen": 2140340224 }, { "epoch": 0.3, "learning_rate": 7.098379072380036e-05, "loss": 2.5306, "theoretical_loss": 3.4111592533370976, "tokens_seen": 2140471296 }, { "epoch": 0.3, "learning_rate": 7.097576632964211e-05, "loss": 2.6231, "theoretical_loss": 3.4111420693634438, "tokens_seen": 2140602368 }, { "epoch": 0.3, "learning_rate": 7.096774193548388e-05, "loss": 2.5484, "theoretical_loss": 3.411124886736551, "tokens_seen": 2140733440 }, { "epoch": 0.3, "learning_rate": 7.095971754132563e-05, "loss": 2.5626, "theoretical_loss": 3.4111077054562315, "tokens_seen": 2140864512 }, { "epoch": 0.3, "learning_rate": 7.09516931471674e-05, "loss": 2.5805, "theoretical_loss": 3.4110905255222965, "tokens_seen": 2140995584 }, { "epoch": 0.3, "learning_rate": 7.094366875300915e-05, "loss": 2.5912, "theoretical_loss": 3.4110733469345584, "tokens_seen": 2141126656 }, { "epoch": 0.3, "learning_rate": 7.09356443588509e-05, "loss": 2.5502, "theoretical_loss": 3.4110561696928294, "tokens_seen": 2141257728 }, { "epoch": 0.3, "learning_rate": 7.092761996469267e-05, "loss": 2.6747, "theoretical_loss": 3.4110389937969217, "tokens_seen": 2141388800 }, { "debugging/Self-BLEU-5": 0.3562156871264047, "debugging/distinct-1-grams": 0.8053620964343855, "debugging/distinct-2-grams": 0.9818306010928961, "debugging/entropy-1-grams": 5.331752904391725, "debugging/entropy-2-grams": 5.892259632920162, "debugging/length": 535.8, "debugging/num_segments": 5, "debugging/score": 0.007634856930105227, "debugging/score_std": 0.0023886908693567656, "epoch": 0.3, "objective/train/docs_used": 1179232, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.219043016433716, "objective/train/theoretical_loss": 3.4110218192466473, "objective/train/tokens_used": 511978976, "theoretical_loss": 3.4110218192466473, "tokens_seen": 2141519872 }, { "epoch": 0.3, "learning_rate": 7.091959557053442e-05, "loss": 2.6092, "theoretical_loss": 3.4110218192466473, "tokens_seen": 2141519872 }, { "epoch": 0.3, "learning_rate": 7.091157117637619e-05, "loss": 2.7279, "theoretical_loss": 3.4110046460418184, "tokens_seen": 2141650944 }, { "epoch": 0.3, "learning_rate": 7.090354678221794e-05, "loss": 2.6784, "theoretical_loss": 3.4109874741822477, "tokens_seen": 2141782016 }, { "epoch": 0.3, "learning_rate": 7.089552238805971e-05, "loss": 2.6052, "theoretical_loss": 3.410970303667747, "tokens_seen": 2141913088 }, { "epoch": 0.3, "learning_rate": 7.088749799390146e-05, "loss": 2.5963, "theoretical_loss": 3.4109531344981283, "tokens_seen": 2142044160 }, { "epoch": 0.3, "learning_rate": 7.087947359974321e-05, "loss": 2.6833, "theoretical_loss": 3.4109359666732053, "tokens_seen": 2142175232 }, { "epoch": 0.3, "learning_rate": 7.087144920558498e-05, "loss": 2.6276, "theoretical_loss": 3.410918800192789, "tokens_seen": 2142306304 }, { "epoch": 0.3, "learning_rate": 7.086342481142673e-05, "loss": 2.5189, "theoretical_loss": 3.4109016350566934, "tokens_seen": 2142437376 }, { "epoch": 0.3, "learning_rate": 7.08554004172685e-05, "loss": 2.5976, "theoretical_loss": 3.410884471264729, "tokens_seen": 2142568448 }, { "epoch": 0.3, "learning_rate": 7.084737602311025e-05, "loss": 2.5985, "theoretical_loss": 3.41086730881671, "tokens_seen": 2142699520 }, { "epoch": 0.3, "learning_rate": 7.083935162895202e-05, "loss": 2.6729, "theoretical_loss": 3.4108501477124484, "tokens_seen": 2142830592 }, { "epoch": 0.3, "learning_rate": 7.083132723479377e-05, "loss": 2.5663, "theoretical_loss": 3.4108329879517565, "tokens_seen": 2142961664 }, { "epoch": 0.3, "learning_rate": 7.082330284063553e-05, "loss": 2.6607, "theoretical_loss": 3.4108158295344473, "tokens_seen": 2143092736 }, { "epoch": 0.3, "objective/train/docs_used": 1180509, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.578857421875, "objective/train/theoretical_loss": 3.4108072508295026, "objective/train/tokens_used": 513617376, "theoretical_loss": 3.4108072508295026, "tokens_seen": 2143158272 }, { "epoch": 0.3, "learning_rate": 7.081527844647729e-05, "loss": 2.563, "theoretical_loss": 3.4107986724603334, "tokens_seen": 2143223808 }, { "epoch": 0.3, "learning_rate": 7.080725405231905e-05, "loss": 2.4593, "theoretical_loss": 3.4107815167292275, "tokens_seen": 2143354880 }, { "epoch": 0.3, "learning_rate": 7.079922965816081e-05, "loss": 2.6285, "theoretical_loss": 3.4107643623409425, "tokens_seen": 2143485952 }, { "epoch": 0.3, "learning_rate": 7.079120526400257e-05, "loss": 2.6234, "theoretical_loss": 3.410747209295291, "tokens_seen": 2143617024 }, { "epoch": 0.3, "learning_rate": 7.078318086984433e-05, "loss": 2.558, "theoretical_loss": 3.4107300575920854, "tokens_seen": 2143748096 }, { "epoch": 0.3, "learning_rate": 7.077515647568609e-05, "loss": 2.4592, "theoretical_loss": 3.4107129072311393, "tokens_seen": 2143879168 }, { "epoch": 0.3, "learning_rate": 7.076713208152784e-05, "loss": 2.6204, "theoretical_loss": 3.410695758212266, "tokens_seen": 2144010240 }, { "epoch": 0.3, "learning_rate": 7.07591076873696e-05, "loss": 2.7057, "theoretical_loss": 3.4106786105352773, "tokens_seen": 2144141312 }, { "epoch": 0.3, "learning_rate": 7.075108329321136e-05, "loss": 2.6211, "theoretical_loss": 3.410661464199986, "tokens_seen": 2144272384 }, { "epoch": 0.3, "learning_rate": 7.074305889905313e-05, "loss": 2.5688, "theoretical_loss": 3.410644319206207, "tokens_seen": 2144403456 }, { "epoch": 0.3, "learning_rate": 7.073503450489488e-05, "loss": 2.6275, "theoretical_loss": 3.4106271755537514, "tokens_seen": 2144534528 }, { "epoch": 0.3, "learning_rate": 7.072701011073665e-05, "loss": 2.5701, "theoretical_loss": 3.410610033242433, "tokens_seen": 2144665600 }, { "epoch": 0.3, "objective/train/docs_used": 1181762, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5804944038391113, "objective/train/theoretical_loss": 3.4105928922720654, "objective/train/tokens_used": 515255776, "theoretical_loss": 3.4105928922720654, "tokens_seen": 2144796672 }, { "epoch": 0.3, "learning_rate": 7.07189857165784e-05, "loss": 2.5957, "theoretical_loss": 3.4105928922720654, "tokens_seen": 2144796672 }, { "epoch": 0.3, "learning_rate": 7.071096132242017e-05, "loss": 2.6133, "theoretical_loss": 3.410575752642461, "tokens_seen": 2144927744 }, { "epoch": 0.3, "learning_rate": 7.070293692826192e-05, "loss": 2.5896, "theoretical_loss": 3.4105586143534334, "tokens_seen": 2145058816 }, { "epoch": 0.3, "learning_rate": 7.069491253410367e-05, "loss": 2.4899, "theoretical_loss": 3.410541477404796, "tokens_seen": 2145189888 }, { "epoch": 0.3, "learning_rate": 7.068688813994544e-05, "loss": 2.6221, "theoretical_loss": 3.4105243417963615, "tokens_seen": 2145320960 }, { "epoch": 0.3, "learning_rate": 7.067886374578719e-05, "loss": 2.593, "theoretical_loss": 3.4105072075279432, "tokens_seen": 2145452032 }, { "epoch": 0.3, "learning_rate": 7.067083935162896e-05, "loss": 2.6936, "theoretical_loss": 3.410490074599356, "tokens_seen": 2145583104 }, { "epoch": 0.3, "learning_rate": 7.066281495747071e-05, "loss": 2.5297, "theoretical_loss": 3.4104729430104115, "tokens_seen": 2145714176 }, { "epoch": 0.3, "learning_rate": 7.065479056331248e-05, "loss": 2.6632, "theoretical_loss": 3.4104558127609232, "tokens_seen": 2145845248 }, { "epoch": 0.3, "learning_rate": 7.064676616915423e-05, "loss": 2.6771, "theoretical_loss": 3.410438683850706, "tokens_seen": 2145976320 }, { "epoch": 0.3, "learning_rate": 7.063874177499598e-05, "loss": 2.667, "theoretical_loss": 3.410421556279572, "tokens_seen": 2146107392 }, { "epoch": 0.3, "learning_rate": 7.063071738083775e-05, "loss": 2.5011, "theoretical_loss": 3.4104044300473353, "tokens_seen": 2146238464 }, { "epoch": 0.3, "learning_rate": 7.06226929866795e-05, "loss": 2.626, "theoretical_loss": 3.4103873051538094, "tokens_seen": 2146369536 }, { "epoch": 0.3, "objective/train/docs_used": 1182263, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.857555866241455, "objective/train/theoretical_loss": 3.410378743209005, "objective/train/tokens_used": 516894176, "theoretical_loss": 3.410378743209005, "tokens_seen": 2146435072 }, { "epoch": 0.3, "learning_rate": 7.061466859252127e-05, "loss": 2.6154, "theoretical_loss": 3.4103701815988083, "tokens_seen": 2146500608 }, { "epoch": 0.3, "learning_rate": 7.060664419836302e-05, "loss": 2.4926, "theoretical_loss": 3.4103530593821447, "tokens_seen": 2146631680 }, { "epoch": 0.3, "learning_rate": 7.059861980420479e-05, "loss": 2.7829, "theoretical_loss": 3.4103359385036334, "tokens_seen": 2146762752 }, { "epoch": 0.3, "learning_rate": 7.059059541004654e-05, "loss": 2.5032, "theoretical_loss": 3.4103188189630878, "tokens_seen": 2146893824 }, { "epoch": 0.3, "learning_rate": 7.05825710158883e-05, "loss": 2.5789, "theoretical_loss": 3.4103017007603205, "tokens_seen": 2147024896 }, { "epoch": 0.3, "learning_rate": 7.057454662173006e-05, "loss": 2.5056, "theoretical_loss": 3.410284583895147, "tokens_seen": 2147155968 }, { "epoch": 0.3, "learning_rate": 7.056652222757182e-05, "loss": 2.5706, "theoretical_loss": 3.4102674683673806, "tokens_seen": 2147287040 }, { "epoch": 0.3, "learning_rate": 7.055849783341358e-05, "loss": 2.6252, "theoretical_loss": 3.4102503541768345, "tokens_seen": 2147418112 }, { "epoch": 0.3, "learning_rate": 7.055047343925534e-05, "loss": 2.5292, "theoretical_loss": 3.4102332413233234, "tokens_seen": 2147549184 }, { "epoch": 0.3, "learning_rate": 7.05424490450971e-05, "loss": 2.6922, "theoretical_loss": 3.4102161298066607, "tokens_seen": 2147680256 }, { "epoch": 0.3, "learning_rate": 7.053442465093886e-05, "loss": 2.6755, "theoretical_loss": 3.410199019626661, "tokens_seen": 2147811328 }, { "epoch": 0.3, "learning_rate": 7.052640025678061e-05, "loss": 2.5742, "theoretical_loss": 3.410181910783138, "tokens_seen": 2147942400 }, { "epoch": 0.3, "objective/train/docs_used": 1183255, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4469528198242188, "objective/train/theoretical_loss": 3.4101648032759053, "objective/train/tokens_used": 518532576, "theoretical_loss": 3.4101648032759053, "tokens_seen": 2148073472 }, { "epoch": 0.3, "learning_rate": 7.051837586262238e-05, "loss": 2.4785, "theoretical_loss": 3.4101648032759053, "tokens_seen": 2148073472 }, { "epoch": 0.3, "learning_rate": 7.051035146846413e-05, "loss": 2.6102, "theoretical_loss": 3.4101476971047777, "tokens_seen": 2148204544 }, { "epoch": 0.3, "learning_rate": 7.05023270743059e-05, "loss": 2.5222, "theoretical_loss": 3.410130592269569, "tokens_seen": 2148335616 }, { "epoch": 0.3, "learning_rate": 7.049430268014765e-05, "loss": 2.665, "theoretical_loss": 3.4101134887700937, "tokens_seen": 2148466688 }, { "epoch": 0.3, "learning_rate": 7.048627828598942e-05, "loss": 2.6828, "theoretical_loss": 3.410096386606166, "tokens_seen": 2148597760 }, { "epoch": 0.3, "learning_rate": 7.047825389183117e-05, "loss": 2.5421, "theoretical_loss": 3.4100792857775994, "tokens_seen": 2148728832 }, { "epoch": 0.3, "learning_rate": 7.047022949767292e-05, "loss": 2.6022, "theoretical_loss": 3.4100621862842084, "tokens_seen": 2148859904 }, { "epoch": 0.3, "learning_rate": 7.046220510351469e-05, "loss": 2.6454, "theoretical_loss": 3.410045088125808, "tokens_seen": 2148990976 }, { "epoch": 0.3, "learning_rate": 7.045418070935644e-05, "loss": 2.5891, "theoretical_loss": 3.4100279913022122, "tokens_seen": 2149122048 }, { "epoch": 0.3, "learning_rate": 7.044615631519821e-05, "loss": 2.6476, "theoretical_loss": 3.410010895813236, "tokens_seen": 2149253120 }, { "epoch": 0.3, "learning_rate": 7.043813192103996e-05, "loss": 2.4586, "theoretical_loss": 3.4099938016586924, "tokens_seen": 2149384192 }, { "epoch": 0.3, "learning_rate": 7.043010752688173e-05, "loss": 2.6862, "theoretical_loss": 3.4099767088383968, "tokens_seen": 2149515264 }, { "epoch": 0.3, "learning_rate": 7.042208313272348e-05, "loss": 2.5898, "theoretical_loss": 3.409959617352164, "tokens_seen": 2149646336 }, { "epoch": 0.3, "objective/train/docs_used": 1183895, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.68380069732666, "objective/train/theoretical_loss": 3.4099510721092625, "objective/train/tokens_used": 520170976, "theoretical_loss": 3.4099510721092625, "tokens_seen": 2149711872 }, { "epoch": 0.3, "learning_rate": 7.041405873856523e-05, "loss": 2.6714, "theoretical_loss": 3.4099425271998074, "tokens_seen": 2149777408 }, { "epoch": 0.3, "learning_rate": 7.0406034344407e-05, "loss": 2.5034, "theoretical_loss": 3.409925438381143, "tokens_seen": 2149908480 }, { "epoch": 0.3, "learning_rate": 7.039800995024875e-05, "loss": 2.5538, "theoretical_loss": 3.409908350895984, "tokens_seen": 2150039552 }, { "epoch": 0.3, "learning_rate": 7.038998555609052e-05, "loss": 2.4173, "theoretical_loss": 3.4098912647441466, "tokens_seen": 2150170624 }, { "epoch": 0.3, "learning_rate": 7.038196116193227e-05, "loss": 2.7148, "theoretical_loss": 3.4098741799254446, "tokens_seen": 2150301696 }, { "epoch": 0.3, "learning_rate": 7.037393676777404e-05, "loss": 2.4907, "theoretical_loss": 3.4098570964396924, "tokens_seen": 2150432768 }, { "epoch": 0.3, "learning_rate": 7.03659123736158e-05, "loss": 2.6022, "theoretical_loss": 3.409840014286705, "tokens_seen": 2150563840 }, { "epoch": 0.3, "learning_rate": 7.035788797945756e-05, "loss": 2.487, "theoretical_loss": 3.409822933466298, "tokens_seen": 2150694912 }, { "epoch": 0.3, "learning_rate": 7.034986358529931e-05, "loss": 2.4843, "theoretical_loss": 3.4098058539782854, "tokens_seen": 2150825984 }, { "epoch": 0.3, "learning_rate": 7.034183919114107e-05, "loss": 2.5202, "theoretical_loss": 3.4097887758224825, "tokens_seen": 2150957056 }, { "epoch": 0.3, "learning_rate": 7.033381479698283e-05, "loss": 2.6072, "theoretical_loss": 3.4097716989987035, "tokens_seen": 2151088128 }, { "epoch": 0.3, "learning_rate": 7.032579040282459e-05, "loss": 2.613, "theoretical_loss": 3.4097546235067644, "tokens_seen": 2151219200 }, { "epoch": 0.3, "objective/train/docs_used": 1184816, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 1.916866660118103, "objective/train/theoretical_loss": 3.409737549346479, "objective/train/tokens_used": 521809376, "theoretical_loss": 3.409737549346479, "tokens_seen": 2151350272 }, { "epoch": 0.3, "learning_rate": 7.031776600866635e-05, "loss": 2.57, "theoretical_loss": 3.409737549346479, "tokens_seen": 2151350272 }, { "epoch": 0.3, "learning_rate": 7.03097416145081e-05, "loss": 2.4954, "theoretical_loss": 3.409720476517664, "tokens_seen": 2151481344 }, { "epoch": 0.3, "learning_rate": 7.030171722034987e-05, "loss": 2.6051, "theoretical_loss": 3.4097034050201334, "tokens_seen": 2151612416 }, { "epoch": 0.3, "learning_rate": 7.029369282619163e-05, "loss": 2.4134, "theoretical_loss": 3.4096863348537014, "tokens_seen": 2151743488 }, { "epoch": 0.3, "learning_rate": 7.028566843203338e-05, "loss": 2.695, "theoretical_loss": 3.409669266018185, "tokens_seen": 2151874560 }, { "epoch": 0.3, "learning_rate": 7.027764403787515e-05, "loss": 2.5432, "theoretical_loss": 3.409652198513398, "tokens_seen": 2152005632 }, { "epoch": 0.3, "learning_rate": 7.02696196437169e-05, "loss": 2.6862, "theoretical_loss": 3.4096351323391563, "tokens_seen": 2152136704 }, { "epoch": 0.3, "learning_rate": 7.026159524955867e-05, "loss": 2.6014, "theoretical_loss": 3.4096180674952747, "tokens_seen": 2152267776 }, { "epoch": 0.3, "learning_rate": 7.025357085540042e-05, "loss": 2.6281, "theoretical_loss": 3.409601003981569, "tokens_seen": 2152398848 }, { "epoch": 0.3, "learning_rate": 7.024554646124219e-05, "loss": 2.7042, "theoretical_loss": 3.4095839417978544, "tokens_seen": 2152529920 }, { "epoch": 0.3, "learning_rate": 7.023752206708394e-05, "loss": 2.5141, "theoretical_loss": 3.409566880943945, "tokens_seen": 2152660992 }, { "epoch": 0.3, "learning_rate": 7.022949767292569e-05, "loss": 2.5674, "theoretical_loss": 3.4095498214196587, "tokens_seen": 2152792064 }, { "epoch": 0.3, "learning_rate": 7.022147327876746e-05, "loss": 2.5695, "theoretical_loss": 3.409532763224809, "tokens_seen": 2152923136 }, { "epoch": 0.3, "objective/train/docs_used": 1185361, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4470932483673096, "objective/train/theoretical_loss": 3.4095242346258656, "objective/train/tokens_used": 523447776, "theoretical_loss": 3.4095242346258656, "tokens_seen": 2152988672 }, { "epoch": 0.3, "learning_rate": 7.021344888460921e-05, "loss": 2.6913, "theoretical_loss": 3.409515706359212, "tokens_seen": 2153054208 }, { "epoch": 0.3, "learning_rate": 7.020542449045098e-05, "loss": 2.7435, "theoretical_loss": 3.409498650822683, "tokens_seen": 2153185280 }, { "epoch": 0.31, "learning_rate": 7.019740009629273e-05, "loss": 2.4549, "theoretical_loss": 3.409481596615038, "tokens_seen": 2153316352 }, { "epoch": 0.31, "learning_rate": 7.01893757021345e-05, "loss": 2.5406, "theoretical_loss": 3.4094645437360915, "tokens_seen": 2153447424 }, { "epoch": 0.31, "learning_rate": 7.018135130797625e-05, "loss": 2.6382, "theoretical_loss": 3.4094474921856603, "tokens_seen": 2153578496 }, { "epoch": 0.31, "learning_rate": 7.0173326913818e-05, "loss": 2.6475, "theoretical_loss": 3.4094304419635595, "tokens_seen": 2153709568 }, { "epoch": 0.31, "learning_rate": 7.016530251965977e-05, "loss": 2.8102, "theoretical_loss": 3.409413393069605, "tokens_seen": 2153840640 }, { "epoch": 0.31, "learning_rate": 7.015727812550152e-05, "loss": 2.6188, "theoretical_loss": 3.4093963455036125, "tokens_seen": 2153971712 }, { "epoch": 0.31, "learning_rate": 7.014925373134329e-05, "loss": 2.609, "theoretical_loss": 3.4093792992653977, "tokens_seen": 2154102784 }, { "epoch": 0.31, "learning_rate": 7.014122933718504e-05, "loss": 2.7287, "theoretical_loss": 3.409362254354776, "tokens_seen": 2154233856 }, { "epoch": 0.31, "learning_rate": 7.013320494302681e-05, "loss": 2.5641, "theoretical_loss": 3.409345210771564, "tokens_seen": 2154364928 }, { "epoch": 0.31, "learning_rate": 7.012518054886856e-05, "loss": 2.5944, "theoretical_loss": 3.409328168515577, "tokens_seen": 2154496000 }, { "epoch": 0.31, "objective/train/docs_used": 1186484, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.775474786758423, "objective/train/theoretical_loss": 3.4093111275866312, "objective/train/tokens_used": 525086176, "theoretical_loss": 3.4093111275866312, "tokens_seen": 2154627072 }, { "epoch": 0.31, "learning_rate": 7.011715615471032e-05, "loss": 2.6251, "theoretical_loss": 3.4093111275866312, "tokens_seen": 2154627072 }, { "epoch": 0.31, "learning_rate": 7.010913176055208e-05, "loss": 2.663, "theoretical_loss": 3.4092940879845424, "tokens_seen": 2154758144 }, { "epoch": 0.31, "learning_rate": 7.010110736639384e-05, "loss": 2.6242, "theoretical_loss": 3.409277049709127, "tokens_seen": 2154889216 }, { "epoch": 0.31, "learning_rate": 7.00930829722356e-05, "loss": 2.548, "theoretical_loss": 3.4092600127602006, "tokens_seen": 2155020288 }, { "epoch": 0.31, "learning_rate": 7.008505857807736e-05, "loss": 2.61, "theoretical_loss": 3.4092429771375787, "tokens_seen": 2155151360 }, { "epoch": 0.31, "learning_rate": 7.007703418391912e-05, "loss": 2.5879, "theoretical_loss": 3.4092259428410783, "tokens_seen": 2155282432 }, { "epoch": 0.31, "learning_rate": 7.006900978976088e-05, "loss": 2.5991, "theoretical_loss": 3.409208909870516, "tokens_seen": 2155413504 }, { "epoch": 0.31, "learning_rate": 7.006098539560264e-05, "loss": 2.5815, "theoretical_loss": 3.4091918782257062, "tokens_seen": 2155544576 }, { "epoch": 0.31, "learning_rate": 7.00529610014444e-05, "loss": 2.6043, "theoretical_loss": 3.409174847906467, "tokens_seen": 2155675648 }, { "epoch": 0.31, "learning_rate": 7.004493660728615e-05, "loss": 2.6625, "theoretical_loss": 3.409157818912613, "tokens_seen": 2155806720 }, { "epoch": 0.31, "learning_rate": 7.003691221312792e-05, "loss": 2.7572, "theoretical_loss": 3.4091407912439617, "tokens_seen": 2155937792 }, { "epoch": 0.31, "learning_rate": 7.002888781896967e-05, "loss": 2.5151, "theoretical_loss": 3.409123764900329, "tokens_seen": 2156068864 }, { "epoch": 0.31, "learning_rate": 7.002086342481144e-05, "loss": 2.6222, "theoretical_loss": 3.409106739881531, "tokens_seen": 2156199936 }, { "epoch": 0.31, "objective/train/docs_used": 1187201, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.76190185546875, "objective/train/theoretical_loss": 3.4090982278688875, "objective/train/tokens_used": 526724576, "theoretical_loss": 3.4090982278688875, "tokens_seen": 2156265472 }, { "epoch": 0.31, "learning_rate": 7.001283903065319e-05, "loss": 2.651, "theoretical_loss": 3.4090897161873848, "tokens_seen": 2156331008 }, { "epoch": 0.31, "learning_rate": 7.000481463649496e-05, "loss": 2.5518, "theoretical_loss": 3.4090726938177056, "tokens_seen": 2156462080 }, { "epoch": 0.31, "learning_rate": 6.999679024233671e-05, "loss": 2.4781, "theoretical_loss": 3.409055672772311, "tokens_seen": 2156593152 }, { "epoch": 0.31, "learning_rate": 6.998876584817846e-05, "loss": 2.6187, "theoretical_loss": 3.409038653051017, "tokens_seen": 2156724224 }, { "epoch": 0.31, "learning_rate": 6.998074145402023e-05, "loss": 2.6385, "theoretical_loss": 3.40902163465364, "tokens_seen": 2156855296 }, { "epoch": 0.31, "learning_rate": 6.997271705986198e-05, "loss": 2.6592, "theoretical_loss": 3.4090046175799973, "tokens_seen": 2156986368 }, { "epoch": 0.31, "learning_rate": 6.996469266570375e-05, "loss": 2.6731, "theoretical_loss": 3.408987601829905, "tokens_seen": 2157117440 }, { "epoch": 0.31, "learning_rate": 6.99566682715455e-05, "loss": 2.6475, "theoretical_loss": 3.4089705874031795, "tokens_seen": 2157248512 }, { "epoch": 0.31, "learning_rate": 6.994864387738727e-05, "loss": 2.436, "theoretical_loss": 3.4089535742996375, "tokens_seen": 2157379584 }, { "epoch": 0.31, "learning_rate": 6.994061948322902e-05, "loss": 2.6633, "theoretical_loss": 3.408936562519096, "tokens_seen": 2157510656 }, { "epoch": 0.31, "learning_rate": 6.993259508907077e-05, "loss": 2.6681, "theoretical_loss": 3.4089195520613718, "tokens_seen": 2157641728 }, { "epoch": 0.31, "learning_rate": 6.992457069491254e-05, "loss": 2.573, "theoretical_loss": 3.4089025429262816, "tokens_seen": 2157772800 }, { "epoch": 0.31, "objective/train/docs_used": 1188127, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6738452911376953, "objective/train/theoretical_loss": 3.4088855351136425, "objective/train/tokens_used": 528362976, "theoretical_loss": 3.4088855351136425, "tokens_seen": 2157903872 }, { "epoch": 0.31, "learning_rate": 6.99165463007543e-05, "loss": 2.5564, "theoretical_loss": 3.4088855351136425, "tokens_seen": 2157903872 }, { "epoch": 0.31, "learning_rate": 6.990852190659606e-05, "loss": 2.6039, "theoretical_loss": 3.4088685286232705, "tokens_seen": 2158034944 }, { "epoch": 0.31, "learning_rate": 6.990049751243781e-05, "loss": 2.5855, "theoretical_loss": 3.4088515234549837, "tokens_seen": 2158166016 }, { "epoch": 0.31, "learning_rate": 6.989247311827958e-05, "loss": 2.6119, "theoretical_loss": 3.408834519608598, "tokens_seen": 2158297088 }, { "epoch": 0.31, "learning_rate": 6.988444872412133e-05, "loss": 2.5422, "theoretical_loss": 3.408817517083931, "tokens_seen": 2158428160 }, { "epoch": 0.31, "learning_rate": 6.987642432996309e-05, "loss": 2.6445, "theoretical_loss": 3.4088005158807992, "tokens_seen": 2158559232 }, { "epoch": 0.31, "learning_rate": 6.986839993580485e-05, "loss": 2.6504, "theoretical_loss": 3.40878351599902, "tokens_seen": 2158690304 }, { "epoch": 0.31, "learning_rate": 6.98603755416466e-05, "loss": 2.6189, "theoretical_loss": 3.408766517438411, "tokens_seen": 2158821376 }, { "epoch": 0.31, "learning_rate": 6.985235114748837e-05, "loss": 2.5501, "theoretical_loss": 3.4087495201987883, "tokens_seen": 2158952448 }, { "epoch": 0.31, "learning_rate": 6.984432675333013e-05, "loss": 2.4872, "theoretical_loss": 3.4087325242799693, "tokens_seen": 2159083520 }, { "epoch": 0.31, "learning_rate": 6.98363023591719e-05, "loss": 2.5379, "theoretical_loss": 3.4087155296817717, "tokens_seen": 2159214592 }, { "epoch": 0.31, "learning_rate": 6.982827796501365e-05, "loss": 2.4032, "theoretical_loss": 3.4086985364040125, "tokens_seen": 2159345664 }, { "epoch": 0.31, "learning_rate": 6.98202535708554e-05, "loss": 2.6065, "theoretical_loss": 3.408681544446509, "tokens_seen": 2159476736 }, { "epoch": 0.31, "objective/train/docs_used": 1188847, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7533798217773438, "objective/train/theoretical_loss": 3.4086730489627954, "objective/train/tokens_used": 530001376, "theoretical_loss": 3.4086730489627954, "tokens_seen": 2159542272 }, { "epoch": 0.31, "learning_rate": 6.981222917669717e-05, "loss": 2.5673, "theoretical_loss": 3.4086645538090776, "tokens_seen": 2159607808 }, { "epoch": 0.31, "learning_rate": 6.980420478253892e-05, "loss": 2.5625, "theoretical_loss": 3.4086475644915373, "tokens_seen": 2159738880 }, { "epoch": 0.31, "learning_rate": 6.979618038838069e-05, "loss": 2.5184, "theoretical_loss": 3.408630576493704, "tokens_seen": 2159869952 }, { "epoch": 0.31, "learning_rate": 6.978815599422244e-05, "loss": 2.6005, "theoretical_loss": 3.408613589815396, "tokens_seen": 2160001024 }, { "epoch": 0.31, "learning_rate": 6.97801316000642e-05, "loss": 2.4839, "theoretical_loss": 3.4085966044564304, "tokens_seen": 2160132096 }, { "epoch": 0.31, "learning_rate": 6.977210720590596e-05, "loss": 2.5711, "theoretical_loss": 3.4085796204166243, "tokens_seen": 2160263168 }, { "epoch": 0.31, "learning_rate": 6.976408281174773e-05, "loss": 2.6037, "theoretical_loss": 3.4085626376957965, "tokens_seen": 2160394240 }, { "epoch": 0.31, "learning_rate": 6.975605841758948e-05, "loss": 2.5763, "theoretical_loss": 3.408545656293763, "tokens_seen": 2160525312 }, { "epoch": 0.31, "learning_rate": 6.974803402343123e-05, "loss": 2.6665, "theoretical_loss": 3.4085286762103424, "tokens_seen": 2160656384 }, { "epoch": 0.31, "learning_rate": 6.9740009629273e-05, "loss": 2.4983, "theoretical_loss": 3.408511697445352, "tokens_seen": 2160787456 }, { "epoch": 0.31, "learning_rate": 6.973198523511475e-05, "loss": 2.5276, "theoretical_loss": 3.4084947199986093, "tokens_seen": 2160918528 }, { "epoch": 0.31, "learning_rate": 6.972396084095652e-05, "loss": 2.6489, "theoretical_loss": 3.4084777438699323, "tokens_seen": 2161049600 }, { "epoch": 0.31, "objective/train/docs_used": 1190045, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.518320083618164, "objective/train/theoretical_loss": 3.408460769059139, "objective/train/tokens_used": 531639776, "theoretical_loss": 3.408460769059139, "tokens_seen": 2161180672 }, { "epoch": 0.31, "learning_rate": 6.971593644679827e-05, "loss": 2.538, "theoretical_loss": 3.408460769059139, "tokens_seen": 2161180672 }, { "epoch": 0.31, "learning_rate": 6.970791205264004e-05, "loss": 2.5788, "theoretical_loss": 3.408443795566046, "tokens_seen": 2161311744 }, { "epoch": 0.31, "learning_rate": 6.969988765848179e-05, "loss": 2.6509, "theoretical_loss": 3.4084268233904726, "tokens_seen": 2161442816 }, { "epoch": 0.31, "learning_rate": 6.969186326432354e-05, "loss": 2.578, "theoretical_loss": 3.4084098525322357, "tokens_seen": 2161573888 }, { "epoch": 0.31, "learning_rate": 6.968383887016531e-05, "loss": 2.4955, "theoretical_loss": 3.408392882991153, "tokens_seen": 2161704960 }, { "epoch": 0.31, "learning_rate": 6.967581447600706e-05, "loss": 2.6306, "theoretical_loss": 3.408375914767044, "tokens_seen": 2161836032 }, { "epoch": 0.31, "learning_rate": 6.966779008184883e-05, "loss": 2.6397, "theoretical_loss": 3.4083589478597247, "tokens_seen": 2161967104 }, { "epoch": 0.31, "learning_rate": 6.965976568769058e-05, "loss": 2.5939, "theoretical_loss": 3.4083419822690137, "tokens_seen": 2162098176 }, { "epoch": 0.31, "learning_rate": 6.965174129353235e-05, "loss": 2.6069, "theoretical_loss": 3.4083250179947293, "tokens_seen": 2162229248 }, { "epoch": 0.31, "learning_rate": 6.96437168993741e-05, "loss": 2.5248, "theoretical_loss": 3.4083080550366898, "tokens_seen": 2162360320 }, { "epoch": 0.31, "learning_rate": 6.963569250521586e-05, "loss": 2.6181, "theoretical_loss": 3.4082910933947126, "tokens_seen": 2162491392 }, { "epoch": 0.31, "learning_rate": 6.962766811105762e-05, "loss": 2.5547, "theoretical_loss": 3.408274133068617, "tokens_seen": 2162622464 }, { "epoch": 0.31, "learning_rate": 6.961964371689938e-05, "loss": 2.4597, "theoretical_loss": 3.4082571740582197, "tokens_seen": 2162753536 }, { "epoch": 0.31, "objective/train/docs_used": 1191405, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.051525115966797, "objective/train/theoretical_loss": 3.4082486950463515, "objective/train/tokens_used": 533278176, "theoretical_loss": 3.4082486950463515, "tokens_seen": 2162819072 }, { "epoch": 0.31, "learning_rate": 6.961161932274114e-05, "loss": 2.5429, "theoretical_loss": 3.4082402163633394, "tokens_seen": 2162884608 }, { "epoch": 0.31, "learning_rate": 6.96035949285829e-05, "loss": 2.5639, "theoretical_loss": 3.408223259983795, "tokens_seen": 2163015680 }, { "epoch": 0.31, "learning_rate": 6.959557053442466e-05, "loss": 2.5452, "theoretical_loss": 3.4082063049194042, "tokens_seen": 2163146752 }, { "epoch": 0.31, "learning_rate": 6.958754614026642e-05, "loss": 2.5338, "theoretical_loss": 3.4081893511699852, "tokens_seen": 2163277824 }, { "epoch": 0.31, "learning_rate": 6.957952174610817e-05, "loss": 2.605, "theoretical_loss": 3.408172398735357, "tokens_seen": 2163408896 }, { "epoch": 0.31, "learning_rate": 6.957149735194994e-05, "loss": 2.5398, "theoretical_loss": 3.4081554476153375, "tokens_seen": 2163539968 }, { "epoch": 0.31, "learning_rate": 6.956347295779169e-05, "loss": 2.5628, "theoretical_loss": 3.4081384978097446, "tokens_seen": 2163671040 }, { "epoch": 0.31, "learning_rate": 6.955544856363346e-05, "loss": 2.5798, "theoretical_loss": 3.408121549318398, "tokens_seen": 2163802112 }, { "epoch": 0.31, "learning_rate": 6.954742416947521e-05, "loss": 2.5445, "theoretical_loss": 3.4081046021411154, "tokens_seen": 2163933184 }, { "epoch": 0.31, "learning_rate": 6.953939977531698e-05, "loss": 2.5407, "theoretical_loss": 3.4080876562777154, "tokens_seen": 2164064256 }, { "epoch": 0.31, "learning_rate": 6.953137538115873e-05, "loss": 2.5254, "theoretical_loss": 3.4080707117280165, "tokens_seen": 2164195328 }, { "epoch": 0.31, "learning_rate": 6.952335098700048e-05, "loss": 2.5325, "theoretical_loss": 3.408053768491838, "tokens_seen": 2164326400 }, { "epoch": 0.31, "objective/train/docs_used": 1192132, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.1488561630249023, "objective/train/theoretical_loss": 3.4080368265689973, "objective/train/tokens_used": 534916576, "theoretical_loss": 3.4080368265689973, "tokens_seen": 2164457472 }, { "epoch": 0.31, "learning_rate": 6.951532659284225e-05, "loss": 2.4803, "theoretical_loss": 3.4080368265689973, "tokens_seen": 2164457472 }, { "epoch": 0.31, "learning_rate": 6.9507302198684e-05, "loss": 2.5546, "theoretical_loss": 3.4080198859593143, "tokens_seen": 2164588544 }, { "epoch": 0.31, "learning_rate": 6.949927780452577e-05, "loss": 2.723, "theoretical_loss": 3.4080029466626067, "tokens_seen": 2164719616 }, { "epoch": 0.31, "learning_rate": 6.949125341036752e-05, "loss": 2.4956, "theoretical_loss": 3.407986008678694, "tokens_seen": 2164850688 }, { "epoch": 0.31, "learning_rate": 6.948322901620929e-05, "loss": 2.4595, "theoretical_loss": 3.4079690720073943, "tokens_seen": 2164981760 }, { "epoch": 0.31, "learning_rate": 6.947520462205104e-05, "loss": 2.5846, "theoretical_loss": 3.4079521366485275, "tokens_seen": 2165112832 }, { "epoch": 0.31, "learning_rate": 6.94671802278928e-05, "loss": 2.6315, "theoretical_loss": 3.4079352026019114, "tokens_seen": 2165243904 }, { "epoch": 0.31, "learning_rate": 6.945915583373456e-05, "loss": 2.595, "theoretical_loss": 3.4079182698673653, "tokens_seen": 2165374976 }, { "epoch": 0.31, "learning_rate": 6.945113143957631e-05, "loss": 2.6572, "theoretical_loss": 3.407901338444708, "tokens_seen": 2165506048 }, { "epoch": 0.31, "learning_rate": 6.944310704541808e-05, "loss": 2.4416, "theoretical_loss": 3.407884408333759, "tokens_seen": 2165637120 }, { "epoch": 0.31, "learning_rate": 6.943508265125983e-05, "loss": 2.4823, "theoretical_loss": 3.407867479534336, "tokens_seen": 2165768192 }, { "epoch": 0.31, "learning_rate": 6.94270582571016e-05, "loss": 2.4664, "theoretical_loss": 3.40785055204626, "tokens_seen": 2165899264 }, { "epoch": 0.31, "learning_rate": 6.941903386294335e-05, "loss": 2.5716, "theoretical_loss": 3.4078336258693485, "tokens_seen": 2166030336 }, { "epoch": 0.31, "objective/train/docs_used": 1193548, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6069717407226562, "objective/train/theoretical_loss": 3.407825163272523, "objective/train/tokens_used": 536554976, "theoretical_loss": 3.407825163272523, "tokens_seen": 2166095872 }, { "epoch": 0.31, "learning_rate": 6.941100946878512e-05, "loss": 2.4828, "theoretical_loss": 3.407816701003421, "tokens_seen": 2166161408 }, { "epoch": 0.31, "learning_rate": 6.940298507462687e-05, "loss": 2.5359, "theoretical_loss": 3.4077997774482967, "tokens_seen": 2166292480 }, { "epoch": 0.31, "learning_rate": 6.939496068046863e-05, "loss": 2.4822, "theoretical_loss": 3.407782855203795, "tokens_seen": 2166423552 }, { "epoch": 0.31, "learning_rate": 6.93869362863104e-05, "loss": 2.5632, "theoretical_loss": 3.4077659342697344, "tokens_seen": 2166554624 }, { "epoch": 0.31, "learning_rate": 6.937891189215215e-05, "loss": 2.41, "theoretical_loss": 3.407749014645935, "tokens_seen": 2166685696 }, { "epoch": 0.31, "learning_rate": 6.937088749799391e-05, "loss": 2.5752, "theoretical_loss": 3.407732096332216, "tokens_seen": 2166816768 }, { "epoch": 0.31, "learning_rate": 6.936286310383567e-05, "loss": 2.6156, "theoretical_loss": 3.407715179328396, "tokens_seen": 2166947840 }, { "epoch": 0.31, "learning_rate": 6.935483870967743e-05, "loss": 2.512, "theoretical_loss": 3.4076982636342956, "tokens_seen": 2167078912 }, { "epoch": 0.31, "learning_rate": 6.934681431551919e-05, "loss": 2.2435, "theoretical_loss": 3.407681349249733, "tokens_seen": 2167209984 }, { "epoch": 0.31, "learning_rate": 6.933878992136094e-05, "loss": 2.4444, "theoretical_loss": 3.4076644361745276, "tokens_seen": 2167341056 }, { "epoch": 0.31, "learning_rate": 6.93307655272027e-05, "loss": 2.6329, "theoretical_loss": 3.4076475244084996, "tokens_seen": 2167472128 }, { "epoch": 0.31, "learning_rate": 6.932274113304446e-05, "loss": 2.535, "theoretical_loss": 3.4076306139514685, "tokens_seen": 2167603200 }, { "epoch": 0.31, "objective/train/docs_used": 1194194, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.670583724975586, "objective/train/theoretical_loss": 3.4076137048032535, "objective/train/tokens_used": 538193376, "theoretical_loss": 3.4076137048032535, "tokens_seen": 2167734272 }, { "epoch": 0.31, "learning_rate": 6.931471673888623e-05, "loss": 2.5218, "theoretical_loss": 3.4076137048032535, "tokens_seen": 2167734272 }, { "epoch": 0.31, "learning_rate": 6.930669234472798e-05, "loss": 2.4784, "theoretical_loss": 3.4075967969636745, "tokens_seen": 2167865344 }, { "epoch": 0.31, "learning_rate": 6.929866795056975e-05, "loss": 2.3791, "theoretical_loss": 3.40757989043255, "tokens_seen": 2167996416 }, { "epoch": 0.31, "learning_rate": 6.92906435564115e-05, "loss": 2.5507, "theoretical_loss": 3.4075629852097014, "tokens_seen": 2168127488 }, { "epoch": 0.31, "learning_rate": 6.928261916225325e-05, "loss": 2.6381, "theoretical_loss": 3.4075460812949467, "tokens_seen": 2168258560 }, { "epoch": 0.31, "learning_rate": 6.927459476809502e-05, "loss": 2.5207, "theoretical_loss": 3.4075291786881072, "tokens_seen": 2168389632 }, { "epoch": 0.31, "learning_rate": 6.926657037393677e-05, "loss": 2.4885, "theoretical_loss": 3.407512277389001, "tokens_seen": 2168520704 }, { "epoch": 0.31, "learning_rate": 6.925854597977854e-05, "loss": 2.5377, "theoretical_loss": 3.4074953773974492, "tokens_seen": 2168651776 }, { "epoch": 0.31, "learning_rate": 6.925052158562029e-05, "loss": 2.3995, "theoretical_loss": 3.4074784787132715, "tokens_seen": 2168782848 }, { "epoch": 0.31, "learning_rate": 6.924249719146206e-05, "loss": 2.4598, "theoretical_loss": 3.407461581336287, "tokens_seen": 2168913920 }, { "epoch": 0.31, "learning_rate": 6.923447279730381e-05, "loss": 2.4744, "theoretical_loss": 3.407444685266316, "tokens_seen": 2169044992 }, { "epoch": 0.31, "learning_rate": 6.922644840314556e-05, "loss": 2.4483, "theoretical_loss": 3.4074277905031782, "tokens_seen": 2169176064 }, { "epoch": 0.31, "learning_rate": 6.921842400898733e-05, "loss": 2.4571, "theoretical_loss": 3.4074108970466943, "tokens_seen": 2169307136 }, { "epoch": 0.31, "objective/train/docs_used": 1195647, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.518240451812744, "objective/train/theoretical_loss": 3.407402450808391, "objective/train/tokens_used": 539831776, "theoretical_loss": 3.407402450808391, "tokens_seen": 2169372672 }, { "epoch": 0.31, "learning_rate": 6.921039961482908e-05, "loss": 2.5114, "theoretical_loss": 3.4073940048966835, "tokens_seen": 2169438208 }, { "epoch": 0.31, "learning_rate": 6.920237522067085e-05, "loss": 2.4289, "theoretical_loss": 3.4073771140529665, "tokens_seen": 2169569280 }, { "epoch": 0.31, "learning_rate": 6.91943508265126e-05, "loss": 2.5673, "theoretical_loss": 3.4073602245153625, "tokens_seen": 2169700352 }, { "epoch": 0.32, "learning_rate": 6.918632643235437e-05, "loss": 2.6631, "theoretical_loss": 3.4073433362836925, "tokens_seen": 2169831424 }, { "epoch": 0.32, "learning_rate": 6.917830203819612e-05, "loss": 2.4073, "theoretical_loss": 3.407326449357776, "tokens_seen": 2169962496 }, { "epoch": 0.32, "learning_rate": 6.917027764403788e-05, "loss": 2.6058, "theoretical_loss": 3.4073095637374338, "tokens_seen": 2170093568 }, { "epoch": 0.32, "learning_rate": 6.916225324987964e-05, "loss": 2.6248, "theoretical_loss": 3.407292679422486, "tokens_seen": 2170224640 }, { "epoch": 0.32, "learning_rate": 6.91542288557214e-05, "loss": 2.614, "theoretical_loss": 3.407275796412752, "tokens_seen": 2170355712 }, { "epoch": 0.32, "learning_rate": 6.914620446156316e-05, "loss": 2.5169, "theoretical_loss": 3.407258914708053, "tokens_seen": 2170486784 }, { "epoch": 0.32, "learning_rate": 6.913818006740492e-05, "loss": 2.6178, "theoretical_loss": 3.407242034308209, "tokens_seen": 2170617856 }, { "epoch": 0.32, "learning_rate": 6.913015567324668e-05, "loss": 2.3758, "theoretical_loss": 3.40722515521304, "tokens_seen": 2170748928 }, { "epoch": 0.32, "learning_rate": 6.912213127908844e-05, "loss": 2.5497, "theoretical_loss": 3.4072082774223675, "tokens_seen": 2170880000 }, { "epoch": 0.32, "objective/train/docs_used": 1196267, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4913604259490967, "objective/train/theoretical_loss": 3.4071914009360107, "objective/train/tokens_used": 541470176, "theoretical_loss": 3.4071914009360107, "tokens_seen": 2171011072 }, { "epoch": 0.32, "learning_rate": 6.911410688493019e-05, "loss": 2.4815, "theoretical_loss": 3.4071914009360107, "tokens_seen": 2171011072 }, { "epoch": 0.32, "learning_rate": 6.910608249077196e-05, "loss": 2.5382, "theoretical_loss": 3.4071745257537907, "tokens_seen": 2171142144 }, { "epoch": 0.32, "learning_rate": 6.909805809661371e-05, "loss": 2.5229, "theoretical_loss": 3.407157651875528, "tokens_seen": 2171273216 }, { "epoch": 0.32, "learning_rate": 6.909003370245548e-05, "loss": 2.3727, "theoretical_loss": 3.407140779301043, "tokens_seen": 2171404288 }, { "epoch": 0.32, "learning_rate": 6.908200930829723e-05, "loss": 2.4773, "theoretical_loss": 3.4071239080301563, "tokens_seen": 2171535360 }, { "epoch": 0.32, "learning_rate": 6.907398491413898e-05, "loss": 2.5604, "theoretical_loss": 3.407107038062688, "tokens_seen": 2171666432 }, { "epoch": 0.32, "learning_rate": 6.906596051998075e-05, "loss": 2.4941, "theoretical_loss": 3.40709016939846, "tokens_seen": 2171797504 }, { "epoch": 0.32, "learning_rate": 6.90579361258225e-05, "loss": 2.6267, "theoretical_loss": 3.4070733020372916, "tokens_seen": 2171928576 }, { "epoch": 0.32, "learning_rate": 6.904991173166427e-05, "loss": 2.5602, "theoretical_loss": 3.4070564359790048, "tokens_seen": 2172059648 }, { "epoch": 0.32, "learning_rate": 6.904188733750602e-05, "loss": 2.6802, "theoretical_loss": 3.407039571223419, "tokens_seen": 2172190720 }, { "epoch": 0.32, "learning_rate": 6.903386294334779e-05, "loss": 2.5658, "theoretical_loss": 3.407022707770356, "tokens_seen": 2172321792 }, { "epoch": 0.32, "learning_rate": 6.902583854918954e-05, "loss": 2.3892, "theoretical_loss": 3.4070058456196364, "tokens_seen": 2172452864 }, { "epoch": 0.32, "learning_rate": 6.90178141550313e-05, "loss": 2.3801, "theoretical_loss": 3.406988984771081, "tokens_seen": 2172583936 }, { "epoch": 0.32, "objective/train/docs_used": 1197448, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.376589298248291, "objective/train/theoretical_loss": 3.4069805548350587, "objective/train/tokens_used": 543108576, "theoretical_loss": 3.4069805548350587, "tokens_seen": 2172649472 }, { "epoch": 0.32, "learning_rate": 6.900978976087306e-05, "loss": 2.4687, "theoretical_loss": 3.40697212522451, "tokens_seen": 2172715008 }, { "epoch": 0.32, "learning_rate": 6.900176536671482e-05, "loss": 2.6256, "theoretical_loss": 3.4069552669797454, "tokens_seen": 2172846080 }, { "epoch": 0.32, "learning_rate": 6.899374097255658e-05, "loss": 2.5264, "theoretical_loss": 3.4069384100366076, "tokens_seen": 2172977152 }, { "epoch": 0.32, "learning_rate": 6.898571657839833e-05, "loss": 2.4931, "theoretical_loss": 3.406921554394918, "tokens_seen": 2173108224 }, { "epoch": 0.32, "learning_rate": 6.897769218424009e-05, "loss": 2.3636, "theoretical_loss": 3.406904700054497, "tokens_seen": 2173239296 }, { "epoch": 0.32, "learning_rate": 6.896966779008185e-05, "loss": 2.4235, "theoretical_loss": 3.4068878470151662, "tokens_seen": 2173370368 }, { "epoch": 0.32, "learning_rate": 6.896164339592361e-05, "loss": 2.4344, "theoretical_loss": 3.406870995276747, "tokens_seen": 2173501440 }, { "epoch": 0.32, "learning_rate": 6.895361900176537e-05, "loss": 2.4578, "theoretical_loss": 3.406854144839059, "tokens_seen": 2173632512 }, { "epoch": 0.32, "learning_rate": 6.894559460760713e-05, "loss": 2.5416, "theoretical_loss": 3.406837295701925, "tokens_seen": 2173763584 }, { "epoch": 0.32, "learning_rate": 6.89375702134489e-05, "loss": 2.5487, "theoretical_loss": 3.406820447865166, "tokens_seen": 2173894656 }, { "epoch": 0.32, "learning_rate": 6.892954581929065e-05, "loss": 2.5288, "theoretical_loss": 3.4068036013286025, "tokens_seen": 2174025728 }, { "epoch": 0.32, "learning_rate": 6.89215214251324e-05, "loss": 2.4718, "theoretical_loss": 3.406786756092056, "tokens_seen": 2174156800 }, { "epoch": 0.32, "objective/train/docs_used": 1198116, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.546083927154541, "objective/train/theoretical_loss": 3.4067699121553483, "objective/train/tokens_used": 544746976, "theoretical_loss": 3.4067699121553483, "tokens_seen": 2174287872 }, { "epoch": 0.32, "learning_rate": 6.891349703097417e-05, "loss": 2.4965, "theoretical_loss": 3.4067699121553483, "tokens_seen": 2174287872 }, { "epoch": 0.32, "learning_rate": 6.890547263681592e-05, "loss": 2.5112, "theoretical_loss": 3.4067530695183, "tokens_seen": 2174418944 }, { "epoch": 0.32, "learning_rate": 6.889744824265769e-05, "loss": 2.4467, "theoretical_loss": 3.4067362281807334, "tokens_seen": 2174550016 }, { "epoch": 0.32, "learning_rate": 6.888942384849944e-05, "loss": 2.5762, "theoretical_loss": 3.4067193881424696, "tokens_seen": 2174681088 }, { "epoch": 0.32, "learning_rate": 6.88813994543412e-05, "loss": 2.5666, "theoretical_loss": 3.406702549403329, "tokens_seen": 2174812160 }, { "epoch": 0.32, "learning_rate": 6.887337506018296e-05, "loss": 2.4124, "theoretical_loss": 3.406685711963135, "tokens_seen": 2174943232 }, { "epoch": 0.32, "learning_rate": 6.886535066602471e-05, "loss": 2.4991, "theoretical_loss": 3.406668875821708, "tokens_seen": 2175074304 }, { "epoch": 0.32, "learning_rate": 6.885732627186648e-05, "loss": 2.3663, "theoretical_loss": 3.406652040978869, "tokens_seen": 2175205376 }, { "epoch": 0.32, "learning_rate": 6.884930187770823e-05, "loss": 2.589, "theoretical_loss": 3.406635207434441, "tokens_seen": 2175336448 }, { "epoch": 0.32, "learning_rate": 6.884127748355e-05, "loss": 2.4368, "theoretical_loss": 3.4066183751882444, "tokens_seen": 2175467520 }, { "epoch": 0.32, "learning_rate": 6.883325308939175e-05, "loss": 2.568, "theoretical_loss": 3.4066015442401016, "tokens_seen": 2175598592 }, { "epoch": 0.32, "learning_rate": 6.88252286952335e-05, "loss": 2.515, "theoretical_loss": 3.406584714589834, "tokens_seen": 2175729664 }, { "epoch": 0.32, "learning_rate": 6.881720430107527e-05, "loss": 2.5158, "theoretical_loss": 3.4065678862372635, "tokens_seen": 2175860736 }, { "epoch": 0.32, "objective/train/docs_used": 1199462, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.0797388553619385, "objective/train/theoretical_loss": 3.406559472547559, "objective/train/tokens_used": 546385376, "theoretical_loss": 3.406559472547559, "tokens_seen": 2175926272 }, { "epoch": 0.32, "learning_rate": 6.880917990691703e-05, "loss": 2.4784, "theoretical_loss": 3.406551059182212, "tokens_seen": 2175991808 }, { "epoch": 0.32, "learning_rate": 6.880115551275879e-05, "loss": 2.5083, "theoretical_loss": 3.406534233424501, "tokens_seen": 2176122880 }, { "epoch": 0.32, "learning_rate": 6.879313111860055e-05, "loss": 2.5584, "theoretical_loss": 3.4065174089639525, "tokens_seen": 2176253952 }, { "epoch": 0.32, "learning_rate": 6.87851067244423e-05, "loss": 2.573, "theoretical_loss": 3.406500585800388, "tokens_seen": 2176385024 }, { "epoch": 0.32, "learning_rate": 6.877708233028407e-05, "loss": 2.5617, "theoretical_loss": 3.40648376393363, "tokens_seen": 2176516096 }, { "epoch": 0.32, "learning_rate": 6.876905793612582e-05, "loss": 2.5937, "theoretical_loss": 3.4064669433635, "tokens_seen": 2176647168 }, { "epoch": 0.32, "learning_rate": 6.876103354196759e-05, "loss": 2.571, "theoretical_loss": 3.4064501240898206, "tokens_seen": 2176778240 }, { "epoch": 0.32, "learning_rate": 6.875300914780934e-05, "loss": 2.515, "theoretical_loss": 3.406433306112413, "tokens_seen": 2176909312 }, { "epoch": 0.32, "learning_rate": 6.87449847536511e-05, "loss": 2.5359, "theoretical_loss": 3.4064164894310998, "tokens_seen": 2177040384 }, { "epoch": 0.32, "learning_rate": 6.873696035949286e-05, "loss": 2.5196, "theoretical_loss": 3.406399674045703, "tokens_seen": 2177171456 }, { "epoch": 0.32, "learning_rate": 6.872893596533461e-05, "loss": 2.5101, "theoretical_loss": 3.406382859956045, "tokens_seen": 2177302528 }, { "epoch": 0.32, "learning_rate": 6.872091157117638e-05, "loss": 2.4146, "theoretical_loss": 3.406366047161947, "tokens_seen": 2177433600 }, { "epoch": 0.32, "objective/train/docs_used": 1200438, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.211505174636841, "objective/train/theoretical_loss": 3.406349235663232, "objective/train/tokens_used": 548023776, "theoretical_loss": 3.406349235663232, "tokens_seen": 2177564672 }, { "epoch": 0.32, "learning_rate": 6.871288717701813e-05, "loss": 2.4019, "theoretical_loss": 3.406349235663232, "tokens_seen": 2177564672 }, { "epoch": 0.32, "learning_rate": 6.87048627828599e-05, "loss": 2.4612, "theoretical_loss": 3.4063324254597225, "tokens_seen": 2177695744 }, { "epoch": 0.32, "learning_rate": 6.869683838870165e-05, "loss": 2.6133, "theoretical_loss": 3.40631561655124, "tokens_seen": 2177826816 }, { "epoch": 0.32, "learning_rate": 6.86888139945434e-05, "loss": 2.4651, "theoretical_loss": 3.406298808937607, "tokens_seen": 2177957888 }, { "epoch": 0.32, "learning_rate": 6.868078960038517e-05, "loss": 2.4464, "theoretical_loss": 3.406282002618646, "tokens_seen": 2178088960 }, { "epoch": 0.32, "learning_rate": 6.867276520622692e-05, "loss": 2.4255, "theoretical_loss": 3.4062651975941796, "tokens_seen": 2178220032 }, { "epoch": 0.32, "learning_rate": 6.866474081206869e-05, "loss": 2.4586, "theoretical_loss": 3.40624839386403, "tokens_seen": 2178351104 }, { "epoch": 0.32, "learning_rate": 6.865671641791044e-05, "loss": 2.4815, "theoretical_loss": 3.4062315914280195, "tokens_seen": 2178482176 }, { "epoch": 0.32, "learning_rate": 6.864869202375221e-05, "loss": 2.6005, "theoretical_loss": 3.4062147902859707, "tokens_seen": 2178613248 }, { "epoch": 0.32, "learning_rate": 6.864066762959396e-05, "loss": 2.5733, "theoretical_loss": 3.4061979904377058, "tokens_seen": 2178744320 }, { "epoch": 0.32, "learning_rate": 6.863264323543572e-05, "loss": 2.4796, "theoretical_loss": 3.4061811918830482, "tokens_seen": 2178875392 }, { "epoch": 0.32, "learning_rate": 6.862461884127748e-05, "loss": 2.5401, "theoretical_loss": 3.406164394621819, "tokens_seen": 2179006464 }, { "epoch": 0.32, "learning_rate": 6.861659444711924e-05, "loss": 2.4538, "theoretical_loss": 3.4061475986538428, "tokens_seen": 2179137536 }, { "epoch": 0.32, "objective/train/docs_used": 1201179, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.695457696914673, "objective/train/theoretical_loss": 3.4061392011547684, "objective/train/tokens_used": 549662176, "theoretical_loss": 3.4061392011547684, "tokens_seen": 2179203072 }, { "epoch": 0.32, "learning_rate": 6.8608570052961e-05, "loss": 2.5881, "theoretical_loss": 3.4061308039789404, "tokens_seen": 2179268608 }, { "epoch": 0.32, "learning_rate": 6.860054565880276e-05, "loss": 2.4436, "theoretical_loss": 3.406114010596936, "tokens_seen": 2179399680 }, { "epoch": 0.32, "learning_rate": 6.859252126464452e-05, "loss": 2.5682, "theoretical_loss": 3.406097218507651, "tokens_seen": 2179530752 }, { "epoch": 0.32, "learning_rate": 6.858449687048628e-05, "loss": 2.3501, "theoretical_loss": 3.4060804277109087, "tokens_seen": 2179661824 }, { "epoch": 0.32, "learning_rate": 6.857647247632803e-05, "loss": 2.4841, "theoretical_loss": 3.406063638206532, "tokens_seen": 2179792896 }, { "epoch": 0.32, "learning_rate": 6.85684480821698e-05, "loss": 2.4954, "theoretical_loss": 3.4060468499943437, "tokens_seen": 2179923968 }, { "epoch": 0.32, "learning_rate": 6.856042368801155e-05, "loss": 2.5969, "theoretical_loss": 3.4060300630741667, "tokens_seen": 2180055040 }, { "epoch": 0.32, "learning_rate": 6.855239929385332e-05, "loss": 2.5807, "theoretical_loss": 3.4060132774458243, "tokens_seen": 2180186112 }, { "epoch": 0.32, "learning_rate": 6.854437489969507e-05, "loss": 2.5297, "theoretical_loss": 3.4059964931091384, "tokens_seen": 2180317184 }, { "epoch": 0.32, "learning_rate": 6.853635050553684e-05, "loss": 2.5044, "theoretical_loss": 3.4059797100639324, "tokens_seen": 2180448256 }, { "epoch": 0.32, "learning_rate": 6.852832611137859e-05, "loss": 2.506, "theoretical_loss": 3.4059629283100294, "tokens_seen": 2180579328 }, { "epoch": 0.32, "learning_rate": 6.852030171722034e-05, "loss": 2.4052, "theoretical_loss": 3.405946147847253, "tokens_seen": 2180710400 }, { "epoch": 0.32, "objective/train/docs_used": 1202357, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7669031620025635, "objective/train/theoretical_loss": 3.4059293686754253, "objective/train/tokens_used": 551300576, "theoretical_loss": 3.4059293686754253, "tokens_seen": 2180841472 }, { "epoch": 0.32, "learning_rate": 6.851227732306211e-05, "loss": 2.4613, "theoretical_loss": 3.4059293686754253, "tokens_seen": 2180841472 }, { "epoch": 0.32, "learning_rate": 6.850425292890386e-05, "loss": 2.529, "theoretical_loss": 3.40591259079437, "tokens_seen": 2180972544 }, { "epoch": 0.32, "learning_rate": 6.849622853474563e-05, "loss": 2.6914, "theoretical_loss": 3.40589581420391, "tokens_seen": 2181103616 }, { "epoch": 0.32, "learning_rate": 6.848820414058738e-05, "loss": 2.5318, "theoretical_loss": 3.4058790389038687, "tokens_seen": 2181234688 }, { "epoch": 0.32, "learning_rate": 6.848017974642915e-05, "loss": 2.4421, "theoretical_loss": 3.4058622648940693, "tokens_seen": 2181365760 }, { "epoch": 0.32, "learning_rate": 6.84721553522709e-05, "loss": 2.4674, "theoretical_loss": 3.4058454921743344, "tokens_seen": 2181496832 }, { "epoch": 0.32, "learning_rate": 6.846413095811267e-05, "loss": 2.3825, "theoretical_loss": 3.405828720744488, "tokens_seen": 2181627904 }, { "epoch": 0.32, "learning_rate": 6.845610656395442e-05, "loss": 2.5162, "theoretical_loss": 3.4058119506043534, "tokens_seen": 2181758976 }, { "epoch": 0.32, "learning_rate": 6.844808216979617e-05, "loss": 2.3782, "theoretical_loss": 3.4057951817537537, "tokens_seen": 2181890048 }, { "epoch": 0.32, "learning_rate": 6.844005777563794e-05, "loss": 2.4567, "theoretical_loss": 3.405778414192512, "tokens_seen": 2182021120 }, { "epoch": 0.32, "learning_rate": 6.84320333814797e-05, "loss": 2.4216, "theoretical_loss": 3.4057616479204524, "tokens_seen": 2182152192 }, { "epoch": 0.32, "learning_rate": 6.842400898732146e-05, "loss": 2.4975, "theoretical_loss": 3.405744882937398, "tokens_seen": 2182283264 }, { "epoch": 0.32, "learning_rate": 6.841598459316321e-05, "loss": 2.41, "theoretical_loss": 3.405728119243172, "tokens_seen": 2182414336 }, { "epoch": 0.32, "objective/train/docs_used": 1202887, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.703336000442505, "objective/train/theoretical_loss": 3.405719737879315, "objective/train/tokens_used": 552938976, "theoretical_loss": 3.405719737879315, "tokens_seen": 2182479872 }, { "epoch": 0.32, "learning_rate": 6.840796019900498e-05, "loss": 2.6198, "theoretical_loss": 3.4057113568375987, "tokens_seen": 2182545408 }, { "epoch": 0.32, "learning_rate": 6.839993580484673e-05, "loss": 2.5698, "theoretical_loss": 3.405694595720501, "tokens_seen": 2182676480 }, { "epoch": 0.32, "learning_rate": 6.839191141068849e-05, "loss": 2.4929, "theoretical_loss": 3.4056778358917024, "tokens_seen": 2182807552 }, { "epoch": 0.32, "learning_rate": 6.838388701653025e-05, "loss": 2.5502, "theoretical_loss": 3.405661077351027, "tokens_seen": 2182938624 }, { "epoch": 0.32, "learning_rate": 6.8375862622372e-05, "loss": 2.6634, "theoretical_loss": 3.405644320098298, "tokens_seen": 2183069696 }, { "epoch": 0.32, "learning_rate": 6.836783822821377e-05, "loss": 2.5867, "theoretical_loss": 3.40562756413334, "tokens_seen": 2183200768 }, { "epoch": 0.32, "learning_rate": 6.835981383405553e-05, "loss": 2.5164, "theoretical_loss": 3.405610809455976, "tokens_seen": 2183331840 }, { "epoch": 0.32, "learning_rate": 6.835178943989729e-05, "loss": 2.4997, "theoretical_loss": 3.4055940560660294, "tokens_seen": 2183462912 }, { "epoch": 0.32, "learning_rate": 6.834376504573905e-05, "loss": 2.4337, "theoretical_loss": 3.405577303963325, "tokens_seen": 2183593984 }, { "epoch": 0.32, "learning_rate": 6.83357406515808e-05, "loss": 2.4849, "theoretical_loss": 3.4055605531476854, "tokens_seen": 2183725056 }, { "epoch": 0.32, "learning_rate": 6.832771625742257e-05, "loss": 2.5738, "theoretical_loss": 3.4055438036189356, "tokens_seen": 2183856128 }, { "epoch": 0.32, "learning_rate": 6.831969186326432e-05, "loss": 2.4967, "theoretical_loss": 3.405527055376899, "tokens_seen": 2183987200 }, { "epoch": 0.32, "objective/train/docs_used": 1204090, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4657957553863525, "objective/train/theoretical_loss": 3.4055103084213996, "objective/train/tokens_used": 554577376, "theoretical_loss": 3.4055103084213996, "tokens_seen": 2184118272 }, { "epoch": 0.32, "learning_rate": 6.831166746910609e-05, "loss": 2.5754, "theoretical_loss": 3.4055103084213996, "tokens_seen": 2184118272 }, { "epoch": 0.32, "learning_rate": 6.830364307494784e-05, "loss": 2.4762, "theoretical_loss": 3.4054935627522616, "tokens_seen": 2184249344 }, { "epoch": 0.32, "learning_rate": 6.82956186807896e-05, "loss": 2.5145, "theoretical_loss": 3.4054768183693085, "tokens_seen": 2184380416 }, { "epoch": 0.32, "learning_rate": 6.828759428663136e-05, "loss": 2.5413, "theoretical_loss": 3.405460075272365, "tokens_seen": 2184511488 }, { "epoch": 0.32, "learning_rate": 6.827956989247311e-05, "loss": 2.6336, "theoretical_loss": 3.4054433334612546, "tokens_seen": 2184642560 }, { "epoch": 0.32, "learning_rate": 6.827154549831488e-05, "loss": 2.5995, "theoretical_loss": 3.4054265929358016, "tokens_seen": 2184773632 }, { "epoch": 0.32, "learning_rate": 6.826352110415663e-05, "loss": 2.5385, "theoretical_loss": 3.40540985369583, "tokens_seen": 2184904704 }, { "epoch": 0.32, "learning_rate": 6.82554967099984e-05, "loss": 2.3402, "theoretical_loss": 3.4053931157411648, "tokens_seen": 2185035776 }, { "epoch": 0.32, "learning_rate": 6.824747231584015e-05, "loss": 2.4948, "theoretical_loss": 3.405376379071629, "tokens_seen": 2185166848 }, { "epoch": 0.32, "learning_rate": 6.823944792168192e-05, "loss": 2.6132, "theoretical_loss": 3.405359643687048, "tokens_seen": 2185297920 }, { "epoch": 0.32, "learning_rate": 6.823142352752367e-05, "loss": 2.4663, "theoretical_loss": 3.405342909587245, "tokens_seen": 2185428992 }, { "epoch": 0.32, "learning_rate": 6.822339913336542e-05, "loss": 2.5485, "theoretical_loss": 3.405326176772045, "tokens_seen": 2185560064 }, { "epoch": 0.32, "learning_rate": 6.821537473920719e-05, "loss": 2.5661, "theoretical_loss": 3.405309445241272, "tokens_seen": 2185691136 }, { "epoch": 0.32, "objective/train/docs_used": 1205389, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.9753358364105225, "objective/train/theoretical_loss": 3.4053010799574914, "objective/train/tokens_used": 556215776, "theoretical_loss": 3.4053010799574914, "tokens_seen": 2185756672 }, { "epoch": 0.32, "learning_rate": 6.820735034504894e-05, "loss": 2.5152, "theoretical_loss": 3.405292714994751, "tokens_seen": 2185822208 }, { "epoch": 0.32, "learning_rate": 6.819932595089071e-05, "loss": 2.7423, "theoretical_loss": 3.405275986032306, "tokens_seen": 2185953280 }, { "epoch": 0.32, "learning_rate": 6.819130155673246e-05, "loss": 2.5975, "theoretical_loss": 3.4052592583537615, "tokens_seen": 2186084352 }, { "epoch": 0.32, "learning_rate": 6.818327716257423e-05, "loss": 2.7106, "theoretical_loss": 3.4052425319589417, "tokens_seen": 2186215424 }, { "epoch": 0.33, "learning_rate": 6.817525276841598e-05, "loss": 2.5553, "theoretical_loss": 3.4052258068476715, "tokens_seen": 2186346496 }, { "epoch": 0.33, "learning_rate": 6.816722837425775e-05, "loss": 2.5986, "theoretical_loss": 3.4052090830197757, "tokens_seen": 2186477568 }, { "epoch": 0.33, "learning_rate": 6.81592039800995e-05, "loss": 2.6214, "theoretical_loss": 3.405192360475078, "tokens_seen": 2186608640 }, { "epoch": 0.33, "learning_rate": 6.815117958594126e-05, "loss": 2.517, "theoretical_loss": 3.405175639213404, "tokens_seen": 2186739712 }, { "epoch": 0.33, "learning_rate": 6.814315519178302e-05, "loss": 2.5678, "theoretical_loss": 3.4051589192345775, "tokens_seen": 2186870784 }, { "epoch": 0.33, "learning_rate": 6.813513079762478e-05, "loss": 2.471, "theoretical_loss": 3.405142200538424, "tokens_seen": 2187001856 }, { "epoch": 0.33, "learning_rate": 6.812710640346654e-05, "loss": 2.6426, "theoretical_loss": 3.405125483124768, "tokens_seen": 2187132928 }, { "epoch": 0.33, "learning_rate": 6.81190820093083e-05, "loss": 2.5891, "theoretical_loss": 3.4051087669934343, "tokens_seen": 2187264000 }, { "epoch": 0.33, "objective/train/docs_used": 1206012, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.534045696258545, "objective/train/theoretical_loss": 3.405092052144247, "objective/train/tokens_used": 557854176, "theoretical_loss": 3.405092052144247, "tokens_seen": 2187395072 }, { "epoch": 0.33, "learning_rate": 6.811105761515006e-05, "loss": 2.6112, "theoretical_loss": 3.405092052144247, "tokens_seen": 2187395072 }, { "epoch": 0.33, "learning_rate": 6.810303322099182e-05, "loss": 2.5409, "theoretical_loss": 3.405075338577032, "tokens_seen": 2187526144 }, { "epoch": 0.33, "learning_rate": 6.809500882683357e-05, "loss": 2.596, "theoretical_loss": 3.4050586262916136, "tokens_seen": 2187657216 }, { "epoch": 0.33, "learning_rate": 6.808698443267534e-05, "loss": 2.4925, "theoretical_loss": 3.4050419152878164, "tokens_seen": 2187788288 }, { "epoch": 0.33, "learning_rate": 6.807896003851709e-05, "loss": 2.5335, "theoretical_loss": 3.405025205565466, "tokens_seen": 2187919360 }, { "epoch": 0.33, "learning_rate": 6.807093564435886e-05, "loss": 2.5448, "theoretical_loss": 3.405008497124387, "tokens_seen": 2188050432 }, { "epoch": 0.33, "learning_rate": 6.806291125020061e-05, "loss": 2.5736, "theoretical_loss": 3.404991789964405, "tokens_seen": 2188181504 }, { "epoch": 0.33, "learning_rate": 6.805488685604238e-05, "loss": 2.6279, "theoretical_loss": 3.404975084085344, "tokens_seen": 2188312576 }, { "epoch": 0.33, "learning_rate": 6.804686246188413e-05, "loss": 2.425, "theoretical_loss": 3.40495837948703, "tokens_seen": 2188443648 }, { "epoch": 0.33, "learning_rate": 6.803883806772588e-05, "loss": 2.4579, "theoretical_loss": 3.4049416761692872, "tokens_seen": 2188574720 }, { "epoch": 0.33, "learning_rate": 6.803081367356765e-05, "loss": 2.5621, "theoretical_loss": 3.4049249741319416, "tokens_seen": 2188705792 }, { "epoch": 0.33, "learning_rate": 6.80227892794094e-05, "loss": 2.4914, "theoretical_loss": 3.4049082733748186, "tokens_seen": 2188836864 }, { "epoch": 0.33, "learning_rate": 6.801476488525117e-05, "loss": 2.4783, "theoretical_loss": 3.4048915738977423, "tokens_seen": 2188967936 }, { "epoch": 0.33, "objective/train/docs_used": 1207268, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3856353759765625, "objective/train/theoretical_loss": 3.4048832246391676, "objective/train/tokens_used": 559492576, "theoretical_loss": 3.4048832246391676, "tokens_seen": 2189033472 }, { "epoch": 0.33, "learning_rate": 6.800674049109292e-05, "loss": 2.4133, "theoretical_loss": 3.4048748757005387, "tokens_seen": 2189099008 }, { "epoch": 0.33, "learning_rate": 6.799871609693469e-05, "loss": 2.5576, "theoretical_loss": 3.4048581787830328, "tokens_seen": 2189230080 }, { "epoch": 0.33, "learning_rate": 6.799069170277644e-05, "loss": 2.5191, "theoretical_loss": 3.4048414831450504, "tokens_seen": 2189361152 }, { "epoch": 0.33, "learning_rate": 6.79826673086182e-05, "loss": 2.5269, "theoretical_loss": 3.4048247887864163, "tokens_seen": 2189492224 }, { "epoch": 0.33, "learning_rate": 6.797464291445996e-05, "loss": 2.4617, "theoretical_loss": 3.404808095706956, "tokens_seen": 2189623296 }, { "epoch": 0.33, "learning_rate": 6.796661852030171e-05, "loss": 2.5635, "theoretical_loss": 3.4047914039064953, "tokens_seen": 2189754368 }, { "epoch": 0.33, "learning_rate": 6.795859412614348e-05, "loss": 2.5235, "theoretical_loss": 3.404774713384859, "tokens_seen": 2189885440 }, { "epoch": 0.33, "learning_rate": 6.795056973198523e-05, "loss": 2.5943, "theoretical_loss": 3.404758024141873, "tokens_seen": 2190016512 }, { "epoch": 0.33, "learning_rate": 6.7942545337827e-05, "loss": 2.5848, "theoretical_loss": 3.4047413361773633, "tokens_seen": 2190147584 }, { "epoch": 0.33, "learning_rate": 6.793452094366875e-05, "loss": 2.5442, "theoretical_loss": 3.4047246494911545, "tokens_seen": 2190278656 }, { "epoch": 0.33, "learning_rate": 6.792649654951051e-05, "loss": 2.4471, "theoretical_loss": 3.404707964083073, "tokens_seen": 2190409728 }, { "epoch": 0.33, "learning_rate": 6.791847215535227e-05, "loss": 2.3276, "theoretical_loss": 3.4046912799529436, "tokens_seen": 2190540800 }, { "epoch": 0.33, "objective/train/docs_used": 1207761, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.551201581954956, "objective/train/theoretical_loss": 3.404674597100593, "objective/train/tokens_used": 561130976, "theoretical_loss": 3.404674597100593, "tokens_seen": 2190671872 }, { "epoch": 0.33, "learning_rate": 6.791044776119403e-05, "loss": 2.5127, "theoretical_loss": 3.404674597100593, "tokens_seen": 2190671872 }, { "epoch": 0.33, "learning_rate": 6.79024233670358e-05, "loss": 2.5596, "theoretical_loss": 3.4046579155258456, "tokens_seen": 2190802944 }, { "epoch": 0.33, "learning_rate": 6.789439897287755e-05, "loss": 2.4916, "theoretical_loss": 3.4046412352285285, "tokens_seen": 2190934016 }, { "epoch": 0.33, "learning_rate": 6.788637457871931e-05, "loss": 2.4777, "theoretical_loss": 3.4046245562084665, "tokens_seen": 2191065088 }, { "epoch": 0.33, "learning_rate": 6.787835018456107e-05, "loss": 2.4466, "theoretical_loss": 3.404607878465486, "tokens_seen": 2191196160 }, { "epoch": 0.33, "learning_rate": 6.787032579040283e-05, "loss": 2.5021, "theoretical_loss": 3.4045912019994122, "tokens_seen": 2191327232 }, { "epoch": 0.33, "learning_rate": 6.786230139624459e-05, "loss": 2.5078, "theoretical_loss": 3.4045745268100718, "tokens_seen": 2191458304 }, { "epoch": 0.33, "learning_rate": 6.785427700208634e-05, "loss": 2.3974, "theoretical_loss": 3.40455785289729, "tokens_seen": 2191589376 }, { "epoch": 0.33, "learning_rate": 6.78462526079281e-05, "loss": 2.5153, "theoretical_loss": 3.404541180260893, "tokens_seen": 2191720448 }, { "epoch": 0.33, "learning_rate": 6.783822821376986e-05, "loss": 2.6615, "theoretical_loss": 3.404524508900707, "tokens_seen": 2191851520 }, { "epoch": 0.33, "learning_rate": 6.783020381961163e-05, "loss": 2.6367, "theoretical_loss": 3.404507838816557, "tokens_seen": 2191982592 }, { "epoch": 0.33, "learning_rate": 6.782217942545338e-05, "loss": 2.4062, "theoretical_loss": 3.40449117000827, "tokens_seen": 2192113664 }, { "epoch": 0.33, "learning_rate": 6.781415503129515e-05, "loss": 2.4533, "theoretical_loss": 3.404474502475672, "tokens_seen": 2192244736 }, { "epoch": 0.33, "objective/train/docs_used": 1208978, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7618091106414795, "objective/train/theoretical_loss": 3.4044661691877023, "objective/train/tokens_used": 562769376, "theoretical_loss": 3.4044661691877023, "tokens_seen": 2192310272 }, { "epoch": 0.33, "learning_rate": 6.78061306371369e-05, "loss": 2.6423, "theoretical_loss": 3.4044578362185893, "tokens_seen": 2192375808 }, { "epoch": 0.33, "learning_rate": 6.779810624297865e-05, "loss": 2.3443, "theoretical_loss": 3.404441171236847, "tokens_seen": 2192506880 }, { "epoch": 0.33, "learning_rate": 6.779008184882042e-05, "loss": 2.4826, "theoretical_loss": 3.4044245075302726, "tokens_seen": 2192637952 }, { "epoch": 0.33, "learning_rate": 6.778205745466217e-05, "loss": 2.7126, "theoretical_loss": 3.4044078450986914, "tokens_seen": 2192769024 }, { "epoch": 0.33, "learning_rate": 6.777403306050394e-05, "loss": 2.4795, "theoretical_loss": 3.40439118394193, "tokens_seen": 2192900096 }, { "epoch": 0.33, "learning_rate": 6.776600866634569e-05, "loss": 2.4872, "theoretical_loss": 3.4043745240598144, "tokens_seen": 2193031168 }, { "epoch": 0.33, "learning_rate": 6.775798427218746e-05, "loss": 2.5142, "theoretical_loss": 3.4043578654521713, "tokens_seen": 2193162240 }, { "epoch": 0.33, "learning_rate": 6.774995987802921e-05, "loss": 2.5768, "theoretical_loss": 3.404341208118827, "tokens_seen": 2193293312 }, { "epoch": 0.33, "learning_rate": 6.774193548387096e-05, "loss": 2.3759, "theoretical_loss": 3.4043245520596077, "tokens_seen": 2193424384 }, { "epoch": 0.33, "learning_rate": 6.773391108971273e-05, "loss": 2.4352, "theoretical_loss": 3.4043078972743395, "tokens_seen": 2193555456 }, { "epoch": 0.33, "learning_rate": 6.772588669555448e-05, "loss": 2.562, "theoretical_loss": 3.4042912437628496, "tokens_seen": 2193686528 }, { "epoch": 0.33, "learning_rate": 6.771786230139625e-05, "loss": 2.5281, "theoretical_loss": 3.4042745915249637, "tokens_seen": 2193817600 }, { "epoch": 0.33, "objective/train/docs_used": 1209592, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4534859657287598, "objective/train/theoretical_loss": 3.404257940560509, "objective/train/tokens_used": 564407776, "theoretical_loss": 3.404257940560509, "tokens_seen": 2193948672 }, { "epoch": 0.33, "learning_rate": 6.7709837907238e-05, "loss": 2.5249, "theoretical_loss": 3.404257940560509, "tokens_seen": 2193948672 }, { "epoch": 0.33, "learning_rate": 6.770181351307977e-05, "loss": 2.4806, "theoretical_loss": 3.4042412908693116, "tokens_seen": 2194079744 }, { "epoch": 0.33, "learning_rate": 6.769378911892152e-05, "loss": 2.5328, "theoretical_loss": 3.404224642451198, "tokens_seen": 2194210816 }, { "epoch": 0.33, "learning_rate": 6.768576472476328e-05, "loss": 2.4127, "theoretical_loss": 3.4042079953059954, "tokens_seen": 2194341888 }, { "epoch": 0.33, "learning_rate": 6.767774033060504e-05, "loss": 2.3612, "theoretical_loss": 3.40419134943353, "tokens_seen": 2194472960 }, { "epoch": 0.33, "learning_rate": 6.76697159364468e-05, "loss": 2.5744, "theoretical_loss": 3.4041747048336286, "tokens_seen": 2194604032 }, { "epoch": 0.33, "learning_rate": 6.766169154228856e-05, "loss": 2.5248, "theoretical_loss": 3.4041580615061178, "tokens_seen": 2194735104 }, { "epoch": 0.33, "learning_rate": 6.765366714813032e-05, "loss": 2.4543, "theoretical_loss": 3.4041414194508244, "tokens_seen": 2194866176 }, { "epoch": 0.33, "learning_rate": 6.764564275397208e-05, "loss": 2.5248, "theoretical_loss": 3.4041247786675752, "tokens_seen": 2194997248 }, { "epoch": 0.33, "learning_rate": 6.763761835981384e-05, "loss": 2.5018, "theoretical_loss": 3.404108139156197, "tokens_seen": 2195128320 }, { "epoch": 0.33, "learning_rate": 6.762959396565559e-05, "loss": 2.4872, "theoretical_loss": 3.4040915009165174, "tokens_seen": 2195259392 }, { "epoch": 0.33, "learning_rate": 6.762156957149736e-05, "loss": 2.5658, "theoretical_loss": 3.404074863948362, "tokens_seen": 2195390464 }, { "epoch": 0.33, "learning_rate": 6.761354517733911e-05, "loss": 2.5886, "theoretical_loss": 3.404058228251558, "tokens_seen": 2195521536 }, { "epoch": 0.33, "objective/train/docs_used": 1211067, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.898007869720459, "objective/train/theoretical_loss": 3.404049910879859, "objective/train/tokens_used": 566046176, "theoretical_loss": 3.404049910879859, "tokens_seen": 2195587072 }, { "epoch": 0.33, "learning_rate": 6.760552078318088e-05, "loss": 2.5062, "theoretical_loss": 3.404041593825933, "tokens_seen": 2195652608 }, { "epoch": 0.33, "learning_rate": 6.759749638902263e-05, "loss": 2.4295, "theoretical_loss": 3.4040249606713138, "tokens_seen": 2195783680 }, { "epoch": 0.33, "learning_rate": 6.75894719948644e-05, "loss": 2.5987, "theoretical_loss": 3.404008328787527, "tokens_seen": 2195914752 }, { "epoch": 0.33, "learning_rate": 6.758144760070615e-05, "loss": 2.5701, "theoretical_loss": 3.4039916981744005, "tokens_seen": 2196045824 }, { "epoch": 0.33, "learning_rate": 6.75734232065479e-05, "loss": 2.5974, "theoretical_loss": 3.4039750688317603, "tokens_seen": 2196176896 }, { "epoch": 0.33, "learning_rate": 6.756539881238967e-05, "loss": 2.5712, "theoretical_loss": 3.4039584407594345, "tokens_seen": 2196307968 }, { "epoch": 0.33, "learning_rate": 6.755737441823142e-05, "loss": 2.4627, "theoretical_loss": 3.4039418139572493, "tokens_seen": 2196439040 }, { "epoch": 0.33, "learning_rate": 6.754935002407319e-05, "loss": 2.61, "theoretical_loss": 3.403925188425033, "tokens_seen": 2196570112 }, { "epoch": 0.33, "learning_rate": 6.754132562991494e-05, "loss": 2.4044, "theoretical_loss": 3.4039085641626117, "tokens_seen": 2196701184 }, { "epoch": 0.33, "learning_rate": 6.753330123575671e-05, "loss": 2.4572, "theoretical_loss": 3.4038919411698134, "tokens_seen": 2196832256 }, { "epoch": 0.33, "learning_rate": 6.752527684159846e-05, "loss": 2.3802, "theoretical_loss": 3.403875319446465, "tokens_seen": 2196963328 }, { "epoch": 0.33, "learning_rate": 6.751725244744023e-05, "loss": 2.538, "theoretical_loss": 3.4038586989923942, "tokens_seen": 2197094400 }, { "epoch": 0.33, "objective/train/docs_used": 1211765, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.350731134414673, "objective/train/theoretical_loss": 3.4038420798074283, "objective/train/tokens_used": 567684576, "theoretical_loss": 3.4038420798074283, "tokens_seen": 2197225472 }, { "epoch": 0.33, "learning_rate": 6.750922805328198e-05, "loss": 2.6054, "theoretical_loss": 3.4038420798074283, "tokens_seen": 2197225472 }, { "epoch": 0.33, "learning_rate": 6.750120365912373e-05, "loss": 2.5121, "theoretical_loss": 3.403825461891395, "tokens_seen": 2197356544 }, { "epoch": 0.33, "learning_rate": 6.74931792649655e-05, "loss": 2.3871, "theoretical_loss": 3.4038088452441206, "tokens_seen": 2197487616 }, { "epoch": 0.33, "learning_rate": 6.748515487080725e-05, "loss": 2.5174, "theoretical_loss": 3.4037922298654335, "tokens_seen": 2197618688 }, { "epoch": 0.33, "learning_rate": 6.747713047664902e-05, "loss": 2.5223, "theoretical_loss": 3.403775615755161, "tokens_seen": 2197749760 }, { "epoch": 0.33, "learning_rate": 6.746910608249077e-05, "loss": 2.4984, "theoretical_loss": 3.4037590029131306, "tokens_seen": 2197880832 }, { "epoch": 0.33, "learning_rate": 6.746108168833254e-05, "loss": 2.3523, "theoretical_loss": 3.40374239133917, "tokens_seen": 2198011904 }, { "epoch": 0.33, "learning_rate": 6.74530572941743e-05, "loss": 2.5873, "theoretical_loss": 3.4037257810331063, "tokens_seen": 2198142976 }, { "epoch": 0.33, "learning_rate": 6.744503290001605e-05, "loss": 2.3596, "theoretical_loss": 3.403709171994768, "tokens_seen": 2198274048 }, { "epoch": 0.33, "learning_rate": 6.743700850585781e-05, "loss": 2.5814, "theoretical_loss": 3.4036925642239817, "tokens_seen": 2198405120 }, { "epoch": 0.33, "learning_rate": 6.742898411169957e-05, "loss": 2.3935, "theoretical_loss": 3.403675957720576, "tokens_seen": 2198536192 }, { "epoch": 0.33, "learning_rate": 6.742095971754133e-05, "loss": 2.5633, "theoretical_loss": 3.403659352484378, "tokens_seen": 2198667264 }, { "epoch": 0.33, "learning_rate": 6.741293532338309e-05, "loss": 2.3273, "theoretical_loss": 3.403642748515216, "tokens_seen": 2198798336 }, { "epoch": 0.33, "objective/train/docs_used": 1213097, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.580597162246704, "objective/train/theoretical_loss": 3.40363444700572, "objective/train/tokens_used": 569322976, "theoretical_loss": 3.40363444700572, "tokens_seen": 2198863872 }, { "epoch": 0.33, "learning_rate": 6.740491092922485e-05, "loss": 2.4982, "theoretical_loss": 3.403626145812918, "tokens_seen": 2198929408 }, { "epoch": 0.33, "learning_rate": 6.73968865350666e-05, "loss": 2.6289, "theoretical_loss": 3.4036095443773107, "tokens_seen": 2199060480 }, { "epoch": 0.33, "learning_rate": 6.738886214090836e-05, "loss": 2.5871, "theoretical_loss": 3.403592944208223, "tokens_seen": 2199191552 }, { "epoch": 0.33, "learning_rate": 6.738083774675013e-05, "loss": 2.4835, "theoretical_loss": 3.403576345305482, "tokens_seen": 2199322624 }, { "epoch": 0.33, "learning_rate": 6.737281335259188e-05, "loss": 2.4306, "theoretical_loss": 3.4035597476689166, "tokens_seen": 2199453696 }, { "epoch": 0.33, "learning_rate": 6.736478895843365e-05, "loss": 2.6061, "theoretical_loss": 3.4035431512983543, "tokens_seen": 2199584768 }, { "epoch": 0.33, "learning_rate": 6.73567645642754e-05, "loss": 2.5906, "theoretical_loss": 3.403526556193623, "tokens_seen": 2199715840 }, { "epoch": 0.33, "learning_rate": 6.734874017011717e-05, "loss": 2.4728, "theoretical_loss": 3.403509962354551, "tokens_seen": 2199846912 }, { "epoch": 0.33, "learning_rate": 6.734071577595892e-05, "loss": 2.4921, "theoretical_loss": 3.403493369780966, "tokens_seen": 2199977984 }, { "epoch": 0.33, "learning_rate": 6.733269138180067e-05, "loss": 2.6282, "theoretical_loss": 3.403476778472696, "tokens_seen": 2200109056 }, { "epoch": 0.33, "learning_rate": 6.732466698764244e-05, "loss": 2.442, "theoretical_loss": 3.4034601884295697, "tokens_seen": 2200240128 }, { "epoch": 0.33, "learning_rate": 6.731664259348419e-05, "loss": 2.4406, "theoretical_loss": 3.403443599651415, "tokens_seen": 2200371200 }, { "epoch": 0.33, "objective/train/docs_used": 1214287, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.592477560043335, "objective/train/theoretical_loss": 3.40342701213806, "objective/train/tokens_used": 570961376, "theoretical_loss": 3.40342701213806, "tokens_seen": 2200502272 }, { "epoch": 0.33, "learning_rate": 6.730861819932596e-05, "loss": 2.493, "theoretical_loss": 3.40342701213806, "tokens_seen": 2200502272 }, { "epoch": 0.33, "learning_rate": 6.730059380516771e-05, "loss": 2.3522, "theoretical_loss": 3.403410425889333, "tokens_seen": 2200633344 }, { "epoch": 0.33, "learning_rate": 6.729256941100948e-05, "loss": 2.3456, "theoretical_loss": 3.4033938409050624, "tokens_seen": 2200764416 }, { "epoch": 0.33, "learning_rate": 6.728454501685123e-05, "loss": 2.5727, "theoretical_loss": 3.403377257185076, "tokens_seen": 2200895488 }, { "epoch": 0.33, "learning_rate": 6.727652062269298e-05, "loss": 2.4925, "theoretical_loss": 3.4033606747292033, "tokens_seen": 2201026560 }, { "epoch": 0.33, "learning_rate": 6.726849622853475e-05, "loss": 2.5255, "theoretical_loss": 3.403344093537272, "tokens_seen": 2201157632 }, { "epoch": 0.33, "learning_rate": 6.72604718343765e-05, "loss": 2.3977, "theoretical_loss": 3.4033275136091095, "tokens_seen": 2201288704 }, { "epoch": 0.33, "learning_rate": 6.725244744021827e-05, "loss": 2.4072, "theoretical_loss": 3.4033109349445456, "tokens_seen": 2201419776 }, { "epoch": 0.33, "learning_rate": 6.724442304606002e-05, "loss": 2.414, "theoretical_loss": 3.4032943575434085, "tokens_seen": 2201550848 }, { "epoch": 0.33, "learning_rate": 6.723639865190179e-05, "loss": 2.488, "theoretical_loss": 3.4032777814055266, "tokens_seen": 2201681920 }, { "epoch": 0.33, "learning_rate": 6.722837425774354e-05, "loss": 2.4856, "theoretical_loss": 3.403261206530728, "tokens_seen": 2201812992 }, { "epoch": 0.33, "learning_rate": 6.722034986358531e-05, "loss": 2.5547, "theoretical_loss": 3.4032446329188417, "tokens_seen": 2201944064 }, { "epoch": 0.33, "learning_rate": 6.721232546942706e-05, "loss": 2.5414, "theoretical_loss": 3.4032280605696967, "tokens_seen": 2202075136 }, { "epoch": 0.33, "objective/train/docs_used": 1214811, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5659379959106445, "objective/train/theoretical_loss": 3.403219774868598, "objective/train/tokens_used": 572599776, "theoretical_loss": 3.403219774868598, "tokens_seen": 2202140672 }, { "epoch": 0.33, "learning_rate": 6.720430107526882e-05, "loss": 2.3994, "theoretical_loss": 3.403211489483121, "tokens_seen": 2202206208 }, { "epoch": 0.33, "learning_rate": 6.719627668111058e-05, "loss": 2.5061, "theoretical_loss": 3.4031949196589433, "tokens_seen": 2202337280 }, { "epoch": 0.33, "learning_rate": 6.718825228695234e-05, "loss": 2.4373, "theoretical_loss": 3.4031783510969924, "tokens_seen": 2202468352 }, { "epoch": 0.33, "learning_rate": 6.71802278927941e-05, "loss": 2.4786, "theoretical_loss": 3.4031617837970973, "tokens_seen": 2202599424 }, { "epoch": 0.34, "learning_rate": 6.717220349863586e-05, "loss": 2.429, "theoretical_loss": 3.4031452177590866, "tokens_seen": 2202730496 }, { "epoch": 0.34, "learning_rate": 6.716417910447762e-05, "loss": 2.5734, "theoretical_loss": 3.4031286529827893, "tokens_seen": 2202861568 }, { "epoch": 0.34, "learning_rate": 6.715615471031938e-05, "loss": 2.4266, "theoretical_loss": 3.4031120894680336, "tokens_seen": 2202992640 }, { "epoch": 0.34, "learning_rate": 6.714813031616113e-05, "loss": 2.4192, "theoretical_loss": 3.4030955272146493, "tokens_seen": 2203123712 }, { "epoch": 0.34, "learning_rate": 6.71401059220029e-05, "loss": 2.5598, "theoretical_loss": 3.4030789662224645, "tokens_seen": 2203254784 }, { "epoch": 0.34, "learning_rate": 6.713208152784465e-05, "loss": 2.6017, "theoretical_loss": 3.4030624064913084, "tokens_seen": 2203385856 }, { "epoch": 0.34, "learning_rate": 6.712405713368642e-05, "loss": 2.5483, "theoretical_loss": 3.40304584802101, "tokens_seen": 2203516928 }, { "epoch": 0.34, "learning_rate": 6.711603273952817e-05, "loss": 2.4384, "theoretical_loss": 3.4030292908113986, "tokens_seen": 2203648000 }, { "epoch": 0.34, "objective/train/docs_used": 1215803, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7451982498168945, "objective/train/theoretical_loss": 3.403012734862303, "objective/train/tokens_used": 574238176, "theoretical_loss": 3.403012734862303, "tokens_seen": 2203779072 }, { "epoch": 0.34, "learning_rate": 6.710800834536994e-05, "loss": 2.5776, "theoretical_loss": 3.403012734862303, "tokens_seen": 2203779072 }, { "epoch": 0.34, "learning_rate": 6.709998395121169e-05, "loss": 2.4848, "theoretical_loss": 3.402996180173552, "tokens_seen": 2203910144 }, { "epoch": 0.34, "learning_rate": 6.709195955705344e-05, "loss": 2.5394, "theoretical_loss": 3.4029796267449752, "tokens_seen": 2204041216 }, { "epoch": 0.34, "learning_rate": 6.708393516289521e-05, "loss": 2.5138, "theoretical_loss": 3.4029630745764012, "tokens_seen": 2204172288 }, { "epoch": 0.34, "learning_rate": 6.707591076873696e-05, "loss": 2.4304, "theoretical_loss": 3.4029465236676595, "tokens_seen": 2204303360 }, { "epoch": 0.34, "learning_rate": 6.706788637457873e-05, "loss": 2.3949, "theoretical_loss": 3.40292997401858, "tokens_seen": 2204434432 }, { "epoch": 0.34, "learning_rate": 6.705986198042048e-05, "loss": 2.422, "theoretical_loss": 3.4029134256289906, "tokens_seen": 2204565504 }, { "epoch": 0.34, "learning_rate": 6.705183758626225e-05, "loss": 2.4923, "theoretical_loss": 3.402896878498721, "tokens_seen": 2204696576 }, { "epoch": 0.34, "learning_rate": 6.7043813192104e-05, "loss": 2.3677, "theoretical_loss": 3.402880332627601, "tokens_seen": 2204827648 }, { "epoch": 0.34, "learning_rate": 6.703578879794575e-05, "loss": 2.5273, "theoretical_loss": 3.4028637880154595, "tokens_seen": 2204958720 }, { "epoch": 0.34, "learning_rate": 6.702776440378752e-05, "loss": 2.5261, "theoretical_loss": 3.4028472446621265, "tokens_seen": 2205089792 }, { "epoch": 0.34, "learning_rate": 6.701974000962927e-05, "loss": 2.4574, "theoretical_loss": 3.4028307025674303, "tokens_seen": 2205220864 }, { "epoch": 0.34, "learning_rate": 6.701171561547104e-05, "loss": 2.5115, "theoretical_loss": 3.402814161731201, "tokens_seen": 2205351936 }, { "epoch": 0.34, "objective/train/docs_used": 1216279, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.408686637878418, "objective/train/theoretical_loss": 3.4028058917849586, "objective/train/tokens_used": 575876576, "theoretical_loss": 3.4028058917849586, "tokens_seen": 2205417472 }, { "epoch": 0.34, "learning_rate": 6.70036912213128e-05, "loss": 2.556, "theoretical_loss": 3.4027976221532685, "tokens_seen": 2205483008 }, { "epoch": 0.34, "learning_rate": 6.699566682715456e-05, "loss": 2.4738, "theoretical_loss": 3.402781083833462, "tokens_seen": 2205614080 }, { "epoch": 0.34, "learning_rate": 6.698764243299631e-05, "loss": 2.4675, "theoretical_loss": 3.4027645467716106, "tokens_seen": 2205745152 }, { "epoch": 0.34, "learning_rate": 6.697961803883807e-05, "loss": 2.4454, "theoretical_loss": 3.4027480109675436, "tokens_seen": 2205876224 }, { "epoch": 0.34, "learning_rate": 6.697159364467983e-05, "loss": 2.6172, "theoretical_loss": 3.402731476421092, "tokens_seen": 2206007296 }, { "epoch": 0.34, "learning_rate": 6.696356925052159e-05, "loss": 2.403, "theoretical_loss": 3.402714943132084, "tokens_seen": 2206138368 }, { "epoch": 0.34, "learning_rate": 6.695554485636335e-05, "loss": 2.431, "theoretical_loss": 3.40269841110035, "tokens_seen": 2206269440 }, { "epoch": 0.34, "learning_rate": 6.694752046220511e-05, "loss": 2.4604, "theoretical_loss": 3.40268188032572, "tokens_seen": 2206400512 }, { "epoch": 0.34, "learning_rate": 6.693949606804687e-05, "loss": 2.4266, "theoretical_loss": 3.402665350808023, "tokens_seen": 2206531584 }, { "epoch": 0.34, "learning_rate": 6.693147167388863e-05, "loss": 2.56, "theoretical_loss": 3.4026488225470892, "tokens_seen": 2206662656 }, { "epoch": 0.34, "learning_rate": 6.69234472797304e-05, "loss": 2.5656, "theoretical_loss": 3.402632295542748, "tokens_seen": 2206793728 }, { "epoch": 0.34, "learning_rate": 6.691542288557215e-05, "loss": 2.4333, "theoretical_loss": 3.40261576979483, "tokens_seen": 2206924800 }, { "epoch": 0.34, "objective/train/docs_used": 1217338, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7593905925750732, "objective/train/theoretical_loss": 3.4025992453031644, "objective/train/tokens_used": 577514976, "theoretical_loss": 3.4025992453031644, "tokens_seen": 2207055872 }, { "epoch": 0.34, "learning_rate": 6.69073984914139e-05, "loss": 2.5397, "theoretical_loss": 3.4025992453031644, "tokens_seen": 2207055872 }, { "epoch": 0.34, "learning_rate": 6.689937409725567e-05, "loss": 2.6356, "theoretical_loss": 3.402582722067581, "tokens_seen": 2207186944 }, { "epoch": 0.34, "learning_rate": 6.689134970309742e-05, "loss": 2.5292, "theoretical_loss": 3.4025662000879104, "tokens_seen": 2207318016 }, { "epoch": 0.34, "learning_rate": 6.688332530893919e-05, "loss": 2.4998, "theoretical_loss": 3.4025496793639824, "tokens_seen": 2207449088 }, { "epoch": 0.34, "learning_rate": 6.687530091478094e-05, "loss": 2.4771, "theoretical_loss": 3.4025331598956265, "tokens_seen": 2207580160 }, { "epoch": 0.34, "learning_rate": 6.68672765206227e-05, "loss": 2.6201, "theoretical_loss": 3.402516641682673, "tokens_seen": 2207711232 }, { "epoch": 0.34, "learning_rate": 6.685925212646446e-05, "loss": 2.583, "theoretical_loss": 3.4025001247249524, "tokens_seen": 2207842304 }, { "epoch": 0.34, "learning_rate": 6.685122773230621e-05, "loss": 2.4937, "theoretical_loss": 3.402483609022294, "tokens_seen": 2207973376 }, { "epoch": 0.34, "learning_rate": 6.684320333814798e-05, "loss": 2.3755, "theoretical_loss": 3.402467094574529, "tokens_seen": 2208104448 }, { "epoch": 0.34, "learning_rate": 6.683517894398973e-05, "loss": 2.5587, "theoretical_loss": 3.4024505813814865, "tokens_seen": 2208235520 }, { "epoch": 0.34, "learning_rate": 6.68271545498315e-05, "loss": 2.5725, "theoretical_loss": 3.402434069442997, "tokens_seen": 2208366592 }, { "epoch": 0.34, "learning_rate": 6.681913015567325e-05, "loss": 2.5799, "theoretical_loss": 3.4024175587588914, "tokens_seen": 2208497664 }, { "epoch": 0.34, "learning_rate": 6.681110576151502e-05, "loss": 2.401, "theoretical_loss": 3.402401049328999, "tokens_seen": 2208628736 }, { "epoch": 0.34, "objective/train/docs_used": 1217846, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5893185138702393, "objective/train/theoretical_loss": 3.40239279508433, "objective/train/tokens_used": 579153376, "theoretical_loss": 3.40239279508433, "tokens_seen": 2208694272 }, { "epoch": 0.34, "learning_rate": 6.680308136735677e-05, "loss": 2.4925, "theoretical_loss": 3.402384541153151, "tokens_seen": 2208759808 }, { "epoch": 0.34, "learning_rate": 6.679505697319852e-05, "loss": 2.4016, "theoretical_loss": 3.402368034231177, "tokens_seen": 2208890880 }, { "epoch": 0.34, "learning_rate": 6.678703257904029e-05, "loss": 2.3912, "theoretical_loss": 3.4023515285629076, "tokens_seen": 2209021952 }, { "epoch": 0.34, "learning_rate": 6.677900818488204e-05, "loss": 2.4989, "theoretical_loss": 3.402335024148173, "tokens_seen": 2209153024 }, { "epoch": 0.34, "learning_rate": 6.677098379072381e-05, "loss": 2.553, "theoretical_loss": 3.4023185209868045, "tokens_seen": 2209284096 }, { "epoch": 0.34, "learning_rate": 6.676295939656556e-05, "loss": 2.5332, "theoretical_loss": 3.4023020190786317, "tokens_seen": 2209415168 }, { "epoch": 0.34, "learning_rate": 6.675493500240733e-05, "loss": 2.5386, "theoretical_loss": 3.4022855184234855, "tokens_seen": 2209546240 }, { "epoch": 0.34, "learning_rate": 6.674691060824908e-05, "loss": 2.4758, "theoretical_loss": 3.4022690190211957, "tokens_seen": 2209677312 }, { "epoch": 0.34, "learning_rate": 6.673888621409084e-05, "loss": 2.482, "theoretical_loss": 3.402252520871594, "tokens_seen": 2209808384 }, { "epoch": 0.34, "learning_rate": 6.67308618199326e-05, "loss": 2.5679, "theoretical_loss": 3.4022360239745106, "tokens_seen": 2209939456 }, { "epoch": 0.34, "learning_rate": 6.672283742577436e-05, "loss": 2.5476, "theoretical_loss": 3.4022195283297756, "tokens_seen": 2210070528 }, { "epoch": 0.34, "learning_rate": 6.671481303161612e-05, "loss": 2.669, "theoretical_loss": 3.4022030339372202, "tokens_seen": 2210201600 }, { "epoch": 0.34, "objective/train/docs_used": 1219246, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7586281299591064, "objective/train/theoretical_loss": 3.402186540796675, "objective/train/tokens_used": 580791776, "theoretical_loss": 3.402186540796675, "tokens_seen": 2210332672 }, { "epoch": 0.34, "learning_rate": 6.670678863745788e-05, "loss": 2.4485, "theoretical_loss": 3.402186540796675, "tokens_seen": 2210332672 }, { "epoch": 0.34, "learning_rate": 6.669876424329964e-05, "loss": 2.3993, "theoretical_loss": 3.4021700489079705, "tokens_seen": 2210463744 }, { "epoch": 0.34, "learning_rate": 6.66907398491414e-05, "loss": 2.4686, "theoretical_loss": 3.402153558270938, "tokens_seen": 2210594816 }, { "epoch": 0.34, "learning_rate": 6.668271545498315e-05, "loss": 2.4943, "theoretical_loss": 3.4021370688854073, "tokens_seen": 2210725888 }, { "epoch": 0.34, "learning_rate": 6.667469106082492e-05, "loss": 2.5063, "theoretical_loss": 3.4021205807512107, "tokens_seen": 2210856960 }, { "epoch": 0.34, "learning_rate": 6.666666666666667e-05, "loss": 2.4981, "theoretical_loss": 3.4021040938681777, "tokens_seen": 2210988032 }, { "epoch": 0.34, "learning_rate": 6.665864227250844e-05, "loss": 2.5183, "theoretical_loss": 3.40208760823614, "tokens_seen": 2211119104 }, { "epoch": 0.34, "learning_rate": 6.665061787835019e-05, "loss": 2.4552, "theoretical_loss": 3.402071123854928, "tokens_seen": 2211250176 }, { "epoch": 0.34, "learning_rate": 6.664259348419196e-05, "loss": 2.5751, "theoretical_loss": 3.4020546407243732, "tokens_seen": 2211381248 }, { "epoch": 0.34, "learning_rate": 6.663456909003371e-05, "loss": 2.5156, "theoretical_loss": 3.4020381588443063, "tokens_seen": 2211512320 }, { "epoch": 0.34, "learning_rate": 6.662654469587546e-05, "loss": 2.5568, "theoretical_loss": 3.402021678214558, "tokens_seen": 2211643392 }, { "epoch": 0.34, "learning_rate": 6.661852030171723e-05, "loss": 2.5904, "theoretical_loss": 3.40200519883496, "tokens_seen": 2211774464 }, { "epoch": 0.34, "learning_rate": 6.661049590755898e-05, "loss": 2.5605, "theoretical_loss": 3.4019887207053428, "tokens_seen": 2211905536 }, { "epoch": 0.34, "objective/train/docs_used": 1220526, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.844651937484741, "objective/train/theoretical_loss": 3.4019804821092245, "objective/train/tokens_used": 582430176, "theoretical_loss": 3.4019804821092245, "tokens_seen": 2211971072 }, { "epoch": 0.34, "learning_rate": 6.660247151340075e-05, "loss": 2.461, "theoretical_loss": 3.401972243825538, "tokens_seen": 2212036608 }, { "epoch": 0.34, "learning_rate": 6.65944471192425e-05, "loss": 2.6401, "theoretical_loss": 3.4019557681953767, "tokens_seen": 2212167680 }, { "epoch": 0.34, "learning_rate": 6.658642272508427e-05, "loss": 2.4676, "theoretical_loss": 3.4019392938146895, "tokens_seen": 2212298752 }, { "epoch": 0.34, "learning_rate": 6.657839833092602e-05, "loss": 2.5706, "theoretical_loss": 3.4019228206833088, "tokens_seen": 2212429824 }, { "epoch": 0.34, "learning_rate": 6.657037393676779e-05, "loss": 2.3169, "theoretical_loss": 3.401906348801065, "tokens_seen": 2212560896 }, { "epoch": 0.34, "learning_rate": 6.656234954260954e-05, "loss": 2.6073, "theoretical_loss": 3.401889878167789, "tokens_seen": 2212691968 }, { "epoch": 0.34, "learning_rate": 6.65543251484513e-05, "loss": 2.539, "theoretical_loss": 3.401873408783313, "tokens_seen": 2212823040 }, { "epoch": 0.34, "learning_rate": 6.654630075429306e-05, "loss": 2.5384, "theoretical_loss": 3.401856940647468, "tokens_seen": 2212954112 }, { "epoch": 0.34, "learning_rate": 6.653827636013481e-05, "loss": 2.5948, "theoretical_loss": 3.4018404737600854, "tokens_seen": 2213085184 }, { "epoch": 0.34, "learning_rate": 6.653025196597658e-05, "loss": 2.5626, "theoretical_loss": 3.4018240081209963, "tokens_seen": 2213216256 }, { "epoch": 0.34, "learning_rate": 6.652222757181833e-05, "loss": 2.561, "theoretical_loss": 3.401807543730033, "tokens_seen": 2213347328 }, { "epoch": 0.34, "learning_rate": 6.65142031776601e-05, "loss": 2.4798, "theoretical_loss": 3.401791080587026, "tokens_seen": 2213478400 }, { "epoch": 0.34, "objective/train/docs_used": 1221127, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5792646408081055, "objective/train/theoretical_loss": 3.4017746186918076, "objective/train/tokens_used": 584068576, "theoretical_loss": 3.4017746186918076, "tokens_seen": 2213609472 }, { "epoch": 0.34, "learning_rate": 6.650617878350185e-05, "loss": 2.6339, "theoretical_loss": 3.4017746186918076, "tokens_seen": 2213609472 }, { "epoch": 0.34, "learning_rate": 6.649815438934361e-05, "loss": 2.4253, "theoretical_loss": 3.401758158044209, "tokens_seen": 2213740544 }, { "epoch": 0.34, "learning_rate": 6.649012999518537e-05, "loss": 2.5425, "theoretical_loss": 3.4017416986440616, "tokens_seen": 2213871616 }, { "epoch": 0.34, "learning_rate": 6.648210560102713e-05, "loss": 2.5688, "theoretical_loss": 3.4017252404911975, "tokens_seen": 2214002688 }, { "epoch": 0.34, "learning_rate": 6.64740812068689e-05, "loss": 2.5209, "theoretical_loss": 3.4017087835854474, "tokens_seen": 2214133760 }, { "epoch": 0.34, "learning_rate": 6.646605681271065e-05, "loss": 2.5333, "theoretical_loss": 3.401692327926644, "tokens_seen": 2214264832 }, { "epoch": 0.34, "learning_rate": 6.645803241855241e-05, "loss": 2.4757, "theoretical_loss": 3.4016758735146193, "tokens_seen": 2214395904 }, { "epoch": 0.34, "learning_rate": 6.645000802439417e-05, "loss": 2.5999, "theoretical_loss": 3.401659420349204, "tokens_seen": 2214526976 }, { "epoch": 0.34, "learning_rate": 6.644198363023592e-05, "loss": 2.4715, "theoretical_loss": 3.40164296843023, "tokens_seen": 2214658048 }, { "epoch": 0.34, "learning_rate": 6.643395923607769e-05, "loss": 2.4961, "theoretical_loss": 3.40162651775753, "tokens_seen": 2214789120 }, { "epoch": 0.34, "learning_rate": 6.642593484191944e-05, "loss": 2.4219, "theoretical_loss": 3.4016100683309345, "tokens_seen": 2214920192 }, { "epoch": 0.34, "learning_rate": 6.64179104477612e-05, "loss": 2.4605, "theoretical_loss": 3.401593620150277, "tokens_seen": 2215051264 }, { "epoch": 0.34, "learning_rate": 6.640988605360296e-05, "loss": 2.5053, "theoretical_loss": 3.401577173215388, "tokens_seen": 2215182336 }, { "epoch": 0.34, "objective/train/docs_used": 1222538, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.598327398300171, "objective/train/theoretical_loss": 3.4015689502150543, "objective/train/tokens_used": 585706976, "theoretical_loss": 3.4015689502150543, "tokens_seen": 2215247872 }, { "epoch": 0.34, "learning_rate": 6.640186165944471e-05, "loss": 2.5099, "theoretical_loss": 3.4015607275260997, "tokens_seen": 2215313408 }, { "epoch": 0.34, "learning_rate": 6.639383726528648e-05, "loss": 2.517, "theoretical_loss": 3.401544283082245, "tokens_seen": 2215444480 }, { "epoch": 0.34, "learning_rate": 6.638581287112823e-05, "loss": 2.6561, "theoretical_loss": 3.401527839883655, "tokens_seen": 2215575552 }, { "epoch": 0.34, "learning_rate": 6.637778847697e-05, "loss": 2.4735, "theoretical_loss": 3.401511397930162, "tokens_seen": 2215706624 }, { "epoch": 0.34, "learning_rate": 6.636976408281175e-05, "loss": 2.5296, "theoretical_loss": 3.401494957221598, "tokens_seen": 2215837696 }, { "epoch": 0.34, "learning_rate": 6.636173968865352e-05, "loss": 2.6785, "theoretical_loss": 3.4014785177577953, "tokens_seen": 2215968768 }, { "epoch": 0.34, "learning_rate": 6.635371529449527e-05, "loss": 2.4534, "theoretical_loss": 3.401462079538586, "tokens_seen": 2216099840 }, { "epoch": 0.34, "learning_rate": 6.634569090033703e-05, "loss": 2.4605, "theoretical_loss": 3.401445642563802, "tokens_seen": 2216230912 }, { "epoch": 0.34, "learning_rate": 6.633766650617879e-05, "loss": 2.4743, "theoretical_loss": 3.401429206833276, "tokens_seen": 2216361984 }, { "epoch": 0.34, "learning_rate": 6.632964211202055e-05, "loss": 2.5421, "theoretical_loss": 3.40141277234684, "tokens_seen": 2216493056 }, { "epoch": 0.34, "learning_rate": 6.632161771786231e-05, "loss": 2.4795, "theoretical_loss": 3.401396339104326, "tokens_seen": 2216624128 }, { "epoch": 0.34, "learning_rate": 6.631359332370407e-05, "loss": 2.3755, "theoretical_loss": 3.401379907105566, "tokens_seen": 2216755200 }, { "epoch": 0.34, "objective/train/docs_used": 1222919, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.506207227706909, "objective/train/theoretical_loss": 3.4013634763503937, "objective/train/tokens_used": 587345376, "theoretical_loss": 3.4013634763503937, "tokens_seen": 2216886272 }, { "epoch": 0.34, "learning_rate": 6.630556892954582e-05, "loss": 2.54, "theoretical_loss": 3.4013634763503937, "tokens_seen": 2216886272 }, { "epoch": 0.34, "learning_rate": 6.629754453538758e-05, "loss": 2.6115, "theoretical_loss": 3.40134704683864, "tokens_seen": 2217017344 }, { "epoch": 0.34, "learning_rate": 6.628952014122934e-05, "loss": 2.6062, "theoretical_loss": 3.4013306185701384, "tokens_seen": 2217148416 }, { "epoch": 0.34, "learning_rate": 6.62814957470711e-05, "loss": 2.392, "theoretical_loss": 3.4013141915447207, "tokens_seen": 2217279488 }, { "epoch": 0.34, "learning_rate": 6.627347135291286e-05, "loss": 2.62, "theoretical_loss": 3.4012977657622194, "tokens_seen": 2217410560 }, { "epoch": 0.34, "learning_rate": 6.626544695875462e-05, "loss": 2.5453, "theoretical_loss": 3.401281341222467, "tokens_seen": 2217541632 }, { "epoch": 0.34, "learning_rate": 6.625742256459638e-05, "loss": 2.5269, "theoretical_loss": 3.4012649179252965, "tokens_seen": 2217672704 }, { "epoch": 0.34, "learning_rate": 6.624939817043813e-05, "loss": 2.4656, "theoretical_loss": 3.4012484958705396, "tokens_seen": 2217803776 }, { "epoch": 0.34, "learning_rate": 6.62413737762799e-05, "loss": 2.4634, "theoretical_loss": 3.40123207505803, "tokens_seen": 2217934848 }, { "epoch": 0.34, "learning_rate": 6.623334938212165e-05, "loss": 2.5549, "theoretical_loss": 3.401215655487599, "tokens_seen": 2218065920 }, { "epoch": 0.34, "learning_rate": 6.622532498796342e-05, "loss": 2.4937, "theoretical_loss": 3.4011992371590805, "tokens_seen": 2218196992 }, { "epoch": 0.34, "learning_rate": 6.621730059380517e-05, "loss": 2.5013, "theoretical_loss": 3.4011828200723064, "tokens_seen": 2218328064 }, { "epoch": 0.34, "learning_rate": 6.620927619964692e-05, "loss": 2.6017, "theoretical_loss": 3.40116640422711, "tokens_seen": 2218459136 }, { "epoch": 0.34, "objective/train/docs_used": 1223993, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.066260814666748, "objective/train/theoretical_loss": 3.4011581967700506, "objective/train/tokens_used": 588983776, "theoretical_loss": 3.4011581967700506, "tokens_seen": 2218524672 }, { "epoch": 0.34, "learning_rate": 6.620125180548869e-05, "loss": 2.5717, "theoretical_loss": 3.401149989623323, "tokens_seen": 2218590208 }, { "epoch": 0.34, "learning_rate": 6.619322741133044e-05, "loss": 2.7128, "theoretical_loss": 3.4011335762607797, "tokens_seen": 2218721280 }, { "epoch": 0.34, "learning_rate": 6.618520301717221e-05, "loss": 2.6112, "theoretical_loss": 3.401117164139312, "tokens_seen": 2218852352 }, { "epoch": 0.34, "learning_rate": 6.617717862301396e-05, "loss": 2.5601, "theoretical_loss": 3.4011007532587527, "tokens_seen": 2218983424 }, { "epoch": 0.34, "learning_rate": 6.616915422885573e-05, "loss": 2.466, "theoretical_loss": 3.401084343618935, "tokens_seen": 2219114496 }, { "epoch": 0.35, "learning_rate": 6.616112983469748e-05, "loss": 2.5962, "theoretical_loss": 3.4010679352196918, "tokens_seen": 2219245568 }, { "epoch": 0.35, "learning_rate": 6.615310544053924e-05, "loss": 2.6519, "theoretical_loss": 3.4010515280608558, "tokens_seen": 2219376640 }, { "epoch": 0.35, "learning_rate": 6.6145081046381e-05, "loss": 2.6243, "theoretical_loss": 3.40103512214226, "tokens_seen": 2219507712 }, { "epoch": 0.35, "learning_rate": 6.613705665222276e-05, "loss": 2.4531, "theoretical_loss": 3.401018717463738, "tokens_seen": 2219638784 }, { "epoch": 0.35, "learning_rate": 6.612903225806452e-05, "loss": 2.6186, "theoretical_loss": 3.401002314025122, "tokens_seen": 2219769856 }, { "epoch": 0.35, "learning_rate": 6.612100786390628e-05, "loss": 2.3972, "theoretical_loss": 3.4009859118262455, "tokens_seen": 2219900928 }, { "epoch": 0.35, "learning_rate": 6.611298346974803e-05, "loss": 2.6535, "theoretical_loss": 3.400969510866942, "tokens_seen": 2220032000 }, { "epoch": 0.35, "objective/train/docs_used": 1224628, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7098729610443115, "objective/train/theoretical_loss": 3.400953111147044, "objective/train/tokens_used": 590622176, "theoretical_loss": 3.400953111147044, "tokens_seen": 2220163072 }, { "epoch": 0.35, "learning_rate": 6.61049590755898e-05, "loss": 2.4975, "theoretical_loss": 3.400953111147044, "tokens_seen": 2220163072 }, { "epoch": 0.35, "learning_rate": 6.609693468143155e-05, "loss": 2.5683, "theoretical_loss": 3.4009367126663848, "tokens_seen": 2220294144 }, { "epoch": 0.35, "learning_rate": 6.608891028727332e-05, "loss": 2.6508, "theoretical_loss": 3.4009203154247976, "tokens_seen": 2220425216 }, { "epoch": 0.35, "learning_rate": 6.608088589311507e-05, "loss": 2.6241, "theoretical_loss": 3.4009039194221162, "tokens_seen": 2220556288 }, { "epoch": 0.35, "learning_rate": 6.607286149895684e-05, "loss": 2.5669, "theoretical_loss": 3.4008875246581733, "tokens_seen": 2220687360 }, { "epoch": 0.35, "learning_rate": 6.606483710479859e-05, "loss": 2.4826, "theoretical_loss": 3.400871131132802, "tokens_seen": 2220818432 }, { "epoch": 0.35, "learning_rate": 6.605681271064034e-05, "loss": 2.3772, "theoretical_loss": 3.4008547388458363, "tokens_seen": 2220949504 }, { "epoch": 0.35, "learning_rate": 6.604878831648211e-05, "loss": 2.6465, "theoretical_loss": 3.400838347797109, "tokens_seen": 2221080576 }, { "epoch": 0.35, "learning_rate": 6.604076392232386e-05, "loss": 2.5433, "theoretical_loss": 3.4008219579864543, "tokens_seen": 2221211648 }, { "epoch": 0.35, "learning_rate": 6.603273952816563e-05, "loss": 2.4825, "theoretical_loss": 3.4008055694137047, "tokens_seen": 2221342720 }, { "epoch": 0.35, "learning_rate": 6.602471513400738e-05, "loss": 2.5396, "theoretical_loss": 3.400789182078694, "tokens_seen": 2221473792 }, { "epoch": 0.35, "learning_rate": 6.601669073984913e-05, "loss": 2.5519, "theoretical_loss": 3.400772795981256, "tokens_seen": 2221604864 }, { "epoch": 0.35, "learning_rate": 6.60086663456909e-05, "loss": 2.5309, "theoretical_loss": 3.4007564111212236, "tokens_seen": 2221735936 }, { "epoch": 0.35, "objective/train/docs_used": 1225688, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.183415651321411, "objective/train/theoretical_loss": 3.400748219155183, "objective/train/tokens_used": 592260576, "theoretical_loss": 3.400748219155183, "tokens_seen": 2221801472 }, { "epoch": 0.35, "learning_rate": 6.600064195153265e-05, "loss": 2.4961, "theoretical_loss": 3.400740027498431, "tokens_seen": 2221867008 }, { "epoch": 0.35, "learning_rate": 6.599261755737442e-05, "loss": 2.662, "theoretical_loss": 3.400723645112712, "tokens_seen": 2221998080 }, { "epoch": 0.35, "learning_rate": 6.598459316321617e-05, "loss": 2.5004, "theoretical_loss": 3.400707263963899, "tokens_seen": 2222129152 }, { "epoch": 0.35, "learning_rate": 6.597656876905794e-05, "loss": 2.5408, "theoretical_loss": 3.400690884051827, "tokens_seen": 2222260224 }, { "epoch": 0.35, "learning_rate": 6.59685443748997e-05, "loss": 2.4789, "theoretical_loss": 3.4006745053763288, "tokens_seen": 2222391296 }, { "epoch": 0.35, "learning_rate": 6.596051998074145e-05, "loss": 2.5802, "theoretical_loss": 3.4006581279372385, "tokens_seen": 2222522368 }, { "epoch": 0.35, "learning_rate": 6.595249558658321e-05, "loss": 2.4772, "theoretical_loss": 3.4006417517343897, "tokens_seen": 2222653440 }, { "epoch": 0.35, "learning_rate": 6.594447119242497e-05, "loss": 2.7164, "theoretical_loss": 3.4006253767676164, "tokens_seen": 2222784512 }, { "epoch": 0.35, "learning_rate": 6.593644679826673e-05, "loss": 2.5849, "theoretical_loss": 3.4006090030367524, "tokens_seen": 2222915584 }, { "epoch": 0.35, "learning_rate": 6.592842240410849e-05, "loss": 2.5573, "theoretical_loss": 3.4005926305416314, "tokens_seen": 2223046656 }, { "epoch": 0.35, "learning_rate": 6.592039800995025e-05, "loss": 2.5864, "theoretical_loss": 3.4005762592820874, "tokens_seen": 2223177728 }, { "epoch": 0.35, "learning_rate": 6.5912373615792e-05, "loss": 2.4563, "theoretical_loss": 3.4005598892579543, "tokens_seen": 2223308800 }, { "epoch": 0.35, "objective/train/docs_used": 1226788, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7860403060913086, "objective/train/theoretical_loss": 3.400543520469066, "objective/train/tokens_used": 593898976, "theoretical_loss": 3.400543520469066, "tokens_seen": 2223439872 }, { "epoch": 0.35, "learning_rate": 6.590434922163376e-05, "loss": 2.7357, "theoretical_loss": 3.400543520469066, "tokens_seen": 2223439872 }, { "epoch": 0.35, "learning_rate": 6.589632482747553e-05, "loss": 2.5975, "theoretical_loss": 3.4005271529152563, "tokens_seen": 2223570944 }, { "epoch": 0.35, "learning_rate": 6.588830043331728e-05, "loss": 2.5676, "theoretical_loss": 3.4005107865963597, "tokens_seen": 2223702016 }, { "epoch": 0.35, "learning_rate": 6.588027603915905e-05, "loss": 2.3776, "theoretical_loss": 3.4004944215122097, "tokens_seen": 2223833088 }, { "epoch": 0.35, "learning_rate": 6.58722516450008e-05, "loss": 2.4822, "theoretical_loss": 3.400478057662641, "tokens_seen": 2223964160 }, { "epoch": 0.35, "learning_rate": 6.586422725084257e-05, "loss": 2.6994, "theoretical_loss": 3.400461695047487, "tokens_seen": 2224095232 }, { "epoch": 0.35, "learning_rate": 6.585620285668432e-05, "loss": 2.5839, "theoretical_loss": 3.4004453336665827, "tokens_seen": 2224226304 }, { "epoch": 0.35, "learning_rate": 6.584817846252607e-05, "loss": 2.5777, "theoretical_loss": 3.4004289735197615, "tokens_seen": 2224357376 }, { "epoch": 0.35, "learning_rate": 6.584015406836784e-05, "loss": 2.6066, "theoretical_loss": 3.4004126146068576, "tokens_seen": 2224488448 }, { "epoch": 0.35, "learning_rate": 6.583212967420959e-05, "loss": 2.553, "theoretical_loss": 3.4003962569277055, "tokens_seen": 2224619520 }, { "epoch": 0.35, "learning_rate": 6.582410528005136e-05, "loss": 2.5847, "theoretical_loss": 3.40037990048214, "tokens_seen": 2224750592 }, { "epoch": 0.35, "learning_rate": 6.581608088589311e-05, "loss": 2.6418, "theoretical_loss": 3.4003635452699945, "tokens_seen": 2224881664 }, { "epoch": 0.35, "learning_rate": 6.580805649173488e-05, "loss": 2.5826, "theoretical_loss": 3.400347191291104, "tokens_seen": 2225012736 }, { "epoch": 0.35, "objective/train/docs_used": 1227393, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.0688304901123047, "objective/train/theoretical_loss": 3.4003390147640773, "objective/train/tokens_used": 595537376, "theoretical_loss": 3.4003390147640773, "tokens_seen": 2225078272 }, { "epoch": 0.35, "learning_rate": 6.580003209757663e-05, "loss": 2.7564, "theoretical_loss": 3.4003308385453024, "tokens_seen": 2225143808 }, { "epoch": 0.35, "learning_rate": 6.579200770341838e-05, "loss": 2.5797, "theoretical_loss": 3.400314487032424, "tokens_seen": 2225274880 }, { "epoch": 0.35, "learning_rate": 6.578398330926015e-05, "loss": 2.5858, "theoretical_loss": 3.400298136752304, "tokens_seen": 2225405952 }, { "epoch": 0.35, "learning_rate": 6.57759589151019e-05, "loss": 2.5401, "theoretical_loss": 3.400281787704776, "tokens_seen": 2225537024 }, { "epoch": 0.35, "learning_rate": 6.576793452094367e-05, "loss": 2.6559, "theoretical_loss": 3.4002654398896754, "tokens_seen": 2225668096 }, { "epoch": 0.35, "learning_rate": 6.575991012678542e-05, "loss": 2.4982, "theoretical_loss": 3.4002490933068357, "tokens_seen": 2225799168 }, { "epoch": 0.35, "learning_rate": 6.575188573262719e-05, "loss": 2.4992, "theoretical_loss": 3.4002327479560925, "tokens_seen": 2225930240 }, { "epoch": 0.35, "learning_rate": 6.574386133846894e-05, "loss": 2.5146, "theoretical_loss": 3.4002164038372795, "tokens_seen": 2226061312 }, { "epoch": 0.35, "learning_rate": 6.57358369443107e-05, "loss": 2.6021, "theoretical_loss": 3.4002000609502314, "tokens_seen": 2226192384 }, { "epoch": 0.35, "learning_rate": 6.572781255015246e-05, "loss": 2.5102, "theoretical_loss": 3.4001837192947835, "tokens_seen": 2226323456 }, { "epoch": 0.35, "learning_rate": 6.571978815599422e-05, "loss": 2.5191, "theoretical_loss": 3.40016737887077, "tokens_seen": 2226454528 }, { "epoch": 0.35, "learning_rate": 6.571176376183598e-05, "loss": 2.5819, "theoretical_loss": 3.4001510396780255, "tokens_seen": 2226585600 }, { "epoch": 0.35, "objective/train/docs_used": 1228788, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.50097393989563, "objective/train/theoretical_loss": 3.400134701716385, "objective/train/tokens_used": 597175776, "theoretical_loss": 3.400134701716385, "tokens_seen": 2226716672 }, { "epoch": 0.35, "learning_rate": 6.570373936767774e-05, "loss": 2.5784, "theoretical_loss": 3.400134701716385, "tokens_seen": 2226716672 }, { "epoch": 0.35, "learning_rate": 6.56957149735195e-05, "loss": 2.359, "theoretical_loss": 3.400118364985684, "tokens_seen": 2226847744 }, { "epoch": 0.35, "learning_rate": 6.568769057936126e-05, "loss": 2.6825, "theoretical_loss": 3.4001020294857556, "tokens_seen": 2226978816 }, { "epoch": 0.35, "learning_rate": 6.567966618520301e-05, "loss": 2.4631, "theoretical_loss": 3.400085695216436, "tokens_seen": 2227109888 }, { "epoch": 0.35, "learning_rate": 6.567164179104478e-05, "loss": 2.4789, "theoretical_loss": 3.40006936217756, "tokens_seen": 2227240960 }, { "epoch": 0.35, "learning_rate": 6.566361739688653e-05, "loss": 2.3606, "theoretical_loss": 3.400053030368962, "tokens_seen": 2227372032 }, { "epoch": 0.35, "learning_rate": 6.56555930027283e-05, "loss": 2.4429, "theoretical_loss": 3.400036699790477, "tokens_seen": 2227503104 }, { "epoch": 0.35, "learning_rate": 6.564756860857005e-05, "loss": 2.398, "theoretical_loss": 3.4000203704419403, "tokens_seen": 2227634176 }, { "epoch": 0.35, "learning_rate": 6.563954421441182e-05, "loss": 2.6781, "theoretical_loss": 3.4000040423231863, "tokens_seen": 2227765248 }, { "epoch": 0.35, "learning_rate": 6.563151982025357e-05, "loss": 2.4473, "theoretical_loss": 3.399987715434051, "tokens_seen": 2227896320 }, { "epoch": 0.35, "learning_rate": 6.562349542609534e-05, "loss": 2.6737, "theoretical_loss": 3.3999713897743686, "tokens_seen": 2228027392 }, { "epoch": 0.35, "learning_rate": 6.561547103193709e-05, "loss": 2.5061, "theoretical_loss": 3.399955065343975, "tokens_seen": 2228158464 }, { "epoch": 0.35, "learning_rate": 6.560744663777884e-05, "loss": 2.609, "theoretical_loss": 3.3999387421427043, "tokens_seen": 2228289536 }, { "epoch": 0.35, "objective/train/docs_used": 1229475, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.80594801902771, "objective/train/theoretical_loss": 3.3999305810029385, "objective/train/tokens_used": 598814176, "theoretical_loss": 3.3999305810029385, "tokens_seen": 2228355072 }, { "epoch": 0.35, "learning_rate": 6.559942224362061e-05, "loss": 2.5804, "theoretical_loss": 3.3999224201703924, "tokens_seen": 2228420608 }, { "epoch": 0.35, "learning_rate": 6.559139784946236e-05, "loss": 2.6122, "theoretical_loss": 3.3999060994268744, "tokens_seen": 2228551680 }, { "epoch": 0.35, "learning_rate": 6.558337345530413e-05, "loss": 2.3885, "theoretical_loss": 3.399889779911985, "tokens_seen": 2228682752 }, { "epoch": 0.35, "learning_rate": 6.557534906114588e-05, "loss": 2.5099, "theoretical_loss": 3.3998734616255604, "tokens_seen": 2228813824 }, { "epoch": 0.35, "learning_rate": 6.556732466698765e-05, "loss": 2.57, "theoretical_loss": 3.3998571445674353, "tokens_seen": 2228944896 }, { "epoch": 0.35, "learning_rate": 6.55593002728294e-05, "loss": 2.554, "theoretical_loss": 3.399840828737445, "tokens_seen": 2229075968 }, { "epoch": 0.35, "learning_rate": 6.555127587867115e-05, "loss": 2.4743, "theoretical_loss": 3.3998245141354255, "tokens_seen": 2229207040 }, { "epoch": 0.35, "learning_rate": 6.554325148451292e-05, "loss": 2.4917, "theoretical_loss": 3.3998082007612114, "tokens_seen": 2229338112 }, { "epoch": 0.35, "learning_rate": 6.553522709035467e-05, "loss": 2.6752, "theoretical_loss": 3.3997918886146383, "tokens_seen": 2229469184 }, { "epoch": 0.35, "learning_rate": 6.552720269619644e-05, "loss": 2.4935, "theoretical_loss": 3.399775577695541, "tokens_seen": 2229600256 }, { "epoch": 0.35, "learning_rate": 6.55191783020382e-05, "loss": 2.5293, "theoretical_loss": 3.3997592680037565, "tokens_seen": 2229731328 }, { "epoch": 0.35, "learning_rate": 6.551115390787996e-05, "loss": 2.4314, "theoretical_loss": 3.3997429595391195, "tokens_seen": 2229862400 }, { "epoch": 0.35, "objective/train/docs_used": 1230697, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2199413776397705, "objective/train/theoretical_loss": 3.3997266523014655, "objective/train/tokens_used": 600452576, "theoretical_loss": 3.3997266523014655, "tokens_seen": 2229993472 }, { "epoch": 0.35, "learning_rate": 6.550312951372171e-05, "loss": 2.3866, "theoretical_loss": 3.3997266523014655, "tokens_seen": 2229993472 }, { "epoch": 0.35, "learning_rate": 6.549510511956347e-05, "loss": 2.6148, "theoretical_loss": 3.39971034629063, "tokens_seen": 2230124544 }, { "epoch": 0.35, "learning_rate": 6.548708072540523e-05, "loss": 2.4489, "theoretical_loss": 3.399694041506449, "tokens_seen": 2230255616 }, { "epoch": 0.35, "learning_rate": 6.547905633124699e-05, "loss": 2.6336, "theoretical_loss": 3.3996777379487577, "tokens_seen": 2230386688 }, { "epoch": 0.35, "learning_rate": 6.547103193708875e-05, "loss": 2.5653, "theoretical_loss": 3.3996614356173924, "tokens_seen": 2230517760 }, { "epoch": 0.35, "learning_rate": 6.54630075429305e-05, "loss": 2.5572, "theoretical_loss": 3.3996451345121876, "tokens_seen": 2230648832 }, { "epoch": 0.35, "learning_rate": 6.545498314877227e-05, "loss": 2.4734, "theoretical_loss": 3.3996288346329804, "tokens_seen": 2230779904 }, { "epoch": 0.35, "learning_rate": 6.544695875461403e-05, "loss": 2.5562, "theoretical_loss": 3.399612535979606, "tokens_seen": 2230910976 }, { "epoch": 0.35, "learning_rate": 6.543893436045578e-05, "loss": 2.5883, "theoretical_loss": 3.3995962385519, "tokens_seen": 2231042048 }, { "epoch": 0.35, "learning_rate": 6.543090996629755e-05, "loss": 2.6082, "theoretical_loss": 3.399579942349698, "tokens_seen": 2231173120 }, { "epoch": 0.35, "learning_rate": 6.54228855721393e-05, "loss": 2.5796, "theoretical_loss": 3.399563647372837, "tokens_seen": 2231304192 }, { "epoch": 0.35, "learning_rate": 6.541486117798107e-05, "loss": 2.5459, "theoretical_loss": 3.3995473536211516, "tokens_seen": 2231435264 }, { "epoch": 0.35, "learning_rate": 6.540683678382282e-05, "loss": 2.6118, "theoretical_loss": 3.3995310610944784, "tokens_seen": 2231566336 }, { "epoch": 0.35, "objective/train/docs_used": 1231344, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.63580322265625, "objective/train/theoretical_loss": 3.3995229152904702, "objective/train/tokens_used": 602090976, "theoretical_loss": 3.3995229152904702, "tokens_seen": 2231631872 }, { "epoch": 0.35, "learning_rate": 6.539881238966459e-05, "loss": 2.5398, "theoretical_loss": 3.3995147697926535, "tokens_seen": 2231697408 }, { "epoch": 0.35, "learning_rate": 6.539078799550634e-05, "loss": 2.4789, "theoretical_loss": 3.399498479715512, "tokens_seen": 2231828480 }, { "epoch": 0.35, "learning_rate": 6.538276360134809e-05, "loss": 2.7088, "theoretical_loss": 3.399482190862891, "tokens_seen": 2231959552 }, { "epoch": 0.35, "learning_rate": 6.537473920718986e-05, "loss": 2.4904, "theoretical_loss": 3.3994659032346264, "tokens_seen": 2232090624 }, { "epoch": 0.35, "learning_rate": 6.536671481303161e-05, "loss": 2.6163, "theoretical_loss": 3.3994496168305535, "tokens_seen": 2232221696 }, { "epoch": 0.35, "learning_rate": 6.535869041887338e-05, "loss": 2.6065, "theoretical_loss": 3.399433331650509, "tokens_seen": 2232352768 }, { "epoch": 0.35, "learning_rate": 6.535066602471513e-05, "loss": 2.6058, "theoretical_loss": 3.399417047694329, "tokens_seen": 2232483840 }, { "epoch": 0.35, "learning_rate": 6.53426416305569e-05, "loss": 2.5078, "theoretical_loss": 3.3994007649618494, "tokens_seen": 2232614912 }, { "epoch": 0.35, "learning_rate": 6.533461723639865e-05, "loss": 2.5577, "theoretical_loss": 3.399384483452907, "tokens_seen": 2232745984 }, { "epoch": 0.35, "learning_rate": 6.532659284224042e-05, "loss": 2.5347, "theoretical_loss": 3.3993682031673376, "tokens_seen": 2232877056 }, { "epoch": 0.35, "learning_rate": 6.531856844808217e-05, "loss": 2.6003, "theoretical_loss": 3.399351924104977, "tokens_seen": 2233008128 }, { "epoch": 0.35, "learning_rate": 6.531054405392392e-05, "loss": 2.4763, "theoretical_loss": 3.399335646265663, "tokens_seen": 2233139200 }, { "epoch": 0.35, "objective/train/docs_used": 1232383, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.546870708465576, "objective/train/theoretical_loss": 3.3993193696492305, "objective/train/tokens_used": 603729376, "theoretical_loss": 3.3993193696492305, "tokens_seen": 2233270272 }, { "epoch": 0.35, "learning_rate": 6.530251965976569e-05, "loss": 2.5945, "theoretical_loss": 3.3993193696492305, "tokens_seen": 2233270272 }, { "epoch": 0.35, "learning_rate": 6.529449526560744e-05, "loss": 2.5399, "theoretical_loss": 3.399303094255516, "tokens_seen": 2233401344 }, { "epoch": 0.35, "learning_rate": 6.528647087144921e-05, "loss": 2.5345, "theoretical_loss": 3.3992868200843565, "tokens_seen": 2233532416 }, { "epoch": 0.35, "learning_rate": 6.527844647729096e-05, "loss": 2.5707, "theoretical_loss": 3.399270547135588, "tokens_seen": 2233663488 }, { "epoch": 0.35, "learning_rate": 6.527042208313273e-05, "loss": 2.4072, "theoretical_loss": 3.3992542754090476, "tokens_seen": 2233794560 }, { "epoch": 0.35, "learning_rate": 6.526239768897448e-05, "loss": 2.6307, "theoretical_loss": 3.399238004904571, "tokens_seen": 2233925632 }, { "epoch": 0.35, "learning_rate": 6.525437329481624e-05, "loss": 2.5343, "theoretical_loss": 3.399221735621995, "tokens_seen": 2234056704 }, { "epoch": 0.35, "learning_rate": 6.5246348900658e-05, "loss": 2.5621, "theoretical_loss": 3.3992054675611563, "tokens_seen": 2234187776 }, { "epoch": 0.35, "learning_rate": 6.523832450649976e-05, "loss": 2.5342, "theoretical_loss": 3.3991892007218913, "tokens_seen": 2234318848 }, { "epoch": 0.35, "learning_rate": 6.523030011234152e-05, "loss": 2.6233, "theoretical_loss": 3.399172935104037, "tokens_seen": 2234449920 }, { "epoch": 0.35, "learning_rate": 6.522227571818328e-05, "loss": 2.464, "theoretical_loss": 3.3991566707074297, "tokens_seen": 2234580992 }, { "epoch": 0.35, "learning_rate": 6.521425132402504e-05, "loss": 2.5425, "theoretical_loss": 3.399140407531906, "tokens_seen": 2234712064 }, { "epoch": 0.35, "learning_rate": 6.52062269298668e-05, "loss": 2.6549, "theoretical_loss": 3.3991241455773027, "tokens_seen": 2234843136 }, { "epoch": 0.35, "objective/train/docs_used": 1233495, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.886728525161743, "objective/train/theoretical_loss": 3.3991160150577953, "objective/train/tokens_used": 605367776, "theoretical_loss": 3.3991160150577953, "tokens_seen": 2234908672 }, { "epoch": 0.35, "learning_rate": 6.519820253570855e-05, "loss": 2.5483, "theoretical_loss": 3.3991078848434566, "tokens_seen": 2234974208 }, { "epoch": 0.35, "learning_rate": 6.519017814155032e-05, "loss": 2.5349, "theoretical_loss": 3.399091625330205, "tokens_seen": 2235105280 }, { "epoch": 0.35, "learning_rate": 6.518215374739207e-05, "loss": 2.5666, "theoretical_loss": 3.3990753670373834, "tokens_seen": 2235236352 }, { "epoch": 0.35, "learning_rate": 6.517412935323384e-05, "loss": 2.554, "theoretical_loss": 3.3990591099648295, "tokens_seen": 2235367424 }, { "epoch": 0.35, "learning_rate": 6.516610495907559e-05, "loss": 2.4128, "theoretical_loss": 3.3990428541123805, "tokens_seen": 2235498496 }, { "epoch": 0.35, "learning_rate": 6.515808056491736e-05, "loss": 2.6454, "theoretical_loss": 3.399026599479873, "tokens_seen": 2235629568 }, { "epoch": 0.36, "learning_rate": 6.515005617075911e-05, "loss": 2.5925, "theoretical_loss": 3.3990103460671435, "tokens_seen": 2235760640 }, { "epoch": 0.36, "learning_rate": 6.514203177660086e-05, "loss": 2.4704, "theoretical_loss": 3.398994093874029, "tokens_seen": 2235891712 }, { "epoch": 0.36, "learning_rate": 6.513400738244263e-05, "loss": 2.4929, "theoretical_loss": 3.398977842900367, "tokens_seen": 2236022784 }, { "epoch": 0.36, "learning_rate": 6.512598298828438e-05, "loss": 2.7658, "theoretical_loss": 3.3989615931459944, "tokens_seen": 2236153856 }, { "epoch": 0.36, "learning_rate": 6.511795859412615e-05, "loss": 2.6872, "theoretical_loss": 3.3989453446107483, "tokens_seen": 2236284928 }, { "epoch": 0.36, "learning_rate": 6.51099341999679e-05, "loss": 2.4934, "theoretical_loss": 3.3989290972944657, "tokens_seen": 2236416000 }, { "epoch": 0.36, "objective/train/docs_used": 1234013, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2317135334014893, "objective/train/theoretical_loss": 3.3989128511969833, "objective/train/tokens_used": 607006176, "theoretical_loss": 3.3989128511969833, "tokens_seen": 2236547072 }, { "epoch": 0.36, "learning_rate": 6.510190980580967e-05, "loss": 2.5527, "theoretical_loss": 3.3989128511969833, "tokens_seen": 2236547072 }, { "epoch": 0.36, "learning_rate": 6.509388541165142e-05, "loss": 2.527, "theoretical_loss": 3.398896606318139, "tokens_seen": 2236678144 }, { "epoch": 0.36, "learning_rate": 6.508586101749317e-05, "loss": 2.5679, "theoretical_loss": 3.3988803626577697, "tokens_seen": 2236809216 }, { "epoch": 0.36, "learning_rate": 6.507783662333494e-05, "loss": 2.615, "theoretical_loss": 3.398864120215712, "tokens_seen": 2236940288 }, { "epoch": 0.36, "learning_rate": 6.50698122291767e-05, "loss": 2.3885, "theoretical_loss": 3.398847878991804, "tokens_seen": 2237071360 }, { "epoch": 0.36, "learning_rate": 6.506178783501846e-05, "loss": 2.5451, "theoretical_loss": 3.398831638985883, "tokens_seen": 2237202432 }, { "epoch": 0.36, "learning_rate": 6.505376344086021e-05, "loss": 2.547, "theoretical_loss": 3.398815400197785, "tokens_seen": 2237333504 }, { "epoch": 0.36, "learning_rate": 6.504573904670198e-05, "loss": 2.504, "theoretical_loss": 3.3987991626273493, "tokens_seen": 2237464576 }, { "epoch": 0.36, "learning_rate": 6.503771465254373e-05, "loss": 2.6316, "theoretical_loss": 3.3987829262744116, "tokens_seen": 2237595648 }, { "epoch": 0.36, "learning_rate": 6.50296902583855e-05, "loss": 2.5986, "theoretical_loss": 3.3987666911388104, "tokens_seen": 2237726720 }, { "epoch": 0.36, "learning_rate": 6.502166586422725e-05, "loss": 2.3704, "theoretical_loss": 3.398750457220382, "tokens_seen": 2237857792 }, { "epoch": 0.36, "learning_rate": 6.501364147006901e-05, "loss": 2.733, "theoretical_loss": 3.3987342245189653, "tokens_seen": 2237988864 }, { "epoch": 0.36, "learning_rate": 6.500561707591077e-05, "loss": 2.6359, "theoretical_loss": 3.398717993034397, "tokens_seen": 2238119936 }, { "epoch": 0.36, "objective/train/docs_used": 1235138, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.893874406814575, "objective/train/theoretical_loss": 3.39870987774838, "objective/train/tokens_used": 608644576, "theoretical_loss": 3.39870987774838, "tokens_seen": 2238185472 }, { "epoch": 0.36, "learning_rate": 6.499759268175253e-05, "loss": 2.5401, "theoretical_loss": 3.3987017627665144, "tokens_seen": 2238251008 }, { "epoch": 0.36, "learning_rate": 6.49895682875943e-05, "loss": 2.3885, "theoretical_loss": 3.398685533715155, "tokens_seen": 2238382080 }, { "epoch": 0.36, "learning_rate": 6.498154389343605e-05, "loss": 2.6437, "theoretical_loss": 3.3986693058801576, "tokens_seen": 2238513152 }, { "epoch": 0.36, "learning_rate": 6.497351949927781e-05, "loss": 2.5716, "theoretical_loss": 3.3986530792613583, "tokens_seen": 2238644224 }, { "epoch": 0.36, "learning_rate": 6.496549510511957e-05, "loss": 2.6233, "theoretical_loss": 3.3986368538585956, "tokens_seen": 2238775296 }, { "epoch": 0.36, "learning_rate": 6.495747071096132e-05, "loss": 2.4871, "theoretical_loss": 3.398620629671707, "tokens_seen": 2238906368 }, { "epoch": 0.36, "learning_rate": 6.494944631680309e-05, "loss": 2.4061, "theoretical_loss": 3.3986044067005303, "tokens_seen": 2239037440 }, { "epoch": 0.36, "learning_rate": 6.494142192264484e-05, "loss": 2.4802, "theoretical_loss": 3.398588184944903, "tokens_seen": 2239168512 }, { "epoch": 0.36, "learning_rate": 6.49333975284866e-05, "loss": 2.4283, "theoretical_loss": 3.3985719644046632, "tokens_seen": 2239299584 }, { "epoch": 0.36, "learning_rate": 6.492537313432836e-05, "loss": 2.5663, "theoretical_loss": 3.3985557450796486, "tokens_seen": 2239430656 }, { "epoch": 0.36, "learning_rate": 6.491734874017013e-05, "loss": 2.4968, "theoretical_loss": 3.3985395269696967, "tokens_seen": 2239561728 }, { "epoch": 0.36, "learning_rate": 6.490932434601188e-05, "loss": 2.6019, "theoretical_loss": 3.398523310074646, "tokens_seen": 2239692800 }, { "epoch": 0.36, "objective/train/docs_used": 1235870, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.631380558013916, "objective/train/theoretical_loss": 3.3985070943943336, "objective/train/tokens_used": 610282976, "theoretical_loss": 3.3985070943943336, "tokens_seen": 2239823872 }, { "epoch": 0.36, "learning_rate": 6.490129995185363e-05, "loss": 2.4624, "theoretical_loss": 3.3985070943943336, "tokens_seen": 2239823872 }, { "epoch": 0.36, "learning_rate": 6.48932755576954e-05, "loss": 2.6135, "theoretical_loss": 3.398490879928598, "tokens_seen": 2239954944 }, { "epoch": 0.36, "learning_rate": 6.488525116353715e-05, "loss": 2.3966, "theoretical_loss": 3.398474666677277, "tokens_seen": 2240086016 }, { "epoch": 0.36, "learning_rate": 6.487722676937892e-05, "loss": 2.5696, "theoretical_loss": 3.398458454640209, "tokens_seen": 2240217088 }, { "epoch": 0.36, "learning_rate": 6.486920237522067e-05, "loss": 2.5395, "theoretical_loss": 3.3984422438172315, "tokens_seen": 2240348160 }, { "epoch": 0.36, "learning_rate": 6.486117798106244e-05, "loss": 2.5404, "theoretical_loss": 3.3984260342081822, "tokens_seen": 2240479232 }, { "epoch": 0.36, "learning_rate": 6.485315358690419e-05, "loss": 2.5419, "theoretical_loss": 3.3984098258129003, "tokens_seen": 2240610304 }, { "epoch": 0.36, "learning_rate": 6.484512919274594e-05, "loss": 2.4847, "theoretical_loss": 3.398393618631223, "tokens_seen": 2240741376 }, { "epoch": 0.36, "learning_rate": 6.483710479858771e-05, "loss": 2.6941, "theoretical_loss": 3.398377412662989, "tokens_seen": 2240872448 }, { "epoch": 0.36, "learning_rate": 6.482908040442946e-05, "loss": 2.4717, "theoretical_loss": 3.398361207908036, "tokens_seen": 2241003520 }, { "epoch": 0.36, "learning_rate": 6.482105601027123e-05, "loss": 2.543, "theoretical_loss": 3.3983450043662025, "tokens_seen": 2241134592 }, { "epoch": 0.36, "learning_rate": 6.481303161611298e-05, "loss": 2.4651, "theoretical_loss": 3.398328802037327, "tokens_seen": 2241265664 }, { "epoch": 0.36, "learning_rate": 6.480500722195475e-05, "loss": 2.4543, "theoretical_loss": 3.398312600921247, "tokens_seen": 2241396736 }, { "epoch": 0.36, "objective/train/docs_used": 1236973, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.588179588317871, "objective/train/theoretical_loss": 3.3983045008179555, "objective/train/tokens_used": 611921376, "theoretical_loss": 3.3983045008179555, "tokens_seen": 2241462272 }, { "epoch": 0.36, "learning_rate": 6.47969828277965e-05, "loss": 2.3791, "theoretical_loss": 3.3982964010178014, "tokens_seen": 2241527808 }, { "epoch": 0.36, "learning_rate": 6.478895843363826e-05, "loss": 2.6522, "theoretical_loss": 3.398280202326829, "tokens_seen": 2241658880 }, { "epoch": 0.36, "learning_rate": 6.478093403948002e-05, "loss": 2.4594, "theoretical_loss": 3.398264004848167, "tokens_seen": 2241789952 }, { "epoch": 0.36, "learning_rate": 6.477290964532178e-05, "loss": 2.5694, "theoretical_loss": 3.3982478085816545, "tokens_seen": 2241921024 }, { "epoch": 0.36, "learning_rate": 6.476488525116354e-05, "loss": 2.6404, "theoretical_loss": 3.3982316135271295, "tokens_seen": 2242052096 }, { "epoch": 0.36, "learning_rate": 6.47568608570053e-05, "loss": 2.548, "theoretical_loss": 3.3982154196844316, "tokens_seen": 2242183168 } ], "max_steps": 12588, "num_train_epochs": 9223372036854775807, "total_flos": 3.02212454547456e+17, "trial_name": null, "trial_params": null }