{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.7583963425356075, "eval_steps": 500, "global_step": 40000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 1.3358560800552368, "learning_rate": 2.9999642387366694e-06, "loss": 1.8182, "step": 100 }, { "epoch": 0.01, "grad_norm": 1.0696768760681152, "learning_rate": 2.999856956651834e-06, "loss": 1.5829, "step": 200 }, { "epoch": 0.01, "grad_norm": 1.1315983533859253, "learning_rate": 2.9996781588608853e-06, "loss": 1.5088, "step": 300 }, { "epoch": 0.02, "grad_norm": 1.2688828706741333, "learning_rate": 2.9994278538892026e-06, "loss": 1.4755, "step": 400 }, { "epoch": 0.02, "grad_norm": 1.3950492143630981, "learning_rate": 2.999106053671749e-06, "loss": 1.44, "step": 500 }, { "epoch": 0.02, "eval_loss": 1.4453312158584595, "eval_runtime": 1866.1237, "eval_samples_per_second": 5.137, "eval_steps_per_second": 0.643, "step": 500 }, { "epoch": 0.03, "grad_norm": 1.4245915412902832, "learning_rate": 2.9987127735525e-06, "loss": 1.428, "step": 600 }, { "epoch": 0.03, "grad_norm": 1.3934013843536377, "learning_rate": 2.9982480322837146e-06, "loss": 1.4205, "step": 700 }, { "epoch": 0.04, "grad_norm": 1.345449447631836, "learning_rate": 2.99771185202504e-06, "loss": 1.4117, "step": 800 }, { "epoch": 0.04, "grad_norm": 1.4815948009490967, "learning_rate": 2.9971042583424534e-06, "loss": 1.4029, "step": 900 }, { "epoch": 0.04, "grad_norm": 1.416758418083191, "learning_rate": 2.996425280207045e-06, "loss": 1.4103, "step": 1000 }, { "epoch": 0.04, "eval_loss": 1.4033704996109009, "eval_runtime": 1866.432, "eval_samples_per_second": 5.136, "eval_steps_per_second": 0.642, "step": 1000 }, { "epoch": 0.05, "grad_norm": 1.5439317226409912, "learning_rate": 2.995674949993636e-06, "loss": 1.3992, "step": 1100 }, { "epoch": 0.05, "grad_norm": 1.6817480325698853, "learning_rate": 2.9948533034792356e-06, "loss": 1.3857, "step": 1200 }, { "epoch": 0.06, "grad_norm": 1.5548583269119263, "learning_rate": 2.993960379841333e-06, "loss": 1.3889, "step": 1300 }, { "epoch": 0.06, "grad_norm": 1.800005316734314, "learning_rate": 2.992996221656031e-06, "loss": 1.3798, "step": 1400 }, { "epoch": 0.07, "grad_norm": 1.5900113582611084, "learning_rate": 2.9919608748960165e-06, "loss": 1.3915, "step": 1500 }, { "epoch": 0.07, "eval_loss": 1.3856122493743896, "eval_runtime": 1867.7341, "eval_samples_per_second": 5.132, "eval_steps_per_second": 0.642, "step": 1500 }, { "epoch": 0.07, "grad_norm": 1.593215823173523, "learning_rate": 2.990854388928367e-06, "loss": 1.3707, "step": 1600 }, { "epoch": 0.07, "grad_norm": 1.6342238187789917, "learning_rate": 2.9896768165121967e-06, "loss": 1.3909, "step": 1700 }, { "epoch": 0.08, "grad_norm": 1.7908762693405151, "learning_rate": 2.9884282137961426e-06, "loss": 1.3758, "step": 1800 }, { "epoch": 0.08, "grad_norm": 1.5925847291946411, "learning_rate": 2.987108640315685e-06, "loss": 1.3816, "step": 1900 }, { "epoch": 0.09, "grad_norm": 1.8841816186904907, "learning_rate": 2.9857181589903103e-06, "loss": 1.383, "step": 2000 }, { "epoch": 0.09, "eval_loss": 1.3752671480178833, "eval_runtime": 1867.3688, "eval_samples_per_second": 5.133, "eval_steps_per_second": 0.642, "step": 2000 }, { "epoch": 0.09, "grad_norm": 1.688745379447937, "learning_rate": 2.984256836120511e-06, "loss": 1.3796, "step": 2100 }, { "epoch": 0.1, "grad_norm": 1.676340103149414, "learning_rate": 2.9827247413846217e-06, "loss": 1.3742, "step": 2200 }, { "epoch": 0.1, "grad_norm": 2.0635907649993896, "learning_rate": 2.981121947835501e-06, "loss": 1.3712, "step": 2300 }, { "epoch": 0.11, "grad_norm": 1.799085021018982, "learning_rate": 2.979448531897045e-06, "loss": 1.3694, "step": 2400 }, { "epoch": 0.11, "grad_norm": 1.8315014839172363, "learning_rate": 2.9777045733605437e-06, "loss": 1.3558, "step": 2500 }, { "epoch": 0.11, "eval_loss": 1.3671655654907227, "eval_runtime": 1867.272, "eval_samples_per_second": 5.134, "eval_steps_per_second": 0.642, "step": 2500 }, { "epoch": 0.11, "grad_norm": 1.8167532682418823, "learning_rate": 2.9758901553808787e-06, "loss": 1.3579, "step": 2600 }, { "epoch": 0.12, "grad_norm": 1.8836266994476318, "learning_rate": 2.9740053644725552e-06, "loss": 1.3714, "step": 2700 }, { "epoch": 0.12, "grad_norm": 1.823400855064392, "learning_rate": 2.972050290505579e-06, "loss": 1.3609, "step": 2800 }, { "epoch": 0.13, "grad_norm": 1.8716390132904053, "learning_rate": 2.970025026701169e-06, "loss": 1.3685, "step": 2900 }, { "epoch": 0.13, "grad_norm": 1.7368168830871582, "learning_rate": 2.967929669627316e-06, "loss": 1.3665, "step": 3000 }, { "epoch": 0.13, "eval_loss": 1.3610810041427612, "eval_runtime": 1867.7829, "eval_samples_per_second": 5.132, "eval_steps_per_second": 0.642, "step": 3000 }, { "epoch": 0.14, "grad_norm": 1.8250432014465332, "learning_rate": 2.9657643191941737e-06, "loss": 1.3549, "step": 3100 }, { "epoch": 0.14, "grad_norm": 1.9420477151870728, "learning_rate": 2.9635290786492985e-06, "loss": 1.3531, "step": 3200 }, { "epoch": 0.15, "grad_norm": 2.0373518466949463, "learning_rate": 2.9612240545727255e-06, "loss": 1.3619, "step": 3300 }, { "epoch": 0.15, "grad_norm": 1.9139854907989502, "learning_rate": 2.9588493568718843e-06, "loss": 1.3665, "step": 3400 }, { "epoch": 0.15, "grad_norm": 1.9374817609786987, "learning_rate": 2.9564050987763614e-06, "loss": 1.3619, "step": 3500 }, { "epoch": 0.15, "eval_loss": 1.3560757637023926, "eval_runtime": 1867.3981, "eval_samples_per_second": 5.133, "eval_steps_per_second": 0.642, "step": 3500 }, { "epoch": 0.16, "grad_norm": 1.8168325424194336, "learning_rate": 2.9538913968325007e-06, "loss": 1.3527, "step": 3600 }, { "epoch": 0.16, "grad_norm": 1.948146939277649, "learning_rate": 2.951308370897845e-06, "loss": 1.3496, "step": 3700 }, { "epoch": 0.17, "grad_norm": 2.068612575531006, "learning_rate": 2.948656144135421e-06, "loss": 1.358, "step": 3800 }, { "epoch": 0.17, "grad_norm": 1.9631004333496094, "learning_rate": 2.945934843007869e-06, "loss": 1.3504, "step": 3900 }, { "epoch": 0.18, "grad_norm": 2.0034778118133545, "learning_rate": 2.9431445972714102e-06, "loss": 1.3582, "step": 4000 }, { "epoch": 0.18, "eval_loss": 1.3521347045898438, "eval_runtime": 1867.5478, "eval_samples_per_second": 5.133, "eval_steps_per_second": 0.642, "step": 4000 }, { "epoch": 0.18, "grad_norm": 1.872603178024292, "learning_rate": 2.940285539969662e-06, "loss": 1.3483, "step": 4100 }, { "epoch": 0.18, "grad_norm": 1.8477139472961426, "learning_rate": 2.9373578074272917e-06, "loss": 1.3623, "step": 4200 }, { "epoch": 0.19, "grad_norm": 2.0872058868408203, "learning_rate": 2.9343615392435187e-06, "loss": 1.3483, "step": 4300 }, { "epoch": 0.19, "grad_norm": 1.931931495666504, "learning_rate": 2.931296878285457e-06, "loss": 1.3499, "step": 4400 }, { "epoch": 0.2, "grad_norm": 1.9555952548980713, "learning_rate": 2.928163970681304e-06, "loss": 1.3525, "step": 4500 }, { "epoch": 0.2, "eval_loss": 1.348177433013916, "eval_runtime": 1865.9442, "eval_samples_per_second": 5.137, "eval_steps_per_second": 0.643, "step": 4500 }, { "epoch": 0.2, "grad_norm": 1.9107462167739868, "learning_rate": 2.9249629658133703e-06, "loss": 1.3446, "step": 4600 }, { "epoch": 0.21, "grad_norm": 2.1549768447875977, "learning_rate": 2.9216940163109613e-06, "loss": 1.3485, "step": 4700 }, { "epoch": 0.21, "grad_norm": 1.9788259267807007, "learning_rate": 2.9183572780430938e-06, "loss": 1.3386, "step": 4800 }, { "epoch": 0.22, "grad_norm": 2.008772850036621, "learning_rate": 2.9149529101110707e-06, "loss": 1.3449, "step": 4900 }, { "epoch": 0.22, "grad_norm": 1.932020664215088, "learning_rate": 2.9114810748408887e-06, "loss": 1.3471, "step": 5000 }, { "epoch": 0.22, "eval_loss": 1.3448920249938965, "eval_runtime": 1866.903, "eval_samples_per_second": 5.135, "eval_steps_per_second": 0.642, "step": 5000 }, { "epoch": 0.22, "grad_norm": 1.780934453010559, "learning_rate": 2.9079419377755016e-06, "loss": 1.3408, "step": 5100 }, { "epoch": 0.23, "grad_norm": 2.0991992950439453, "learning_rate": 2.904335667666926e-06, "loss": 1.3377, "step": 5200 }, { "epoch": 0.23, "grad_norm": 1.7961390018463135, "learning_rate": 2.9006624364681955e-06, "loss": 1.3446, "step": 5300 }, { "epoch": 0.24, "grad_norm": 1.951217770576477, "learning_rate": 2.896922419325161e-06, "loss": 1.3434, "step": 5400 }, { "epoch": 0.24, "grad_norm": 1.9937087297439575, "learning_rate": 2.89311579456814e-06, "loss": 1.3383, "step": 5500 }, { "epoch": 0.24, "eval_loss": 1.3420648574829102, "eval_runtime": 1864.5499, "eval_samples_per_second": 5.141, "eval_steps_per_second": 0.643, "step": 5500 }, { "epoch": 0.25, "grad_norm": 1.9863967895507812, "learning_rate": 2.889242743703412e-06, "loss": 1.3553, "step": 5600 }, { "epoch": 0.25, "grad_norm": 2.026629686355591, "learning_rate": 2.8853034514045667e-06, "loss": 1.3384, "step": 5700 }, { "epoch": 0.25, "grad_norm": 2.1524720191955566, "learning_rate": 2.8812981055036967e-06, "loss": 1.339, "step": 5800 }, { "epoch": 0.26, "grad_norm": 2.011669635772705, "learning_rate": 2.8772268969824412e-06, "loss": 1.3449, "step": 5900 }, { "epoch": 0.26, "grad_norm": 1.96876060962677, "learning_rate": 2.87309001996288e-06, "loss": 1.3352, "step": 6000 }, { "epoch": 0.26, "eval_loss": 1.3392791748046875, "eval_runtime": 1863.5996, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 6000 }, { "epoch": 0.27, "grad_norm": 2.030050754547119, "learning_rate": 2.868887671698278e-06, "loss": 1.3473, "step": 6100 }, { "epoch": 0.27, "grad_norm": 1.9417366981506348, "learning_rate": 2.8646200525636784e-06, "loss": 1.3467, "step": 6200 }, { "epoch": 0.28, "grad_norm": 1.8963119983673096, "learning_rate": 2.860287366046351e-06, "loss": 1.3298, "step": 6300 }, { "epoch": 0.28, "grad_norm": 2.0446488857269287, "learning_rate": 2.855889818736086e-06, "loss": 1.3364, "step": 6400 }, { "epoch": 0.29, "grad_norm": 1.9520761966705322, "learning_rate": 2.8514276203153476e-06, "loss": 1.3328, "step": 6500 }, { "epoch": 0.29, "eval_loss": 1.3369933366775513, "eval_runtime": 1864.5392, "eval_samples_per_second": 5.141, "eval_steps_per_second": 0.643, "step": 6500 }, { "epoch": 0.29, "grad_norm": 1.9833135604858398, "learning_rate": 2.846900983549272e-06, "loss": 1.3461, "step": 6600 }, { "epoch": 0.29, "grad_norm": 1.8837144374847412, "learning_rate": 2.8423101242755264e-06, "loss": 1.3354, "step": 6700 }, { "epoch": 0.3, "grad_norm": 1.9261667728424072, "learning_rate": 2.837655261394013e-06, "loss": 1.3349, "step": 6800 }, { "epoch": 0.3, "grad_norm": 1.8928134441375732, "learning_rate": 2.832936616856434e-06, "loss": 1.3247, "step": 6900 }, { "epoch": 0.31, "grad_norm": 2.0842361450195312, "learning_rate": 2.828154415655711e-06, "loss": 1.3396, "step": 7000 }, { "epoch": 0.31, "eval_loss": 1.3346595764160156, "eval_runtime": 1864.6191, "eval_samples_per_second": 5.141, "eval_steps_per_second": 0.643, "step": 7000 }, { "epoch": 0.31, "grad_norm": 2.0268609523773193, "learning_rate": 2.823308885815251e-06, "loss": 1.3577, "step": 7100 }, { "epoch": 0.32, "grad_norm": 1.9914246797561646, "learning_rate": 2.8184002583780802e-06, "loss": 1.3273, "step": 7200 }, { "epoch": 0.32, "grad_norm": 2.065075159072876, "learning_rate": 2.813428767395822e-06, "loss": 1.3449, "step": 7300 }, { "epoch": 0.33, "grad_norm": 2.0914437770843506, "learning_rate": 2.8083946499175413e-06, "loss": 1.3415, "step": 7400 }, { "epoch": 0.33, "grad_norm": 1.9424386024475098, "learning_rate": 2.803298145978439e-06, "loss": 1.3316, "step": 7500 }, { "epoch": 0.33, "eval_loss": 1.3324897289276123, "eval_runtime": 1864.3018, "eval_samples_per_second": 5.142, "eval_steps_per_second": 0.643, "step": 7500 }, { "epoch": 0.33, "grad_norm": 2.083033561706543, "learning_rate": 2.7981394985884074e-06, "loss": 1.3213, "step": 7600 }, { "epoch": 0.34, "grad_norm": 1.8912067413330078, "learning_rate": 2.792918953720444e-06, "loss": 1.3284, "step": 7700 }, { "epoch": 0.34, "grad_norm": 1.8995610475540161, "learning_rate": 2.7876367602989206e-06, "loss": 1.3256, "step": 7800 }, { "epoch": 0.35, "grad_norm": 1.9921075105667114, "learning_rate": 2.7822931701877186e-06, "loss": 1.3339, "step": 7900 }, { "epoch": 0.35, "grad_norm": 1.9593900442123413, "learning_rate": 2.7768884381782147e-06, "loss": 1.3357, "step": 8000 }, { "epoch": 0.35, "eval_loss": 1.3304702043533325, "eval_runtime": 1863.8735, "eval_samples_per_second": 5.143, "eval_steps_per_second": 0.643, "step": 8000 }, { "epoch": 0.36, "grad_norm": 2.0294156074523926, "learning_rate": 2.7714228219771354e-06, "loss": 1.3248, "step": 8100 }, { "epoch": 0.36, "grad_norm": 1.9358181953430176, "learning_rate": 2.765896582194267e-06, "loss": 1.3321, "step": 8200 }, { "epoch": 0.36, "grad_norm": 1.9804461002349854, "learning_rate": 2.7603099823300316e-06, "loss": 1.328, "step": 8300 }, { "epoch": 0.37, "grad_norm": 1.9724797010421753, "learning_rate": 2.754663288762921e-06, "loss": 1.3313, "step": 8400 }, { "epoch": 0.37, "grad_norm": 2.1071040630340576, "learning_rate": 2.748956770736796e-06, "loss": 1.3164, "step": 8500 }, { "epoch": 0.37, "eval_loss": 1.3289612531661987, "eval_runtime": 1864.6483, "eval_samples_per_second": 5.141, "eval_steps_per_second": 0.643, "step": 8500 }, { "epoch": 0.38, "grad_norm": 1.8682750463485718, "learning_rate": 2.743190700348048e-06, "loss": 1.3318, "step": 8600 }, { "epoch": 0.38, "grad_norm": 2.1705667972564697, "learning_rate": 2.7373653525326257e-06, "loss": 1.3203, "step": 8700 }, { "epoch": 0.39, "grad_norm": 2.2123374938964844, "learning_rate": 2.7314810050529265e-06, "loss": 1.3279, "step": 8800 }, { "epoch": 0.39, "grad_norm": 2.0958218574523926, "learning_rate": 2.7255379384845483e-06, "loss": 1.332, "step": 8900 }, { "epoch": 0.4, "grad_norm": 2.0944671630859375, "learning_rate": 2.7195364362029173e-06, "loss": 1.3231, "step": 9000 }, { "epoch": 0.4, "eval_loss": 1.3273080587387085, "eval_runtime": 1864.6036, "eval_samples_per_second": 5.141, "eval_steps_per_second": 0.643, "step": 9000 }, { "epoch": 0.4, "grad_norm": 2.045405149459839, "learning_rate": 2.713476784369771e-06, "loss": 1.3179, "step": 9100 }, { "epoch": 0.4, "grad_norm": 1.985823392868042, "learning_rate": 2.7073592719195155e-06, "loss": 1.3318, "step": 9200 }, { "epoch": 0.41, "grad_norm": 2.064948081970215, "learning_rate": 2.7011841905454495e-06, "loss": 1.3181, "step": 9300 }, { "epoch": 0.41, "grad_norm": 2.136582612991333, "learning_rate": 2.694951834685854e-06, "loss": 1.3181, "step": 9400 }, { "epoch": 0.42, "grad_norm": 2.394834041595459, "learning_rate": 2.6886625015099553e-06, "loss": 1.3287, "step": 9500 }, { "epoch": 0.42, "eval_loss": 1.3257533311843872, "eval_runtime": 1871.6529, "eval_samples_per_second": 5.122, "eval_steps_per_second": 0.641, "step": 9500 }, { "epoch": 0.42, "grad_norm": 2.0295205116271973, "learning_rate": 2.682316490903753e-06, "loss": 1.3316, "step": 9600 }, { "epoch": 0.43, "grad_norm": 1.883170247077942, "learning_rate": 2.6759141054557214e-06, "loss": 1.345, "step": 9700 }, { "epoch": 0.43, "grad_norm": 2.172785997390747, "learning_rate": 2.6694556504423836e-06, "loss": 1.3197, "step": 9800 }, { "epoch": 0.44, "grad_norm": 2.065781354904175, "learning_rate": 2.662941433813754e-06, "loss": 1.3165, "step": 9900 }, { "epoch": 0.44, "grad_norm": 1.825239896774292, "learning_rate": 2.6563717661786536e-06, "loss": 1.3308, "step": 10000 }, { "epoch": 0.44, "eval_loss": 1.324156641960144, "eval_runtime": 1865.6117, "eval_samples_per_second": 5.138, "eval_steps_per_second": 0.643, "step": 10000 }, { "epoch": 0.44, "grad_norm": 2.128471612930298, "learning_rate": 2.649746960789902e-06, "loss": 1.3176, "step": 10100 }, { "epoch": 0.45, "grad_norm": 1.994718074798584, "learning_rate": 2.6430673335293788e-06, "loss": 1.3245, "step": 10200 }, { "epoch": 0.45, "grad_norm": 2.0164434909820557, "learning_rate": 2.6363332028929633e-06, "loss": 1.3179, "step": 10300 }, { "epoch": 0.46, "grad_norm": 1.879969835281372, "learning_rate": 2.629544889975348e-06, "loss": 1.3346, "step": 10400 }, { "epoch": 0.46, "grad_norm": 1.9689232110977173, "learning_rate": 2.6227027184547265e-06, "loss": 1.311, "step": 10500 }, { "epoch": 0.46, "eval_loss": 1.3230102062225342, "eval_runtime": 1865.9111, "eval_samples_per_second": 5.137, "eval_steps_per_second": 0.643, "step": 10500 }, { "epoch": 0.47, "grad_norm": 2.0227608680725098, "learning_rate": 2.615807014577363e-06, "loss": 1.3226, "step": 10600 }, { "epoch": 0.47, "grad_norm": 1.9242812395095825, "learning_rate": 2.608858107142033e-06, "loss": 1.3254, "step": 10700 }, { "epoch": 0.47, "grad_norm": 1.9587324857711792, "learning_rate": 2.601856327484348e-06, "loss": 1.3219, "step": 10800 }, { "epoch": 0.48, "grad_norm": 2.0176897048950195, "learning_rate": 2.594802009460957e-06, "loss": 1.3133, "step": 10900 }, { "epoch": 0.48, "grad_norm": 1.871654987335205, "learning_rate": 2.5876954894336257e-06, "loss": 1.3242, "step": 11000 }, { "epoch": 0.48, "eval_loss": 1.321611762046814, "eval_runtime": 1864.3465, "eval_samples_per_second": 5.142, "eval_steps_per_second": 0.643, "step": 11000 }, { "epoch": 0.49, "grad_norm": 2.1732594966888428, "learning_rate": 2.580537106253199e-06, "loss": 1.3123, "step": 11100 }, { "epoch": 0.49, "grad_norm": 1.9368730783462524, "learning_rate": 2.5733272012434456e-06, "loss": 1.323, "step": 11200 }, { "epoch": 0.5, "grad_norm": 2.197732448577881, "learning_rate": 2.56606611818478e-06, "loss": 1.316, "step": 11300 }, { "epoch": 0.5, "grad_norm": 2.0765159130096436, "learning_rate": 2.5587542032978735e-06, "loss": 1.326, "step": 11400 }, { "epoch": 0.51, "grad_norm": 1.9759124517440796, "learning_rate": 2.551391805227145e-06, "loss": 1.319, "step": 11500 }, { "epoch": 0.51, "eval_loss": 1.3201991319656372, "eval_runtime": 1867.1093, "eval_samples_per_second": 5.134, "eval_steps_per_second": 0.642, "step": 11500 }, { "epoch": 0.51, "grad_norm": 2.119062900543213, "learning_rate": 2.5439792750241362e-06, "loss": 1.319, "step": 11600 }, { "epoch": 0.51, "grad_norm": 1.9623346328735352, "learning_rate": 2.5365169661307723e-06, "loss": 1.3188, "step": 11700 }, { "epoch": 0.52, "grad_norm": 2.069474935531616, "learning_rate": 2.529005234362512e-06, "loss": 1.3236, "step": 11800 }, { "epoch": 0.52, "grad_norm": 1.8530821800231934, "learning_rate": 2.521444437891378e-06, "loss": 1.3172, "step": 11900 }, { "epoch": 0.53, "grad_norm": 2.1172099113464355, "learning_rate": 2.513834937228883e-06, "loss": 1.3183, "step": 12000 }, { "epoch": 0.53, "eval_loss": 1.31929349899292, "eval_runtime": 1867.4942, "eval_samples_per_second": 5.133, "eval_steps_per_second": 0.642, "step": 12000 }, { "epoch": 0.53, "grad_norm": 1.981194019317627, "learning_rate": 2.5061770952088354e-06, "loss": 1.3278, "step": 12100 }, { "epoch": 0.54, "grad_norm": 1.9915848970413208, "learning_rate": 2.4984712769700423e-06, "loss": 1.3143, "step": 12200 }, { "epoch": 0.54, "grad_norm": 2.114997625350952, "learning_rate": 2.490717849938897e-06, "loss": 1.3207, "step": 12300 }, { "epoch": 0.55, "grad_norm": 1.9648516178131104, "learning_rate": 2.4829171838118613e-06, "loss": 1.3129, "step": 12400 }, { "epoch": 0.55, "grad_norm": 1.9119199514389038, "learning_rate": 2.4750696505378355e-06, "loss": 1.3211, "step": 12500 }, { "epoch": 0.55, "eval_loss": 1.3180123567581177, "eval_runtime": 1866.9983, "eval_samples_per_second": 5.134, "eval_steps_per_second": 0.642, "step": 12500 }, { "epoch": 0.55, "grad_norm": 2.158143997192383, "learning_rate": 2.4671756243004243e-06, "loss": 1.3124, "step": 12600 }, { "epoch": 0.56, "grad_norm": 2.059196949005127, "learning_rate": 2.4592354815000963e-06, "loss": 1.3229, "step": 12700 }, { "epoch": 0.56, "grad_norm": 1.9812512397766113, "learning_rate": 2.451249600736235e-06, "loss": 1.3159, "step": 12800 }, { "epoch": 0.57, "grad_norm": 1.9367738962173462, "learning_rate": 2.4432183627890867e-06, "loss": 1.3101, "step": 12900 }, { "epoch": 0.57, "grad_norm": 2.1279842853546143, "learning_rate": 2.4351421506016047e-06, "loss": 1.3158, "step": 13000 }, { "epoch": 0.57, "eval_loss": 1.3169833421707153, "eval_runtime": 1870.0395, "eval_samples_per_second": 5.126, "eval_steps_per_second": 0.641, "step": 13000 }, { "epoch": 0.58, "grad_norm": 2.213517189025879, "learning_rate": 2.4270213492611903e-06, "loss": 1.3113, "step": 13100 }, { "epoch": 0.58, "grad_norm": 2.060974597930908, "learning_rate": 2.41885634598133e-06, "loss": 1.3248, "step": 13200 }, { "epoch": 0.58, "grad_norm": 2.004950523376465, "learning_rate": 2.4106475300831343e-06, "loss": 1.321, "step": 13300 }, { "epoch": 0.59, "grad_norm": 2.081263303756714, "learning_rate": 2.402395292976772e-06, "loss": 1.3058, "step": 13400 }, { "epoch": 0.59, "grad_norm": 2.141289472579956, "learning_rate": 2.394100028142809e-06, "loss": 1.3054, "step": 13500 }, { "epoch": 0.59, "eval_loss": 1.3159672021865845, "eval_runtime": 1868.4185, "eval_samples_per_second": 5.131, "eval_steps_per_second": 0.642, "step": 13500 }, { "epoch": 0.6, "grad_norm": 2.041181802749634, "learning_rate": 2.3857621311134457e-06, "loss": 1.324, "step": 13600 }, { "epoch": 0.6, "grad_norm": 1.9553183317184448, "learning_rate": 2.3773819994536564e-06, "loss": 1.3115, "step": 13700 }, { "epoch": 0.61, "grad_norm": 2.0408129692077637, "learning_rate": 2.368960032742235e-06, "loss": 1.3007, "step": 13800 }, { "epoch": 0.61, "grad_norm": 2.0877633094787598, "learning_rate": 2.36049663255274e-06, "loss": 1.3174, "step": 13900 }, { "epoch": 0.62, "grad_norm": 2.013047218322754, "learning_rate": 2.35199220243435e-06, "loss": 1.3249, "step": 14000 }, { "epoch": 0.62, "eval_loss": 1.3149954080581665, "eval_runtime": 1864.575, "eval_samples_per_second": 5.141, "eval_steps_per_second": 0.643, "step": 14000 }, { "epoch": 0.62, "grad_norm": 1.9625511169433594, "learning_rate": 2.3434471478926168e-06, "loss": 1.3053, "step": 14100 }, { "epoch": 0.62, "grad_norm": 1.9895579814910889, "learning_rate": 2.334861876370135e-06, "loss": 1.3097, "step": 14200 }, { "epoch": 0.63, "grad_norm": 2.0086371898651123, "learning_rate": 2.3262367972271126e-06, "loss": 1.3075, "step": 14300 }, { "epoch": 0.63, "grad_norm": 2.0822415351867676, "learning_rate": 2.317572321721851e-06, "loss": 1.3122, "step": 14400 }, { "epoch": 0.64, "grad_norm": 2.1191747188568115, "learning_rate": 2.3088688629911378e-06, "loss": 1.3029, "step": 14500 }, { "epoch": 0.64, "eval_loss": 1.313984990119934, "eval_runtime": 1868.3094, "eval_samples_per_second": 5.131, "eval_steps_per_second": 0.642, "step": 14500 }, { "epoch": 0.64, "grad_norm": 2.0153069496154785, "learning_rate": 2.3001268360305446e-06, "loss": 1.31, "step": 14600 }, { "epoch": 0.65, "grad_norm": 2.031879425048828, "learning_rate": 2.2913466576746433e-06, "loss": 1.3113, "step": 14700 }, { "epoch": 0.65, "grad_norm": 1.9688141345977783, "learning_rate": 2.2825287465771276e-06, "loss": 1.3114, "step": 14800 }, { "epoch": 0.66, "grad_norm": 2.0744152069091797, "learning_rate": 2.2736735231908515e-06, "loss": 1.3068, "step": 14900 }, { "epoch": 0.66, "grad_norm": 2.004817485809326, "learning_rate": 2.2647814097477816e-06, "loss": 1.3148, "step": 15000 }, { "epoch": 0.66, "eval_loss": 1.3130929470062256, "eval_runtime": 1864.9246, "eval_samples_per_second": 5.14, "eval_steps_per_second": 0.643, "step": 15000 }, { "epoch": 0.66, "grad_norm": 2.018002510070801, "learning_rate": 2.2558528302388663e-06, "loss": 1.318, "step": 15100 }, { "epoch": 0.67, "grad_norm": 2.0592522621154785, "learning_rate": 2.2468882103938155e-06, "loss": 1.3139, "step": 15200 }, { "epoch": 0.67, "grad_norm": 1.9203455448150635, "learning_rate": 2.237887977660804e-06, "loss": 1.3163, "step": 15300 }, { "epoch": 0.68, "grad_norm": 2.1021227836608887, "learning_rate": 2.2288525611860886e-06, "loss": 1.3182, "step": 15400 }, { "epoch": 0.68, "grad_norm": 2.050011396408081, "learning_rate": 2.219782391793547e-06, "loss": 1.3036, "step": 15500 }, { "epoch": 0.68, "eval_loss": 1.312229871749878, "eval_runtime": 1865.3014, "eval_samples_per_second": 5.139, "eval_steps_per_second": 0.643, "step": 15500 }, { "epoch": 0.69, "grad_norm": 2.0257480144500732, "learning_rate": 2.2106779019641336e-06, "loss": 1.2984, "step": 15600 }, { "epoch": 0.69, "grad_norm": 1.9984742403030396, "learning_rate": 2.2015395258152596e-06, "loss": 1.3206, "step": 15700 }, { "epoch": 0.69, "grad_norm": 2.094935417175293, "learning_rate": 2.1923676990800927e-06, "loss": 1.3157, "step": 15800 }, { "epoch": 0.7, "grad_norm": 2.0469040870666504, "learning_rate": 2.1831628590867812e-06, "loss": 1.3048, "step": 15900 }, { "epoch": 0.7, "grad_norm": 2.0622622966766357, "learning_rate": 2.1739254447376006e-06, "loss": 1.3138, "step": 16000 }, { "epoch": 0.7, "eval_loss": 1.3114036321640015, "eval_runtime": 1865.761, "eval_samples_per_second": 5.138, "eval_steps_per_second": 0.643, "step": 16000 }, { "epoch": 0.71, "grad_norm": 2.0791163444519043, "learning_rate": 2.1646558964880277e-06, "loss": 1.3108, "step": 16100 }, { "epoch": 0.71, "grad_norm": 2.027860403060913, "learning_rate": 2.1553546563257362e-06, "loss": 1.3116, "step": 16200 }, { "epoch": 0.72, "grad_norm": 2.0047738552093506, "learning_rate": 2.146022167749525e-06, "loss": 1.3048, "step": 16300 }, { "epoch": 0.72, "grad_norm": 2.213186025619507, "learning_rate": 2.136658875748169e-06, "loss": 1.3071, "step": 16400 }, { "epoch": 0.73, "grad_norm": 2.0164272785186768, "learning_rate": 2.1272652267792036e-06, "loss": 1.314, "step": 16500 }, { "epoch": 0.73, "eval_loss": 1.310592532157898, "eval_runtime": 1865.0707, "eval_samples_per_second": 5.14, "eval_steps_per_second": 0.643, "step": 16500 }, { "epoch": 0.73, "grad_norm": 2.063606023788452, "learning_rate": 2.117841668747633e-06, "loss": 1.3171, "step": 16600 }, { "epoch": 0.73, "grad_norm": 2.066218137741089, "learning_rate": 2.1083886509845794e-06, "loss": 1.3095, "step": 16700 }, { "epoch": 0.74, "grad_norm": 2.00728702545166, "learning_rate": 2.098906624225852e-06, "loss": 1.3082, "step": 16800 }, { "epoch": 0.74, "grad_norm": 1.946075439453125, "learning_rate": 2.089396040590459e-06, "loss": 1.3106, "step": 16900 }, { "epoch": 0.75, "grad_norm": 2.2883059978485107, "learning_rate": 2.079857353559047e-06, "loss": 1.3045, "step": 17000 }, { "epoch": 0.75, "eval_loss": 1.3099678754806519, "eval_runtime": 1865.3447, "eval_samples_per_second": 5.139, "eval_steps_per_second": 0.643, "step": 17000 }, { "epoch": 0.75, "grad_norm": 1.9902862310409546, "learning_rate": 2.070291017952282e-06, "loss": 1.3188, "step": 17100 }, { "epoch": 0.76, "grad_norm": 2.0283963680267334, "learning_rate": 2.0606974899091596e-06, "loss": 1.3115, "step": 17200 }, { "epoch": 0.76, "grad_norm": 2.249575138092041, "learning_rate": 2.0510772268652563e-06, "loss": 1.3095, "step": 17300 }, { "epoch": 0.76, "grad_norm": 2.022587537765503, "learning_rate": 2.0414306875309195e-06, "loss": 1.2989, "step": 17400 }, { "epoch": 0.77, "grad_norm": 2.3016037940979004, "learning_rate": 2.0317583318693924e-06, "loss": 1.303, "step": 17500 }, { "epoch": 0.77, "eval_loss": 1.3091822862625122, "eval_runtime": 1863.7921, "eval_samples_per_second": 5.143, "eval_steps_per_second": 0.643, "step": 17500 }, { "epoch": 0.77, "grad_norm": 2.1690895557403564, "learning_rate": 2.022060621074887e-06, "loss": 1.3057, "step": 17600 }, { "epoch": 0.78, "grad_norm": 2.1167852878570557, "learning_rate": 2.012338017550587e-06, "loss": 1.297, "step": 17700 }, { "epoch": 0.78, "grad_norm": 2.035123109817505, "learning_rate": 2.0025909848866085e-06, "loss": 1.305, "step": 17800 }, { "epoch": 0.79, "grad_norm": 1.931913137435913, "learning_rate": 1.9928199878378854e-06, "loss": 1.2938, "step": 17900 }, { "epoch": 0.79, "grad_norm": 2.0481088161468506, "learning_rate": 1.983025492302016e-06, "loss": 1.2977, "step": 18000 }, { "epoch": 0.79, "eval_loss": 1.3084911108016968, "eval_runtime": 1866.3126, "eval_samples_per_second": 5.136, "eval_steps_per_second": 0.642, "step": 18000 }, { "epoch": 0.8, "grad_norm": 1.9030753374099731, "learning_rate": 1.973207965297045e-06, "loss": 1.3115, "step": 18100 }, { "epoch": 0.8, "grad_norm": 1.9687721729278564, "learning_rate": 1.963367874939198e-06, "loss": 1.3026, "step": 18200 }, { "epoch": 0.8, "grad_norm": 2.2594008445739746, "learning_rate": 1.9535056904205588e-06, "loss": 1.3134, "step": 18300 }, { "epoch": 0.81, "grad_norm": 2.038999080657959, "learning_rate": 1.943621881986696e-06, "loss": 1.3039, "step": 18400 }, { "epoch": 0.81, "grad_norm": 1.9501850605010986, "learning_rate": 1.933716920914245e-06, "loss": 1.3096, "step": 18500 }, { "epoch": 0.81, "eval_loss": 1.3078255653381348, "eval_runtime": 1865.4912, "eval_samples_per_second": 5.139, "eval_steps_per_second": 0.643, "step": 18500 }, { "epoch": 0.82, "grad_norm": 2.107854127883911, "learning_rate": 1.923791279488435e-06, "loss": 1.307, "step": 18600 }, { "epoch": 0.82, "grad_norm": 2.067185640335083, "learning_rate": 1.913845430980567e-06, "loss": 1.3225, "step": 18700 }, { "epoch": 0.83, "grad_norm": 1.9632668495178223, "learning_rate": 1.9038798496254518e-06, "loss": 1.3095, "step": 18800 }, { "epoch": 0.83, "grad_norm": 2.140648603439331, "learning_rate": 1.8938950105987948e-06, "loss": 1.2998, "step": 18900 }, { "epoch": 0.84, "grad_norm": 1.9392528533935547, "learning_rate": 1.8838913899945394e-06, "loss": 1.3073, "step": 19000 }, { "epoch": 0.84, "eval_loss": 1.3071683645248413, "eval_runtime": 1864.0962, "eval_samples_per_second": 5.142, "eval_steps_per_second": 0.643, "step": 19000 }, { "epoch": 0.84, "grad_norm": 2.2147011756896973, "learning_rate": 1.8738694648021666e-06, "loss": 1.3017, "step": 19100 }, { "epoch": 0.84, "grad_norm": 2.112633228302002, "learning_rate": 1.863829712883951e-06, "loss": 1.312, "step": 19200 }, { "epoch": 0.85, "grad_norm": 2.0416507720947266, "learning_rate": 1.8537726129521755e-06, "loss": 1.3036, "step": 19300 }, { "epoch": 0.85, "grad_norm": 1.988303303718567, "learning_rate": 1.8436986445463049e-06, "loss": 1.3009, "step": 19400 }, { "epoch": 0.86, "grad_norm": 2.023023843765259, "learning_rate": 1.8336082880101228e-06, "loss": 1.3137, "step": 19500 }, { "epoch": 0.86, "eval_loss": 1.3065180778503418, "eval_runtime": 1866.3466, "eval_samples_per_second": 5.136, "eval_steps_per_second": 0.642, "step": 19500 }, { "epoch": 0.86, "grad_norm": 2.112020254135132, "learning_rate": 1.8235020244688246e-06, "loss": 1.3037, "step": 19600 }, { "epoch": 0.87, "grad_norm": 2.056504249572754, "learning_rate": 1.8133803358060798e-06, "loss": 1.3141, "step": 19700 }, { "epoch": 0.87, "grad_norm": 2.0338857173919678, "learning_rate": 1.8032437046410528e-06, "loss": 1.3012, "step": 19800 }, { "epoch": 0.87, "grad_norm": 2.071561098098755, "learning_rate": 1.7930926143053926e-06, "loss": 1.3063, "step": 19900 }, { "epoch": 0.88, "grad_norm": 2.080360174179077, "learning_rate": 1.7829275488201849e-06, "loss": 1.3083, "step": 20000 }, { "epoch": 0.88, "eval_loss": 1.3060280084609985, "eval_runtime": 1870.2241, "eval_samples_per_second": 5.126, "eval_steps_per_second": 0.641, "step": 20000 }, { "epoch": 0.88, "grad_norm": 2.0536909103393555, "learning_rate": 1.7727489928728747e-06, "loss": 1.3115, "step": 20100 }, { "epoch": 0.89, "grad_norm": 2.07513689994812, "learning_rate": 1.762557431794155e-06, "loss": 1.3028, "step": 20200 }, { "epoch": 0.89, "grad_norm": 2.2155916690826416, "learning_rate": 1.7523533515348249e-06, "loss": 1.3061, "step": 20300 }, { "epoch": 0.9, "grad_norm": 2.2243480682373047, "learning_rate": 1.7421372386426185e-06, "loss": 1.312, "step": 20400 }, { "epoch": 0.9, "grad_norm": 2.010845184326172, "learning_rate": 1.7319095802390087e-06, "loss": 1.3199, "step": 20500 }, { "epoch": 0.9, "eval_loss": 1.3053687810897827, "eval_runtime": 1865.2464, "eval_samples_per_second": 5.139, "eval_steps_per_second": 0.643, "step": 20500 }, { "epoch": 0.91, "grad_norm": 1.9459248781204224, "learning_rate": 1.7216708639959744e-06, "loss": 1.3091, "step": 20600 }, { "epoch": 0.91, "grad_norm": 2.0189294815063477, "learning_rate": 1.711421578112754e-06, "loss": 1.2958, "step": 20700 }, { "epoch": 0.91, "grad_norm": 2.0753591060638428, "learning_rate": 1.701162211292561e-06, "loss": 1.3041, "step": 20800 }, { "epoch": 0.92, "grad_norm": 2.1882078647613525, "learning_rate": 1.6908932527192886e-06, "loss": 1.3104, "step": 20900 }, { "epoch": 0.92, "grad_norm": 2.0360772609710693, "learning_rate": 1.6806151920341779e-06, "loss": 1.31, "step": 21000 }, { "epoch": 0.92, "eval_loss": 1.3050304651260376, "eval_runtime": 1868.8769, "eval_samples_per_second": 5.129, "eval_steps_per_second": 0.642, "step": 21000 }, { "epoch": 0.93, "grad_norm": 2.137718439102173, "learning_rate": 1.6703285193124748e-06, "loss": 1.3027, "step": 21100 }, { "epoch": 0.93, "grad_norm": 2.0401344299316406, "learning_rate": 1.660033725040063e-06, "loss": 1.3037, "step": 21200 }, { "epoch": 0.94, "grad_norm": 1.9637211561203003, "learning_rate": 1.649731300090074e-06, "loss": 1.2974, "step": 21300 }, { "epoch": 0.94, "grad_norm": 2.090510368347168, "learning_rate": 1.6394217356994827e-06, "loss": 1.2934, "step": 21400 }, { "epoch": 0.95, "grad_norm": 2.3479647636413574, "learning_rate": 1.6291055234456859e-06, "loss": 1.3201, "step": 21500 }, { "epoch": 0.95, "eval_loss": 1.3043450117111206, "eval_runtime": 1866.4601, "eval_samples_per_second": 5.136, "eval_steps_per_second": 0.642, "step": 21500 }, { "epoch": 0.95, "grad_norm": 2.1579744815826416, "learning_rate": 1.6187831552230603e-06, "loss": 1.2983, "step": 21600 }, { "epoch": 0.95, "grad_norm": 2.2169971466064453, "learning_rate": 1.608455123219511e-06, "loss": 1.3011, "step": 21700 }, { "epoch": 0.96, "grad_norm": 1.9283170700073242, "learning_rate": 1.5981219198929995e-06, "loss": 1.3036, "step": 21800 }, { "epoch": 0.96, "grad_norm": 1.9650486707687378, "learning_rate": 1.5877840379480678e-06, "loss": 1.3017, "step": 21900 }, { "epoch": 0.97, "grad_norm": 2.2148048877716064, "learning_rate": 1.5774419703123393e-06, "loss": 1.3118, "step": 22000 }, { "epoch": 0.97, "eval_loss": 1.303900122642517, "eval_runtime": 1863.2857, "eval_samples_per_second": 5.145, "eval_steps_per_second": 0.643, "step": 22000 }, { "epoch": 0.97, "grad_norm": 2.119337558746338, "learning_rate": 1.56709621011302e-06, "loss": 1.3038, "step": 22100 }, { "epoch": 0.98, "grad_norm": 2.062990427017212, "learning_rate": 1.5567472506533827e-06, "loss": 1.2878, "step": 22200 }, { "epoch": 0.98, "grad_norm": 2.1745545864105225, "learning_rate": 1.546395585389247e-06, "loss": 1.2964, "step": 22300 }, { "epoch": 0.98, "grad_norm": 2.1263065338134766, "learning_rate": 1.5360417079054494e-06, "loss": 1.295, "step": 22400 }, { "epoch": 0.99, "grad_norm": 2.044600248336792, "learning_rate": 1.5256861118923083e-06, "loss": 1.3049, "step": 22500 }, { "epoch": 0.99, "eval_loss": 1.303480625152588, "eval_runtime": 1862.9597, "eval_samples_per_second": 5.146, "eval_steps_per_second": 0.644, "step": 22500 }, { "epoch": 0.99, "grad_norm": 2.0431716442108154, "learning_rate": 1.5153292911220858e-06, "loss": 1.2992, "step": 22600 }, { "epoch": 1.0, "grad_norm": 2.1952879428863525, "learning_rate": 1.5049717394254412e-06, "loss": 1.2952, "step": 22700 }, { "epoch": 1.0, "grad_norm": 1.9967477321624756, "learning_rate": 1.4946139506678864e-06, "loss": 1.2967, "step": 22800 }, { "epoch": 1.01, "grad_norm": 2.1949024200439453, "learning_rate": 1.4842564187262365e-06, "loss": 1.3107, "step": 22900 }, { "epoch": 1.01, "grad_norm": 2.114365339279175, "learning_rate": 1.47389963746506e-06, "loss": 1.2946, "step": 23000 }, { "epoch": 1.01, "eval_loss": 1.3029588460922241, "eval_runtime": 1862.6203, "eval_samples_per_second": 5.147, "eval_steps_per_second": 0.644, "step": 23000 }, { "epoch": 1.02, "grad_norm": 2.3793764114379883, "learning_rate": 1.4635441007131343e-06, "loss": 1.3098, "step": 23100 }, { "epoch": 1.02, "grad_norm": 2.0778090953826904, "learning_rate": 1.453190302239894e-06, "loss": 1.3051, "step": 23200 }, { "epoch": 1.02, "grad_norm": 2.019946336746216, "learning_rate": 1.442838735731892e-06, "loss": 1.2966, "step": 23300 }, { "epoch": 1.03, "grad_norm": 2.1698460578918457, "learning_rate": 1.432489894769254e-06, "loss": 1.3029, "step": 23400 }, { "epoch": 1.03, "grad_norm": 2.034184217453003, "learning_rate": 1.4221442728021506e-06, "loss": 1.316, "step": 23500 }, { "epoch": 1.03, "eval_loss": 1.3026907444000244, "eval_runtime": 1863.5533, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 23500 }, { "epoch": 1.04, "grad_norm": 2.025113582611084, "learning_rate": 1.4118023631272633e-06, "loss": 1.2935, "step": 23600 }, { "epoch": 1.04, "grad_norm": 2.073225259780884, "learning_rate": 1.4014646588642656e-06, "loss": 1.2967, "step": 23700 }, { "epoch": 1.05, "grad_norm": 2.120889186859131, "learning_rate": 1.391131652932311e-06, "loss": 1.3013, "step": 23800 }, { "epoch": 1.05, "grad_norm": 2.005122423171997, "learning_rate": 1.3808038380265258e-06, "loss": 1.2928, "step": 23900 }, { "epoch": 1.06, "grad_norm": 2.147322654724121, "learning_rate": 1.3704817065945235e-06, "loss": 1.3059, "step": 24000 }, { "epoch": 1.06, "eval_loss": 1.3021385669708252, "eval_runtime": 1865.741, "eval_samples_per_second": 5.138, "eval_steps_per_second": 0.643, "step": 24000 }, { "epoch": 1.06, "grad_norm": 2.336851119995117, "learning_rate": 1.360165750812916e-06, "loss": 1.2986, "step": 24100 }, { "epoch": 1.06, "grad_norm": 2.077028274536133, "learning_rate": 1.3498564625638525e-06, "loss": 1.3062, "step": 24200 }, { "epoch": 1.07, "grad_norm": 2.137953519821167, "learning_rate": 1.3395543334115616e-06, "loss": 1.3071, "step": 24300 }, { "epoch": 1.07, "grad_norm": 2.054823875427246, "learning_rate": 1.3292598545789157e-06, "loss": 1.2981, "step": 24400 }, { "epoch": 1.08, "grad_norm": 2.045687198638916, "learning_rate": 1.3189735169240044e-06, "loss": 1.2971, "step": 24500 }, { "epoch": 1.08, "eval_loss": 1.3017216920852661, "eval_runtime": 1863.3072, "eval_samples_per_second": 5.145, "eval_steps_per_second": 0.643, "step": 24500 }, { "epoch": 1.08, "grad_norm": 2.2879574298858643, "learning_rate": 1.308695810916735e-06, "loss": 1.2943, "step": 24600 }, { "epoch": 1.09, "grad_norm": 2.1517672538757324, "learning_rate": 1.2984272266154414e-06, "loss": 1.3083, "step": 24700 }, { "epoch": 1.09, "grad_norm": 2.0498697757720947, "learning_rate": 1.288168253643521e-06, "loss": 1.2873, "step": 24800 }, { "epoch": 1.09, "grad_norm": 2.1879069805145264, "learning_rate": 1.2779193811660837e-06, "loss": 1.2902, "step": 24900 }, { "epoch": 1.1, "grad_norm": 2.2728075981140137, "learning_rate": 1.2676810978666353e-06, "loss": 1.3001, "step": 25000 }, { "epoch": 1.1, "eval_loss": 1.3014191389083862, "eval_runtime": 1863.092, "eval_samples_per_second": 5.145, "eval_steps_per_second": 0.644, "step": 25000 }, { "epoch": 1.1, "grad_norm": 2.2328622341156006, "learning_rate": 1.2574538919237673e-06, "loss": 1.3028, "step": 25100 }, { "epoch": 1.11, "grad_norm": 2.2706058025360107, "learning_rate": 1.2472382509878873e-06, "loss": 1.2931, "step": 25200 }, { "epoch": 1.11, "grad_norm": 2.234330892562866, "learning_rate": 1.2370346621579623e-06, "loss": 1.3025, "step": 25300 }, { "epoch": 1.12, "grad_norm": 2.125455379486084, "learning_rate": 1.2268436119582958e-06, "loss": 1.3031, "step": 25400 }, { "epoch": 1.12, "grad_norm": 2.0227627754211426, "learning_rate": 1.2166655863153263e-06, "loss": 1.2743, "step": 25500 }, { "epoch": 1.12, "eval_loss": 1.3010908365249634, "eval_runtime": 1862.8931, "eval_samples_per_second": 5.146, "eval_steps_per_second": 0.644, "step": 25500 }, { "epoch": 1.13, "grad_norm": 2.0580368041992188, "learning_rate": 1.2065010705344618e-06, "loss": 1.3054, "step": 25600 }, { "epoch": 1.13, "grad_norm": 2.1647911071777344, "learning_rate": 1.1963505492769355e-06, "loss": 1.2916, "step": 25700 }, { "epoch": 1.13, "grad_norm": 2.0057008266448975, "learning_rate": 1.1862145065366998e-06, "loss": 1.3016, "step": 25800 }, { "epoch": 1.14, "grad_norm": 2.097618341445923, "learning_rate": 1.1760934256173447e-06, "loss": 1.2946, "step": 25900 }, { "epoch": 1.14, "grad_norm": 2.1340222358703613, "learning_rate": 1.1659877891090587e-06, "loss": 1.2983, "step": 26000 }, { "epoch": 1.14, "eval_loss": 1.300696849822998, "eval_runtime": 1862.6712, "eval_samples_per_second": 5.146, "eval_steps_per_second": 0.644, "step": 26000 }, { "epoch": 1.15, "grad_norm": 2.054945468902588, "learning_rate": 1.155898078865611e-06, "loss": 1.3035, "step": 26100 }, { "epoch": 1.15, "grad_norm": 2.0940968990325928, "learning_rate": 1.1458247759813828e-06, "loss": 1.2955, "step": 26200 }, { "epoch": 1.16, "grad_norm": 2.199674129486084, "learning_rate": 1.135768360768423e-06, "loss": 1.3013, "step": 26300 }, { "epoch": 1.16, "grad_norm": 2.2380006313323975, "learning_rate": 1.125729312733549e-06, "loss": 1.3017, "step": 26400 }, { "epoch": 1.16, "grad_norm": 2.0469789505004883, "learning_rate": 1.1157081105554801e-06, "loss": 1.3048, "step": 26500 }, { "epoch": 1.16, "eval_loss": 1.3003997802734375, "eval_runtime": 1862.5897, "eval_samples_per_second": 5.147, "eval_steps_per_second": 0.644, "step": 26500 }, { "epoch": 1.17, "grad_norm": 2.136064052581787, "learning_rate": 1.1057052320620172e-06, "loss": 1.2849, "step": 26600 }, { "epoch": 1.17, "grad_norm": 2.2330269813537598, "learning_rate": 1.0957211542072556e-06, "loss": 1.2946, "step": 26700 }, { "epoch": 1.18, "grad_norm": 2.45259690284729, "learning_rate": 1.085756353048846e-06, "loss": 1.2956, "step": 26800 }, { "epoch": 1.18, "grad_norm": 2.133361339569092, "learning_rate": 1.0758113037252912e-06, "loss": 1.3041, "step": 26900 }, { "epoch": 1.19, "grad_norm": 2.1824896335601807, "learning_rate": 1.065886480433296e-06, "loss": 1.2945, "step": 27000 }, { "epoch": 1.19, "eval_loss": 1.2999972105026245, "eval_runtime": 1863.119, "eval_samples_per_second": 5.145, "eval_steps_per_second": 0.644, "step": 27000 }, { "epoch": 1.19, "grad_norm": 2.218424081802368, "learning_rate": 1.0559823564051523e-06, "loss": 1.3011, "step": 27100 }, { "epoch": 1.2, "grad_norm": 2.0408236980438232, "learning_rate": 1.0460994038861768e-06, "loss": 1.2935, "step": 27200 }, { "epoch": 1.2, "grad_norm": 2.0885488986968994, "learning_rate": 1.0362380941121925e-06, "loss": 1.3001, "step": 27300 }, { "epoch": 1.2, "grad_norm": 2.130805015563965, "learning_rate": 1.0263988972870613e-06, "loss": 1.3081, "step": 27400 }, { "epoch": 1.21, "grad_norm": 2.0574839115142822, "learning_rate": 1.0165822825602596e-06, "loss": 1.3071, "step": 27500 }, { "epoch": 1.21, "eval_loss": 1.2996830940246582, "eval_runtime": 1863.286, "eval_samples_per_second": 5.145, "eval_steps_per_second": 0.643, "step": 27500 }, { "epoch": 1.21, "grad_norm": 2.075486421585083, "learning_rate": 1.0067887180045145e-06, "loss": 1.296, "step": 27600 }, { "epoch": 1.22, "grad_norm": 2.07733154296875, "learning_rate": 9.970186705934803e-07, "loss": 1.3015, "step": 27700 }, { "epoch": 1.22, "grad_norm": 2.1551625728607178, "learning_rate": 9.872726061794744e-07, "loss": 1.2951, "step": 27800 }, { "epoch": 1.23, "grad_norm": 2.174243927001953, "learning_rate": 9.775509894712656e-07, "loss": 1.3052, "step": 27900 }, { "epoch": 1.23, "grad_norm": 2.11954665184021, "learning_rate": 9.678542840119123e-07, "loss": 1.294, "step": 28000 }, { "epoch": 1.23, "eval_loss": 1.2994180917739868, "eval_runtime": 1864.2204, "eval_samples_per_second": 5.142, "eval_steps_per_second": 0.643, "step": 28000 }, { "epoch": 1.24, "grad_norm": 2.0479559898376465, "learning_rate": 9.581829521566663e-07, "loss": 1.3078, "step": 28100 }, { "epoch": 1.24, "grad_norm": 2.1119577884674072, "learning_rate": 9.485374550509197e-07, "loss": 1.2878, "step": 28200 }, { "epoch": 1.24, "grad_norm": 2.1557509899139404, "learning_rate": 9.389182526082229e-07, "loss": 1.2866, "step": 28300 }, { "epoch": 1.25, "grad_norm": 2.226323127746582, "learning_rate": 9.293258034883504e-07, "loss": 1.2977, "step": 28400 }, { "epoch": 1.25, "grad_norm": 2.1167123317718506, "learning_rate": 9.197605650754353e-07, "loss": 1.2937, "step": 28500 }, { "epoch": 1.25, "eval_loss": 1.2991981506347656, "eval_runtime": 1863.4013, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 28500 }, { "epoch": 1.26, "grad_norm": 2.254310131072998, "learning_rate": 9.10222993456156e-07, "loss": 1.3016, "step": 28600 }, { "epoch": 1.26, "grad_norm": 2.0988237857818604, "learning_rate": 9.007135433979935e-07, "loss": 1.3, "step": 28700 }, { "epoch": 1.27, "grad_norm": 2.0965945720672607, "learning_rate": 8.912326683275442e-07, "loss": 1.2877, "step": 28800 }, { "epoch": 1.27, "grad_norm": 2.037484884262085, "learning_rate": 8.817808203089022e-07, "loss": 1.3084, "step": 28900 }, { "epoch": 1.27, "grad_norm": 2.0629284381866455, "learning_rate": 8.723584500221013e-07, "loss": 1.2857, "step": 29000 }, { "epoch": 1.27, "eval_loss": 1.298945426940918, "eval_runtime": 1863.2462, "eval_samples_per_second": 5.145, "eval_steps_per_second": 0.644, "step": 29000 }, { "epoch": 1.28, "grad_norm": 2.1192641258239746, "learning_rate": 8.62966006741628e-07, "loss": 1.3011, "step": 29100 }, { "epoch": 1.28, "grad_norm": 2.2492525577545166, "learning_rate": 8.536039383150001e-07, "loss": 1.2882, "step": 29200 }, { "epoch": 1.29, "grad_norm": 2.1896586418151855, "learning_rate": 8.442726911414094e-07, "loss": 1.2933, "step": 29300 }, { "epoch": 1.29, "grad_norm": 2.073396921157837, "learning_rate": 8.349727101504388e-07, "loss": 1.3119, "step": 29400 }, { "epoch": 1.3, "grad_norm": 2.1311280727386475, "learning_rate": 8.257044387808482e-07, "loss": 1.2921, "step": 29500 }, { "epoch": 1.3, "eval_loss": 1.2986724376678467, "eval_runtime": 1862.6142, "eval_samples_per_second": 5.147, "eval_steps_per_second": 0.644, "step": 29500 }, { "epoch": 1.3, "grad_norm": 2.1984448432922363, "learning_rate": 8.164683189594269e-07, "loss": 1.2933, "step": 29600 }, { "epoch": 1.31, "grad_norm": 2.1039230823516846, "learning_rate": 8.072647910799269e-07, "loss": 1.3129, "step": 29700 }, { "epoch": 1.31, "grad_norm": 1.9746812582015991, "learning_rate": 7.980942939820596e-07, "loss": 1.305, "step": 29800 }, { "epoch": 1.31, "grad_norm": 2.099011182785034, "learning_rate": 7.889572649305736e-07, "loss": 1.2956, "step": 29900 }, { "epoch": 1.32, "grad_norm": 2.3024837970733643, "learning_rate": 7.798541395944053e-07, "loss": 1.2794, "step": 30000 }, { "epoch": 1.32, "eval_loss": 1.2984648942947388, "eval_runtime": 1861.962, "eval_samples_per_second": 5.148, "eval_steps_per_second": 0.644, "step": 30000 }, { "epoch": 1.32, "grad_norm": 2.1850967407226562, "learning_rate": 7.707853520259052e-07, "loss": 1.2854, "step": 30100 }, { "epoch": 1.33, "grad_norm": 2.239342212677002, "learning_rate": 7.617513346401388e-07, "loss": 1.2919, "step": 30200 }, { "epoch": 1.33, "grad_norm": 2.3949005603790283, "learning_rate": 7.527525181942738e-07, "loss": 1.3048, "step": 30300 }, { "epoch": 1.34, "grad_norm": 2.2545089721679688, "learning_rate": 7.437893317670351e-07, "loss": 1.2909, "step": 30400 }, { "epoch": 1.34, "grad_norm": 2.2839066982269287, "learning_rate": 7.348622027382499e-07, "loss": 1.299, "step": 30500 }, { "epoch": 1.34, "eval_loss": 1.2982053756713867, "eval_runtime": 1862.6374, "eval_samples_per_second": 5.146, "eval_steps_per_second": 0.644, "step": 30500 }, { "epoch": 1.35, "grad_norm": 2.43034029006958, "learning_rate": 7.259715567684677e-07, "loss": 1.2973, "step": 30600 }, { "epoch": 1.35, "grad_norm": 2.251255989074707, "learning_rate": 7.171178177786646e-07, "loss": 1.2973, "step": 30700 }, { "epoch": 1.35, "grad_norm": 2.132671356201172, "learning_rate": 7.083014079300282e-07, "loss": 1.2931, "step": 30800 }, { "epoch": 1.36, "grad_norm": 2.243849277496338, "learning_rate": 6.995227476038316e-07, "loss": 1.2967, "step": 30900 }, { "epoch": 1.36, "grad_norm": 2.284914016723633, "learning_rate": 6.907822553813857e-07, "loss": 1.2925, "step": 31000 }, { "epoch": 1.36, "eval_loss": 1.2980060577392578, "eval_runtime": 1862.987, "eval_samples_per_second": 5.146, "eval_steps_per_second": 0.644, "step": 31000 }, { "epoch": 1.37, "grad_norm": 2.1762306690216064, "learning_rate": 6.820803480240832e-07, "loss": 1.3035, "step": 31100 }, { "epoch": 1.37, "grad_norm": 2.1974294185638428, "learning_rate": 6.734174404535233e-07, "loss": 1.2982, "step": 31200 }, { "epoch": 1.38, "grad_norm": 2.1154937744140625, "learning_rate": 6.647939457317332e-07, "loss": 1.2987, "step": 31300 }, { "epoch": 1.38, "grad_norm": 2.2262349128723145, "learning_rate": 6.562102750414655e-07, "loss": 1.2894, "step": 31400 }, { "epoch": 1.38, "grad_norm": 2.2507781982421875, "learning_rate": 6.476668376665978e-07, "loss": 1.3024, "step": 31500 }, { "epoch": 1.38, "eval_loss": 1.2977538108825684, "eval_runtime": 1863.5376, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 31500 }, { "epoch": 1.39, "grad_norm": 2.0441462993621826, "learning_rate": 6.391640409726157e-07, "loss": 1.2939, "step": 31600 }, { "epoch": 1.39, "grad_norm": 2.172423839569092, "learning_rate": 6.307022903871869e-07, "loss": 1.3002, "step": 31700 }, { "epoch": 1.4, "grad_norm": 1.932667851448059, "learning_rate": 6.222819893808335e-07, "loss": 1.292, "step": 31800 }, { "epoch": 1.4, "grad_norm": 2.173680305480957, "learning_rate": 6.139035394476905e-07, "loss": 1.3025, "step": 31900 }, { "epoch": 1.41, "grad_norm": 2.1886558532714844, "learning_rate": 6.055673400863636e-07, "loss": 1.3103, "step": 32000 }, { "epoch": 1.41, "eval_loss": 1.2976477146148682, "eval_runtime": 1862.7508, "eval_samples_per_second": 5.146, "eval_steps_per_second": 0.644, "step": 32000 }, { "epoch": 1.41, "grad_norm": 2.3404366970062256, "learning_rate": 5.972737887808805e-07, "loss": 1.2915, "step": 32100 }, { "epoch": 1.42, "grad_norm": 2.2503886222839355, "learning_rate": 5.890232809817378e-07, "loss": 1.2887, "step": 32200 }, { "epoch": 1.42, "grad_norm": 2.1247682571411133, "learning_rate": 5.808162100870441e-07, "loss": 1.297, "step": 32300 }, { "epoch": 1.42, "grad_norm": 2.1436526775360107, "learning_rate": 5.72652967423765e-07, "loss": 1.3008, "step": 32400 }, { "epoch": 1.43, "grad_norm": 2.1058590412139893, "learning_rate": 5.645339422290599e-07, "loss": 1.2956, "step": 32500 }, { "epoch": 1.43, "eval_loss": 1.2974357604980469, "eval_runtime": 1866.987, "eval_samples_per_second": 5.134, "eval_steps_per_second": 0.642, "step": 32500 }, { "epoch": 1.43, "grad_norm": 2.1536617279052734, "learning_rate": 5.564595216317266e-07, "loss": 1.299, "step": 32600 }, { "epoch": 1.44, "grad_norm": 2.031917095184326, "learning_rate": 5.484300906337403e-07, "loss": 1.2828, "step": 32700 }, { "epoch": 1.44, "grad_norm": 1.9297816753387451, "learning_rate": 5.404460320918966e-07, "loss": 1.2937, "step": 32800 }, { "epoch": 1.45, "grad_norm": 2.253413677215576, "learning_rate": 5.325077266995542e-07, "loss": 1.3078, "step": 32900 }, { "epoch": 1.45, "grad_norm": 2.250991106033325, "learning_rate": 5.24615552968487e-07, "loss": 1.2788, "step": 33000 }, { "epoch": 1.45, "eval_loss": 1.297324776649475, "eval_runtime": 1864.5246, "eval_samples_per_second": 5.141, "eval_steps_per_second": 0.643, "step": 33000 }, { "epoch": 1.46, "grad_norm": 2.186206579208374, "learning_rate": 5.167698872108312e-07, "loss": 1.3048, "step": 33100 }, { "epoch": 1.46, "grad_norm": 2.1334028244018555, "learning_rate": 5.089711035211465e-07, "loss": 1.279, "step": 33200 }, { "epoch": 1.46, "grad_norm": 2.334712266921997, "learning_rate": 5.012195737585756e-07, "loss": 1.2949, "step": 33300 }, { "epoch": 1.47, "grad_norm": 2.2277867794036865, "learning_rate": 4.935156675291152e-07, "loss": 1.2998, "step": 33400 }, { "epoch": 1.47, "grad_norm": 2.1463205814361572, "learning_rate": 4.858597521679899e-07, "loss": 1.2878, "step": 33500 }, { "epoch": 1.47, "eval_loss": 1.2971347570419312, "eval_runtime": 1865.0446, "eval_samples_per_second": 5.14, "eval_steps_per_second": 0.643, "step": 33500 }, { "epoch": 1.48, "grad_norm": 2.0899240970611572, "learning_rate": 4.782521927221414e-07, "loss": 1.2927, "step": 33600 }, { "epoch": 1.48, "grad_norm": 2.1801438331604004, "learning_rate": 4.70693351932817e-07, "loss": 1.2822, "step": 33700 }, { "epoch": 1.49, "grad_norm": 2.0565810203552246, "learning_rate": 4.631835902182795e-07, "loss": 1.2998, "step": 33800 }, { "epoch": 1.49, "grad_norm": 2.0865461826324463, "learning_rate": 4.55723265656616e-07, "loss": 1.2912, "step": 33900 }, { "epoch": 1.49, "grad_norm": 2.186192750930786, "learning_rate": 4.483127339686686e-07, "loss": 1.2893, "step": 34000 }, { "epoch": 1.49, "eval_loss": 1.297006368637085, "eval_runtime": 1864.1525, "eval_samples_per_second": 5.142, "eval_steps_per_second": 0.643, "step": 34000 }, { "epoch": 1.5, "grad_norm": 2.418661594390869, "learning_rate": 4.409523485010707e-07, "loss": 1.2823, "step": 34100 }, { "epoch": 1.5, "grad_norm": 2.213181972503662, "learning_rate": 4.3364246020940003e-07, "loss": 1.2921, "step": 34200 }, { "epoch": 1.51, "grad_norm": 2.1837573051452637, "learning_rate": 4.263834176414426e-07, "loss": 1.2834, "step": 34300 }, { "epoch": 1.51, "grad_norm": 2.2444188594818115, "learning_rate": 4.191755669205763e-07, "loss": 1.3072, "step": 34400 }, { "epoch": 1.52, "grad_norm": 2.1974008083343506, "learning_rate": 4.120192517292637e-07, "loss": 1.295, "step": 34500 }, { "epoch": 1.52, "eval_loss": 1.2968517541885376, "eval_runtime": 1863.4227, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 34500 }, { "epoch": 1.52, "grad_norm": 2.2060582637786865, "learning_rate": 4.0491481329266823e-07, "loss": 1.294, "step": 34600 }, { "epoch": 1.53, "grad_norm": 2.2055234909057617, "learning_rate": 3.978625903623813e-07, "loss": 1.301, "step": 34700 }, { "epoch": 1.53, "grad_norm": 2.3005611896514893, "learning_rate": 3.9086291920027183e-07, "loss": 1.2909, "step": 34800 }, { "epoch": 1.53, "grad_norm": 2.2892158031463623, "learning_rate": 3.839161335624504e-07, "loss": 1.2898, "step": 34900 }, { "epoch": 1.54, "grad_norm": 2.2120680809020996, "learning_rate": 3.7702256468335877e-07, "loss": 1.2937, "step": 35000 }, { "epoch": 1.54, "eval_loss": 1.2967385053634644, "eval_runtime": 1863.4438, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 35000 }, { "epoch": 1.54, "grad_norm": 2.06485652923584, "learning_rate": 3.7018254125997167e-07, "loss": 1.3034, "step": 35100 }, { "epoch": 1.55, "grad_norm": 2.2690818309783936, "learning_rate": 3.6339638943612876e-07, "loss": 1.2816, "step": 35200 }, { "epoch": 1.55, "grad_norm": 2.1384482383728027, "learning_rate": 3.566644327869784e-07, "loss": 1.2958, "step": 35300 }, { "epoch": 1.56, "grad_norm": 2.262807607650757, "learning_rate": 3.4998699230355445e-07, "loss": 1.2906, "step": 35400 }, { "epoch": 1.56, "grad_norm": 2.1555404663085938, "learning_rate": 3.4336438637746716e-07, "loss": 1.2911, "step": 35500 }, { "epoch": 1.56, "eval_loss": 1.2966289520263672, "eval_runtime": 1863.3587, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 35500 }, { "epoch": 1.56, "grad_norm": 2.088890552520752, "learning_rate": 3.3679693078572135e-07, "loss": 1.2905, "step": 35600 }, { "epoch": 1.57, "grad_norm": 2.362851142883301, "learning_rate": 3.302849386756631e-07, "loss": 1.2877, "step": 35700 }, { "epoch": 1.57, "grad_norm": 2.3154048919677734, "learning_rate": 3.2382872055004437e-07, "loss": 1.288, "step": 35800 }, { "epoch": 1.58, "grad_norm": 2.204124927520752, "learning_rate": 3.1742858425222153e-07, "loss": 1.2888, "step": 35900 }, { "epoch": 1.58, "grad_norm": 2.0219850540161133, "learning_rate": 3.1108483495147276e-07, "loss": 1.2962, "step": 36000 }, { "epoch": 1.58, "eval_loss": 1.2965224981307983, "eval_runtime": 1863.8916, "eval_samples_per_second": 5.143, "eval_steps_per_second": 0.643, "step": 36000 }, { "epoch": 1.59, "grad_norm": 2.456171989440918, "learning_rate": 3.0479777512845107e-07, "loss": 1.291, "step": 36100 }, { "epoch": 1.59, "grad_norm": 2.043088674545288, "learning_rate": 2.9856770456075904e-07, "loss": 1.2926, "step": 36200 }, { "epoch": 1.6, "grad_norm": 2.3121893405914307, "learning_rate": 2.9239492030865573e-07, "loss": 1.2933, "step": 36300 }, { "epoch": 1.6, "grad_norm": 2.117414712905884, "learning_rate": 2.862797167008913e-07, "loss": 1.2842, "step": 36400 }, { "epoch": 1.6, "grad_norm": 2.2130415439605713, "learning_rate": 2.802223853206752e-07, "loss": 1.2952, "step": 36500 }, { "epoch": 1.6, "eval_loss": 1.2964164018630981, "eval_runtime": 1863.5413, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 36500 }, { "epoch": 1.61, "grad_norm": 2.1747307777404785, "learning_rate": 2.7422321499177006e-07, "loss": 1.2953, "step": 36600 }, { "epoch": 1.61, "grad_norm": 2.238208293914795, "learning_rate": 2.6828249176472284e-07, "loss": 1.2901, "step": 36700 }, { "epoch": 1.62, "grad_norm": 2.170766830444336, "learning_rate": 2.624004989032238e-07, "loss": 1.3061, "step": 36800 }, { "epoch": 1.62, "grad_norm": 2.2250659465789795, "learning_rate": 2.565775168706008e-07, "loss": 1.3015, "step": 36900 }, { "epoch": 1.63, "grad_norm": 2.16519832611084, "learning_rate": 2.5081382331644484e-07, "loss": 1.3065, "step": 37000 }, { "epoch": 1.63, "eval_loss": 1.2963379621505737, "eval_runtime": 1863.6216, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 37000 }, { "epoch": 1.63, "grad_norm": 2.3075404167175293, "learning_rate": 2.451096930633741e-07, "loss": 1.2933, "step": 37100 }, { "epoch": 1.64, "grad_norm": 2.204376459121704, "learning_rate": 2.3946539809392637e-07, "loss": 1.2938, "step": 37200 }, { "epoch": 1.64, "grad_norm": 2.1348817348480225, "learning_rate": 2.33881207537594e-07, "loss": 1.2975, "step": 37300 }, { "epoch": 1.64, "grad_norm": 2.2030017375946045, "learning_rate": 2.283573876579882e-07, "loss": 1.2873, "step": 37400 }, { "epoch": 1.65, "grad_norm": 2.280266523361206, "learning_rate": 2.2289420184014636e-07, "loss": 1.2862, "step": 37500 }, { "epoch": 1.65, "eval_loss": 1.2962608337402344, "eval_runtime": 1864.1974, "eval_samples_per_second": 5.142, "eval_steps_per_second": 0.643, "step": 37500 }, { "epoch": 1.65, "grad_norm": 2.2561440467834473, "learning_rate": 2.1749191057796996e-07, "loss": 1.2926, "step": 37600 }, { "epoch": 1.66, "grad_norm": 2.064223289489746, "learning_rate": 2.1215077146180688e-07, "loss": 1.2911, "step": 37700 }, { "epoch": 1.66, "grad_norm": 2.0723884105682373, "learning_rate": 2.0687103916616612e-07, "loss": 1.2927, "step": 37800 }, { "epoch": 1.67, "grad_norm": 2.1863532066345215, "learning_rate": 2.0165296543757777e-07, "loss": 1.2958, "step": 37900 }, { "epoch": 1.67, "grad_norm": 2.242121458053589, "learning_rate": 1.9649679908258606e-07, "loss": 1.2856, "step": 38000 }, { "epoch": 1.67, "eval_loss": 1.2961931228637695, "eval_runtime": 1863.7107, "eval_samples_per_second": 5.144, "eval_steps_per_second": 0.643, "step": 38000 }, { "epoch": 1.67, "grad_norm": 2.2258660793304443, "learning_rate": 1.914027859558884e-07, "loss": 1.2895, "step": 38100 }, { "epoch": 1.68, "grad_norm": 2.194655418395996, "learning_rate": 1.8637116894861117e-07, "loss": 1.299, "step": 38200 }, { "epoch": 1.68, "grad_norm": 2.083817958831787, "learning_rate": 1.8140218797672898e-07, "loss": 1.2868, "step": 38300 }, { "epoch": 1.69, "grad_norm": 2.1037402153015137, "learning_rate": 1.7649607996962425e-07, "loss": 1.2833, "step": 38400 }, { "epoch": 1.69, "grad_norm": 2.3146462440490723, "learning_rate": 1.71653078858791e-07, "loss": 1.295, "step": 38500 }, { "epoch": 1.69, "eval_loss": 1.2961418628692627, "eval_runtime": 1863.9344, "eval_samples_per_second": 5.143, "eval_steps_per_second": 0.643, "step": 38500 }, { "epoch": 1.7, "grad_norm": 2.178917646408081, "learning_rate": 1.6687341556667956e-07, "loss": 1.2967, "step": 38600 }, { "epoch": 1.7, "grad_norm": 2.230252742767334, "learning_rate": 1.6215731799568707e-07, "loss": 1.2912, "step": 38700 }, { "epoch": 1.71, "grad_norm": 2.027996301651001, "learning_rate": 1.575050110172896e-07, "loss": 1.2931, "step": 38800 }, { "epoch": 1.71, "grad_norm": 2.343449831008911, "learning_rate": 1.5291671646132082e-07, "loss": 1.2896, "step": 38900 }, { "epoch": 1.71, "grad_norm": 2.1190574169158936, "learning_rate": 1.483926531053933e-07, "loss": 1.2926, "step": 39000 }, { "epoch": 1.71, "eval_loss": 1.2960731983184814, "eval_runtime": 1864.8274, "eval_samples_per_second": 5.14, "eval_steps_per_second": 0.643, "step": 39000 }, { "epoch": 1.72, "grad_norm": 2.0614373683929443, "learning_rate": 1.4393303666446873e-07, "loss": 1.2939, "step": 39100 }, { "epoch": 1.72, "grad_norm": 2.185314178466797, "learning_rate": 1.395380797805706e-07, "loss": 1.293, "step": 39200 }, { "epoch": 1.73, "grad_norm": 2.1229093074798584, "learning_rate": 1.3520799201264644e-07, "loss": 1.3041, "step": 39300 }, { "epoch": 1.73, "grad_norm": 2.213289976119995, "learning_rate": 1.3094297982657484e-07, "loss": 1.2956, "step": 39400 }, { "epoch": 1.74, "grad_norm": 2.1949591636657715, "learning_rate": 1.267432465853209e-07, "loss": 1.2928, "step": 39500 }, { "epoch": 1.74, "eval_loss": 1.2960156202316284, "eval_runtime": 1862.2042, "eval_samples_per_second": 5.148, "eval_steps_per_second": 0.644, "step": 39500 }, { "epoch": 1.74, "grad_norm": 2.0892333984375, "learning_rate": 1.2260899253924008e-07, "loss": 1.2991, "step": 39600 }, { "epoch": 1.75, "grad_norm": 2.1460952758789062, "learning_rate": 1.1854041481652849e-07, "loss": 1.2949, "step": 39700 }, { "epoch": 1.75, "grad_norm": 2.241546630859375, "learning_rate": 1.1453770741382607e-07, "loss": 1.2836, "step": 39800 }, { "epoch": 1.75, "grad_norm": 2.225529670715332, "learning_rate": 1.1060106118696345e-07, "loss": 1.2923, "step": 39900 }, { "epoch": 1.76, "grad_norm": 2.106523036956787, "learning_rate": 1.0673066384186425e-07, "loss": 1.2966, "step": 40000 }, { "epoch": 1.76, "eval_loss": 1.295967698097229, "eval_runtime": 1866.8915, "eval_samples_per_second": 5.135, "eval_steps_per_second": 0.642, "step": 40000 } ], "logging_steps": 100, "max_steps": 45496, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 1000, "total_flos": 1.403599875735552e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }