{ "best_metric": null, "best_model_checkpoint": null, "epoch": 10.0, "eval_steps": 500, "global_step": 3710, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.026954177897574125, "grad_norm": 93.09680938720703, "learning_rate": 7.692307692307694e-06, "loss": 5.4416, "step": 10 }, { "epoch": 0.05390835579514825, "grad_norm": 1869.5386962890625, "learning_rate": 1.5384615384615387e-05, "loss": 4.4275, "step": 20 }, { "epoch": 0.08086253369272237, "grad_norm": 29.11223602294922, "learning_rate": 2.307692307692308e-05, "loss": 2.2122, "step": 30 }, { "epoch": 0.1078167115902965, "grad_norm": 45.186805725097656, "learning_rate": 3.0769230769230774e-05, "loss": 1.9827, "step": 40 }, { "epoch": 0.1347708894878706, "grad_norm": 88.7030029296875, "learning_rate": 3.846153846153846e-05, "loss": 1.9525, "step": 50 }, { "epoch": 0.16172506738544473, "grad_norm": 11.193861961364746, "learning_rate": 4.615384615384616e-05, "loss": 1.401, "step": 60 }, { "epoch": 0.18867924528301888, "grad_norm": 1.629717469215393, "learning_rate": 4.993141289437586e-05, "loss": 1.4541, "step": 70 }, { "epoch": 0.215633423180593, "grad_norm": 6.375326156616211, "learning_rate": 4.9794238683127575e-05, "loss": 1.2295, "step": 80 }, { "epoch": 0.24258760107816713, "grad_norm": 2.1516671180725098, "learning_rate": 4.965706447187929e-05, "loss": 1.8626, "step": 90 }, { "epoch": 0.2695417789757412, "grad_norm": 2.126923084259033, "learning_rate": 4.9519890260631e-05, "loss": 1.5149, "step": 100 }, { "epoch": 0.29649595687331537, "grad_norm": 3.0875656604766846, "learning_rate": 4.938271604938271e-05, "loss": 1.7612, "step": 110 }, { "epoch": 0.32345013477088946, "grad_norm": 3.2731056213378906, "learning_rate": 4.924554183813443e-05, "loss": 1.1813, "step": 120 }, { "epoch": 0.3504043126684636, "grad_norm": 1.5794568061828613, "learning_rate": 4.9108367626886145e-05, "loss": 1.2228, "step": 130 }, { "epoch": 0.37735849056603776, "grad_norm": 1.0991185903549194, "learning_rate": 4.8971193415637865e-05, "loss": 2.1294, "step": 140 }, { "epoch": 0.40431266846361186, "grad_norm": 0.7588101029396057, "learning_rate": 4.883401920438958e-05, "loss": 1.6584, "step": 150 }, { "epoch": 0.431266846361186, "grad_norm": 0.6242086887359619, "learning_rate": 4.86968449931413e-05, "loss": 1.1689, "step": 160 }, { "epoch": 0.4582210242587601, "grad_norm": 5.792896270751953, "learning_rate": 4.855967078189301e-05, "loss": 1.3696, "step": 170 }, { "epoch": 0.48517520215633425, "grad_norm": 2.379605770111084, "learning_rate": 4.842249657064472e-05, "loss": 1.2535, "step": 180 }, { "epoch": 0.5121293800539084, "grad_norm": 0.7756444811820984, "learning_rate": 4.8285322359396435e-05, "loss": 1.2733, "step": 190 }, { "epoch": 0.5390835579514824, "grad_norm": 0.9577376246452332, "learning_rate": 4.814814814814815e-05, "loss": 1.366, "step": 200 }, { "epoch": 0.5660377358490566, "grad_norm": 0.7913112640380859, "learning_rate": 4.801097393689987e-05, "loss": 0.8251, "step": 210 }, { "epoch": 0.5929919137466307, "grad_norm": 1.1061924695968628, "learning_rate": 4.787379972565158e-05, "loss": 1.0164, "step": 220 }, { "epoch": 0.6199460916442049, "grad_norm": 0.854999840259552, "learning_rate": 4.773662551440329e-05, "loss": 1.2698, "step": 230 }, { "epoch": 0.6469002695417789, "grad_norm": 0.9977573156356812, "learning_rate": 4.7599451303155006e-05, "loss": 1.2353, "step": 240 }, { "epoch": 0.6738544474393531, "grad_norm": 1.3502521514892578, "learning_rate": 4.7462277091906725e-05, "loss": 1.3206, "step": 250 }, { "epoch": 0.7008086253369272, "grad_norm": 0.7636290192604065, "learning_rate": 4.732510288065844e-05, "loss": 1.2825, "step": 260 }, { "epoch": 0.7277628032345014, "grad_norm": 1.700835108757019, "learning_rate": 4.718792866941015e-05, "loss": 1.0656, "step": 270 }, { "epoch": 0.7547169811320755, "grad_norm": 0.7844202518463135, "learning_rate": 4.7050754458161864e-05, "loss": 1.4337, "step": 280 }, { "epoch": 0.7816711590296496, "grad_norm": 2.4619789123535156, "learning_rate": 4.691358024691358e-05, "loss": 1.2912, "step": 290 }, { "epoch": 0.8086253369272237, "grad_norm": 2.110649347305298, "learning_rate": 4.6776406035665296e-05, "loss": 1.1539, "step": 300 }, { "epoch": 0.8355795148247979, "grad_norm": 1.690564751625061, "learning_rate": 4.6639231824417016e-05, "loss": 1.8889, "step": 310 }, { "epoch": 0.862533692722372, "grad_norm": 2.4102346897125244, "learning_rate": 4.650205761316873e-05, "loss": 0.7576, "step": 320 }, { "epoch": 0.889487870619946, "grad_norm": 1.0781276226043701, "learning_rate": 4.636488340192044e-05, "loss": 1.167, "step": 330 }, { "epoch": 0.9164420485175202, "grad_norm": 1.4452636241912842, "learning_rate": 4.622770919067216e-05, "loss": 1.4974, "step": 340 }, { "epoch": 0.9433962264150944, "grad_norm": 0.7777872681617737, "learning_rate": 4.609053497942387e-05, "loss": 0.7603, "step": 350 }, { "epoch": 0.9703504043126685, "grad_norm": 1.9176746606826782, "learning_rate": 4.5953360768175586e-05, "loss": 1.2179, "step": 360 }, { "epoch": 0.9973045822102425, "grad_norm": 0.7132334113121033, "learning_rate": 4.58161865569273e-05, "loss": 1.3606, "step": 370 }, { "epoch": 1.0242587601078168, "grad_norm": 2.0499353408813477, "learning_rate": 4.567901234567901e-05, "loss": 1.5535, "step": 380 }, { "epoch": 1.0512129380053907, "grad_norm": 1.237725019454956, "learning_rate": 4.554183813443073e-05, "loss": 1.2777, "step": 390 }, { "epoch": 1.0781671159029649, "grad_norm": 0.8941486477851868, "learning_rate": 4.5404663923182444e-05, "loss": 1.098, "step": 400 }, { "epoch": 1.105121293800539, "grad_norm": 3.052698850631714, "learning_rate": 4.5267489711934157e-05, "loss": 0.8203, "step": 410 }, { "epoch": 1.1320754716981132, "grad_norm": 1.9386261701583862, "learning_rate": 4.513031550068587e-05, "loss": 1.3465, "step": 420 }, { "epoch": 1.1590296495956873, "grad_norm": 1.5312304496765137, "learning_rate": 4.499314128943759e-05, "loss": 1.5988, "step": 430 }, { "epoch": 1.1859838274932615, "grad_norm": 1.5553970336914062, "learning_rate": 4.48559670781893e-05, "loss": 1.3158, "step": 440 }, { "epoch": 1.2129380053908356, "grad_norm": 0.9215840101242065, "learning_rate": 4.4718792866941014e-05, "loss": 1.1357, "step": 450 }, { "epoch": 1.2398921832884098, "grad_norm": 0.8981826901435852, "learning_rate": 4.4581618655692734e-05, "loss": 1.4697, "step": 460 }, { "epoch": 1.266846361185984, "grad_norm": 1.3505191802978516, "learning_rate": 4.4444444444444447e-05, "loss": 1.3321, "step": 470 }, { "epoch": 1.2938005390835579, "grad_norm": 1.836127758026123, "learning_rate": 4.4307270233196166e-05, "loss": 0.7458, "step": 480 }, { "epoch": 1.320754716981132, "grad_norm": 1.3519006967544556, "learning_rate": 4.417009602194788e-05, "loss": 0.9444, "step": 490 }, { "epoch": 1.3477088948787062, "grad_norm": 0.8336719870567322, "learning_rate": 4.403292181069959e-05, "loss": 1.2264, "step": 500 }, { "epoch": 1.3746630727762803, "grad_norm": 2.3419370651245117, "learning_rate": 4.3895747599451304e-05, "loss": 1.4784, "step": 510 }, { "epoch": 1.4016172506738545, "grad_norm": 1.679734468460083, "learning_rate": 4.3758573388203024e-05, "loss": 0.9167, "step": 520 }, { "epoch": 1.4285714285714286, "grad_norm": 0.7462900876998901, "learning_rate": 4.3621399176954737e-05, "loss": 0.8118, "step": 530 }, { "epoch": 1.4555256064690028, "grad_norm": 1.11771559715271, "learning_rate": 4.348422496570645e-05, "loss": 1.2103, "step": 540 }, { "epoch": 1.482479784366577, "grad_norm": 1.0246716737747192, "learning_rate": 4.334705075445816e-05, "loss": 0.9964, "step": 550 }, { "epoch": 1.509433962264151, "grad_norm": 1.9072849750518799, "learning_rate": 4.3209876543209875e-05, "loss": 0.9468, "step": 560 }, { "epoch": 1.536388140161725, "grad_norm": 2.465437889099121, "learning_rate": 4.3072702331961594e-05, "loss": 1.0798, "step": 570 }, { "epoch": 1.5633423180592994, "grad_norm": 2.021115779876709, "learning_rate": 4.293552812071331e-05, "loss": 1.5825, "step": 580 }, { "epoch": 1.5902964959568733, "grad_norm": 1.403136968612671, "learning_rate": 4.279835390946502e-05, "loss": 1.0247, "step": 590 }, { "epoch": 1.6172506738544474, "grad_norm": 2.3880553245544434, "learning_rate": 4.266117969821673e-05, "loss": 0.7598, "step": 600 }, { "epoch": 1.6442048517520216, "grad_norm": 1.278221607208252, "learning_rate": 4.252400548696845e-05, "loss": 1.2538, "step": 610 }, { "epoch": 1.6711590296495957, "grad_norm": 1.4277830123901367, "learning_rate": 4.2386831275720165e-05, "loss": 1.0456, "step": 620 }, { "epoch": 1.6981132075471699, "grad_norm": 1.370888113975525, "learning_rate": 4.2249657064471884e-05, "loss": 1.1437, "step": 630 }, { "epoch": 1.7250673854447438, "grad_norm": 1.2015599012374878, "learning_rate": 4.21124828532236e-05, "loss": 1.309, "step": 640 }, { "epoch": 1.7520215633423182, "grad_norm": 0.6258445978164673, "learning_rate": 4.197530864197531e-05, "loss": 0.7159, "step": 650 }, { "epoch": 1.778975741239892, "grad_norm": 1.4056499004364014, "learning_rate": 4.183813443072703e-05, "loss": 0.9217, "step": 660 }, { "epoch": 1.8059299191374663, "grad_norm": 2.2113542556762695, "learning_rate": 4.170096021947874e-05, "loss": 0.9003, "step": 670 }, { "epoch": 1.8328840970350404, "grad_norm": 1.0116914510726929, "learning_rate": 4.1563786008230455e-05, "loss": 1.0246, "step": 680 }, { "epoch": 1.8598382749326146, "grad_norm": 1.2756783962249756, "learning_rate": 4.142661179698217e-05, "loss": 1.2909, "step": 690 }, { "epoch": 1.8867924528301887, "grad_norm": 1.2654672861099243, "learning_rate": 4.128943758573389e-05, "loss": 1.2136, "step": 700 }, { "epoch": 1.9137466307277629, "grad_norm": 0.7074457406997681, "learning_rate": 4.11522633744856e-05, "loss": 1.4202, "step": 710 }, { "epoch": 1.940700808625337, "grad_norm": 2.5393640995025635, "learning_rate": 4.101508916323731e-05, "loss": 0.959, "step": 720 }, { "epoch": 1.967654986522911, "grad_norm": 1.194204330444336, "learning_rate": 4.0877914951989025e-05, "loss": 0.7979, "step": 730 }, { "epoch": 1.9946091644204853, "grad_norm": 2.2830913066864014, "learning_rate": 4.074074074074074e-05, "loss": 0.9024, "step": 740 }, { "epoch": 2.0215633423180592, "grad_norm": 1.4763576984405518, "learning_rate": 4.060356652949246e-05, "loss": 0.8169, "step": 750 }, { "epoch": 2.0485175202156336, "grad_norm": 2.3150315284729004, "learning_rate": 4.046639231824417e-05, "loss": 1.1729, "step": 760 }, { "epoch": 2.0754716981132075, "grad_norm": 1.5322928428649902, "learning_rate": 4.032921810699588e-05, "loss": 0.8163, "step": 770 }, { "epoch": 2.1024258760107815, "grad_norm": 1.3889187574386597, "learning_rate": 4.01920438957476e-05, "loss": 1.1553, "step": 780 }, { "epoch": 2.129380053908356, "grad_norm": 1.2974004745483398, "learning_rate": 4.0054869684499315e-05, "loss": 1.5599, "step": 790 }, { "epoch": 2.1563342318059298, "grad_norm": 1.3070930242538452, "learning_rate": 3.9917695473251035e-05, "loss": 0.8604, "step": 800 }, { "epoch": 2.183288409703504, "grad_norm": 2.8242597579956055, "learning_rate": 3.978052126200275e-05, "loss": 1.1088, "step": 810 }, { "epoch": 2.210242587601078, "grad_norm": 1.1547167301177979, "learning_rate": 3.964334705075446e-05, "loss": 0.9212, "step": 820 }, { "epoch": 2.2371967654986524, "grad_norm": 1.5743651390075684, "learning_rate": 3.950617283950617e-05, "loss": 1.0367, "step": 830 }, { "epoch": 2.2641509433962264, "grad_norm": 1.0381134748458862, "learning_rate": 3.936899862825789e-05, "loss": 1.1694, "step": 840 }, { "epoch": 2.2911051212938007, "grad_norm": 1.475040316581726, "learning_rate": 3.9231824417009605e-05, "loss": 0.7031, "step": 850 }, { "epoch": 2.3180592991913747, "grad_norm": 1.6957072019577026, "learning_rate": 3.909465020576132e-05, "loss": 0.6501, "step": 860 }, { "epoch": 2.3450134770889486, "grad_norm": 1.2005444765090942, "learning_rate": 3.895747599451303e-05, "loss": 1.1483, "step": 870 }, { "epoch": 2.371967654986523, "grad_norm": 0.9146741628646851, "learning_rate": 3.8820301783264744e-05, "loss": 0.6192, "step": 880 }, { "epoch": 2.398921832884097, "grad_norm": 2.0229978561401367, "learning_rate": 3.868312757201646e-05, "loss": 1.0959, "step": 890 }, { "epoch": 2.4258760107816713, "grad_norm": 1.864313006401062, "learning_rate": 3.8545953360768176e-05, "loss": 1.1319, "step": 900 }, { "epoch": 2.452830188679245, "grad_norm": 2.050306558609009, "learning_rate": 3.840877914951989e-05, "loss": 0.8544, "step": 910 }, { "epoch": 2.4797843665768196, "grad_norm": 0.8470428586006165, "learning_rate": 3.82716049382716e-05, "loss": 0.7248, "step": 920 }, { "epoch": 2.5067385444743935, "grad_norm": 1.329987645149231, "learning_rate": 3.813443072702332e-05, "loss": 0.95, "step": 930 }, { "epoch": 2.533692722371968, "grad_norm": 1.2895963191986084, "learning_rate": 3.7997256515775034e-05, "loss": 0.8183, "step": 940 }, { "epoch": 2.560646900269542, "grad_norm": 1.2707000970840454, "learning_rate": 3.786008230452675e-05, "loss": 1.0259, "step": 950 }, { "epoch": 2.5876010781671157, "grad_norm": 1.5575352907180786, "learning_rate": 3.7722908093278466e-05, "loss": 0.8121, "step": 960 }, { "epoch": 2.61455525606469, "grad_norm": 1.4395464658737183, "learning_rate": 3.758573388203018e-05, "loss": 0.9093, "step": 970 }, { "epoch": 2.641509433962264, "grad_norm": 1.8523131608963013, "learning_rate": 3.74485596707819e-05, "loss": 1.4859, "step": 980 }, { "epoch": 2.6684636118598384, "grad_norm": 1.0986195802688599, "learning_rate": 3.731138545953361e-05, "loss": 0.9775, "step": 990 }, { "epoch": 2.6954177897574123, "grad_norm": 1.1339648962020874, "learning_rate": 3.7174211248285324e-05, "loss": 1.1358, "step": 1000 }, { "epoch": 2.7223719676549867, "grad_norm": 2.267329216003418, "learning_rate": 3.7037037037037037e-05, "loss": 0.8947, "step": 1010 }, { "epoch": 2.7493261455525606, "grad_norm": 1.501046061515808, "learning_rate": 3.6899862825788756e-05, "loss": 1.3137, "step": 1020 }, { "epoch": 2.776280323450135, "grad_norm": 3.182382345199585, "learning_rate": 3.676268861454047e-05, "loss": 1.4114, "step": 1030 }, { "epoch": 2.803234501347709, "grad_norm": 1.9981257915496826, "learning_rate": 3.662551440329218e-05, "loss": 1.3174, "step": 1040 }, { "epoch": 2.830188679245283, "grad_norm": 2.4771640300750732, "learning_rate": 3.6488340192043894e-05, "loss": 1.7184, "step": 1050 }, { "epoch": 2.857142857142857, "grad_norm": 1.307492971420288, "learning_rate": 3.635116598079561e-05, "loss": 1.0643, "step": 1060 }, { "epoch": 2.884097035040431, "grad_norm": 0.7365540862083435, "learning_rate": 3.6213991769547327e-05, "loss": 0.9895, "step": 1070 }, { "epoch": 2.9110512129380055, "grad_norm": 1.7630441188812256, "learning_rate": 3.607681755829904e-05, "loss": 1.0862, "step": 1080 }, { "epoch": 2.9380053908355794, "grad_norm": 1.731484055519104, "learning_rate": 3.593964334705075e-05, "loss": 0.7582, "step": 1090 }, { "epoch": 2.964959568733154, "grad_norm": 1.7666038274765015, "learning_rate": 3.580246913580247e-05, "loss": 0.9193, "step": 1100 }, { "epoch": 2.9919137466307277, "grad_norm": 1.3342130184173584, "learning_rate": 3.566529492455419e-05, "loss": 0.9391, "step": 1110 }, { "epoch": 3.018867924528302, "grad_norm": 1.7891361713409424, "learning_rate": 3.5528120713305904e-05, "loss": 0.4177, "step": 1120 }, { "epoch": 3.045822102425876, "grad_norm": 1.730653166770935, "learning_rate": 3.539094650205762e-05, "loss": 1.4771, "step": 1130 }, { "epoch": 3.07277628032345, "grad_norm": 1.5244202613830566, "learning_rate": 3.525377229080933e-05, "loss": 0.6919, "step": 1140 }, { "epoch": 3.0997304582210243, "grad_norm": 1.608145833015442, "learning_rate": 3.511659807956104e-05, "loss": 1.0934, "step": 1150 }, { "epoch": 3.1266846361185983, "grad_norm": 2.0203192234039307, "learning_rate": 3.497942386831276e-05, "loss": 0.8308, "step": 1160 }, { "epoch": 3.1536388140161726, "grad_norm": 1.7629802227020264, "learning_rate": 3.4842249657064474e-05, "loss": 0.8283, "step": 1170 }, { "epoch": 3.1805929919137466, "grad_norm": 1.6114568710327148, "learning_rate": 3.470507544581619e-05, "loss": 1.1672, "step": 1180 }, { "epoch": 3.207547169811321, "grad_norm": 1.5010507106781006, "learning_rate": 3.45679012345679e-05, "loss": 1.5866, "step": 1190 }, { "epoch": 3.234501347708895, "grad_norm": 1.4426625967025757, "learning_rate": 3.443072702331962e-05, "loss": 0.8659, "step": 1200 }, { "epoch": 3.2614555256064692, "grad_norm": 1.3204811811447144, "learning_rate": 3.429355281207133e-05, "loss": 0.643, "step": 1210 }, { "epoch": 3.288409703504043, "grad_norm": 0.9203002452850342, "learning_rate": 3.4156378600823045e-05, "loss": 0.6825, "step": 1220 }, { "epoch": 3.315363881401617, "grad_norm": 1.3211963176727295, "learning_rate": 3.401920438957476e-05, "loss": 1.1164, "step": 1230 }, { "epoch": 3.3423180592991915, "grad_norm": 1.4665788412094116, "learning_rate": 3.388203017832647e-05, "loss": 0.7564, "step": 1240 }, { "epoch": 3.3692722371967654, "grad_norm": 1.3217601776123047, "learning_rate": 3.374485596707819e-05, "loss": 0.8292, "step": 1250 }, { "epoch": 3.3962264150943398, "grad_norm": 1.1328483819961548, "learning_rate": 3.360768175582991e-05, "loss": 0.9806, "step": 1260 }, { "epoch": 3.4231805929919137, "grad_norm": 1.3456640243530273, "learning_rate": 3.347050754458162e-05, "loss": 0.8297, "step": 1270 }, { "epoch": 3.450134770889488, "grad_norm": 0.9857800006866455, "learning_rate": 3.3333333333333335e-05, "loss": 0.6432, "step": 1280 }, { "epoch": 3.477088948787062, "grad_norm": 1.5837385654449463, "learning_rate": 3.3196159122085054e-05, "loss": 0.6766, "step": 1290 }, { "epoch": 3.5040431266846364, "grad_norm": 0.9640551805496216, "learning_rate": 3.305898491083677e-05, "loss": 0.9636, "step": 1300 }, { "epoch": 3.5309973045822103, "grad_norm": 1.6151617765426636, "learning_rate": 3.292181069958848e-05, "loss": 0.5541, "step": 1310 }, { "epoch": 3.557951482479784, "grad_norm": 1.0056880712509155, "learning_rate": 3.278463648834019e-05, "loss": 0.8266, "step": 1320 }, { "epoch": 3.5849056603773586, "grad_norm": 2.169956684112549, "learning_rate": 3.2647462277091905e-05, "loss": 1.4843, "step": 1330 }, { "epoch": 3.6118598382749325, "grad_norm": 1.395251989364624, "learning_rate": 3.2510288065843625e-05, "loss": 0.4211, "step": 1340 }, { "epoch": 3.638814016172507, "grad_norm": 2.2781965732574463, "learning_rate": 3.237311385459534e-05, "loss": 0.799, "step": 1350 }, { "epoch": 3.665768194070081, "grad_norm": 2.2836267948150635, "learning_rate": 3.223593964334705e-05, "loss": 0.7618, "step": 1360 }, { "epoch": 3.6927223719676547, "grad_norm": 2.4063193798065186, "learning_rate": 3.209876543209876e-05, "loss": 1.3887, "step": 1370 }, { "epoch": 3.719676549865229, "grad_norm": 1.7423830032348633, "learning_rate": 3.196159122085048e-05, "loss": 0.8476, "step": 1380 }, { "epoch": 3.7466307277628035, "grad_norm": 2.047558307647705, "learning_rate": 3.1824417009602195e-05, "loss": 1.0701, "step": 1390 }, { "epoch": 3.7735849056603774, "grad_norm": 1.421557068824768, "learning_rate": 3.168724279835391e-05, "loss": 1.6139, "step": 1400 }, { "epoch": 3.8005390835579513, "grad_norm": 1.6829692125320435, "learning_rate": 3.155006858710563e-05, "loss": 0.8253, "step": 1410 }, { "epoch": 3.8274932614555257, "grad_norm": 0.9380494356155396, "learning_rate": 3.141289437585734e-05, "loss": 1.0122, "step": 1420 }, { "epoch": 3.8544474393530996, "grad_norm": 1.209517478942871, "learning_rate": 3.127572016460906e-05, "loss": 1.6921, "step": 1430 }, { "epoch": 3.881401617250674, "grad_norm": 1.2071969509124756, "learning_rate": 3.113854595336077e-05, "loss": 1.3358, "step": 1440 }, { "epoch": 3.908355795148248, "grad_norm": 1.6536906957626343, "learning_rate": 3.1001371742112486e-05, "loss": 0.7539, "step": 1450 }, { "epoch": 3.935309973045822, "grad_norm": 3.013587713241577, "learning_rate": 3.08641975308642e-05, "loss": 1.5292, "step": 1460 }, { "epoch": 3.9622641509433962, "grad_norm": 1.5310529470443726, "learning_rate": 3.072702331961592e-05, "loss": 0.8488, "step": 1470 }, { "epoch": 3.9892183288409706, "grad_norm": 1.6104665994644165, "learning_rate": 3.058984910836763e-05, "loss": 0.5511, "step": 1480 }, { "epoch": 4.0161725067385445, "grad_norm": 1.7617496252059937, "learning_rate": 3.0452674897119343e-05, "loss": 1.5566, "step": 1490 }, { "epoch": 4.0431266846361185, "grad_norm": 1.6060303449630737, "learning_rate": 3.0315500685871056e-05, "loss": 0.6705, "step": 1500 }, { "epoch": 4.070080862533692, "grad_norm": 1.2487995624542236, "learning_rate": 3.017832647462277e-05, "loss": 0.8386, "step": 1510 }, { "epoch": 4.097035040431267, "grad_norm": 1.5747429132461548, "learning_rate": 3.0041152263374488e-05, "loss": 1.5059, "step": 1520 }, { "epoch": 4.123989218328841, "grad_norm": 2.865582227706909, "learning_rate": 2.99039780521262e-05, "loss": 0.801, "step": 1530 }, { "epoch": 4.150943396226415, "grad_norm": 2.48677659034729, "learning_rate": 2.9766803840877917e-05, "loss": 1.1271, "step": 1540 }, { "epoch": 4.177897574123989, "grad_norm": 2.485933303833008, "learning_rate": 2.962962962962963e-05, "loss": 1.0503, "step": 1550 }, { "epoch": 4.204851752021563, "grad_norm": 2.0622434616088867, "learning_rate": 2.949245541838135e-05, "loss": 0.5068, "step": 1560 }, { "epoch": 4.231805929919138, "grad_norm": 0.7956051826477051, "learning_rate": 2.9355281207133062e-05, "loss": 1.0424, "step": 1570 }, { "epoch": 4.258760107816712, "grad_norm": 1.3042256832122803, "learning_rate": 2.9218106995884775e-05, "loss": 0.5118, "step": 1580 }, { "epoch": 4.285714285714286, "grad_norm": 2.2554943561553955, "learning_rate": 2.9080932784636488e-05, "loss": 0.9746, "step": 1590 }, { "epoch": 4.3126684636118595, "grad_norm": 2.549921751022339, "learning_rate": 2.8943758573388204e-05, "loss": 0.7228, "step": 1600 }, { "epoch": 4.339622641509434, "grad_norm": 1.333503246307373, "learning_rate": 2.880658436213992e-05, "loss": 0.9424, "step": 1610 }, { "epoch": 4.366576819407008, "grad_norm": 1.5245575904846191, "learning_rate": 2.8669410150891636e-05, "loss": 1.0232, "step": 1620 }, { "epoch": 4.393530997304582, "grad_norm": 5.889776229858398, "learning_rate": 2.853223593964335e-05, "loss": 0.9108, "step": 1630 }, { "epoch": 4.420485175202156, "grad_norm": 1.444460153579712, "learning_rate": 2.839506172839506e-05, "loss": 0.7664, "step": 1640 }, { "epoch": 4.44743935309973, "grad_norm": 2.633338689804077, "learning_rate": 2.825788751714678e-05, "loss": 0.9278, "step": 1650 }, { "epoch": 4.474393530997305, "grad_norm": 0.8140472769737244, "learning_rate": 2.8120713305898494e-05, "loss": 0.7608, "step": 1660 }, { "epoch": 4.501347708894879, "grad_norm": 1.0764524936676025, "learning_rate": 2.7983539094650207e-05, "loss": 0.8768, "step": 1670 }, { "epoch": 4.528301886792453, "grad_norm": 1.2656837701797485, "learning_rate": 2.784636488340192e-05, "loss": 1.2142, "step": 1680 }, { "epoch": 4.555256064690027, "grad_norm": 1.1997019052505493, "learning_rate": 2.7709190672153635e-05, "loss": 1.1219, "step": 1690 }, { "epoch": 4.5822102425876015, "grad_norm": 1.1328818798065186, "learning_rate": 2.757201646090535e-05, "loss": 0.7957, "step": 1700 }, { "epoch": 4.609164420485175, "grad_norm": 0.9673919677734375, "learning_rate": 2.7434842249657068e-05, "loss": 0.833, "step": 1710 }, { "epoch": 4.636118598382749, "grad_norm": 1.556420087814331, "learning_rate": 2.729766803840878e-05, "loss": 0.644, "step": 1720 }, { "epoch": 4.663072776280323, "grad_norm": 1.8422091007232666, "learning_rate": 2.7160493827160493e-05, "loss": 1.0444, "step": 1730 }, { "epoch": 4.690026954177897, "grad_norm": 1.5853588581085205, "learning_rate": 2.7023319615912206e-05, "loss": 0.8302, "step": 1740 }, { "epoch": 4.716981132075472, "grad_norm": 2.221689462661743, "learning_rate": 2.6886145404663926e-05, "loss": 0.8505, "step": 1750 }, { "epoch": 4.743935309973046, "grad_norm": 1.9763739109039307, "learning_rate": 2.6748971193415638e-05, "loss": 0.7795, "step": 1760 }, { "epoch": 4.77088948787062, "grad_norm": 2.094904661178589, "learning_rate": 2.6611796982167354e-05, "loss": 0.6406, "step": 1770 }, { "epoch": 4.797843665768194, "grad_norm": 2.045687437057495, "learning_rate": 2.6474622770919067e-05, "loss": 0.7323, "step": 1780 }, { "epoch": 4.824797843665769, "grad_norm": 2.2809979915618896, "learning_rate": 2.6337448559670787e-05, "loss": 0.783, "step": 1790 }, { "epoch": 4.8517520215633425, "grad_norm": 1.7300888299942017, "learning_rate": 2.62002743484225e-05, "loss": 0.6659, "step": 1800 }, { "epoch": 4.878706199460916, "grad_norm": 3.9680426120758057, "learning_rate": 2.6063100137174212e-05, "loss": 0.6866, "step": 1810 }, { "epoch": 4.90566037735849, "grad_norm": 1.8751779794692993, "learning_rate": 2.5925925925925925e-05, "loss": 0.9165, "step": 1820 }, { "epoch": 4.932614555256064, "grad_norm": 1.671337366104126, "learning_rate": 2.5788751714677638e-05, "loss": 0.846, "step": 1830 }, { "epoch": 4.959568733153639, "grad_norm": 2.150132179260254, "learning_rate": 2.5651577503429357e-05, "loss": 1.2381, "step": 1840 }, { "epoch": 4.986522911051213, "grad_norm": 1.2727786302566528, "learning_rate": 2.551440329218107e-05, "loss": 0.7399, "step": 1850 }, { "epoch": 5.013477088948787, "grad_norm": 1.6058365106582642, "learning_rate": 2.5377229080932786e-05, "loss": 1.3821, "step": 1860 }, { "epoch": 5.040431266846361, "grad_norm": 0.873196542263031, "learning_rate": 2.52400548696845e-05, "loss": 1.0094, "step": 1870 }, { "epoch": 5.067385444743936, "grad_norm": 2.6777074337005615, "learning_rate": 2.510288065843622e-05, "loss": 0.956, "step": 1880 }, { "epoch": 5.09433962264151, "grad_norm": 2.2035608291625977, "learning_rate": 2.496570644718793e-05, "loss": 1.0327, "step": 1890 }, { "epoch": 5.121293800539084, "grad_norm": 1.9999809265136719, "learning_rate": 2.4828532235939644e-05, "loss": 0.9534, "step": 1900 }, { "epoch": 5.1482479784366575, "grad_norm": 1.1194771528244019, "learning_rate": 2.4691358024691357e-05, "loss": 0.6969, "step": 1910 }, { "epoch": 5.175202156334231, "grad_norm": 2.1058928966522217, "learning_rate": 2.4554183813443073e-05, "loss": 1.077, "step": 1920 }, { "epoch": 5.202156334231806, "grad_norm": 1.584088921546936, "learning_rate": 2.441700960219479e-05, "loss": 1.0808, "step": 1930 }, { "epoch": 5.22911051212938, "grad_norm": 2.4530153274536133, "learning_rate": 2.4279835390946505e-05, "loss": 1.1974, "step": 1940 }, { "epoch": 5.256064690026954, "grad_norm": 1.3323501348495483, "learning_rate": 2.4142661179698218e-05, "loss": 1.0003, "step": 1950 }, { "epoch": 5.283018867924528, "grad_norm": 2.1293039321899414, "learning_rate": 2.4005486968449934e-05, "loss": 1.2501, "step": 1960 }, { "epoch": 5.309973045822103, "grad_norm": 2.1004316806793213, "learning_rate": 2.3868312757201647e-05, "loss": 1.0852, "step": 1970 }, { "epoch": 5.336927223719677, "grad_norm": 2.6623051166534424, "learning_rate": 2.3731138545953363e-05, "loss": 0.6813, "step": 1980 }, { "epoch": 5.363881401617251, "grad_norm": 1.0680763721466064, "learning_rate": 2.3593964334705075e-05, "loss": 0.6413, "step": 1990 }, { "epoch": 5.390835579514825, "grad_norm": 1.4404453039169312, "learning_rate": 2.345679012345679e-05, "loss": 0.8624, "step": 2000 }, { "epoch": 5.4177897574123985, "grad_norm": 2.0535359382629395, "learning_rate": 2.3319615912208508e-05, "loss": 0.8049, "step": 2010 }, { "epoch": 5.444743935309973, "grad_norm": 1.0573314428329468, "learning_rate": 2.318244170096022e-05, "loss": 0.8271, "step": 2020 }, { "epoch": 5.471698113207547, "grad_norm": 3.0816051959991455, "learning_rate": 2.3045267489711937e-05, "loss": 1.2149, "step": 2030 }, { "epoch": 5.498652291105121, "grad_norm": 0.8514117002487183, "learning_rate": 2.290809327846365e-05, "loss": 0.6105, "step": 2040 }, { "epoch": 5.525606469002695, "grad_norm": 2.2031140327453613, "learning_rate": 2.2770919067215366e-05, "loss": 0.7386, "step": 2050 }, { "epoch": 5.55256064690027, "grad_norm": 0.6656814217567444, "learning_rate": 2.2633744855967078e-05, "loss": 0.64, "step": 2060 }, { "epoch": 5.579514824797844, "grad_norm": 1.3257490396499634, "learning_rate": 2.2496570644718794e-05, "loss": 0.5246, "step": 2070 }, { "epoch": 5.606469002695418, "grad_norm": 1.9381606578826904, "learning_rate": 2.2359396433470507e-05, "loss": 0.8002, "step": 2080 }, { "epoch": 5.633423180592992, "grad_norm": 1.5016483068466187, "learning_rate": 2.2222222222222223e-05, "loss": 0.7496, "step": 2090 }, { "epoch": 5.660377358490566, "grad_norm": 1.4559197425842285, "learning_rate": 2.208504801097394e-05, "loss": 0.4904, "step": 2100 }, { "epoch": 5.6873315363881405, "grad_norm": 1.7636557817459106, "learning_rate": 2.1947873799725652e-05, "loss": 0.6664, "step": 2110 }, { "epoch": 5.714285714285714, "grad_norm": 1.5143455266952515, "learning_rate": 2.1810699588477368e-05, "loss": 0.9208, "step": 2120 }, { "epoch": 5.741239892183288, "grad_norm": 2.0838751792907715, "learning_rate": 2.167352537722908e-05, "loss": 0.3976, "step": 2130 }, { "epoch": 5.768194070080862, "grad_norm": 1.7315948009490967, "learning_rate": 2.1536351165980797e-05, "loss": 0.6964, "step": 2140 }, { "epoch": 5.795148247978437, "grad_norm": 2.732347249984741, "learning_rate": 2.139917695473251e-05, "loss": 1.4496, "step": 2150 }, { "epoch": 5.822102425876011, "grad_norm": 1.661921501159668, "learning_rate": 2.1262002743484226e-05, "loss": 0.8408, "step": 2160 }, { "epoch": 5.849056603773585, "grad_norm": 1.123995065689087, "learning_rate": 2.1124828532235942e-05, "loss": 0.5594, "step": 2170 }, { "epoch": 5.876010781671159, "grad_norm": 1.3693888187408447, "learning_rate": 2.0987654320987655e-05, "loss": 1.0366, "step": 2180 }, { "epoch": 5.902964959568733, "grad_norm": 1.589400053024292, "learning_rate": 2.085048010973937e-05, "loss": 1.1677, "step": 2190 }, { "epoch": 5.929919137466308, "grad_norm": 1.3596928119659424, "learning_rate": 2.0713305898491084e-05, "loss": 0.6217, "step": 2200 }, { "epoch": 5.9568733153638815, "grad_norm": 1.5818856954574585, "learning_rate": 2.05761316872428e-05, "loss": 1.0682, "step": 2210 }, { "epoch": 5.9838274932614555, "grad_norm": 2.694117546081543, "learning_rate": 2.0438957475994513e-05, "loss": 0.9281, "step": 2220 }, { "epoch": 6.010781671159029, "grad_norm": 1.463112473487854, "learning_rate": 2.030178326474623e-05, "loss": 0.7868, "step": 2230 }, { "epoch": 6.037735849056604, "grad_norm": 1.8323993682861328, "learning_rate": 2.016460905349794e-05, "loss": 0.6472, "step": 2240 }, { "epoch": 6.064690026954178, "grad_norm": 2.3480756282806396, "learning_rate": 2.0027434842249658e-05, "loss": 0.6044, "step": 2250 }, { "epoch": 6.091644204851752, "grad_norm": 3.0445172786712646, "learning_rate": 1.9890260631001374e-05, "loss": 0.5516, "step": 2260 }, { "epoch": 6.118598382749326, "grad_norm": 2.489713430404663, "learning_rate": 1.9753086419753087e-05, "loss": 0.5861, "step": 2270 }, { "epoch": 6.1455525606469, "grad_norm": 2.1993355751037598, "learning_rate": 1.9615912208504803e-05, "loss": 0.6938, "step": 2280 }, { "epoch": 6.172506738544475, "grad_norm": 1.914015769958496, "learning_rate": 1.9478737997256515e-05, "loss": 0.8575, "step": 2290 }, { "epoch": 6.199460916442049, "grad_norm": 2.359715223312378, "learning_rate": 1.934156378600823e-05, "loss": 0.6444, "step": 2300 }, { "epoch": 6.226415094339623, "grad_norm": 0.865803062915802, "learning_rate": 1.9204389574759944e-05, "loss": 0.6543, "step": 2310 }, { "epoch": 6.2533692722371965, "grad_norm": 1.363655686378479, "learning_rate": 1.906721536351166e-05, "loss": 0.9914, "step": 2320 }, { "epoch": 6.280323450134771, "grad_norm": 2.174215316772461, "learning_rate": 1.8930041152263377e-05, "loss": 0.6777, "step": 2330 }, { "epoch": 6.307277628032345, "grad_norm": 1.1655148267745972, "learning_rate": 1.879286694101509e-05, "loss": 0.6198, "step": 2340 }, { "epoch": 6.334231805929919, "grad_norm": 2.751349449157715, "learning_rate": 1.8655692729766806e-05, "loss": 1.3854, "step": 2350 }, { "epoch": 6.361185983827493, "grad_norm": 1.7437028884887695, "learning_rate": 1.8518518518518518e-05, "loss": 0.6498, "step": 2360 }, { "epoch": 6.388140161725067, "grad_norm": 2.6876518726348877, "learning_rate": 1.8381344307270234e-05, "loss": 0.8044, "step": 2370 }, { "epoch": 6.415094339622642, "grad_norm": 2.60074520111084, "learning_rate": 1.8244170096021947e-05, "loss": 1.155, "step": 2380 }, { "epoch": 6.442048517520216, "grad_norm": 2.644638776779175, "learning_rate": 1.8106995884773663e-05, "loss": 0.8515, "step": 2390 }, { "epoch": 6.46900269541779, "grad_norm": 3.353593587875366, "learning_rate": 1.7969821673525376e-05, "loss": 0.5867, "step": 2400 }, { "epoch": 6.495956873315364, "grad_norm": 1.9784125089645386, "learning_rate": 1.7832647462277096e-05, "loss": 0.6024, "step": 2410 }, { "epoch": 6.5229110512129385, "grad_norm": 1.7953509092330933, "learning_rate": 1.769547325102881e-05, "loss": 1.2158, "step": 2420 }, { "epoch": 6.549865229110512, "grad_norm": 2.232206344604492, "learning_rate": 1.755829903978052e-05, "loss": 0.8265, "step": 2430 }, { "epoch": 6.576819407008086, "grad_norm": 1.054999828338623, "learning_rate": 1.7421124828532237e-05, "loss": 0.7372, "step": 2440 }, { "epoch": 6.60377358490566, "grad_norm": 1.8623944520950317, "learning_rate": 1.728395061728395e-05, "loss": 1.4106, "step": 2450 }, { "epoch": 6.630727762803234, "grad_norm": 2.689140796661377, "learning_rate": 1.7146776406035666e-05, "loss": 0.9934, "step": 2460 }, { "epoch": 6.657681940700809, "grad_norm": 1.7552543878555298, "learning_rate": 1.700960219478738e-05, "loss": 0.8231, "step": 2470 }, { "epoch": 6.684636118598383, "grad_norm": 3.4416518211364746, "learning_rate": 1.6872427983539095e-05, "loss": 0.7788, "step": 2480 }, { "epoch": 6.711590296495957, "grad_norm": 3.726334571838379, "learning_rate": 1.673525377229081e-05, "loss": 0.7407, "step": 2490 }, { "epoch": 6.738544474393531, "grad_norm": 2.932966470718384, "learning_rate": 1.6598079561042527e-05, "loss": 1.0143, "step": 2500 }, { "epoch": 6.765498652291106, "grad_norm": 1.8736401796340942, "learning_rate": 1.646090534979424e-05, "loss": 0.8365, "step": 2510 }, { "epoch": 6.7924528301886795, "grad_norm": 2.0926668643951416, "learning_rate": 1.6323731138545953e-05, "loss": 0.7736, "step": 2520 }, { "epoch": 6.819407008086253, "grad_norm": 3.3576009273529053, "learning_rate": 1.618655692729767e-05, "loss": 1.1465, "step": 2530 }, { "epoch": 6.846361185983827, "grad_norm": 2.4661567211151123, "learning_rate": 1.604938271604938e-05, "loss": 0.7863, "step": 2540 }, { "epoch": 6.873315363881401, "grad_norm": 1.88754141330719, "learning_rate": 1.5912208504801098e-05, "loss": 0.9081, "step": 2550 }, { "epoch": 6.900269541778976, "grad_norm": 47.142337799072266, "learning_rate": 1.5775034293552814e-05, "loss": 0.8261, "step": 2560 }, { "epoch": 6.92722371967655, "grad_norm": 2.473158359527588, "learning_rate": 1.563786008230453e-05, "loss": 0.4822, "step": 2570 }, { "epoch": 6.954177897574124, "grad_norm": 1.6194536685943604, "learning_rate": 1.5500685871056243e-05, "loss": 0.856, "step": 2580 }, { "epoch": 6.981132075471698, "grad_norm": 5.963684558868408, "learning_rate": 1.536351165980796e-05, "loss": 0.7661, "step": 2590 }, { "epoch": 7.008086253369272, "grad_norm": 1.8639130592346191, "learning_rate": 1.5226337448559672e-05, "loss": 1.388, "step": 2600 }, { "epoch": 7.035040431266847, "grad_norm": 3.447125196456909, "learning_rate": 1.5089163237311384e-05, "loss": 0.7474, "step": 2610 }, { "epoch": 7.061994609164421, "grad_norm": 2.3289992809295654, "learning_rate": 1.49519890260631e-05, "loss": 0.5907, "step": 2620 }, { "epoch": 7.0889487870619945, "grad_norm": 1.342872977256775, "learning_rate": 1.4814814814814815e-05, "loss": 0.8798, "step": 2630 }, { "epoch": 7.115902964959568, "grad_norm": 1.3832533359527588, "learning_rate": 1.4677640603566531e-05, "loss": 1.0841, "step": 2640 }, { "epoch": 7.142857142857143, "grad_norm": 1.0350087881088257, "learning_rate": 1.4540466392318244e-05, "loss": 1.2334, "step": 2650 }, { "epoch": 7.169811320754717, "grad_norm": 4.020228862762451, "learning_rate": 1.440329218106996e-05, "loss": 0.8947, "step": 2660 }, { "epoch": 7.196765498652291, "grad_norm": 2.982022523880005, "learning_rate": 1.4266117969821674e-05, "loss": 0.9667, "step": 2670 }, { "epoch": 7.223719676549865, "grad_norm": 2.171691417694092, "learning_rate": 1.412894375857339e-05, "loss": 0.7494, "step": 2680 }, { "epoch": 7.250673854447439, "grad_norm": 2.717907667160034, "learning_rate": 1.3991769547325103e-05, "loss": 0.6898, "step": 2690 }, { "epoch": 7.277628032345014, "grad_norm": 2.5439579486846924, "learning_rate": 1.3854595336076818e-05, "loss": 0.4311, "step": 2700 }, { "epoch": 7.304582210242588, "grad_norm": 0.7530654072761536, "learning_rate": 1.3717421124828534e-05, "loss": 0.9008, "step": 2710 }, { "epoch": 7.331536388140162, "grad_norm": 0.637667179107666, "learning_rate": 1.3580246913580247e-05, "loss": 1.087, "step": 2720 }, { "epoch": 7.3584905660377355, "grad_norm": 1.1592116355895996, "learning_rate": 1.3443072702331963e-05, "loss": 0.6196, "step": 2730 }, { "epoch": 7.38544474393531, "grad_norm": 2.054795503616333, "learning_rate": 1.3305898491083677e-05, "loss": 0.7915, "step": 2740 }, { "epoch": 7.412398921832884, "grad_norm": 1.4349403381347656, "learning_rate": 1.3168724279835393e-05, "loss": 0.6572, "step": 2750 }, { "epoch": 7.439353099730458, "grad_norm": 1.0293610095977783, "learning_rate": 1.3031550068587106e-05, "loss": 1.0812, "step": 2760 }, { "epoch": 7.466307277628032, "grad_norm": 2.4686598777770996, "learning_rate": 1.2894375857338819e-05, "loss": 0.9702, "step": 2770 }, { "epoch": 7.493261455525606, "grad_norm": 1.4416760206222534, "learning_rate": 1.2757201646090535e-05, "loss": 1.0412, "step": 2780 }, { "epoch": 7.520215633423181, "grad_norm": 3.614410400390625, "learning_rate": 1.262002743484225e-05, "loss": 1.2559, "step": 2790 }, { "epoch": 7.547169811320755, "grad_norm": 0.8973541855812073, "learning_rate": 1.2482853223593966e-05, "loss": 0.4049, "step": 2800 }, { "epoch": 7.574123989218329, "grad_norm": 1.4669396877288818, "learning_rate": 1.2345679012345678e-05, "loss": 0.764, "step": 2810 }, { "epoch": 7.601078167115903, "grad_norm": 2.6035633087158203, "learning_rate": 1.2208504801097394e-05, "loss": 0.7421, "step": 2820 }, { "epoch": 7.628032345013477, "grad_norm": 3.8187856674194336, "learning_rate": 1.2071330589849109e-05, "loss": 0.8026, "step": 2830 }, { "epoch": 7.654986522911051, "grad_norm": 1.5178192853927612, "learning_rate": 1.1934156378600823e-05, "loss": 0.8348, "step": 2840 }, { "epoch": 7.681940700808625, "grad_norm": 1.0527846813201904, "learning_rate": 1.1796982167352538e-05, "loss": 0.7379, "step": 2850 }, { "epoch": 7.708894878706199, "grad_norm": 2.7532355785369873, "learning_rate": 1.1659807956104254e-05, "loss": 1.4819, "step": 2860 }, { "epoch": 7.735849056603773, "grad_norm": 1.9591217041015625, "learning_rate": 1.1522633744855968e-05, "loss": 0.4296, "step": 2870 }, { "epoch": 7.762803234501348, "grad_norm": 2.7292425632476807, "learning_rate": 1.1385459533607683e-05, "loss": 0.6324, "step": 2880 }, { "epoch": 7.789757412398922, "grad_norm": 2.3577399253845215, "learning_rate": 1.1248285322359397e-05, "loss": 0.5892, "step": 2890 }, { "epoch": 7.816711590296496, "grad_norm": 1.4059489965438843, "learning_rate": 1.1111111111111112e-05, "loss": 0.5519, "step": 2900 }, { "epoch": 7.84366576819407, "grad_norm": 1.9480534791946411, "learning_rate": 1.0973936899862826e-05, "loss": 0.9436, "step": 2910 }, { "epoch": 7.870619946091644, "grad_norm": 2.3746042251586914, "learning_rate": 1.083676268861454e-05, "loss": 0.3955, "step": 2920 }, { "epoch": 7.8975741239892185, "grad_norm": 1.2576045989990234, "learning_rate": 1.0699588477366255e-05, "loss": 0.604, "step": 2930 }, { "epoch": 7.9245283018867925, "grad_norm": 2.509427309036255, "learning_rate": 1.0562414266117971e-05, "loss": 0.7358, "step": 2940 }, { "epoch": 7.951482479784366, "grad_norm": 1.8080178499221802, "learning_rate": 1.0425240054869686e-05, "loss": 0.6726, "step": 2950 }, { "epoch": 7.97843665768194, "grad_norm": 3.254493474960327, "learning_rate": 1.02880658436214e-05, "loss": 0.6716, "step": 2960 }, { "epoch": 8.005390835579515, "grad_norm": 1.1409560441970825, "learning_rate": 1.0150891632373114e-05, "loss": 0.8359, "step": 2970 }, { "epoch": 8.032345013477089, "grad_norm": 2.8074634075164795, "learning_rate": 1.0013717421124829e-05, "loss": 0.8532, "step": 2980 }, { "epoch": 8.059299191374663, "grad_norm": 1.6796783208847046, "learning_rate": 9.876543209876543e-06, "loss": 1.1366, "step": 2990 }, { "epoch": 8.086253369272237, "grad_norm": 0.6373213529586792, "learning_rate": 9.739368998628258e-06, "loss": 0.5121, "step": 3000 }, { "epoch": 8.11320754716981, "grad_norm": 1.404941201210022, "learning_rate": 9.602194787379972e-06, "loss": 0.7168, "step": 3010 }, { "epoch": 8.140161725067385, "grad_norm": 3.0173532962799072, "learning_rate": 9.465020576131688e-06, "loss": 0.6495, "step": 3020 }, { "epoch": 8.167115902964959, "grad_norm": 0.9373369812965393, "learning_rate": 9.327846364883403e-06, "loss": 1.1245, "step": 3030 }, { "epoch": 8.194070080862534, "grad_norm": 1.8687936067581177, "learning_rate": 9.190672153635117e-06, "loss": 0.7253, "step": 3040 }, { "epoch": 8.221024258760108, "grad_norm": 0.9592246413230896, "learning_rate": 9.053497942386832e-06, "loss": 1.2584, "step": 3050 }, { "epoch": 8.247978436657682, "grad_norm": 2.668527126312256, "learning_rate": 8.916323731138548e-06, "loss": 0.6645, "step": 3060 }, { "epoch": 8.274932614555256, "grad_norm": 0.8130900859832764, "learning_rate": 8.77914951989026e-06, "loss": 0.5244, "step": 3070 }, { "epoch": 8.30188679245283, "grad_norm": 1.980900764465332, "learning_rate": 8.641975308641975e-06, "loss": 0.5843, "step": 3080 }, { "epoch": 8.328840970350404, "grad_norm": 1.7426378726959229, "learning_rate": 8.50480109739369e-06, "loss": 0.9521, "step": 3090 }, { "epoch": 8.355795148247978, "grad_norm": 2.6671996116638184, "learning_rate": 8.367626886145406e-06, "loss": 0.8797, "step": 3100 }, { "epoch": 8.382749326145552, "grad_norm": 2.629798650741577, "learning_rate": 8.23045267489712e-06, "loss": 0.5629, "step": 3110 }, { "epoch": 8.409703504043126, "grad_norm": 1.723059058189392, "learning_rate": 8.093278463648834e-06, "loss": 0.5738, "step": 3120 }, { "epoch": 8.436657681940702, "grad_norm": 0.5404053926467896, "learning_rate": 7.956104252400549e-06, "loss": 0.72, "step": 3130 }, { "epoch": 8.463611859838275, "grad_norm": 2.8987836837768555, "learning_rate": 7.818930041152265e-06, "loss": 0.7432, "step": 3140 }, { "epoch": 8.49056603773585, "grad_norm": 1.2646089792251587, "learning_rate": 7.68175582990398e-06, "loss": 0.5356, "step": 3150 }, { "epoch": 8.517520215633423, "grad_norm": 3.3997156620025635, "learning_rate": 7.544581618655692e-06, "loss": 0.8056, "step": 3160 }, { "epoch": 8.544474393530997, "grad_norm": 2.264604330062866, "learning_rate": 7.4074074074074075e-06, "loss": 1.0019, "step": 3170 }, { "epoch": 8.571428571428571, "grad_norm": 1.962279200553894, "learning_rate": 7.270233196159122e-06, "loss": 0.6352, "step": 3180 }, { "epoch": 8.598382749326145, "grad_norm": 2.4050302505493164, "learning_rate": 7.133058984910837e-06, "loss": 0.6712, "step": 3190 }, { "epoch": 8.625336927223719, "grad_norm": 1.2828004360198975, "learning_rate": 6.995884773662552e-06, "loss": 0.625, "step": 3200 }, { "epoch": 8.652291105121293, "grad_norm": 3.306525230407715, "learning_rate": 6.858710562414267e-06, "loss": 0.7727, "step": 3210 }, { "epoch": 8.679245283018869, "grad_norm": 1.3902239799499512, "learning_rate": 6.721536351165981e-06, "loss": 0.4857, "step": 3220 }, { "epoch": 8.706199460916443, "grad_norm": 3.665642261505127, "learning_rate": 6.584362139917697e-06, "loss": 0.9572, "step": 3230 }, { "epoch": 8.733153638814017, "grad_norm": 2.6966567039489746, "learning_rate": 6.447187928669409e-06, "loss": 0.739, "step": 3240 }, { "epoch": 8.76010781671159, "grad_norm": 2.127206325531006, "learning_rate": 6.310013717421125e-06, "loss": 0.8983, "step": 3250 }, { "epoch": 8.787061994609164, "grad_norm": 2.1118884086608887, "learning_rate": 6.172839506172839e-06, "loss": 1.0385, "step": 3260 }, { "epoch": 8.814016172506738, "grad_norm": 1.6121476888656616, "learning_rate": 6.0356652949245544e-06, "loss": 0.4564, "step": 3270 }, { "epoch": 8.840970350404312, "grad_norm": 3.346813440322876, "learning_rate": 5.898491083676269e-06, "loss": 0.9239, "step": 3280 }, { "epoch": 8.867924528301886, "grad_norm": 2.808685541152954, "learning_rate": 5.761316872427984e-06, "loss": 1.0246, "step": 3290 }, { "epoch": 8.89487870619946, "grad_norm": 0.6503840684890747, "learning_rate": 5.624142661179699e-06, "loss": 0.4066, "step": 3300 }, { "epoch": 8.921832884097036, "grad_norm": 2.2090868949890137, "learning_rate": 5.486968449931413e-06, "loss": 0.8856, "step": 3310 }, { "epoch": 8.94878706199461, "grad_norm": 2.4998106956481934, "learning_rate": 5.3497942386831275e-06, "loss": 1.0539, "step": 3320 }, { "epoch": 8.975741239892184, "grad_norm": 1.3785419464111328, "learning_rate": 5.212620027434843e-06, "loss": 1.0206, "step": 3330 }, { "epoch": 9.002695417789758, "grad_norm": 2.4248452186584473, "learning_rate": 5.075445816186557e-06, "loss": 0.6985, "step": 3340 }, { "epoch": 9.029649595687331, "grad_norm": 2.76381254196167, "learning_rate": 4.938271604938272e-06, "loss": 0.8552, "step": 3350 }, { "epoch": 9.056603773584905, "grad_norm": 0.44387710094451904, "learning_rate": 4.801097393689986e-06, "loss": 0.7421, "step": 3360 }, { "epoch": 9.08355795148248, "grad_norm": 1.8279584646224976, "learning_rate": 4.663923182441701e-06, "loss": 0.6471, "step": 3370 }, { "epoch": 9.110512129380053, "grad_norm": 4.515883922576904, "learning_rate": 4.526748971193416e-06, "loss": 0.8336, "step": 3380 }, { "epoch": 9.137466307277627, "grad_norm": 2.7579915523529053, "learning_rate": 4.38957475994513e-06, "loss": 1.0069, "step": 3390 }, { "epoch": 9.164420485175203, "grad_norm": 2.561863660812378, "learning_rate": 4.252400548696845e-06, "loss": 0.8613, "step": 3400 }, { "epoch": 9.191374663072777, "grad_norm": 3.5827927589416504, "learning_rate": 4.11522633744856e-06, "loss": 0.7177, "step": 3410 }, { "epoch": 9.21832884097035, "grad_norm": 1.109462022781372, "learning_rate": 3.9780521262002744e-06, "loss": 0.6787, "step": 3420 }, { "epoch": 9.245283018867925, "grad_norm": 2.3508827686309814, "learning_rate": 3.84087791495199e-06, "loss": 0.7854, "step": 3430 }, { "epoch": 9.272237196765499, "grad_norm": 3.206239938735962, "learning_rate": 3.7037037037037037e-06, "loss": 0.6083, "step": 3440 }, { "epoch": 9.299191374663073, "grad_norm": 5.175552845001221, "learning_rate": 3.5665294924554186e-06, "loss": 0.6398, "step": 3450 }, { "epoch": 9.326145552560646, "grad_norm": 1.54671311378479, "learning_rate": 3.4293552812071335e-06, "loss": 1.3045, "step": 3460 }, { "epoch": 9.35309973045822, "grad_norm": 1.7365717887878418, "learning_rate": 3.2921810699588483e-06, "loss": 0.6332, "step": 3470 }, { "epoch": 9.380053908355794, "grad_norm": 1.5034387111663818, "learning_rate": 3.1550068587105624e-06, "loss": 0.6763, "step": 3480 }, { "epoch": 9.40700808625337, "grad_norm": 3.266697645187378, "learning_rate": 3.0178326474622772e-06, "loss": 0.8033, "step": 3490 }, { "epoch": 9.433962264150944, "grad_norm": 1.4214359521865845, "learning_rate": 2.880658436213992e-06, "loss": 0.748, "step": 3500 }, { "epoch": 9.460916442048518, "grad_norm": 2.0403311252593994, "learning_rate": 2.7434842249657065e-06, "loss": 0.862, "step": 3510 }, { "epoch": 9.487870619946092, "grad_norm": 1.678673505783081, "learning_rate": 2.6063100137174214e-06, "loss": 0.6221, "step": 3520 }, { "epoch": 9.514824797843666, "grad_norm": 2.317265033721924, "learning_rate": 2.469135802469136e-06, "loss": 0.697, "step": 3530 }, { "epoch": 9.54177897574124, "grad_norm": 2.5591745376586914, "learning_rate": 2.3319615912208507e-06, "loss": 0.6209, "step": 3540 }, { "epoch": 9.568733153638814, "grad_norm": 4.573774337768555, "learning_rate": 2.194787379972565e-06, "loss": 0.6689, "step": 3550 }, { "epoch": 9.595687331536388, "grad_norm": 1.756940484046936, "learning_rate": 2.05761316872428e-06, "loss": 0.9016, "step": 3560 }, { "epoch": 9.622641509433961, "grad_norm": 1.913479208946228, "learning_rate": 1.920438957475995e-06, "loss": 0.7732, "step": 3570 }, { "epoch": 9.649595687331537, "grad_norm": 3.303154230117798, "learning_rate": 1.7832647462277093e-06, "loss": 1.0092, "step": 3580 }, { "epoch": 9.676549865229111, "grad_norm": 2.583827018737793, "learning_rate": 1.6460905349794242e-06, "loss": 1.1913, "step": 3590 }, { "epoch": 9.703504043126685, "grad_norm": 0.6483349800109863, "learning_rate": 1.5089163237311386e-06, "loss": 0.8234, "step": 3600 }, { "epoch": 9.730458221024259, "grad_norm": 3.4046685695648193, "learning_rate": 1.3717421124828533e-06, "loss": 1.1062, "step": 3610 }, { "epoch": 9.757412398921833, "grad_norm": 1.3275524377822876, "learning_rate": 1.234567901234568e-06, "loss": 0.5139, "step": 3620 }, { "epoch": 9.784366576819407, "grad_norm": 2.8111588954925537, "learning_rate": 1.0973936899862826e-06, "loss": 0.7716, "step": 3630 }, { "epoch": 9.81132075471698, "grad_norm": 1.3000844717025757, "learning_rate": 9.602194787379974e-07, "loss": 0.4894, "step": 3640 }, { "epoch": 9.838274932614555, "grad_norm": 2.7694694995880127, "learning_rate": 8.230452674897121e-07, "loss": 0.5647, "step": 3650 }, { "epoch": 9.865229110512129, "grad_norm": 1.6062043905258179, "learning_rate": 6.858710562414266e-07, "loss": 0.9145, "step": 3660 }, { "epoch": 9.892183288409704, "grad_norm": 1.323443055152893, "learning_rate": 5.486968449931413e-07, "loss": 0.7909, "step": 3670 }, { "epoch": 9.919137466307278, "grad_norm": 3.0170159339904785, "learning_rate": 4.1152263374485604e-07, "loss": 0.7927, "step": 3680 }, { "epoch": 9.946091644204852, "grad_norm": 3.5468873977661133, "learning_rate": 2.7434842249657064e-07, "loss": 0.7568, "step": 3690 }, { "epoch": 9.973045822102426, "grad_norm": 1.8293559551239014, "learning_rate": 1.3717421124828532e-07, "loss": 0.6616, "step": 3700 }, { "epoch": 10.0, "grad_norm": 2.379469633102417, "learning_rate": 0.0, "loss": 0.3192, "step": 3710 } ], "logging_steps": 10, "max_steps": 3710, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3062070293667840.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }