diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,4293 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 9.98769987699877, + "eval_steps": 500, + "global_step": 6090, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.016400164001640016, + "grad_norm": 0.49903953075408936, + "learning_rate": 0.00019999866944080257, + "loss": 1.1198, + "step": 10 + }, + { + "epoch": 0.03280032800328003, + "grad_norm": 0.38635629415512085, + "learning_rate": 0.00019999467779861793, + "loss": 0.957, + "step": 20 + }, + { + "epoch": 0.04920049200492005, + "grad_norm": 0.3574012815952301, + "learning_rate": 0.0001999880251796685, + "loss": 0.8963, + "step": 30 + }, + { + "epoch": 0.06560065600656007, + "grad_norm": 0.36165040731430054, + "learning_rate": 0.00019997871176098827, + "loss": 0.8893, + "step": 40 + }, + { + "epoch": 0.08200082000820008, + "grad_norm": 0.3633221983909607, + "learning_rate": 0.0001999667377904184, + "loss": 0.8788, + "step": 50 + }, + { + "epoch": 0.0984009840098401, + "grad_norm": 0.3308263123035431, + "learning_rate": 0.00019995210358660038, + "loss": 0.8597, + "step": 60 + }, + { + "epoch": 0.11480114801148011, + "grad_norm": 0.37404152750968933, + "learning_rate": 0.0001999348095389677, + "loss": 0.8394, + "step": 70 + }, + { + "epoch": 0.13120131201312013, + "grad_norm": 0.36840370297431946, + "learning_rate": 0.0001999148561077355, + "loss": 0.8305, + "step": 80 + }, + { + "epoch": 0.14760147601476015, + "grad_norm": 0.33561620116233826, + "learning_rate": 0.00019989224382388813, + "loss": 0.836, + "step": 90 + }, + { + "epoch": 0.16400164001640016, + "grad_norm": 0.34898167848587036, + "learning_rate": 0.00019986697328916526, + "loss": 0.8259, + "step": 100 + }, + { + "epoch": 0.18040180401804018, + "grad_norm": 0.32518813014030457, + "learning_rate": 0.00019983904517604576, + "loss": 0.8255, + "step": 110 + }, + { + "epoch": 0.1968019680196802, + "grad_norm": 0.34965017437934875, + "learning_rate": 0.00019980846022772978, + "loss": 0.8192, + "step": 120 + }, + { + "epoch": 0.2132021320213202, + "grad_norm": 0.3208518326282501, + "learning_rate": 0.00019977521925811903, + "loss": 0.8177, + "step": 130 + }, + { + "epoch": 0.22960229602296023, + "grad_norm": 0.3215318024158478, + "learning_rate": 0.000199739323151795, + "loss": 0.8046, + "step": 140 + }, + { + "epoch": 0.24600246002460024, + "grad_norm": 0.3197932541370392, + "learning_rate": 0.0001997007728639956, + "loss": 0.8031, + "step": 150 + }, + { + "epoch": 0.26240262402624026, + "grad_norm": 0.3127509355545044, + "learning_rate": 0.0001996595694205897, + "loss": 0.7911, + "step": 160 + }, + { + "epoch": 0.2788027880278803, + "grad_norm": 0.3898150324821472, + "learning_rate": 0.00019961571391804962, + "loss": 0.8013, + "step": 170 + }, + { + "epoch": 0.2952029520295203, + "grad_norm": 0.32089415192604065, + "learning_rate": 0.00019956920752342225, + "loss": 0.7959, + "step": 180 + }, + { + "epoch": 0.3116031160311603, + "grad_norm": 0.3503674864768982, + "learning_rate": 0.0001995200514742978, + "loss": 0.7869, + "step": 190 + }, + { + "epoch": 0.3280032800328003, + "grad_norm": 0.36325424909591675, + "learning_rate": 0.00019946824707877693, + "loss": 0.7813, + "step": 200 + }, + { + "epoch": 0.34440344403444034, + "grad_norm": 0.3318350613117218, + "learning_rate": 0.00019941379571543596, + "loss": 0.796, + "step": 210 + }, + { + "epoch": 0.36080360803608036, + "grad_norm": 0.3133643865585327, + "learning_rate": 0.00019935669883329013, + "loss": 0.7832, + "step": 220 + }, + { + "epoch": 0.3772037720377204, + "grad_norm": 0.3414137363433838, + "learning_rate": 0.00019929695795175507, + "loss": 0.7907, + "step": 230 + }, + { + "epoch": 0.3936039360393604, + "grad_norm": 0.32717064023017883, + "learning_rate": 0.00019923457466060636, + "loss": 0.7824, + "step": 240 + }, + { + "epoch": 0.4100041000410004, + "grad_norm": 0.3150332272052765, + "learning_rate": 0.00019916955061993725, + "loss": 0.7992, + "step": 250 + }, + { + "epoch": 0.4264042640426404, + "grad_norm": 0.3353354334831238, + "learning_rate": 0.00019910188756011446, + "loss": 0.7782, + "step": 260 + }, + { + "epoch": 0.44280442804428044, + "grad_norm": 0.2986792325973511, + "learning_rate": 0.00019903158728173205, + "loss": 0.7609, + "step": 270 + }, + { + "epoch": 0.45920459204592046, + "grad_norm": 0.3223719596862793, + "learning_rate": 0.00019895865165556377, + "loss": 0.7785, + "step": 280 + }, + { + "epoch": 0.4756047560475605, + "grad_norm": 0.3108189105987549, + "learning_rate": 0.00019888308262251285, + "loss": 0.7762, + "step": 290 + }, + { + "epoch": 0.4920049200492005, + "grad_norm": 0.3153323829174042, + "learning_rate": 0.00019880488219356087, + "loss": 0.7797, + "step": 300 + }, + { + "epoch": 0.5084050840508405, + "grad_norm": 0.33102768659591675, + "learning_rate": 0.00019872405244971374, + "loss": 0.7731, + "step": 310 + }, + { + "epoch": 0.5248052480524805, + "grad_norm": 0.29496216773986816, + "learning_rate": 0.00019864059554194669, + "loss": 0.7648, + "step": 320 + }, + { + "epoch": 0.5412054120541205, + "grad_norm": 0.3221481740474701, + "learning_rate": 0.00019855451369114676, + "loss": 0.7762, + "step": 330 + }, + { + "epoch": 0.5576055760557606, + "grad_norm": 0.3474864065647125, + "learning_rate": 0.000198465809188054, + "loss": 0.7691, + "step": 340 + }, + { + "epoch": 0.5740057400574006, + "grad_norm": 0.36715272068977356, + "learning_rate": 0.00019837448439320027, + "loss": 0.7751, + "step": 350 + }, + { + "epoch": 0.5904059040590406, + "grad_norm": 0.33955511450767517, + "learning_rate": 0.00019828054173684644, + "loss": 0.7791, + "step": 360 + }, + { + "epoch": 0.6068060680606806, + "grad_norm": 0.3461228311061859, + "learning_rate": 0.00019818398371891784, + "loss": 0.7728, + "step": 370 + }, + { + "epoch": 0.6232062320623206, + "grad_norm": 0.32821011543273926, + "learning_rate": 0.0001980848129089376, + "loss": 0.7736, + "step": 380 + }, + { + "epoch": 0.6396063960639606, + "grad_norm": 0.3576536476612091, + "learning_rate": 0.00019798303194595846, + "loss": 0.756, + "step": 390 + }, + { + "epoch": 0.6560065600656007, + "grad_norm": 0.3349688649177551, + "learning_rate": 0.00019787864353849232, + "loss": 0.765, + "step": 400 + }, + { + "epoch": 0.6724067240672407, + "grad_norm": 0.2960722744464874, + "learning_rate": 0.00019777165046443824, + "loss": 0.7676, + "step": 410 + }, + { + "epoch": 0.6888068880688807, + "grad_norm": 0.31184637546539307, + "learning_rate": 0.00019766205557100868, + "loss": 0.7366, + "step": 420 + }, + { + "epoch": 0.7052070520705207, + "grad_norm": 0.3365538418292999, + "learning_rate": 0.00019754986177465344, + "loss": 0.7647, + "step": 430 + }, + { + "epoch": 0.7216072160721607, + "grad_norm": 0.3039809763431549, + "learning_rate": 0.00019743507206098233, + "loss": 0.741, + "step": 440 + }, + { + "epoch": 0.7380073800738007, + "grad_norm": 0.32564470171928406, + "learning_rate": 0.00019731768948468549, + "loss": 0.7576, + "step": 450 + }, + { + "epoch": 0.7544075440754408, + "grad_norm": 0.33124250173568726, + "learning_rate": 0.00019719771716945227, + "loss": 0.7392, + "step": 460 + }, + { + "epoch": 0.7708077080770808, + "grad_norm": 0.3306678235530853, + "learning_rate": 0.00019707515830788802, + "loss": 0.7513, + "step": 470 + }, + { + "epoch": 0.7872078720787208, + "grad_norm": 0.31717410683631897, + "learning_rate": 0.00019695001616142915, + "loss": 0.7419, + "step": 480 + }, + { + "epoch": 0.8036080360803608, + "grad_norm": 0.3117675483226776, + "learning_rate": 0.00019682229406025635, + "loss": 0.7543, + "step": 490 + }, + { + "epoch": 0.8200082000820008, + "grad_norm": 0.3408834636211395, + "learning_rate": 0.0001966919954032059, + "loss": 0.7356, + "step": 500 + }, + { + "epoch": 0.8364083640836408, + "grad_norm": 0.3456558883190155, + "learning_rate": 0.0001965591236576794, + "loss": 0.7509, + "step": 510 + }, + { + "epoch": 0.8528085280852808, + "grad_norm": 0.30214881896972656, + "learning_rate": 0.00019642368235955125, + "loss": 0.7483, + "step": 520 + }, + { + "epoch": 0.8692086920869209, + "grad_norm": 0.3334808945655823, + "learning_rate": 0.0001962856751130748, + "loss": 0.7532, + "step": 530 + }, + { + "epoch": 0.8856088560885609, + "grad_norm": 0.3257792890071869, + "learning_rate": 0.00019614510559078625, + "loss": 0.7465, + "step": 540 + }, + { + "epoch": 0.9020090200902009, + "grad_norm": 0.4038065969944, + "learning_rate": 0.00019600197753340699, + "loss": 0.7555, + "step": 550 + }, + { + "epoch": 0.9184091840918409, + "grad_norm": 0.32042616605758667, + "learning_rate": 0.00019585629474974415, + "loss": 0.7395, + "step": 560 + }, + { + "epoch": 0.9348093480934809, + "grad_norm": 0.3661757707595825, + "learning_rate": 0.00019570806111658898, + "loss": 0.7447, + "step": 570 + }, + { + "epoch": 0.951209512095121, + "grad_norm": 0.31112104654312134, + "learning_rate": 0.0001955572805786141, + "loss": 0.7333, + "step": 580 + }, + { + "epoch": 0.967609676096761, + "grad_norm": 0.31758418679237366, + "learning_rate": 0.000195403957148268, + "loss": 0.7243, + "step": 590 + }, + { + "epoch": 0.984009840098401, + "grad_norm": 0.3203936219215393, + "learning_rate": 0.00019524809490566877, + "loss": 0.744, + "step": 600 + }, + { + "epoch": 1.000410004100041, + "grad_norm": 0.3630532920360565, + "learning_rate": 0.00019508969799849523, + "loss": 0.7249, + "step": 610 + }, + { + "epoch": 1.016810168101681, + "grad_norm": 0.31734269857406616, + "learning_rate": 0.00019492877064187654, + "loss": 0.7088, + "step": 620 + }, + { + "epoch": 1.033210332103321, + "grad_norm": 0.3637007176876068, + "learning_rate": 0.00019476531711828027, + "loss": 0.7139, + "step": 630 + }, + { + "epoch": 1.049610496104961, + "grad_norm": 0.3162935972213745, + "learning_rate": 0.00019459934177739813, + "loss": 0.7064, + "step": 640 + }, + { + "epoch": 1.066010660106601, + "grad_norm": 0.32640552520751953, + "learning_rate": 0.0001944308490360305, + "loss": 0.7161, + "step": 650 + }, + { + "epoch": 1.082410824108241, + "grad_norm": 0.34453850984573364, + "learning_rate": 0.0001942598433779687, + "loss": 0.7095, + "step": 660 + }, + { + "epoch": 1.098810988109881, + "grad_norm": 0.3382568061351776, + "learning_rate": 0.00019408632935387577, + "loss": 0.7092, + "step": 670 + }, + { + "epoch": 1.115211152111521, + "grad_norm": 0.3298267722129822, + "learning_rate": 0.0001939103115811653, + "loss": 0.7154, + "step": 680 + }, + { + "epoch": 1.1316113161131611, + "grad_norm": 0.33924031257629395, + "learning_rate": 0.00019373179474387858, + "loss": 0.6945, + "step": 690 + }, + { + "epoch": 1.1480114801148011, + "grad_norm": 0.3305610418319702, + "learning_rate": 0.0001935507835925601, + "loss": 0.7156, + "step": 700 + }, + { + "epoch": 1.1644116441164412, + "grad_norm": 0.3400373160839081, + "learning_rate": 0.00019336728294413083, + "loss": 0.7091, + "step": 710 + }, + { + "epoch": 1.1808118081180812, + "grad_norm": 0.3354150056838989, + "learning_rate": 0.00019318129768176032, + "loss": 0.7019, + "step": 720 + }, + { + "epoch": 1.1972119721197212, + "grad_norm": 0.3427719175815582, + "learning_rate": 0.00019299283275473657, + "loss": 0.7157, + "step": 730 + }, + { + "epoch": 1.2136121361213612, + "grad_norm": 0.3088740408420563, + "learning_rate": 0.00019280189317833445, + "loss": 0.7115, + "step": 740 + }, + { + "epoch": 1.2300123001230012, + "grad_norm": 0.34402570128440857, + "learning_rate": 0.0001926084840336821, + "loss": 0.7197, + "step": 750 + }, + { + "epoch": 1.2464124641246412, + "grad_norm": 0.3542528450489044, + "learning_rate": 0.0001924126104676259, + "loss": 0.7103, + "step": 760 + }, + { + "epoch": 1.2628126281262813, + "grad_norm": 0.3533455431461334, + "learning_rate": 0.00019221427769259333, + "loss": 0.7209, + "step": 770 + }, + { + "epoch": 1.2792127921279213, + "grad_norm": 0.3470947742462158, + "learning_rate": 0.00019201349098645434, + "loss": 0.7137, + "step": 780 + }, + { + "epoch": 1.2956129561295613, + "grad_norm": 0.3277159035205841, + "learning_rate": 0.0001918102556923809, + "loss": 0.7036, + "step": 790 + }, + { + "epoch": 1.3120131201312013, + "grad_norm": 0.35741642117500305, + "learning_rate": 0.0001916045772187048, + "loss": 0.7013, + "step": 800 + }, + { + "epoch": 1.3284132841328413, + "grad_norm": 0.33222687244415283, + "learning_rate": 0.0001913964610387738, + "loss": 0.7052, + "step": 810 + }, + { + "epoch": 1.3448134481344813, + "grad_norm": 0.3761126399040222, + "learning_rate": 0.00019118591269080574, + "loss": 0.695, + "step": 820 + }, + { + "epoch": 1.3612136121361214, + "grad_norm": 0.34687796235084534, + "learning_rate": 0.00019097293777774153, + "loss": 0.7053, + "step": 830 + }, + { + "epoch": 1.3776137761377614, + "grad_norm": 0.3391115367412567, + "learning_rate": 0.00019075754196709572, + "loss": 0.6842, + "step": 840 + }, + { + "epoch": 1.3940139401394014, + "grad_norm": 0.35560715198516846, + "learning_rate": 0.00019053973099080585, + "loss": 0.6925, + "step": 850 + }, + { + "epoch": 1.4104141041410414, + "grad_norm": 0.35144075751304626, + "learning_rate": 0.00019031951064507987, + "loss": 0.7063, + "step": 860 + }, + { + "epoch": 1.4268142681426814, + "grad_norm": 0.3776865005493164, + "learning_rate": 0.0001900968867902419, + "loss": 0.7157, + "step": 870 + }, + { + "epoch": 1.4432144321443214, + "grad_norm": 0.3723245859146118, + "learning_rate": 0.00018987186535057634, + "loss": 0.7064, + "step": 880 + }, + { + "epoch": 1.4596145961459615, + "grad_norm": 0.3623642921447754, + "learning_rate": 0.0001896444523141701, + "loss": 0.7056, + "step": 890 + }, + { + "epoch": 1.4760147601476015, + "grad_norm": 0.34806951880455017, + "learning_rate": 0.0001894146537327533, + "loss": 0.7086, + "step": 900 + }, + { + "epoch": 1.4924149241492415, + "grad_norm": 0.36614471673965454, + "learning_rate": 0.00018918247572153823, + "loss": 0.7032, + "step": 910 + }, + { + "epoch": 1.5088150881508815, + "grad_norm": 0.351053923368454, + "learning_rate": 0.00018894792445905674, + "loss": 0.6988, + "step": 920 + }, + { + "epoch": 1.5252152521525215, + "grad_norm": 0.37270405888557434, + "learning_rate": 0.00018871100618699554, + "loss": 0.6926, + "step": 930 + }, + { + "epoch": 1.5416154161541615, + "grad_norm": 0.3521835207939148, + "learning_rate": 0.00018847172721003043, + "loss": 0.7004, + "step": 940 + }, + { + "epoch": 1.5580155801558015, + "grad_norm": 0.37083443999290466, + "learning_rate": 0.00018823009389565818, + "loss": 0.7141, + "step": 950 + }, + { + "epoch": 1.5744157441574416, + "grad_norm": 0.36266541481018066, + "learning_rate": 0.00018798611267402746, + "loss": 0.7029, + "step": 960 + }, + { + "epoch": 1.5908159081590816, + "grad_norm": 0.34403783082962036, + "learning_rate": 0.0001877397900377674, + "loss": 0.7022, + "step": 970 + }, + { + "epoch": 1.6072160721607216, + "grad_norm": 0.37788769602775574, + "learning_rate": 0.00018749113254181498, + "loss": 0.7049, + "step": 980 + }, + { + "epoch": 1.6236162361623616, + "grad_norm": 0.35607653856277466, + "learning_rate": 0.00018724014680324057, + "loss": 0.6957, + "step": 990 + }, + { + "epoch": 1.6400164001640016, + "grad_norm": 0.34468725323677063, + "learning_rate": 0.00018698683950107184, + "loss": 0.7148, + "step": 1000 + }, + { + "epoch": 1.6564165641656416, + "grad_norm": 0.36872032284736633, + "learning_rate": 0.00018673121737611598, + "loss": 0.6922, + "step": 1010 + }, + { + "epoch": 1.6728167281672817, + "grad_norm": 0.34786421060562134, + "learning_rate": 0.00018647328723078038, + "loss": 0.7021, + "step": 1020 + }, + { + "epoch": 1.6892168921689217, + "grad_norm": 0.370241641998291, + "learning_rate": 0.00018621305592889163, + "loss": 0.6789, + "step": 1030 + }, + { + "epoch": 1.7056170561705617, + "grad_norm": 0.3378838300704956, + "learning_rate": 0.00018595053039551274, + "loss": 0.6741, + "step": 1040 + }, + { + "epoch": 1.7220172201722017, + "grad_norm": 0.3661263883113861, + "learning_rate": 0.00018568571761675893, + "loss": 0.6808, + "step": 1050 + }, + { + "epoch": 1.7384173841738417, + "grad_norm": 0.3620483875274658, + "learning_rate": 0.0001854186246396118, + "loss": 0.6895, + "step": 1060 + }, + { + "epoch": 1.7548175481754817, + "grad_norm": 0.3466210663318634, + "learning_rate": 0.00018514925857173177, + "loss": 0.6876, + "step": 1070 + }, + { + "epoch": 1.7712177121771218, + "grad_norm": 0.36353030800819397, + "learning_rate": 0.0001848776265812687, + "loss": 0.6876, + "step": 1080 + }, + { + "epoch": 1.7876178761787618, + "grad_norm": 0.3688850998878479, + "learning_rate": 0.00018460373589667154, + "loss": 0.6881, + "step": 1090 + }, + { + "epoch": 1.8040180401804018, + "grad_norm": 0.35844072699546814, + "learning_rate": 0.00018432759380649562, + "loss": 0.7038, + "step": 1100 + }, + { + "epoch": 1.8204182041820418, + "grad_norm": 0.38838833570480347, + "learning_rate": 0.00018404920765920896, + "loss": 0.6961, + "step": 1110 + }, + { + "epoch": 1.8368183681836818, + "grad_norm": 0.34356170892715454, + "learning_rate": 0.00018376858486299647, + "loss": 0.6849, + "step": 1120 + }, + { + "epoch": 1.8532185321853218, + "grad_norm": 0.3590626120567322, + "learning_rate": 0.00018348573288556308, + "loss": 0.68, + "step": 1130 + }, + { + "epoch": 1.8696186961869619, + "grad_norm": 0.3639477491378784, + "learning_rate": 0.00018320065925393468, + "loss": 0.6996, + "step": 1140 + }, + { + "epoch": 1.8860188601886019, + "grad_norm": 0.39745190739631653, + "learning_rate": 0.00018291337155425821, + "loss": 0.6901, + "step": 1150 + }, + { + "epoch": 1.902419024190242, + "grad_norm": 0.37302690744400024, + "learning_rate": 0.0001826238774315995, + "loss": 0.6784, + "step": 1160 + }, + { + "epoch": 1.918819188191882, + "grad_norm": 0.37480470538139343, + "learning_rate": 0.00018233218458973984, + "loss": 0.6806, + "step": 1170 + }, + { + "epoch": 1.935219352193522, + "grad_norm": 0.37153083086013794, + "learning_rate": 0.00018203830079097113, + "loss": 0.6853, + "step": 1180 + }, + { + "epoch": 1.951619516195162, + "grad_norm": 0.3530969023704529, + "learning_rate": 0.00018174223385588917, + "loss": 0.6785, + "step": 1190 + }, + { + "epoch": 1.968019680196802, + "grad_norm": 0.37324872612953186, + "learning_rate": 0.00018144399166318572, + "loss": 0.6733, + "step": 1200 + }, + { + "epoch": 1.984419844198442, + "grad_norm": 0.3680344223976135, + "learning_rate": 0.00018114358214943853, + "loss": 0.6885, + "step": 1210 + }, + { + "epoch": 2.000820008200082, + "grad_norm": 0.42146405577659607, + "learning_rate": 0.00018084101330890048, + "loss": 0.6759, + "step": 1220 + }, + { + "epoch": 2.0172201722017222, + "grad_norm": 0.3808436989784241, + "learning_rate": 0.00018053629319328662, + "loss": 0.655, + "step": 1230 + }, + { + "epoch": 2.033620336203362, + "grad_norm": 0.39610400795936584, + "learning_rate": 0.00018022942991156, + "loss": 0.6675, + "step": 1240 + }, + { + "epoch": 2.0500205002050023, + "grad_norm": 0.3906943202018738, + "learning_rate": 0.00017992043162971588, + "loss": 0.6451, + "step": 1250 + }, + { + "epoch": 2.066420664206642, + "grad_norm": 0.4022831916809082, + "learning_rate": 0.00017960930657056438, + "loss": 0.6699, + "step": 1260 + }, + { + "epoch": 2.0828208282082823, + "grad_norm": 0.41200950741767883, + "learning_rate": 0.00017929606301351168, + "loss": 0.6559, + "step": 1270 + }, + { + "epoch": 2.099220992209922, + "grad_norm": 0.37226754426956177, + "learning_rate": 0.00017898070929433965, + "loss": 0.655, + "step": 1280 + }, + { + "epoch": 2.1156211562115623, + "grad_norm": 0.3640751540660858, + "learning_rate": 0.00017866325380498416, + "loss": 0.6471, + "step": 1290 + }, + { + "epoch": 2.132021320213202, + "grad_norm": 0.37838858366012573, + "learning_rate": 0.00017834370499331166, + "loss": 0.651, + "step": 1300 + }, + { + "epoch": 2.1484214842148424, + "grad_norm": 0.3831733167171478, + "learning_rate": 0.0001780220713628943, + "loss": 0.6504, + "step": 1310 + }, + { + "epoch": 2.164821648216482, + "grad_norm": 0.39556068181991577, + "learning_rate": 0.0001776983614727838, + "loss": 0.6501, + "step": 1320 + }, + { + "epoch": 2.181221812218122, + "grad_norm": 0.3833105266094208, + "learning_rate": 0.00017737258393728364, + "loss": 0.6598, + "step": 1330 + }, + { + "epoch": 2.197621976219762, + "grad_norm": 0.40265655517578125, + "learning_rate": 0.00017704474742571969, + "loss": 0.663, + "step": 1340 + }, + { + "epoch": 2.2140221402214024, + "grad_norm": 0.4098280072212219, + "learning_rate": 0.00017671486066220965, + "loss": 0.6567, + "step": 1350 + }, + { + "epoch": 2.230422304223042, + "grad_norm": 0.4019940197467804, + "learning_rate": 0.0001763829324254309, + "loss": 0.6717, + "step": 1360 + }, + { + "epoch": 2.246822468224682, + "grad_norm": 0.3930509388446808, + "learning_rate": 0.00017604897154838685, + "loss": 0.6544, + "step": 1370 + }, + { + "epoch": 2.2632226322263223, + "grad_norm": 0.40629473328590393, + "learning_rate": 0.00017571298691817177, + "loss": 0.65, + "step": 1380 + }, + { + "epoch": 2.2796227962279625, + "grad_norm": 0.41943445801734924, + "learning_rate": 0.00017537498747573443, + "loss": 0.6612, + "step": 1390 + }, + { + "epoch": 2.2960229602296023, + "grad_norm": 0.3832157552242279, + "learning_rate": 0.00017503498221564025, + "loss": 0.6599, + "step": 1400 + }, + { + "epoch": 2.312423124231242, + "grad_norm": 0.3786883056163788, + "learning_rate": 0.0001746929801858317, + "loss": 0.6636, + "step": 1410 + }, + { + "epoch": 2.3288232882328823, + "grad_norm": 0.40903323888778687, + "learning_rate": 0.0001743489904873877, + "loss": 0.6462, + "step": 1420 + }, + { + "epoch": 2.3452234522345226, + "grad_norm": 0.3858684003353119, + "learning_rate": 0.00017400302227428143, + "loss": 0.6556, + "step": 1430 + }, + { + "epoch": 2.3616236162361623, + "grad_norm": 0.43149909377098083, + "learning_rate": 0.0001736550847531366, + "loss": 0.6591, + "step": 1440 + }, + { + "epoch": 2.378023780237802, + "grad_norm": 0.43920159339904785, + "learning_rate": 0.00017330518718298264, + "loss": 0.6609, + "step": 1450 + }, + { + "epoch": 2.3944239442394424, + "grad_norm": 0.39907246828079224, + "learning_rate": 0.0001729533388750081, + "loss": 0.6572, + "step": 1460 + }, + { + "epoch": 2.4108241082410826, + "grad_norm": 0.3920808434486389, + "learning_rate": 0.0001725995491923131, + "loss": 0.6497, + "step": 1470 + }, + { + "epoch": 2.4272242722427224, + "grad_norm": 0.39895495772361755, + "learning_rate": 0.00017224382754965989, + "loss": 0.646, + "step": 1480 + }, + { + "epoch": 2.443624436244362, + "grad_norm": 0.41222187876701355, + "learning_rate": 0.00017188618341322254, + "loss": 0.6458, + "step": 1490 + }, + { + "epoch": 2.4600246002460024, + "grad_norm": 0.41394343972206116, + "learning_rate": 0.00017152662630033505, + "loss": 0.6528, + "step": 1500 + }, + { + "epoch": 2.4764247642476427, + "grad_norm": 0.4161871373653412, + "learning_rate": 0.00017116516577923775, + "loss": 0.6414, + "step": 1510 + }, + { + "epoch": 2.4928249282492825, + "grad_norm": 0.43303370475769043, + "learning_rate": 0.00017080181146882317, + "loss": 0.6573, + "step": 1520 + }, + { + "epoch": 2.5092250922509223, + "grad_norm": 0.3973419964313507, + "learning_rate": 0.00017043657303837963, + "loss": 0.6541, + "step": 1530 + }, + { + "epoch": 2.5256252562525625, + "grad_norm": 0.3919370472431183, + "learning_rate": 0.00017006946020733425, + "loss": 0.6626, + "step": 1540 + }, + { + "epoch": 2.5420254202542027, + "grad_norm": 0.4177446663379669, + "learning_rate": 0.00016970048274499408, + "loss": 0.6478, + "step": 1550 + }, + { + "epoch": 2.5584255842558425, + "grad_norm": 0.39967086911201477, + "learning_rate": 0.0001693296504702862, + "loss": 0.6679, + "step": 1560 + }, + { + "epoch": 2.5748257482574823, + "grad_norm": 0.4124162495136261, + "learning_rate": 0.00016895697325149657, + "loss": 0.6597, + "step": 1570 + }, + { + "epoch": 2.5912259122591226, + "grad_norm": 0.37739571928977966, + "learning_rate": 0.00016858246100600715, + "loss": 0.6578, + "step": 1580 + }, + { + "epoch": 2.607626076260763, + "grad_norm": 0.4335273802280426, + "learning_rate": 0.00016820612370003221, + "loss": 0.6608, + "step": 1590 + }, + { + "epoch": 2.6240262402624026, + "grad_norm": 0.3718789517879486, + "learning_rate": 0.00016782797134835305, + "loss": 0.6447, + "step": 1600 + }, + { + "epoch": 2.6404264042640424, + "grad_norm": 0.41410380601882935, + "learning_rate": 0.0001674480140140514, + "loss": 0.6562, + "step": 1610 + }, + { + "epoch": 2.6568265682656826, + "grad_norm": 0.4115227460861206, + "learning_rate": 0.00016706626180824186, + "loss": 0.6451, + "step": 1620 + }, + { + "epoch": 2.673226732267323, + "grad_norm": 0.4145969748497009, + "learning_rate": 0.00016668272488980254, + "loss": 0.6326, + "step": 1630 + }, + { + "epoch": 2.6896268962689627, + "grad_norm": 0.42359670996665955, + "learning_rate": 0.00016629741346510496, + "loss": 0.6324, + "step": 1640 + }, + { + "epoch": 2.7060270602706025, + "grad_norm": 0.38336479663848877, + "learning_rate": 0.0001659103377877423, + "loss": 0.6508, + "step": 1650 + }, + { + "epoch": 2.7224272242722427, + "grad_norm": 0.4128510057926178, + "learning_rate": 0.00016552150815825662, + "loss": 0.6538, + "step": 1660 + }, + { + "epoch": 2.738827388273883, + "grad_norm": 0.47073811292648315, + "learning_rate": 0.0001651309349238647, + "loss": 0.6574, + "step": 1670 + }, + { + "epoch": 2.7552275522755227, + "grad_norm": 0.4234805703163147, + "learning_rate": 0.00016473862847818277, + "loss": 0.6643, + "step": 1680 + }, + { + "epoch": 2.7716277162771625, + "grad_norm": 0.4076490104198456, + "learning_rate": 0.0001643445992609498, + "loss": 0.6509, + "step": 1690 + }, + { + "epoch": 2.7880278802788028, + "grad_norm": 0.4320470988750458, + "learning_rate": 0.00016394885775774975, + "loss": 0.643, + "step": 1700 + }, + { + "epoch": 2.804428044280443, + "grad_norm": 0.409150630235672, + "learning_rate": 0.00016355141449973256, + "loss": 0.6452, + "step": 1710 + }, + { + "epoch": 2.820828208282083, + "grad_norm": 0.46668359637260437, + "learning_rate": 0.0001631522800633339, + "loss": 0.645, + "step": 1720 + }, + { + "epoch": 2.8372283722837226, + "grad_norm": 0.4314485192298889, + "learning_rate": 0.00016275146506999365, + "loss": 0.6374, + "step": 1730 + }, + { + "epoch": 2.853628536285363, + "grad_norm": 0.41864487528800964, + "learning_rate": 0.00016234898018587337, + "loss": 0.6343, + "step": 1740 + }, + { + "epoch": 2.870028700287003, + "grad_norm": 0.4255831241607666, + "learning_rate": 0.0001619448361215723, + "loss": 0.6614, + "step": 1750 + }, + { + "epoch": 2.886428864288643, + "grad_norm": 0.411807656288147, + "learning_rate": 0.0001615390436318425, + "loss": 0.6372, + "step": 1760 + }, + { + "epoch": 2.9028290282902827, + "grad_norm": 0.41166943311691284, + "learning_rate": 0.0001611316135153026, + "loss": 0.6505, + "step": 1770 + }, + { + "epoch": 2.919229192291923, + "grad_norm": 0.41985031962394714, + "learning_rate": 0.0001607225566141503, + "loss": 0.6515, + "step": 1780 + }, + { + "epoch": 2.935629356293563, + "grad_norm": 0.40023133158683777, + "learning_rate": 0.0001603118838138741, + "loss": 0.6466, + "step": 1790 + }, + { + "epoch": 2.952029520295203, + "grad_norm": 0.4138244688510895, + "learning_rate": 0.0001598996060429634, + "loss": 0.6475, + "step": 1800 + }, + { + "epoch": 2.9684296842968427, + "grad_norm": 0.39823102951049805, + "learning_rate": 0.0001594857342726178, + "loss": 0.6419, + "step": 1810 + }, + { + "epoch": 2.984829848298483, + "grad_norm": 0.3888455033302307, + "learning_rate": 0.0001590702795164551, + "loss": 0.6416, + "step": 1820 + }, + { + "epoch": 3.001230012300123, + "grad_norm": 0.5585771799087524, + "learning_rate": 0.0001586532528302183, + "loss": 0.6565, + "step": 1830 + }, + { + "epoch": 3.017630176301763, + "grad_norm": 0.4332481026649475, + "learning_rate": 0.00015823466531148124, + "loss": 0.619, + "step": 1840 + }, + { + "epoch": 3.0340303403034032, + "grad_norm": 0.44493553042411804, + "learning_rate": 0.0001578145280993533, + "loss": 0.6057, + "step": 1850 + }, + { + "epoch": 3.050430504305043, + "grad_norm": 0.43917590379714966, + "learning_rate": 0.0001573928523741832, + "loss": 0.6058, + "step": 1860 + }, + { + "epoch": 3.0668306683066833, + "grad_norm": 0.40770649909973145, + "learning_rate": 0.0001569696493572612, + "loss": 0.6092, + "step": 1870 + }, + { + "epoch": 3.083230832308323, + "grad_norm": 0.44579923152923584, + "learning_rate": 0.0001565449303105207, + "loss": 0.6272, + "step": 1880 + }, + { + "epoch": 3.0996309963099633, + "grad_norm": 0.4329695403575897, + "learning_rate": 0.00015611870653623825, + "loss": 0.6238, + "step": 1890 + }, + { + "epoch": 3.116031160311603, + "grad_norm": 0.44702407717704773, + "learning_rate": 0.00015569098937673318, + "loss": 0.6124, + "step": 1900 + }, + { + "epoch": 3.1324313243132433, + "grad_norm": 0.43535304069519043, + "learning_rate": 0.00015526179021406553, + "loss": 0.6179, + "step": 1910 + }, + { + "epoch": 3.148831488314883, + "grad_norm": 0.4820752739906311, + "learning_rate": 0.0001548311204697331, + "loss": 0.6228, + "step": 1920 + }, + { + "epoch": 3.1652316523165234, + "grad_norm": 0.4629499614238739, + "learning_rate": 0.00015439899160436772, + "loss": 0.6189, + "step": 1930 + }, + { + "epoch": 3.181631816318163, + "grad_norm": 0.4845089316368103, + "learning_rate": 0.00015396541511743012, + "loss": 0.6229, + "step": 1940 + }, + { + "epoch": 3.1980319803198034, + "grad_norm": 0.46592435240745544, + "learning_rate": 0.00015353040254690393, + "loss": 0.6178, + "step": 1950 + }, + { + "epoch": 3.214432144321443, + "grad_norm": 0.44851964712142944, + "learning_rate": 0.0001530939654689887, + "loss": 0.6246, + "step": 1960 + }, + { + "epoch": 3.2308323083230834, + "grad_norm": 0.4483705461025238, + "learning_rate": 0.00015265611549779176, + "loss": 0.6204, + "step": 1970 + }, + { + "epoch": 3.2472324723247232, + "grad_norm": 0.45693984627723694, + "learning_rate": 0.00015221686428501928, + "loss": 0.6061, + "step": 1980 + }, + { + "epoch": 3.2636326363263635, + "grad_norm": 0.46255984902381897, + "learning_rate": 0.0001517762235196661, + "loss": 0.6058, + "step": 1990 + }, + { + "epoch": 3.2800328003280033, + "grad_norm": 0.471983939409256, + "learning_rate": 0.00015133420492770462, + "loss": 0.6157, + "step": 2000 + }, + { + "epoch": 3.296432964329643, + "grad_norm": 0.49640896916389465, + "learning_rate": 0.0001508908202717729, + "loss": 0.6237, + "step": 2010 + }, + { + "epoch": 3.3128331283312833, + "grad_norm": 0.4903806149959564, + "learning_rate": 0.00015044608135086164, + "loss": 0.6339, + "step": 2020 + }, + { + "epoch": 3.3292332923329235, + "grad_norm": 0.47123804688453674, + "learning_rate": 0.00015000000000000001, + "loss": 0.6319, + "step": 2030 + }, + { + "epoch": 3.3456334563345633, + "grad_norm": 0.47124335169792175, + "learning_rate": 0.00014955258808994096, + "loss": 0.6119, + "step": 2040 + }, + { + "epoch": 3.362033620336203, + "grad_norm": 0.4579004645347595, + "learning_rate": 0.00014910385752684506, + "loss": 0.6042, + "step": 2050 + }, + { + "epoch": 3.3784337843378434, + "grad_norm": 0.4778992235660553, + "learning_rate": 0.0001486538202519639, + "loss": 0.6187, + "step": 2060 + }, + { + "epoch": 3.3948339483394836, + "grad_norm": 0.4603955149650574, + "learning_rate": 0.0001482024882413222, + "loss": 0.6166, + "step": 2070 + }, + { + "epoch": 3.4112341123411234, + "grad_norm": 0.4642026126384735, + "learning_rate": 0.00014774987350539911, + "loss": 0.6183, + "step": 2080 + }, + { + "epoch": 3.427634276342763, + "grad_norm": 0.5194045305252075, + "learning_rate": 0.00014729598808880861, + "loss": 0.617, + "step": 2090 + }, + { + "epoch": 3.4440344403444034, + "grad_norm": 0.4535214900970459, + "learning_rate": 0.00014684084406997903, + "loss": 0.6301, + "step": 2100 + }, + { + "epoch": 3.4604346043460437, + "grad_norm": 0.47860169410705566, + "learning_rate": 0.00014638445356083155, + "loss": 0.6113, + "step": 2110 + }, + { + "epoch": 3.4768347683476835, + "grad_norm": 0.4803871810436249, + "learning_rate": 0.000145926828706458, + "loss": 0.6122, + "step": 2120 + }, + { + "epoch": 3.4932349323493233, + "grad_norm": 0.47383010387420654, + "learning_rate": 0.00014546798168479756, + "loss": 0.6137, + "step": 2130 + }, + { + "epoch": 3.5096350963509635, + "grad_norm": 0.47791990637779236, + "learning_rate": 0.0001450079247063127, + "loss": 0.6215, + "step": 2140 + }, + { + "epoch": 3.5260352603526037, + "grad_norm": 0.4606546461582184, + "learning_rate": 0.0001445466700136643, + "loss": 0.6022, + "step": 2150 + }, + { + "epoch": 3.5424354243542435, + "grad_norm": 0.4335190951824188, + "learning_rate": 0.00014408422988138584, + "loss": 0.6089, + "step": 2160 + }, + { + "epoch": 3.5588355883558833, + "grad_norm": 0.45156562328338623, + "learning_rate": 0.00014362061661555675, + "loss": 0.6251, + "step": 2170 + }, + { + "epoch": 3.5752357523575236, + "grad_norm": 0.43693843483924866, + "learning_rate": 0.00014315584255347494, + "loss": 0.62, + "step": 2180 + }, + { + "epoch": 3.591635916359164, + "grad_norm": 0.46648937463760376, + "learning_rate": 0.00014268992006332846, + "loss": 0.6336, + "step": 2190 + }, + { + "epoch": 3.6080360803608036, + "grad_norm": 0.47864094376564026, + "learning_rate": 0.00014222286154386642, + "loss": 0.6079, + "step": 2200 + }, + { + "epoch": 3.6244362443624434, + "grad_norm": 0.43990063667297363, + "learning_rate": 0.00014175467942406894, + "loss": 0.6321, + "step": 2210 + }, + { + "epoch": 3.6408364083640836, + "grad_norm": 0.4700396656990051, + "learning_rate": 0.0001412853861628166, + "loss": 0.6195, + "step": 2220 + }, + { + "epoch": 3.657236572365724, + "grad_norm": 0.45286816358566284, + "learning_rate": 0.00014081499424855863, + "loss": 0.6236, + "step": 2230 + }, + { + "epoch": 3.6736367363673637, + "grad_norm": 0.4623110294342041, + "learning_rate": 0.00014034351619898088, + "loss": 0.6269, + "step": 2240 + }, + { + "epoch": 3.6900369003690034, + "grad_norm": 0.5022940039634705, + "learning_rate": 0.00013987096456067236, + "loss": 0.6107, + "step": 2250 + }, + { + "epoch": 3.7064370643706437, + "grad_norm": 0.49338841438293457, + "learning_rate": 0.00013939735190879172, + "loss": 0.5996, + "step": 2260 + }, + { + "epoch": 3.722837228372284, + "grad_norm": 0.4769245982170105, + "learning_rate": 0.0001389226908467323, + "loss": 0.6074, + "step": 2270 + }, + { + "epoch": 3.7392373923739237, + "grad_norm": 0.48450419306755066, + "learning_rate": 0.00013844699400578696, + "loss": 0.6092, + "step": 2280 + }, + { + "epoch": 3.7556375563755635, + "grad_norm": 0.46382635831832886, + "learning_rate": 0.00013797027404481184, + "loss": 0.6154, + "step": 2290 + }, + { + "epoch": 3.7720377203772038, + "grad_norm": 0.4822761118412018, + "learning_rate": 0.00013749254364988956, + "loss": 0.612, + "step": 2300 + }, + { + "epoch": 3.788437884378844, + "grad_norm": 0.4664019048213959, + "learning_rate": 0.00013701381553399145, + "loss": 0.6003, + "step": 2310 + }, + { + "epoch": 3.804838048380484, + "grad_norm": 0.4592795670032501, + "learning_rate": 0.00013653410243663952, + "loss": 0.6122, + "step": 2320 + }, + { + "epoch": 3.8212382123821236, + "grad_norm": 0.4780454635620117, + "learning_rate": 0.00013605341712356723, + "loss": 0.6216, + "step": 2330 + }, + { + "epoch": 3.837638376383764, + "grad_norm": 0.48119214177131653, + "learning_rate": 0.00013557177238637986, + "loss": 0.6119, + "step": 2340 + }, + { + "epoch": 3.854038540385404, + "grad_norm": 0.44965630769729614, + "learning_rate": 0.00013508918104221412, + "loss": 0.619, + "step": 2350 + }, + { + "epoch": 3.870438704387044, + "grad_norm": 0.4915274977684021, + "learning_rate": 0.00013460565593339705, + "loss": 0.6101, + "step": 2360 + }, + { + "epoch": 3.8868388683886836, + "grad_norm": 0.4430049955844879, + "learning_rate": 0.00013412120992710425, + "loss": 0.6125, + "step": 2370 + }, + { + "epoch": 3.903239032390324, + "grad_norm": 0.4661041796207428, + "learning_rate": 0.0001336358559150175, + "loss": 0.6177, + "step": 2380 + }, + { + "epoch": 3.919639196391964, + "grad_norm": 0.48300737142562866, + "learning_rate": 0.00013314960681298175, + "loss": 0.6039, + "step": 2390 + }, + { + "epoch": 3.936039360393604, + "grad_norm": 0.45341914892196655, + "learning_rate": 0.00013266247556066122, + "loss": 0.5989, + "step": 2400 + }, + { + "epoch": 3.9524395243952437, + "grad_norm": 0.46126899123191833, + "learning_rate": 0.00013217447512119533, + "loss": 0.6188, + "step": 2410 + }, + { + "epoch": 3.968839688396884, + "grad_norm": 0.4723737835884094, + "learning_rate": 0.0001316856184808535, + "loss": 0.6091, + "step": 2420 + }, + { + "epoch": 3.985239852398524, + "grad_norm": 0.4723500907421112, + "learning_rate": 0.0001311959186486898, + "loss": 0.6178, + "step": 2430 + }, + { + "epoch": 4.001640016400164, + "grad_norm": 0.4641455113887787, + "learning_rate": 0.00013070538865619642, + "loss": 0.6144, + "step": 2440 + }, + { + "epoch": 4.018040180401804, + "grad_norm": 0.5465942621231079, + "learning_rate": 0.00013021404155695725, + "loss": 0.5762, + "step": 2450 + }, + { + "epoch": 4.0344403444034445, + "grad_norm": 0.5195038318634033, + "learning_rate": 0.00012972189042630044, + "loss": 0.5936, + "step": 2460 + }, + { + "epoch": 4.050840508405084, + "grad_norm": 0.5215120911598206, + "learning_rate": 0.00012922894836095013, + "loss": 0.5797, + "step": 2470 + }, + { + "epoch": 4.067240672406724, + "grad_norm": 0.5008783340454102, + "learning_rate": 0.00012873522847867835, + "loss": 0.5887, + "step": 2480 + }, + { + "epoch": 4.083640836408364, + "grad_norm": 0.47933313250541687, + "learning_rate": 0.0001282407439179557, + "loss": 0.5831, + "step": 2490 + }, + { + "epoch": 4.1000410004100045, + "grad_norm": 0.5141870379447937, + "learning_rate": 0.00012774550783760182, + "loss": 0.5765, + "step": 2500 + }, + { + "epoch": 4.116441164411644, + "grad_norm": 0.4889928996562958, + "learning_rate": 0.0001272495334164351, + "loss": 0.5834, + "step": 2510 + }, + { + "epoch": 4.132841328413284, + "grad_norm": 0.5281891822814941, + "learning_rate": 0.00012675283385292212, + "loss": 0.5778, + "step": 2520 + }, + { + "epoch": 4.149241492414924, + "grad_norm": 0.5182332396507263, + "learning_rate": 0.00012625542236482628, + "loss": 0.5882, + "step": 2530 + }, + { + "epoch": 4.165641656416565, + "grad_norm": 0.554658055305481, + "learning_rate": 0.00012575731218885625, + "loss": 0.5871, + "step": 2540 + }, + { + "epoch": 4.182041820418204, + "grad_norm": 0.5192410349845886, + "learning_rate": 0.00012525851658031352, + "loss": 0.5768, + "step": 2550 + }, + { + "epoch": 4.198441984419844, + "grad_norm": 0.5177326202392578, + "learning_rate": 0.0001247590488127398, + "loss": 0.5849, + "step": 2560 + }, + { + "epoch": 4.214842148421484, + "grad_norm": 0.5142077803611755, + "learning_rate": 0.0001242589221775637, + "loss": 0.5744, + "step": 2570 + }, + { + "epoch": 4.231242312423125, + "grad_norm": 0.48389384150505066, + "learning_rate": 0.00012375814998374712, + "loss": 0.5977, + "step": 2580 + }, + { + "epoch": 4.2476424764247644, + "grad_norm": 0.5631837248802185, + "learning_rate": 0.00012325674555743106, + "loss": 0.5827, + "step": 2590 + }, + { + "epoch": 4.264042640426404, + "grad_norm": 0.5213803052902222, + "learning_rate": 0.0001227547222415809, + "loss": 0.5815, + "step": 2600 + }, + { + "epoch": 4.280442804428044, + "grad_norm": 0.5063258409500122, + "learning_rate": 0.00012225209339563145, + "loss": 0.591, + "step": 2610 + }, + { + "epoch": 4.296842968429685, + "grad_norm": 0.503495454788208, + "learning_rate": 0.0001217488723951314, + "loss": 0.5872, + "step": 2620 + }, + { + "epoch": 4.3132431324313245, + "grad_norm": 0.5165495276451111, + "learning_rate": 0.00012124507263138736, + "loss": 0.5739, + "step": 2630 + }, + { + "epoch": 4.329643296432964, + "grad_norm": 0.4980062246322632, + "learning_rate": 0.00012074070751110751, + "loss": 0.5795, + "step": 2640 + }, + { + "epoch": 4.346043460434604, + "grad_norm": 0.5419421792030334, + "learning_rate": 0.00012023579045604485, + "loss": 0.5844, + "step": 2650 + }, + { + "epoch": 4.362443624436244, + "grad_norm": 0.5244302749633789, + "learning_rate": 0.00011973033490264001, + "loss": 0.5777, + "step": 2660 + }, + { + "epoch": 4.378843788437885, + "grad_norm": 0.5156043171882629, + "learning_rate": 0.0001192243543016637, + "loss": 0.5832, + "step": 2670 + }, + { + "epoch": 4.395243952439524, + "grad_norm": 0.5119035840034485, + "learning_rate": 0.00011871786211785876, + "loss": 0.59, + "step": 2680 + }, + { + "epoch": 4.411644116441164, + "grad_norm": 0.5079758167266846, + "learning_rate": 0.00011821087182958186, + "loss": 0.5652, + "step": 2690 + }, + { + "epoch": 4.428044280442805, + "grad_norm": 0.5077067613601685, + "learning_rate": 0.00011770339692844483, + "loss": 0.6014, + "step": 2700 + }, + { + "epoch": 4.444444444444445, + "grad_norm": 0.5312401056289673, + "learning_rate": 0.00011719545091895564, + "loss": 0.594, + "step": 2710 + }, + { + "epoch": 4.460844608446084, + "grad_norm": 0.5242477655410767, + "learning_rate": 0.00011668704731815892, + "loss": 0.5914, + "step": 2720 + }, + { + "epoch": 4.477244772447724, + "grad_norm": 0.48659422993659973, + "learning_rate": 0.0001161781996552765, + "loss": 0.5927, + "step": 2730 + }, + { + "epoch": 4.493644936449364, + "grad_norm": 0.5395858287811279, + "learning_rate": 0.00011566892147134705, + "loss": 0.5961, + "step": 2740 + }, + { + "epoch": 4.510045100451005, + "grad_norm": 0.503237783908844, + "learning_rate": 0.00011515922631886605, + "loss": 0.5763, + "step": 2750 + }, + { + "epoch": 4.5264452644526445, + "grad_norm": 0.5137144923210144, + "learning_rate": 0.00011464912776142494, + "loss": 0.5912, + "step": 2760 + }, + { + "epoch": 4.542845428454284, + "grad_norm": 0.5336122512817383, + "learning_rate": 0.00011413863937335028, + "loss": 0.5857, + "step": 2770 + }, + { + "epoch": 4.559245592455925, + "grad_norm": 0.5271728038787842, + "learning_rate": 0.00011362777473934248, + "loss": 0.5915, + "step": 2780 + }, + { + "epoch": 4.575645756457565, + "grad_norm": 0.5467897057533264, + "learning_rate": 0.00011311654745411425, + "loss": 0.5871, + "step": 2790 + }, + { + "epoch": 4.592045920459205, + "grad_norm": 0.5065737962722778, + "learning_rate": 0.00011260497112202895, + "loss": 0.5867, + "step": 2800 + }, + { + "epoch": 4.608446084460844, + "grad_norm": 0.5274237990379333, + "learning_rate": 0.00011209305935673844, + "loss": 0.5885, + "step": 2810 + }, + { + "epoch": 4.624846248462484, + "grad_norm": 0.5007864832878113, + "learning_rate": 0.00011158082578082089, + "loss": 0.5965, + "step": 2820 + }, + { + "epoch": 4.641246412464125, + "grad_norm": 0.48857831954956055, + "learning_rate": 0.0001110682840254182, + "loss": 0.5898, + "step": 2830 + }, + { + "epoch": 4.657646576465765, + "grad_norm": 0.5346599221229553, + "learning_rate": 0.00011055544772987335, + "loss": 0.5941, + "step": 2840 + }, + { + "epoch": 4.674046740467404, + "grad_norm": 0.4968061149120331, + "learning_rate": 0.00011004233054136725, + "loss": 0.5852, + "step": 2850 + }, + { + "epoch": 4.690446904469045, + "grad_norm": 0.5415840744972229, + "learning_rate": 0.00010952894611455584, + "loss": 0.5891, + "step": 2860 + }, + { + "epoch": 4.706847068470685, + "grad_norm": 0.5450642704963684, + "learning_rate": 0.00010901530811120655, + "loss": 0.5758, + "step": 2870 + }, + { + "epoch": 4.723247232472325, + "grad_norm": 0.5052106976509094, + "learning_rate": 0.00010850143019983474, + "loss": 0.5736, + "step": 2880 + }, + { + "epoch": 4.7396473964739645, + "grad_norm": 0.5136608481407166, + "learning_rate": 0.00010798732605534006, + "loss": 0.5769, + "step": 2890 + }, + { + "epoch": 4.756047560475604, + "grad_norm": 0.5369486808776855, + "learning_rate": 0.00010747300935864243, + "loss": 0.5762, + "step": 2900 + }, + { + "epoch": 4.772447724477245, + "grad_norm": 0.5404406189918518, + "learning_rate": 0.00010695849379631813, + "loss": 0.5862, + "step": 2910 + }, + { + "epoch": 4.788847888478885, + "grad_norm": 0.5491173267364502, + "learning_rate": 0.0001064437930602354, + "loss": 0.5783, + "step": 2920 + }, + { + "epoch": 4.805248052480525, + "grad_norm": 0.5794305801391602, + "learning_rate": 0.00010592892084719024, + "loss": 0.596, + "step": 2930 + }, + { + "epoch": 4.821648216482165, + "grad_norm": 0.5437578558921814, + "learning_rate": 0.00010541389085854176, + "loss": 0.5892, + "step": 2940 + }, + { + "epoch": 4.838048380483805, + "grad_norm": 0.5468413233757019, + "learning_rate": 0.00010489871679984777, + "loss": 0.57, + "step": 2950 + }, + { + "epoch": 4.854448544485445, + "grad_norm": 0.5444806814193726, + "learning_rate": 0.00010438341238049991, + "loss": 0.5754, + "step": 2960 + }, + { + "epoch": 4.870848708487085, + "grad_norm": 0.5279709696769714, + "learning_rate": 0.00010386799131335889, + "loss": 0.5757, + "step": 2970 + }, + { + "epoch": 4.887248872488724, + "grad_norm": 0.5229681134223938, + "learning_rate": 0.00010335246731438948, + "loss": 0.5851, + "step": 2980 + }, + { + "epoch": 4.903649036490365, + "grad_norm": 0.5356857180595398, + "learning_rate": 0.00010283685410229571, + "loss": 0.5846, + "step": 2990 + }, + { + "epoch": 4.920049200492005, + "grad_norm": 0.5306787490844727, + "learning_rate": 0.00010232116539815558, + "loss": 0.5595, + "step": 3000 + }, + { + "epoch": 4.936449364493645, + "grad_norm": 0.5103681683540344, + "learning_rate": 0.00010180541492505604, + "loss": 0.5878, + "step": 3010 + }, + { + "epoch": 4.952849528495285, + "grad_norm": 0.528273344039917, + "learning_rate": 0.00010128961640772785, + "loss": 0.5684, + "step": 3020 + }, + { + "epoch": 4.969249692496925, + "grad_norm": 0.5309355854988098, + "learning_rate": 0.00010077378357218021, + "loss": 0.5949, + "step": 3030 + }, + { + "epoch": 4.985649856498565, + "grad_norm": 0.5428647994995117, + "learning_rate": 0.00010025793014533558, + "loss": 0.5893, + "step": 3040 + }, + { + "epoch": 5.002050020500205, + "grad_norm": 0.519814133644104, + "learning_rate": 9.974206985466442e-05, + "loss": 0.5898, + "step": 3050 + }, + { + "epoch": 5.018450184501845, + "grad_norm": 0.5455579161643982, + "learning_rate": 9.92262164278198e-05, + "loss": 0.5495, + "step": 3060 + }, + { + "epoch": 5.034850348503485, + "grad_norm": 0.5572001338005066, + "learning_rate": 9.871038359227214e-05, + "loss": 0.543, + "step": 3070 + }, + { + "epoch": 5.051250512505125, + "grad_norm": 0.5723180174827576, + "learning_rate": 9.819458507494394e-05, + "loss": 0.5557, + "step": 3080 + }, + { + "epoch": 5.067650676506765, + "grad_norm": 0.5615038275718689, + "learning_rate": 9.767883460184443e-05, + "loss": 0.5505, + "step": 3090 + }, + { + "epoch": 5.0840508405084055, + "grad_norm": 0.5597648024559021, + "learning_rate": 9.71631458977043e-05, + "loss": 0.5599, + "step": 3100 + }, + { + "epoch": 5.100451004510045, + "grad_norm": 0.5660073161125183, + "learning_rate": 9.66475326856105e-05, + "loss": 0.5609, + "step": 3110 + }, + { + "epoch": 5.116851168511685, + "grad_norm": 0.5740461945533752, + "learning_rate": 9.613200868664112e-05, + "loss": 0.5505, + "step": 3120 + }, + { + "epoch": 5.133251332513325, + "grad_norm": 0.5641244649887085, + "learning_rate": 9.561658761950007e-05, + "loss": 0.5725, + "step": 3130 + }, + { + "epoch": 5.149651496514966, + "grad_norm": 0.5780383944511414, + "learning_rate": 9.510128320015224e-05, + "loss": 0.5584, + "step": 3140 + }, + { + "epoch": 5.166051660516605, + "grad_norm": 0.5865011215209961, + "learning_rate": 9.458610914145826e-05, + "loss": 0.555, + "step": 3150 + }, + { + "epoch": 5.182451824518245, + "grad_norm": 0.5646878480911255, + "learning_rate": 9.40710791528098e-05, + "loss": 0.5524, + "step": 3160 + }, + { + "epoch": 5.198851988519885, + "grad_norm": 0.6092641353607178, + "learning_rate": 9.355620693976461e-05, + "loss": 0.5704, + "step": 3170 + }, + { + "epoch": 5.215252152521526, + "grad_norm": 0.5585269927978516, + "learning_rate": 9.304150620368188e-05, + "loss": 0.5652, + "step": 3180 + }, + { + "epoch": 5.231652316523165, + "grad_norm": 0.5374350547790527, + "learning_rate": 9.252699064135758e-05, + "loss": 0.5591, + "step": 3190 + }, + { + "epoch": 5.248052480524805, + "grad_norm": 0.5892452001571655, + "learning_rate": 9.201267394465998e-05, + "loss": 0.568, + "step": 3200 + }, + { + "epoch": 5.264452644526445, + "grad_norm": 0.5627169609069824, + "learning_rate": 9.149856980016529e-05, + "loss": 0.5573, + "step": 3210 + }, + { + "epoch": 5.280852808528086, + "grad_norm": 0.6352980732917786, + "learning_rate": 9.098469188879349e-05, + "loss": 0.5548, + "step": 3220 + }, + { + "epoch": 5.2972529725297255, + "grad_norm": 0.5747233629226685, + "learning_rate": 9.047105388544417e-05, + "loss": 0.5802, + "step": 3230 + }, + { + "epoch": 5.313653136531365, + "grad_norm": 0.66510009765625, + "learning_rate": 8.995766945863277e-05, + "loss": 0.5583, + "step": 3240 + }, + { + "epoch": 5.330053300533005, + "grad_norm": 0.6103816628456116, + "learning_rate": 8.944455227012666e-05, + "loss": 0.5622, + "step": 3250 + }, + { + "epoch": 5.346453464534646, + "grad_norm": 0.53193598985672, + "learning_rate": 8.89317159745818e-05, + "loss": 0.5622, + "step": 3260 + }, + { + "epoch": 5.3628536285362856, + "grad_norm": 0.5971400737762451, + "learning_rate": 8.841917421917912e-05, + "loss": 0.5407, + "step": 3270 + }, + { + "epoch": 5.379253792537925, + "grad_norm": 0.5562443733215332, + "learning_rate": 8.790694064326157e-05, + "loss": 0.5544, + "step": 3280 + }, + { + "epoch": 5.395653956539565, + "grad_norm": 0.6092881560325623, + "learning_rate": 8.739502887797107e-05, + "loss": 0.5543, + "step": 3290 + }, + { + "epoch": 5.412054120541206, + "grad_norm": 0.5391745567321777, + "learning_rate": 8.688345254588578e-05, + "loss": 0.5553, + "step": 3300 + }, + { + "epoch": 5.428454284542846, + "grad_norm": 0.558480978012085, + "learning_rate": 8.637222526065756e-05, + "loss": 0.5614, + "step": 3310 + }, + { + "epoch": 5.444854448544485, + "grad_norm": 0.5935754776000977, + "learning_rate": 8.586136062664974e-05, + "loss": 0.5569, + "step": 3320 + }, + { + "epoch": 5.461254612546125, + "grad_norm": 0.5984821319580078, + "learning_rate": 8.535087223857508e-05, + "loss": 0.5625, + "step": 3330 + }, + { + "epoch": 5.477654776547766, + "grad_norm": 0.5659565925598145, + "learning_rate": 8.484077368113399e-05, + "loss": 0.5483, + "step": 3340 + }, + { + "epoch": 5.494054940549406, + "grad_norm": 0.5952328443527222, + "learning_rate": 8.433107852865298e-05, + "loss": 0.5546, + "step": 3350 + }, + { + "epoch": 5.5104551045510455, + "grad_norm": 0.5910452604293823, + "learning_rate": 8.382180034472353e-05, + "loss": 0.5371, + "step": 3360 + }, + { + "epoch": 5.526855268552685, + "grad_norm": 0.5623068809509277, + "learning_rate": 8.33129526818411e-05, + "loss": 0.5463, + "step": 3370 + }, + { + "epoch": 5.543255432554325, + "grad_norm": 0.557976484298706, + "learning_rate": 8.280454908104439e-05, + "loss": 0.555, + "step": 3380 + }, + { + "epoch": 5.559655596555966, + "grad_norm": 0.5730746984481812, + "learning_rate": 8.229660307155518e-05, + "loss": 0.5612, + "step": 3390 + }, + { + "epoch": 5.5760557605576055, + "grad_norm": 0.5673303604125977, + "learning_rate": 8.178912817041817e-05, + "loss": 0.5583, + "step": 3400 + }, + { + "epoch": 5.592455924559245, + "grad_norm": 0.5584287047386169, + "learning_rate": 8.128213788214126e-05, + "loss": 0.5471, + "step": 3410 + }, + { + "epoch": 5.608856088560886, + "grad_norm": 0.5875297784805298, + "learning_rate": 8.077564569833632e-05, + "loss": 0.5576, + "step": 3420 + }, + { + "epoch": 5.625256252562526, + "grad_norm": 0.5899044275283813, + "learning_rate": 8.026966509736001e-05, + "loss": 0.5437, + "step": 3430 + }, + { + "epoch": 5.641656416564166, + "grad_norm": 0.6080020666122437, + "learning_rate": 7.976420954395518e-05, + "loss": 0.5552, + "step": 3440 + }, + { + "epoch": 5.658056580565805, + "grad_norm": 0.594315230846405, + "learning_rate": 7.92592924888925e-05, + "loss": 0.5587, + "step": 3450 + }, + { + "epoch": 5.674456744567445, + "grad_norm": 0.5834687948226929, + "learning_rate": 7.875492736861266e-05, + "loss": 0.57, + "step": 3460 + }, + { + "epoch": 5.690856908569086, + "grad_norm": 0.5709730982780457, + "learning_rate": 7.825112760486861e-05, + "loss": 0.5737, + "step": 3470 + }, + { + "epoch": 5.707257072570726, + "grad_norm": 0.5856568217277527, + "learning_rate": 7.774790660436858e-05, + "loss": 0.5484, + "step": 3480 + }, + { + "epoch": 5.7236572365723655, + "grad_norm": 0.5535131692886353, + "learning_rate": 7.724527775841914e-05, + "loss": 0.5483, + "step": 3490 + }, + { + "epoch": 5.740057400574006, + "grad_norm": 0.5895147323608398, + "learning_rate": 7.674325444256899e-05, + "loss": 0.5463, + "step": 3500 + }, + { + "epoch": 5.756457564575646, + "grad_norm": 0.57289719581604, + "learning_rate": 7.624185001625292e-05, + "loss": 0.5569, + "step": 3510 + }, + { + "epoch": 5.772857728577286, + "grad_norm": 0.5502966046333313, + "learning_rate": 7.574107782243634e-05, + "loss": 0.5696, + "step": 3520 + }, + { + "epoch": 5.7892578925789255, + "grad_norm": 0.5804471969604492, + "learning_rate": 7.524095118726025e-05, + "loss": 0.5506, + "step": 3530 + }, + { + "epoch": 5.805658056580565, + "grad_norm": 0.5519583225250244, + "learning_rate": 7.474148341968652e-05, + "loss": 0.5571, + "step": 3540 + }, + { + "epoch": 5.822058220582206, + "grad_norm": 0.5806939005851746, + "learning_rate": 7.42426878111438e-05, + "loss": 0.5556, + "step": 3550 + }, + { + "epoch": 5.838458384583846, + "grad_norm": 0.5681191682815552, + "learning_rate": 7.374457763517376e-05, + "loss": 0.5557, + "step": 3560 + }, + { + "epoch": 5.854858548585486, + "grad_norm": 0.6004670262336731, + "learning_rate": 7.324716614707793e-05, + "loss": 0.5686, + "step": 3570 + }, + { + "epoch": 5.871258712587126, + "grad_norm": 0.5577644109725952, + "learning_rate": 7.275046658356494e-05, + "loss": 0.5619, + "step": 3580 + }, + { + "epoch": 5.887658876588766, + "grad_norm": 0.5934208631515503, + "learning_rate": 7.225449216239821e-05, + "loss": 0.5593, + "step": 3590 + }, + { + "epoch": 5.904059040590406, + "grad_norm": 0.5490043759346008, + "learning_rate": 7.175925608204428e-05, + "loss": 0.5404, + "step": 3600 + }, + { + "epoch": 5.920459204592046, + "grad_norm": 0.662482500076294, + "learning_rate": 7.126477152132164e-05, + "loss": 0.5525, + "step": 3610 + }, + { + "epoch": 5.9368593685936855, + "grad_norm": 0.5766283869743347, + "learning_rate": 7.077105163904987e-05, + "loss": 0.5706, + "step": 3620 + }, + { + "epoch": 5.953259532595326, + "grad_norm": 0.5751728415489197, + "learning_rate": 7.027810957369957e-05, + "loss": 0.5626, + "step": 3630 + }, + { + "epoch": 5.969659696596966, + "grad_norm": 0.5690221190452576, + "learning_rate": 6.978595844304271e-05, + "loss": 0.558, + "step": 3640 + }, + { + "epoch": 5.986059860598606, + "grad_norm": 0.5637174248695374, + "learning_rate": 6.92946113438036e-05, + "loss": 0.5665, + "step": 3650 + }, + { + "epoch": 6.002460024600246, + "grad_norm": 0.5749226808547974, + "learning_rate": 6.880408135131022e-05, + "loss": 0.5687, + "step": 3660 + }, + { + "epoch": 6.018860188601886, + "grad_norm": 0.581462562084198, + "learning_rate": 6.831438151914649e-05, + "loss": 0.5272, + "step": 3670 + }, + { + "epoch": 6.035260352603526, + "grad_norm": 0.585290253162384, + "learning_rate": 6.782552487880468e-05, + "loss": 0.5347, + "step": 3680 + }, + { + "epoch": 6.051660516605166, + "grad_norm": 0.6168552041053772, + "learning_rate": 6.733752443933878e-05, + "loss": 0.5224, + "step": 3690 + }, + { + "epoch": 6.0680606806068065, + "grad_norm": 0.5952621698379517, + "learning_rate": 6.685039318701826e-05, + "loss": 0.5267, + "step": 3700 + }, + { + "epoch": 6.084460844608446, + "grad_norm": 0.708752453327179, + "learning_rate": 6.636414408498249e-05, + "loss": 0.5323, + "step": 3710 + }, + { + "epoch": 6.100861008610086, + "grad_norm": 0.6072937250137329, + "learning_rate": 6.587879007289576e-05, + "loss": 0.5361, + "step": 3720 + }, + { + "epoch": 6.117261172611726, + "grad_norm": 0.6426242589950562, + "learning_rate": 6.539434406660296e-05, + "loss": 0.5439, + "step": 3730 + }, + { + "epoch": 6.1336613366133665, + "grad_norm": 0.6066299080848694, + "learning_rate": 6.491081895778588e-05, + "loss": 0.5256, + "step": 3740 + }, + { + "epoch": 6.150061500615006, + "grad_norm": 0.578899621963501, + "learning_rate": 6.442822761362015e-05, + "loss": 0.5373, + "step": 3750 + }, + { + "epoch": 6.166461664616646, + "grad_norm": 0.6230922341346741, + "learning_rate": 6.394658287643278e-05, + "loss": 0.5249, + "step": 3760 + }, + { + "epoch": 6.182861828618286, + "grad_norm": 0.610723614692688, + "learning_rate": 6.34658975633605e-05, + "loss": 0.5417, + "step": 3770 + }, + { + "epoch": 6.199261992619927, + "grad_norm": 0.6284223198890686, + "learning_rate": 6.298618446600856e-05, + "loss": 0.5361, + "step": 3780 + }, + { + "epoch": 6.215662156621566, + "grad_norm": 0.6320038437843323, + "learning_rate": 6.250745635011048e-05, + "loss": 0.5344, + "step": 3790 + }, + { + "epoch": 6.232062320623206, + "grad_norm": 0.5748194456100464, + "learning_rate": 6.202972595518817e-05, + "loss": 0.5285, + "step": 3800 + }, + { + "epoch": 6.248462484624846, + "grad_norm": 0.6720646023750305, + "learning_rate": 6.155300599421306e-05, + "loss": 0.5399, + "step": 3810 + }, + { + "epoch": 6.264862648626487, + "grad_norm": 0.6088182330131531, + "learning_rate": 6.107730915326772e-05, + "loss": 0.5387, + "step": 3820 + }, + { + "epoch": 6.2812628126281265, + "grad_norm": 0.6509405970573425, + "learning_rate": 6.0602648091208324e-05, + "loss": 0.5356, + "step": 3830 + }, + { + "epoch": 6.297662976629766, + "grad_norm": 0.6377142667770386, + "learning_rate": 6.012903543932766e-05, + "loss": 0.5328, + "step": 3840 + }, + { + "epoch": 6.314063140631406, + "grad_norm": 0.638660728931427, + "learning_rate": 5.965648380101916e-05, + "loss": 0.5403, + "step": 3850 + }, + { + "epoch": 6.330463304633047, + "grad_norm": 0.6343023180961609, + "learning_rate": 5.918500575144138e-05, + "loss": 0.5316, + "step": 3860 + }, + { + "epoch": 6.3468634686346865, + "grad_norm": 0.6120122075080872, + "learning_rate": 5.871461383718344e-05, + "loss": 0.5475, + "step": 3870 + }, + { + "epoch": 6.363263632636326, + "grad_norm": 0.664526104927063, + "learning_rate": 5.8245320575931085e-05, + "loss": 0.5539, + "step": 3880 + }, + { + "epoch": 6.379663796637966, + "grad_norm": 0.6195924878120422, + "learning_rate": 5.777713845613364e-05, + "loss": 0.5314, + "step": 3890 + }, + { + "epoch": 6.396063960639607, + "grad_norm": 0.6664077639579773, + "learning_rate": 5.7310079936671545e-05, + "loss": 0.5309, + "step": 3900 + }, + { + "epoch": 6.412464124641247, + "grad_norm": 0.6408997178077698, + "learning_rate": 5.684415744652509e-05, + "loss": 0.5346, + "step": 3910 + }, + { + "epoch": 6.428864288642886, + "grad_norm": 0.596552312374115, + "learning_rate": 5.6379383384443255e-05, + "loss": 0.5424, + "step": 3920 + }, + { + "epoch": 6.445264452644526, + "grad_norm": 0.6852882504463196, + "learning_rate": 5.59157701186142e-05, + "loss": 0.5367, + "step": 3930 + }, + { + "epoch": 6.461664616646167, + "grad_norm": 0.6832631826400757, + "learning_rate": 5.545332998633572e-05, + "loss": 0.5311, + "step": 3940 + }, + { + "epoch": 6.478064780647807, + "grad_norm": 0.6201120018959045, + "learning_rate": 5.499207529368734e-05, + "loss": 0.5398, + "step": 3950 + }, + { + "epoch": 6.4944649446494465, + "grad_norm": 0.6140998601913452, + "learning_rate": 5.453201831520245e-05, + "loss": 0.5388, + "step": 3960 + }, + { + "epoch": 6.510865108651086, + "grad_norm": 0.6650799512863159, + "learning_rate": 5.4073171293542016e-05, + "loss": 0.5419, + "step": 3970 + }, + { + "epoch": 6.527265272652727, + "grad_norm": 0.6716285347938538, + "learning_rate": 5.3615546439168485e-05, + "loss": 0.5234, + "step": 3980 + }, + { + "epoch": 6.543665436654367, + "grad_norm": 0.6595677733421326, + "learning_rate": 5.3159155930021e-05, + "loss": 0.5194, + "step": 3990 + }, + { + "epoch": 6.5600656006560065, + "grad_norm": 0.5784692168235779, + "learning_rate": 5.270401191119143e-05, + "loss": 0.5222, + "step": 4000 + }, + { + "epoch": 6.576465764657646, + "grad_norm": 0.6419820189476013, + "learning_rate": 5.2250126494600916e-05, + "loss": 0.5352, + "step": 4010 + }, + { + "epoch": 6.592865928659286, + "grad_norm": 0.6142451763153076, + "learning_rate": 5.179751175867784e-05, + "loss": 0.5357, + "step": 4020 + }, + { + "epoch": 6.609266092660927, + "grad_norm": 0.6860802173614502, + "learning_rate": 5.1346179748036116e-05, + "loss": 0.531, + "step": 4030 + }, + { + "epoch": 6.625666256662567, + "grad_norm": 0.6106059551239014, + "learning_rate": 5.0896142473154987e-05, + "loss": 0.5333, + "step": 4040 + }, + { + "epoch": 6.642066420664206, + "grad_norm": 0.6644913554191589, + "learning_rate": 5.044741191005908e-05, + "loss": 0.5403, + "step": 4050 + }, + { + "epoch": 6.658466584665847, + "grad_norm": 0.5942371487617493, + "learning_rate": 5.000000000000002e-05, + "loss": 0.5467, + "step": 4060 + }, + { + "epoch": 6.674866748667487, + "grad_norm": 0.617970883846283, + "learning_rate": 4.9553918649138386e-05, + "loss": 0.5325, + "step": 4070 + }, + { + "epoch": 6.691266912669127, + "grad_norm": 0.6382017135620117, + "learning_rate": 4.910917972822713e-05, + "loss": 0.525, + "step": 4080 + }, + { + "epoch": 6.7076670766707664, + "grad_norm": 0.6390166282653809, + "learning_rate": 4.866579507229545e-05, + "loss": 0.5356, + "step": 4090 + }, + { + "epoch": 6.724067240672406, + "grad_norm": 0.6630433201789856, + "learning_rate": 4.822377648033394e-05, + "loss": 0.541, + "step": 4100 + }, + { + "epoch": 6.740467404674047, + "grad_norm": 0.6468001008033752, + "learning_rate": 4.7783135714980744e-05, + "loss": 0.5399, + "step": 4110 + }, + { + "epoch": 6.756867568675687, + "grad_norm": 0.620833694934845, + "learning_rate": 4.734388450220825e-05, + "loss": 0.5265, + "step": 4120 + }, + { + "epoch": 6.7732677326773265, + "grad_norm": 0.61692875623703, + "learning_rate": 4.6906034531011346e-05, + "loss": 0.5303, + "step": 4130 + }, + { + "epoch": 6.789667896678967, + "grad_norm": 0.6430515646934509, + "learning_rate": 4.646959745309609e-05, + "loss": 0.543, + "step": 4140 + }, + { + "epoch": 6.806068060680607, + "grad_norm": 0.653596043586731, + "learning_rate": 4.603458488256992e-05, + "loss": 0.5384, + "step": 4150 + }, + { + "epoch": 6.822468224682247, + "grad_norm": 0.6125525832176208, + "learning_rate": 4.560100839563229e-05, + "loss": 0.5417, + "step": 4160 + }, + { + "epoch": 6.838868388683887, + "grad_norm": 0.5968248844146729, + "learning_rate": 4.516887953026691e-05, + "loss": 0.5326, + "step": 4170 + }, + { + "epoch": 6.855268552685526, + "grad_norm": 0.5980390906333923, + "learning_rate": 4.4738209785934505e-05, + "loss": 0.5313, + "step": 4180 + }, + { + "epoch": 6.871668716687167, + "grad_norm": 0.614548921585083, + "learning_rate": 4.430901062326681e-05, + "loss": 0.5321, + "step": 4190 + }, + { + "epoch": 6.888068880688807, + "grad_norm": 0.5848116278648376, + "learning_rate": 4.388129346376178e-05, + "loss": 0.542, + "step": 4200 + }, + { + "epoch": 6.904469044690447, + "grad_norm": 0.6172247529029846, + "learning_rate": 4.345506968947931e-05, + "loss": 0.5439, + "step": 4210 + }, + { + "epoch": 6.920869208692087, + "grad_norm": 0.6525529026985168, + "learning_rate": 4.303035064273878e-05, + "loss": 0.5319, + "step": 4220 + }, + { + "epoch": 6.937269372693727, + "grad_norm": 0.614452064037323, + "learning_rate": 4.260714762581677e-05, + "loss": 0.5287, + "step": 4230 + }, + { + "epoch": 6.953669536695367, + "grad_norm": 0.6268100142478943, + "learning_rate": 4.21854719006467e-05, + "loss": 0.5337, + "step": 4240 + }, + { + "epoch": 6.970069700697007, + "grad_norm": 0.6285504698753357, + "learning_rate": 4.1765334688518766e-05, + "loss": 0.5402, + "step": 4250 + }, + { + "epoch": 6.9864698646986465, + "grad_norm": 0.6280767917633057, + "learning_rate": 4.13467471697817e-05, + "loss": 0.533, + "step": 4260 + }, + { + "epoch": 7.002870028700287, + "grad_norm": 0.5814547538757324, + "learning_rate": 4.092972048354491e-05, + "loss": 0.5346, + "step": 4270 + }, + { + "epoch": 7.019270192701927, + "grad_norm": 0.6583351492881775, + "learning_rate": 4.0514265727382215e-05, + "loss": 0.5261, + "step": 4280 + }, + { + "epoch": 7.035670356703567, + "grad_norm": 0.6630005240440369, + "learning_rate": 4.010039395703664e-05, + "loss": 0.5196, + "step": 4290 + }, + { + "epoch": 7.0520705207052075, + "grad_norm": 0.6557034254074097, + "learning_rate": 3.968811618612592e-05, + "loss": 0.5161, + "step": 4300 + }, + { + "epoch": 7.068470684706847, + "grad_norm": 0.6600791215896606, + "learning_rate": 3.927744338584972e-05, + "loss": 0.5066, + "step": 4310 + }, + { + "epoch": 7.084870848708487, + "grad_norm": 0.6123402118682861, + "learning_rate": 3.8868386484697417e-05, + "loss": 0.5195, + "step": 4320 + }, + { + "epoch": 7.101271012710127, + "grad_norm": 0.6454523801803589, + "learning_rate": 3.84609563681575e-05, + "loss": 0.5113, + "step": 4330 + }, + { + "epoch": 7.1176711767117675, + "grad_norm": 0.6421491503715515, + "learning_rate": 3.80551638784277e-05, + "loss": 0.5134, + "step": 4340 + }, + { + "epoch": 7.134071340713407, + "grad_norm": 0.6418097615242004, + "learning_rate": 3.7651019814126654e-05, + "loss": 0.5185, + "step": 4350 + }, + { + "epoch": 7.150471504715047, + "grad_norm": 0.6425509452819824, + "learning_rate": 3.724853493000635e-05, + "loss": 0.5251, + "step": 4360 + }, + { + "epoch": 7.166871668716687, + "grad_norm": 0.6572188138961792, + "learning_rate": 3.6847719936666124e-05, + "loss": 0.5124, + "step": 4370 + }, + { + "epoch": 7.183271832718328, + "grad_norm": 0.6694904565811157, + "learning_rate": 3.6448585500267485e-05, + "loss": 0.5174, + "step": 4380 + }, + { + "epoch": 7.199671996719967, + "grad_norm": 0.6733431816101074, + "learning_rate": 3.605114224225028e-05, + "loss": 0.5227, + "step": 4390 + }, + { + "epoch": 7.216072160721607, + "grad_norm": 0.6532800793647766, + "learning_rate": 3.565540073905025e-05, + "loss": 0.5128, + "step": 4400 + }, + { + "epoch": 7.232472324723247, + "grad_norm": 0.6475611329078674, + "learning_rate": 3.5261371521817244e-05, + "loss": 0.5282, + "step": 4410 + }, + { + "epoch": 7.248872488724888, + "grad_norm": 0.6134998202323914, + "learning_rate": 3.486906507613531e-05, + "loss": 0.5118, + "step": 4420 + }, + { + "epoch": 7.2652726527265274, + "grad_norm": 0.648802638053894, + "learning_rate": 3.4478491841743397e-05, + "loss": 0.5253, + "step": 4430 + }, + { + "epoch": 7.281672816728167, + "grad_norm": 0.6617629528045654, + "learning_rate": 3.408966221225773e-05, + "loss": 0.5177, + "step": 4440 + }, + { + "epoch": 7.298072980729807, + "grad_norm": 0.6465151906013489, + "learning_rate": 3.370258653489505e-05, + "loss": 0.5133, + "step": 4450 + }, + { + "epoch": 7.314473144731448, + "grad_norm": 0.6327465176582336, + "learning_rate": 3.331727511019749e-05, + "loss": 0.5155, + "step": 4460 + }, + { + "epoch": 7.3308733087330875, + "grad_norm": 0.6604763865470886, + "learning_rate": 3.293373819175816e-05, + "loss": 0.519, + "step": 4470 + }, + { + "epoch": 7.347273472734727, + "grad_norm": 0.681736171245575, + "learning_rate": 3.2551985985948616e-05, + "loss": 0.5243, + "step": 4480 + }, + { + "epoch": 7.363673636736367, + "grad_norm": 0.6531623601913452, + "learning_rate": 3.217202865164697e-05, + "loss": 0.5201, + "step": 4490 + }, + { + "epoch": 7.380073800738008, + "grad_norm": 0.674201250076294, + "learning_rate": 3.1793876299967816e-05, + "loss": 0.5252, + "step": 4500 + }, + { + "epoch": 7.396473964739648, + "grad_norm": 0.6714997291564941, + "learning_rate": 3.141753899399289e-05, + "loss": 0.5253, + "step": 4510 + }, + { + "epoch": 7.412874128741287, + "grad_norm": 0.6872825622558594, + "learning_rate": 3.104302674850346e-05, + "loss": 0.5311, + "step": 4520 + }, + { + "epoch": 7.429274292742927, + "grad_norm": 0.7338685989379883, + "learning_rate": 3.0670349529713816e-05, + "loss": 0.515, + "step": 4530 + }, + { + "epoch": 7.445674456744568, + "grad_norm": 0.6552906036376953, + "learning_rate": 3.0299517255005937e-05, + "loss": 0.5077, + "step": 4540 + }, + { + "epoch": 7.462074620746208, + "grad_norm": 0.6388695240020752, + "learning_rate": 2.993053979266577e-05, + "loss": 0.5296, + "step": 4550 + }, + { + "epoch": 7.478474784747847, + "grad_norm": 0.6324784159660339, + "learning_rate": 2.9563426961620367e-05, + "loss": 0.5249, + "step": 4560 + }, + { + "epoch": 7.494874948749487, + "grad_norm": 0.6526088714599609, + "learning_rate": 2.9198188531176863e-05, + "loss": 0.5164, + "step": 4570 + }, + { + "epoch": 7.511275112751127, + "grad_norm": 0.6632060408592224, + "learning_rate": 2.883483422076225e-05, + "loss": 0.5185, + "step": 4580 + }, + { + "epoch": 7.527675276752768, + "grad_norm": 0.643341600894928, + "learning_rate": 2.8473373699664997e-05, + "loss": 0.5162, + "step": 4590 + }, + { + "epoch": 7.5440754407544075, + "grad_norm": 0.6466885805130005, + "learning_rate": 2.811381658677744e-05, + "loss": 0.5103, + "step": 4600 + }, + { + "epoch": 7.560475604756047, + "grad_norm": 0.6679097414016724, + "learning_rate": 2.7756172450340134e-05, + "loss": 0.5238, + "step": 4610 + }, + { + "epoch": 7.576875768757688, + "grad_norm": 0.6607965230941772, + "learning_rate": 2.7400450807686938e-05, + "loss": 0.5104, + "step": 4620 + }, + { + "epoch": 7.593275932759328, + "grad_norm": 0.64109867811203, + "learning_rate": 2.70466611249919e-05, + "loss": 0.5199, + "step": 4630 + }, + { + "epoch": 7.609676096760968, + "grad_norm": 0.6699521541595459, + "learning_rate": 2.669481281701739e-05, + "loss": 0.5339, + "step": 4640 + }, + { + "epoch": 7.626076260762607, + "grad_norm": 0.6698071360588074, + "learning_rate": 2.6344915246863412e-05, + "loss": 0.5128, + "step": 4650 + }, + { + "epoch": 7.642476424764247, + "grad_norm": 0.6424985527992249, + "learning_rate": 2.5996977725718607e-05, + "loss": 0.5053, + "step": 4660 + }, + { + "epoch": 7.658876588765888, + "grad_norm": 0.7101566195487976, + "learning_rate": 2.5651009512612312e-05, + "loss": 0.5311, + "step": 4670 + }, + { + "epoch": 7.675276752767528, + "grad_norm": 0.6432877779006958, + "learning_rate": 2.5307019814168342e-05, + "loss": 0.5216, + "step": 4680 + }, + { + "epoch": 7.691676916769167, + "grad_norm": 0.6850785613059998, + "learning_rate": 2.496501778435977e-05, + "loss": 0.5108, + "step": 4690 + }, + { + "epoch": 7.708077080770808, + "grad_norm": 0.6734909415245056, + "learning_rate": 2.462501252426559e-05, + "loss": 0.5186, + "step": 4700 + }, + { + "epoch": 7.724477244772448, + "grad_norm": 0.7187743782997131, + "learning_rate": 2.4287013081828257e-05, + "loss": 0.5182, + "step": 4710 + }, + { + "epoch": 7.740877408774088, + "grad_norm": 0.7147987484931946, + "learning_rate": 2.3951028451613144e-05, + "loss": 0.5193, + "step": 4720 + }, + { + "epoch": 7.7572775727757275, + "grad_norm": 0.7066251635551453, + "learning_rate": 2.3617067574569087e-05, + "loss": 0.5175, + "step": 4730 + }, + { + "epoch": 7.773677736777367, + "grad_norm": 0.7206938862800598, + "learning_rate": 2.328513933779034e-05, + "loss": 0.5255, + "step": 4740 + }, + { + "epoch": 7.790077900779008, + "grad_norm": 0.6903772354125977, + "learning_rate": 2.2955252574280328e-05, + "loss": 0.512, + "step": 4750 + }, + { + "epoch": 7.806478064780648, + "grad_norm": 0.6938795447349548, + "learning_rate": 2.2627416062716366e-05, + "loss": 0.513, + "step": 4760 + }, + { + "epoch": 7.822878228782288, + "grad_norm": 0.6563030481338501, + "learning_rate": 2.2301638527216194e-05, + "loss": 0.5255, + "step": 4770 + }, + { + "epoch": 7.839278392783928, + "grad_norm": 0.676189124584198, + "learning_rate": 2.1977928637105692e-05, + "loss": 0.5331, + "step": 4780 + }, + { + "epoch": 7.855678556785568, + "grad_norm": 0.6740292906761169, + "learning_rate": 2.1656295006688353e-05, + "loss": 0.5161, + "step": 4790 + }, + { + "epoch": 7.872078720787208, + "grad_norm": 0.6368678212165833, + "learning_rate": 2.1336746195015846e-05, + "loss": 0.5181, + "step": 4800 + }, + { + "epoch": 7.888478884788848, + "grad_norm": 0.6666802763938904, + "learning_rate": 2.1019290705660356e-05, + "loss": 0.5248, + "step": 4810 + }, + { + "epoch": 7.904879048790487, + "grad_norm": 0.7021058797836304, + "learning_rate": 2.070393698648836e-05, + "loss": 0.5076, + "step": 4820 + }, + { + "epoch": 7.921279212792128, + "grad_norm": 0.6414440870285034, + "learning_rate": 2.0390693429435627e-05, + "loss": 0.5091, + "step": 4830 + }, + { + "epoch": 7.937679376793768, + "grad_norm": 0.6494581699371338, + "learning_rate": 2.0079568370284128e-05, + "loss": 0.5205, + "step": 4840 + }, + { + "epoch": 7.954079540795408, + "grad_norm": 0.6568921208381653, + "learning_rate": 1.977057008844e-05, + "loss": 0.5161, + "step": 4850 + }, + { + "epoch": 7.970479704797048, + "grad_norm": 0.6623067855834961, + "learning_rate": 1.946370680671341e-05, + "loss": 0.5134, + "step": 4860 + }, + { + "epoch": 7.986879868798688, + "grad_norm": 0.6465177536010742, + "learning_rate": 1.9158986691099544e-05, + "loss": 0.5097, + "step": 4870 + }, + { + "epoch": 8.003280032800328, + "grad_norm": 0.6603899002075195, + "learning_rate": 1.885641785056149e-05, + "loss": 0.5152, + "step": 4880 + }, + { + "epoch": 8.019680196801968, + "grad_norm": 0.6793326735496521, + "learning_rate": 1.85560083368143e-05, + "loss": 0.5201, + "step": 4890 + }, + { + "epoch": 8.036080360803608, + "grad_norm": 0.6705737709999084, + "learning_rate": 1.825776614411082e-05, + "loss": 0.4945, + "step": 4900 + }, + { + "epoch": 8.052480524805247, + "grad_norm": 0.6642212271690369, + "learning_rate": 1.7961699209028905e-05, + "loss": 0.5019, + "step": 4910 + }, + { + "epoch": 8.068880688806889, + "grad_norm": 0.6646486520767212, + "learning_rate": 1.766781541026018e-05, + "loss": 0.5245, + "step": 4920 + }, + { + "epoch": 8.085280852808529, + "grad_norm": 0.6514326930046082, + "learning_rate": 1.7376122568400532e-05, + "loss": 0.5083, + "step": 4930 + }, + { + "epoch": 8.101681016810169, + "grad_norm": 0.6762531995773315, + "learning_rate": 1.708662844574178e-05, + "loss": 0.5144, + "step": 4940 + }, + { + "epoch": 8.118081180811808, + "grad_norm": 0.6511978507041931, + "learning_rate": 1.679934074606533e-05, + "loss": 0.5045, + "step": 4950 + }, + { + "epoch": 8.134481344813448, + "grad_norm": 0.6896832585334778, + "learning_rate": 1.6514267114436945e-05, + "loss": 0.5089, + "step": 4960 + }, + { + "epoch": 8.150881508815088, + "grad_norm": 0.6686916947364807, + "learning_rate": 1.6231415137003537e-05, + "loss": 0.5048, + "step": 4970 + }, + { + "epoch": 8.167281672816728, + "grad_norm": 0.6836830377578735, + "learning_rate": 1.5950792340791043e-05, + "loss": 0.5072, + "step": 4980 + }, + { + "epoch": 8.183681836818367, + "grad_norm": 0.6605989336967468, + "learning_rate": 1.5672406193504384e-05, + "loss": 0.5115, + "step": 4990 + }, + { + "epoch": 8.200082000820009, + "grad_norm": 0.6973133087158203, + "learning_rate": 1.5396264103328474e-05, + "loss": 0.5132, + "step": 5000 + }, + { + "epoch": 8.216482164821649, + "grad_norm": 0.7265208959579468, + "learning_rate": 1.5122373418731306e-05, + "loss": 0.5036, + "step": 5010 + }, + { + "epoch": 8.232882328823289, + "grad_norm": 0.6954894661903381, + "learning_rate": 1.4850741428268244e-05, + "loss": 0.5037, + "step": 5020 + }, + { + "epoch": 8.249282492824928, + "grad_norm": 0.655786395072937, + "learning_rate": 1.4581375360388183e-05, + "loss": 0.515, + "step": 5030 + }, + { + "epoch": 8.265682656826568, + "grad_norm": 0.674171507358551, + "learning_rate": 1.4314282383241096e-05, + "loss": 0.4967, + "step": 5040 + }, + { + "epoch": 8.282082820828208, + "grad_norm": 0.7670585513114929, + "learning_rate": 1.4049469604487297e-05, + "loss": 0.4991, + "step": 5050 + }, + { + "epoch": 8.298482984829848, + "grad_norm": 0.6614052057266235, + "learning_rate": 1.3786944071108398e-05, + "loss": 0.5169, + "step": 5060 + }, + { + "epoch": 8.314883148831488, + "grad_norm": 0.7165561318397522, + "learning_rate": 1.3526712769219618e-05, + "loss": 0.502, + "step": 5070 + }, + { + "epoch": 8.33128331283313, + "grad_norm": 0.7127593159675598, + "learning_rate": 1.3268782623884047e-05, + "loss": 0.5091, + "step": 5080 + }, + { + "epoch": 8.347683476834769, + "grad_norm": 0.6743724346160889, + "learning_rate": 1.301316049892818e-05, + "loss": 0.5019, + "step": 5090 + }, + { + "epoch": 8.364083640836409, + "grad_norm": 0.6609333157539368, + "learning_rate": 1.2759853196759453e-05, + "loss": 0.5052, + "step": 5100 + }, + { + "epoch": 8.380483804838049, + "grad_norm": 0.7156481146812439, + "learning_rate": 1.2508867458185037e-05, + "loss": 0.51, + "step": 5110 + }, + { + "epoch": 8.396883968839688, + "grad_norm": 0.7323744893074036, + "learning_rate": 1.2260209962232628e-05, + "loss": 0.5092, + "step": 5120 + }, + { + "epoch": 8.413284132841328, + "grad_norm": 0.6540841460227966, + "learning_rate": 1.201388732597255e-05, + "loss": 0.507, + "step": 5130 + }, + { + "epoch": 8.429684296842968, + "grad_norm": 0.6970362663269043, + "learning_rate": 1.1769906104341832e-05, + "loss": 0.5216, + "step": 5140 + }, + { + "epoch": 8.446084460844608, + "grad_norm": 0.6555099487304688, + "learning_rate": 1.1528272789969618e-05, + "loss": 0.5001, + "step": 5150 + }, + { + "epoch": 8.46248462484625, + "grad_norm": 0.6940327286720276, + "learning_rate": 1.1288993813004467e-05, + "loss": 0.5025, + "step": 5160 + }, + { + "epoch": 8.478884788847889, + "grad_norm": 0.6825557351112366, + "learning_rate": 1.1052075540943296e-05, + "loss": 0.5089, + "step": 5170 + }, + { + "epoch": 8.495284952849529, + "grad_norm": 0.6553847789764404, + "learning_rate": 1.0817524278461776e-05, + "loss": 0.5052, + "step": 5180 + }, + { + "epoch": 8.511685116851169, + "grad_norm": 0.6892216205596924, + "learning_rate": 1.0585346267246743e-05, + "loss": 0.5158, + "step": 5190 + }, + { + "epoch": 8.528085280852808, + "grad_norm": 0.6756864786148071, + "learning_rate": 1.0355547685829926e-05, + "loss": 0.5133, + "step": 5200 + }, + { + "epoch": 8.544485444854448, + "grad_norm": 0.7741680145263672, + "learning_rate": 1.0128134649423671e-05, + "loss": 0.5167, + "step": 5210 + }, + { + "epoch": 8.560885608856088, + "grad_norm": 0.6552737355232239, + "learning_rate": 9.903113209758096e-06, + "loss": 0.5094, + "step": 5220 + }, + { + "epoch": 8.577285772857728, + "grad_norm": 0.6564160585403442, + "learning_rate": 9.680489354920152e-06, + "loss": 0.5048, + "step": 5230 + }, + { + "epoch": 8.59368593685937, + "grad_norm": 0.7316927313804626, + "learning_rate": 9.460269009194167e-06, + "loss": 0.499, + "step": 5240 + }, + { + "epoch": 8.61008610086101, + "grad_norm": 0.7389256358146667, + "learning_rate": 9.242458032904311e-06, + "loss": 0.5022, + "step": 5250 + }, + { + "epoch": 8.626486264862649, + "grad_norm": 0.6633957624435425, + "learning_rate": 9.027062222258487e-06, + "loss": 0.5042, + "step": 5260 + }, + { + "epoch": 8.642886428864289, + "grad_norm": 0.640006422996521, + "learning_rate": 8.814087309194251e-06, + "loss": 0.5068, + "step": 5270 + }, + { + "epoch": 8.659286592865929, + "grad_norm": 0.6830674409866333, + "learning_rate": 8.603538961226232e-06, + "loss": 0.502, + "step": 5280 + }, + { + "epoch": 8.675686756867568, + "grad_norm": 0.6639739871025085, + "learning_rate": 8.395422781295192e-06, + "loss": 0.5177, + "step": 5290 + }, + { + "epoch": 8.692086920869208, + "grad_norm": 0.6661989092826843, + "learning_rate": 8.189744307619118e-06, + "loss": 0.5014, + "step": 5300 + }, + { + "epoch": 8.708487084870848, + "grad_norm": 0.6645560264587402, + "learning_rate": 7.986509013545673e-06, + "loss": 0.5017, + "step": 5310 + }, + { + "epoch": 8.724887248872488, + "grad_norm": 0.7042533159255981, + "learning_rate": 7.785722307406684e-06, + "loss": 0.5163, + "step": 5320 + }, + { + "epoch": 8.74128741287413, + "grad_norm": 0.698277473449707, + "learning_rate": 7.587389532374123e-06, + "loss": 0.5054, + "step": 5330 + }, + { + "epoch": 8.75768757687577, + "grad_norm": 0.6191078424453735, + "learning_rate": 7.3915159663179075e-06, + "loss": 0.4992, + "step": 5340 + }, + { + "epoch": 8.774087740877409, + "grad_norm": 0.6769982576370239, + "learning_rate": 7.198106821665585e-06, + "loss": 0.5147, + "step": 5350 + }, + { + "epoch": 8.790487904879049, + "grad_norm": 0.6643761396408081, + "learning_rate": 7.007167245263435e-06, + "loss": 0.5105, + "step": 5360 + }, + { + "epoch": 8.806888068880689, + "grad_norm": 0.6664229035377502, + "learning_rate": 6.818702318239689e-06, + "loss": 0.5021, + "step": 5370 + }, + { + "epoch": 8.823288232882328, + "grad_norm": 0.6411841511726379, + "learning_rate": 6.632717055869164e-06, + "loss": 0.5076, + "step": 5380 + }, + { + "epoch": 8.839688396883968, + "grad_norm": 0.6643224954605103, + "learning_rate": 6.4492164074399065e-06, + "loss": 0.5044, + "step": 5390 + }, + { + "epoch": 8.85608856088561, + "grad_norm": 0.6722341775894165, + "learning_rate": 6.268205256121396e-06, + "loss": 0.5092, + "step": 5400 + }, + { + "epoch": 8.87248872488725, + "grad_norm": 0.6829484105110168, + "learning_rate": 6.089688418834727e-06, + "loss": 0.5164, + "step": 5410 + }, + { + "epoch": 8.88888888888889, + "grad_norm": 0.7037432789802551, + "learning_rate": 5.913670646124236e-06, + "loss": 0.5142, + "step": 5420 + }, + { + "epoch": 8.905289052890529, + "grad_norm": 0.6822744607925415, + "learning_rate": 5.7401566220313005e-06, + "loss": 0.5018, + "step": 5430 + }, + { + "epoch": 8.921689216892169, + "grad_norm": 0.6738746166229248, + "learning_rate": 5.569150963969494e-06, + "loss": 0.5013, + "step": 5440 + }, + { + "epoch": 8.938089380893809, + "grad_norm": 0.6965447664260864, + "learning_rate": 5.400658222601873e-06, + "loss": 0.5014, + "step": 5450 + }, + { + "epoch": 8.954489544895448, + "grad_norm": 0.7140269875526428, + "learning_rate": 5.2346828817197655e-06, + "loss": 0.5119, + "step": 5460 + }, + { + "epoch": 8.970889708897088, + "grad_norm": 0.6803216934204102, + "learning_rate": 5.071229358123464e-06, + "loss": 0.5021, + "step": 5470 + }, + { + "epoch": 8.987289872898728, + "grad_norm": 0.6827109456062317, + "learning_rate": 4.910302001504807e-06, + "loss": 0.5157, + "step": 5480 + }, + { + "epoch": 9.00369003690037, + "grad_norm": 0.6659403443336487, + "learning_rate": 4.7519050943312325e-06, + "loss": 0.4975, + "step": 5490 + }, + { + "epoch": 9.02009020090201, + "grad_norm": 0.7039642333984375, + "learning_rate": 4.596042851732008e-06, + "loss": 0.4949, + "step": 5500 + }, + { + "epoch": 9.03649036490365, + "grad_norm": 0.6756150722503662, + "learning_rate": 4.442719421385922e-06, + "loss": 0.4925, + "step": 5510 + }, + { + "epoch": 9.052890528905289, + "grad_norm": 0.684483528137207, + "learning_rate": 4.291938883411007e-06, + "loss": 0.4997, + "step": 5520 + }, + { + "epoch": 9.069290692906929, + "grad_norm": 0.6780720353126526, + "learning_rate": 4.143705250255869e-06, + "loss": 0.5104, + "step": 5530 + }, + { + "epoch": 9.085690856908569, + "grad_norm": 0.6675730347633362, + "learning_rate": 3.99802246659301e-06, + "loss": 0.5181, + "step": 5540 + }, + { + "epoch": 9.102091020910208, + "grad_norm": 0.6699404120445251, + "learning_rate": 3.85489440921376e-06, + "loss": 0.506, + "step": 5550 + }, + { + "epoch": 9.11849118491185, + "grad_norm": 0.6987513899803162, + "learning_rate": 3.7143248869252022e-06, + "loss": 0.5058, + "step": 5560 + }, + { + "epoch": 9.13489134891349, + "grad_norm": 0.6789569854736328, + "learning_rate": 3.5763176404487564e-06, + "loss": 0.4977, + "step": 5570 + }, + { + "epoch": 9.15129151291513, + "grad_norm": 0.6613333821296692, + "learning_rate": 3.440876342320609e-06, + "loss": 0.5096, + "step": 5580 + }, + { + "epoch": 9.16769167691677, + "grad_norm": 0.689217746257782, + "learning_rate": 3.308004596794101e-06, + "loss": 0.4949, + "step": 5590 + }, + { + "epoch": 9.18409184091841, + "grad_norm": 0.7100328803062439, + "learning_rate": 3.1777059397436692e-06, + "loss": 0.5044, + "step": 5600 + }, + { + "epoch": 9.200492004920049, + "grad_norm": 0.6680495738983154, + "learning_rate": 3.049983838570858e-06, + "loss": 0.5037, + "step": 5610 + }, + { + "epoch": 9.216892168921689, + "grad_norm": 0.6641035676002502, + "learning_rate": 2.9248416921119794e-06, + "loss": 0.4964, + "step": 5620 + }, + { + "epoch": 9.233292332923329, + "grad_norm": 0.7248251438140869, + "learning_rate": 2.8022828305477423e-06, + "loss": 0.5017, + "step": 5630 + }, + { + "epoch": 9.24969249692497, + "grad_norm": 0.7348821759223938, + "learning_rate": 2.682310515314512e-06, + "loss": 0.5077, + "step": 5640 + }, + { + "epoch": 9.26609266092661, + "grad_norm": 0.6882705688476562, + "learning_rate": 2.5649279390176806e-06, + "loss": 0.4926, + "step": 5650 + }, + { + "epoch": 9.28249282492825, + "grad_norm": 0.6873205900192261, + "learning_rate": 2.4501382253465543e-06, + "loss": 0.5009, + "step": 5660 + }, + { + "epoch": 9.29889298892989, + "grad_norm": 0.703940212726593, + "learning_rate": 2.3379444289913342e-06, + "loss": 0.5011, + "step": 5670 + }, + { + "epoch": 9.31529315293153, + "grad_norm": 0.6873738169670105, + "learning_rate": 2.228349535561769e-06, + "loss": 0.4991, + "step": 5680 + }, + { + "epoch": 9.331693316933169, + "grad_norm": 0.6344056725502014, + "learning_rate": 2.1213564615077065e-06, + "loss": 0.5099, + "step": 5690 + }, + { + "epoch": 9.348093480934809, + "grad_norm": 0.6923158168792725, + "learning_rate": 2.016968054041546e-06, + "loss": 0.5068, + "step": 5700 + }, + { + "epoch": 9.364493644936449, + "grad_norm": 0.6982558965682983, + "learning_rate": 1.915187091062387e-06, + "loss": 0.5052, + "step": 5710 + }, + { + "epoch": 9.38089380893809, + "grad_norm": 0.6965045928955078, + "learning_rate": 1.816016281082178e-06, + "loss": 0.4963, + "step": 5720 + }, + { + "epoch": 9.39729397293973, + "grad_norm": 0.704952597618103, + "learning_rate": 1.7194582631535617e-06, + "loss": 0.4922, + "step": 5730 + }, + { + "epoch": 9.41369413694137, + "grad_norm": 0.673637866973877, + "learning_rate": 1.6255156067997323e-06, + "loss": 0.4943, + "step": 5740 + }, + { + "epoch": 9.43009430094301, + "grad_norm": 0.6612069010734558, + "learning_rate": 1.5341908119459792e-06, + "loss": 0.4973, + "step": 5750 + }, + { + "epoch": 9.44649446494465, + "grad_norm": 0.7053012847900391, + "learning_rate": 1.4454863088532388e-06, + "loss": 0.5049, + "step": 5760 + }, + { + "epoch": 9.46289462894629, + "grad_norm": 0.6800163388252258, + "learning_rate": 1.3594044580533482e-06, + "loss": 0.5009, + "step": 5770 + }, + { + "epoch": 9.479294792947929, + "grad_norm": 0.7071276307106018, + "learning_rate": 1.2759475502862828e-06, + "loss": 0.5016, + "step": 5780 + }, + { + "epoch": 9.495694956949569, + "grad_norm": 0.693181037902832, + "learning_rate": 1.19511780643915e-06, + "loss": 0.5006, + "step": 5790 + }, + { + "epoch": 9.512095120951209, + "grad_norm": 0.7130532264709473, + "learning_rate": 1.1169173774871478e-06, + "loss": 0.5072, + "step": 5800 + }, + { + "epoch": 9.52849528495285, + "grad_norm": 0.6810155510902405, + "learning_rate": 1.0413483444362771e-06, + "loss": 0.5014, + "step": 5810 + }, + { + "epoch": 9.54489544895449, + "grad_norm": 0.6402135491371155, + "learning_rate": 9.684127182679526e-07, + "loss": 0.4956, + "step": 5820 + }, + { + "epoch": 9.56129561295613, + "grad_norm": 0.6540482640266418, + "learning_rate": 8.981124398855678e-07, + "loss": 0.5035, + "step": 5830 + }, + { + "epoch": 9.57769577695777, + "grad_norm": 0.6759128570556641, + "learning_rate": 8.304493800627589e-07, + "loss": 0.4896, + "step": 5840 + }, + { + "epoch": 9.59409594095941, + "grad_norm": 0.67359459400177, + "learning_rate": 7.654253393936439e-07, + "loss": 0.5071, + "step": 5850 + }, + { + "epoch": 9.61049610496105, + "grad_norm": 0.740897536277771, + "learning_rate": 7.030420482449395e-07, + "loss": 0.5013, + "step": 5860 + }, + { + "epoch": 9.626896268962689, + "grad_norm": 0.687887966632843, + "learning_rate": 6.433011667098754e-07, + "loss": 0.5017, + "step": 5870 + }, + { + "epoch": 9.64329643296433, + "grad_norm": 0.6390769481658936, + "learning_rate": 5.862042845640403e-07, + "loss": 0.5032, + "step": 5880 + }, + { + "epoch": 9.65969659696597, + "grad_norm": 0.6661211848258972, + "learning_rate": 5.317529212230721e-07, + "loss": 0.5026, + "step": 5890 + }, + { + "epoch": 9.67609676096761, + "grad_norm": 0.7113086581230164, + "learning_rate": 4.799485257022118e-07, + "loss": 0.4996, + "step": 5900 + }, + { + "epoch": 9.69249692496925, + "grad_norm": 0.6506823301315308, + "learning_rate": 4.307924765777682e-07, + "loss": 0.5032, + "step": 5910 + }, + { + "epoch": 9.70889708897089, + "grad_norm": 0.6711443066596985, + "learning_rate": 3.842860819504024e-07, + "loss": 0.502, + "step": 5920 + }, + { + "epoch": 9.72529725297253, + "grad_norm": 0.7337656021118164, + "learning_rate": 3.404305794103224e-07, + "loss": 0.5138, + "step": 5930 + }, + { + "epoch": 9.74169741697417, + "grad_norm": 0.6719791889190674, + "learning_rate": 2.9922713600439854e-07, + "loss": 0.5015, + "step": 5940 + }, + { + "epoch": 9.758097580975809, + "grad_norm": 0.6952204704284668, + "learning_rate": 2.606768482050215e-07, + "loss": 0.5161, + "step": 5950 + }, + { + "epoch": 9.774497744977449, + "grad_norm": 0.7182480096817017, + "learning_rate": 2.2478074188099219e-07, + "loss": 0.5056, + "step": 5960 + }, + { + "epoch": 9.79089790897909, + "grad_norm": 0.7192881107330322, + "learning_rate": 1.915397722702217e-07, + "loss": 0.4972, + "step": 5970 + }, + { + "epoch": 9.80729807298073, + "grad_norm": 0.6714368462562561, + "learning_rate": 1.609548239542402e-07, + "loss": 0.4958, + "step": 5980 + }, + { + "epoch": 9.82369823698237, + "grad_norm": 0.6657147407531738, + "learning_rate": 1.3302671083474938e-07, + "loss": 0.4943, + "step": 5990 + }, + { + "epoch": 9.84009840098401, + "grad_norm": 0.6431368589401245, + "learning_rate": 1.0775617611189503e-07, + "loss": 0.5079, + "step": 6000 + }, + { + "epoch": 9.85649856498565, + "grad_norm": 0.7010581493377686, + "learning_rate": 8.514389226452757e-08, + "loss": 0.52, + "step": 6010 + }, + { + "epoch": 9.87289872898729, + "grad_norm": 0.6842231750488281, + "learning_rate": 6.519046103230508e-08, + "loss": 0.5011, + "step": 6020 + }, + { + "epoch": 9.88929889298893, + "grad_norm": 0.6943760514259338, + "learning_rate": 4.789641339963957e-08, + "loss": 0.5041, + "step": 6030 + }, + { + "epoch": 9.90569905699057, + "grad_norm": 0.6967430710792542, + "learning_rate": 3.3262209581619297e-08, + "loss": 0.5096, + "step": 6040 + }, + { + "epoch": 9.92209922099221, + "grad_norm": 0.6971690058708191, + "learning_rate": 2.1288239011729717e-08, + "loss": 0.5076, + "step": 6050 + }, + { + "epoch": 9.93849938499385, + "grad_norm": 0.668303906917572, + "learning_rate": 1.1974820331517312e-08, + "loss": 0.497, + "step": 6060 + }, + { + "epoch": 9.95489954899549, + "grad_norm": 0.6784859895706177, + "learning_rate": 5.3222013820741765e-09, + "loss": 0.5058, + "step": 6070 + }, + { + "epoch": 9.97129971299713, + "grad_norm": 0.6811095476150513, + "learning_rate": 1.3305591974543953e-09, + "loss": 0.501, + "step": 6080 + }, + { + "epoch": 9.98769987699877, + "grad_norm": 0.6882724165916443, + "learning_rate": 0.0, + "loss": 0.5079, + "step": 6090 + }, + { + "epoch": 9.98769987699877, + "step": 6090, + "total_flos": 4.6640782508241715e+17, + "train_loss": 0.5962841084046513, + "train_runtime": 32440.4443, + "train_samples_per_second": 2.255, + "train_steps_per_second": 0.188 + } + ], + "logging_steps": 10, + "max_steps": 6090, + "num_input_tokens_seen": 0, + "num_train_epochs": 10, + "save_steps": 500, + "total_flos": 4.6640782508241715e+17, + "train_batch_size": 3, + "trial_name": null, + "trial_params": null +}