{ "best_metric": null, "best_model_checkpoint": null, "epoch": 200.0, "eval_steps": 500, "global_step": 54000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.04, "grad_norm": 0.2127438336610794, "learning_rate": 0.00199962962962963, "loss": 2.6182, "step": 10 }, { "epoch": 0.07, "grad_norm": 0.1781795769929886, "learning_rate": 0.0019992592592592592, "loss": 2.5831, "step": 20 }, { "epoch": 0.11, "grad_norm": 0.20666161179542542, "learning_rate": 0.001998888888888889, "loss": 2.5437, "step": 30 }, { "epoch": 0.15, "grad_norm": 0.1834264099597931, "learning_rate": 0.001998518518518519, "loss": 2.6241, "step": 40 }, { "epoch": 0.19, "grad_norm": 0.1979055106639862, "learning_rate": 0.001998148148148148, "loss": 2.5936, "step": 50 }, { "epoch": 0.22, "grad_norm": 0.21300563216209412, "learning_rate": 0.001997777777777778, "loss": 2.5713, "step": 60 }, { "epoch": 0.26, "grad_norm": 0.17894431948661804, "learning_rate": 0.0019974074074074074, "loss": 2.5696, "step": 70 }, { "epoch": 0.3, "grad_norm": 0.17115385830402374, "learning_rate": 0.001997037037037037, "loss": 2.5648, "step": 80 }, { "epoch": 0.33, "grad_norm": 0.17225870490074158, "learning_rate": 0.0019966666666666666, "loss": 2.6359, "step": 90 }, { "epoch": 0.37, "grad_norm": 0.18164905905723572, "learning_rate": 0.0019962962962962964, "loss": 2.5752, "step": 100 }, { "epoch": 0.41, "grad_norm": 0.16691413521766663, "learning_rate": 0.0019959259259259257, "loss": 2.568, "step": 110 }, { "epoch": 0.44, "grad_norm": 0.169966459274292, "learning_rate": 0.0019955555555555555, "loss": 2.5995, "step": 120 }, { "epoch": 0.48, "grad_norm": 0.18239422142505646, "learning_rate": 0.0019951851851851853, "loss": 2.6098, "step": 130 }, { "epoch": 0.52, "grad_norm": 0.2049872726202011, "learning_rate": 0.0019948148148148147, "loss": 2.6161, "step": 140 }, { "epoch": 0.56, "grad_norm": 0.1980140656232834, "learning_rate": 0.0019944444444444445, "loss": 2.5886, "step": 150 }, { "epoch": 0.59, "grad_norm": 0.1836952418088913, "learning_rate": 0.0019940740740740743, "loss": 2.5801, "step": 160 }, { "epoch": 0.63, "grad_norm": 0.17529241740703583, "learning_rate": 0.0019937037037037037, "loss": 2.5513, "step": 170 }, { "epoch": 0.67, "grad_norm": 0.15523068606853485, "learning_rate": 0.0019933333333333335, "loss": 2.5621, "step": 180 }, { "epoch": 0.7, "grad_norm": 0.167140930891037, "learning_rate": 0.001992962962962963, "loss": 2.5803, "step": 190 }, { "epoch": 0.74, "grad_norm": 0.16198833286762238, "learning_rate": 0.0019925925925925927, "loss": 2.5546, "step": 200 }, { "epoch": 0.78, "grad_norm": 0.17828112840652466, "learning_rate": 0.001992222222222222, "loss": 2.5633, "step": 210 }, { "epoch": 0.81, "grad_norm": 0.16426187753677368, "learning_rate": 0.001991851851851852, "loss": 2.5959, "step": 220 }, { "epoch": 0.85, "grad_norm": 0.1646105945110321, "learning_rate": 0.0019914814814814817, "loss": 2.5519, "step": 230 }, { "epoch": 0.89, "grad_norm": 0.15640555322170258, "learning_rate": 0.001991111111111111, "loss": 2.5908, "step": 240 }, { "epoch": 0.93, "grad_norm": 0.16206321120262146, "learning_rate": 0.001990740740740741, "loss": 2.5684, "step": 250 }, { "epoch": 0.96, "grad_norm": 0.15909776091575623, "learning_rate": 0.00199037037037037, "loss": 2.6021, "step": 260 }, { "epoch": 1.0, "grad_norm": 0.2778194844722748, "learning_rate": 0.00199, "loss": 2.6248, "step": 270 }, { "epoch": 1.04, "grad_norm": 0.16800597310066223, "learning_rate": 0.00198962962962963, "loss": 2.4247, "step": 280 }, { "epoch": 1.07, "grad_norm": 0.16354629397392273, "learning_rate": 0.001989259259259259, "loss": 2.4314, "step": 290 }, { "epoch": 1.11, "grad_norm": 0.18757009506225586, "learning_rate": 0.001988888888888889, "loss": 2.4616, "step": 300 }, { "epoch": 1.15, "grad_norm": 0.16495901346206665, "learning_rate": 0.001988518518518519, "loss": 2.4934, "step": 310 }, { "epoch": 1.19, "grad_norm": 0.172115758061409, "learning_rate": 0.001988148148148148, "loss": 2.4561, "step": 320 }, { "epoch": 1.22, "grad_norm": 0.1990496963262558, "learning_rate": 0.001987777777777778, "loss": 2.4718, "step": 330 }, { "epoch": 1.26, "grad_norm": 0.19888705015182495, "learning_rate": 0.0019874074074074074, "loss": 2.5014, "step": 340 }, { "epoch": 1.3, "grad_norm": 0.20610582828521729, "learning_rate": 0.001987037037037037, "loss": 2.4453, "step": 350 }, { "epoch": 1.33, "grad_norm": 0.19915615022182465, "learning_rate": 0.0019866666666666665, "loss": 2.474, "step": 360 }, { "epoch": 1.37, "grad_norm": 0.1926460564136505, "learning_rate": 0.0019862962962962963, "loss": 2.5039, "step": 370 }, { "epoch": 1.41, "grad_norm": 0.20390328764915466, "learning_rate": 0.0019859259259259257, "loss": 2.5152, "step": 380 }, { "epoch": 1.44, "grad_norm": 0.21577057242393494, "learning_rate": 0.0019855555555555555, "loss": 2.5675, "step": 390 }, { "epoch": 1.48, "grad_norm": 0.20676636695861816, "learning_rate": 0.0019851851851851853, "loss": 2.4926, "step": 400 }, { "epoch": 1.52, "grad_norm": 0.19866999983787537, "learning_rate": 0.0019848148148148147, "loss": 2.4871, "step": 410 }, { "epoch": 1.56, "grad_norm": 0.19230778515338898, "learning_rate": 0.0019844444444444445, "loss": 2.4865, "step": 420 }, { "epoch": 1.59, "grad_norm": 0.19253689050674438, "learning_rate": 0.0019840740740740743, "loss": 2.4906, "step": 430 }, { "epoch": 1.63, "grad_norm": 0.19537055492401123, "learning_rate": 0.0019837037037037037, "loss": 2.5039, "step": 440 }, { "epoch": 1.67, "grad_norm": 0.2013939619064331, "learning_rate": 0.0019833333333333335, "loss": 2.5191, "step": 450 }, { "epoch": 1.7, "grad_norm": 0.1931568831205368, "learning_rate": 0.0019829629629629633, "loss": 2.5574, "step": 460 }, { "epoch": 1.74, "grad_norm": 0.18673449754714966, "learning_rate": 0.0019825925925925927, "loss": 2.53, "step": 470 }, { "epoch": 1.78, "grad_norm": 0.1869499236345291, "learning_rate": 0.0019822222222222225, "loss": 2.5651, "step": 480 }, { "epoch": 1.81, "grad_norm": 0.18973563611507416, "learning_rate": 0.001981851851851852, "loss": 2.5289, "step": 490 }, { "epoch": 1.85, "grad_norm": 0.17751389741897583, "learning_rate": 0.0019814814814814816, "loss": 2.4998, "step": 500 }, { "epoch": 1.89, "grad_norm": 0.18833574652671814, "learning_rate": 0.001981111111111111, "loss": 2.539, "step": 510 }, { "epoch": 1.93, "grad_norm": 0.1748533695936203, "learning_rate": 0.001980740740740741, "loss": 2.5488, "step": 520 }, { "epoch": 1.96, "grad_norm": 0.1988702416419983, "learning_rate": 0.00198037037037037, "loss": 2.5585, "step": 530 }, { "epoch": 2.0, "grad_norm": 0.3576868772506714, "learning_rate": 0.00198, "loss": 2.5076, "step": 540 }, { "epoch": 2.04, "grad_norm": 0.20755210518836975, "learning_rate": 0.00197962962962963, "loss": 2.3629, "step": 550 }, { "epoch": 2.07, "grad_norm": 0.21379078924655914, "learning_rate": 0.001979259259259259, "loss": 2.3609, "step": 560 }, { "epoch": 2.11, "grad_norm": 0.2395298331975937, "learning_rate": 0.001978888888888889, "loss": 2.3526, "step": 570 }, { "epoch": 2.15, "grad_norm": 0.2168968915939331, "learning_rate": 0.0019785185185185188, "loss": 2.366, "step": 580 }, { "epoch": 2.19, "grad_norm": 0.20609894394874573, "learning_rate": 0.001978148148148148, "loss": 2.3096, "step": 590 }, { "epoch": 2.22, "grad_norm": 0.24114735424518585, "learning_rate": 0.001977777777777778, "loss": 2.3611, "step": 600 }, { "epoch": 2.26, "grad_norm": 0.23323725163936615, "learning_rate": 0.0019774074074074078, "loss": 2.3785, "step": 610 }, { "epoch": 2.3, "grad_norm": 0.23909179866313934, "learning_rate": 0.001977037037037037, "loss": 2.4195, "step": 620 }, { "epoch": 2.33, "grad_norm": 0.2208935022354126, "learning_rate": 0.0019766666666666665, "loss": 2.3907, "step": 630 }, { "epoch": 2.37, "grad_norm": 0.21565470099449158, "learning_rate": 0.0019762962962962963, "loss": 2.3982, "step": 640 }, { "epoch": 2.41, "grad_norm": 0.24911703169345856, "learning_rate": 0.001975925925925926, "loss": 2.37, "step": 650 }, { "epoch": 2.44, "grad_norm": 0.2562112510204315, "learning_rate": 0.0019755555555555555, "loss": 2.3608, "step": 660 }, { "epoch": 2.48, "grad_norm": 0.2207982838153839, "learning_rate": 0.0019751851851851853, "loss": 2.4214, "step": 670 }, { "epoch": 2.52, "grad_norm": 0.25064149498939514, "learning_rate": 0.0019748148148148147, "loss": 2.409, "step": 680 }, { "epoch": 2.56, "grad_norm": 0.21787331998348236, "learning_rate": 0.0019744444444444445, "loss": 2.4376, "step": 690 }, { "epoch": 2.59, "grad_norm": 0.23485612869262695, "learning_rate": 0.0019740740740740743, "loss": 2.4168, "step": 700 }, { "epoch": 2.63, "grad_norm": 0.21844473481178284, "learning_rate": 0.0019737037037037036, "loss": 2.4271, "step": 710 }, { "epoch": 2.67, "grad_norm": 0.21845415234565735, "learning_rate": 0.0019733333333333334, "loss": 2.4636, "step": 720 }, { "epoch": 2.7, "grad_norm": 0.2322779893875122, "learning_rate": 0.0019729629629629633, "loss": 2.4407, "step": 730 }, { "epoch": 2.74, "grad_norm": 0.24358735978603363, "learning_rate": 0.0019725925925925926, "loss": 2.4425, "step": 740 }, { "epoch": 2.78, "grad_norm": 0.2179432511329651, "learning_rate": 0.0019722222222222224, "loss": 2.4417, "step": 750 }, { "epoch": 2.81, "grad_norm": 0.2172284722328186, "learning_rate": 0.001971851851851852, "loss": 2.4661, "step": 760 }, { "epoch": 2.85, "grad_norm": 0.23072928190231323, "learning_rate": 0.0019714814814814816, "loss": 2.4636, "step": 770 }, { "epoch": 2.89, "grad_norm": 0.21614006161689758, "learning_rate": 0.001971111111111111, "loss": 2.4556, "step": 780 }, { "epoch": 2.93, "grad_norm": 0.21256545186042786, "learning_rate": 0.001970740740740741, "loss": 2.454, "step": 790 }, { "epoch": 2.96, "grad_norm": 0.21115510165691376, "learning_rate": 0.00197037037037037, "loss": 2.4724, "step": 800 }, { "epoch": 3.0, "grad_norm": 0.532010555267334, "learning_rate": 0.00197, "loss": 2.4652, "step": 810 }, { "epoch": 3.04, "grad_norm": 0.2634429931640625, "learning_rate": 0.0019696296296296298, "loss": 2.2645, "step": 820 }, { "epoch": 3.07, "grad_norm": 0.24721555411815643, "learning_rate": 0.001969259259259259, "loss": 2.2422, "step": 830 }, { "epoch": 3.11, "grad_norm": 0.2868567109107971, "learning_rate": 0.001968888888888889, "loss": 2.2501, "step": 840 }, { "epoch": 3.15, "grad_norm": 0.2567349076271057, "learning_rate": 0.0019685185185185188, "loss": 2.2504, "step": 850 }, { "epoch": 3.19, "grad_norm": 0.28046759963035583, "learning_rate": 0.001968148148148148, "loss": 2.2738, "step": 860 }, { "epoch": 3.22, "grad_norm": 0.26420778036117554, "learning_rate": 0.001967777777777778, "loss": 2.3218, "step": 870 }, { "epoch": 3.26, "grad_norm": 0.2746962308883667, "learning_rate": 0.0019674074074074077, "loss": 2.293, "step": 880 }, { "epoch": 3.3, "grad_norm": 0.2583375871181488, "learning_rate": 0.001967037037037037, "loss": 2.3095, "step": 890 }, { "epoch": 3.33, "grad_norm": 0.26904764771461487, "learning_rate": 0.0019666666666666665, "loss": 2.3102, "step": 900 }, { "epoch": 3.37, "grad_norm": 0.27271828055381775, "learning_rate": 0.0019662962962962963, "loss": 2.2845, "step": 910 }, { "epoch": 3.41, "grad_norm": 0.25372961163520813, "learning_rate": 0.001965925925925926, "loss": 2.3796, "step": 920 }, { "epoch": 3.44, "grad_norm": 0.2448766678571701, "learning_rate": 0.0019655555555555555, "loss": 2.3591, "step": 930 }, { "epoch": 3.48, "grad_norm": 0.24286913871765137, "learning_rate": 0.0019651851851851853, "loss": 2.313, "step": 940 }, { "epoch": 3.52, "grad_norm": 0.3014277219772339, "learning_rate": 0.0019648148148148146, "loss": 2.3547, "step": 950 }, { "epoch": 3.56, "grad_norm": 0.2519340515136719, "learning_rate": 0.0019644444444444444, "loss": 2.3517, "step": 960 }, { "epoch": 3.59, "grad_norm": 0.25753921270370483, "learning_rate": 0.0019640740740740742, "loss": 2.3593, "step": 970 }, { "epoch": 3.63, "grad_norm": 0.25435492396354675, "learning_rate": 0.0019637037037037036, "loss": 2.3535, "step": 980 }, { "epoch": 3.67, "grad_norm": 0.2480631172657013, "learning_rate": 0.0019633333333333334, "loss": 2.3681, "step": 990 }, { "epoch": 3.7, "grad_norm": 0.2389717549085617, "learning_rate": 0.0019629629629629632, "loss": 2.3574, "step": 1000 }, { "epoch": 3.74, "grad_norm": 0.23806588351726532, "learning_rate": 0.0019625925925925926, "loss": 2.3473, "step": 1010 }, { "epoch": 3.78, "grad_norm": 0.25649794936180115, "learning_rate": 0.0019622222222222224, "loss": 2.3862, "step": 1020 }, { "epoch": 3.81, "grad_norm": 0.2556024491786957, "learning_rate": 0.001961851851851852, "loss": 2.3727, "step": 1030 }, { "epoch": 3.85, "grad_norm": 0.23823349177837372, "learning_rate": 0.0019614814814814816, "loss": 2.39, "step": 1040 }, { "epoch": 3.89, "grad_norm": 0.24092094600200653, "learning_rate": 0.001961111111111111, "loss": 2.369, "step": 1050 }, { "epoch": 3.93, "grad_norm": 0.24226288497447968, "learning_rate": 0.0019607407407407408, "loss": 2.387, "step": 1060 }, { "epoch": 3.96, "grad_norm": 0.23601235449314117, "learning_rate": 0.0019603703703703706, "loss": 2.4098, "step": 1070 }, { "epoch": 4.0, "grad_norm": 0.49935537576675415, "learning_rate": 0.00196, "loss": 2.4015, "step": 1080 }, { "epoch": 4.04, "grad_norm": 0.25792741775512695, "learning_rate": 0.0019596296296296297, "loss": 2.1849, "step": 1090 }, { "epoch": 4.07, "grad_norm": 0.2823100686073303, "learning_rate": 0.001959259259259259, "loss": 2.1466, "step": 1100 }, { "epoch": 4.11, "grad_norm": 0.2801184058189392, "learning_rate": 0.001958888888888889, "loss": 2.1778, "step": 1110 }, { "epoch": 4.15, "grad_norm": 0.283256471157074, "learning_rate": 0.0019585185185185187, "loss": 2.1968, "step": 1120 }, { "epoch": 4.19, "grad_norm": 0.3168976306915283, "learning_rate": 0.001958148148148148, "loss": 2.232, "step": 1130 }, { "epoch": 4.22, "grad_norm": 0.34400302171707153, "learning_rate": 0.001957777777777778, "loss": 2.2086, "step": 1140 }, { "epoch": 4.26, "grad_norm": 0.2858618497848511, "learning_rate": 0.0019574074074074077, "loss": 2.2432, "step": 1150 }, { "epoch": 4.3, "grad_norm": 0.286922812461853, "learning_rate": 0.001957037037037037, "loss": 2.2277, "step": 1160 }, { "epoch": 4.33, "grad_norm": 0.3100804090499878, "learning_rate": 0.001956666666666667, "loss": 2.2501, "step": 1170 }, { "epoch": 4.37, "grad_norm": 0.2689295709133148, "learning_rate": 0.0019562962962962963, "loss": 2.2456, "step": 1180 }, { "epoch": 4.41, "grad_norm": 0.2740320563316345, "learning_rate": 0.001955925925925926, "loss": 2.2712, "step": 1190 }, { "epoch": 4.44, "grad_norm": 0.2822621166706085, "learning_rate": 0.0019555555555555554, "loss": 2.2705, "step": 1200 }, { "epoch": 4.48, "grad_norm": 0.2842746376991272, "learning_rate": 0.0019551851851851852, "loss": 2.267, "step": 1210 }, { "epoch": 4.52, "grad_norm": 0.2541738450527191, "learning_rate": 0.0019548148148148146, "loss": 2.2729, "step": 1220 }, { "epoch": 4.56, "grad_norm": 0.28076863288879395, "learning_rate": 0.0019544444444444444, "loss": 2.3218, "step": 1230 }, { "epoch": 4.59, "grad_norm": 0.30424174666404724, "learning_rate": 0.0019540740740740742, "loss": 2.282, "step": 1240 }, { "epoch": 4.63, "grad_norm": 0.28858065605163574, "learning_rate": 0.0019537037037037036, "loss": 2.2981, "step": 1250 }, { "epoch": 4.67, "grad_norm": 0.25463220477104187, "learning_rate": 0.0019533333333333334, "loss": 2.2858, "step": 1260 }, { "epoch": 4.7, "grad_norm": 0.26512467861175537, "learning_rate": 0.001952962962962963, "loss": 2.3391, "step": 1270 }, { "epoch": 4.74, "grad_norm": 0.26096561551094055, "learning_rate": 0.0019525925925925928, "loss": 2.2934, "step": 1280 }, { "epoch": 4.78, "grad_norm": 0.26782888174057007, "learning_rate": 0.0019522222222222224, "loss": 2.2905, "step": 1290 }, { "epoch": 4.81, "grad_norm": 0.28986856341362, "learning_rate": 0.001951851851851852, "loss": 2.3185, "step": 1300 }, { "epoch": 4.85, "grad_norm": 0.25891393423080444, "learning_rate": 0.0019514814814814813, "loss": 2.3583, "step": 1310 }, { "epoch": 4.89, "grad_norm": 0.293237566947937, "learning_rate": 0.0019511111111111111, "loss": 2.3291, "step": 1320 }, { "epoch": 4.93, "grad_norm": 0.275447815656662, "learning_rate": 0.0019507407407407407, "loss": 2.3604, "step": 1330 }, { "epoch": 4.96, "grad_norm": 0.2680230438709259, "learning_rate": 0.0019503703703703703, "loss": 2.3833, "step": 1340 }, { "epoch": 5.0, "grad_norm": 0.5771416425704956, "learning_rate": 0.00195, "loss": 2.3233, "step": 1350 }, { "epoch": 5.04, "grad_norm": 0.30692097544670105, "learning_rate": 0.0019496296296296297, "loss": 2.0699, "step": 1360 }, { "epoch": 5.07, "grad_norm": 0.2993278205394745, "learning_rate": 0.0019492592592592593, "loss": 2.0887, "step": 1370 }, { "epoch": 5.11, "grad_norm": 0.3134785592556, "learning_rate": 0.001948888888888889, "loss": 2.154, "step": 1380 }, { "epoch": 5.15, "grad_norm": 0.28041109442710876, "learning_rate": 0.0019485185185185185, "loss": 2.1398, "step": 1390 }, { "epoch": 5.19, "grad_norm": 0.33000850677490234, "learning_rate": 0.0019481481481481483, "loss": 2.1235, "step": 1400 }, { "epoch": 5.22, "grad_norm": 0.32690125703811646, "learning_rate": 0.0019477777777777779, "loss": 2.1624, "step": 1410 }, { "epoch": 5.26, "grad_norm": 0.3051457405090332, "learning_rate": 0.0019474074074074075, "loss": 2.1376, "step": 1420 }, { "epoch": 5.3, "grad_norm": 0.30885449051856995, "learning_rate": 0.0019470370370370373, "loss": 2.1909, "step": 1430 }, { "epoch": 5.33, "grad_norm": 0.30936262011528015, "learning_rate": 0.0019466666666666669, "loss": 2.2156, "step": 1440 }, { "epoch": 5.37, "grad_norm": 0.29614242911338806, "learning_rate": 0.0019462962962962962, "loss": 2.2312, "step": 1450 }, { "epoch": 5.41, "grad_norm": 0.3082076907157898, "learning_rate": 0.0019459259259259258, "loss": 2.1844, "step": 1460 }, { "epoch": 5.44, "grad_norm": 0.3007611334323883, "learning_rate": 0.0019455555555555556, "loss": 2.2135, "step": 1470 }, { "epoch": 5.48, "grad_norm": 0.29505500197410583, "learning_rate": 0.0019451851851851852, "loss": 2.2212, "step": 1480 }, { "epoch": 5.52, "grad_norm": 0.2663390040397644, "learning_rate": 0.0019448148148148148, "loss": 2.2553, "step": 1490 }, { "epoch": 5.56, "grad_norm": 0.28654712438583374, "learning_rate": 0.0019444444444444444, "loss": 2.2783, "step": 1500 }, { "epoch": 5.59, "grad_norm": 0.29491278529167175, "learning_rate": 0.0019440740740740742, "loss": 2.2367, "step": 1510 }, { "epoch": 5.63, "grad_norm": 0.31238043308258057, "learning_rate": 0.0019437037037037038, "loss": 2.2897, "step": 1520 }, { "epoch": 5.67, "grad_norm": 0.30257317423820496, "learning_rate": 0.0019433333333333334, "loss": 2.2855, "step": 1530 }, { "epoch": 5.7, "grad_norm": 0.2944425046443939, "learning_rate": 0.001942962962962963, "loss": 2.242, "step": 1540 }, { "epoch": 5.74, "grad_norm": 0.2931998372077942, "learning_rate": 0.0019425925925925928, "loss": 2.2303, "step": 1550 }, { "epoch": 5.78, "grad_norm": 0.2865775227546692, "learning_rate": 0.0019422222222222224, "loss": 2.2459, "step": 1560 }, { "epoch": 5.81, "grad_norm": 0.28233101963996887, "learning_rate": 0.001941851851851852, "loss": 2.3051, "step": 1570 }, { "epoch": 5.85, "grad_norm": 0.2974388599395752, "learning_rate": 0.0019414814814814815, "loss": 2.3188, "step": 1580 }, { "epoch": 5.89, "grad_norm": 0.287343293428421, "learning_rate": 0.0019411111111111111, "loss": 2.2731, "step": 1590 }, { "epoch": 5.93, "grad_norm": 0.2774428427219391, "learning_rate": 0.0019407407407407407, "loss": 2.308, "step": 1600 }, { "epoch": 5.96, "grad_norm": 0.29502561688423157, "learning_rate": 0.0019403703703703703, "loss": 2.3099, "step": 1610 }, { "epoch": 6.0, "grad_norm": 0.7296469807624817, "learning_rate": 0.0019399999999999999, "loss": 2.3111, "step": 1620 }, { "epoch": 6.04, "grad_norm": 0.3095107078552246, "learning_rate": 0.0019396296296296297, "loss": 2.0247, "step": 1630 }, { "epoch": 6.07, "grad_norm": 0.3360276520252228, "learning_rate": 0.0019392592592592593, "loss": 2.0108, "step": 1640 }, { "epoch": 6.11, "grad_norm": 0.2921351194381714, "learning_rate": 0.0019388888888888889, "loss": 2.0452, "step": 1650 }, { "epoch": 6.15, "grad_norm": 0.3128146231174469, "learning_rate": 0.0019385185185185187, "loss": 2.0899, "step": 1660 }, { "epoch": 6.19, "grad_norm": 0.32681846618652344, "learning_rate": 0.0019381481481481483, "loss": 2.1214, "step": 1670 }, { "epoch": 6.22, "grad_norm": 0.3218587040901184, "learning_rate": 0.0019377777777777778, "loss": 2.1333, "step": 1680 }, { "epoch": 6.26, "grad_norm": 0.32519564032554626, "learning_rate": 0.0019374074074074074, "loss": 2.1453, "step": 1690 }, { "epoch": 6.3, "grad_norm": 0.3455524742603302, "learning_rate": 0.0019370370370370372, "loss": 2.1466, "step": 1700 }, { "epoch": 6.33, "grad_norm": 0.3287692666053772, "learning_rate": 0.0019366666666666668, "loss": 2.1663, "step": 1710 }, { "epoch": 6.37, "grad_norm": 0.34251290559768677, "learning_rate": 0.0019362962962962964, "loss": 2.1636, "step": 1720 }, { "epoch": 6.41, "grad_norm": 0.3383674621582031, "learning_rate": 0.0019359259259259258, "loss": 2.1728, "step": 1730 }, { "epoch": 6.44, "grad_norm": 0.29128894209861755, "learning_rate": 0.0019355555555555556, "loss": 2.1815, "step": 1740 }, { "epoch": 6.48, "grad_norm": 0.3406875431537628, "learning_rate": 0.0019351851851851852, "loss": 2.228, "step": 1750 }, { "epoch": 6.52, "grad_norm": 0.3243136405944824, "learning_rate": 0.0019348148148148148, "loss": 2.1952, "step": 1760 }, { "epoch": 6.56, "grad_norm": 0.29255208373069763, "learning_rate": 0.0019344444444444444, "loss": 2.213, "step": 1770 }, { "epoch": 6.59, "grad_norm": 0.3365947902202606, "learning_rate": 0.0019340740740740742, "loss": 2.2172, "step": 1780 }, { "epoch": 6.63, "grad_norm": 0.3175502121448517, "learning_rate": 0.0019337037037037038, "loss": 2.2419, "step": 1790 }, { "epoch": 6.67, "grad_norm": 0.27796781063079834, "learning_rate": 0.0019333333333333333, "loss": 2.1788, "step": 1800 }, { "epoch": 6.7, "grad_norm": 0.3237530589103699, "learning_rate": 0.001932962962962963, "loss": 2.1958, "step": 1810 }, { "epoch": 6.74, "grad_norm": 0.2936450242996216, "learning_rate": 0.0019325925925925927, "loss": 2.2351, "step": 1820 }, { "epoch": 6.78, "grad_norm": 0.3066917657852173, "learning_rate": 0.0019322222222222223, "loss": 2.2373, "step": 1830 }, { "epoch": 6.81, "grad_norm": 0.32224395871162415, "learning_rate": 0.001931851851851852, "loss": 2.2343, "step": 1840 }, { "epoch": 6.85, "grad_norm": 0.29986292123794556, "learning_rate": 0.0019314814814814817, "loss": 2.2505, "step": 1850 }, { "epoch": 6.89, "grad_norm": 0.29892221093177795, "learning_rate": 0.001931111111111111, "loss": 2.2092, "step": 1860 }, { "epoch": 6.93, "grad_norm": 0.2913380265235901, "learning_rate": 0.0019307407407407407, "loss": 2.2298, "step": 1870 }, { "epoch": 6.96, "grad_norm": 0.28266847133636475, "learning_rate": 0.0019303703703703703, "loss": 2.2531, "step": 1880 }, { "epoch": 7.0, "grad_norm": 0.5149755477905273, "learning_rate": 0.00193, "loss": 2.2141, "step": 1890 }, { "epoch": 7.04, "grad_norm": 0.34365060925483704, "learning_rate": 0.0019296296296296297, "loss": 2.0277, "step": 1900 }, { "epoch": 7.07, "grad_norm": 0.3547360599040985, "learning_rate": 0.0019292592592592593, "loss": 2.0065, "step": 1910 }, { "epoch": 7.11, "grad_norm": 0.36707526445388794, "learning_rate": 0.0019288888888888888, "loss": 2.0481, "step": 1920 }, { "epoch": 7.15, "grad_norm": 0.3768918812274933, "learning_rate": 0.0019285185185185186, "loss": 2.0431, "step": 1930 }, { "epoch": 7.19, "grad_norm": 0.36633971333503723, "learning_rate": 0.0019281481481481482, "loss": 2.0451, "step": 1940 }, { "epoch": 7.22, "grad_norm": 0.33258959650993347, "learning_rate": 0.0019277777777777778, "loss": 2.0471, "step": 1950 }, { "epoch": 7.26, "grad_norm": 0.31032317876815796, "learning_rate": 0.0019274074074074074, "loss": 2.0915, "step": 1960 }, { "epoch": 7.3, "grad_norm": 0.3308986723423004, "learning_rate": 0.0019270370370370372, "loss": 2.1063, "step": 1970 }, { "epoch": 7.33, "grad_norm": 0.32878440618515015, "learning_rate": 0.0019266666666666668, "loss": 2.1293, "step": 1980 }, { "epoch": 7.37, "grad_norm": 0.3539588153362274, "learning_rate": 0.0019262962962962964, "loss": 2.1302, "step": 1990 }, { "epoch": 7.41, "grad_norm": 0.33087241649627686, "learning_rate": 0.0019259259259259258, "loss": 2.1474, "step": 2000 }, { "epoch": 7.44, "grad_norm": 0.35217225551605225, "learning_rate": 0.0019255555555555556, "loss": 2.1043, "step": 2010 }, { "epoch": 7.48, "grad_norm": 0.3334253132343292, "learning_rate": 0.0019251851851851852, "loss": 2.1372, "step": 2020 }, { "epoch": 7.52, "grad_norm": 0.32482457160949707, "learning_rate": 0.0019248148148148147, "loss": 2.1861, "step": 2030 }, { "epoch": 7.56, "grad_norm": 0.3670805096626282, "learning_rate": 0.0019244444444444443, "loss": 2.1611, "step": 2040 }, { "epoch": 7.59, "grad_norm": 0.3149818181991577, "learning_rate": 0.0019240740740740741, "loss": 2.1459, "step": 2050 }, { "epoch": 7.63, "grad_norm": 0.3138531446456909, "learning_rate": 0.0019237037037037037, "loss": 2.1299, "step": 2060 }, { "epoch": 7.67, "grad_norm": 0.30926617980003357, "learning_rate": 0.0019233333333333333, "loss": 2.1817, "step": 2070 }, { "epoch": 7.7, "grad_norm": 0.3508884310722351, "learning_rate": 0.0019229629629629631, "loss": 2.197, "step": 2080 }, { "epoch": 7.74, "grad_norm": 0.3276742696762085, "learning_rate": 0.0019225925925925927, "loss": 2.1813, "step": 2090 }, { "epoch": 7.78, "grad_norm": 0.31571853160858154, "learning_rate": 0.0019222222222222223, "loss": 2.1706, "step": 2100 }, { "epoch": 7.81, "grad_norm": 0.300502210855484, "learning_rate": 0.0019218518518518519, "loss": 2.163, "step": 2110 }, { "epoch": 7.85, "grad_norm": 0.32625290751457214, "learning_rate": 0.0019214814814814817, "loss": 2.189, "step": 2120 }, { "epoch": 7.89, "grad_norm": 0.28779491782188416, "learning_rate": 0.0019211111111111113, "loss": 2.1737, "step": 2130 }, { "epoch": 7.93, "grad_norm": 0.32611775398254395, "learning_rate": 0.0019207407407407407, "loss": 2.2241, "step": 2140 }, { "epoch": 7.96, "grad_norm": 0.3058241009712219, "learning_rate": 0.0019203703703703702, "loss": 2.2035, "step": 2150 }, { "epoch": 8.0, "grad_norm": 0.5639318823814392, "learning_rate": 0.00192, "loss": 2.1991, "step": 2160 }, { "epoch": 8.04, "grad_norm": 0.34727299213409424, "learning_rate": 0.0019196296296296296, "loss": 1.9146, "step": 2170 }, { "epoch": 8.07, "grad_norm": 0.3747304081916809, "learning_rate": 0.0019192592592592592, "loss": 1.9499, "step": 2180 }, { "epoch": 8.11, "grad_norm": 0.3590710759162903, "learning_rate": 0.0019188888888888888, "loss": 2.0072, "step": 2190 }, { "epoch": 8.15, "grad_norm": 0.413136750459671, "learning_rate": 0.0019185185185185186, "loss": 2.0058, "step": 2200 }, { "epoch": 8.19, "grad_norm": 0.3658411502838135, "learning_rate": 0.0019181481481481482, "loss": 2.0421, "step": 2210 }, { "epoch": 8.22, "grad_norm": 0.3345026671886444, "learning_rate": 0.0019177777777777778, "loss": 2.0118, "step": 2220 }, { "epoch": 8.26, "grad_norm": 0.3682136833667755, "learning_rate": 0.0019174074074074074, "loss": 2.0397, "step": 2230 }, { "epoch": 8.3, "grad_norm": 0.37295275926589966, "learning_rate": 0.0019170370370370372, "loss": 2.0797, "step": 2240 }, { "epoch": 8.33, "grad_norm": 0.3477710783481598, "learning_rate": 0.0019166666666666668, "loss": 2.0644, "step": 2250 }, { "epoch": 8.37, "grad_norm": 0.33700767159461975, "learning_rate": 0.0019162962962962964, "loss": 2.0802, "step": 2260 }, { "epoch": 8.41, "grad_norm": 0.3817288875579834, "learning_rate": 0.0019159259259259262, "loss": 2.1217, "step": 2270 }, { "epoch": 8.44, "grad_norm": 0.3545536994934082, "learning_rate": 0.0019155555555555555, "loss": 2.1185, "step": 2280 }, { "epoch": 8.48, "grad_norm": 0.39838239550590515, "learning_rate": 0.0019151851851851851, "loss": 2.0784, "step": 2290 }, { "epoch": 8.52, "grad_norm": 0.35482680797576904, "learning_rate": 0.0019148148148148147, "loss": 2.0967, "step": 2300 }, { "epoch": 8.56, "grad_norm": 0.3527957797050476, "learning_rate": 0.0019144444444444445, "loss": 2.1285, "step": 2310 }, { "epoch": 8.59, "grad_norm": 0.33589911460876465, "learning_rate": 0.0019140740740740741, "loss": 2.1375, "step": 2320 }, { "epoch": 8.63, "grad_norm": 0.32576510310173035, "learning_rate": 0.0019137037037037037, "loss": 2.1766, "step": 2330 }, { "epoch": 8.67, "grad_norm": 0.31789445877075195, "learning_rate": 0.0019133333333333333, "loss": 2.1351, "step": 2340 }, { "epoch": 8.7, "grad_norm": 0.3491995632648468, "learning_rate": 0.001912962962962963, "loss": 2.1502, "step": 2350 }, { "epoch": 8.74, "grad_norm": 0.33717969059944153, "learning_rate": 0.0019125925925925927, "loss": 2.1529, "step": 2360 }, { "epoch": 8.78, "grad_norm": 0.31113317608833313, "learning_rate": 0.0019122222222222223, "loss": 2.1498, "step": 2370 }, { "epoch": 8.81, "grad_norm": 0.34073254466056824, "learning_rate": 0.0019118518518518519, "loss": 2.162, "step": 2380 }, { "epoch": 8.85, "grad_norm": 0.32688942551612854, "learning_rate": 0.0019114814814814817, "loss": 2.164, "step": 2390 }, { "epoch": 8.89, "grad_norm": 0.3454456329345703, "learning_rate": 0.0019111111111111113, "loss": 2.168, "step": 2400 }, { "epoch": 8.93, "grad_norm": 0.326792448759079, "learning_rate": 0.0019107407407407408, "loss": 2.1524, "step": 2410 }, { "epoch": 8.96, "grad_norm": 0.32493969798088074, "learning_rate": 0.0019103703703703702, "loss": 2.1982, "step": 2420 }, { "epoch": 9.0, "grad_norm": 0.569231390953064, "learning_rate": 0.00191, "loss": 2.1692, "step": 2430 }, { "epoch": 9.04, "grad_norm": 0.37039652466773987, "learning_rate": 0.0019096296296296296, "loss": 1.9196, "step": 2440 }, { "epoch": 9.07, "grad_norm": 0.36663028597831726, "learning_rate": 0.0019092592592592592, "loss": 1.9251, "step": 2450 }, { "epoch": 9.11, "grad_norm": 0.40725961327552795, "learning_rate": 0.001908888888888889, "loss": 1.9387, "step": 2460 }, { "epoch": 9.15, "grad_norm": 0.38818269968032837, "learning_rate": 0.0019085185185185186, "loss": 1.9633, "step": 2470 }, { "epoch": 9.19, "grad_norm": 0.3413904905319214, "learning_rate": 0.0019081481481481482, "loss": 1.9779, "step": 2480 }, { "epoch": 9.22, "grad_norm": 0.3483598232269287, "learning_rate": 0.0019077777777777778, "loss": 2.0182, "step": 2490 }, { "epoch": 9.26, "grad_norm": 0.3658324182033539, "learning_rate": 0.0019074074074074076, "loss": 2.0443, "step": 2500 }, { "epoch": 9.3, "grad_norm": 0.3571145534515381, "learning_rate": 0.0019070370370370372, "loss": 2.0266, "step": 2510 }, { "epoch": 9.33, "grad_norm": 0.3722548186779022, "learning_rate": 0.0019066666666666668, "loss": 2.0479, "step": 2520 }, { "epoch": 9.37, "grad_norm": 0.39253392815589905, "learning_rate": 0.0019062962962962963, "loss": 2.0347, "step": 2530 }, { "epoch": 9.41, "grad_norm": 0.36997315287590027, "learning_rate": 0.0019059259259259261, "loss": 2.0454, "step": 2540 }, { "epoch": 9.44, "grad_norm": 0.3789735734462738, "learning_rate": 0.0019055555555555555, "loss": 2.0611, "step": 2550 }, { "epoch": 9.48, "grad_norm": 0.3231625556945801, "learning_rate": 0.001905185185185185, "loss": 2.126, "step": 2560 }, { "epoch": 9.52, "grad_norm": 0.3473018407821655, "learning_rate": 0.0019048148148148147, "loss": 2.0589, "step": 2570 }, { "epoch": 9.56, "grad_norm": 0.3326464295387268, "learning_rate": 0.0019044444444444445, "loss": 2.1153, "step": 2580 }, { "epoch": 9.59, "grad_norm": 0.3558256924152374, "learning_rate": 0.001904074074074074, "loss": 2.0861, "step": 2590 }, { "epoch": 9.63, "grad_norm": 0.35280081629753113, "learning_rate": 0.0019037037037037037, "loss": 2.099, "step": 2600 }, { "epoch": 9.67, "grad_norm": 0.3500099182128906, "learning_rate": 0.0019033333333333333, "loss": 2.116, "step": 2610 }, { "epoch": 9.7, "grad_norm": 0.35508620738983154, "learning_rate": 0.001902962962962963, "loss": 2.1118, "step": 2620 }, { "epoch": 9.74, "grad_norm": 0.34156134724617004, "learning_rate": 0.0019025925925925927, "loss": 2.1196, "step": 2630 }, { "epoch": 9.78, "grad_norm": 0.35427629947662354, "learning_rate": 0.0019022222222222222, "loss": 2.1033, "step": 2640 }, { "epoch": 9.81, "grad_norm": 0.3459043502807617, "learning_rate": 0.0019018518518518518, "loss": 2.124, "step": 2650 }, { "epoch": 9.85, "grad_norm": 0.38089868426322937, "learning_rate": 0.0019014814814814816, "loss": 2.1434, "step": 2660 }, { "epoch": 9.89, "grad_norm": 0.3299236595630646, "learning_rate": 0.0019011111111111112, "loss": 2.1155, "step": 2670 }, { "epoch": 9.93, "grad_norm": 0.37226101756095886, "learning_rate": 0.0019007407407407408, "loss": 2.1338, "step": 2680 }, { "epoch": 9.96, "grad_norm": 0.34492045640945435, "learning_rate": 0.0019003703703703704, "loss": 2.1391, "step": 2690 }, { "epoch": 10.0, "grad_norm": 0.718737781047821, "learning_rate": 0.0019, "loss": 2.119, "step": 2700 }, { "epoch": 10.04, "grad_norm": 0.39133042097091675, "learning_rate": 0.0018996296296296296, "loss": 1.9413, "step": 2710 }, { "epoch": 10.07, "grad_norm": 0.38486966490745544, "learning_rate": 0.0018992592592592592, "loss": 1.9115, "step": 2720 }, { "epoch": 10.11, "grad_norm": 0.42015212774276733, "learning_rate": 0.001898888888888889, "loss": 1.9019, "step": 2730 }, { "epoch": 10.15, "grad_norm": 0.4082113206386566, "learning_rate": 0.0018985185185185186, "loss": 1.9221, "step": 2740 }, { "epoch": 10.19, "grad_norm": 0.37983325123786926, "learning_rate": 0.0018981481481481482, "loss": 1.942, "step": 2750 }, { "epoch": 10.22, "grad_norm": 0.37726789712905884, "learning_rate": 0.0018977777777777777, "loss": 1.9894, "step": 2760 }, { "epoch": 10.26, "grad_norm": 0.3986751139163971, "learning_rate": 0.0018974074074074075, "loss": 2.0071, "step": 2770 }, { "epoch": 10.3, "grad_norm": 0.37349823117256165, "learning_rate": 0.0018970370370370371, "loss": 1.9686, "step": 2780 }, { "epoch": 10.33, "grad_norm": 0.37317484617233276, "learning_rate": 0.0018966666666666667, "loss": 2.023, "step": 2790 }, { "epoch": 10.37, "grad_norm": 0.4159048795700073, "learning_rate": 0.0018962962962962963, "loss": 2.0305, "step": 2800 }, { "epoch": 10.41, "grad_norm": 0.36336109042167664, "learning_rate": 0.0018959259259259261, "loss": 2.0059, "step": 2810 }, { "epoch": 10.44, "grad_norm": 0.3658113181591034, "learning_rate": 0.0018955555555555557, "loss": 2.0488, "step": 2820 }, { "epoch": 10.48, "grad_norm": 0.41594254970550537, "learning_rate": 0.001895185185185185, "loss": 2.0423, "step": 2830 }, { "epoch": 10.52, "grad_norm": 0.3627219498157501, "learning_rate": 0.0018948148148148147, "loss": 2.0593, "step": 2840 }, { "epoch": 10.56, "grad_norm": 0.360855370759964, "learning_rate": 0.0018944444444444445, "loss": 2.0557, "step": 2850 }, { "epoch": 10.59, "grad_norm": 0.3695986866950989, "learning_rate": 0.001894074074074074, "loss": 2.0679, "step": 2860 }, { "epoch": 10.63, "grad_norm": 0.3898347020149231, "learning_rate": 0.0018937037037037037, "loss": 2.1022, "step": 2870 }, { "epoch": 10.67, "grad_norm": 0.3742794096469879, "learning_rate": 0.0018933333333333335, "loss": 2.092, "step": 2880 }, { "epoch": 10.7, "grad_norm": 0.3694680333137512, "learning_rate": 0.001892962962962963, "loss": 2.0915, "step": 2890 }, { "epoch": 10.74, "grad_norm": 0.3656015992164612, "learning_rate": 0.0018925925925925926, "loss": 2.0832, "step": 2900 }, { "epoch": 10.78, "grad_norm": 0.3568435609340668, "learning_rate": 0.0018922222222222222, "loss": 2.0981, "step": 2910 }, { "epoch": 10.81, "grad_norm": 0.3529919981956482, "learning_rate": 0.001891851851851852, "loss": 2.0947, "step": 2920 }, { "epoch": 10.85, "grad_norm": 0.37247973680496216, "learning_rate": 0.0018914814814814816, "loss": 2.1217, "step": 2930 }, { "epoch": 10.89, "grad_norm": 0.36205926537513733, "learning_rate": 0.0018911111111111112, "loss": 2.095, "step": 2940 }, { "epoch": 10.93, "grad_norm": 0.3482193052768707, "learning_rate": 0.0018907407407407408, "loss": 2.0999, "step": 2950 }, { "epoch": 10.96, "grad_norm": 0.3682381510734558, "learning_rate": 0.0018903703703703706, "loss": 2.097, "step": 2960 }, { "epoch": 11.0, "grad_norm": 0.7415529489517212, "learning_rate": 0.00189, "loss": 2.0817, "step": 2970 }, { "epoch": 11.04, "grad_norm": 0.3742087781429291, "learning_rate": 0.0018896296296296296, "loss": 1.8732, "step": 2980 }, { "epoch": 11.07, "grad_norm": 0.42774471640586853, "learning_rate": 0.0018892592592592591, "loss": 1.9006, "step": 2990 }, { "epoch": 11.11, "grad_norm": 0.4058782458305359, "learning_rate": 0.001888888888888889, "loss": 1.8626, "step": 3000 }, { "epoch": 11.15, "grad_norm": 0.3706398904323578, "learning_rate": 0.0018885185185185185, "loss": 1.9556, "step": 3010 }, { "epoch": 11.19, "grad_norm": 0.38733819127082825, "learning_rate": 0.0018881481481481481, "loss": 1.9707, "step": 3020 }, { "epoch": 11.22, "grad_norm": 0.4044550061225891, "learning_rate": 0.0018877777777777777, "loss": 1.9411, "step": 3030 }, { "epoch": 11.26, "grad_norm": 0.43712499737739563, "learning_rate": 0.0018874074074074075, "loss": 1.9652, "step": 3040 }, { "epoch": 11.3, "grad_norm": 0.3710927367210388, "learning_rate": 0.0018870370370370371, "loss": 1.9604, "step": 3050 }, { "epoch": 11.33, "grad_norm": 0.4272994101047516, "learning_rate": 0.0018866666666666667, "loss": 1.9831, "step": 3060 }, { "epoch": 11.37, "grad_norm": 0.39658141136169434, "learning_rate": 0.0018862962962962965, "loss": 1.946, "step": 3070 }, { "epoch": 11.41, "grad_norm": 0.3768123686313629, "learning_rate": 0.001885925925925926, "loss": 2.0135, "step": 3080 }, { "epoch": 11.44, "grad_norm": 0.405793696641922, "learning_rate": 0.0018855555555555557, "loss": 2.0506, "step": 3090 }, { "epoch": 11.48, "grad_norm": 0.3863692581653595, "learning_rate": 0.001885185185185185, "loss": 2.0011, "step": 3100 }, { "epoch": 11.52, "grad_norm": 0.3689294457435608, "learning_rate": 0.0018848148148148149, "loss": 1.9958, "step": 3110 }, { "epoch": 11.56, "grad_norm": 0.3637295365333557, "learning_rate": 0.0018844444444444444, "loss": 2.0012, "step": 3120 }, { "epoch": 11.59, "grad_norm": 0.38232067227363586, "learning_rate": 0.001884074074074074, "loss": 2.029, "step": 3130 }, { "epoch": 11.63, "grad_norm": 0.3879601061344147, "learning_rate": 0.0018837037037037036, "loss": 2.0689, "step": 3140 }, { "epoch": 11.67, "grad_norm": 0.35336965322494507, "learning_rate": 0.0018833333333333334, "loss": 2.0494, "step": 3150 }, { "epoch": 11.7, "grad_norm": 0.39037132263183594, "learning_rate": 0.001882962962962963, "loss": 2.0513, "step": 3160 }, { "epoch": 11.74, "grad_norm": 0.383658766746521, "learning_rate": 0.0018825925925925926, "loss": 2.0459, "step": 3170 }, { "epoch": 11.78, "grad_norm": 0.3601329028606415, "learning_rate": 0.0018822222222222222, "loss": 2.0795, "step": 3180 }, { "epoch": 11.81, "grad_norm": 0.36117351055145264, "learning_rate": 0.001881851851851852, "loss": 2.0726, "step": 3190 }, { "epoch": 11.85, "grad_norm": 0.3798144459724426, "learning_rate": 0.0018814814814814816, "loss": 2.0436, "step": 3200 }, { "epoch": 11.89, "grad_norm": 0.35812708735466003, "learning_rate": 0.0018811111111111112, "loss": 2.1049, "step": 3210 }, { "epoch": 11.93, "grad_norm": 0.38053765892982483, "learning_rate": 0.0018807407407407408, "loss": 2.0561, "step": 3220 }, { "epoch": 11.96, "grad_norm": 0.3914078176021576, "learning_rate": 0.0018803703703703706, "loss": 2.0689, "step": 3230 }, { "epoch": 12.0, "grad_norm": 0.6013017296791077, "learning_rate": 0.00188, "loss": 2.1096, "step": 3240 }, { "epoch": 12.04, "grad_norm": 0.3909628093242645, "learning_rate": 0.0018796296296296295, "loss": 1.842, "step": 3250 }, { "epoch": 12.07, "grad_norm": 0.4217047393321991, "learning_rate": 0.0018792592592592591, "loss": 1.8258, "step": 3260 }, { "epoch": 12.11, "grad_norm": 0.4064037799835205, "learning_rate": 0.001878888888888889, "loss": 1.8367, "step": 3270 }, { "epoch": 12.15, "grad_norm": 0.4413491189479828, "learning_rate": 0.0018785185185185185, "loss": 1.8863, "step": 3280 }, { "epoch": 12.19, "grad_norm": 0.4350806474685669, "learning_rate": 0.001878148148148148, "loss": 1.8945, "step": 3290 }, { "epoch": 12.22, "grad_norm": 0.3987176716327667, "learning_rate": 0.001877777777777778, "loss": 1.9026, "step": 3300 }, { "epoch": 12.26, "grad_norm": 0.41014865040779114, "learning_rate": 0.0018774074074074075, "loss": 1.9443, "step": 3310 }, { "epoch": 12.3, "grad_norm": 0.38420772552490234, "learning_rate": 0.001877037037037037, "loss": 1.9551, "step": 3320 }, { "epoch": 12.33, "grad_norm": 0.41501930356025696, "learning_rate": 0.0018766666666666667, "loss": 1.9956, "step": 3330 }, { "epoch": 12.37, "grad_norm": 0.386685848236084, "learning_rate": 0.0018762962962962965, "loss": 1.9536, "step": 3340 }, { "epoch": 12.41, "grad_norm": 0.4385821521282196, "learning_rate": 0.001875925925925926, "loss": 1.9822, "step": 3350 }, { "epoch": 12.44, "grad_norm": 0.4114198088645935, "learning_rate": 0.0018755555555555557, "loss": 2.0037, "step": 3360 }, { "epoch": 12.48, "grad_norm": 0.4060194194316864, "learning_rate": 0.0018751851851851852, "loss": 1.9864, "step": 3370 }, { "epoch": 12.52, "grad_norm": 0.3886150121688843, "learning_rate": 0.0018748148148148148, "loss": 1.9556, "step": 3380 }, { "epoch": 12.56, "grad_norm": 0.36917322874069214, "learning_rate": 0.0018744444444444444, "loss": 2.0239, "step": 3390 }, { "epoch": 12.59, "grad_norm": 0.3943292498588562, "learning_rate": 0.001874074074074074, "loss": 2.0027, "step": 3400 }, { "epoch": 12.63, "grad_norm": 0.3857436180114746, "learning_rate": 0.0018737037037037036, "loss": 1.9949, "step": 3410 }, { "epoch": 12.67, "grad_norm": 0.4206480383872986, "learning_rate": 0.0018733333333333334, "loss": 2.0477, "step": 3420 }, { "epoch": 12.7, "grad_norm": 0.3751334249973297, "learning_rate": 0.001872962962962963, "loss": 2.0233, "step": 3430 }, { "epoch": 12.74, "grad_norm": 0.38285306096076965, "learning_rate": 0.0018725925925925926, "loss": 2.0445, "step": 3440 }, { "epoch": 12.78, "grad_norm": 0.37726810574531555, "learning_rate": 0.0018722222222222222, "loss": 2.0564, "step": 3450 }, { "epoch": 12.81, "grad_norm": 0.3757777512073517, "learning_rate": 0.001871851851851852, "loss": 2.0471, "step": 3460 }, { "epoch": 12.85, "grad_norm": 0.4291413128376007, "learning_rate": 0.0018714814814814816, "loss": 1.9982, "step": 3470 }, { "epoch": 12.89, "grad_norm": 0.37266525626182556, "learning_rate": 0.0018711111111111112, "loss": 2.0529, "step": 3480 }, { "epoch": 12.93, "grad_norm": 0.37555840611457825, "learning_rate": 0.001870740740740741, "loss": 2.0596, "step": 3490 }, { "epoch": 12.96, "grad_norm": 0.38363468647003174, "learning_rate": 0.0018703703703703705, "loss": 2.0597, "step": 3500 }, { "epoch": 13.0, "grad_norm": 0.8527718186378479, "learning_rate": 0.0018700000000000001, "loss": 2.0715, "step": 3510 }, { "epoch": 13.04, "grad_norm": 0.3870679438114166, "learning_rate": 0.0018696296296296295, "loss": 1.783, "step": 3520 }, { "epoch": 13.07, "grad_norm": 0.4346376955509186, "learning_rate": 0.0018692592592592593, "loss": 1.8299, "step": 3530 }, { "epoch": 13.11, "grad_norm": 0.40322786569595337, "learning_rate": 0.001868888888888889, "loss": 1.8313, "step": 3540 }, { "epoch": 13.15, "grad_norm": 0.43011534214019775, "learning_rate": 0.0018685185185185185, "loss": 1.8661, "step": 3550 }, { "epoch": 13.19, "grad_norm": 0.40157124400138855, "learning_rate": 0.001868148148148148, "loss": 1.8585, "step": 3560 }, { "epoch": 13.22, "grad_norm": 0.45722874999046326, "learning_rate": 0.0018677777777777779, "loss": 1.9132, "step": 3570 }, { "epoch": 13.26, "grad_norm": 0.42185381054878235, "learning_rate": 0.0018674074074074075, "loss": 1.9007, "step": 3580 }, { "epoch": 13.3, "grad_norm": 0.4317697584629059, "learning_rate": 0.001867037037037037, "loss": 1.9352, "step": 3590 }, { "epoch": 13.33, "grad_norm": 0.3869325518608093, "learning_rate": 0.0018666666666666666, "loss": 1.9519, "step": 3600 }, { "epoch": 13.37, "grad_norm": 0.4048895537853241, "learning_rate": 0.0018662962962962965, "loss": 1.9419, "step": 3610 }, { "epoch": 13.41, "grad_norm": 0.457479864358902, "learning_rate": 0.001865925925925926, "loss": 1.9694, "step": 3620 }, { "epoch": 13.44, "grad_norm": 0.4000145196914673, "learning_rate": 0.0018655555555555556, "loss": 1.9519, "step": 3630 }, { "epoch": 13.48, "grad_norm": 0.4235948920249939, "learning_rate": 0.0018651851851851852, "loss": 1.9422, "step": 3640 }, { "epoch": 13.52, "grad_norm": 0.3991868197917938, "learning_rate": 0.001864814814814815, "loss": 1.9906, "step": 3650 }, { "epoch": 13.56, "grad_norm": 0.3810255825519562, "learning_rate": 0.0018644444444444444, "loss": 1.9545, "step": 3660 }, { "epoch": 13.59, "grad_norm": 0.4063350558280945, "learning_rate": 0.001864074074074074, "loss": 1.9896, "step": 3670 }, { "epoch": 13.63, "grad_norm": 0.39621415734291077, "learning_rate": 0.0018637037037037036, "loss": 2.0098, "step": 3680 }, { "epoch": 13.67, "grad_norm": 0.4067321717739105, "learning_rate": 0.0018633333333333334, "loss": 1.993, "step": 3690 }, { "epoch": 13.7, "grad_norm": 0.4024389088153839, "learning_rate": 0.001862962962962963, "loss": 1.9938, "step": 3700 }, { "epoch": 13.74, "grad_norm": 0.42660489678382874, "learning_rate": 0.0018625925925925926, "loss": 2.0147, "step": 3710 }, { "epoch": 13.78, "grad_norm": 0.3961524963378906, "learning_rate": 0.0018622222222222224, "loss": 2.0363, "step": 3720 }, { "epoch": 13.81, "grad_norm": 0.3870633542537689, "learning_rate": 0.001861851851851852, "loss": 2.0368, "step": 3730 }, { "epoch": 13.85, "grad_norm": 0.3615890443325043, "learning_rate": 0.0018614814814814815, "loss": 1.9881, "step": 3740 }, { "epoch": 13.89, "grad_norm": 0.3907649517059326, "learning_rate": 0.0018611111111111111, "loss": 2.0409, "step": 3750 }, { "epoch": 13.93, "grad_norm": 0.3881995677947998, "learning_rate": 0.001860740740740741, "loss": 2.045, "step": 3760 }, { "epoch": 13.96, "grad_norm": 0.4178624451160431, "learning_rate": 0.0018603703703703705, "loss": 2.0016, "step": 3770 }, { "epoch": 14.0, "grad_norm": 0.7995258569717407, "learning_rate": 0.00186, "loss": 2.0585, "step": 3780 }, { "epoch": 14.04, "grad_norm": 0.4634484648704529, "learning_rate": 0.0018596296296296295, "loss": 1.8123, "step": 3790 }, { "epoch": 14.07, "grad_norm": 0.4777580499649048, "learning_rate": 0.0018592592592592593, "loss": 1.8001, "step": 3800 }, { "epoch": 14.11, "grad_norm": 0.43421947956085205, "learning_rate": 0.0018588888888888889, "loss": 1.8164, "step": 3810 }, { "epoch": 14.15, "grad_norm": 0.4807451665401459, "learning_rate": 0.0018585185185185185, "loss": 1.8032, "step": 3820 }, { "epoch": 14.19, "grad_norm": 0.4432723820209503, "learning_rate": 0.001858148148148148, "loss": 1.8449, "step": 3830 }, { "epoch": 14.22, "grad_norm": 0.44446131587028503, "learning_rate": 0.0018577777777777779, "loss": 1.8489, "step": 3840 }, { "epoch": 14.26, "grad_norm": 0.4247501790523529, "learning_rate": 0.0018574074074074074, "loss": 1.8724, "step": 3850 }, { "epoch": 14.3, "grad_norm": 0.4298627972602844, "learning_rate": 0.001857037037037037, "loss": 1.879, "step": 3860 }, { "epoch": 14.33, "grad_norm": 0.4358549416065216, "learning_rate": 0.0018566666666666666, "loss": 1.9109, "step": 3870 }, { "epoch": 14.37, "grad_norm": 0.4526694715023041, "learning_rate": 0.0018562962962962964, "loss": 1.9652, "step": 3880 }, { "epoch": 14.41, "grad_norm": 0.4284061789512634, "learning_rate": 0.001855925925925926, "loss": 1.9399, "step": 3890 }, { "epoch": 14.44, "grad_norm": 0.4858032763004303, "learning_rate": 0.0018555555555555556, "loss": 1.9424, "step": 3900 }, { "epoch": 14.48, "grad_norm": 0.4293293356895447, "learning_rate": 0.0018551851851851854, "loss": 1.9548, "step": 3910 }, { "epoch": 14.52, "grad_norm": 0.4032863676548004, "learning_rate": 0.001854814814814815, "loss": 1.9461, "step": 3920 }, { "epoch": 14.56, "grad_norm": 0.40789926052093506, "learning_rate": 0.0018544444444444444, "loss": 1.9562, "step": 3930 }, { "epoch": 14.59, "grad_norm": 0.4035639762878418, "learning_rate": 0.001854074074074074, "loss": 1.9913, "step": 3940 }, { "epoch": 14.63, "grad_norm": 0.4079335629940033, "learning_rate": 0.0018537037037037038, "loss": 1.9682, "step": 3950 }, { "epoch": 14.67, "grad_norm": 0.4082600772380829, "learning_rate": 0.0018533333333333334, "loss": 1.9547, "step": 3960 }, { "epoch": 14.7, "grad_norm": 0.4207163453102112, "learning_rate": 0.001852962962962963, "loss": 1.973, "step": 3970 }, { "epoch": 14.74, "grad_norm": 0.4074302017688751, "learning_rate": 0.0018525925925925925, "loss": 2.0217, "step": 3980 }, { "epoch": 14.78, "grad_norm": 0.40997204184532166, "learning_rate": 0.0018522222222222223, "loss": 1.9972, "step": 3990 }, { "epoch": 14.81, "grad_norm": 0.3812522888183594, "learning_rate": 0.001851851851851852, "loss": 1.9702, "step": 4000 }, { "epoch": 14.85, "grad_norm": 0.42103034257888794, "learning_rate": 0.0018514814814814815, "loss": 2.0201, "step": 4010 }, { "epoch": 14.89, "grad_norm": 0.40356218814849854, "learning_rate": 0.001851111111111111, "loss": 2.0157, "step": 4020 }, { "epoch": 14.93, "grad_norm": 0.4146977365016937, "learning_rate": 0.001850740740740741, "loss": 1.9972, "step": 4030 }, { "epoch": 14.96, "grad_norm": 0.3836473822593689, "learning_rate": 0.0018503703703703705, "loss": 2.0188, "step": 4040 }, { "epoch": 15.0, "grad_norm": 0.8165007829666138, "learning_rate": 0.00185, "loss": 1.9983, "step": 4050 }, { "epoch": 15.04, "grad_norm": 0.4440079927444458, "learning_rate": 0.0018496296296296297, "loss": 1.7477, "step": 4060 }, { "epoch": 15.07, "grad_norm": 0.4328083097934723, "learning_rate": 0.0018492592592592593, "loss": 1.7601, "step": 4070 }, { "epoch": 15.11, "grad_norm": 0.4315027594566345, "learning_rate": 0.0018488888888888888, "loss": 1.7958, "step": 4080 }, { "epoch": 15.15, "grad_norm": 0.4953678548336029, "learning_rate": 0.0018485185185185184, "loss": 1.8026, "step": 4090 }, { "epoch": 15.19, "grad_norm": 0.42505890130996704, "learning_rate": 0.0018481481481481482, "loss": 1.8054, "step": 4100 }, { "epoch": 15.22, "grad_norm": 0.44902315735816956, "learning_rate": 0.0018477777777777778, "loss": 1.8499, "step": 4110 }, { "epoch": 15.26, "grad_norm": 0.45028263330459595, "learning_rate": 0.0018474074074074074, "loss": 1.8985, "step": 4120 }, { "epoch": 15.3, "grad_norm": 0.44876259565353394, "learning_rate": 0.001847037037037037, "loss": 1.8558, "step": 4130 }, { "epoch": 15.33, "grad_norm": 0.43394652009010315, "learning_rate": 0.0018466666666666668, "loss": 1.8795, "step": 4140 }, { "epoch": 15.37, "grad_norm": 0.445832222700119, "learning_rate": 0.0018462962962962964, "loss": 1.8889, "step": 4150 }, { "epoch": 15.41, "grad_norm": 0.46300771832466125, "learning_rate": 0.001845925925925926, "loss": 1.924, "step": 4160 }, { "epoch": 15.44, "grad_norm": 0.4595911204814911, "learning_rate": 0.0018455555555555556, "loss": 1.9116, "step": 4170 }, { "epoch": 15.48, "grad_norm": 0.4411725103855133, "learning_rate": 0.0018451851851851854, "loss": 1.9208, "step": 4180 }, { "epoch": 15.52, "grad_norm": 0.4301586151123047, "learning_rate": 0.001844814814814815, "loss": 1.9635, "step": 4190 }, { "epoch": 15.56, "grad_norm": 0.4307883679866791, "learning_rate": 0.0018444444444444446, "loss": 1.9143, "step": 4200 }, { "epoch": 15.59, "grad_norm": 0.43274444341659546, "learning_rate": 0.001844074074074074, "loss": 1.943, "step": 4210 }, { "epoch": 15.63, "grad_norm": 0.40284019708633423, "learning_rate": 0.0018437037037037037, "loss": 1.9672, "step": 4220 }, { "epoch": 15.67, "grad_norm": 0.41831302642822266, "learning_rate": 0.0018433333333333333, "loss": 1.985, "step": 4230 }, { "epoch": 15.7, "grad_norm": 0.39616280794143677, "learning_rate": 0.001842962962962963, "loss": 1.9658, "step": 4240 }, { "epoch": 15.74, "grad_norm": 0.4373377859592438, "learning_rate": 0.0018425925925925925, "loss": 1.992, "step": 4250 }, { "epoch": 15.78, "grad_norm": 0.43265146017074585, "learning_rate": 0.0018422222222222223, "loss": 1.9469, "step": 4260 }, { "epoch": 15.81, "grad_norm": 0.4245070815086365, "learning_rate": 0.001841851851851852, "loss": 1.9895, "step": 4270 }, { "epoch": 15.85, "grad_norm": 0.4061969220638275, "learning_rate": 0.0018414814814814815, "loss": 1.9768, "step": 4280 }, { "epoch": 15.89, "grad_norm": 0.41055822372436523, "learning_rate": 0.001841111111111111, "loss": 1.9773, "step": 4290 }, { "epoch": 15.93, "grad_norm": 0.40104153752326965, "learning_rate": 0.0018407407407407409, "loss": 1.9641, "step": 4300 }, { "epoch": 15.96, "grad_norm": 0.3944847881793976, "learning_rate": 0.0018403703703703705, "loss": 1.9745, "step": 4310 }, { "epoch": 16.0, "grad_norm": 0.8410804867744446, "learning_rate": 0.00184, "loss": 1.9942, "step": 4320 }, { "epoch": 16.04, "grad_norm": 0.4624631106853485, "learning_rate": 0.0018396296296296299, "loss": 1.7683, "step": 4330 }, { "epoch": 16.07, "grad_norm": 0.43549641966819763, "learning_rate": 0.0018392592592592592, "loss": 1.7841, "step": 4340 }, { "epoch": 16.11, "grad_norm": 0.43923529982566833, "learning_rate": 0.0018388888888888888, "loss": 1.8057, "step": 4350 }, { "epoch": 16.15, "grad_norm": 0.4524279832839966, "learning_rate": 0.0018385185185185184, "loss": 1.7818, "step": 4360 }, { "epoch": 16.19, "grad_norm": 0.4260402023792267, "learning_rate": 0.0018381481481481482, "loss": 1.8019, "step": 4370 }, { "epoch": 16.22, "grad_norm": 0.47032997012138367, "learning_rate": 0.0018377777777777778, "loss": 1.7883, "step": 4380 }, { "epoch": 16.26, "grad_norm": 0.4660649597644806, "learning_rate": 0.0018374074074074074, "loss": 1.819, "step": 4390 }, { "epoch": 16.3, "grad_norm": 0.4598945081233978, "learning_rate": 0.001837037037037037, "loss": 1.8146, "step": 4400 }, { "epoch": 16.33, "grad_norm": 0.47421982884407043, "learning_rate": 0.0018366666666666668, "loss": 1.8558, "step": 4410 }, { "epoch": 16.37, "grad_norm": 0.43668022751808167, "learning_rate": 0.0018362962962962964, "loss": 1.903, "step": 4420 }, { "epoch": 16.41, "grad_norm": 0.42682287096977234, "learning_rate": 0.001835925925925926, "loss": 1.9004, "step": 4430 }, { "epoch": 16.44, "grad_norm": 0.46274223923683167, "learning_rate": 0.0018355555555555556, "loss": 1.8872, "step": 4440 }, { "epoch": 16.48, "grad_norm": 0.43078264594078064, "learning_rate": 0.0018351851851851854, "loss": 1.9173, "step": 4450 }, { "epoch": 16.52, "grad_norm": 0.4077586233615875, "learning_rate": 0.001834814814814815, "loss": 1.938, "step": 4460 }, { "epoch": 16.56, "grad_norm": 0.42057204246520996, "learning_rate": 0.0018344444444444445, "loss": 1.9126, "step": 4470 }, { "epoch": 16.59, "grad_norm": 0.4092435836791992, "learning_rate": 0.001834074074074074, "loss": 1.9148, "step": 4480 }, { "epoch": 16.63, "grad_norm": 0.44807711243629456, "learning_rate": 0.0018337037037037037, "loss": 1.9139, "step": 4490 }, { "epoch": 16.67, "grad_norm": 0.4302557408809662, "learning_rate": 0.0018333333333333333, "loss": 1.9476, "step": 4500 }, { "epoch": 16.7, "grad_norm": 0.41777893900871277, "learning_rate": 0.0018329629629629629, "loss": 1.9249, "step": 4510 }, { "epoch": 16.74, "grad_norm": 0.44393467903137207, "learning_rate": 0.0018325925925925927, "loss": 1.9524, "step": 4520 }, { "epoch": 16.78, "grad_norm": 0.4212867319583893, "learning_rate": 0.0018322222222222223, "loss": 1.9233, "step": 4530 }, { "epoch": 16.81, "grad_norm": 0.3991056978702545, "learning_rate": 0.0018318518518518519, "loss": 1.9497, "step": 4540 }, { "epoch": 16.85, "grad_norm": 0.3818279504776001, "learning_rate": 0.0018314814814814815, "loss": 1.9828, "step": 4550 }, { "epoch": 16.89, "grad_norm": 0.4082184135913849, "learning_rate": 0.0018311111111111113, "loss": 1.9505, "step": 4560 }, { "epoch": 16.93, "grad_norm": 0.40178772807121277, "learning_rate": 0.0018307407407407409, "loss": 1.9886, "step": 4570 }, { "epoch": 16.96, "grad_norm": 0.4248037040233612, "learning_rate": 0.0018303703703703704, "loss": 1.9412, "step": 4580 }, { "epoch": 17.0, "grad_norm": 0.8885304927825928, "learning_rate": 0.00183, "loss": 1.9689, "step": 4590 }, { "epoch": 17.04, "grad_norm": 0.4501792788505554, "learning_rate": 0.0018296296296296298, "loss": 1.7377, "step": 4600 }, { "epoch": 17.07, "grad_norm": 0.483440637588501, "learning_rate": 0.0018292592592592594, "loss": 1.7509, "step": 4610 }, { "epoch": 17.11, "grad_norm": 0.46696922183036804, "learning_rate": 0.0018288888888888888, "loss": 1.7556, "step": 4620 }, { "epoch": 17.15, "grad_norm": 0.4699177145957947, "learning_rate": 0.0018285185185185184, "loss": 1.7667, "step": 4630 }, { "epoch": 17.19, "grad_norm": 0.47239282727241516, "learning_rate": 0.0018281481481481482, "loss": 1.7972, "step": 4640 }, { "epoch": 17.22, "grad_norm": 0.47145387530326843, "learning_rate": 0.0018277777777777778, "loss": 1.8257, "step": 4650 }, { "epoch": 17.26, "grad_norm": 0.47286316752433777, "learning_rate": 0.0018274074074074074, "loss": 1.8362, "step": 4660 }, { "epoch": 17.3, "grad_norm": 0.4284460246562958, "learning_rate": 0.001827037037037037, "loss": 1.8442, "step": 4670 }, { "epoch": 17.33, "grad_norm": 0.44590720534324646, "learning_rate": 0.0018266666666666668, "loss": 1.856, "step": 4680 }, { "epoch": 17.37, "grad_norm": 0.4675563871860504, "learning_rate": 0.0018262962962962963, "loss": 1.8532, "step": 4690 }, { "epoch": 17.41, "grad_norm": 0.464427649974823, "learning_rate": 0.001825925925925926, "loss": 1.8839, "step": 4700 }, { "epoch": 17.44, "grad_norm": 0.4539407193660736, "learning_rate": 0.0018255555555555557, "loss": 1.8841, "step": 4710 }, { "epoch": 17.48, "grad_norm": 0.4451824724674225, "learning_rate": 0.0018251851851851853, "loss": 1.8744, "step": 4720 }, { "epoch": 17.52, "grad_norm": 0.43316957354545593, "learning_rate": 0.001824814814814815, "loss": 1.9018, "step": 4730 }, { "epoch": 17.56, "grad_norm": 0.4676300287246704, "learning_rate": 0.0018244444444444445, "loss": 1.892, "step": 4740 }, { "epoch": 17.59, "grad_norm": 0.46603700518608093, "learning_rate": 0.0018240740740740743, "loss": 1.8955, "step": 4750 }, { "epoch": 17.63, "grad_norm": 0.4689687192440033, "learning_rate": 0.0018237037037037037, "loss": 1.9119, "step": 4760 }, { "epoch": 17.67, "grad_norm": 0.4662397801876068, "learning_rate": 0.0018233333333333333, "loss": 1.9195, "step": 4770 }, { "epoch": 17.7, "grad_norm": 0.4355107545852661, "learning_rate": 0.0018229629629629629, "loss": 1.9403, "step": 4780 }, { "epoch": 17.74, "grad_norm": 0.41498440504074097, "learning_rate": 0.0018225925925925927, "loss": 1.9187, "step": 4790 }, { "epoch": 17.78, "grad_norm": 0.44109728932380676, "learning_rate": 0.0018222222222222223, "loss": 1.9468, "step": 4800 }, { "epoch": 17.81, "grad_norm": 0.4080424904823303, "learning_rate": 0.0018218518518518518, "loss": 1.9261, "step": 4810 }, { "epoch": 17.85, "grad_norm": 0.4469340145587921, "learning_rate": 0.0018214814814814814, "loss": 1.9199, "step": 4820 }, { "epoch": 17.89, "grad_norm": 0.42027363181114197, "learning_rate": 0.0018211111111111112, "loss": 1.9378, "step": 4830 }, { "epoch": 17.93, "grad_norm": 0.42399153113365173, "learning_rate": 0.0018207407407407408, "loss": 1.9417, "step": 4840 }, { "epoch": 17.96, "grad_norm": 0.46285194158554077, "learning_rate": 0.0018203703703703704, "loss": 1.944, "step": 4850 }, { "epoch": 18.0, "grad_norm": 0.9817782640457153, "learning_rate": 0.00182, "loss": 1.93, "step": 4860 }, { "epoch": 18.04, "grad_norm": 0.48128384351730347, "learning_rate": 0.0018196296296296298, "loss": 1.6866, "step": 4870 }, { "epoch": 18.07, "grad_norm": 0.4536389112472534, "learning_rate": 0.0018192592592592594, "loss": 1.6788, "step": 4880 }, { "epoch": 18.11, "grad_norm": 0.4597219228744507, "learning_rate": 0.001818888888888889, "loss": 1.7284, "step": 4890 }, { "epoch": 18.15, "grad_norm": 0.47937148809432983, "learning_rate": 0.0018185185185185184, "loss": 1.7616, "step": 4900 }, { "epoch": 18.19, "grad_norm": 0.4905545711517334, "learning_rate": 0.0018181481481481482, "loss": 1.7659, "step": 4910 }, { "epoch": 18.22, "grad_norm": 0.45469731092453003, "learning_rate": 0.0018177777777777778, "loss": 1.7832, "step": 4920 }, { "epoch": 18.26, "grad_norm": 0.5039942860603333, "learning_rate": 0.0018174074074074073, "loss": 1.7958, "step": 4930 }, { "epoch": 18.3, "grad_norm": 0.45666229724884033, "learning_rate": 0.0018170370370370371, "loss": 1.8288, "step": 4940 }, { "epoch": 18.33, "grad_norm": 0.4742020070552826, "learning_rate": 0.0018166666666666667, "loss": 1.8486, "step": 4950 }, { "epoch": 18.37, "grad_norm": 0.48586955666542053, "learning_rate": 0.0018162962962962963, "loss": 1.8158, "step": 4960 }, { "epoch": 18.41, "grad_norm": 0.44601643085479736, "learning_rate": 0.001815925925925926, "loss": 1.8467, "step": 4970 }, { "epoch": 18.44, "grad_norm": 0.4457637667655945, "learning_rate": 0.0018155555555555557, "loss": 1.8542, "step": 4980 }, { "epoch": 18.48, "grad_norm": 0.45777076482772827, "learning_rate": 0.0018151851851851853, "loss": 1.8653, "step": 4990 }, { "epoch": 18.52, "grad_norm": 0.4599011540412903, "learning_rate": 0.001814814814814815, "loss": 1.8432, "step": 5000 }, { "epoch": 18.56, "grad_norm": 0.4542379379272461, "learning_rate": 0.0018144444444444445, "loss": 1.9042, "step": 5010 }, { "epoch": 18.59, "grad_norm": 0.4289243817329407, "learning_rate": 0.0018140740740740743, "loss": 1.8927, "step": 5020 }, { "epoch": 18.63, "grad_norm": 0.4727242887020111, "learning_rate": 0.0018137037037037037, "loss": 1.8896, "step": 5030 }, { "epoch": 18.67, "grad_norm": 0.49049800634384155, "learning_rate": 0.0018133333333333332, "loss": 1.8991, "step": 5040 }, { "epoch": 18.7, "grad_norm": 0.42483386397361755, "learning_rate": 0.0018129629629629628, "loss": 1.8808, "step": 5050 }, { "epoch": 18.74, "grad_norm": 0.44468677043914795, "learning_rate": 0.0018125925925925926, "loss": 1.9129, "step": 5060 }, { "epoch": 18.78, "grad_norm": 0.42822033166885376, "learning_rate": 0.0018122222222222222, "loss": 1.9334, "step": 5070 }, { "epoch": 18.81, "grad_norm": 0.45312270522117615, "learning_rate": 0.0018118518518518518, "loss": 1.9465, "step": 5080 }, { "epoch": 18.85, "grad_norm": 0.4553111493587494, "learning_rate": 0.0018114814814814814, "loss": 1.8999, "step": 5090 }, { "epoch": 18.89, "grad_norm": 0.44948896765708923, "learning_rate": 0.0018111111111111112, "loss": 1.9649, "step": 5100 }, { "epoch": 18.93, "grad_norm": 0.4670604467391968, "learning_rate": 0.0018107407407407408, "loss": 1.9572, "step": 5110 }, { "epoch": 18.96, "grad_norm": 0.43330103158950806, "learning_rate": 0.0018103703703703704, "loss": 1.9322, "step": 5120 }, { "epoch": 19.0, "grad_norm": 0.921258807182312, "learning_rate": 0.0018100000000000002, "loss": 1.9589, "step": 5130 }, { "epoch": 19.04, "grad_norm": 0.5031507611274719, "learning_rate": 0.0018096296296296298, "loss": 1.6821, "step": 5140 }, { "epoch": 19.07, "grad_norm": 0.48202696442604065, "learning_rate": 0.0018092592592592594, "loss": 1.7146, "step": 5150 }, { "epoch": 19.11, "grad_norm": 0.4963012933731079, "learning_rate": 0.001808888888888889, "loss": 1.7043, "step": 5160 }, { "epoch": 19.15, "grad_norm": 0.4779762327671051, "learning_rate": 0.0018085185185185185, "loss": 1.7551, "step": 5170 }, { "epoch": 19.19, "grad_norm": 0.4379749298095703, "learning_rate": 0.0018081481481481481, "loss": 1.7755, "step": 5180 }, { "epoch": 19.22, "grad_norm": 0.49266839027404785, "learning_rate": 0.0018077777777777777, "loss": 1.7835, "step": 5190 }, { "epoch": 19.26, "grad_norm": 0.4939051568508148, "learning_rate": 0.0018074074074074073, "loss": 1.7979, "step": 5200 }, { "epoch": 19.3, "grad_norm": 0.4597000181674957, "learning_rate": 0.0018070370370370371, "loss": 1.8011, "step": 5210 }, { "epoch": 19.33, "grad_norm": 0.47228899598121643, "learning_rate": 0.0018066666666666667, "loss": 1.8119, "step": 5220 }, { "epoch": 19.37, "grad_norm": 0.4497470259666443, "learning_rate": 0.0018062962962962963, "loss": 1.8311, "step": 5230 }, { "epoch": 19.41, "grad_norm": 0.46324270963668823, "learning_rate": 0.0018059259259259259, "loss": 1.8474, "step": 5240 }, { "epoch": 19.44, "grad_norm": 0.4588213264942169, "learning_rate": 0.0018055555555555557, "loss": 1.837, "step": 5250 }, { "epoch": 19.48, "grad_norm": 0.47411489486694336, "learning_rate": 0.0018051851851851853, "loss": 1.8464, "step": 5260 }, { "epoch": 19.52, "grad_norm": 0.45634156465530396, "learning_rate": 0.0018048148148148149, "loss": 1.8478, "step": 5270 }, { "epoch": 19.56, "grad_norm": 0.42189350724220276, "learning_rate": 0.0018044444444444445, "loss": 1.8705, "step": 5280 }, { "epoch": 19.59, "grad_norm": 0.45213887095451355, "learning_rate": 0.0018040740740740743, "loss": 1.8743, "step": 5290 }, { "epoch": 19.63, "grad_norm": 0.4699808955192566, "learning_rate": 0.0018037037037037038, "loss": 1.8755, "step": 5300 }, { "epoch": 19.67, "grad_norm": 0.5012890100479126, "learning_rate": 0.0018033333333333332, "loss": 1.8705, "step": 5310 }, { "epoch": 19.7, "grad_norm": 0.39823225140571594, "learning_rate": 0.0018029629629629628, "loss": 1.8902, "step": 5320 }, { "epoch": 19.74, "grad_norm": 0.4447478652000427, "learning_rate": 0.0018025925925925926, "loss": 1.8827, "step": 5330 }, { "epoch": 19.78, "grad_norm": 0.4571211338043213, "learning_rate": 0.0018022222222222222, "loss": 1.8891, "step": 5340 }, { "epoch": 19.81, "grad_norm": 0.5060418248176575, "learning_rate": 0.0018018518518518518, "loss": 1.894, "step": 5350 }, { "epoch": 19.85, "grad_norm": 0.44250112771987915, "learning_rate": 0.0018014814814814816, "loss": 1.9192, "step": 5360 }, { "epoch": 19.89, "grad_norm": 0.48073089122772217, "learning_rate": 0.0018011111111111112, "loss": 1.9161, "step": 5370 }, { "epoch": 19.93, "grad_norm": 0.44471997022628784, "learning_rate": 0.0018007407407407408, "loss": 1.8936, "step": 5380 }, { "epoch": 19.96, "grad_norm": 0.44000044465065, "learning_rate": 0.0018003703703703704, "loss": 1.9215, "step": 5390 }, { "epoch": 20.0, "grad_norm": 1.005425214767456, "learning_rate": 0.0018000000000000002, "loss": 1.9003, "step": 5400 }, { "epoch": 20.04, "grad_norm": 0.48888400197029114, "learning_rate": 0.0017996296296296298, "loss": 1.6326, "step": 5410 }, { "epoch": 20.07, "grad_norm": 0.5140829682350159, "learning_rate": 0.0017992592592592593, "loss": 1.6589, "step": 5420 }, { "epoch": 20.11, "grad_norm": 0.5162104964256287, "learning_rate": 0.001798888888888889, "loss": 1.7067, "step": 5430 }, { "epoch": 20.15, "grad_norm": 0.47877031564712524, "learning_rate": 0.0017985185185185187, "loss": 1.728, "step": 5440 }, { "epoch": 20.19, "grad_norm": 0.5037879347801208, "learning_rate": 0.0017981481481481481, "loss": 1.7498, "step": 5450 }, { "epoch": 20.22, "grad_norm": 0.483064740896225, "learning_rate": 0.0017977777777777777, "loss": 1.7433, "step": 5460 }, { "epoch": 20.26, "grad_norm": 0.5271421670913696, "learning_rate": 0.0017974074074074073, "loss": 1.7561, "step": 5470 }, { "epoch": 20.3, "grad_norm": 0.4639941155910492, "learning_rate": 0.001797037037037037, "loss": 1.805, "step": 5480 }, { "epoch": 20.33, "grad_norm": 0.47345638275146484, "learning_rate": 0.0017966666666666667, "loss": 1.837, "step": 5490 }, { "epoch": 20.37, "grad_norm": 0.5081591606140137, "learning_rate": 0.0017962962962962963, "loss": 1.7829, "step": 5500 }, { "epoch": 20.41, "grad_norm": 0.4264516532421112, "learning_rate": 0.0017959259259259259, "loss": 1.8165, "step": 5510 }, { "epoch": 20.44, "grad_norm": 0.4831700623035431, "learning_rate": 0.0017955555555555557, "loss": 1.8385, "step": 5520 }, { "epoch": 20.48, "grad_norm": 0.4971249997615814, "learning_rate": 0.0017951851851851853, "loss": 1.8195, "step": 5530 }, { "epoch": 20.52, "grad_norm": 0.4875461161136627, "learning_rate": 0.0017948148148148148, "loss": 1.8137, "step": 5540 }, { "epoch": 20.56, "grad_norm": 0.47468191385269165, "learning_rate": 0.0017944444444444446, "loss": 1.8548, "step": 5550 }, { "epoch": 20.59, "grad_norm": 0.4697520136833191, "learning_rate": 0.0017940740740740742, "loss": 1.8786, "step": 5560 }, { "epoch": 20.63, "grad_norm": 0.4431191384792328, "learning_rate": 0.0017937037037037038, "loss": 1.8743, "step": 5570 }, { "epoch": 20.67, "grad_norm": 0.46886029839515686, "learning_rate": 0.0017933333333333332, "loss": 1.8492, "step": 5580 }, { "epoch": 20.7, "grad_norm": 0.46815305948257446, "learning_rate": 0.001792962962962963, "loss": 1.8523, "step": 5590 }, { "epoch": 20.74, "grad_norm": 0.42444488406181335, "learning_rate": 0.0017925925925925926, "loss": 1.894, "step": 5600 }, { "epoch": 20.78, "grad_norm": 0.4468463361263275, "learning_rate": 0.0017922222222222222, "loss": 1.8849, "step": 5610 }, { "epoch": 20.81, "grad_norm": 0.4831134080886841, "learning_rate": 0.0017918518518518518, "loss": 1.917, "step": 5620 }, { "epoch": 20.85, "grad_norm": 0.4968695342540741, "learning_rate": 0.0017914814814814816, "loss": 1.8697, "step": 5630 }, { "epoch": 20.89, "grad_norm": 0.4390285015106201, "learning_rate": 0.0017911111111111112, "loss": 1.8904, "step": 5640 }, { "epoch": 20.93, "grad_norm": 0.47561535239219666, "learning_rate": 0.0017907407407407407, "loss": 1.8959, "step": 5650 }, { "epoch": 20.96, "grad_norm": 0.45114925503730774, "learning_rate": 0.0017903703703703703, "loss": 1.895, "step": 5660 }, { "epoch": 21.0, "grad_norm": 0.8482184410095215, "learning_rate": 0.0017900000000000001, "loss": 1.9381, "step": 5670 }, { "epoch": 21.04, "grad_norm": 0.4956987202167511, "learning_rate": 0.0017896296296296297, "loss": 1.6155, "step": 5680 }, { "epoch": 21.07, "grad_norm": 0.5676151514053345, "learning_rate": 0.0017892592592592593, "loss": 1.6748, "step": 5690 }, { "epoch": 21.11, "grad_norm": 0.4920210838317871, "learning_rate": 0.001788888888888889, "loss": 1.7132, "step": 5700 }, { "epoch": 21.15, "grad_norm": 0.5253728628158569, "learning_rate": 0.0017885185185185187, "loss": 1.6675, "step": 5710 }, { "epoch": 21.19, "grad_norm": 0.5597649812698364, "learning_rate": 0.001788148148148148, "loss": 1.7294, "step": 5720 }, { "epoch": 21.22, "grad_norm": 0.5059242248535156, "learning_rate": 0.0017877777777777777, "loss": 1.7161, "step": 5730 }, { "epoch": 21.26, "grad_norm": 0.5013036131858826, "learning_rate": 0.0017874074074074075, "loss": 1.7621, "step": 5740 }, { "epoch": 21.3, "grad_norm": 0.4777660071849823, "learning_rate": 0.001787037037037037, "loss": 1.761, "step": 5750 }, { "epoch": 21.33, "grad_norm": 0.5135553479194641, "learning_rate": 0.0017866666666666667, "loss": 1.7918, "step": 5760 }, { "epoch": 21.37, "grad_norm": 0.48970091342926025, "learning_rate": 0.0017862962962962962, "loss": 1.7895, "step": 5770 }, { "epoch": 21.41, "grad_norm": 0.49350282549858093, "learning_rate": 0.001785925925925926, "loss": 1.8056, "step": 5780 }, { "epoch": 21.44, "grad_norm": 0.49979352951049805, "learning_rate": 0.0017855555555555556, "loss": 1.78, "step": 5790 }, { "epoch": 21.48, "grad_norm": 0.4719812273979187, "learning_rate": 0.0017851851851851852, "loss": 1.8209, "step": 5800 }, { "epoch": 21.52, "grad_norm": 0.48076552152633667, "learning_rate": 0.0017848148148148148, "loss": 1.7865, "step": 5810 }, { "epoch": 21.56, "grad_norm": 0.4831288754940033, "learning_rate": 0.0017844444444444446, "loss": 1.8441, "step": 5820 }, { "epoch": 21.59, "grad_norm": 0.46618470549583435, "learning_rate": 0.0017840740740740742, "loss": 1.8559, "step": 5830 }, { "epoch": 21.63, "grad_norm": 0.47354695200920105, "learning_rate": 0.0017837037037037038, "loss": 1.8284, "step": 5840 }, { "epoch": 21.67, "grad_norm": 0.464826762676239, "learning_rate": 0.0017833333333333334, "loss": 1.8633, "step": 5850 }, { "epoch": 21.7, "grad_norm": 0.43753430247306824, "learning_rate": 0.001782962962962963, "loss": 1.8726, "step": 5860 }, { "epoch": 21.74, "grad_norm": 0.4837232232093811, "learning_rate": 0.0017825925925925926, "loss": 1.8374, "step": 5870 }, { "epoch": 21.78, "grad_norm": 0.47672173380851746, "learning_rate": 0.0017822222222222222, "loss": 1.8707, "step": 5880 }, { "epoch": 21.81, "grad_norm": 0.48062843084335327, "learning_rate": 0.0017818518518518517, "loss": 1.891, "step": 5890 }, { "epoch": 21.85, "grad_norm": 0.47505030035972595, "learning_rate": 0.0017814814814814815, "loss": 1.8794, "step": 5900 }, { "epoch": 21.89, "grad_norm": 0.48005613684654236, "learning_rate": 0.0017811111111111111, "loss": 1.8771, "step": 5910 }, { "epoch": 21.93, "grad_norm": 0.475238561630249, "learning_rate": 0.0017807407407407407, "loss": 1.8643, "step": 5920 }, { "epoch": 21.96, "grad_norm": 0.44499918818473816, "learning_rate": 0.0017803703703703703, "loss": 1.9119, "step": 5930 }, { "epoch": 22.0, "grad_norm": 0.859517514705658, "learning_rate": 0.0017800000000000001, "loss": 1.9101, "step": 5940 }, { "epoch": 22.04, "grad_norm": 0.47651925683021545, "learning_rate": 0.0017796296296296297, "loss": 1.6509, "step": 5950 }, { "epoch": 22.07, "grad_norm": 0.5325976014137268, "learning_rate": 0.0017792592592592593, "loss": 1.6124, "step": 5960 }, { "epoch": 22.11, "grad_norm": 0.5133865475654602, "learning_rate": 0.001778888888888889, "loss": 1.6649, "step": 5970 }, { "epoch": 22.15, "grad_norm": 0.44351667165756226, "learning_rate": 0.0017785185185185187, "loss": 1.6458, "step": 5980 }, { "epoch": 22.19, "grad_norm": 0.5221470594406128, "learning_rate": 0.0017781481481481483, "loss": 1.6832, "step": 5990 }, { "epoch": 22.22, "grad_norm": 0.5310438275337219, "learning_rate": 0.0017777777777777776, "loss": 1.745, "step": 6000 }, { "epoch": 22.26, "grad_norm": 0.5029959082603455, "learning_rate": 0.0017774074074074075, "loss": 1.7047, "step": 6010 }, { "epoch": 22.3, "grad_norm": 0.5572134852409363, "learning_rate": 0.001777037037037037, "loss": 1.7418, "step": 6020 }, { "epoch": 22.33, "grad_norm": 0.46236780285835266, "learning_rate": 0.0017766666666666666, "loss": 1.7839, "step": 6030 }, { "epoch": 22.37, "grad_norm": 0.4862767457962036, "learning_rate": 0.0017762962962962962, "loss": 1.7749, "step": 6040 }, { "epoch": 22.41, "grad_norm": 0.46328261494636536, "learning_rate": 0.001775925925925926, "loss": 1.7915, "step": 6050 }, { "epoch": 22.44, "grad_norm": 0.5205667614936829, "learning_rate": 0.0017755555555555556, "loss": 1.8303, "step": 6060 }, { "epoch": 22.48, "grad_norm": 0.525134801864624, "learning_rate": 0.0017751851851851852, "loss": 1.7764, "step": 6070 }, { "epoch": 22.52, "grad_norm": 0.5078030824661255, "learning_rate": 0.0017748148148148148, "loss": 1.8026, "step": 6080 }, { "epoch": 22.56, "grad_norm": 0.4789275527000427, "learning_rate": 0.0017744444444444446, "loss": 1.8189, "step": 6090 }, { "epoch": 22.59, "grad_norm": 0.46829742193222046, "learning_rate": 0.0017740740740740742, "loss": 1.8302, "step": 6100 }, { "epoch": 22.63, "grad_norm": 0.4511633813381195, "learning_rate": 0.0017737037037037038, "loss": 1.8382, "step": 6110 }, { "epoch": 22.67, "grad_norm": 0.4943828284740448, "learning_rate": 0.0017733333333333334, "loss": 1.8424, "step": 6120 }, { "epoch": 22.7, "grad_norm": 0.44993260502815247, "learning_rate": 0.0017729629629629632, "loss": 1.8131, "step": 6130 }, { "epoch": 22.74, "grad_norm": 0.49377501010894775, "learning_rate": 0.0017725925925925925, "loss": 1.8772, "step": 6140 }, { "epoch": 22.78, "grad_norm": 0.46665674448013306, "learning_rate": 0.0017722222222222221, "loss": 1.8264, "step": 6150 }, { "epoch": 22.81, "grad_norm": 0.46312686800956726, "learning_rate": 0.001771851851851852, "loss": 1.8625, "step": 6160 }, { "epoch": 22.85, "grad_norm": 0.4879915714263916, "learning_rate": 0.0017714814814814815, "loss": 1.8923, "step": 6170 }, { "epoch": 22.89, "grad_norm": 0.4814596474170685, "learning_rate": 0.001771111111111111, "loss": 1.878, "step": 6180 }, { "epoch": 22.93, "grad_norm": 0.4914632737636566, "learning_rate": 0.0017707407407407407, "loss": 1.8592, "step": 6190 }, { "epoch": 22.96, "grad_norm": 0.4641704559326172, "learning_rate": 0.0017703703703703705, "loss": 1.8921, "step": 6200 }, { "epoch": 23.0, "grad_norm": 0.7784886360168457, "learning_rate": 0.00177, "loss": 1.8705, "step": 6210 }, { "epoch": 23.04, "grad_norm": 0.5497826933860779, "learning_rate": 0.0017696296296296297, "loss": 1.6352, "step": 6220 }, { "epoch": 23.07, "grad_norm": 0.5098663568496704, "learning_rate": 0.0017692592592592593, "loss": 1.629, "step": 6230 }, { "epoch": 23.11, "grad_norm": 0.5080466270446777, "learning_rate": 0.001768888888888889, "loss": 1.6593, "step": 6240 }, { "epoch": 23.15, "grad_norm": 0.4842182993888855, "learning_rate": 0.0017685185185185187, "loss": 1.6483, "step": 6250 }, { "epoch": 23.19, "grad_norm": 0.5256596207618713, "learning_rate": 0.0017681481481481482, "loss": 1.6967, "step": 6260 }, { "epoch": 23.22, "grad_norm": 0.5058240294456482, "learning_rate": 0.0017677777777777776, "loss": 1.7019, "step": 6270 }, { "epoch": 23.26, "grad_norm": 0.5488624572753906, "learning_rate": 0.0017674074074074074, "loss": 1.7202, "step": 6280 }, { "epoch": 23.3, "grad_norm": 0.5268415212631226, "learning_rate": 0.001767037037037037, "loss": 1.7774, "step": 6290 }, { "epoch": 23.33, "grad_norm": 0.48284679651260376, "learning_rate": 0.0017666666666666666, "loss": 1.7642, "step": 6300 }, { "epoch": 23.37, "grad_norm": 0.5055042505264282, "learning_rate": 0.0017662962962962962, "loss": 1.7453, "step": 6310 }, { "epoch": 23.41, "grad_norm": 0.4818604588508606, "learning_rate": 0.001765925925925926, "loss": 1.7657, "step": 6320 }, { "epoch": 23.44, "grad_norm": 0.46134889125823975, "learning_rate": 0.0017655555555555556, "loss": 1.7814, "step": 6330 }, { "epoch": 23.48, "grad_norm": 0.4772760570049286, "learning_rate": 0.0017651851851851852, "loss": 1.7934, "step": 6340 }, { "epoch": 23.52, "grad_norm": 0.4994891583919525, "learning_rate": 0.001764814814814815, "loss": 1.8017, "step": 6350 }, { "epoch": 23.56, "grad_norm": 0.4992477297782898, "learning_rate": 0.0017644444444444446, "loss": 1.792, "step": 6360 }, { "epoch": 23.59, "grad_norm": 0.47414326667785645, "learning_rate": 0.0017640740740740742, "loss": 1.8296, "step": 6370 }, { "epoch": 23.63, "grad_norm": 0.500939130783081, "learning_rate": 0.0017637037037037037, "loss": 1.8026, "step": 6380 }, { "epoch": 23.67, "grad_norm": 0.5158637166023254, "learning_rate": 0.0017633333333333335, "loss": 1.8275, "step": 6390 }, { "epoch": 23.7, "grad_norm": 0.5043916702270508, "learning_rate": 0.0017629629629629631, "loss": 1.8181, "step": 6400 }, { "epoch": 23.74, "grad_norm": 0.502905547618866, "learning_rate": 0.0017625925925925925, "loss": 1.8219, "step": 6410 }, { "epoch": 23.78, "grad_norm": 0.4975557029247284, "learning_rate": 0.001762222222222222, "loss": 1.845, "step": 6420 }, { "epoch": 23.81, "grad_norm": 0.4911981225013733, "learning_rate": 0.001761851851851852, "loss": 1.8564, "step": 6430 }, { "epoch": 23.85, "grad_norm": 0.4847852289676666, "learning_rate": 0.0017614814814814815, "loss": 1.8248, "step": 6440 }, { "epoch": 23.89, "grad_norm": 0.4612410366535187, "learning_rate": 0.001761111111111111, "loss": 1.8405, "step": 6450 }, { "epoch": 23.93, "grad_norm": 0.5118724703788757, "learning_rate": 0.0017607407407407407, "loss": 1.8545, "step": 6460 }, { "epoch": 23.96, "grad_norm": 0.46121078729629517, "learning_rate": 0.0017603703703703705, "loss": 1.8519, "step": 6470 }, { "epoch": 24.0, "grad_norm": 0.771907389163971, "learning_rate": 0.00176, "loss": 1.8582, "step": 6480 }, { "epoch": 24.04, "grad_norm": 0.5113037824630737, "learning_rate": 0.0017596296296296297, "loss": 1.588, "step": 6490 }, { "epoch": 24.07, "grad_norm": 0.47875529527664185, "learning_rate": 0.0017592592592592592, "loss": 1.6267, "step": 6500 }, { "epoch": 24.11, "grad_norm": 0.506752610206604, "learning_rate": 0.001758888888888889, "loss": 1.6279, "step": 6510 }, { "epoch": 24.15, "grad_norm": 0.5407618284225464, "learning_rate": 0.0017585185185185186, "loss": 1.63, "step": 6520 }, { "epoch": 24.19, "grad_norm": 0.5361552238464355, "learning_rate": 0.0017581481481481482, "loss": 1.6463, "step": 6530 }, { "epoch": 24.22, "grad_norm": 0.5410870909690857, "learning_rate": 0.001757777777777778, "loss": 1.6656, "step": 6540 }, { "epoch": 24.26, "grad_norm": 0.5300707817077637, "learning_rate": 0.0017574074074074074, "loss": 1.7206, "step": 6550 }, { "epoch": 24.3, "grad_norm": 0.5036062598228455, "learning_rate": 0.001757037037037037, "loss": 1.7108, "step": 6560 }, { "epoch": 24.33, "grad_norm": 0.47783538699150085, "learning_rate": 0.0017566666666666666, "loss": 1.735, "step": 6570 }, { "epoch": 24.37, "grad_norm": 0.5215782523155212, "learning_rate": 0.0017562962962962964, "loss": 1.7043, "step": 6580 }, { "epoch": 24.41, "grad_norm": 0.5307631492614746, "learning_rate": 0.001755925925925926, "loss": 1.7816, "step": 6590 }, { "epoch": 24.44, "grad_norm": 0.5181083083152771, "learning_rate": 0.0017555555555555556, "loss": 1.7592, "step": 6600 }, { "epoch": 24.48, "grad_norm": 0.521377444267273, "learning_rate": 0.0017551851851851851, "loss": 1.8097, "step": 6610 }, { "epoch": 24.52, "grad_norm": 0.5385915040969849, "learning_rate": 0.001754814814814815, "loss": 1.7616, "step": 6620 }, { "epoch": 24.56, "grad_norm": 0.5241475105285645, "learning_rate": 0.0017544444444444445, "loss": 1.7928, "step": 6630 }, { "epoch": 24.59, "grad_norm": 0.5101531147956848, "learning_rate": 0.0017540740740740741, "loss": 1.7963, "step": 6640 }, { "epoch": 24.63, "grad_norm": 0.5142462849617004, "learning_rate": 0.0017537037037037037, "loss": 1.803, "step": 6650 }, { "epoch": 24.67, "grad_norm": 0.5254778265953064, "learning_rate": 0.0017533333333333335, "loss": 1.8004, "step": 6660 }, { "epoch": 24.7, "grad_norm": 0.5626793503761292, "learning_rate": 0.0017529629629629631, "loss": 1.8089, "step": 6670 }, { "epoch": 24.74, "grad_norm": 0.5398451089859009, "learning_rate": 0.0017525925925925927, "loss": 1.8282, "step": 6680 }, { "epoch": 24.78, "grad_norm": 0.5404179096221924, "learning_rate": 0.001752222222222222, "loss": 1.8205, "step": 6690 }, { "epoch": 24.81, "grad_norm": 0.44393420219421387, "learning_rate": 0.0017518518518518519, "loss": 1.8549, "step": 6700 }, { "epoch": 24.85, "grad_norm": 0.5345913767814636, "learning_rate": 0.0017514814814814815, "loss": 1.8387, "step": 6710 }, { "epoch": 24.89, "grad_norm": 0.4756942689418793, "learning_rate": 0.001751111111111111, "loss": 1.8561, "step": 6720 }, { "epoch": 24.93, "grad_norm": 0.4690358340740204, "learning_rate": 0.0017507407407407406, "loss": 1.8477, "step": 6730 }, { "epoch": 24.96, "grad_norm": 0.4660181701183319, "learning_rate": 0.0017503703703703704, "loss": 1.8604, "step": 6740 }, { "epoch": 25.0, "grad_norm": 1.1090754270553589, "learning_rate": 0.00175, "loss": 1.8425, "step": 6750 }, { "epoch": 25.04, "grad_norm": 0.532478392124176, "learning_rate": 0.0017496296296296296, "loss": 1.5765, "step": 6760 }, { "epoch": 25.07, "grad_norm": 0.5390456318855286, "learning_rate": 0.0017492592592592594, "loss": 1.6065, "step": 6770 }, { "epoch": 25.11, "grad_norm": 0.5295791029930115, "learning_rate": 0.001748888888888889, "loss": 1.6317, "step": 6780 }, { "epoch": 25.15, "grad_norm": 0.5816089510917664, "learning_rate": 0.0017485185185185186, "loss": 1.6569, "step": 6790 }, { "epoch": 25.19, "grad_norm": 0.5486976504325867, "learning_rate": 0.0017481481481481482, "loss": 1.6866, "step": 6800 }, { "epoch": 25.22, "grad_norm": 0.5331709384918213, "learning_rate": 0.001747777777777778, "loss": 1.7122, "step": 6810 }, { "epoch": 25.26, "grad_norm": 0.46710267663002014, "learning_rate": 0.0017474074074074074, "loss": 1.6933, "step": 6820 }, { "epoch": 25.3, "grad_norm": 0.526747465133667, "learning_rate": 0.001747037037037037, "loss": 1.7145, "step": 6830 }, { "epoch": 25.33, "grad_norm": 0.5332999229431152, "learning_rate": 0.0017466666666666665, "loss": 1.6894, "step": 6840 }, { "epoch": 25.37, "grad_norm": 0.5216143131256104, "learning_rate": 0.0017462962962962964, "loss": 1.7418, "step": 6850 }, { "epoch": 25.41, "grad_norm": 0.5023478865623474, "learning_rate": 0.001745925925925926, "loss": 1.737, "step": 6860 }, { "epoch": 25.44, "grad_norm": 0.48997578024864197, "learning_rate": 0.0017455555555555555, "loss": 1.744, "step": 6870 }, { "epoch": 25.48, "grad_norm": 0.49648550152778625, "learning_rate": 0.0017451851851851851, "loss": 1.7504, "step": 6880 }, { "epoch": 25.52, "grad_norm": 0.5312948822975159, "learning_rate": 0.001744814814814815, "loss": 1.7733, "step": 6890 }, { "epoch": 25.56, "grad_norm": 0.5200966596603394, "learning_rate": 0.0017444444444444445, "loss": 1.758, "step": 6900 }, { "epoch": 25.59, "grad_norm": 0.519443929195404, "learning_rate": 0.001744074074074074, "loss": 1.7788, "step": 6910 }, { "epoch": 25.63, "grad_norm": 0.5155293345451355, "learning_rate": 0.0017437037037037037, "loss": 1.7849, "step": 6920 }, { "epoch": 25.67, "grad_norm": 0.4716353714466095, "learning_rate": 0.0017433333333333335, "loss": 1.8222, "step": 6930 }, { "epoch": 25.7, "grad_norm": 0.4919406771659851, "learning_rate": 0.001742962962962963, "loss": 1.8008, "step": 6940 }, { "epoch": 25.74, "grad_norm": 0.5622554421424866, "learning_rate": 0.0017425925925925927, "loss": 1.8081, "step": 6950 }, { "epoch": 25.78, "grad_norm": 0.4926491677761078, "learning_rate": 0.001742222222222222, "loss": 1.8174, "step": 6960 }, { "epoch": 25.81, "grad_norm": 0.4717676043510437, "learning_rate": 0.0017418518518518519, "loss": 1.8291, "step": 6970 }, { "epoch": 25.85, "grad_norm": 0.4829355776309967, "learning_rate": 0.0017414814814814814, "loss": 1.8098, "step": 6980 }, { "epoch": 25.89, "grad_norm": 0.5100350379943848, "learning_rate": 0.001741111111111111, "loss": 1.8398, "step": 6990 }, { "epoch": 25.93, "grad_norm": 0.504905104637146, "learning_rate": 0.0017407407407407408, "loss": 1.8189, "step": 7000 }, { "epoch": 25.96, "grad_norm": 0.5000597834587097, "learning_rate": 0.0017403703703703704, "loss": 1.8175, "step": 7010 }, { "epoch": 26.0, "grad_norm": 0.7448571920394897, "learning_rate": 0.00174, "loss": 1.86, "step": 7020 }, { "epoch": 26.04, "grad_norm": 0.504486620426178, "learning_rate": 0.0017396296296296296, "loss": 1.5784, "step": 7030 }, { "epoch": 26.07, "grad_norm": 0.5133050084114075, "learning_rate": 0.0017392592592592594, "loss": 1.5856, "step": 7040 }, { "epoch": 26.11, "grad_norm": 0.49264371395111084, "learning_rate": 0.001738888888888889, "loss": 1.5844, "step": 7050 }, { "epoch": 26.15, "grad_norm": 0.5140932202339172, "learning_rate": 0.0017385185185185186, "loss": 1.6197, "step": 7060 }, { "epoch": 26.19, "grad_norm": 0.5410853624343872, "learning_rate": 0.0017381481481481482, "loss": 1.6744, "step": 7070 }, { "epoch": 26.22, "grad_norm": 0.636168897151947, "learning_rate": 0.001737777777777778, "loss": 1.6568, "step": 7080 }, { "epoch": 26.26, "grad_norm": 0.5204961895942688, "learning_rate": 0.0017374074074074076, "loss": 1.6965, "step": 7090 }, { "epoch": 26.3, "grad_norm": 0.512642502784729, "learning_rate": 0.001737037037037037, "loss": 1.691, "step": 7100 }, { "epoch": 26.33, "grad_norm": 0.5050673484802246, "learning_rate": 0.0017366666666666665, "loss": 1.6974, "step": 7110 }, { "epoch": 26.37, "grad_norm": 0.5110860466957092, "learning_rate": 0.0017362962962962963, "loss": 1.7239, "step": 7120 }, { "epoch": 26.41, "grad_norm": 0.5326704382896423, "learning_rate": 0.001735925925925926, "loss": 1.7303, "step": 7130 }, { "epoch": 26.44, "grad_norm": 0.5366049408912659, "learning_rate": 0.0017355555555555555, "loss": 1.7422, "step": 7140 }, { "epoch": 26.48, "grad_norm": 0.5421983599662781, "learning_rate": 0.001735185185185185, "loss": 1.7498, "step": 7150 }, { "epoch": 26.52, "grad_norm": 0.5443106889724731, "learning_rate": 0.001734814814814815, "loss": 1.7636, "step": 7160 }, { "epoch": 26.56, "grad_norm": 0.5099042654037476, "learning_rate": 0.0017344444444444445, "loss": 1.7464, "step": 7170 }, { "epoch": 26.59, "grad_norm": 0.5240099430084229, "learning_rate": 0.001734074074074074, "loss": 1.7317, "step": 7180 }, { "epoch": 26.63, "grad_norm": 0.4959777891635895, "learning_rate": 0.0017337037037037039, "loss": 1.7725, "step": 7190 }, { "epoch": 26.67, "grad_norm": 0.5010444521903992, "learning_rate": 0.0017333333333333335, "loss": 1.7842, "step": 7200 }, { "epoch": 26.7, "grad_norm": 0.545404314994812, "learning_rate": 0.001732962962962963, "loss": 1.8179, "step": 7210 }, { "epoch": 26.74, "grad_norm": 0.5156935453414917, "learning_rate": 0.0017325925925925926, "loss": 1.7742, "step": 7220 }, { "epoch": 26.78, "grad_norm": 0.518345057964325, "learning_rate": 0.0017322222222222225, "loss": 1.8186, "step": 7230 }, { "epoch": 26.81, "grad_norm": 0.5182226300239563, "learning_rate": 0.0017318518518518518, "loss": 1.8013, "step": 7240 }, { "epoch": 26.85, "grad_norm": 0.49864670634269714, "learning_rate": 0.0017314814814814814, "loss": 1.8095, "step": 7250 }, { "epoch": 26.89, "grad_norm": 0.4981412887573242, "learning_rate": 0.001731111111111111, "loss": 1.7865, "step": 7260 }, { "epoch": 26.93, "grad_norm": 0.49631133675575256, "learning_rate": 0.0017307407407407408, "loss": 1.8195, "step": 7270 }, { "epoch": 26.96, "grad_norm": 0.5223541259765625, "learning_rate": 0.0017303703703703704, "loss": 1.815, "step": 7280 }, { "epoch": 27.0, "grad_norm": 0.8307838439941406, "learning_rate": 0.00173, "loss": 1.8576, "step": 7290 }, { "epoch": 27.04, "grad_norm": 0.5112069845199585, "learning_rate": 0.0017296296296296296, "loss": 1.5567, "step": 7300 }, { "epoch": 27.07, "grad_norm": 0.5340370535850525, "learning_rate": 0.0017292592592592594, "loss": 1.5654, "step": 7310 }, { "epoch": 27.11, "grad_norm": 0.5549976825714111, "learning_rate": 0.001728888888888889, "loss": 1.6092, "step": 7320 }, { "epoch": 27.15, "grad_norm": 0.5123783349990845, "learning_rate": 0.0017285185185185186, "loss": 1.6213, "step": 7330 }, { "epoch": 27.19, "grad_norm": 0.5847094058990479, "learning_rate": 0.0017281481481481481, "loss": 1.6237, "step": 7340 }, { "epoch": 27.22, "grad_norm": 0.5540896654129028, "learning_rate": 0.001727777777777778, "loss": 1.6704, "step": 7350 }, { "epoch": 27.26, "grad_norm": 0.5649771690368652, "learning_rate": 0.0017274074074074075, "loss": 1.6731, "step": 7360 }, { "epoch": 27.3, "grad_norm": 0.5565537810325623, "learning_rate": 0.001727037037037037, "loss": 1.6729, "step": 7370 }, { "epoch": 27.33, "grad_norm": 0.5422229766845703, "learning_rate": 0.0017266666666666667, "loss": 1.686, "step": 7380 }, { "epoch": 27.37, "grad_norm": 0.4892452657222748, "learning_rate": 0.0017262962962962963, "loss": 1.7129, "step": 7390 }, { "epoch": 27.41, "grad_norm": 0.5745701789855957, "learning_rate": 0.0017259259259259259, "loss": 1.7245, "step": 7400 }, { "epoch": 27.44, "grad_norm": 0.5544735789299011, "learning_rate": 0.0017255555555555555, "loss": 1.7314, "step": 7410 }, { "epoch": 27.48, "grad_norm": 0.5073575973510742, "learning_rate": 0.0017251851851851853, "loss": 1.7392, "step": 7420 }, { "epoch": 27.52, "grad_norm": 0.5089659690856934, "learning_rate": 0.0017248148148148149, "loss": 1.7203, "step": 7430 }, { "epoch": 27.56, "grad_norm": 0.51652991771698, "learning_rate": 0.0017244444444444445, "loss": 1.7763, "step": 7440 }, { "epoch": 27.59, "grad_norm": 0.5411249399185181, "learning_rate": 0.001724074074074074, "loss": 1.7672, "step": 7450 }, { "epoch": 27.63, "grad_norm": 0.5420709252357483, "learning_rate": 0.0017237037037037039, "loss": 1.7659, "step": 7460 }, { "epoch": 27.67, "grad_norm": 0.5166801810264587, "learning_rate": 0.0017233333333333334, "loss": 1.7749, "step": 7470 }, { "epoch": 27.7, "grad_norm": 0.5379444360733032, "learning_rate": 0.001722962962962963, "loss": 1.7884, "step": 7480 }, { "epoch": 27.74, "grad_norm": 0.5639204978942871, "learning_rate": 0.0017225925925925926, "loss": 1.7857, "step": 7490 }, { "epoch": 27.78, "grad_norm": 0.5733610987663269, "learning_rate": 0.0017222222222222224, "loss": 1.7809, "step": 7500 }, { "epoch": 27.81, "grad_norm": 0.5559260845184326, "learning_rate": 0.0017218518518518518, "loss": 1.7841, "step": 7510 }, { "epoch": 27.85, "grad_norm": 0.5033745169639587, "learning_rate": 0.0017214814814814814, "loss": 1.7862, "step": 7520 }, { "epoch": 27.89, "grad_norm": 0.4976441264152527, "learning_rate": 0.001721111111111111, "loss": 1.7642, "step": 7530 }, { "epoch": 27.93, "grad_norm": 0.48872360587120056, "learning_rate": 0.0017207407407407408, "loss": 1.8113, "step": 7540 }, { "epoch": 27.96, "grad_norm": 0.5321819186210632, "learning_rate": 0.0017203703703703704, "loss": 1.8143, "step": 7550 }, { "epoch": 28.0, "grad_norm": 1.1776005029678345, "learning_rate": 0.00172, "loss": 1.8321, "step": 7560 }, { "epoch": 28.04, "grad_norm": 0.5432651042938232, "learning_rate": 0.0017196296296296298, "loss": 1.5738, "step": 7570 }, { "epoch": 28.07, "grad_norm": 0.512981653213501, "learning_rate": 0.0017192592592592594, "loss": 1.5647, "step": 7580 }, { "epoch": 28.11, "grad_norm": 0.5530872344970703, "learning_rate": 0.001718888888888889, "loss": 1.5745, "step": 7590 }, { "epoch": 28.15, "grad_norm": 0.5536510944366455, "learning_rate": 0.0017185185185185185, "loss": 1.603, "step": 7600 }, { "epoch": 28.19, "grad_norm": 0.5366228222846985, "learning_rate": 0.0017181481481481483, "loss": 1.6581, "step": 7610 }, { "epoch": 28.22, "grad_norm": 0.631415843963623, "learning_rate": 0.001717777777777778, "loss": 1.6106, "step": 7620 }, { "epoch": 28.26, "grad_norm": 0.5335027575492859, "learning_rate": 0.0017174074074074075, "loss": 1.6436, "step": 7630 }, { "epoch": 28.3, "grad_norm": 0.5051824450492859, "learning_rate": 0.001717037037037037, "loss": 1.6577, "step": 7640 }, { "epoch": 28.33, "grad_norm": 0.5768982768058777, "learning_rate": 0.0017166666666666667, "loss": 1.6765, "step": 7650 }, { "epoch": 28.37, "grad_norm": 0.5530310869216919, "learning_rate": 0.0017162962962962963, "loss": 1.6831, "step": 7660 }, { "epoch": 28.41, "grad_norm": 0.5534545183181763, "learning_rate": 0.0017159259259259259, "loss": 1.7275, "step": 7670 }, { "epoch": 28.44, "grad_norm": 0.5321471095085144, "learning_rate": 0.0017155555555555555, "loss": 1.7241, "step": 7680 }, { "epoch": 28.48, "grad_norm": 0.5455407500267029, "learning_rate": 0.0017151851851851853, "loss": 1.7278, "step": 7690 }, { "epoch": 28.52, "grad_norm": 0.5250757336616516, "learning_rate": 0.0017148148148148148, "loss": 1.7389, "step": 7700 }, { "epoch": 28.56, "grad_norm": 0.5096060037612915, "learning_rate": 0.0017144444444444444, "loss": 1.7569, "step": 7710 }, { "epoch": 28.59, "grad_norm": 0.5587198138237, "learning_rate": 0.001714074074074074, "loss": 1.7363, "step": 7720 }, { "epoch": 28.63, "grad_norm": 0.510917067527771, "learning_rate": 0.0017137037037037038, "loss": 1.7823, "step": 7730 }, { "epoch": 28.67, "grad_norm": 0.5196210741996765, "learning_rate": 0.0017133333333333334, "loss": 1.7751, "step": 7740 }, { "epoch": 28.7, "grad_norm": 0.5318698883056641, "learning_rate": 0.001712962962962963, "loss": 1.7504, "step": 7750 }, { "epoch": 28.74, "grad_norm": 0.5652453303337097, "learning_rate": 0.0017125925925925926, "loss": 1.7604, "step": 7760 }, { "epoch": 28.78, "grad_norm": 0.5259050130844116, "learning_rate": 0.0017122222222222224, "loss": 1.7665, "step": 7770 }, { "epoch": 28.81, "grad_norm": 0.5152835249900818, "learning_rate": 0.001711851851851852, "loss": 1.7487, "step": 7780 }, { "epoch": 28.85, "grad_norm": 0.5386359691619873, "learning_rate": 0.0017114814814814814, "loss": 1.8161, "step": 7790 }, { "epoch": 28.89, "grad_norm": 0.514400064945221, "learning_rate": 0.0017111111111111112, "loss": 1.795, "step": 7800 }, { "epoch": 28.93, "grad_norm": 0.5309016108512878, "learning_rate": 0.0017107407407407408, "loss": 1.7683, "step": 7810 }, { "epoch": 28.96, "grad_norm": 0.5700574517250061, "learning_rate": 0.0017103703703703703, "loss": 1.8102, "step": 7820 }, { "epoch": 29.0, "grad_norm": 1.172451138496399, "learning_rate": 0.00171, "loss": 1.7893, "step": 7830 }, { "epoch": 29.04, "grad_norm": 0.5429467558860779, "learning_rate": 0.0017096296296296297, "loss": 1.5388, "step": 7840 }, { "epoch": 29.07, "grad_norm": 0.5805901288986206, "learning_rate": 0.0017092592592592593, "loss": 1.5459, "step": 7850 }, { "epoch": 29.11, "grad_norm": 0.539729118347168, "learning_rate": 0.001708888888888889, "loss": 1.5985, "step": 7860 }, { "epoch": 29.15, "grad_norm": 0.5773619413375854, "learning_rate": 0.0017085185185185185, "loss": 1.5833, "step": 7870 }, { "epoch": 29.19, "grad_norm": 0.564813494682312, "learning_rate": 0.0017081481481481483, "loss": 1.6251, "step": 7880 }, { "epoch": 29.22, "grad_norm": 0.5339441299438477, "learning_rate": 0.001707777777777778, "loss": 1.6363, "step": 7890 }, { "epoch": 29.26, "grad_norm": 0.5444590449333191, "learning_rate": 0.0017074074074074075, "loss": 1.653, "step": 7900 }, { "epoch": 29.3, "grad_norm": 0.5453532934188843, "learning_rate": 0.001707037037037037, "loss": 1.6631, "step": 7910 }, { "epoch": 29.33, "grad_norm": 0.5633497834205627, "learning_rate": 0.0017066666666666669, "loss": 1.6852, "step": 7920 }, { "epoch": 29.37, "grad_norm": 0.506697416305542, "learning_rate": 0.0017062962962962963, "loss": 1.6729, "step": 7930 }, { "epoch": 29.41, "grad_norm": 0.558829665184021, "learning_rate": 0.0017059259259259258, "loss": 1.6851, "step": 7940 }, { "epoch": 29.44, "grad_norm": 0.5217764377593994, "learning_rate": 0.0017055555555555554, "loss": 1.6967, "step": 7950 }, { "epoch": 29.48, "grad_norm": 0.5348731875419617, "learning_rate": 0.0017051851851851852, "loss": 1.6994, "step": 7960 }, { "epoch": 29.52, "grad_norm": 0.5449881553649902, "learning_rate": 0.0017048148148148148, "loss": 1.6906, "step": 7970 }, { "epoch": 29.56, "grad_norm": 0.5640777349472046, "learning_rate": 0.0017044444444444444, "loss": 1.7228, "step": 7980 }, { "epoch": 29.59, "grad_norm": 0.5632838606834412, "learning_rate": 0.0017040740740740742, "loss": 1.7475, "step": 7990 }, { "epoch": 29.63, "grad_norm": 0.514080822467804, "learning_rate": 0.0017037037037037038, "loss": 1.7825, "step": 8000 }, { "epoch": 29.67, "grad_norm": 0.5150419473648071, "learning_rate": 0.0017033333333333334, "loss": 1.735, "step": 8010 }, { "epoch": 29.7, "grad_norm": 0.5598604083061218, "learning_rate": 0.001702962962962963, "loss": 1.7627, "step": 8020 }, { "epoch": 29.74, "grad_norm": 0.5505332350730896, "learning_rate": 0.0017025925925925928, "loss": 1.7434, "step": 8030 }, { "epoch": 29.78, "grad_norm": 0.5757072567939758, "learning_rate": 0.0017022222222222224, "loss": 1.7673, "step": 8040 }, { "epoch": 29.81, "grad_norm": 0.5394684672355652, "learning_rate": 0.001701851851851852, "loss": 1.7808, "step": 8050 }, { "epoch": 29.85, "grad_norm": 0.5051487684249878, "learning_rate": 0.0017014814814814813, "loss": 1.779, "step": 8060 }, { "epoch": 29.89, "grad_norm": 0.45365673303604126, "learning_rate": 0.0017011111111111111, "loss": 1.7741, "step": 8070 }, { "epoch": 29.93, "grad_norm": 0.5558077692985535, "learning_rate": 0.0017007407407407407, "loss": 1.7883, "step": 8080 }, { "epoch": 29.96, "grad_norm": 0.5742713212966919, "learning_rate": 0.0017003703703703703, "loss": 1.7801, "step": 8090 }, { "epoch": 30.0, "grad_norm": 0.9552031755447388, "learning_rate": 0.0017, "loss": 1.7876, "step": 8100 }, { "epoch": 30.04, "grad_norm": 0.5495716333389282, "learning_rate": 0.0016996296296296297, "loss": 1.4899, "step": 8110 }, { "epoch": 30.07, "grad_norm": 0.5539374947547913, "learning_rate": 0.0016992592592592593, "loss": 1.5593, "step": 8120 }, { "epoch": 30.11, "grad_norm": 0.6192070841789246, "learning_rate": 0.0016988888888888889, "loss": 1.5557, "step": 8130 }, { "epoch": 30.15, "grad_norm": 0.5558332204818726, "learning_rate": 0.0016985185185185185, "loss": 1.5884, "step": 8140 }, { "epoch": 30.19, "grad_norm": 0.5470682978630066, "learning_rate": 0.0016981481481481483, "loss": 1.5894, "step": 8150 }, { "epoch": 30.22, "grad_norm": 0.547297477722168, "learning_rate": 0.0016977777777777779, "loss": 1.6244, "step": 8160 }, { "epoch": 30.26, "grad_norm": 0.577240526676178, "learning_rate": 0.0016974074074074075, "loss": 1.6502, "step": 8170 }, { "epoch": 30.3, "grad_norm": 0.5513033270835876, "learning_rate": 0.0016970370370370373, "loss": 1.6443, "step": 8180 }, { "epoch": 30.33, "grad_norm": 0.5598578453063965, "learning_rate": 0.0016966666666666669, "loss": 1.6471, "step": 8190 }, { "epoch": 30.37, "grad_norm": 0.5605131387710571, "learning_rate": 0.0016962962962962962, "loss": 1.6755, "step": 8200 }, { "epoch": 30.41, "grad_norm": 0.5713827013969421, "learning_rate": 0.0016959259259259258, "loss": 1.6753, "step": 8210 }, { "epoch": 30.44, "grad_norm": 0.5782068967819214, "learning_rate": 0.0016955555555555556, "loss": 1.7284, "step": 8220 }, { "epoch": 30.48, "grad_norm": 0.5492303371429443, "learning_rate": 0.0016951851851851852, "loss": 1.6863, "step": 8230 }, { "epoch": 30.52, "grad_norm": 0.5332817435264587, "learning_rate": 0.0016948148148148148, "loss": 1.702, "step": 8240 }, { "epoch": 30.56, "grad_norm": 0.5532167553901672, "learning_rate": 0.0016944444444444444, "loss": 1.7317, "step": 8250 }, { "epoch": 30.59, "grad_norm": 0.566382110118866, "learning_rate": 0.0016940740740740742, "loss": 1.7365, "step": 8260 }, { "epoch": 30.63, "grad_norm": 0.542638897895813, "learning_rate": 0.0016937037037037038, "loss": 1.7288, "step": 8270 }, { "epoch": 30.67, "grad_norm": 0.5863054990768433, "learning_rate": 0.0016933333333333334, "loss": 1.7207, "step": 8280 }, { "epoch": 30.7, "grad_norm": 0.5963066816329956, "learning_rate": 0.001692962962962963, "loss": 1.7264, "step": 8290 }, { "epoch": 30.74, "grad_norm": 0.5126426219940186, "learning_rate": 0.0016925925925925928, "loss": 1.7056, "step": 8300 }, { "epoch": 30.78, "grad_norm": 0.5691437125205994, "learning_rate": 0.0016922222222222223, "loss": 1.7685, "step": 8310 }, { "epoch": 30.81, "grad_norm": 0.4958675801753998, "learning_rate": 0.001691851851851852, "loss": 1.7537, "step": 8320 }, { "epoch": 30.85, "grad_norm": 0.5244887471199036, "learning_rate": 0.0016914814814814815, "loss": 1.7765, "step": 8330 }, { "epoch": 30.89, "grad_norm": 0.4961899220943451, "learning_rate": 0.0016911111111111111, "loss": 1.7801, "step": 8340 }, { "epoch": 30.93, "grad_norm": 0.5399650931358337, "learning_rate": 0.0016907407407407407, "loss": 1.7591, "step": 8350 }, { "epoch": 30.96, "grad_norm": 0.549034595489502, "learning_rate": 0.0016903703703703703, "loss": 1.7858, "step": 8360 }, { "epoch": 31.0, "grad_norm": 1.2998181581497192, "learning_rate": 0.0016899999999999999, "loss": 1.8002, "step": 8370 }, { "epoch": 31.04, "grad_norm": 0.5629531145095825, "learning_rate": 0.0016896296296296297, "loss": 1.5119, "step": 8380 }, { "epoch": 31.07, "grad_norm": 0.5641016960144043, "learning_rate": 0.0016892592592592593, "loss": 1.5397, "step": 8390 }, { "epoch": 31.11, "grad_norm": 0.5819301605224609, "learning_rate": 0.0016888888888888889, "loss": 1.5523, "step": 8400 }, { "epoch": 31.15, "grad_norm": 0.5689669251441956, "learning_rate": 0.0016885185185185187, "loss": 1.5829, "step": 8410 }, { "epoch": 31.19, "grad_norm": 0.5574280023574829, "learning_rate": 0.0016881481481481483, "loss": 1.5748, "step": 8420 }, { "epoch": 31.22, "grad_norm": 0.6067005395889282, "learning_rate": 0.0016877777777777778, "loss": 1.6045, "step": 8430 }, { "epoch": 31.26, "grad_norm": 0.5879372358322144, "learning_rate": 0.0016874074074074074, "loss": 1.6307, "step": 8440 }, { "epoch": 31.3, "grad_norm": 0.6153671145439148, "learning_rate": 0.0016870370370370372, "loss": 1.6503, "step": 8450 }, { "epoch": 31.33, "grad_norm": 0.5140969753265381, "learning_rate": 0.0016866666666666668, "loss": 1.6719, "step": 8460 }, { "epoch": 31.37, "grad_norm": 0.5702369809150696, "learning_rate": 0.0016862962962962964, "loss": 1.6727, "step": 8470 }, { "epoch": 31.41, "grad_norm": 0.5763192176818848, "learning_rate": 0.0016859259259259258, "loss": 1.6691, "step": 8480 }, { "epoch": 31.44, "grad_norm": 0.5894213318824768, "learning_rate": 0.0016855555555555556, "loss": 1.69, "step": 8490 }, { "epoch": 31.48, "grad_norm": 0.5990567207336426, "learning_rate": 0.0016851851851851852, "loss": 1.6665, "step": 8500 }, { "epoch": 31.52, "grad_norm": 0.5143067240715027, "learning_rate": 0.0016848148148148148, "loss": 1.678, "step": 8510 }, { "epoch": 31.56, "grad_norm": 0.5345697402954102, "learning_rate": 0.0016844444444444444, "loss": 1.708, "step": 8520 }, { "epoch": 31.59, "grad_norm": 0.5773447751998901, "learning_rate": 0.0016840740740740742, "loss": 1.7029, "step": 8530 }, { "epoch": 31.63, "grad_norm": 0.5366548299789429, "learning_rate": 0.0016837037037037038, "loss": 1.7183, "step": 8540 }, { "epoch": 31.67, "grad_norm": 0.5319424867630005, "learning_rate": 0.0016833333333333333, "loss": 1.7219, "step": 8550 }, { "epoch": 31.7, "grad_norm": 0.5273670554161072, "learning_rate": 0.001682962962962963, "loss": 1.7296, "step": 8560 }, { "epoch": 31.74, "grad_norm": 0.5865380764007568, "learning_rate": 0.0016825925925925927, "loss": 1.7321, "step": 8570 }, { "epoch": 31.78, "grad_norm": 0.5639680624008179, "learning_rate": 0.0016822222222222223, "loss": 1.7335, "step": 8580 }, { "epoch": 31.81, "grad_norm": 0.5686609148979187, "learning_rate": 0.001681851851851852, "loss": 1.7393, "step": 8590 }, { "epoch": 31.85, "grad_norm": 0.5990113019943237, "learning_rate": 0.0016814814814814817, "loss": 1.7704, "step": 8600 }, { "epoch": 31.89, "grad_norm": 0.4920467734336853, "learning_rate": 0.001681111111111111, "loss": 1.7624, "step": 8610 }, { "epoch": 31.93, "grad_norm": 0.5353578925132751, "learning_rate": 0.0016807407407407407, "loss": 1.7495, "step": 8620 }, { "epoch": 31.96, "grad_norm": 0.503509521484375, "learning_rate": 0.0016803703703703703, "loss": 1.7685, "step": 8630 }, { "epoch": 32.0, "grad_norm": 1.060388207435608, "learning_rate": 0.00168, "loss": 1.8034, "step": 8640 }, { "epoch": 32.04, "grad_norm": 0.5331477522850037, "learning_rate": 0.0016796296296296297, "loss": 1.502, "step": 8650 }, { "epoch": 32.07, "grad_norm": 0.5541171431541443, "learning_rate": 0.0016792592592592592, "loss": 1.5502, "step": 8660 }, { "epoch": 32.11, "grad_norm": 0.5719496607780457, "learning_rate": 0.0016788888888888888, "loss": 1.5465, "step": 8670 }, { "epoch": 32.15, "grad_norm": 0.5539078116416931, "learning_rate": 0.0016785185185185186, "loss": 1.5404, "step": 8680 }, { "epoch": 32.19, "grad_norm": 0.5710481405258179, "learning_rate": 0.0016781481481481482, "loss": 1.5532, "step": 8690 }, { "epoch": 32.22, "grad_norm": 0.5715047717094421, "learning_rate": 0.0016777777777777778, "loss": 1.5976, "step": 8700 }, { "epoch": 32.26, "grad_norm": 0.5798360109329224, "learning_rate": 0.0016774074074074074, "loss": 1.6342, "step": 8710 }, { "epoch": 32.3, "grad_norm": 0.5830899477005005, "learning_rate": 0.0016770370370370372, "loss": 1.6217, "step": 8720 }, { "epoch": 32.33, "grad_norm": 0.5676963329315186, "learning_rate": 0.0016766666666666668, "loss": 1.6163, "step": 8730 }, { "epoch": 32.37, "grad_norm": 0.586168646812439, "learning_rate": 0.0016762962962962964, "loss": 1.6336, "step": 8740 }, { "epoch": 32.41, "grad_norm": 0.5690294504165649, "learning_rate": 0.0016759259259259258, "loss": 1.6528, "step": 8750 }, { "epoch": 32.44, "grad_norm": 0.5214428901672363, "learning_rate": 0.0016755555555555556, "loss": 1.6664, "step": 8760 }, { "epoch": 32.48, "grad_norm": 0.5383786559104919, "learning_rate": 0.0016751851851851852, "loss": 1.6719, "step": 8770 }, { "epoch": 32.52, "grad_norm": 0.6171760559082031, "learning_rate": 0.0016748148148148147, "loss": 1.6992, "step": 8780 }, { "epoch": 32.56, "grad_norm": 0.5024405121803284, "learning_rate": 0.0016744444444444443, "loss": 1.6858, "step": 8790 }, { "epoch": 32.59, "grad_norm": 0.5647597908973694, "learning_rate": 0.0016740740740740741, "loss": 1.6923, "step": 8800 }, { "epoch": 32.63, "grad_norm": 0.6205365657806396, "learning_rate": 0.0016737037037037037, "loss": 1.7016, "step": 8810 }, { "epoch": 32.67, "grad_norm": 0.5677077770233154, "learning_rate": 0.0016733333333333333, "loss": 1.7227, "step": 8820 }, { "epoch": 32.7, "grad_norm": 0.5691731572151184, "learning_rate": 0.0016729629629629631, "loss": 1.6843, "step": 8830 }, { "epoch": 32.74, "grad_norm": 0.5742966532707214, "learning_rate": 0.0016725925925925927, "loss": 1.7423, "step": 8840 }, { "epoch": 32.78, "grad_norm": 0.558135986328125, "learning_rate": 0.0016722222222222223, "loss": 1.7457, "step": 8850 }, { "epoch": 32.81, "grad_norm": 0.5381272435188293, "learning_rate": 0.0016718518518518519, "loss": 1.7229, "step": 8860 }, { "epoch": 32.85, "grad_norm": 0.53874272108078, "learning_rate": 0.0016714814814814817, "loss": 1.7586, "step": 8870 }, { "epoch": 32.89, "grad_norm": 0.559933602809906, "learning_rate": 0.0016711111111111113, "loss": 1.7902, "step": 8880 }, { "epoch": 32.93, "grad_norm": 0.5527853965759277, "learning_rate": 0.0016707407407407406, "loss": 1.7616, "step": 8890 }, { "epoch": 32.96, "grad_norm": 0.600522518157959, "learning_rate": 0.0016703703703703702, "loss": 1.7471, "step": 8900 }, { "epoch": 33.0, "grad_norm": 1.1233681440353394, "learning_rate": 0.00167, "loss": 1.7853, "step": 8910 }, { "epoch": 33.04, "grad_norm": 0.5869570970535278, "learning_rate": 0.0016696296296296296, "loss": 1.5254, "step": 8920 }, { "epoch": 33.07, "grad_norm": 0.5793870687484741, "learning_rate": 0.0016692592592592592, "loss": 1.52, "step": 8930 }, { "epoch": 33.11, "grad_norm": 0.5554567575454712, "learning_rate": 0.0016688888888888888, "loss": 1.5477, "step": 8940 }, { "epoch": 33.15, "grad_norm": 0.5523855686187744, "learning_rate": 0.0016685185185185186, "loss": 1.5726, "step": 8950 }, { "epoch": 33.19, "grad_norm": 0.5525711178779602, "learning_rate": 0.0016681481481481482, "loss": 1.5859, "step": 8960 }, { "epoch": 33.22, "grad_norm": 0.593908429145813, "learning_rate": 0.0016677777777777778, "loss": 1.5876, "step": 8970 }, { "epoch": 33.26, "grad_norm": 0.5583921074867249, "learning_rate": 0.0016674074074074074, "loss": 1.5976, "step": 8980 }, { "epoch": 33.3, "grad_norm": 0.5856320858001709, "learning_rate": 0.0016670370370370372, "loss": 1.6249, "step": 8990 }, { "epoch": 33.33, "grad_norm": 0.5795212984085083, "learning_rate": 0.0016666666666666668, "loss": 1.6162, "step": 9000 }, { "epoch": 33.37, "grad_norm": 0.6002980470657349, "learning_rate": 0.0016662962962962964, "loss": 1.6169, "step": 9010 }, { "epoch": 33.41, "grad_norm": 0.60965895652771, "learning_rate": 0.0016659259259259262, "loss": 1.6507, "step": 9020 }, { "epoch": 33.44, "grad_norm": 0.568634569644928, "learning_rate": 0.0016655555555555555, "loss": 1.6778, "step": 9030 }, { "epoch": 33.48, "grad_norm": 0.5564097166061401, "learning_rate": 0.0016651851851851851, "loss": 1.6964, "step": 9040 }, { "epoch": 33.52, "grad_norm": 0.5448118448257446, "learning_rate": 0.0016648148148148147, "loss": 1.6412, "step": 9050 }, { "epoch": 33.56, "grad_norm": 0.5949601531028748, "learning_rate": 0.0016644444444444445, "loss": 1.6647, "step": 9060 }, { "epoch": 33.59, "grad_norm": 0.5293440222740173, "learning_rate": 0.0016640740740740741, "loss": 1.6952, "step": 9070 }, { "epoch": 33.63, "grad_norm": 0.6145957112312317, "learning_rate": 0.0016637037037037037, "loss": 1.7021, "step": 9080 }, { "epoch": 33.67, "grad_norm": 0.5578882098197937, "learning_rate": 0.0016633333333333333, "loss": 1.7205, "step": 9090 }, { "epoch": 33.7, "grad_norm": 0.5359580516815186, "learning_rate": 0.001662962962962963, "loss": 1.7085, "step": 9100 }, { "epoch": 33.74, "grad_norm": 0.5765175223350525, "learning_rate": 0.0016625925925925927, "loss": 1.6787, "step": 9110 }, { "epoch": 33.78, "grad_norm": 0.517185389995575, "learning_rate": 0.0016622222222222223, "loss": 1.7226, "step": 9120 }, { "epoch": 33.81, "grad_norm": 0.5444340109825134, "learning_rate": 0.0016618518518518519, "loss": 1.7066, "step": 9130 }, { "epoch": 33.85, "grad_norm": 0.5430245399475098, "learning_rate": 0.0016614814814814817, "loss": 1.6819, "step": 9140 }, { "epoch": 33.89, "grad_norm": 0.5591315627098083, "learning_rate": 0.0016611111111111113, "loss": 1.7322, "step": 9150 }, { "epoch": 33.93, "grad_norm": 0.5514317154884338, "learning_rate": 0.0016607407407407408, "loss": 1.7251, "step": 9160 }, { "epoch": 33.96, "grad_norm": 0.6007348895072937, "learning_rate": 0.0016603703703703702, "loss": 1.7792, "step": 9170 }, { "epoch": 34.0, "grad_norm": 0.8949159979820251, "learning_rate": 0.00166, "loss": 1.7497, "step": 9180 }, { "epoch": 34.04, "grad_norm": 0.5425311326980591, "learning_rate": 0.0016596296296296296, "loss": 1.4758, "step": 9190 }, { "epoch": 34.07, "grad_norm": 0.606648325920105, "learning_rate": 0.0016592592592592592, "loss": 1.5247, "step": 9200 }, { "epoch": 34.11, "grad_norm": 0.542310357093811, "learning_rate": 0.001658888888888889, "loss": 1.5363, "step": 9210 }, { "epoch": 34.15, "grad_norm": 0.5714516043663025, "learning_rate": 0.0016585185185185186, "loss": 1.5374, "step": 9220 }, { "epoch": 34.19, "grad_norm": 0.6230427622795105, "learning_rate": 0.0016581481481481482, "loss": 1.5204, "step": 9230 }, { "epoch": 34.22, "grad_norm": 0.5695710182189941, "learning_rate": 0.0016577777777777778, "loss": 1.576, "step": 9240 }, { "epoch": 34.26, "grad_norm": 0.5765951871871948, "learning_rate": 0.0016574074074074076, "loss": 1.5759, "step": 9250 }, { "epoch": 34.3, "grad_norm": 0.5947939157485962, "learning_rate": 0.0016570370370370372, "loss": 1.608, "step": 9260 }, { "epoch": 34.33, "grad_norm": 0.607772171497345, "learning_rate": 0.0016566666666666667, "loss": 1.6108, "step": 9270 }, { "epoch": 34.37, "grad_norm": 0.6625462770462036, "learning_rate": 0.0016562962962962963, "loss": 1.6323, "step": 9280 }, { "epoch": 34.41, "grad_norm": 0.5714110732078552, "learning_rate": 0.0016559259259259261, "loss": 1.6922, "step": 9290 }, { "epoch": 34.44, "grad_norm": 0.5694931149482727, "learning_rate": 0.0016555555555555555, "loss": 1.6251, "step": 9300 }, { "epoch": 34.48, "grad_norm": 0.6467699408531189, "learning_rate": 0.001655185185185185, "loss": 1.6503, "step": 9310 }, { "epoch": 34.52, "grad_norm": 0.6535106301307678, "learning_rate": 0.0016548148148148147, "loss": 1.6817, "step": 9320 }, { "epoch": 34.56, "grad_norm": 0.5543297529220581, "learning_rate": 0.0016544444444444445, "loss": 1.6625, "step": 9330 }, { "epoch": 34.59, "grad_norm": 0.5453930497169495, "learning_rate": 0.001654074074074074, "loss": 1.6692, "step": 9340 }, { "epoch": 34.63, "grad_norm": 0.5699336528778076, "learning_rate": 0.0016537037037037037, "loss": 1.685, "step": 9350 }, { "epoch": 34.67, "grad_norm": 0.5972739458084106, "learning_rate": 0.0016533333333333333, "loss": 1.6944, "step": 9360 }, { "epoch": 34.7, "grad_norm": 0.6316691040992737, "learning_rate": 0.001652962962962963, "loss": 1.6992, "step": 9370 }, { "epoch": 34.74, "grad_norm": 0.5370445847511292, "learning_rate": 0.0016525925925925927, "loss": 1.7141, "step": 9380 }, { "epoch": 34.78, "grad_norm": 0.5576089024543762, "learning_rate": 0.0016522222222222222, "loss": 1.7085, "step": 9390 }, { "epoch": 34.81, "grad_norm": 0.5875248312950134, "learning_rate": 0.0016518518518518518, "loss": 1.7217, "step": 9400 }, { "epoch": 34.85, "grad_norm": 0.5235693454742432, "learning_rate": 0.0016514814814814816, "loss": 1.7026, "step": 9410 }, { "epoch": 34.89, "grad_norm": 0.6246321797370911, "learning_rate": 0.0016511111111111112, "loss": 1.7451, "step": 9420 }, { "epoch": 34.93, "grad_norm": 0.5568575859069824, "learning_rate": 0.0016507407407407408, "loss": 1.7116, "step": 9430 }, { "epoch": 34.96, "grad_norm": 0.5829682350158691, "learning_rate": 0.0016503703703703704, "loss": 1.7469, "step": 9440 }, { "epoch": 35.0, "grad_norm": 1.1844964027404785, "learning_rate": 0.00165, "loss": 1.7515, "step": 9450 }, { "epoch": 35.04, "grad_norm": 0.5359990000724792, "learning_rate": 0.0016496296296296296, "loss": 1.4927, "step": 9460 }, { "epoch": 35.07, "grad_norm": 0.5757229924201965, "learning_rate": 0.0016492592592592592, "loss": 1.5032, "step": 9470 }, { "epoch": 35.11, "grad_norm": 0.615017294883728, "learning_rate": 0.001648888888888889, "loss": 1.5318, "step": 9480 }, { "epoch": 35.15, "grad_norm": 0.540900707244873, "learning_rate": 0.0016485185185185186, "loss": 1.5309, "step": 9490 }, { "epoch": 35.19, "grad_norm": 0.589608371257782, "learning_rate": 0.0016481481481481482, "loss": 1.5338, "step": 9500 }, { "epoch": 35.22, "grad_norm": 0.5594486594200134, "learning_rate": 0.0016477777777777777, "loss": 1.5788, "step": 9510 }, { "epoch": 35.26, "grad_norm": 0.6130346059799194, "learning_rate": 0.0016474074074074075, "loss": 1.612, "step": 9520 }, { "epoch": 35.3, "grad_norm": 0.5725595355033875, "learning_rate": 0.0016470370370370371, "loss": 1.5903, "step": 9530 }, { "epoch": 35.33, "grad_norm": 0.5720920562744141, "learning_rate": 0.0016466666666666667, "loss": 1.6188, "step": 9540 }, { "epoch": 35.37, "grad_norm": 0.5802409648895264, "learning_rate": 0.0016462962962962963, "loss": 1.5857, "step": 9550 }, { "epoch": 35.41, "grad_norm": 0.6417401432991028, "learning_rate": 0.0016459259259259261, "loss": 1.6185, "step": 9560 }, { "epoch": 35.44, "grad_norm": 0.6042072772979736, "learning_rate": 0.0016455555555555557, "loss": 1.6666, "step": 9570 }, { "epoch": 35.48, "grad_norm": 0.6269539594650269, "learning_rate": 0.001645185185185185, "loss": 1.6404, "step": 9580 }, { "epoch": 35.52, "grad_norm": 0.5669009685516357, "learning_rate": 0.0016448148148148147, "loss": 1.6734, "step": 9590 }, { "epoch": 35.56, "grad_norm": 0.5852543115615845, "learning_rate": 0.0016444444444444445, "loss": 1.6635, "step": 9600 }, { "epoch": 35.59, "grad_norm": 0.6072357296943665, "learning_rate": 0.001644074074074074, "loss": 1.6624, "step": 9610 }, { "epoch": 35.63, "grad_norm": 0.632712721824646, "learning_rate": 0.0016437037037037036, "loss": 1.6884, "step": 9620 }, { "epoch": 35.67, "grad_norm": 0.5859778523445129, "learning_rate": 0.0016433333333333335, "loss": 1.6693, "step": 9630 }, { "epoch": 35.7, "grad_norm": 0.5471202731132507, "learning_rate": 0.001642962962962963, "loss": 1.7173, "step": 9640 }, { "epoch": 35.74, "grad_norm": 0.5577960014343262, "learning_rate": 0.0016425925925925926, "loss": 1.6699, "step": 9650 }, { "epoch": 35.78, "grad_norm": 0.5657382607460022, "learning_rate": 0.0016422222222222222, "loss": 1.7091, "step": 9660 }, { "epoch": 35.81, "grad_norm": 0.5789744257926941, "learning_rate": 0.001641851851851852, "loss": 1.6923, "step": 9670 }, { "epoch": 35.85, "grad_norm": 0.5898125767707825, "learning_rate": 0.0016414814814814816, "loss": 1.6982, "step": 9680 }, { "epoch": 35.89, "grad_norm": 0.5303668975830078, "learning_rate": 0.0016411111111111112, "loss": 1.722, "step": 9690 }, { "epoch": 35.93, "grad_norm": 0.5981086492538452, "learning_rate": 0.0016407407407407408, "loss": 1.7036, "step": 9700 }, { "epoch": 35.96, "grad_norm": 0.5202363729476929, "learning_rate": 0.0016403703703703706, "loss": 1.7071, "step": 9710 }, { "epoch": 36.0, "grad_norm": 1.0010344982147217, "learning_rate": 0.00164, "loss": 1.7195, "step": 9720 }, { "epoch": 36.04, "grad_norm": 0.5903262495994568, "learning_rate": 0.0016396296296296296, "loss": 1.4546, "step": 9730 }, { "epoch": 36.07, "grad_norm": 0.5673249959945679, "learning_rate": 0.0016392592592592591, "loss": 1.4846, "step": 9740 }, { "epoch": 36.11, "grad_norm": 0.623580813407898, "learning_rate": 0.001638888888888889, "loss": 1.5554, "step": 9750 }, { "epoch": 36.15, "grad_norm": 0.5951016545295715, "learning_rate": 0.0016385185185185185, "loss": 1.5375, "step": 9760 }, { "epoch": 36.19, "grad_norm": 0.5958836078643799, "learning_rate": 0.0016381481481481481, "loss": 1.5261, "step": 9770 }, { "epoch": 36.22, "grad_norm": 0.6242657899856567, "learning_rate": 0.0016377777777777777, "loss": 1.5417, "step": 9780 }, { "epoch": 36.26, "grad_norm": 0.6229197978973389, "learning_rate": 0.0016374074074074075, "loss": 1.587, "step": 9790 }, { "epoch": 36.3, "grad_norm": 0.5648860931396484, "learning_rate": 0.001637037037037037, "loss": 1.5499, "step": 9800 }, { "epoch": 36.33, "grad_norm": 0.5891751646995544, "learning_rate": 0.0016366666666666667, "loss": 1.5852, "step": 9810 }, { "epoch": 36.37, "grad_norm": 0.8289972543716431, "learning_rate": 0.0016362962962962965, "loss": 1.5974, "step": 9820 }, { "epoch": 36.41, "grad_norm": 0.6231335401535034, "learning_rate": 0.001635925925925926, "loss": 1.6485, "step": 9830 }, { "epoch": 36.44, "grad_norm": 0.5463785529136658, "learning_rate": 0.0016355555555555557, "loss": 1.6329, "step": 9840 }, { "epoch": 36.48, "grad_norm": 0.5732906460762024, "learning_rate": 0.001635185185185185, "loss": 1.6287, "step": 9850 }, { "epoch": 36.52, "grad_norm": 0.5856965780258179, "learning_rate": 0.0016348148148148149, "loss": 1.6757, "step": 9860 }, { "epoch": 36.56, "grad_norm": 0.5316782593727112, "learning_rate": 0.0016344444444444444, "loss": 1.6571, "step": 9870 }, { "epoch": 36.59, "grad_norm": 0.6029806733131409, "learning_rate": 0.001634074074074074, "loss": 1.6276, "step": 9880 }, { "epoch": 36.63, "grad_norm": 0.5773792862892151, "learning_rate": 0.0016337037037037036, "loss": 1.6763, "step": 9890 }, { "epoch": 36.67, "grad_norm": 0.5454092621803284, "learning_rate": 0.0016333333333333334, "loss": 1.6742, "step": 9900 }, { "epoch": 36.7, "grad_norm": 0.560020923614502, "learning_rate": 0.001632962962962963, "loss": 1.6871, "step": 9910 }, { "epoch": 36.74, "grad_norm": 0.5670827031135559, "learning_rate": 0.0016325925925925926, "loss": 1.6921, "step": 9920 }, { "epoch": 36.78, "grad_norm": 0.5637957453727722, "learning_rate": 0.0016322222222222222, "loss": 1.6752, "step": 9930 }, { "epoch": 36.81, "grad_norm": 0.5903841257095337, "learning_rate": 0.001631851851851852, "loss": 1.6807, "step": 9940 }, { "epoch": 36.85, "grad_norm": 0.5725448727607727, "learning_rate": 0.0016314814814814816, "loss": 1.7029, "step": 9950 }, { "epoch": 36.89, "grad_norm": 0.554770290851593, "learning_rate": 0.0016311111111111112, "loss": 1.7361, "step": 9960 }, { "epoch": 36.93, "grad_norm": 0.656335711479187, "learning_rate": 0.0016307407407407408, "loss": 1.7117, "step": 9970 }, { "epoch": 36.96, "grad_norm": 0.5630307793617249, "learning_rate": 0.0016303703703703706, "loss": 1.6824, "step": 9980 }, { "epoch": 37.0, "grad_norm": 1.2158167362213135, "learning_rate": 0.00163, "loss": 1.7112, "step": 9990 }, { "epoch": 37.04, "grad_norm": 0.5833454132080078, "learning_rate": 0.0016296296296296295, "loss": 1.4483, "step": 10000 }, { "epoch": 37.07, "grad_norm": 0.5893401503562927, "learning_rate": 0.0016292592592592591, "loss": 1.4812, "step": 10010 }, { "epoch": 37.11, "grad_norm": 0.5967376828193665, "learning_rate": 0.001628888888888889, "loss": 1.5071, "step": 10020 }, { "epoch": 37.15, "grad_norm": 0.572011411190033, "learning_rate": 0.0016285185185185185, "loss": 1.4945, "step": 10030 }, { "epoch": 37.19, "grad_norm": 0.6522864699363708, "learning_rate": 0.001628148148148148, "loss": 1.5245, "step": 10040 }, { "epoch": 37.22, "grad_norm": 0.5800381898880005, "learning_rate": 0.001627777777777778, "loss": 1.5355, "step": 10050 }, { "epoch": 37.26, "grad_norm": 0.6092601418495178, "learning_rate": 0.0016274074074074075, "loss": 1.5551, "step": 10060 }, { "epoch": 37.3, "grad_norm": 0.5820883512496948, "learning_rate": 0.001627037037037037, "loss": 1.5907, "step": 10070 }, { "epoch": 37.33, "grad_norm": 0.6052312850952148, "learning_rate": 0.0016266666666666667, "loss": 1.5893, "step": 10080 }, { "epoch": 37.37, "grad_norm": 0.6281132698059082, "learning_rate": 0.0016262962962962965, "loss": 1.5962, "step": 10090 }, { "epoch": 37.41, "grad_norm": 0.6100971102714539, "learning_rate": 0.001625925925925926, "loss": 1.6276, "step": 10100 }, { "epoch": 37.44, "grad_norm": 0.6440417170524597, "learning_rate": 0.0016255555555555557, "loss": 1.6179, "step": 10110 }, { "epoch": 37.48, "grad_norm": 0.5597901940345764, "learning_rate": 0.0016251851851851852, "loss": 1.6376, "step": 10120 }, { "epoch": 37.52, "grad_norm": 0.5792917609214783, "learning_rate": 0.0016248148148148148, "loss": 1.6437, "step": 10130 }, { "epoch": 37.56, "grad_norm": 0.5647159218788147, "learning_rate": 0.0016244444444444444, "loss": 1.6415, "step": 10140 }, { "epoch": 37.59, "grad_norm": 0.560454785823822, "learning_rate": 0.001624074074074074, "loss": 1.6694, "step": 10150 }, { "epoch": 37.63, "grad_norm": 0.5815364718437195, "learning_rate": 0.0016237037037037036, "loss": 1.6779, "step": 10160 }, { "epoch": 37.67, "grad_norm": 0.5997341871261597, "learning_rate": 0.0016233333333333334, "loss": 1.6526, "step": 10170 }, { "epoch": 37.7, "grad_norm": 0.6044934988021851, "learning_rate": 0.001622962962962963, "loss": 1.6668, "step": 10180 }, { "epoch": 37.74, "grad_norm": 0.6119056344032288, "learning_rate": 0.0016225925925925926, "loss": 1.7002, "step": 10190 }, { "epoch": 37.78, "grad_norm": 0.5937202572822571, "learning_rate": 0.0016222222222222222, "loss": 1.656, "step": 10200 }, { "epoch": 37.81, "grad_norm": 0.5740966200828552, "learning_rate": 0.001621851851851852, "loss": 1.7033, "step": 10210 }, { "epoch": 37.85, "grad_norm": 0.5541761517524719, "learning_rate": 0.0016214814814814816, "loss": 1.6999, "step": 10220 }, { "epoch": 37.89, "grad_norm": 0.5576919913291931, "learning_rate": 0.0016211111111111111, "loss": 1.6883, "step": 10230 }, { "epoch": 37.93, "grad_norm": 0.5857290625572205, "learning_rate": 0.001620740740740741, "loss": 1.6925, "step": 10240 }, { "epoch": 37.96, "grad_norm": 0.6111593842506409, "learning_rate": 0.0016203703703703705, "loss": 1.6872, "step": 10250 }, { "epoch": 38.0, "grad_norm": 1.0599267482757568, "learning_rate": 0.0016200000000000001, "loss": 1.683, "step": 10260 }, { "epoch": 38.04, "grad_norm": 0.5836325883865356, "learning_rate": 0.0016196296296296295, "loss": 1.4644, "step": 10270 }, { "epoch": 38.07, "grad_norm": 0.5915476679801941, "learning_rate": 0.0016192592592592593, "loss": 1.4787, "step": 10280 }, { "epoch": 38.11, "grad_norm": 0.63363116979599, "learning_rate": 0.001618888888888889, "loss": 1.456, "step": 10290 }, { "epoch": 38.15, "grad_norm": 0.6005881428718567, "learning_rate": 0.0016185185185185185, "loss": 1.4905, "step": 10300 }, { "epoch": 38.19, "grad_norm": 0.6480348110198975, "learning_rate": 0.001618148148148148, "loss": 1.5057, "step": 10310 }, { "epoch": 38.22, "grad_norm": 0.5606887936592102, "learning_rate": 0.0016177777777777779, "loss": 1.5551, "step": 10320 }, { "epoch": 38.26, "grad_norm": 0.5907540917396545, "learning_rate": 0.0016174074074074075, "loss": 1.5862, "step": 10330 }, { "epoch": 38.3, "grad_norm": 0.5859313011169434, "learning_rate": 0.001617037037037037, "loss": 1.5688, "step": 10340 }, { "epoch": 38.33, "grad_norm": 0.7137908935546875, "learning_rate": 0.0016166666666666666, "loss": 1.5659, "step": 10350 }, { "epoch": 38.37, "grad_norm": 0.6301664710044861, "learning_rate": 0.0016162962962962964, "loss": 1.609, "step": 10360 }, { "epoch": 38.41, "grad_norm": 0.64017254114151, "learning_rate": 0.001615925925925926, "loss": 1.583, "step": 10370 }, { "epoch": 38.44, "grad_norm": 0.5742957592010498, "learning_rate": 0.0016155555555555556, "loss": 1.6008, "step": 10380 }, { "epoch": 38.48, "grad_norm": 0.5781733393669128, "learning_rate": 0.0016151851851851852, "loss": 1.6219, "step": 10390 }, { "epoch": 38.52, "grad_norm": 0.6394752264022827, "learning_rate": 0.001614814814814815, "loss": 1.6126, "step": 10400 }, { "epoch": 38.56, "grad_norm": 0.5630576610565186, "learning_rate": 0.0016144444444444444, "loss": 1.6292, "step": 10410 }, { "epoch": 38.59, "grad_norm": 0.6065866947174072, "learning_rate": 0.001614074074074074, "loss": 1.6424, "step": 10420 }, { "epoch": 38.63, "grad_norm": 0.5589773058891296, "learning_rate": 0.0016137037037037036, "loss": 1.6514, "step": 10430 }, { "epoch": 38.67, "grad_norm": 0.5719542503356934, "learning_rate": 0.0016133333333333334, "loss": 1.6464, "step": 10440 }, { "epoch": 38.7, "grad_norm": 0.595180869102478, "learning_rate": 0.001612962962962963, "loss": 1.6467, "step": 10450 }, { "epoch": 38.74, "grad_norm": 0.575110137462616, "learning_rate": 0.0016125925925925925, "loss": 1.6728, "step": 10460 }, { "epoch": 38.78, "grad_norm": 0.6241551041603088, "learning_rate": 0.0016122222222222224, "loss": 1.6654, "step": 10470 }, { "epoch": 38.81, "grad_norm": 0.5750718116760254, "learning_rate": 0.001611851851851852, "loss": 1.6792, "step": 10480 }, { "epoch": 38.85, "grad_norm": 0.6219273805618286, "learning_rate": 0.0016114814814814815, "loss": 1.6844, "step": 10490 }, { "epoch": 38.89, "grad_norm": 0.5741206407546997, "learning_rate": 0.0016111111111111111, "loss": 1.696, "step": 10500 }, { "epoch": 38.93, "grad_norm": 0.5772682428359985, "learning_rate": 0.001610740740740741, "loss": 1.7026, "step": 10510 }, { "epoch": 38.96, "grad_norm": 0.598931610584259, "learning_rate": 0.0016103703703703705, "loss": 1.6795, "step": 10520 }, { "epoch": 39.0, "grad_norm": 0.944146454334259, "learning_rate": 0.00161, "loss": 1.6987, "step": 10530 }, { "epoch": 39.04, "grad_norm": 0.578942596912384, "learning_rate": 0.0016096296296296295, "loss": 1.4329, "step": 10540 }, { "epoch": 39.07, "grad_norm": 0.6079661250114441, "learning_rate": 0.0016092592592592593, "loss": 1.4834, "step": 10550 }, { "epoch": 39.11, "grad_norm": 0.6036855578422546, "learning_rate": 0.0016088888888888889, "loss": 1.5066, "step": 10560 }, { "epoch": 39.15, "grad_norm": 0.620212972164154, "learning_rate": 0.0016085185185185185, "loss": 1.4874, "step": 10570 }, { "epoch": 39.19, "grad_norm": 0.5837911367416382, "learning_rate": 0.001608148148148148, "loss": 1.5096, "step": 10580 }, { "epoch": 39.22, "grad_norm": 0.581173837184906, "learning_rate": 0.0016077777777777779, "loss": 1.505, "step": 10590 }, { "epoch": 39.26, "grad_norm": 0.6712148189544678, "learning_rate": 0.0016074074074074074, "loss": 1.511, "step": 10600 }, { "epoch": 39.3, "grad_norm": 0.5917540192604065, "learning_rate": 0.001607037037037037, "loss": 1.5772, "step": 10610 }, { "epoch": 39.33, "grad_norm": 0.6199811100959778, "learning_rate": 0.0016066666666666666, "loss": 1.5893, "step": 10620 }, { "epoch": 39.37, "grad_norm": 0.5958033800125122, "learning_rate": 0.0016062962962962964, "loss": 1.5871, "step": 10630 }, { "epoch": 39.41, "grad_norm": 0.578544020652771, "learning_rate": 0.001605925925925926, "loss": 1.592, "step": 10640 }, { "epoch": 39.44, "grad_norm": 0.6394162178039551, "learning_rate": 0.0016055555555555556, "loss": 1.6051, "step": 10650 }, { "epoch": 39.48, "grad_norm": 0.6192345023155212, "learning_rate": 0.0016051851851851854, "loss": 1.6229, "step": 10660 }, { "epoch": 39.52, "grad_norm": 0.5998772382736206, "learning_rate": 0.001604814814814815, "loss": 1.6152, "step": 10670 }, { "epoch": 39.56, "grad_norm": 0.5876280665397644, "learning_rate": 0.0016044444444444444, "loss": 1.6249, "step": 10680 }, { "epoch": 39.59, "grad_norm": 0.5959920287132263, "learning_rate": 0.001604074074074074, "loss": 1.6276, "step": 10690 }, { "epoch": 39.63, "grad_norm": 0.6072013974189758, "learning_rate": 0.0016037037037037038, "loss": 1.6075, "step": 10700 }, { "epoch": 39.67, "grad_norm": 0.6390188932418823, "learning_rate": 0.0016033333333333333, "loss": 1.6377, "step": 10710 }, { "epoch": 39.7, "grad_norm": 0.6121541857719421, "learning_rate": 0.001602962962962963, "loss": 1.665, "step": 10720 }, { "epoch": 39.74, "grad_norm": 0.5736008882522583, "learning_rate": 0.0016025925925925925, "loss": 1.6434, "step": 10730 }, { "epoch": 39.78, "grad_norm": 0.5839884877204895, "learning_rate": 0.0016022222222222223, "loss": 1.6786, "step": 10740 }, { "epoch": 39.81, "grad_norm": 0.6004469394683838, "learning_rate": 0.001601851851851852, "loss": 1.6428, "step": 10750 }, { "epoch": 39.85, "grad_norm": 0.5822156667709351, "learning_rate": 0.0016014814814814815, "loss": 1.6589, "step": 10760 }, { "epoch": 39.89, "grad_norm": 0.6007235646247864, "learning_rate": 0.001601111111111111, "loss": 1.682, "step": 10770 }, { "epoch": 39.93, "grad_norm": 0.663915753364563, "learning_rate": 0.001600740740740741, "loss": 1.6864, "step": 10780 }, { "epoch": 39.96, "grad_norm": 0.5694162249565125, "learning_rate": 0.0016003703703703705, "loss": 1.6844, "step": 10790 }, { "epoch": 40.0, "grad_norm": 0.9879453182220459, "learning_rate": 0.0016, "loss": 1.7021, "step": 10800 }, { "epoch": 40.04, "grad_norm": 0.6228801012039185, "learning_rate": 0.0015996296296296297, "loss": 1.4537, "step": 10810 }, { "epoch": 40.07, "grad_norm": 0.6306723356246948, "learning_rate": 0.0015992592592592593, "loss": 1.4237, "step": 10820 }, { "epoch": 40.11, "grad_norm": 0.5553726553916931, "learning_rate": 0.0015988888888888888, "loss": 1.4835, "step": 10830 }, { "epoch": 40.15, "grad_norm": 0.57485032081604, "learning_rate": 0.0015985185185185184, "loss": 1.4796, "step": 10840 }, { "epoch": 40.19, "grad_norm": 0.6278731226921082, "learning_rate": 0.0015981481481481482, "loss": 1.4915, "step": 10850 }, { "epoch": 40.22, "grad_norm": 0.6191769242286682, "learning_rate": 0.0015977777777777778, "loss": 1.509, "step": 10860 }, { "epoch": 40.26, "grad_norm": 0.593605637550354, "learning_rate": 0.0015974074074074074, "loss": 1.5293, "step": 10870 }, { "epoch": 40.3, "grad_norm": 0.6278729438781738, "learning_rate": 0.001597037037037037, "loss": 1.5459, "step": 10880 }, { "epoch": 40.33, "grad_norm": 0.6142277121543884, "learning_rate": 0.0015966666666666668, "loss": 1.5509, "step": 10890 }, { "epoch": 40.37, "grad_norm": 0.6126216650009155, "learning_rate": 0.0015962962962962964, "loss": 1.5465, "step": 10900 }, { "epoch": 40.41, "grad_norm": 0.6625998616218567, "learning_rate": 0.001595925925925926, "loss": 1.5905, "step": 10910 }, { "epoch": 40.44, "grad_norm": 0.6081230044364929, "learning_rate": 0.0015955555555555556, "loss": 1.5928, "step": 10920 }, { "epoch": 40.48, "grad_norm": 0.5991567969322205, "learning_rate": 0.0015951851851851854, "loss": 1.5605, "step": 10930 }, { "epoch": 40.52, "grad_norm": 0.6089213490486145, "learning_rate": 0.001594814814814815, "loss": 1.5856, "step": 10940 }, { "epoch": 40.56, "grad_norm": 0.6468093395233154, "learning_rate": 0.0015944444444444446, "loss": 1.6097, "step": 10950 }, { "epoch": 40.59, "grad_norm": 0.6019711494445801, "learning_rate": 0.001594074074074074, "loss": 1.613, "step": 10960 }, { "epoch": 40.63, "grad_norm": 0.6264703869819641, "learning_rate": 0.0015937037037037037, "loss": 1.6485, "step": 10970 }, { "epoch": 40.67, "grad_norm": 0.5833860635757446, "learning_rate": 0.0015933333333333333, "loss": 1.6455, "step": 10980 }, { "epoch": 40.7, "grad_norm": 0.6113573908805847, "learning_rate": 0.001592962962962963, "loss": 1.6659, "step": 10990 }, { "epoch": 40.74, "grad_norm": 0.5929760336875916, "learning_rate": 0.0015925925925925925, "loss": 1.6641, "step": 11000 }, { "epoch": 40.78, "grad_norm": 0.5363314151763916, "learning_rate": 0.0015922222222222223, "loss": 1.6415, "step": 11010 }, { "epoch": 40.81, "grad_norm": 0.6000963449478149, "learning_rate": 0.0015918518518518519, "loss": 1.648, "step": 11020 }, { "epoch": 40.85, "grad_norm": 0.6014578342437744, "learning_rate": 0.0015914814814814815, "loss": 1.692, "step": 11030 }, { "epoch": 40.89, "grad_norm": 0.5798655152320862, "learning_rate": 0.001591111111111111, "loss": 1.6799, "step": 11040 }, { "epoch": 40.93, "grad_norm": 0.5583765506744385, "learning_rate": 0.0015907407407407409, "loss": 1.6767, "step": 11050 }, { "epoch": 40.96, "grad_norm": 0.5604484677314758, "learning_rate": 0.0015903703703703705, "loss": 1.7067, "step": 11060 }, { "epoch": 41.0, "grad_norm": 1.2024364471435547, "learning_rate": 0.00159, "loss": 1.6975, "step": 11070 }, { "epoch": 41.04, "grad_norm": 0.6150010824203491, "learning_rate": 0.0015896296296296299, "loss": 1.4372, "step": 11080 }, { "epoch": 41.07, "grad_norm": 0.605194091796875, "learning_rate": 0.0015892592592592592, "loss": 1.4411, "step": 11090 }, { "epoch": 41.11, "grad_norm": 0.621233344078064, "learning_rate": 0.0015888888888888888, "loss": 1.4452, "step": 11100 }, { "epoch": 41.15, "grad_norm": 0.5805311799049377, "learning_rate": 0.0015885185185185184, "loss": 1.478, "step": 11110 }, { "epoch": 41.19, "grad_norm": 0.5650319457054138, "learning_rate": 0.0015881481481481482, "loss": 1.5149, "step": 11120 }, { "epoch": 41.22, "grad_norm": 0.6075964570045471, "learning_rate": 0.0015877777777777778, "loss": 1.5134, "step": 11130 }, { "epoch": 41.26, "grad_norm": 0.6080381870269775, "learning_rate": 0.0015874074074074074, "loss": 1.5133, "step": 11140 }, { "epoch": 41.3, "grad_norm": 0.5945024490356445, "learning_rate": 0.001587037037037037, "loss": 1.5214, "step": 11150 }, { "epoch": 41.33, "grad_norm": 0.5842931270599365, "learning_rate": 0.0015866666666666668, "loss": 1.5149, "step": 11160 }, { "epoch": 41.37, "grad_norm": 0.6802297830581665, "learning_rate": 0.0015862962962962964, "loss": 1.5337, "step": 11170 }, { "epoch": 41.41, "grad_norm": 0.6695743799209595, "learning_rate": 0.001585925925925926, "loss": 1.5723, "step": 11180 }, { "epoch": 41.44, "grad_norm": 0.6051291227340698, "learning_rate": 0.0015855555555555555, "loss": 1.5818, "step": 11190 }, { "epoch": 41.48, "grad_norm": 0.623485803604126, "learning_rate": 0.0015851851851851854, "loss": 1.5934, "step": 11200 }, { "epoch": 41.52, "grad_norm": 0.5692495703697205, "learning_rate": 0.001584814814814815, "loss": 1.6444, "step": 11210 }, { "epoch": 41.56, "grad_norm": 0.5983277559280396, "learning_rate": 0.0015844444444444445, "loss": 1.6346, "step": 11220 }, { "epoch": 41.59, "grad_norm": 0.6303779482841492, "learning_rate": 0.001584074074074074, "loss": 1.6532, "step": 11230 }, { "epoch": 41.63, "grad_norm": 0.6377682685852051, "learning_rate": 0.0015837037037037037, "loss": 1.639, "step": 11240 }, { "epoch": 41.67, "grad_norm": 0.5868995189666748, "learning_rate": 0.0015833333333333333, "loss": 1.6179, "step": 11250 }, { "epoch": 41.7, "grad_norm": 0.645334005355835, "learning_rate": 0.0015829629629629629, "loss": 1.6623, "step": 11260 }, { "epoch": 41.74, "grad_norm": 0.6188551783561707, "learning_rate": 0.0015825925925925927, "loss": 1.6446, "step": 11270 }, { "epoch": 41.78, "grad_norm": 0.6530919075012207, "learning_rate": 0.0015822222222222223, "loss": 1.6387, "step": 11280 }, { "epoch": 41.81, "grad_norm": 0.6173209547996521, "learning_rate": 0.0015818518518518519, "loss": 1.6248, "step": 11290 }, { "epoch": 41.85, "grad_norm": 0.6064965128898621, "learning_rate": 0.0015814814814814815, "loss": 1.6507, "step": 11300 }, { "epoch": 41.89, "grad_norm": 0.5441533923149109, "learning_rate": 0.0015811111111111113, "loss": 1.6184, "step": 11310 }, { "epoch": 41.93, "grad_norm": 0.6284180283546448, "learning_rate": 0.0015807407407407408, "loss": 1.6524, "step": 11320 }, { "epoch": 41.96, "grad_norm": 0.6238906979560852, "learning_rate": 0.0015803703703703704, "loss": 1.6529, "step": 11330 }, { "epoch": 42.0, "grad_norm": 1.1080102920532227, "learning_rate": 0.00158, "loss": 1.6716, "step": 11340 }, { "epoch": 42.04, "grad_norm": 0.6129464507102966, "learning_rate": 0.0015796296296296298, "loss": 1.4015, "step": 11350 }, { "epoch": 42.07, "grad_norm": 0.6170362234115601, "learning_rate": 0.0015792592592592594, "loss": 1.4541, "step": 11360 }, { "epoch": 42.11, "grad_norm": 0.6223025321960449, "learning_rate": 0.0015788888888888888, "loss": 1.442, "step": 11370 }, { "epoch": 42.15, "grad_norm": 0.6103693246841431, "learning_rate": 0.0015785185185185184, "loss": 1.4537, "step": 11380 }, { "epoch": 42.19, "grad_norm": 0.6398111581802368, "learning_rate": 0.0015781481481481482, "loss": 1.5009, "step": 11390 }, { "epoch": 42.22, "grad_norm": 0.6270093321800232, "learning_rate": 0.0015777777777777778, "loss": 1.4739, "step": 11400 }, { "epoch": 42.26, "grad_norm": 0.6488049626350403, "learning_rate": 0.0015774074074074074, "loss": 1.5274, "step": 11410 }, { "epoch": 42.3, "grad_norm": 0.6524102687835693, "learning_rate": 0.001577037037037037, "loss": 1.5243, "step": 11420 }, { "epoch": 42.33, "grad_norm": 0.655152440071106, "learning_rate": 0.0015766666666666668, "loss": 1.5228, "step": 11430 }, { "epoch": 42.37, "grad_norm": 0.675579309463501, "learning_rate": 0.0015762962962962963, "loss": 1.5568, "step": 11440 }, { "epoch": 42.41, "grad_norm": 0.6234245300292969, "learning_rate": 0.001575925925925926, "loss": 1.6002, "step": 11450 }, { "epoch": 42.44, "grad_norm": 0.5919743776321411, "learning_rate": 0.0015755555555555557, "loss": 1.5659, "step": 11460 }, { "epoch": 42.48, "grad_norm": 0.6351736187934875, "learning_rate": 0.0015751851851851853, "loss": 1.5863, "step": 11470 }, { "epoch": 42.52, "grad_norm": 0.6661149263381958, "learning_rate": 0.001574814814814815, "loss": 1.584, "step": 11480 }, { "epoch": 42.56, "grad_norm": 0.6044712662696838, "learning_rate": 0.0015744444444444445, "loss": 1.611, "step": 11490 }, { "epoch": 42.59, "grad_norm": 0.6325424313545227, "learning_rate": 0.0015740740740740743, "loss": 1.597, "step": 11500 }, { "epoch": 42.63, "grad_norm": 0.5968678593635559, "learning_rate": 0.0015737037037037037, "loss": 1.6211, "step": 11510 }, { "epoch": 42.67, "grad_norm": 0.60098797082901, "learning_rate": 0.0015733333333333333, "loss": 1.6103, "step": 11520 }, { "epoch": 42.7, "grad_norm": 0.6495356559753418, "learning_rate": 0.0015729629629629629, "loss": 1.6439, "step": 11530 }, { "epoch": 42.74, "grad_norm": 0.5992751717567444, "learning_rate": 0.0015725925925925927, "loss": 1.6211, "step": 11540 }, { "epoch": 42.78, "grad_norm": 0.5891815423965454, "learning_rate": 0.0015722222222222223, "loss": 1.6477, "step": 11550 }, { "epoch": 42.81, "grad_norm": 0.6189820766448975, "learning_rate": 0.0015718518518518518, "loss": 1.6335, "step": 11560 }, { "epoch": 42.85, "grad_norm": 0.6549631357192993, "learning_rate": 0.0015714814814814814, "loss": 1.6318, "step": 11570 }, { "epoch": 42.89, "grad_norm": 0.6061648726463318, "learning_rate": 0.0015711111111111112, "loss": 1.6649, "step": 11580 }, { "epoch": 42.93, "grad_norm": 0.6047539710998535, "learning_rate": 0.0015707407407407408, "loss": 1.6516, "step": 11590 }, { "epoch": 42.96, "grad_norm": 0.5719320178031921, "learning_rate": 0.0015703703703703704, "loss": 1.6608, "step": 11600 }, { "epoch": 43.0, "grad_norm": 1.1914095878601074, "learning_rate": 0.00157, "loss": 1.649, "step": 11610 }, { "epoch": 43.04, "grad_norm": 0.6697654724121094, "learning_rate": 0.0015696296296296298, "loss": 1.4124, "step": 11620 }, { "epoch": 43.07, "grad_norm": 0.5861226320266724, "learning_rate": 0.0015692592592592594, "loss": 1.4241, "step": 11630 }, { "epoch": 43.11, "grad_norm": 0.6160594820976257, "learning_rate": 0.001568888888888889, "loss": 1.4394, "step": 11640 }, { "epoch": 43.15, "grad_norm": 0.6051038503646851, "learning_rate": 0.0015685185185185184, "loss": 1.469, "step": 11650 }, { "epoch": 43.19, "grad_norm": 0.6148421764373779, "learning_rate": 0.0015681481481481482, "loss": 1.4866, "step": 11660 }, { "epoch": 43.22, "grad_norm": 0.6597320437431335, "learning_rate": 0.0015677777777777777, "loss": 1.5145, "step": 11670 }, { "epoch": 43.26, "grad_norm": 0.6334295868873596, "learning_rate": 0.0015674074074074073, "loss": 1.4995, "step": 11680 }, { "epoch": 43.3, "grad_norm": 0.6128714680671692, "learning_rate": 0.0015670370370370371, "loss": 1.4923, "step": 11690 }, { "epoch": 43.33, "grad_norm": 0.6361562609672546, "learning_rate": 0.0015666666666666667, "loss": 1.5172, "step": 11700 }, { "epoch": 43.37, "grad_norm": 0.615699052810669, "learning_rate": 0.0015662962962962963, "loss": 1.5052, "step": 11710 }, { "epoch": 43.41, "grad_norm": 0.6639183163642883, "learning_rate": 0.001565925925925926, "loss": 1.5669, "step": 11720 }, { "epoch": 43.44, "grad_norm": 0.6383649706840515, "learning_rate": 0.0015655555555555557, "loss": 1.5403, "step": 11730 }, { "epoch": 43.48, "grad_norm": 0.6247989535331726, "learning_rate": 0.0015651851851851853, "loss": 1.5791, "step": 11740 }, { "epoch": 43.52, "grad_norm": 0.6123582720756531, "learning_rate": 0.0015648148148148149, "loss": 1.5503, "step": 11750 }, { "epoch": 43.56, "grad_norm": 0.6618642807006836, "learning_rate": 0.0015644444444444445, "loss": 1.5792, "step": 11760 }, { "epoch": 43.59, "grad_norm": 0.6254457831382751, "learning_rate": 0.0015640740740740743, "loss": 1.6016, "step": 11770 }, { "epoch": 43.63, "grad_norm": 0.6345099210739136, "learning_rate": 0.0015637037037037037, "loss": 1.5877, "step": 11780 }, { "epoch": 43.67, "grad_norm": 0.6265460848808289, "learning_rate": 0.0015633333333333332, "loss": 1.6274, "step": 11790 }, { "epoch": 43.7, "grad_norm": 0.6505123376846313, "learning_rate": 0.0015629629629629628, "loss": 1.6227, "step": 11800 }, { "epoch": 43.74, "grad_norm": 0.6225472688674927, "learning_rate": 0.0015625925925925926, "loss": 1.6175, "step": 11810 }, { "epoch": 43.78, "grad_norm": 0.6467319130897522, "learning_rate": 0.0015622222222222222, "loss": 1.6325, "step": 11820 }, { "epoch": 43.81, "grad_norm": 0.6350839138031006, "learning_rate": 0.0015618518518518518, "loss": 1.6572, "step": 11830 }, { "epoch": 43.85, "grad_norm": 0.614852249622345, "learning_rate": 0.0015614814814814814, "loss": 1.6469, "step": 11840 }, { "epoch": 43.89, "grad_norm": 0.6368483901023865, "learning_rate": 0.0015611111111111112, "loss": 1.6346, "step": 11850 }, { "epoch": 43.93, "grad_norm": 0.6596633195877075, "learning_rate": 0.0015607407407407408, "loss": 1.6669, "step": 11860 }, { "epoch": 43.96, "grad_norm": 0.5940620303153992, "learning_rate": 0.0015603703703703704, "loss": 1.6529, "step": 11870 }, { "epoch": 44.0, "grad_norm": 1.0889242887496948, "learning_rate": 0.0015600000000000002, "loss": 1.6417, "step": 11880 }, { "epoch": 44.04, "grad_norm": 0.6191521883010864, "learning_rate": 0.0015596296296296298, "loss": 1.4271, "step": 11890 }, { "epoch": 44.07, "grad_norm": 0.6261031031608582, "learning_rate": 0.0015592592592592594, "loss": 1.4191, "step": 11900 }, { "epoch": 44.11, "grad_norm": 0.6673709154129028, "learning_rate": 0.001558888888888889, "loss": 1.4281, "step": 11910 }, { "epoch": 44.15, "grad_norm": 0.621164083480835, "learning_rate": 0.0015585185185185185, "loss": 1.4318, "step": 11920 }, { "epoch": 44.19, "grad_norm": 0.6410382390022278, "learning_rate": 0.0015581481481481481, "loss": 1.498, "step": 11930 }, { "epoch": 44.22, "grad_norm": 0.6146425604820251, "learning_rate": 0.0015577777777777777, "loss": 1.5023, "step": 11940 }, { "epoch": 44.26, "grad_norm": 0.6328328251838684, "learning_rate": 0.0015574074074074073, "loss": 1.5026, "step": 11950 }, { "epoch": 44.3, "grad_norm": 0.677683413028717, "learning_rate": 0.0015570370370370371, "loss": 1.5434, "step": 11960 }, { "epoch": 44.33, "grad_norm": 0.6298605799674988, "learning_rate": 0.0015566666666666667, "loss": 1.5206, "step": 11970 }, { "epoch": 44.37, "grad_norm": 0.6399940848350525, "learning_rate": 0.0015562962962962963, "loss": 1.5417, "step": 11980 }, { "epoch": 44.41, "grad_norm": 0.6250726580619812, "learning_rate": 0.0015559259259259259, "loss": 1.5478, "step": 11990 }, { "epoch": 44.44, "grad_norm": 0.6349297165870667, "learning_rate": 0.0015555555555555557, "loss": 1.5409, "step": 12000 }, { "epoch": 44.48, "grad_norm": 0.686031699180603, "learning_rate": 0.0015551851851851853, "loss": 1.5595, "step": 12010 }, { "epoch": 44.52, "grad_norm": 0.5983814001083374, "learning_rate": 0.0015548148148148149, "loss": 1.5518, "step": 12020 }, { "epoch": 44.56, "grad_norm": 0.6011641621589661, "learning_rate": 0.0015544444444444444, "loss": 1.5647, "step": 12030 }, { "epoch": 44.59, "grad_norm": 0.6552057266235352, "learning_rate": 0.0015540740740740743, "loss": 1.5726, "step": 12040 }, { "epoch": 44.63, "grad_norm": 0.6505006551742554, "learning_rate": 0.0015537037037037038, "loss": 1.5974, "step": 12050 }, { "epoch": 44.67, "grad_norm": 0.6745777726173401, "learning_rate": 0.0015533333333333332, "loss": 1.6191, "step": 12060 }, { "epoch": 44.7, "grad_norm": 0.6342930197715759, "learning_rate": 0.0015529629629629628, "loss": 1.5988, "step": 12070 }, { "epoch": 44.74, "grad_norm": 0.6098774075508118, "learning_rate": 0.0015525925925925926, "loss": 1.611, "step": 12080 }, { "epoch": 44.78, "grad_norm": 0.7015253901481628, "learning_rate": 0.0015522222222222222, "loss": 1.6142, "step": 12090 }, { "epoch": 44.81, "grad_norm": 0.5675494074821472, "learning_rate": 0.0015518518518518518, "loss": 1.6014, "step": 12100 }, { "epoch": 44.85, "grad_norm": 0.6159512996673584, "learning_rate": 0.0015514814814814816, "loss": 1.6346, "step": 12110 }, { "epoch": 44.89, "grad_norm": 0.6535830497741699, "learning_rate": 0.0015511111111111112, "loss": 1.6159, "step": 12120 }, { "epoch": 44.93, "grad_norm": 0.6116699576377869, "learning_rate": 0.0015507407407407408, "loss": 1.6558, "step": 12130 }, { "epoch": 44.96, "grad_norm": 0.6020302176475525, "learning_rate": 0.0015503703703703704, "loss": 1.6704, "step": 12140 }, { "epoch": 45.0, "grad_norm": 1.1520200967788696, "learning_rate": 0.0015500000000000002, "loss": 1.6353, "step": 12150 }, { "epoch": 45.04, "grad_norm": 0.6187354326248169, "learning_rate": 0.0015496296296296298, "loss": 1.4284, "step": 12160 }, { "epoch": 45.07, "grad_norm": 0.6019529104232788, "learning_rate": 0.0015492592592592593, "loss": 1.3639, "step": 12170 }, { "epoch": 45.11, "grad_norm": 0.6266639232635498, "learning_rate": 0.001548888888888889, "loss": 1.4464, "step": 12180 }, { "epoch": 45.15, "grad_norm": 0.6570635437965393, "learning_rate": 0.0015485185185185187, "loss": 1.44, "step": 12190 }, { "epoch": 45.19, "grad_norm": 0.6148927211761475, "learning_rate": 0.001548148148148148, "loss": 1.4382, "step": 12200 }, { "epoch": 45.22, "grad_norm": 0.6563842296600342, "learning_rate": 0.0015477777777777777, "loss": 1.4584, "step": 12210 }, { "epoch": 45.26, "grad_norm": 0.6292750239372253, "learning_rate": 0.0015474074074074073, "loss": 1.4918, "step": 12220 }, { "epoch": 45.3, "grad_norm": 0.6535168886184692, "learning_rate": 0.001547037037037037, "loss": 1.4881, "step": 12230 }, { "epoch": 45.33, "grad_norm": 0.6435755491256714, "learning_rate": 0.0015466666666666667, "loss": 1.5152, "step": 12240 }, { "epoch": 45.37, "grad_norm": 0.6139950156211853, "learning_rate": 0.0015462962962962963, "loss": 1.5355, "step": 12250 }, { "epoch": 45.41, "grad_norm": 0.661777913570404, "learning_rate": 0.0015459259259259259, "loss": 1.5319, "step": 12260 }, { "epoch": 45.44, "grad_norm": 0.7189764976501465, "learning_rate": 0.0015455555555555557, "loss": 1.5786, "step": 12270 }, { "epoch": 45.48, "grad_norm": 0.604444146156311, "learning_rate": 0.0015451851851851852, "loss": 1.5659, "step": 12280 }, { "epoch": 45.52, "grad_norm": 0.6002549529075623, "learning_rate": 0.0015448148148148148, "loss": 1.5513, "step": 12290 }, { "epoch": 45.56, "grad_norm": 0.6639789342880249, "learning_rate": 0.0015444444444444446, "loss": 1.5771, "step": 12300 }, { "epoch": 45.59, "grad_norm": 0.6240954399108887, "learning_rate": 0.0015440740740740742, "loss": 1.5577, "step": 12310 }, { "epoch": 45.63, "grad_norm": 0.6983676552772522, "learning_rate": 0.0015437037037037038, "loss": 1.5728, "step": 12320 }, { "epoch": 45.67, "grad_norm": 0.6203968524932861, "learning_rate": 0.0015433333333333332, "loss": 1.5723, "step": 12330 }, { "epoch": 45.7, "grad_norm": 0.6263325810432434, "learning_rate": 0.001542962962962963, "loss": 1.5891, "step": 12340 }, { "epoch": 45.74, "grad_norm": 0.6136410236358643, "learning_rate": 0.0015425925925925926, "loss": 1.5958, "step": 12350 }, { "epoch": 45.78, "grad_norm": 0.6326088309288025, "learning_rate": 0.0015422222222222222, "loss": 1.6039, "step": 12360 }, { "epoch": 45.81, "grad_norm": 0.649049699306488, "learning_rate": 0.0015418518518518518, "loss": 1.6412, "step": 12370 }, { "epoch": 45.85, "grad_norm": 0.6206584572792053, "learning_rate": 0.0015414814814814816, "loss": 1.6086, "step": 12380 }, { "epoch": 45.89, "grad_norm": 0.6422502994537354, "learning_rate": 0.0015411111111111112, "loss": 1.6182, "step": 12390 }, { "epoch": 45.93, "grad_norm": 0.612797737121582, "learning_rate": 0.0015407407407407407, "loss": 1.6429, "step": 12400 }, { "epoch": 45.96, "grad_norm": 0.652485191822052, "learning_rate": 0.0015403703703703703, "loss": 1.6542, "step": 12410 }, { "epoch": 46.0, "grad_norm": 1.1145904064178467, "learning_rate": 0.0015400000000000001, "loss": 1.6458, "step": 12420 }, { "epoch": 46.04, "grad_norm": 0.6264052391052246, "learning_rate": 0.0015396296296296297, "loss": 1.4118, "step": 12430 }, { "epoch": 46.07, "grad_norm": 0.6265543103218079, "learning_rate": 0.0015392592592592593, "loss": 1.4072, "step": 12440 }, { "epoch": 46.11, "grad_norm": 0.6246814727783203, "learning_rate": 0.001538888888888889, "loss": 1.4276, "step": 12450 }, { "epoch": 46.15, "grad_norm": 0.6075099110603333, "learning_rate": 0.0015385185185185187, "loss": 1.4119, "step": 12460 }, { "epoch": 46.19, "grad_norm": 0.6614938974380493, "learning_rate": 0.001538148148148148, "loss": 1.4499, "step": 12470 }, { "epoch": 46.22, "grad_norm": 0.6482969522476196, "learning_rate": 0.0015377777777777777, "loss": 1.4508, "step": 12480 }, { "epoch": 46.26, "grad_norm": 0.61202073097229, "learning_rate": 0.0015374074074074075, "loss": 1.4833, "step": 12490 }, { "epoch": 46.3, "grad_norm": 0.6306284070014954, "learning_rate": 0.001537037037037037, "loss": 1.5188, "step": 12500 }, { "epoch": 46.33, "grad_norm": 0.6624696850776672, "learning_rate": 0.0015366666666666666, "loss": 1.4991, "step": 12510 }, { "epoch": 46.37, "grad_norm": 0.6851840019226074, "learning_rate": 0.0015362962962962962, "loss": 1.5289, "step": 12520 }, { "epoch": 46.41, "grad_norm": 0.6263905763626099, "learning_rate": 0.001535925925925926, "loss": 1.5316, "step": 12530 }, { "epoch": 46.44, "grad_norm": 0.588344395160675, "learning_rate": 0.0015355555555555556, "loss": 1.5447, "step": 12540 }, { "epoch": 46.48, "grad_norm": 0.6301348805427551, "learning_rate": 0.0015351851851851852, "loss": 1.56, "step": 12550 }, { "epoch": 46.52, "grad_norm": 0.6092745661735535, "learning_rate": 0.0015348148148148148, "loss": 1.5466, "step": 12560 }, { "epoch": 46.56, "grad_norm": 0.6487380266189575, "learning_rate": 0.0015344444444444446, "loss": 1.5944, "step": 12570 }, { "epoch": 46.59, "grad_norm": 0.6388218998908997, "learning_rate": 0.0015340740740740742, "loss": 1.5849, "step": 12580 }, { "epoch": 46.63, "grad_norm": 0.6476507782936096, "learning_rate": 0.0015337037037037038, "loss": 1.541, "step": 12590 }, { "epoch": 46.67, "grad_norm": 0.622490406036377, "learning_rate": 0.0015333333333333334, "loss": 1.5799, "step": 12600 }, { "epoch": 46.7, "grad_norm": 0.6342136263847351, "learning_rate": 0.001532962962962963, "loss": 1.5772, "step": 12610 }, { "epoch": 46.74, "grad_norm": 0.6312634348869324, "learning_rate": 0.0015325925925925926, "loss": 1.5933, "step": 12620 }, { "epoch": 46.78, "grad_norm": 0.604749858379364, "learning_rate": 0.0015322222222222221, "loss": 1.6059, "step": 12630 }, { "epoch": 46.81, "grad_norm": 0.6882821917533875, "learning_rate": 0.0015318518518518517, "loss": 1.6056, "step": 12640 }, { "epoch": 46.85, "grad_norm": 0.656486451625824, "learning_rate": 0.0015314814814814815, "loss": 1.6105, "step": 12650 }, { "epoch": 46.89, "grad_norm": 0.6267317533493042, "learning_rate": 0.0015311111111111111, "loss": 1.6267, "step": 12660 }, { "epoch": 46.93, "grad_norm": 0.6135218739509583, "learning_rate": 0.0015307407407407407, "loss": 1.6251, "step": 12670 }, { "epoch": 46.96, "grad_norm": 0.6487293839454651, "learning_rate": 0.0015303703703703703, "loss": 1.6157, "step": 12680 }, { "epoch": 47.0, "grad_norm": 1.242255687713623, "learning_rate": 0.0015300000000000001, "loss": 1.6046, "step": 12690 }, { "epoch": 47.04, "grad_norm": 0.6860459446907043, "learning_rate": 0.0015296296296296297, "loss": 1.3779, "step": 12700 }, { "epoch": 47.07, "grad_norm": 0.6426923871040344, "learning_rate": 0.0015292592592592593, "loss": 1.4071, "step": 12710 }, { "epoch": 47.11, "grad_norm": 0.6774601936340332, "learning_rate": 0.001528888888888889, "loss": 1.4033, "step": 12720 }, { "epoch": 47.15, "grad_norm": 0.6799710392951965, "learning_rate": 0.0015285185185185187, "loss": 1.4038, "step": 12730 }, { "epoch": 47.19, "grad_norm": 0.6156853437423706, "learning_rate": 0.0015281481481481483, "loss": 1.4496, "step": 12740 }, { "epoch": 47.22, "grad_norm": 0.6396149396896362, "learning_rate": 0.0015277777777777776, "loss": 1.453, "step": 12750 }, { "epoch": 47.26, "grad_norm": 0.6574532389640808, "learning_rate": 0.0015274074074074074, "loss": 1.4712, "step": 12760 }, { "epoch": 47.3, "grad_norm": 0.6416125893592834, "learning_rate": 0.001527037037037037, "loss": 1.4736, "step": 12770 }, { "epoch": 47.33, "grad_norm": 0.6952414512634277, "learning_rate": 0.0015266666666666666, "loss": 1.4879, "step": 12780 }, { "epoch": 47.37, "grad_norm": 0.6132145524024963, "learning_rate": 0.0015262962962962962, "loss": 1.5121, "step": 12790 }, { "epoch": 47.41, "grad_norm": 0.6308304667472839, "learning_rate": 0.001525925925925926, "loss": 1.5141, "step": 12800 }, { "epoch": 47.44, "grad_norm": 0.6352577209472656, "learning_rate": 0.0015255555555555556, "loss": 1.5348, "step": 12810 }, { "epoch": 47.48, "grad_norm": 0.624758243560791, "learning_rate": 0.0015251851851851852, "loss": 1.516, "step": 12820 }, { "epoch": 47.52, "grad_norm": 0.630090594291687, "learning_rate": 0.0015248148148148148, "loss": 1.5644, "step": 12830 }, { "epoch": 47.56, "grad_norm": 0.6315339803695679, "learning_rate": 0.0015244444444444446, "loss": 1.575, "step": 12840 }, { "epoch": 47.59, "grad_norm": 0.6333447098731995, "learning_rate": 0.0015240740740740742, "loss": 1.5693, "step": 12850 }, { "epoch": 47.63, "grad_norm": 0.6272197365760803, "learning_rate": 0.0015237037037037038, "loss": 1.5711, "step": 12860 }, { "epoch": 47.67, "grad_norm": 0.6363614797592163, "learning_rate": 0.0015233333333333334, "loss": 1.5668, "step": 12870 }, { "epoch": 47.7, "grad_norm": 0.6725128293037415, "learning_rate": 0.0015229629629629632, "loss": 1.5753, "step": 12880 }, { "epoch": 47.74, "grad_norm": 0.6107433438301086, "learning_rate": 0.0015225925925925925, "loss": 1.5887, "step": 12890 }, { "epoch": 47.78, "grad_norm": 0.6690382361412048, "learning_rate": 0.0015222222222222221, "loss": 1.6191, "step": 12900 }, { "epoch": 47.81, "grad_norm": 0.617404043674469, "learning_rate": 0.001521851851851852, "loss": 1.5842, "step": 12910 }, { "epoch": 47.85, "grad_norm": 0.6330074667930603, "learning_rate": 0.0015214814814814815, "loss": 1.6184, "step": 12920 }, { "epoch": 47.89, "grad_norm": 0.6891388297080994, "learning_rate": 0.001521111111111111, "loss": 1.6195, "step": 12930 }, { "epoch": 47.93, "grad_norm": 0.61796635389328, "learning_rate": 0.0015207407407407407, "loss": 1.6086, "step": 12940 }, { "epoch": 47.96, "grad_norm": 0.6967139840126038, "learning_rate": 0.0015203703703703705, "loss": 1.613, "step": 12950 }, { "epoch": 48.0, "grad_norm": 1.0406641960144043, "learning_rate": 0.00152, "loss": 1.6041, "step": 12960 }, { "epoch": 48.04, "grad_norm": 0.6598649024963379, "learning_rate": 0.0015196296296296297, "loss": 1.3771, "step": 12970 }, { "epoch": 48.07, "grad_norm": 0.6702351570129395, "learning_rate": 0.0015192592592592593, "loss": 1.3665, "step": 12980 }, { "epoch": 48.11, "grad_norm": 0.6731034517288208, "learning_rate": 0.001518888888888889, "loss": 1.3969, "step": 12990 }, { "epoch": 48.15, "grad_norm": 0.6232157349586487, "learning_rate": 0.0015185185185185187, "loss": 1.4008, "step": 13000 }, { "epoch": 48.19, "grad_norm": 0.6420234441757202, "learning_rate": 0.0015181481481481482, "loss": 1.4486, "step": 13010 }, { "epoch": 48.22, "grad_norm": 0.6267896890640259, "learning_rate": 0.0015177777777777776, "loss": 1.4501, "step": 13020 }, { "epoch": 48.26, "grad_norm": 0.6549137234687805, "learning_rate": 0.0015174074074074074, "loss": 1.4482, "step": 13030 }, { "epoch": 48.3, "grad_norm": 0.7054146528244019, "learning_rate": 0.001517037037037037, "loss": 1.4779, "step": 13040 }, { "epoch": 48.33, "grad_norm": 0.6636216044425964, "learning_rate": 0.0015166666666666666, "loss": 1.4828, "step": 13050 }, { "epoch": 48.37, "grad_norm": 0.6946861743927002, "learning_rate": 0.0015162962962962962, "loss": 1.4607, "step": 13060 }, { "epoch": 48.41, "grad_norm": 0.6364338994026184, "learning_rate": 0.001515925925925926, "loss": 1.5354, "step": 13070 }, { "epoch": 48.44, "grad_norm": 0.6550884246826172, "learning_rate": 0.0015155555555555556, "loss": 1.5186, "step": 13080 }, { "epoch": 48.48, "grad_norm": 0.671321451663971, "learning_rate": 0.0015151851851851852, "loss": 1.5092, "step": 13090 }, { "epoch": 48.52, "grad_norm": 0.6748003959655762, "learning_rate": 0.001514814814814815, "loss": 1.5204, "step": 13100 }, { "epoch": 48.56, "grad_norm": 0.7088521718978882, "learning_rate": 0.0015144444444444446, "loss": 1.5215, "step": 13110 }, { "epoch": 48.59, "grad_norm": 0.7026909589767456, "learning_rate": 0.0015140740740740742, "loss": 1.5686, "step": 13120 }, { "epoch": 48.63, "grad_norm": 0.5976001620292664, "learning_rate": 0.0015137037037037037, "loss": 1.5663, "step": 13130 }, { "epoch": 48.67, "grad_norm": 0.651646077632904, "learning_rate": 0.0015133333333333335, "loss": 1.5718, "step": 13140 }, { "epoch": 48.7, "grad_norm": 0.664674699306488, "learning_rate": 0.0015129629629629631, "loss": 1.5884, "step": 13150 }, { "epoch": 48.74, "grad_norm": 0.6666598320007324, "learning_rate": 0.0015125925925925925, "loss": 1.5799, "step": 13160 }, { "epoch": 48.78, "grad_norm": 0.6095290780067444, "learning_rate": 0.001512222222222222, "loss": 1.5815, "step": 13170 }, { "epoch": 48.81, "grad_norm": 0.6643814444541931, "learning_rate": 0.001511851851851852, "loss": 1.6082, "step": 13180 }, { "epoch": 48.85, "grad_norm": 0.6361218094825745, "learning_rate": 0.0015114814814814815, "loss": 1.6041, "step": 13190 }, { "epoch": 48.89, "grad_norm": 0.6308338642120361, "learning_rate": 0.001511111111111111, "loss": 1.5788, "step": 13200 }, { "epoch": 48.93, "grad_norm": 0.6339203119277954, "learning_rate": 0.0015107407407407407, "loss": 1.6344, "step": 13210 }, { "epoch": 48.96, "grad_norm": 0.6878967881202698, "learning_rate": 0.0015103703703703705, "loss": 1.6565, "step": 13220 }, { "epoch": 49.0, "grad_norm": 1.5557820796966553, "learning_rate": 0.00151, "loss": 1.5785, "step": 13230 }, { "epoch": 49.04, "grad_norm": 0.6234275102615356, "learning_rate": 0.0015096296296296296, "loss": 1.3406, "step": 13240 }, { "epoch": 49.07, "grad_norm": 0.6727663278579712, "learning_rate": 0.0015092592592592592, "loss": 1.3492, "step": 13250 }, { "epoch": 49.11, "grad_norm": 0.6145593523979187, "learning_rate": 0.001508888888888889, "loss": 1.3564, "step": 13260 }, { "epoch": 49.15, "grad_norm": 0.7036429047584534, "learning_rate": 0.0015085185185185186, "loss": 1.378, "step": 13270 }, { "epoch": 49.19, "grad_norm": 0.6589803695678711, "learning_rate": 0.0015081481481481482, "loss": 1.4108, "step": 13280 }, { "epoch": 49.22, "grad_norm": 0.6215453743934631, "learning_rate": 0.001507777777777778, "loss": 1.4335, "step": 13290 }, { "epoch": 49.26, "grad_norm": 0.6513487100601196, "learning_rate": 0.0015074074074074074, "loss": 1.4713, "step": 13300 }, { "epoch": 49.3, "grad_norm": 0.6613112688064575, "learning_rate": 0.001507037037037037, "loss": 1.4948, "step": 13310 }, { "epoch": 49.33, "grad_norm": 0.6666290760040283, "learning_rate": 0.0015066666666666666, "loss": 1.4837, "step": 13320 }, { "epoch": 49.37, "grad_norm": 0.6554436087608337, "learning_rate": 0.0015062962962962964, "loss": 1.4909, "step": 13330 }, { "epoch": 49.41, "grad_norm": 0.635359525680542, "learning_rate": 0.001505925925925926, "loss": 1.4803, "step": 13340 }, { "epoch": 49.44, "grad_norm": 0.6472628712654114, "learning_rate": 0.0015055555555555556, "loss": 1.5059, "step": 13350 }, { "epoch": 49.48, "grad_norm": 0.6820836663246155, "learning_rate": 0.0015051851851851851, "loss": 1.5074, "step": 13360 }, { "epoch": 49.52, "grad_norm": 0.6269857287406921, "learning_rate": 0.001504814814814815, "loss": 1.5613, "step": 13370 }, { "epoch": 49.56, "grad_norm": 0.6553345322608948, "learning_rate": 0.0015044444444444445, "loss": 1.5285, "step": 13380 }, { "epoch": 49.59, "grad_norm": 0.7129311561584473, "learning_rate": 0.0015040740740740741, "loss": 1.5444, "step": 13390 }, { "epoch": 49.63, "grad_norm": 0.6317949891090393, "learning_rate": 0.0015037037037037037, "loss": 1.5464, "step": 13400 }, { "epoch": 49.67, "grad_norm": 0.6542783975601196, "learning_rate": 0.0015033333333333335, "loss": 1.5649, "step": 13410 }, { "epoch": 49.7, "grad_norm": 0.645422101020813, "learning_rate": 0.001502962962962963, "loss": 1.5832, "step": 13420 }, { "epoch": 49.74, "grad_norm": 0.6377230286598206, "learning_rate": 0.0015025925925925927, "loss": 1.6086, "step": 13430 }, { "epoch": 49.78, "grad_norm": 0.6812024712562561, "learning_rate": 0.001502222222222222, "loss": 1.5837, "step": 13440 }, { "epoch": 49.81, "grad_norm": 0.666580855846405, "learning_rate": 0.0015018518518518519, "loss": 1.5813, "step": 13450 }, { "epoch": 49.85, "grad_norm": 0.7089651226997375, "learning_rate": 0.0015014814814814815, "loss": 1.5876, "step": 13460 }, { "epoch": 49.89, "grad_norm": 0.6948536038398743, "learning_rate": 0.001501111111111111, "loss": 1.5941, "step": 13470 }, { "epoch": 49.93, "grad_norm": 0.6506425738334656, "learning_rate": 0.0015007407407407406, "loss": 1.5814, "step": 13480 }, { "epoch": 49.96, "grad_norm": 0.6823866963386536, "learning_rate": 0.0015003703703703704, "loss": 1.6191, "step": 13490 }, { "epoch": 50.0, "grad_norm": 1.197648286819458, "learning_rate": 0.0015, "loss": 1.6013, "step": 13500 }, { "epoch": 50.04, "grad_norm": 0.6759616732597351, "learning_rate": 0.0014996296296296296, "loss": 1.3345, "step": 13510 }, { "epoch": 50.07, "grad_norm": 0.6753141283988953, "learning_rate": 0.0014992592592592594, "loss": 1.3395, "step": 13520 }, { "epoch": 50.11, "grad_norm": 0.6210965514183044, "learning_rate": 0.001498888888888889, "loss": 1.3955, "step": 13530 }, { "epoch": 50.15, "grad_norm": 0.6821613907814026, "learning_rate": 0.0014985185185185186, "loss": 1.4198, "step": 13540 }, { "epoch": 50.19, "grad_norm": 0.7067998647689819, "learning_rate": 0.0014981481481481482, "loss": 1.4207, "step": 13550 }, { "epoch": 50.22, "grad_norm": 0.6579769849777222, "learning_rate": 0.001497777777777778, "loss": 1.4274, "step": 13560 }, { "epoch": 50.26, "grad_norm": 0.666738748550415, "learning_rate": 0.0014974074074074074, "loss": 1.4537, "step": 13570 }, { "epoch": 50.3, "grad_norm": 0.7158663868904114, "learning_rate": 0.001497037037037037, "loss": 1.476, "step": 13580 }, { "epoch": 50.33, "grad_norm": 0.6293248534202576, "learning_rate": 0.0014966666666666665, "loss": 1.476, "step": 13590 }, { "epoch": 50.37, "grad_norm": 0.7110118865966797, "learning_rate": 0.0014962962962962963, "loss": 1.5088, "step": 13600 }, { "epoch": 50.41, "grad_norm": 0.6568295955657959, "learning_rate": 0.001495925925925926, "loss": 1.4764, "step": 13610 }, { "epoch": 50.44, "grad_norm": 0.6455584764480591, "learning_rate": 0.0014955555555555555, "loss": 1.4821, "step": 13620 }, { "epoch": 50.48, "grad_norm": 0.6564566493034363, "learning_rate": 0.0014951851851851851, "loss": 1.4976, "step": 13630 }, { "epoch": 50.52, "grad_norm": 0.6496867537498474, "learning_rate": 0.001494814814814815, "loss": 1.5238, "step": 13640 }, { "epoch": 50.56, "grad_norm": 0.6984143853187561, "learning_rate": 0.0014944444444444445, "loss": 1.5487, "step": 13650 }, { "epoch": 50.59, "grad_norm": 0.6508861184120178, "learning_rate": 0.001494074074074074, "loss": 1.5372, "step": 13660 }, { "epoch": 50.63, "grad_norm": 0.670586109161377, "learning_rate": 0.0014937037037037037, "loss": 1.5353, "step": 13670 }, { "epoch": 50.67, "grad_norm": 0.6368412971496582, "learning_rate": 0.0014933333333333335, "loss": 1.5751, "step": 13680 }, { "epoch": 50.7, "grad_norm": 0.6989796757698059, "learning_rate": 0.001492962962962963, "loss": 1.5406, "step": 13690 }, { "epoch": 50.74, "grad_norm": 0.6799941658973694, "learning_rate": 0.0014925925925925927, "loss": 1.5596, "step": 13700 }, { "epoch": 50.78, "grad_norm": 0.6923843622207642, "learning_rate": 0.001492222222222222, "loss": 1.5707, "step": 13710 }, { "epoch": 50.81, "grad_norm": 0.6336764097213745, "learning_rate": 0.0014918518518518518, "loss": 1.5833, "step": 13720 }, { "epoch": 50.85, "grad_norm": 0.6471168398857117, "learning_rate": 0.0014914814814814814, "loss": 1.5799, "step": 13730 }, { "epoch": 50.89, "grad_norm": 0.6990411877632141, "learning_rate": 0.001491111111111111, "loss": 1.5772, "step": 13740 }, { "epoch": 50.93, "grad_norm": 0.629309356212616, "learning_rate": 0.0014907407407407408, "loss": 1.6318, "step": 13750 }, { "epoch": 50.96, "grad_norm": 0.6218940615653992, "learning_rate": 0.0014903703703703704, "loss": 1.6012, "step": 13760 }, { "epoch": 51.0, "grad_norm": 1.1239666938781738, "learning_rate": 0.00149, "loss": 1.5763, "step": 13770 }, { "epoch": 51.04, "grad_norm": 0.6716301441192627, "learning_rate": 0.0014896296296296296, "loss": 1.3228, "step": 13780 }, { "epoch": 51.07, "grad_norm": 0.6490216851234436, "learning_rate": 0.0014892592592592594, "loss": 1.3701, "step": 13790 }, { "epoch": 51.11, "grad_norm": 0.6851688623428345, "learning_rate": 0.001488888888888889, "loss": 1.3892, "step": 13800 }, { "epoch": 51.15, "grad_norm": 0.6652293801307678, "learning_rate": 0.0014885185185185186, "loss": 1.3798, "step": 13810 }, { "epoch": 51.19, "grad_norm": 0.6425241231918335, "learning_rate": 0.0014881481481481482, "loss": 1.4337, "step": 13820 }, { "epoch": 51.22, "grad_norm": 0.6521861553192139, "learning_rate": 0.001487777777777778, "loss": 1.4287, "step": 13830 }, { "epoch": 51.26, "grad_norm": 0.6515931487083435, "learning_rate": 0.0014874074074074076, "loss": 1.4422, "step": 13840 }, { "epoch": 51.3, "grad_norm": 0.6525137424468994, "learning_rate": 0.001487037037037037, "loss": 1.4593, "step": 13850 }, { "epoch": 51.33, "grad_norm": 0.615257203578949, "learning_rate": 0.0014866666666666665, "loss": 1.4458, "step": 13860 }, { "epoch": 51.37, "grad_norm": 0.7334824800491333, "learning_rate": 0.0014862962962962963, "loss": 1.4622, "step": 13870 }, { "epoch": 51.41, "grad_norm": 0.6889981031417847, "learning_rate": 0.001485925925925926, "loss": 1.5, "step": 13880 }, { "epoch": 51.44, "grad_norm": 0.6559149026870728, "learning_rate": 0.0014855555555555555, "loss": 1.4843, "step": 13890 }, { "epoch": 51.48, "grad_norm": 0.6670100688934326, "learning_rate": 0.001485185185185185, "loss": 1.5008, "step": 13900 }, { "epoch": 51.52, "grad_norm": 0.6622434258460999, "learning_rate": 0.001484814814814815, "loss": 1.5177, "step": 13910 }, { "epoch": 51.56, "grad_norm": 0.6639595627784729, "learning_rate": 0.0014844444444444445, "loss": 1.5215, "step": 13920 }, { "epoch": 51.59, "grad_norm": 0.6925778388977051, "learning_rate": 0.001484074074074074, "loss": 1.5358, "step": 13930 }, { "epoch": 51.63, "grad_norm": 0.6861134171485901, "learning_rate": 0.0014837037037037039, "loss": 1.566, "step": 13940 }, { "epoch": 51.67, "grad_norm": 0.6408432126045227, "learning_rate": 0.0014833333333333335, "loss": 1.5612, "step": 13950 }, { "epoch": 51.7, "grad_norm": 0.679660439491272, "learning_rate": 0.001482962962962963, "loss": 1.558, "step": 13960 }, { "epoch": 51.74, "grad_norm": 0.6439672708511353, "learning_rate": 0.0014825925925925926, "loss": 1.5728, "step": 13970 }, { "epoch": 51.78, "grad_norm": 0.6197627782821655, "learning_rate": 0.0014822222222222224, "loss": 1.5413, "step": 13980 }, { "epoch": 51.81, "grad_norm": 0.6940560340881348, "learning_rate": 0.0014818518518518518, "loss": 1.5327, "step": 13990 }, { "epoch": 51.85, "grad_norm": 0.6489192843437195, "learning_rate": 0.0014814814814814814, "loss": 1.5833, "step": 14000 }, { "epoch": 51.89, "grad_norm": 0.7189292907714844, "learning_rate": 0.001481111111111111, "loss": 1.5644, "step": 14010 }, { "epoch": 51.93, "grad_norm": 0.7071329951286316, "learning_rate": 0.0014807407407407408, "loss": 1.5661, "step": 14020 }, { "epoch": 51.96, "grad_norm": 0.6607621908187866, "learning_rate": 0.0014803703703703704, "loss": 1.5743, "step": 14030 }, { "epoch": 52.0, "grad_norm": 1.4042985439300537, "learning_rate": 0.00148, "loss": 1.5808, "step": 14040 }, { "epoch": 52.04, "grad_norm": 0.625298798084259, "learning_rate": 0.0014796296296296296, "loss": 1.3482, "step": 14050 }, { "epoch": 52.07, "grad_norm": 0.6287106871604919, "learning_rate": 0.0014792592592592594, "loss": 1.3591, "step": 14060 }, { "epoch": 52.11, "grad_norm": 0.6880126595497131, "learning_rate": 0.001478888888888889, "loss": 1.3634, "step": 14070 }, { "epoch": 52.15, "grad_norm": 0.7166730761528015, "learning_rate": 0.0014785185185185185, "loss": 1.3566, "step": 14080 }, { "epoch": 52.19, "grad_norm": 0.7044410109519958, "learning_rate": 0.0014781481481481481, "loss": 1.3779, "step": 14090 }, { "epoch": 52.22, "grad_norm": 0.7081303000450134, "learning_rate": 0.001477777777777778, "loss": 1.4112, "step": 14100 }, { "epoch": 52.26, "grad_norm": 0.7148960828781128, "learning_rate": 0.0014774074074074075, "loss": 1.4383, "step": 14110 }, { "epoch": 52.3, "grad_norm": 0.722838282585144, "learning_rate": 0.001477037037037037, "loss": 1.4278, "step": 14120 }, { "epoch": 52.33, "grad_norm": 0.6913270354270935, "learning_rate": 0.0014766666666666667, "loss": 1.4577, "step": 14130 }, { "epoch": 52.37, "grad_norm": 0.6681016087532043, "learning_rate": 0.0014762962962962963, "loss": 1.4674, "step": 14140 }, { "epoch": 52.41, "grad_norm": 0.7427762746810913, "learning_rate": 0.0014759259259259259, "loss": 1.4696, "step": 14150 }, { "epoch": 52.44, "grad_norm": 0.744890034198761, "learning_rate": 0.0014755555555555555, "loss": 1.5042, "step": 14160 }, { "epoch": 52.48, "grad_norm": 0.7112950682640076, "learning_rate": 0.0014751851851851853, "loss": 1.4885, "step": 14170 }, { "epoch": 52.52, "grad_norm": 0.6994560956954956, "learning_rate": 0.0014748148148148149, "loss": 1.527, "step": 14180 }, { "epoch": 52.56, "grad_norm": 0.6381304264068604, "learning_rate": 0.0014744444444444445, "loss": 1.5282, "step": 14190 }, { "epoch": 52.59, "grad_norm": 0.6289535164833069, "learning_rate": 0.001474074074074074, "loss": 1.5379, "step": 14200 }, { "epoch": 52.63, "grad_norm": 0.7018424868583679, "learning_rate": 0.0014737037037037039, "loss": 1.5422, "step": 14210 }, { "epoch": 52.67, "grad_norm": 0.6578280329704285, "learning_rate": 0.0014733333333333334, "loss": 1.5505, "step": 14220 }, { "epoch": 52.7, "grad_norm": 0.6513351202011108, "learning_rate": 0.001472962962962963, "loss": 1.533, "step": 14230 }, { "epoch": 52.74, "grad_norm": 0.6813905835151672, "learning_rate": 0.0014725925925925926, "loss": 1.5365, "step": 14240 }, { "epoch": 52.78, "grad_norm": 0.7050051093101501, "learning_rate": 0.0014722222222222224, "loss": 1.5358, "step": 14250 }, { "epoch": 52.81, "grad_norm": 0.6147084832191467, "learning_rate": 0.0014718518518518518, "loss": 1.5515, "step": 14260 }, { "epoch": 52.85, "grad_norm": 0.6574007868766785, "learning_rate": 0.0014714814814814814, "loss": 1.5587, "step": 14270 }, { "epoch": 52.89, "grad_norm": 0.74244225025177, "learning_rate": 0.001471111111111111, "loss": 1.5827, "step": 14280 }, { "epoch": 52.93, "grad_norm": 0.6948452591896057, "learning_rate": 0.0014707407407407408, "loss": 1.5675, "step": 14290 }, { "epoch": 52.96, "grad_norm": 0.6986340880393982, "learning_rate": 0.0014703703703703704, "loss": 1.5783, "step": 14300 }, { "epoch": 53.0, "grad_norm": 1.0962846279144287, "learning_rate": 0.00147, "loss": 1.5808, "step": 14310 }, { "epoch": 53.04, "grad_norm": 0.6880574822425842, "learning_rate": 0.0014696296296296298, "loss": 1.3356, "step": 14320 }, { "epoch": 53.07, "grad_norm": 0.6791065335273743, "learning_rate": 0.0014692592592592593, "loss": 1.3505, "step": 14330 }, { "epoch": 53.11, "grad_norm": 0.6797129511833191, "learning_rate": 0.001468888888888889, "loss": 1.3387, "step": 14340 }, { "epoch": 53.15, "grad_norm": 0.6955825686454773, "learning_rate": 0.0014685185185185185, "loss": 1.3706, "step": 14350 }, { "epoch": 53.19, "grad_norm": 0.6407482028007507, "learning_rate": 0.0014681481481481483, "loss": 1.3851, "step": 14360 }, { "epoch": 53.22, "grad_norm": 0.7131444215774536, "learning_rate": 0.001467777777777778, "loss": 1.4148, "step": 14370 }, { "epoch": 53.26, "grad_norm": 0.7091857194900513, "learning_rate": 0.0014674074074074075, "loss": 1.438, "step": 14380 }, { "epoch": 53.3, "grad_norm": 0.6437903046607971, "learning_rate": 0.001467037037037037, "loss": 1.4216, "step": 14390 }, { "epoch": 53.33, "grad_norm": 0.6952517628669739, "learning_rate": 0.0014666666666666667, "loss": 1.4454, "step": 14400 }, { "epoch": 53.37, "grad_norm": 0.6853577494621277, "learning_rate": 0.0014662962962962963, "loss": 1.4695, "step": 14410 }, { "epoch": 53.41, "grad_norm": 0.683782696723938, "learning_rate": 0.0014659259259259259, "loss": 1.4711, "step": 14420 }, { "epoch": 53.44, "grad_norm": 0.6749939918518066, "learning_rate": 0.0014655555555555554, "loss": 1.4837, "step": 14430 }, { "epoch": 53.48, "grad_norm": 0.6679930090904236, "learning_rate": 0.0014651851851851853, "loss": 1.4844, "step": 14440 }, { "epoch": 53.52, "grad_norm": 0.7036862373352051, "learning_rate": 0.0014648148148148148, "loss": 1.478, "step": 14450 }, { "epoch": 53.56, "grad_norm": 0.6320893168449402, "learning_rate": 0.0014644444444444444, "loss": 1.5058, "step": 14460 }, { "epoch": 53.59, "grad_norm": 0.6774411797523499, "learning_rate": 0.001464074074074074, "loss": 1.5123, "step": 14470 }, { "epoch": 53.63, "grad_norm": 0.7401627898216248, "learning_rate": 0.0014637037037037038, "loss": 1.5269, "step": 14480 }, { "epoch": 53.67, "grad_norm": 0.6697722673416138, "learning_rate": 0.0014633333333333334, "loss": 1.5253, "step": 14490 }, { "epoch": 53.7, "grad_norm": 0.7028366923332214, "learning_rate": 0.001462962962962963, "loss": 1.5125, "step": 14500 }, { "epoch": 53.74, "grad_norm": 0.6694210171699524, "learning_rate": 0.0014625925925925926, "loss": 1.5426, "step": 14510 }, { "epoch": 53.78, "grad_norm": 0.6523129343986511, "learning_rate": 0.0014622222222222224, "loss": 1.5194, "step": 14520 }, { "epoch": 53.81, "grad_norm": 0.6751163005828857, "learning_rate": 0.001461851851851852, "loss": 1.5516, "step": 14530 }, { "epoch": 53.85, "grad_norm": 0.6803255081176758, "learning_rate": 0.0014614814814814814, "loss": 1.5671, "step": 14540 }, { "epoch": 53.89, "grad_norm": 0.7053388357162476, "learning_rate": 0.0014611111111111112, "loss": 1.5692, "step": 14550 }, { "epoch": 53.93, "grad_norm": 0.6673864722251892, "learning_rate": 0.0014607407407407407, "loss": 1.581, "step": 14560 }, { "epoch": 53.96, "grad_norm": 0.6886114478111267, "learning_rate": 0.0014603703703703703, "loss": 1.5639, "step": 14570 }, { "epoch": 54.0, "grad_norm": 1.3442867994308472, "learning_rate": 0.00146, "loss": 1.5796, "step": 14580 }, { "epoch": 54.04, "grad_norm": 0.669252336025238, "learning_rate": 0.0014596296296296297, "loss": 1.3192, "step": 14590 }, { "epoch": 54.07, "grad_norm": 0.6675966382026672, "learning_rate": 0.0014592592592592593, "loss": 1.353, "step": 14600 }, { "epoch": 54.11, "grad_norm": 0.7156364321708679, "learning_rate": 0.001458888888888889, "loss": 1.3499, "step": 14610 }, { "epoch": 54.15, "grad_norm": 0.7250548005104065, "learning_rate": 0.0014585185185185185, "loss": 1.3749, "step": 14620 }, { "epoch": 54.19, "grad_norm": 0.6688981652259827, "learning_rate": 0.0014581481481481483, "loss": 1.3711, "step": 14630 }, { "epoch": 54.22, "grad_norm": 0.7125627398490906, "learning_rate": 0.0014577777777777779, "loss": 1.3894, "step": 14640 }, { "epoch": 54.26, "grad_norm": 0.6876236796379089, "learning_rate": 0.0014574074074074075, "loss": 1.3956, "step": 14650 }, { "epoch": 54.3, "grad_norm": 0.7455447912216187, "learning_rate": 0.001457037037037037, "loss": 1.4206, "step": 14660 }, { "epoch": 54.33, "grad_norm": 0.7188883423805237, "learning_rate": 0.0014566666666666669, "loss": 1.418, "step": 14670 }, { "epoch": 54.37, "grad_norm": 0.6575316190719604, "learning_rate": 0.0014562962962962962, "loss": 1.4615, "step": 14680 }, { "epoch": 54.41, "grad_norm": 0.6895075440406799, "learning_rate": 0.0014559259259259258, "loss": 1.4702, "step": 14690 }, { "epoch": 54.44, "grad_norm": 0.7062181830406189, "learning_rate": 0.0014555555555555554, "loss": 1.5067, "step": 14700 }, { "epoch": 54.48, "grad_norm": 0.7163558602333069, "learning_rate": 0.0014551851851851852, "loss": 1.4755, "step": 14710 }, { "epoch": 54.52, "grad_norm": 0.644522488117218, "learning_rate": 0.0014548148148148148, "loss": 1.4929, "step": 14720 }, { "epoch": 54.56, "grad_norm": 0.7128614187240601, "learning_rate": 0.0014544444444444444, "loss": 1.5137, "step": 14730 }, { "epoch": 54.59, "grad_norm": 0.6865217685699463, "learning_rate": 0.0014540740740740742, "loss": 1.4744, "step": 14740 }, { "epoch": 54.63, "grad_norm": 0.6782078742980957, "learning_rate": 0.0014537037037037038, "loss": 1.5026, "step": 14750 }, { "epoch": 54.67, "grad_norm": 0.6496673226356506, "learning_rate": 0.0014533333333333334, "loss": 1.5087, "step": 14760 }, { "epoch": 54.7, "grad_norm": 0.7069528102874756, "learning_rate": 0.001452962962962963, "loss": 1.5458, "step": 14770 }, { "epoch": 54.74, "grad_norm": 0.6799771785736084, "learning_rate": 0.0014525925925925928, "loss": 1.5454, "step": 14780 }, { "epoch": 54.78, "grad_norm": 0.6214430928230286, "learning_rate": 0.0014522222222222224, "loss": 1.5295, "step": 14790 }, { "epoch": 54.81, "grad_norm": 0.6988822221755981, "learning_rate": 0.001451851851851852, "loss": 1.5521, "step": 14800 }, { "epoch": 54.85, "grad_norm": 0.7073298096656799, "learning_rate": 0.0014514814814814813, "loss": 1.5719, "step": 14810 }, { "epoch": 54.89, "grad_norm": 0.6836145520210266, "learning_rate": 0.0014511111111111111, "loss": 1.5707, "step": 14820 }, { "epoch": 54.93, "grad_norm": 0.6364758014678955, "learning_rate": 0.0014507407407407407, "loss": 1.5669, "step": 14830 }, { "epoch": 54.96, "grad_norm": 0.7107850909233093, "learning_rate": 0.0014503703703703703, "loss": 1.5315, "step": 14840 }, { "epoch": 55.0, "grad_norm": 1.1641267538070679, "learning_rate": 0.00145, "loss": 1.5796, "step": 14850 }, { "epoch": 55.04, "grad_norm": 0.7165002822875977, "learning_rate": 0.0014496296296296297, "loss": 1.3244, "step": 14860 }, { "epoch": 55.07, "grad_norm": 0.7347918152809143, "learning_rate": 0.0014492592592592593, "loss": 1.3162, "step": 14870 }, { "epoch": 55.11, "grad_norm": 0.623955488204956, "learning_rate": 0.0014488888888888889, "loss": 1.3492, "step": 14880 }, { "epoch": 55.15, "grad_norm": 0.6705119609832764, "learning_rate": 0.0014485185185185185, "loss": 1.3307, "step": 14890 }, { "epoch": 55.19, "grad_norm": 0.6863866448402405, "learning_rate": 0.0014481481481481483, "loss": 1.4045, "step": 14900 }, { "epoch": 55.22, "grad_norm": 0.727873682975769, "learning_rate": 0.0014477777777777779, "loss": 1.3976, "step": 14910 }, { "epoch": 55.26, "grad_norm": 0.6954161524772644, "learning_rate": 0.0014474074074074075, "loss": 1.4039, "step": 14920 }, { "epoch": 55.3, "grad_norm": 0.6985371112823486, "learning_rate": 0.0014470370370370373, "loss": 1.4561, "step": 14930 }, { "epoch": 55.33, "grad_norm": 0.6702157258987427, "learning_rate": 0.0014466666666666668, "loss": 1.4516, "step": 14940 }, { "epoch": 55.37, "grad_norm": 0.6751401424407959, "learning_rate": 0.0014462962962962962, "loss": 1.4431, "step": 14950 }, { "epoch": 55.41, "grad_norm": 0.6904053092002869, "learning_rate": 0.0014459259259259258, "loss": 1.472, "step": 14960 }, { "epoch": 55.44, "grad_norm": 0.6829718351364136, "learning_rate": 0.0014455555555555556, "loss": 1.4178, "step": 14970 }, { "epoch": 55.48, "grad_norm": 0.7579282522201538, "learning_rate": 0.0014451851851851852, "loss": 1.4848, "step": 14980 }, { "epoch": 55.52, "grad_norm": 0.6680785417556763, "learning_rate": 0.0014448148148148148, "loss": 1.4719, "step": 14990 }, { "epoch": 55.56, "grad_norm": 0.7036536335945129, "learning_rate": 0.0014444444444444444, "loss": 1.4953, "step": 15000 }, { "epoch": 55.59, "grad_norm": 0.6639661192893982, "learning_rate": 0.0014440740740740742, "loss": 1.5032, "step": 15010 }, { "epoch": 55.63, "grad_norm": 0.6836636662483215, "learning_rate": 0.0014437037037037038, "loss": 1.4788, "step": 15020 }, { "epoch": 55.67, "grad_norm": 0.7251687049865723, "learning_rate": 0.0014433333333333334, "loss": 1.5103, "step": 15030 }, { "epoch": 55.7, "grad_norm": 0.7114871144294739, "learning_rate": 0.001442962962962963, "loss": 1.4929, "step": 15040 }, { "epoch": 55.74, "grad_norm": 0.6300182342529297, "learning_rate": 0.0014425925925925928, "loss": 1.537, "step": 15050 }, { "epoch": 55.78, "grad_norm": 0.6930983662605286, "learning_rate": 0.0014422222222222223, "loss": 1.5433, "step": 15060 }, { "epoch": 55.81, "grad_norm": 0.7037534713745117, "learning_rate": 0.001441851851851852, "loss": 1.5347, "step": 15070 }, { "epoch": 55.85, "grad_norm": 0.6565731763839722, "learning_rate": 0.0014414814814814815, "loss": 1.5356, "step": 15080 }, { "epoch": 55.89, "grad_norm": 0.6927472949028015, "learning_rate": 0.001441111111111111, "loss": 1.5415, "step": 15090 }, { "epoch": 55.93, "grad_norm": 0.6429632306098938, "learning_rate": 0.0014407407407407407, "loss": 1.5364, "step": 15100 }, { "epoch": 55.96, "grad_norm": 0.6984735727310181, "learning_rate": 0.0014403703703703703, "loss": 1.5524, "step": 15110 }, { "epoch": 56.0, "grad_norm": 1.1884384155273438, "learning_rate": 0.0014399999999999999, "loss": 1.5601, "step": 15120 }, { "epoch": 56.04, "grad_norm": 0.7488458752632141, "learning_rate": 0.0014396296296296297, "loss": 1.3167, "step": 15130 }, { "epoch": 56.07, "grad_norm": 0.6538281440734863, "learning_rate": 0.0014392592592592593, "loss": 1.3421, "step": 15140 }, { "epoch": 56.11, "grad_norm": 0.711005449295044, "learning_rate": 0.0014388888888888889, "loss": 1.3511, "step": 15150 }, { "epoch": 56.15, "grad_norm": 0.656951904296875, "learning_rate": 0.0014385185185185187, "loss": 1.357, "step": 15160 }, { "epoch": 56.19, "grad_norm": 0.7439039945602417, "learning_rate": 0.0014381481481481483, "loss": 1.3834, "step": 15170 }, { "epoch": 56.22, "grad_norm": 0.6957058906555176, "learning_rate": 0.0014377777777777778, "loss": 1.3929, "step": 15180 }, { "epoch": 56.26, "grad_norm": 0.6609116792678833, "learning_rate": 0.0014374074074074074, "loss": 1.3745, "step": 15190 }, { "epoch": 56.3, "grad_norm": 0.7077909708023071, "learning_rate": 0.0014370370370370372, "loss": 1.4118, "step": 15200 }, { "epoch": 56.33, "grad_norm": 0.68143230676651, "learning_rate": 0.0014366666666666668, "loss": 1.4364, "step": 15210 }, { "epoch": 56.37, "grad_norm": 0.6971974968910217, "learning_rate": 0.0014362962962962964, "loss": 1.4552, "step": 15220 }, { "epoch": 56.41, "grad_norm": 0.6728135347366333, "learning_rate": 0.0014359259259259258, "loss": 1.4186, "step": 15230 }, { "epoch": 56.44, "grad_norm": 0.7541175484657288, "learning_rate": 0.0014355555555555556, "loss": 1.4607, "step": 15240 }, { "epoch": 56.48, "grad_norm": 0.7734980583190918, "learning_rate": 0.0014351851851851852, "loss": 1.4602, "step": 15250 }, { "epoch": 56.52, "grad_norm": 0.7637688517570496, "learning_rate": 0.0014348148148148148, "loss": 1.4519, "step": 15260 }, { "epoch": 56.56, "grad_norm": 0.701262354850769, "learning_rate": 0.0014344444444444444, "loss": 1.4738, "step": 15270 }, { "epoch": 56.59, "grad_norm": 0.7055422067642212, "learning_rate": 0.0014340740740740742, "loss": 1.4991, "step": 15280 }, { "epoch": 56.63, "grad_norm": 0.6790987253189087, "learning_rate": 0.0014337037037037037, "loss": 1.4904, "step": 15290 }, { "epoch": 56.67, "grad_norm": 0.6938120126724243, "learning_rate": 0.0014333333333333333, "loss": 1.5177, "step": 15300 }, { "epoch": 56.7, "grad_norm": 0.6558261513710022, "learning_rate": 0.001432962962962963, "loss": 1.4946, "step": 15310 }, { "epoch": 56.74, "grad_norm": 0.7213912010192871, "learning_rate": 0.0014325925925925927, "loss": 1.5161, "step": 15320 }, { "epoch": 56.78, "grad_norm": 0.6817654967308044, "learning_rate": 0.0014322222222222223, "loss": 1.5097, "step": 15330 }, { "epoch": 56.81, "grad_norm": 0.669110894203186, "learning_rate": 0.001431851851851852, "loss": 1.5273, "step": 15340 }, { "epoch": 56.85, "grad_norm": 0.6431531310081482, "learning_rate": 0.0014314814814814817, "loss": 1.5635, "step": 15350 }, { "epoch": 56.89, "grad_norm": 0.7609390020370483, "learning_rate": 0.001431111111111111, "loss": 1.5422, "step": 15360 }, { "epoch": 56.93, "grad_norm": 0.7726841568946838, "learning_rate": 0.0014307407407407407, "loss": 1.5369, "step": 15370 }, { "epoch": 56.96, "grad_norm": 0.6646488308906555, "learning_rate": 0.0014303703703703703, "loss": 1.5354, "step": 15380 }, { "epoch": 57.0, "grad_norm": 1.5030916929244995, "learning_rate": 0.00143, "loss": 1.5473, "step": 15390 }, { "epoch": 57.04, "grad_norm": 0.7018294930458069, "learning_rate": 0.0014296296296296297, "loss": 1.3131, "step": 15400 }, { "epoch": 57.07, "grad_norm": 0.6784168481826782, "learning_rate": 0.0014292592592592592, "loss": 1.3124, "step": 15410 }, { "epoch": 57.11, "grad_norm": 0.702157735824585, "learning_rate": 0.0014288888888888888, "loss": 1.3219, "step": 15420 }, { "epoch": 57.15, "grad_norm": 0.6698986291885376, "learning_rate": 0.0014285185185185186, "loss": 1.3259, "step": 15430 }, { "epoch": 57.19, "grad_norm": 0.7673943638801575, "learning_rate": 0.0014281481481481482, "loss": 1.3273, "step": 15440 }, { "epoch": 57.22, "grad_norm": 0.6996111869812012, "learning_rate": 0.0014277777777777778, "loss": 1.3506, "step": 15450 }, { "epoch": 57.26, "grad_norm": 0.7185860276222229, "learning_rate": 0.0014274074074074074, "loss": 1.4138, "step": 15460 }, { "epoch": 57.3, "grad_norm": 0.6754123568534851, "learning_rate": 0.0014270370370370372, "loss": 1.4118, "step": 15470 }, { "epoch": 57.33, "grad_norm": 0.7364366054534912, "learning_rate": 0.0014266666666666668, "loss": 1.4377, "step": 15480 }, { "epoch": 57.37, "grad_norm": 0.691130518913269, "learning_rate": 0.0014262962962962964, "loss": 1.4192, "step": 15490 }, { "epoch": 57.41, "grad_norm": 0.703258216381073, "learning_rate": 0.0014259259259259258, "loss": 1.4592, "step": 15500 }, { "epoch": 57.44, "grad_norm": 0.6907350420951843, "learning_rate": 0.0014255555555555556, "loss": 1.4591, "step": 15510 }, { "epoch": 57.48, "grad_norm": 0.6878598928451538, "learning_rate": 0.0014251851851851851, "loss": 1.4409, "step": 15520 }, { "epoch": 57.52, "grad_norm": 0.694089412689209, "learning_rate": 0.0014248148148148147, "loss": 1.4814, "step": 15530 }, { "epoch": 57.56, "grad_norm": 0.677599310874939, "learning_rate": 0.0014244444444444443, "loss": 1.4811, "step": 15540 }, { "epoch": 57.59, "grad_norm": 0.7053378224372864, "learning_rate": 0.0014240740740740741, "loss": 1.5, "step": 15550 }, { "epoch": 57.63, "grad_norm": 0.7823730707168579, "learning_rate": 0.0014237037037037037, "loss": 1.5035, "step": 15560 }, { "epoch": 57.67, "grad_norm": 0.7754201292991638, "learning_rate": 0.0014233333333333333, "loss": 1.491, "step": 15570 }, { "epoch": 57.7, "grad_norm": 0.742430567741394, "learning_rate": 0.0014229629629629631, "loss": 1.4948, "step": 15580 }, { "epoch": 57.74, "grad_norm": 0.7142267823219299, "learning_rate": 0.0014225925925925927, "loss": 1.4721, "step": 15590 }, { "epoch": 57.78, "grad_norm": 0.6918465495109558, "learning_rate": 0.0014222222222222223, "loss": 1.5362, "step": 15600 }, { "epoch": 57.81, "grad_norm": 0.7011120915412903, "learning_rate": 0.0014218518518518519, "loss": 1.5254, "step": 15610 }, { "epoch": 57.85, "grad_norm": 0.6877245903015137, "learning_rate": 0.0014214814814814817, "loss": 1.4972, "step": 15620 }, { "epoch": 57.89, "grad_norm": 0.6386919021606445, "learning_rate": 0.0014211111111111113, "loss": 1.5306, "step": 15630 }, { "epoch": 57.93, "grad_norm": 0.7149176597595215, "learning_rate": 0.0014207407407407406, "loss": 1.5251, "step": 15640 }, { "epoch": 57.96, "grad_norm": 0.6852824091911316, "learning_rate": 0.0014203703703703702, "loss": 1.5594, "step": 15650 }, { "epoch": 58.0, "grad_norm": 1.1446104049682617, "learning_rate": 0.00142, "loss": 1.5287, "step": 15660 }, { "epoch": 58.04, "grad_norm": 0.7290907502174377, "learning_rate": 0.0014196296296296296, "loss": 1.2803, "step": 15670 }, { "epoch": 58.07, "grad_norm": 0.6596196889877319, "learning_rate": 0.0014192592592592592, "loss": 1.3151, "step": 15680 }, { "epoch": 58.11, "grad_norm": 0.6732152104377747, "learning_rate": 0.0014188888888888888, "loss": 1.3397, "step": 15690 }, { "epoch": 58.15, "grad_norm": 0.7772537469863892, "learning_rate": 0.0014185185185185186, "loss": 1.3527, "step": 15700 }, { "epoch": 58.19, "grad_norm": 0.7557883858680725, "learning_rate": 0.0014181481481481482, "loss": 1.3314, "step": 15710 }, { "epoch": 58.22, "grad_norm": 0.7082036733627319, "learning_rate": 0.0014177777777777778, "loss": 1.3414, "step": 15720 }, { "epoch": 58.26, "grad_norm": 0.773300290107727, "learning_rate": 0.0014174074074074074, "loss": 1.3594, "step": 15730 }, { "epoch": 58.3, "grad_norm": 0.6985162496566772, "learning_rate": 0.0014170370370370372, "loss": 1.3939, "step": 15740 }, { "epoch": 58.33, "grad_norm": 0.8249867558479309, "learning_rate": 0.0014166666666666668, "loss": 1.4084, "step": 15750 }, { "epoch": 58.37, "grad_norm": 0.6576926112174988, "learning_rate": 0.0014162962962962964, "loss": 1.4324, "step": 15760 }, { "epoch": 58.41, "grad_norm": 0.7057226896286011, "learning_rate": 0.0014159259259259262, "loss": 1.4467, "step": 15770 }, { "epoch": 58.44, "grad_norm": 0.6919978260993958, "learning_rate": 0.0014155555555555555, "loss": 1.4659, "step": 15780 }, { "epoch": 58.48, "grad_norm": 0.7028883695602417, "learning_rate": 0.0014151851851851851, "loss": 1.4111, "step": 15790 }, { "epoch": 58.52, "grad_norm": 0.7400588989257812, "learning_rate": 0.0014148148148148147, "loss": 1.4447, "step": 15800 }, { "epoch": 58.56, "grad_norm": 0.6596423983573914, "learning_rate": 0.0014144444444444445, "loss": 1.4724, "step": 15810 }, { "epoch": 58.59, "grad_norm": 0.6897073984146118, "learning_rate": 0.001414074074074074, "loss": 1.4845, "step": 15820 }, { "epoch": 58.63, "grad_norm": 0.7325702905654907, "learning_rate": 0.0014137037037037037, "loss": 1.4635, "step": 15830 }, { "epoch": 58.67, "grad_norm": 0.6876033544540405, "learning_rate": 0.0014133333333333333, "loss": 1.5037, "step": 15840 }, { "epoch": 58.7, "grad_norm": 0.7476497888565063, "learning_rate": 0.001412962962962963, "loss": 1.5093, "step": 15850 }, { "epoch": 58.74, "grad_norm": 0.7862426042556763, "learning_rate": 0.0014125925925925927, "loss": 1.4884, "step": 15860 }, { "epoch": 58.78, "grad_norm": 0.6755629777908325, "learning_rate": 0.0014122222222222223, "loss": 1.5251, "step": 15870 }, { "epoch": 58.81, "grad_norm": 0.6785138845443726, "learning_rate": 0.0014118518518518519, "loss": 1.4881, "step": 15880 }, { "epoch": 58.85, "grad_norm": 0.6602843999862671, "learning_rate": 0.0014114814814814817, "loss": 1.4935, "step": 15890 }, { "epoch": 58.89, "grad_norm": 0.6840865015983582, "learning_rate": 0.0014111111111111112, "loss": 1.5311, "step": 15900 }, { "epoch": 58.93, "grad_norm": 0.7246437668800354, "learning_rate": 0.0014107407407407408, "loss": 1.5321, "step": 15910 }, { "epoch": 58.96, "grad_norm": 0.7134275436401367, "learning_rate": 0.0014103703703703702, "loss": 1.5445, "step": 15920 }, { "epoch": 59.0, "grad_norm": 1.0462781190872192, "learning_rate": 0.00141, "loss": 1.5421, "step": 15930 }, { "epoch": 59.04, "grad_norm": 0.7037992477416992, "learning_rate": 0.0014096296296296296, "loss": 1.2916, "step": 15940 }, { "epoch": 59.07, "grad_norm": 0.7080190777778625, "learning_rate": 0.0014092592592592592, "loss": 1.2838, "step": 15950 }, { "epoch": 59.11, "grad_norm": 0.6890296936035156, "learning_rate": 0.001408888888888889, "loss": 1.3009, "step": 15960 }, { "epoch": 59.15, "grad_norm": 0.7467537522315979, "learning_rate": 0.0014085185185185186, "loss": 1.3518, "step": 15970 }, { "epoch": 59.19, "grad_norm": 0.6879865527153015, "learning_rate": 0.0014081481481481482, "loss": 1.3118, "step": 15980 }, { "epoch": 59.22, "grad_norm": 0.7323064208030701, "learning_rate": 0.0014077777777777778, "loss": 1.3767, "step": 15990 }, { "epoch": 59.26, "grad_norm": 0.7376371622085571, "learning_rate": 0.0014074074074074076, "loss": 1.3924, "step": 16000 }, { "epoch": 59.3, "grad_norm": 0.7222793102264404, "learning_rate": 0.0014070370370370372, "loss": 1.3829, "step": 16010 }, { "epoch": 59.33, "grad_norm": 0.7288392186164856, "learning_rate": 0.0014066666666666667, "loss": 1.4339, "step": 16020 }, { "epoch": 59.37, "grad_norm": 0.7137278318405151, "learning_rate": 0.0014062962962962963, "loss": 1.4027, "step": 16030 }, { "epoch": 59.41, "grad_norm": 0.7457606792449951, "learning_rate": 0.0014059259259259261, "loss": 1.4154, "step": 16040 }, { "epoch": 59.44, "grad_norm": 0.7324532866477966, "learning_rate": 0.0014055555555555555, "loss": 1.4614, "step": 16050 }, { "epoch": 59.48, "grad_norm": 0.6916303634643555, "learning_rate": 0.001405185185185185, "loss": 1.3899, "step": 16060 }, { "epoch": 59.52, "grad_norm": 0.7010741233825684, "learning_rate": 0.0014048148148148147, "loss": 1.4602, "step": 16070 }, { "epoch": 59.56, "grad_norm": 0.7448731064796448, "learning_rate": 0.0014044444444444445, "loss": 1.4703, "step": 16080 }, { "epoch": 59.59, "grad_norm": 0.6861727833747864, "learning_rate": 0.001404074074074074, "loss": 1.4591, "step": 16090 }, { "epoch": 59.63, "grad_norm": 0.7314042448997498, "learning_rate": 0.0014037037037037037, "loss": 1.4722, "step": 16100 }, { "epoch": 59.67, "grad_norm": 0.656716525554657, "learning_rate": 0.0014033333333333333, "loss": 1.4766, "step": 16110 }, { "epoch": 59.7, "grad_norm": 0.7024694085121155, "learning_rate": 0.001402962962962963, "loss": 1.5067, "step": 16120 }, { "epoch": 59.74, "grad_norm": 0.7215801477432251, "learning_rate": 0.0014025925925925926, "loss": 1.4881, "step": 16130 }, { "epoch": 59.78, "grad_norm": 0.6713820695877075, "learning_rate": 0.0014022222222222222, "loss": 1.5, "step": 16140 }, { "epoch": 59.81, "grad_norm": 0.6949275732040405, "learning_rate": 0.0014018518518518518, "loss": 1.5227, "step": 16150 }, { "epoch": 59.85, "grad_norm": 0.720103919506073, "learning_rate": 0.0014014814814814816, "loss": 1.5081, "step": 16160 }, { "epoch": 59.89, "grad_norm": 0.7111461162567139, "learning_rate": 0.0014011111111111112, "loss": 1.4886, "step": 16170 }, { "epoch": 59.93, "grad_norm": 0.7205804586410522, "learning_rate": 0.0014007407407407408, "loss": 1.5198, "step": 16180 }, { "epoch": 59.96, "grad_norm": 0.6634063124656677, "learning_rate": 0.0014003703703703704, "loss": 1.5365, "step": 16190 }, { "epoch": 60.0, "grad_norm": 1.360110878944397, "learning_rate": 0.0014, "loss": 1.5187, "step": 16200 }, { "epoch": 60.04, "grad_norm": 0.6875532865524292, "learning_rate": 0.0013996296296296296, "loss": 1.2746, "step": 16210 }, { "epoch": 60.07, "grad_norm": 0.7562872767448425, "learning_rate": 0.0013992592592592592, "loss": 1.2724, "step": 16220 }, { "epoch": 60.11, "grad_norm": 0.7381096482276917, "learning_rate": 0.001398888888888889, "loss": 1.3072, "step": 16230 }, { "epoch": 60.15, "grad_norm": 0.7193554043769836, "learning_rate": 0.0013985185185185186, "loss": 1.3413, "step": 16240 }, { "epoch": 60.19, "grad_norm": 0.6620252728462219, "learning_rate": 0.0013981481481481481, "loss": 1.35, "step": 16250 }, { "epoch": 60.22, "grad_norm": 0.7208623886108398, "learning_rate": 0.0013977777777777777, "loss": 1.3402, "step": 16260 }, { "epoch": 60.26, "grad_norm": 0.7059289813041687, "learning_rate": 0.0013974074074074075, "loss": 1.3602, "step": 16270 }, { "epoch": 60.3, "grad_norm": 0.6971426010131836, "learning_rate": 0.0013970370370370371, "loss": 1.3834, "step": 16280 }, { "epoch": 60.33, "grad_norm": 0.6854264140129089, "learning_rate": 0.0013966666666666667, "loss": 1.3635, "step": 16290 }, { "epoch": 60.37, "grad_norm": 0.7690728306770325, "learning_rate": 0.0013962962962962963, "loss": 1.3863, "step": 16300 }, { "epoch": 60.41, "grad_norm": 0.7954300045967102, "learning_rate": 0.0013959259259259261, "loss": 1.4147, "step": 16310 }, { "epoch": 60.44, "grad_norm": 0.7556511163711548, "learning_rate": 0.0013955555555555557, "loss": 1.4286, "step": 16320 }, { "epoch": 60.48, "grad_norm": 0.6833049058914185, "learning_rate": 0.001395185185185185, "loss": 1.4271, "step": 16330 }, { "epoch": 60.52, "grad_norm": 0.7242010831832886, "learning_rate": 0.0013948148148148147, "loss": 1.4355, "step": 16340 }, { "epoch": 60.56, "grad_norm": 0.7006980776786804, "learning_rate": 0.0013944444444444445, "loss": 1.4785, "step": 16350 }, { "epoch": 60.59, "grad_norm": 0.7061491012573242, "learning_rate": 0.001394074074074074, "loss": 1.4504, "step": 16360 }, { "epoch": 60.63, "grad_norm": 0.710753858089447, "learning_rate": 0.0013937037037037036, "loss": 1.4822, "step": 16370 }, { "epoch": 60.67, "grad_norm": 0.7644755840301514, "learning_rate": 0.0013933333333333334, "loss": 1.457, "step": 16380 }, { "epoch": 60.7, "grad_norm": 0.7168658971786499, "learning_rate": 0.001392962962962963, "loss": 1.5014, "step": 16390 }, { "epoch": 60.74, "grad_norm": 0.7056949138641357, "learning_rate": 0.0013925925925925926, "loss": 1.5104, "step": 16400 }, { "epoch": 60.78, "grad_norm": 0.6816049814224243, "learning_rate": 0.0013922222222222222, "loss": 1.485, "step": 16410 }, { "epoch": 60.81, "grad_norm": 0.7482391595840454, "learning_rate": 0.001391851851851852, "loss": 1.4858, "step": 16420 }, { "epoch": 60.85, "grad_norm": 0.7468370795249939, "learning_rate": 0.0013914814814814816, "loss": 1.4805, "step": 16430 }, { "epoch": 60.89, "grad_norm": 0.7351725697517395, "learning_rate": 0.0013911111111111112, "loss": 1.521, "step": 16440 }, { "epoch": 60.93, "grad_norm": 0.7151160836219788, "learning_rate": 0.0013907407407407408, "loss": 1.4889, "step": 16450 }, { "epoch": 60.96, "grad_norm": 0.7012620568275452, "learning_rate": 0.0013903703703703706, "loss": 1.5615, "step": 16460 }, { "epoch": 61.0, "grad_norm": 1.633631944656372, "learning_rate": 0.00139, "loss": 1.5143, "step": 16470 }, { "epoch": 61.04, "grad_norm": 0.7052876353263855, "learning_rate": 0.0013896296296296295, "loss": 1.2774, "step": 16480 }, { "epoch": 61.07, "grad_norm": 0.6705044507980347, "learning_rate": 0.0013892592592592591, "loss": 1.2894, "step": 16490 }, { "epoch": 61.11, "grad_norm": 0.746444046497345, "learning_rate": 0.001388888888888889, "loss": 1.274, "step": 16500 }, { "epoch": 61.15, "grad_norm": 0.6993348002433777, "learning_rate": 0.0013885185185185185, "loss": 1.3287, "step": 16510 }, { "epoch": 61.19, "grad_norm": 0.7455359697341919, "learning_rate": 0.0013881481481481481, "loss": 1.327, "step": 16520 }, { "epoch": 61.22, "grad_norm": 0.6975919604301453, "learning_rate": 0.0013877777777777777, "loss": 1.3357, "step": 16530 }, { "epoch": 61.26, "grad_norm": 0.7186405062675476, "learning_rate": 0.0013874074074074075, "loss": 1.3618, "step": 16540 }, { "epoch": 61.3, "grad_norm": 0.7653625011444092, "learning_rate": 0.001387037037037037, "loss": 1.3658, "step": 16550 }, { "epoch": 61.33, "grad_norm": 0.7147905230522156, "learning_rate": 0.0013866666666666667, "loss": 1.3879, "step": 16560 }, { "epoch": 61.37, "grad_norm": 0.760710597038269, "learning_rate": 0.0013862962962962965, "loss": 1.3938, "step": 16570 }, { "epoch": 61.41, "grad_norm": 0.6825000047683716, "learning_rate": 0.001385925925925926, "loss": 1.4133, "step": 16580 }, { "epoch": 61.44, "grad_norm": 0.7693556547164917, "learning_rate": 0.0013855555555555557, "loss": 1.4312, "step": 16590 }, { "epoch": 61.48, "grad_norm": 0.7311041355133057, "learning_rate": 0.001385185185185185, "loss": 1.4205, "step": 16600 }, { "epoch": 61.52, "grad_norm": 0.7689315676689148, "learning_rate": 0.0013848148148148148, "loss": 1.4454, "step": 16610 }, { "epoch": 61.56, "grad_norm": 0.7419061064720154, "learning_rate": 0.0013844444444444444, "loss": 1.4293, "step": 16620 }, { "epoch": 61.59, "grad_norm": 0.6940130591392517, "learning_rate": 0.001384074074074074, "loss": 1.4663, "step": 16630 }, { "epoch": 61.63, "grad_norm": 0.721855640411377, "learning_rate": 0.0013837037037037036, "loss": 1.4526, "step": 16640 }, { "epoch": 61.67, "grad_norm": 0.6498404741287231, "learning_rate": 0.0013833333333333334, "loss": 1.4941, "step": 16650 }, { "epoch": 61.7, "grad_norm": 0.7447950839996338, "learning_rate": 0.001382962962962963, "loss": 1.484, "step": 16660 }, { "epoch": 61.74, "grad_norm": 0.6579961776733398, "learning_rate": 0.0013825925925925926, "loss": 1.4905, "step": 16670 }, { "epoch": 61.78, "grad_norm": 0.70018070936203, "learning_rate": 0.0013822222222222222, "loss": 1.4693, "step": 16680 }, { "epoch": 61.81, "grad_norm": 0.7552095055580139, "learning_rate": 0.001381851851851852, "loss": 1.5029, "step": 16690 }, { "epoch": 61.85, "grad_norm": 0.7254689335823059, "learning_rate": 0.0013814814814814816, "loss": 1.4784, "step": 16700 }, { "epoch": 61.89, "grad_norm": 0.6803253889083862, "learning_rate": 0.0013811111111111112, "loss": 1.4925, "step": 16710 }, { "epoch": 61.93, "grad_norm": 0.6918293237686157, "learning_rate": 0.0013807407407407408, "loss": 1.5135, "step": 16720 }, { "epoch": 61.96, "grad_norm": 0.7400352358818054, "learning_rate": 0.0013803703703703706, "loss": 1.4999, "step": 16730 }, { "epoch": 62.0, "grad_norm": 1.246131181716919, "learning_rate": 0.00138, "loss": 1.5048, "step": 16740 }, { "epoch": 62.04, "grad_norm": 0.7409060597419739, "learning_rate": 0.0013796296296296295, "loss": 1.2478, "step": 16750 }, { "epoch": 62.07, "grad_norm": 0.7132163047790527, "learning_rate": 0.0013792592592592591, "loss": 1.2838, "step": 16760 }, { "epoch": 62.11, "grad_norm": 0.7276562452316284, "learning_rate": 0.001378888888888889, "loss": 1.2673, "step": 16770 }, { "epoch": 62.15, "grad_norm": 0.7163761854171753, "learning_rate": 0.0013785185185185185, "loss": 1.3045, "step": 16780 }, { "epoch": 62.19, "grad_norm": 0.7683485746383667, "learning_rate": 0.001378148148148148, "loss": 1.3284, "step": 16790 }, { "epoch": 62.22, "grad_norm": 0.678407609462738, "learning_rate": 0.001377777777777778, "loss": 1.363, "step": 16800 }, { "epoch": 62.26, "grad_norm": 0.652581512928009, "learning_rate": 0.0013774074074074075, "loss": 1.3628, "step": 16810 }, { "epoch": 62.3, "grad_norm": 0.7575969696044922, "learning_rate": 0.001377037037037037, "loss": 1.3731, "step": 16820 }, { "epoch": 62.33, "grad_norm": 0.7279390096664429, "learning_rate": 0.0013766666666666667, "loss": 1.3444, "step": 16830 }, { "epoch": 62.37, "grad_norm": 0.7434338331222534, "learning_rate": 0.0013762962962962965, "loss": 1.3699, "step": 16840 }, { "epoch": 62.41, "grad_norm": 0.7671130895614624, "learning_rate": 0.001375925925925926, "loss": 1.3842, "step": 16850 }, { "epoch": 62.44, "grad_norm": 0.775354266166687, "learning_rate": 0.0013755555555555556, "loss": 1.4063, "step": 16860 }, { "epoch": 62.48, "grad_norm": 0.7300680875778198, "learning_rate": 0.0013751851851851852, "loss": 1.4227, "step": 16870 }, { "epoch": 62.52, "grad_norm": 0.6965778470039368, "learning_rate": 0.0013748148148148148, "loss": 1.4352, "step": 16880 }, { "epoch": 62.56, "grad_norm": 0.7088303565979004, "learning_rate": 0.0013744444444444444, "loss": 1.4484, "step": 16890 }, { "epoch": 62.59, "grad_norm": 0.6896482706069946, "learning_rate": 0.001374074074074074, "loss": 1.46, "step": 16900 }, { "epoch": 62.63, "grad_norm": 0.7502049803733826, "learning_rate": 0.0013737037037037036, "loss": 1.4443, "step": 16910 }, { "epoch": 62.67, "grad_norm": 0.7610279321670532, "learning_rate": 0.0013733333333333334, "loss": 1.4562, "step": 16920 }, { "epoch": 62.7, "grad_norm": 0.7035161852836609, "learning_rate": 0.001372962962962963, "loss": 1.4789, "step": 16930 }, { "epoch": 62.74, "grad_norm": 0.635873556137085, "learning_rate": 0.0013725925925925926, "loss": 1.4699, "step": 16940 }, { "epoch": 62.78, "grad_norm": 0.7646804451942444, "learning_rate": 0.0013722222222222222, "loss": 1.4785, "step": 16950 }, { "epoch": 62.81, "grad_norm": 0.6760735511779785, "learning_rate": 0.001371851851851852, "loss": 1.4836, "step": 16960 }, { "epoch": 62.85, "grad_norm": 0.7437437772750854, "learning_rate": 0.0013714814814814816, "loss": 1.4912, "step": 16970 }, { "epoch": 62.89, "grad_norm": 0.7423031330108643, "learning_rate": 0.0013711111111111111, "loss": 1.4699, "step": 16980 }, { "epoch": 62.93, "grad_norm": 0.7733505368232727, "learning_rate": 0.001370740740740741, "loss": 1.5027, "step": 16990 }, { "epoch": 62.96, "grad_norm": 0.7638499140739441, "learning_rate": 0.0013703703703703705, "loss": 1.4934, "step": 17000 }, { "epoch": 63.0, "grad_norm": 1.214098334312439, "learning_rate": 0.0013700000000000001, "loss": 1.5209, "step": 17010 }, { "epoch": 63.04, "grad_norm": 0.7190075516700745, "learning_rate": 0.0013696296296296295, "loss": 1.2687, "step": 17020 }, { "epoch": 63.07, "grad_norm": 0.7111839652061462, "learning_rate": 0.0013692592592592593, "loss": 1.261, "step": 17030 }, { "epoch": 63.11, "grad_norm": 0.7525911331176758, "learning_rate": 0.0013688888888888889, "loss": 1.2793, "step": 17040 }, { "epoch": 63.15, "grad_norm": 0.720476508140564, "learning_rate": 0.0013685185185185185, "loss": 1.2787, "step": 17050 }, { "epoch": 63.19, "grad_norm": 0.7493318915367126, "learning_rate": 0.001368148148148148, "loss": 1.3218, "step": 17060 }, { "epoch": 63.22, "grad_norm": 0.7084575891494751, "learning_rate": 0.0013677777777777779, "loss": 1.3391, "step": 17070 }, { "epoch": 63.26, "grad_norm": 0.8328226208686829, "learning_rate": 0.0013674074074074075, "loss": 1.3533, "step": 17080 }, { "epoch": 63.3, "grad_norm": 0.7640653848648071, "learning_rate": 0.001367037037037037, "loss": 1.3232, "step": 17090 }, { "epoch": 63.33, "grad_norm": 0.773227334022522, "learning_rate": 0.0013666666666666666, "loss": 1.381, "step": 17100 }, { "epoch": 63.37, "grad_norm": 0.6970317363739014, "learning_rate": 0.0013662962962962964, "loss": 1.3851, "step": 17110 }, { "epoch": 63.41, "grad_norm": 0.7326919436454773, "learning_rate": 0.001365925925925926, "loss": 1.3854, "step": 17120 }, { "epoch": 63.44, "grad_norm": 0.8480170965194702, "learning_rate": 0.0013655555555555556, "loss": 1.4173, "step": 17130 }, { "epoch": 63.48, "grad_norm": 0.7686656713485718, "learning_rate": 0.0013651851851851852, "loss": 1.3834, "step": 17140 }, { "epoch": 63.52, "grad_norm": 0.705543577671051, "learning_rate": 0.001364814814814815, "loss": 1.4647, "step": 17150 }, { "epoch": 63.56, "grad_norm": 0.7954389452934265, "learning_rate": 0.0013644444444444444, "loss": 1.4571, "step": 17160 }, { "epoch": 63.59, "grad_norm": 0.7180191874504089, "learning_rate": 0.001364074074074074, "loss": 1.4375, "step": 17170 }, { "epoch": 63.63, "grad_norm": 0.6709361672401428, "learning_rate": 0.0013637037037037036, "loss": 1.4573, "step": 17180 }, { "epoch": 63.67, "grad_norm": 0.7622054815292358, "learning_rate": 0.0013633333333333334, "loss": 1.4532, "step": 17190 }, { "epoch": 63.7, "grad_norm": 0.7002084851264954, "learning_rate": 0.001362962962962963, "loss": 1.4217, "step": 17200 }, { "epoch": 63.74, "grad_norm": 0.7226837277412415, "learning_rate": 0.0013625925925925925, "loss": 1.4612, "step": 17210 }, { "epoch": 63.78, "grad_norm": 0.6710451245307922, "learning_rate": 0.0013622222222222223, "loss": 1.491, "step": 17220 }, { "epoch": 63.81, "grad_norm": 0.8155320286750793, "learning_rate": 0.001361851851851852, "loss": 1.4587, "step": 17230 }, { "epoch": 63.85, "grad_norm": 0.73924320936203, "learning_rate": 0.0013614814814814815, "loss": 1.4874, "step": 17240 }, { "epoch": 63.89, "grad_norm": 0.7509787082672119, "learning_rate": 0.0013611111111111111, "loss": 1.4736, "step": 17250 }, { "epoch": 63.93, "grad_norm": 0.7284494042396545, "learning_rate": 0.001360740740740741, "loss": 1.4753, "step": 17260 }, { "epoch": 63.96, "grad_norm": 0.6677215695381165, "learning_rate": 0.0013603703703703705, "loss": 1.5084, "step": 17270 }, { "epoch": 64.0, "grad_norm": 1.1213937997817993, "learning_rate": 0.00136, "loss": 1.5176, "step": 17280 }, { "epoch": 64.04, "grad_norm": 0.7577294111251831, "learning_rate": 0.0013596296296296295, "loss": 1.2681, "step": 17290 }, { "epoch": 64.07, "grad_norm": 0.7758843898773193, "learning_rate": 0.0013592592592592593, "loss": 1.2588, "step": 17300 }, { "epoch": 64.11, "grad_norm": 0.71478271484375, "learning_rate": 0.0013588888888888889, "loss": 1.2865, "step": 17310 }, { "epoch": 64.15, "grad_norm": 0.7606163024902344, "learning_rate": 0.0013585185185185185, "loss": 1.302, "step": 17320 }, { "epoch": 64.19, "grad_norm": 0.7798721194267273, "learning_rate": 0.001358148148148148, "loss": 1.3156, "step": 17330 }, { "epoch": 64.22, "grad_norm": 0.756719708442688, "learning_rate": 0.0013577777777777778, "loss": 1.3245, "step": 17340 }, { "epoch": 64.26, "grad_norm": 0.7324285507202148, "learning_rate": 0.0013574074074074074, "loss": 1.3406, "step": 17350 }, { "epoch": 64.3, "grad_norm": 0.7444923520088196, "learning_rate": 0.001357037037037037, "loss": 1.3676, "step": 17360 }, { "epoch": 64.33, "grad_norm": 0.7438486814498901, "learning_rate": 0.0013566666666666666, "loss": 1.3631, "step": 17370 }, { "epoch": 64.37, "grad_norm": 0.7301568984985352, "learning_rate": 0.0013562962962962964, "loss": 1.3979, "step": 17380 }, { "epoch": 64.41, "grad_norm": 0.7605223655700684, "learning_rate": 0.001355925925925926, "loss": 1.3797, "step": 17390 }, { "epoch": 64.44, "grad_norm": 0.7399788498878479, "learning_rate": 0.0013555555555555556, "loss": 1.3976, "step": 17400 }, { "epoch": 64.48, "grad_norm": 0.7360973954200745, "learning_rate": 0.0013551851851851854, "loss": 1.4176, "step": 17410 }, { "epoch": 64.52, "grad_norm": 0.7445151805877686, "learning_rate": 0.001354814814814815, "loss": 1.4123, "step": 17420 }, { "epoch": 64.56, "grad_norm": 0.7263211011886597, "learning_rate": 0.0013544444444444444, "loss": 1.403, "step": 17430 }, { "epoch": 64.59, "grad_norm": 0.7512028813362122, "learning_rate": 0.001354074074074074, "loss": 1.4177, "step": 17440 }, { "epoch": 64.63, "grad_norm": 0.7566067576408386, "learning_rate": 0.0013537037037037038, "loss": 1.4249, "step": 17450 }, { "epoch": 64.67, "grad_norm": 0.7968330383300781, "learning_rate": 0.0013533333333333333, "loss": 1.4237, "step": 17460 }, { "epoch": 64.7, "grad_norm": 0.7006673216819763, "learning_rate": 0.001352962962962963, "loss": 1.4591, "step": 17470 }, { "epoch": 64.74, "grad_norm": 0.7665920257568359, "learning_rate": 0.0013525925925925925, "loss": 1.4532, "step": 17480 }, { "epoch": 64.78, "grad_norm": 0.7508714199066162, "learning_rate": 0.0013522222222222223, "loss": 1.474, "step": 17490 }, { "epoch": 64.81, "grad_norm": 0.7315797209739685, "learning_rate": 0.001351851851851852, "loss": 1.439, "step": 17500 }, { "epoch": 64.85, "grad_norm": 0.8568592071533203, "learning_rate": 0.0013514814814814815, "loss": 1.4562, "step": 17510 }, { "epoch": 64.89, "grad_norm": 0.7268970012664795, "learning_rate": 0.001351111111111111, "loss": 1.5033, "step": 17520 }, { "epoch": 64.93, "grad_norm": 0.7254437804222107, "learning_rate": 0.001350740740740741, "loss": 1.4433, "step": 17530 }, { "epoch": 64.96, "grad_norm": 0.7384520769119263, "learning_rate": 0.0013503703703703705, "loss": 1.4822, "step": 17540 }, { "epoch": 65.0, "grad_norm": 1.2170757055282593, "learning_rate": 0.00135, "loss": 1.4843, "step": 17550 }, { "epoch": 65.04, "grad_norm": 0.715018630027771, "learning_rate": 0.0013496296296296297, "loss": 1.2211, "step": 17560 }, { "epoch": 65.07, "grad_norm": 0.8384785652160645, "learning_rate": 0.0013492592592592592, "loss": 1.2132, "step": 17570 }, { "epoch": 65.11, "grad_norm": 0.6869712471961975, "learning_rate": 0.0013488888888888888, "loss": 1.2573, "step": 17580 }, { "epoch": 65.15, "grad_norm": 0.7423787117004395, "learning_rate": 0.0013485185185185184, "loss": 1.2947, "step": 17590 }, { "epoch": 65.19, "grad_norm": 0.7109069228172302, "learning_rate": 0.0013481481481481482, "loss": 1.3025, "step": 17600 }, { "epoch": 65.22, "grad_norm": 0.7798488736152649, "learning_rate": 0.0013477777777777778, "loss": 1.3109, "step": 17610 }, { "epoch": 65.26, "grad_norm": 0.7772756218910217, "learning_rate": 0.0013474074074074074, "loss": 1.3323, "step": 17620 }, { "epoch": 65.3, "grad_norm": 0.7582665681838989, "learning_rate": 0.001347037037037037, "loss": 1.3301, "step": 17630 }, { "epoch": 65.33, "grad_norm": 0.7635788917541504, "learning_rate": 0.0013466666666666668, "loss": 1.3534, "step": 17640 }, { "epoch": 65.37, "grad_norm": 0.7907354235649109, "learning_rate": 0.0013462962962962964, "loss": 1.3917, "step": 17650 }, { "epoch": 65.41, "grad_norm": 0.7538552284240723, "learning_rate": 0.001345925925925926, "loss": 1.3934, "step": 17660 }, { "epoch": 65.44, "grad_norm": 0.7302442193031311, "learning_rate": 0.0013455555555555556, "loss": 1.3849, "step": 17670 }, { "epoch": 65.48, "grad_norm": 0.7725008130073547, "learning_rate": 0.0013451851851851854, "loss": 1.3807, "step": 17680 }, { "epoch": 65.52, "grad_norm": 0.7958641648292542, "learning_rate": 0.001344814814814815, "loss": 1.4248, "step": 17690 }, { "epoch": 65.56, "grad_norm": 0.770406186580658, "learning_rate": 0.0013444444444444445, "loss": 1.4143, "step": 17700 }, { "epoch": 65.59, "grad_norm": 0.6958473920822144, "learning_rate": 0.001344074074074074, "loss": 1.4358, "step": 17710 }, { "epoch": 65.63, "grad_norm": 0.7644444108009338, "learning_rate": 0.0013437037037037037, "loss": 1.4147, "step": 17720 }, { "epoch": 65.67, "grad_norm": 0.7265663743019104, "learning_rate": 0.0013433333333333333, "loss": 1.4309, "step": 17730 }, { "epoch": 65.7, "grad_norm": 0.7758519649505615, "learning_rate": 0.001342962962962963, "loss": 1.4262, "step": 17740 }, { "epoch": 65.74, "grad_norm": 0.7135465741157532, "learning_rate": 0.0013425925925925925, "loss": 1.4435, "step": 17750 }, { "epoch": 65.78, "grad_norm": 0.7321381568908691, "learning_rate": 0.0013422222222222223, "loss": 1.4586, "step": 17760 }, { "epoch": 65.81, "grad_norm": 0.7740125060081482, "learning_rate": 0.0013418518518518519, "loss": 1.4454, "step": 17770 }, { "epoch": 65.85, "grad_norm": 0.714545488357544, "learning_rate": 0.0013414814814814815, "loss": 1.4606, "step": 17780 }, { "epoch": 65.89, "grad_norm": 0.7659265995025635, "learning_rate": 0.001341111111111111, "loss": 1.4924, "step": 17790 }, { "epoch": 65.93, "grad_norm": 0.6884202361106873, "learning_rate": 0.0013407407407407409, "loss": 1.502, "step": 17800 }, { "epoch": 65.96, "grad_norm": 0.8415507078170776, "learning_rate": 0.0013403703703703705, "loss": 1.4725, "step": 17810 }, { "epoch": 66.0, "grad_norm": 1.2466375827789307, "learning_rate": 0.00134, "loss": 1.4451, "step": 17820 }, { "epoch": 66.04, "grad_norm": 0.7439112663269043, "learning_rate": 0.0013396296296296299, "loss": 1.256, "step": 17830 }, { "epoch": 66.07, "grad_norm": 0.8170116543769836, "learning_rate": 0.0013392592592592592, "loss": 1.2327, "step": 17840 }, { "epoch": 66.11, "grad_norm": 0.707555890083313, "learning_rate": 0.0013388888888888888, "loss": 1.2809, "step": 17850 }, { "epoch": 66.15, "grad_norm": 0.6814079284667969, "learning_rate": 0.0013385185185185184, "loss": 1.276, "step": 17860 }, { "epoch": 66.19, "grad_norm": 0.7349382042884827, "learning_rate": 0.0013381481481481482, "loss": 1.2884, "step": 17870 }, { "epoch": 66.22, "grad_norm": 0.7153812646865845, "learning_rate": 0.0013377777777777778, "loss": 1.322, "step": 17880 }, { "epoch": 66.26, "grad_norm": 0.7231837511062622, "learning_rate": 0.0013374074074074074, "loss": 1.299, "step": 17890 }, { "epoch": 66.3, "grad_norm": 0.7510600686073303, "learning_rate": 0.001337037037037037, "loss": 1.3416, "step": 17900 }, { "epoch": 66.33, "grad_norm": 0.7508620023727417, "learning_rate": 0.0013366666666666668, "loss": 1.3809, "step": 17910 }, { "epoch": 66.37, "grad_norm": 0.7234875559806824, "learning_rate": 0.0013362962962962964, "loss": 1.3698, "step": 17920 }, { "epoch": 66.41, "grad_norm": 0.731718897819519, "learning_rate": 0.001335925925925926, "loss": 1.3692, "step": 17930 }, { "epoch": 66.44, "grad_norm": 0.7301697731018066, "learning_rate": 0.0013355555555555555, "loss": 1.3878, "step": 17940 }, { "epoch": 66.48, "grad_norm": 0.7749512195587158, "learning_rate": 0.0013351851851851853, "loss": 1.3832, "step": 17950 }, { "epoch": 66.52, "grad_norm": 0.8031456470489502, "learning_rate": 0.001334814814814815, "loss": 1.4165, "step": 17960 }, { "epoch": 66.56, "grad_norm": 0.7439578771591187, "learning_rate": 0.0013344444444444445, "loss": 1.3794, "step": 17970 }, { "epoch": 66.59, "grad_norm": 0.7619519233703613, "learning_rate": 0.001334074074074074, "loss": 1.4354, "step": 17980 }, { "epoch": 66.63, "grad_norm": 0.7329776883125305, "learning_rate": 0.0013337037037037037, "loss": 1.4155, "step": 17990 }, { "epoch": 66.67, "grad_norm": 0.7692404389381409, "learning_rate": 0.0013333333333333333, "loss": 1.4214, "step": 18000 }, { "epoch": 66.7, "grad_norm": 0.7139493227005005, "learning_rate": 0.0013329629629629629, "loss": 1.4516, "step": 18010 }, { "epoch": 66.74, "grad_norm": 0.7538207173347473, "learning_rate": 0.0013325925925925927, "loss": 1.432, "step": 18020 }, { "epoch": 66.78, "grad_norm": 0.7339174747467041, "learning_rate": 0.0013322222222222223, "loss": 1.4441, "step": 18030 }, { "epoch": 66.81, "grad_norm": 0.7614271640777588, "learning_rate": 0.0013318518518518519, "loss": 1.4407, "step": 18040 }, { "epoch": 66.85, "grad_norm": 0.7265300750732422, "learning_rate": 0.0013314814814814814, "loss": 1.4497, "step": 18050 }, { "epoch": 66.89, "grad_norm": 0.7015870213508606, "learning_rate": 0.0013311111111111113, "loss": 1.463, "step": 18060 }, { "epoch": 66.93, "grad_norm": 0.736568033695221, "learning_rate": 0.0013307407407407408, "loss": 1.4459, "step": 18070 }, { "epoch": 66.96, "grad_norm": 0.8021724820137024, "learning_rate": 0.0013303703703703704, "loss": 1.4521, "step": 18080 }, { "epoch": 67.0, "grad_norm": 1.27347993850708, "learning_rate": 0.00133, "loss": 1.4864, "step": 18090 }, { "epoch": 67.04, "grad_norm": 0.7476965188980103, "learning_rate": 0.0013296296296296298, "loss": 1.2337, "step": 18100 }, { "epoch": 67.07, "grad_norm": 0.7044015526771545, "learning_rate": 0.0013292592592592594, "loss": 1.2397, "step": 18110 }, { "epoch": 67.11, "grad_norm": 0.7922229170799255, "learning_rate": 0.0013288888888888888, "loss": 1.256, "step": 18120 }, { "epoch": 67.15, "grad_norm": 0.7829376459121704, "learning_rate": 0.0013285185185185184, "loss": 1.2582, "step": 18130 }, { "epoch": 67.19, "grad_norm": 0.7793470621109009, "learning_rate": 0.0013281481481481482, "loss": 1.2719, "step": 18140 }, { "epoch": 67.22, "grad_norm": 0.727636456489563, "learning_rate": 0.0013277777777777778, "loss": 1.3313, "step": 18150 }, { "epoch": 67.26, "grad_norm": 0.7546897530555725, "learning_rate": 0.0013274074074074074, "loss": 1.312, "step": 18160 }, { "epoch": 67.3, "grad_norm": 0.6836079955101013, "learning_rate": 0.001327037037037037, "loss": 1.3276, "step": 18170 }, { "epoch": 67.33, "grad_norm": 0.802686333656311, "learning_rate": 0.0013266666666666667, "loss": 1.3376, "step": 18180 }, { "epoch": 67.37, "grad_norm": 0.8498095870018005, "learning_rate": 0.0013262962962962963, "loss": 1.3466, "step": 18190 }, { "epoch": 67.41, "grad_norm": 0.7194114327430725, "learning_rate": 0.001325925925925926, "loss": 1.3611, "step": 18200 }, { "epoch": 67.44, "grad_norm": 0.7203561663627625, "learning_rate": 0.0013255555555555557, "loss": 1.3319, "step": 18210 }, { "epoch": 67.48, "grad_norm": 0.7772254943847656, "learning_rate": 0.0013251851851851853, "loss": 1.4026, "step": 18220 }, { "epoch": 67.52, "grad_norm": 0.7512720823287964, "learning_rate": 0.001324814814814815, "loss": 1.3907, "step": 18230 }, { "epoch": 67.56, "grad_norm": 0.7536607980728149, "learning_rate": 0.0013244444444444445, "loss": 1.4007, "step": 18240 }, { "epoch": 67.59, "grad_norm": 0.7336398363113403, "learning_rate": 0.0013240740740740743, "loss": 1.3965, "step": 18250 }, { "epoch": 67.63, "grad_norm": 0.7424538135528564, "learning_rate": 0.0013237037037037037, "loss": 1.4303, "step": 18260 }, { "epoch": 67.67, "grad_norm": 0.7541590332984924, "learning_rate": 0.0013233333333333333, "loss": 1.419, "step": 18270 }, { "epoch": 67.7, "grad_norm": 0.7302483320236206, "learning_rate": 0.0013229629629629629, "loss": 1.4236, "step": 18280 }, { "epoch": 67.74, "grad_norm": 0.7059903740882874, "learning_rate": 0.0013225925925925927, "loss": 1.4488, "step": 18290 }, { "epoch": 67.78, "grad_norm": 0.7474135756492615, "learning_rate": 0.0013222222222222222, "loss": 1.4424, "step": 18300 }, { "epoch": 67.81, "grad_norm": 0.7122178077697754, "learning_rate": 0.0013218518518518518, "loss": 1.4612, "step": 18310 }, { "epoch": 67.85, "grad_norm": 0.7674033641815186, "learning_rate": 0.0013214814814814814, "loss": 1.4424, "step": 18320 }, { "epoch": 67.89, "grad_norm": 0.7650808095932007, "learning_rate": 0.0013211111111111112, "loss": 1.4694, "step": 18330 }, { "epoch": 67.93, "grad_norm": 0.7181142568588257, "learning_rate": 0.0013207407407407408, "loss": 1.4476, "step": 18340 }, { "epoch": 67.96, "grad_norm": 0.7057814598083496, "learning_rate": 0.0013203703703703704, "loss": 1.4677, "step": 18350 }, { "epoch": 68.0, "grad_norm": 1.2915374040603638, "learning_rate": 0.00132, "loss": 1.4835, "step": 18360 }, { "epoch": 68.04, "grad_norm": 0.709320068359375, "learning_rate": 0.0013196296296296298, "loss": 1.1855, "step": 18370 }, { "epoch": 68.07, "grad_norm": 0.7228713035583496, "learning_rate": 0.0013192592592592594, "loss": 1.2257, "step": 18380 }, { "epoch": 68.11, "grad_norm": 0.7062861919403076, "learning_rate": 0.001318888888888889, "loss": 1.2602, "step": 18390 }, { "epoch": 68.15, "grad_norm": 0.7732377052307129, "learning_rate": 0.0013185185185185183, "loss": 1.2496, "step": 18400 }, { "epoch": 68.19, "grad_norm": 0.7629472017288208, "learning_rate": 0.0013181481481481482, "loss": 1.2419, "step": 18410 }, { "epoch": 68.22, "grad_norm": 0.7411864399909973, "learning_rate": 0.0013177777777777777, "loss": 1.2923, "step": 18420 }, { "epoch": 68.26, "grad_norm": 0.7572233080863953, "learning_rate": 0.0013174074074074073, "loss": 1.308, "step": 18430 }, { "epoch": 68.3, "grad_norm": 0.7793846726417542, "learning_rate": 0.0013170370370370371, "loss": 1.3047, "step": 18440 }, { "epoch": 68.33, "grad_norm": 0.8033546805381775, "learning_rate": 0.0013166666666666667, "loss": 1.3386, "step": 18450 }, { "epoch": 68.37, "grad_norm": 0.7773953676223755, "learning_rate": 0.0013162962962962963, "loss": 1.3362, "step": 18460 }, { "epoch": 68.41, "grad_norm": 0.7706006169319153, "learning_rate": 0.001315925925925926, "loss": 1.3677, "step": 18470 }, { "epoch": 68.44, "grad_norm": 0.751693844795227, "learning_rate": 0.0013155555555555557, "loss": 1.3489, "step": 18480 }, { "epoch": 68.48, "grad_norm": 0.7412254810333252, "learning_rate": 0.0013151851851851853, "loss": 1.3638, "step": 18490 }, { "epoch": 68.52, "grad_norm": 0.704195499420166, "learning_rate": 0.0013148148148148149, "loss": 1.3764, "step": 18500 }, { "epoch": 68.56, "grad_norm": 0.8066453337669373, "learning_rate": 0.0013144444444444445, "loss": 1.4044, "step": 18510 }, { "epoch": 68.59, "grad_norm": 0.789995014667511, "learning_rate": 0.0013140740740740743, "loss": 1.424, "step": 18520 }, { "epoch": 68.63, "grad_norm": 0.7190313339233398, "learning_rate": 0.0013137037037037036, "loss": 1.3813, "step": 18530 }, { "epoch": 68.67, "grad_norm": 0.8326443433761597, "learning_rate": 0.0013133333333333332, "loss": 1.4548, "step": 18540 }, { "epoch": 68.7, "grad_norm": 0.7189662456512451, "learning_rate": 0.0013129629629629628, "loss": 1.4213, "step": 18550 }, { "epoch": 68.74, "grad_norm": 0.7524738907814026, "learning_rate": 0.0013125925925925926, "loss": 1.4392, "step": 18560 }, { "epoch": 68.78, "grad_norm": 0.7662227749824524, "learning_rate": 0.0013122222222222222, "loss": 1.4195, "step": 18570 }, { "epoch": 68.81, "grad_norm": 0.7383051514625549, "learning_rate": 0.0013118518518518518, "loss": 1.4304, "step": 18580 }, { "epoch": 68.85, "grad_norm": 0.750335693359375, "learning_rate": 0.0013114814814814814, "loss": 1.4339, "step": 18590 }, { "epoch": 68.89, "grad_norm": 0.6686143279075623, "learning_rate": 0.0013111111111111112, "loss": 1.4295, "step": 18600 }, { "epoch": 68.93, "grad_norm": 0.7729794979095459, "learning_rate": 0.0013107407407407408, "loss": 1.468, "step": 18610 }, { "epoch": 68.96, "grad_norm": 0.7255796194076538, "learning_rate": 0.0013103703703703704, "loss": 1.4713, "step": 18620 }, { "epoch": 69.0, "grad_norm": 1.3524179458618164, "learning_rate": 0.0013100000000000002, "loss": 1.4484, "step": 18630 }, { "epoch": 69.04, "grad_norm": 0.7976313233375549, "learning_rate": 0.0013096296296296298, "loss": 1.2158, "step": 18640 }, { "epoch": 69.07, "grad_norm": 0.7665860652923584, "learning_rate": 0.0013092592592592594, "loss": 1.2247, "step": 18650 }, { "epoch": 69.11, "grad_norm": 0.7670441269874573, "learning_rate": 0.001308888888888889, "loss": 1.2384, "step": 18660 }, { "epoch": 69.15, "grad_norm": 0.8014783263206482, "learning_rate": 0.0013085185185185185, "loss": 1.2996, "step": 18670 }, { "epoch": 69.19, "grad_norm": 0.7495048642158508, "learning_rate": 0.0013081481481481481, "loss": 1.2836, "step": 18680 }, { "epoch": 69.22, "grad_norm": 0.7853214740753174, "learning_rate": 0.0013077777777777777, "loss": 1.2771, "step": 18690 }, { "epoch": 69.26, "grad_norm": 0.69195157289505, "learning_rate": 0.0013074074074074073, "loss": 1.2946, "step": 18700 }, { "epoch": 69.3, "grad_norm": 0.7509171962738037, "learning_rate": 0.001307037037037037, "loss": 1.3141, "step": 18710 }, { "epoch": 69.33, "grad_norm": 0.7930737733840942, "learning_rate": 0.0013066666666666667, "loss": 1.2937, "step": 18720 }, { "epoch": 69.37, "grad_norm": 0.7614037990570068, "learning_rate": 0.0013062962962962963, "loss": 1.3544, "step": 18730 }, { "epoch": 69.41, "grad_norm": 0.7565775513648987, "learning_rate": 0.0013059259259259259, "loss": 1.3547, "step": 18740 }, { "epoch": 69.44, "grad_norm": 0.7282271385192871, "learning_rate": 0.0013055555555555557, "loss": 1.3461, "step": 18750 }, { "epoch": 69.48, "grad_norm": 0.717829704284668, "learning_rate": 0.0013051851851851853, "loss": 1.3662, "step": 18760 }, { "epoch": 69.52, "grad_norm": 0.697563111782074, "learning_rate": 0.0013048148148148149, "loss": 1.3676, "step": 18770 }, { "epoch": 69.56, "grad_norm": 0.8086504936218262, "learning_rate": 0.0013044444444444444, "loss": 1.384, "step": 18780 }, { "epoch": 69.59, "grad_norm": 0.754641056060791, "learning_rate": 0.0013040740740740742, "loss": 1.4019, "step": 18790 }, { "epoch": 69.63, "grad_norm": 0.7806912660598755, "learning_rate": 0.0013037037037037038, "loss": 1.4061, "step": 18800 }, { "epoch": 69.67, "grad_norm": 0.7380501627922058, "learning_rate": 0.0013033333333333332, "loss": 1.4203, "step": 18810 }, { "epoch": 69.7, "grad_norm": 0.8265412449836731, "learning_rate": 0.0013029629629629628, "loss": 1.3984, "step": 18820 }, { "epoch": 69.74, "grad_norm": 0.7108408808708191, "learning_rate": 0.0013025925925925926, "loss": 1.4126, "step": 18830 }, { "epoch": 69.78, "grad_norm": 0.7676126956939697, "learning_rate": 0.0013022222222222222, "loss": 1.3999, "step": 18840 }, { "epoch": 69.81, "grad_norm": 0.8513116836547852, "learning_rate": 0.0013018518518518518, "loss": 1.4224, "step": 18850 }, { "epoch": 69.85, "grad_norm": 0.7177796363830566, "learning_rate": 0.0013014814814814816, "loss": 1.4688, "step": 18860 }, { "epoch": 69.89, "grad_norm": 0.7571403384208679, "learning_rate": 0.0013011111111111112, "loss": 1.4465, "step": 18870 }, { "epoch": 69.93, "grad_norm": 0.7542866468429565, "learning_rate": 0.0013007407407407408, "loss": 1.4452, "step": 18880 }, { "epoch": 69.96, "grad_norm": 0.7163647413253784, "learning_rate": 0.0013003703703703704, "loss": 1.4489, "step": 18890 }, { "epoch": 70.0, "grad_norm": 1.3719452619552612, "learning_rate": 0.0013000000000000002, "loss": 1.4496, "step": 18900 }, { "epoch": 70.04, "grad_norm": 0.7573015689849854, "learning_rate": 0.0012996296296296297, "loss": 1.216, "step": 18910 }, { "epoch": 70.07, "grad_norm": 0.7383880019187927, "learning_rate": 0.0012992592592592593, "loss": 1.1958, "step": 18920 }, { "epoch": 70.11, "grad_norm": 0.7401066422462463, "learning_rate": 0.001298888888888889, "loss": 1.2191, "step": 18930 }, { "epoch": 70.15, "grad_norm": 0.7491777539253235, "learning_rate": 0.0012985185185185187, "loss": 1.257, "step": 18940 }, { "epoch": 70.19, "grad_norm": 0.7398974895477295, "learning_rate": 0.001298148148148148, "loss": 1.2388, "step": 18950 }, { "epoch": 70.22, "grad_norm": 0.7718586921691895, "learning_rate": 0.0012977777777777777, "loss": 1.3012, "step": 18960 }, { "epoch": 70.26, "grad_norm": 0.7411701083183289, "learning_rate": 0.0012974074074074073, "loss": 1.2783, "step": 18970 }, { "epoch": 70.3, "grad_norm": 0.764370858669281, "learning_rate": 0.001297037037037037, "loss": 1.2849, "step": 18980 }, { "epoch": 70.33, "grad_norm": 0.7457252740859985, "learning_rate": 0.0012966666666666667, "loss": 1.3388, "step": 18990 }, { "epoch": 70.37, "grad_norm": 0.736430287361145, "learning_rate": 0.0012962962962962963, "loss": 1.3276, "step": 19000 }, { "epoch": 70.41, "grad_norm": 0.8024626970291138, "learning_rate": 0.0012959259259259258, "loss": 1.3272, "step": 19010 }, { "epoch": 70.44, "grad_norm": 0.8038630485534668, "learning_rate": 0.0012955555555555557, "loss": 1.3592, "step": 19020 }, { "epoch": 70.48, "grad_norm": 0.7494184970855713, "learning_rate": 0.0012951851851851852, "loss": 1.3458, "step": 19030 }, { "epoch": 70.52, "grad_norm": 0.7740994095802307, "learning_rate": 0.0012948148148148148, "loss": 1.3552, "step": 19040 }, { "epoch": 70.56, "grad_norm": 0.8263460993766785, "learning_rate": 0.0012944444444444446, "loss": 1.365, "step": 19050 }, { "epoch": 70.59, "grad_norm": 0.7671233415603638, "learning_rate": 0.0012940740740740742, "loss": 1.3681, "step": 19060 }, { "epoch": 70.63, "grad_norm": 0.7569007277488708, "learning_rate": 0.0012937037037037038, "loss": 1.3722, "step": 19070 }, { "epoch": 70.67, "grad_norm": 0.748662531375885, "learning_rate": 0.0012933333333333332, "loss": 1.391, "step": 19080 }, { "epoch": 70.7, "grad_norm": 0.759195864200592, "learning_rate": 0.001292962962962963, "loss": 1.3899, "step": 19090 }, { "epoch": 70.74, "grad_norm": 0.7581557631492615, "learning_rate": 0.0012925925925925926, "loss": 1.4389, "step": 19100 }, { "epoch": 70.78, "grad_norm": 0.7946545481681824, "learning_rate": 0.0012922222222222222, "loss": 1.4117, "step": 19110 }, { "epoch": 70.81, "grad_norm": 0.7786641120910645, "learning_rate": 0.0012918518518518518, "loss": 1.4132, "step": 19120 }, { "epoch": 70.85, "grad_norm": 0.8186221718788147, "learning_rate": 0.0012914814814814816, "loss": 1.44, "step": 19130 }, { "epoch": 70.89, "grad_norm": 0.7414263486862183, "learning_rate": 0.0012911111111111111, "loss": 1.4461, "step": 19140 }, { "epoch": 70.93, "grad_norm": 0.7749980092048645, "learning_rate": 0.0012907407407407407, "loss": 1.4314, "step": 19150 }, { "epoch": 70.96, "grad_norm": 0.7037963271141052, "learning_rate": 0.0012903703703703703, "loss": 1.4367, "step": 19160 }, { "epoch": 71.0, "grad_norm": 1.8344924449920654, "learning_rate": 0.0012900000000000001, "loss": 1.4445, "step": 19170 }, { "epoch": 71.04, "grad_norm": 0.7934958338737488, "learning_rate": 0.0012896296296296297, "loss": 1.2252, "step": 19180 }, { "epoch": 71.07, "grad_norm": 0.724882185459137, "learning_rate": 0.0012892592592592593, "loss": 1.231, "step": 19190 }, { "epoch": 71.11, "grad_norm": 0.6996666789054871, "learning_rate": 0.001288888888888889, "loss": 1.2123, "step": 19200 }, { "epoch": 71.15, "grad_norm": 0.7373991012573242, "learning_rate": 0.0012885185185185187, "loss": 1.2471, "step": 19210 }, { "epoch": 71.19, "grad_norm": 0.8212770223617554, "learning_rate": 0.001288148148148148, "loss": 1.2238, "step": 19220 }, { "epoch": 71.22, "grad_norm": 0.8789666891098022, "learning_rate": 0.0012877777777777777, "loss": 1.2717, "step": 19230 }, { "epoch": 71.26, "grad_norm": 0.7911102771759033, "learning_rate": 0.0012874074074074075, "loss": 1.2665, "step": 19240 }, { "epoch": 71.3, "grad_norm": 0.7351455688476562, "learning_rate": 0.001287037037037037, "loss": 1.3006, "step": 19250 }, { "epoch": 71.33, "grad_norm": 0.7648353576660156, "learning_rate": 0.0012866666666666666, "loss": 1.3268, "step": 19260 }, { "epoch": 71.37, "grad_norm": 0.7373409271240234, "learning_rate": 0.0012862962962962962, "loss": 1.3222, "step": 19270 }, { "epoch": 71.41, "grad_norm": 0.7790569067001343, "learning_rate": 0.001285925925925926, "loss": 1.3216, "step": 19280 }, { "epoch": 71.44, "grad_norm": 0.7489818930625916, "learning_rate": 0.0012855555555555556, "loss": 1.3577, "step": 19290 }, { "epoch": 71.48, "grad_norm": 0.8069300055503845, "learning_rate": 0.0012851851851851852, "loss": 1.331, "step": 19300 }, { "epoch": 71.52, "grad_norm": 0.8932879567146301, "learning_rate": 0.0012848148148148148, "loss": 1.3861, "step": 19310 }, { "epoch": 71.56, "grad_norm": 0.803302526473999, "learning_rate": 0.0012844444444444446, "loss": 1.3809, "step": 19320 }, { "epoch": 71.59, "grad_norm": 0.8255770802497864, "learning_rate": 0.0012840740740740742, "loss": 1.3624, "step": 19330 }, { "epoch": 71.63, "grad_norm": 0.8509892225265503, "learning_rate": 0.0012837037037037038, "loss": 1.3956, "step": 19340 }, { "epoch": 71.67, "grad_norm": 0.7588201761245728, "learning_rate": 0.0012833333333333334, "loss": 1.3989, "step": 19350 }, { "epoch": 71.7, "grad_norm": 0.7393040657043457, "learning_rate": 0.001282962962962963, "loss": 1.3604, "step": 19360 }, { "epoch": 71.74, "grad_norm": 0.7771015167236328, "learning_rate": 0.0012825925925925926, "loss": 1.399, "step": 19370 }, { "epoch": 71.78, "grad_norm": 0.8176193237304688, "learning_rate": 0.0012822222222222221, "loss": 1.4179, "step": 19380 }, { "epoch": 71.81, "grad_norm": 0.8000503778457642, "learning_rate": 0.0012818518518518517, "loss": 1.425, "step": 19390 }, { "epoch": 71.85, "grad_norm": 0.7793208956718445, "learning_rate": 0.0012814814814814815, "loss": 1.4215, "step": 19400 }, { "epoch": 71.89, "grad_norm": 0.7359727025032043, "learning_rate": 0.0012811111111111111, "loss": 1.4261, "step": 19410 }, { "epoch": 71.93, "grad_norm": 0.7612905502319336, "learning_rate": 0.0012807407407407407, "loss": 1.4268, "step": 19420 }, { "epoch": 71.96, "grad_norm": 0.7255610823631287, "learning_rate": 0.0012803703703703703, "loss": 1.4153, "step": 19430 }, { "epoch": 72.0, "grad_norm": 1.6871055364608765, "learning_rate": 0.00128, "loss": 1.4457, "step": 19440 }, { "epoch": 72.04, "grad_norm": 0.7359239459037781, "learning_rate": 0.0012796296296296297, "loss": 1.1798, "step": 19450 }, { "epoch": 72.07, "grad_norm": 0.7320151925086975, "learning_rate": 0.0012792592592592593, "loss": 1.1813, "step": 19460 }, { "epoch": 72.11, "grad_norm": 0.866115927696228, "learning_rate": 0.001278888888888889, "loss": 1.2234, "step": 19470 }, { "epoch": 72.15, "grad_norm": 0.7804827690124512, "learning_rate": 0.0012785185185185187, "loss": 1.224, "step": 19480 }, { "epoch": 72.19, "grad_norm": 0.7378130555152893, "learning_rate": 0.0012781481481481483, "loss": 1.2452, "step": 19490 }, { "epoch": 72.22, "grad_norm": 0.8071990609169006, "learning_rate": 0.0012777777777777776, "loss": 1.2677, "step": 19500 }, { "epoch": 72.26, "grad_norm": 0.7027687430381775, "learning_rate": 0.0012774074074074074, "loss": 1.2874, "step": 19510 }, { "epoch": 72.3, "grad_norm": 0.8697335720062256, "learning_rate": 0.001277037037037037, "loss": 1.2834, "step": 19520 }, { "epoch": 72.33, "grad_norm": 0.7541685104370117, "learning_rate": 0.0012766666666666666, "loss": 1.3215, "step": 19530 }, { "epoch": 72.37, "grad_norm": 0.7561916708946228, "learning_rate": 0.0012762962962962962, "loss": 1.3078, "step": 19540 }, { "epoch": 72.41, "grad_norm": 0.8127706050872803, "learning_rate": 0.001275925925925926, "loss": 1.3113, "step": 19550 }, { "epoch": 72.44, "grad_norm": 0.7472105622291565, "learning_rate": 0.0012755555555555556, "loss": 1.3324, "step": 19560 }, { "epoch": 72.48, "grad_norm": 0.7457278370857239, "learning_rate": 0.0012751851851851852, "loss": 1.3743, "step": 19570 }, { "epoch": 72.52, "grad_norm": 0.8131669759750366, "learning_rate": 0.0012748148148148148, "loss": 1.3736, "step": 19580 }, { "epoch": 72.56, "grad_norm": 0.757375955581665, "learning_rate": 0.0012744444444444446, "loss": 1.3808, "step": 19590 }, { "epoch": 72.59, "grad_norm": 0.779960572719574, "learning_rate": 0.0012740740740740742, "loss": 1.4005, "step": 19600 }, { "epoch": 72.63, "grad_norm": 0.7514369487762451, "learning_rate": 0.0012737037037037038, "loss": 1.3723, "step": 19610 }, { "epoch": 72.67, "grad_norm": 0.8244257569313049, "learning_rate": 0.0012733333333333333, "loss": 1.3948, "step": 19620 }, { "epoch": 72.7, "grad_norm": 0.8477365970611572, "learning_rate": 0.0012729629629629632, "loss": 1.3758, "step": 19630 }, { "epoch": 72.74, "grad_norm": 0.7021734118461609, "learning_rate": 0.0012725925925925925, "loss": 1.3878, "step": 19640 }, { "epoch": 72.78, "grad_norm": 0.7419242858886719, "learning_rate": 0.0012722222222222221, "loss": 1.3916, "step": 19650 }, { "epoch": 72.81, "grad_norm": 0.7143440246582031, "learning_rate": 0.001271851851851852, "loss": 1.3839, "step": 19660 }, { "epoch": 72.85, "grad_norm": 0.7648002505302429, "learning_rate": 0.0012714814814814815, "loss": 1.3988, "step": 19670 }, { "epoch": 72.89, "grad_norm": 0.7516924738883972, "learning_rate": 0.001271111111111111, "loss": 1.4166, "step": 19680 }, { "epoch": 72.93, "grad_norm": 0.7532698512077332, "learning_rate": 0.0012707407407407407, "loss": 1.4171, "step": 19690 }, { "epoch": 72.96, "grad_norm": 0.7403442859649658, "learning_rate": 0.0012703703703703705, "loss": 1.4313, "step": 19700 }, { "epoch": 73.0, "grad_norm": 1.462038278579712, "learning_rate": 0.00127, "loss": 1.4292, "step": 19710 }, { "epoch": 73.04, "grad_norm": 0.7673774361610413, "learning_rate": 0.0012696296296296297, "loss": 1.1586, "step": 19720 }, { "epoch": 73.07, "grad_norm": 0.75072181224823, "learning_rate": 0.0012692592592592593, "loss": 1.222, "step": 19730 }, { "epoch": 73.11, "grad_norm": 0.7945419549942017, "learning_rate": 0.001268888888888889, "loss": 1.2124, "step": 19740 }, { "epoch": 73.15, "grad_norm": 0.7756657600402832, "learning_rate": 0.0012685185185185186, "loss": 1.212, "step": 19750 }, { "epoch": 73.19, "grad_norm": 0.760955274105072, "learning_rate": 0.0012681481481481482, "loss": 1.2436, "step": 19760 }, { "epoch": 73.22, "grad_norm": 0.7394781708717346, "learning_rate": 0.0012677777777777776, "loss": 1.2815, "step": 19770 }, { "epoch": 73.26, "grad_norm": 0.8197513818740845, "learning_rate": 0.0012674074074074074, "loss": 1.2559, "step": 19780 }, { "epoch": 73.3, "grad_norm": 0.8319255113601685, "learning_rate": 0.001267037037037037, "loss": 1.2566, "step": 19790 }, { "epoch": 73.33, "grad_norm": 0.8381143808364868, "learning_rate": 0.0012666666666666666, "loss": 1.2853, "step": 19800 }, { "epoch": 73.37, "grad_norm": 0.7372511625289917, "learning_rate": 0.0012662962962962962, "loss": 1.3118, "step": 19810 }, { "epoch": 73.41, "grad_norm": 0.7255342602729797, "learning_rate": 0.001265925925925926, "loss": 1.3363, "step": 19820 }, { "epoch": 73.44, "grad_norm": 0.7761967182159424, "learning_rate": 0.0012655555555555556, "loss": 1.3206, "step": 19830 }, { "epoch": 73.48, "grad_norm": 0.788222074508667, "learning_rate": 0.0012651851851851852, "loss": 1.3595, "step": 19840 }, { "epoch": 73.52, "grad_norm": 0.8446279168128967, "learning_rate": 0.001264814814814815, "loss": 1.34, "step": 19850 }, { "epoch": 73.56, "grad_norm": 0.7769805788993835, "learning_rate": 0.0012644444444444446, "loss": 1.3626, "step": 19860 }, { "epoch": 73.59, "grad_norm": 0.8095614314079285, "learning_rate": 0.0012640740740740741, "loss": 1.3667, "step": 19870 }, { "epoch": 73.63, "grad_norm": 0.7476008534431458, "learning_rate": 0.0012637037037037037, "loss": 1.3414, "step": 19880 }, { "epoch": 73.67, "grad_norm": 0.8041726350784302, "learning_rate": 0.0012633333333333335, "loss": 1.3698, "step": 19890 }, { "epoch": 73.7, "grad_norm": 0.7896828651428223, "learning_rate": 0.0012629629629629631, "loss": 1.3821, "step": 19900 }, { "epoch": 73.74, "grad_norm": 0.763286828994751, "learning_rate": 0.0012625925925925925, "loss": 1.3715, "step": 19910 }, { "epoch": 73.78, "grad_norm": 0.7226396203041077, "learning_rate": 0.001262222222222222, "loss": 1.3837, "step": 19920 }, { "epoch": 73.81, "grad_norm": 0.8018871545791626, "learning_rate": 0.001261851851851852, "loss": 1.4263, "step": 19930 }, { "epoch": 73.85, "grad_norm": 0.7554224729537964, "learning_rate": 0.0012614814814814815, "loss": 1.4306, "step": 19940 }, { "epoch": 73.89, "grad_norm": 0.780648946762085, "learning_rate": 0.001261111111111111, "loss": 1.4078, "step": 19950 }, { "epoch": 73.93, "grad_norm": 0.7898496389389038, "learning_rate": 0.0012607407407407407, "loss": 1.4097, "step": 19960 }, { "epoch": 73.96, "grad_norm": 0.7502182126045227, "learning_rate": 0.0012603703703703705, "loss": 1.3955, "step": 19970 }, { "epoch": 74.0, "grad_norm": 1.3677808046340942, "learning_rate": 0.00126, "loss": 1.44, "step": 19980 }, { "epoch": 74.04, "grad_norm": 0.8023880124092102, "learning_rate": 0.0012596296296296296, "loss": 1.1658, "step": 19990 }, { "epoch": 74.07, "grad_norm": 0.8229208588600159, "learning_rate": 0.0012592592592592592, "loss": 1.205, "step": 20000 }, { "epoch": 74.11, "grad_norm": 0.7845343351364136, "learning_rate": 0.001258888888888889, "loss": 1.2099, "step": 20010 }, { "epoch": 74.15, "grad_norm": 0.7770639657974243, "learning_rate": 0.0012585185185185186, "loss": 1.2291, "step": 20020 }, { "epoch": 74.19, "grad_norm": 0.7614319920539856, "learning_rate": 0.0012581481481481482, "loss": 1.2247, "step": 20030 }, { "epoch": 74.22, "grad_norm": 0.808239758014679, "learning_rate": 0.001257777777777778, "loss": 1.244, "step": 20040 }, { "epoch": 74.26, "grad_norm": 0.7549338340759277, "learning_rate": 0.0012574074074074074, "loss": 1.2564, "step": 20050 }, { "epoch": 74.3, "grad_norm": 0.8833985328674316, "learning_rate": 0.001257037037037037, "loss": 1.2776, "step": 20060 }, { "epoch": 74.33, "grad_norm": 0.7880859971046448, "learning_rate": 0.0012566666666666666, "loss": 1.3141, "step": 20070 }, { "epoch": 74.37, "grad_norm": 0.8000026941299438, "learning_rate": 0.0012562962962962964, "loss": 1.2965, "step": 20080 }, { "epoch": 74.41, "grad_norm": 0.7329573035240173, "learning_rate": 0.001255925925925926, "loss": 1.3052, "step": 20090 }, { "epoch": 74.44, "grad_norm": 0.7924298048019409, "learning_rate": 0.0012555555555555555, "loss": 1.3378, "step": 20100 }, { "epoch": 74.48, "grad_norm": 0.7924289703369141, "learning_rate": 0.0012551851851851851, "loss": 1.3316, "step": 20110 }, { "epoch": 74.52, "grad_norm": 0.8606981635093689, "learning_rate": 0.001254814814814815, "loss": 1.322, "step": 20120 }, { "epoch": 74.56, "grad_norm": 0.8466325998306274, "learning_rate": 0.0012544444444444445, "loss": 1.3403, "step": 20130 }, { "epoch": 74.59, "grad_norm": 0.7616990804672241, "learning_rate": 0.0012540740740740741, "loss": 1.3709, "step": 20140 }, { "epoch": 74.63, "grad_norm": 0.8217259645462036, "learning_rate": 0.0012537037037037037, "loss": 1.3643, "step": 20150 }, { "epoch": 74.67, "grad_norm": 0.8023414611816406, "learning_rate": 0.0012533333333333335, "loss": 1.3635, "step": 20160 }, { "epoch": 74.7, "grad_norm": 0.7349652051925659, "learning_rate": 0.001252962962962963, "loss": 1.3707, "step": 20170 }, { "epoch": 74.74, "grad_norm": 0.7817662358283997, "learning_rate": 0.0012525925925925927, "loss": 1.4089, "step": 20180 }, { "epoch": 74.78, "grad_norm": 0.819149374961853, "learning_rate": 0.001252222222222222, "loss": 1.3633, "step": 20190 }, { "epoch": 74.81, "grad_norm": 0.811070442199707, "learning_rate": 0.0012518518518518519, "loss": 1.3932, "step": 20200 }, { "epoch": 74.85, "grad_norm": 0.7167898416519165, "learning_rate": 0.0012514814814814815, "loss": 1.412, "step": 20210 }, { "epoch": 74.89, "grad_norm": 0.6857497692108154, "learning_rate": 0.001251111111111111, "loss": 1.3975, "step": 20220 }, { "epoch": 74.93, "grad_norm": 0.7601510882377625, "learning_rate": 0.0012507407407407406, "loss": 1.3972, "step": 20230 }, { "epoch": 74.96, "grad_norm": 0.7811948657035828, "learning_rate": 0.0012503703703703704, "loss": 1.4251, "step": 20240 }, { "epoch": 75.0, "grad_norm": 1.9018360376358032, "learning_rate": 0.00125, "loss": 1.3991, "step": 20250 }, { "epoch": 75.04, "grad_norm": 0.7820109724998474, "learning_rate": 0.0012496296296296296, "loss": 1.1763, "step": 20260 }, { "epoch": 75.07, "grad_norm": 0.799656867980957, "learning_rate": 0.0012492592592592594, "loss": 1.1977, "step": 20270 }, { "epoch": 75.11, "grad_norm": 0.7293298840522766, "learning_rate": 0.001248888888888889, "loss": 1.1909, "step": 20280 }, { "epoch": 75.15, "grad_norm": 0.762310266494751, "learning_rate": 0.0012485185185185186, "loss": 1.1957, "step": 20290 }, { "epoch": 75.19, "grad_norm": 0.8296578526496887, "learning_rate": 0.0012481481481481482, "loss": 1.2131, "step": 20300 }, { "epoch": 75.22, "grad_norm": 0.7799403071403503, "learning_rate": 0.001247777777777778, "loss": 1.2428, "step": 20310 }, { "epoch": 75.26, "grad_norm": 0.8273761868476868, "learning_rate": 0.0012474074074074074, "loss": 1.2384, "step": 20320 }, { "epoch": 75.3, "grad_norm": 0.7832826375961304, "learning_rate": 0.001247037037037037, "loss": 1.2835, "step": 20330 }, { "epoch": 75.33, "grad_norm": 0.7677596807479858, "learning_rate": 0.0012466666666666665, "loss": 1.2618, "step": 20340 }, { "epoch": 75.37, "grad_norm": 0.7945502400398254, "learning_rate": 0.0012462962962962963, "loss": 1.2838, "step": 20350 }, { "epoch": 75.41, "grad_norm": 0.7328624725341797, "learning_rate": 0.001245925925925926, "loss": 1.3011, "step": 20360 }, { "epoch": 75.44, "grad_norm": 0.8257105946540833, "learning_rate": 0.0012455555555555555, "loss": 1.3188, "step": 20370 }, { "epoch": 75.48, "grad_norm": 0.8209635615348816, "learning_rate": 0.0012451851851851851, "loss": 1.3072, "step": 20380 }, { "epoch": 75.52, "grad_norm": 0.8422302603721619, "learning_rate": 0.001244814814814815, "loss": 1.3266, "step": 20390 }, { "epoch": 75.56, "grad_norm": 0.8323236107826233, "learning_rate": 0.0012444444444444445, "loss": 1.3249, "step": 20400 }, { "epoch": 75.59, "grad_norm": 0.7966023683547974, "learning_rate": 0.001244074074074074, "loss": 1.3378, "step": 20410 }, { "epoch": 75.63, "grad_norm": 0.7316237688064575, "learning_rate": 0.0012437037037037037, "loss": 1.3489, "step": 20420 }, { "epoch": 75.67, "grad_norm": 0.8376904726028442, "learning_rate": 0.0012433333333333335, "loss": 1.3569, "step": 20430 }, { "epoch": 75.7, "grad_norm": 0.799994945526123, "learning_rate": 0.001242962962962963, "loss": 1.3984, "step": 20440 }, { "epoch": 75.74, "grad_norm": 0.7840830087661743, "learning_rate": 0.0012425925925925927, "loss": 1.3776, "step": 20450 }, { "epoch": 75.78, "grad_norm": 0.7834445834159851, "learning_rate": 0.001242222222222222, "loss": 1.383, "step": 20460 }, { "epoch": 75.81, "grad_norm": 0.8726301193237305, "learning_rate": 0.0012418518518518518, "loss": 1.3965, "step": 20470 }, { "epoch": 75.85, "grad_norm": 0.8033871054649353, "learning_rate": 0.0012414814814814814, "loss": 1.386, "step": 20480 }, { "epoch": 75.89, "grad_norm": 0.8240883946418762, "learning_rate": 0.001241111111111111, "loss": 1.3963, "step": 20490 }, { "epoch": 75.93, "grad_norm": 0.7927721738815308, "learning_rate": 0.0012407407407407408, "loss": 1.4127, "step": 20500 }, { "epoch": 75.96, "grad_norm": 0.8034483194351196, "learning_rate": 0.0012403703703703704, "loss": 1.3915, "step": 20510 }, { "epoch": 76.0, "grad_norm": 1.4726250171661377, "learning_rate": 0.00124, "loss": 1.4116, "step": 20520 }, { "epoch": 76.04, "grad_norm": 0.775287389755249, "learning_rate": 0.0012396296296296296, "loss": 1.1529, "step": 20530 }, { "epoch": 76.07, "grad_norm": 0.7530531287193298, "learning_rate": 0.0012392592592592594, "loss": 1.1796, "step": 20540 }, { "epoch": 76.11, "grad_norm": 0.8483404517173767, "learning_rate": 0.001238888888888889, "loss": 1.1751, "step": 20550 }, { "epoch": 76.15, "grad_norm": 0.8471804857254028, "learning_rate": 0.0012385185185185186, "loss": 1.1828, "step": 20560 }, { "epoch": 76.19, "grad_norm": 0.7711694836616516, "learning_rate": 0.0012381481481481482, "loss": 1.2362, "step": 20570 }, { "epoch": 76.22, "grad_norm": 0.8558454513549805, "learning_rate": 0.001237777777777778, "loss": 1.2296, "step": 20580 }, { "epoch": 76.26, "grad_norm": 0.7980576753616333, "learning_rate": 0.0012374074074074076, "loss": 1.2518, "step": 20590 }, { "epoch": 76.3, "grad_norm": 0.7709694504737854, "learning_rate": 0.001237037037037037, "loss": 1.2723, "step": 20600 }, { "epoch": 76.33, "grad_norm": 0.7979551553726196, "learning_rate": 0.0012366666666666665, "loss": 1.2802, "step": 20610 }, { "epoch": 76.37, "grad_norm": 0.7723931670188904, "learning_rate": 0.0012362962962962963, "loss": 1.2805, "step": 20620 }, { "epoch": 76.41, "grad_norm": 0.7951886653900146, "learning_rate": 0.001235925925925926, "loss": 1.3112, "step": 20630 }, { "epoch": 76.44, "grad_norm": 0.8366179466247559, "learning_rate": 0.0012355555555555555, "loss": 1.3134, "step": 20640 }, { "epoch": 76.48, "grad_norm": 0.759067177772522, "learning_rate": 0.001235185185185185, "loss": 1.3315, "step": 20650 }, { "epoch": 76.52, "grad_norm": 0.7688968777656555, "learning_rate": 0.0012348148148148149, "loss": 1.3062, "step": 20660 }, { "epoch": 76.56, "grad_norm": 0.7929026484489441, "learning_rate": 0.0012344444444444445, "loss": 1.3243, "step": 20670 }, { "epoch": 76.59, "grad_norm": 0.7894778251647949, "learning_rate": 0.001234074074074074, "loss": 1.3411, "step": 20680 }, { "epoch": 76.63, "grad_norm": 0.7940474152565002, "learning_rate": 0.0012337037037037039, "loss": 1.3307, "step": 20690 }, { "epoch": 76.67, "grad_norm": 0.7494774460792542, "learning_rate": 0.0012333333333333335, "loss": 1.3671, "step": 20700 }, { "epoch": 76.7, "grad_norm": 0.8801915049552917, "learning_rate": 0.001232962962962963, "loss": 1.3683, "step": 20710 }, { "epoch": 76.74, "grad_norm": 0.8223617076873779, "learning_rate": 0.0012325925925925926, "loss": 1.3709, "step": 20720 }, { "epoch": 76.78, "grad_norm": 0.8362076878547668, "learning_rate": 0.0012322222222222224, "loss": 1.3642, "step": 20730 }, { "epoch": 76.81, "grad_norm": 0.8495985865592957, "learning_rate": 0.0012318518518518518, "loss": 1.3719, "step": 20740 }, { "epoch": 76.85, "grad_norm": 0.8225275278091431, "learning_rate": 0.0012314814814814814, "loss": 1.3903, "step": 20750 }, { "epoch": 76.89, "grad_norm": 0.7975557446479797, "learning_rate": 0.001231111111111111, "loss": 1.4067, "step": 20760 }, { "epoch": 76.93, "grad_norm": 0.8219361901283264, "learning_rate": 0.0012307407407407408, "loss": 1.4093, "step": 20770 }, { "epoch": 76.96, "grad_norm": 0.8116274476051331, "learning_rate": 0.0012303703703703704, "loss": 1.4051, "step": 20780 }, { "epoch": 77.0, "grad_norm": 1.761422872543335, "learning_rate": 0.00123, "loss": 1.3978, "step": 20790 }, { "epoch": 77.04, "grad_norm": 0.7580262422561646, "learning_rate": 0.0012296296296296296, "loss": 1.1828, "step": 20800 }, { "epoch": 77.07, "grad_norm": 0.7770316004753113, "learning_rate": 0.0012292592592592594, "loss": 1.1474, "step": 20810 }, { "epoch": 77.11, "grad_norm": 0.7733827233314514, "learning_rate": 0.001228888888888889, "loss": 1.1756, "step": 20820 }, { "epoch": 77.15, "grad_norm": 0.8824875354766846, "learning_rate": 0.0012285185185185185, "loss": 1.213, "step": 20830 }, { "epoch": 77.19, "grad_norm": 0.8360825777053833, "learning_rate": 0.0012281481481481481, "loss": 1.2194, "step": 20840 }, { "epoch": 77.22, "grad_norm": 0.790859580039978, "learning_rate": 0.001227777777777778, "loss": 1.2376, "step": 20850 }, { "epoch": 77.26, "grad_norm": 0.8717347383499146, "learning_rate": 0.0012274074074074075, "loss": 1.2512, "step": 20860 }, { "epoch": 77.3, "grad_norm": 0.7890602350234985, "learning_rate": 0.001227037037037037, "loss": 1.2397, "step": 20870 }, { "epoch": 77.33, "grad_norm": 0.8594658970832825, "learning_rate": 0.0012266666666666667, "loss": 1.2704, "step": 20880 }, { "epoch": 77.37, "grad_norm": 0.8138068914413452, "learning_rate": 0.0012262962962962963, "loss": 1.2879, "step": 20890 }, { "epoch": 77.41, "grad_norm": 0.8145601749420166, "learning_rate": 0.0012259259259259259, "loss": 1.3026, "step": 20900 }, { "epoch": 77.44, "grad_norm": 0.8223507404327393, "learning_rate": 0.0012255555555555555, "loss": 1.3116, "step": 20910 }, { "epoch": 77.48, "grad_norm": 0.8235211372375488, "learning_rate": 0.0012251851851851853, "loss": 1.3071, "step": 20920 }, { "epoch": 77.52, "grad_norm": 0.7376195788383484, "learning_rate": 0.0012248148148148149, "loss": 1.3193, "step": 20930 }, { "epoch": 77.56, "grad_norm": 0.7939662337303162, "learning_rate": 0.0012244444444444445, "loss": 1.3261, "step": 20940 }, { "epoch": 77.59, "grad_norm": 0.8232659101486206, "learning_rate": 0.001224074074074074, "loss": 1.3285, "step": 20950 }, { "epoch": 77.63, "grad_norm": 0.772993266582489, "learning_rate": 0.0012237037037037038, "loss": 1.3368, "step": 20960 }, { "epoch": 77.67, "grad_norm": 0.8146404027938843, "learning_rate": 0.0012233333333333334, "loss": 1.3365, "step": 20970 }, { "epoch": 77.7, "grad_norm": 0.7433125972747803, "learning_rate": 0.001222962962962963, "loss": 1.3431, "step": 20980 }, { "epoch": 77.74, "grad_norm": 0.8123118281364441, "learning_rate": 0.0012225925925925926, "loss": 1.3616, "step": 20990 }, { "epoch": 77.78, "grad_norm": 0.7485589385032654, "learning_rate": 0.0012222222222222224, "loss": 1.3594, "step": 21000 }, { "epoch": 77.81, "grad_norm": 0.8744904398918152, "learning_rate": 0.0012218518518518518, "loss": 1.3389, "step": 21010 }, { "epoch": 77.85, "grad_norm": 0.813572883605957, "learning_rate": 0.0012214814814814814, "loss": 1.3902, "step": 21020 }, { "epoch": 77.89, "grad_norm": 0.8318865895271301, "learning_rate": 0.001221111111111111, "loss": 1.3729, "step": 21030 }, { "epoch": 77.93, "grad_norm": 0.7886088490486145, "learning_rate": 0.0012207407407407408, "loss": 1.3966, "step": 21040 }, { "epoch": 77.96, "grad_norm": 0.7813388109207153, "learning_rate": 0.0012203703703703704, "loss": 1.3769, "step": 21050 }, { "epoch": 78.0, "grad_norm": 1.5267912149429321, "learning_rate": 0.00122, "loss": 1.408, "step": 21060 }, { "epoch": 78.04, "grad_norm": 0.7442604899406433, "learning_rate": 0.0012196296296296298, "loss": 1.1388, "step": 21070 }, { "epoch": 78.07, "grad_norm": 0.7419655919075012, "learning_rate": 0.0012192592592592593, "loss": 1.1539, "step": 21080 }, { "epoch": 78.11, "grad_norm": 0.8307773470878601, "learning_rate": 0.001218888888888889, "loss": 1.1785, "step": 21090 }, { "epoch": 78.15, "grad_norm": 0.7722887992858887, "learning_rate": 0.0012185185185185185, "loss": 1.1858, "step": 21100 }, { "epoch": 78.19, "grad_norm": 0.7633727192878723, "learning_rate": 0.0012181481481481483, "loss": 1.1952, "step": 21110 }, { "epoch": 78.22, "grad_norm": 0.82599937915802, "learning_rate": 0.001217777777777778, "loss": 1.2302, "step": 21120 }, { "epoch": 78.26, "grad_norm": 0.8898715376853943, "learning_rate": 0.0012174074074074075, "loss": 1.222, "step": 21130 }, { "epoch": 78.3, "grad_norm": 0.8221423625946045, "learning_rate": 0.001217037037037037, "loss": 1.2533, "step": 21140 }, { "epoch": 78.33, "grad_norm": 0.8007521629333496, "learning_rate": 0.0012166666666666667, "loss": 1.2614, "step": 21150 }, { "epoch": 78.37, "grad_norm": 0.7668291926383972, "learning_rate": 0.0012162962962962963, "loss": 1.262, "step": 21160 }, { "epoch": 78.41, "grad_norm": 0.8179112672805786, "learning_rate": 0.0012159259259259259, "loss": 1.2717, "step": 21170 }, { "epoch": 78.44, "grad_norm": 0.8355836272239685, "learning_rate": 0.0012155555555555554, "loss": 1.2962, "step": 21180 }, { "epoch": 78.48, "grad_norm": 0.7732020020484924, "learning_rate": 0.0012151851851851852, "loss": 1.322, "step": 21190 }, { "epoch": 78.52, "grad_norm": 0.8221396207809448, "learning_rate": 0.0012148148148148148, "loss": 1.2944, "step": 21200 }, { "epoch": 78.56, "grad_norm": 0.8312259316444397, "learning_rate": 0.0012144444444444444, "loss": 1.3004, "step": 21210 }, { "epoch": 78.59, "grad_norm": 0.8105566501617432, "learning_rate": 0.001214074074074074, "loss": 1.2993, "step": 21220 }, { "epoch": 78.63, "grad_norm": 0.8554087281227112, "learning_rate": 0.0012137037037037038, "loss": 1.3287, "step": 21230 }, { "epoch": 78.67, "grad_norm": 0.8251190185546875, "learning_rate": 0.0012133333333333334, "loss": 1.3235, "step": 21240 }, { "epoch": 78.7, "grad_norm": 0.8064236640930176, "learning_rate": 0.001212962962962963, "loss": 1.3465, "step": 21250 }, { "epoch": 78.74, "grad_norm": 0.75641930103302, "learning_rate": 0.0012125925925925926, "loss": 1.3483, "step": 21260 }, { "epoch": 78.78, "grad_norm": 0.7814801335334778, "learning_rate": 0.0012122222222222224, "loss": 1.3438, "step": 21270 }, { "epoch": 78.81, "grad_norm": 0.8190613389015198, "learning_rate": 0.001211851851851852, "loss": 1.3722, "step": 21280 }, { "epoch": 78.85, "grad_norm": 0.8852077722549438, "learning_rate": 0.0012114814814814814, "loss": 1.3676, "step": 21290 }, { "epoch": 78.89, "grad_norm": 0.8047147393226624, "learning_rate": 0.0012111111111111112, "loss": 1.3839, "step": 21300 }, { "epoch": 78.93, "grad_norm": 0.7829154133796692, "learning_rate": 0.0012107407407407407, "loss": 1.3801, "step": 21310 }, { "epoch": 78.96, "grad_norm": 0.7903611063957214, "learning_rate": 0.0012103703703703703, "loss": 1.409, "step": 21320 }, { "epoch": 79.0, "grad_norm": 1.498512864112854, "learning_rate": 0.00121, "loss": 1.3902, "step": 21330 }, { "epoch": 79.04, "grad_norm": 0.8592652082443237, "learning_rate": 0.0012096296296296297, "loss": 1.1362, "step": 21340 }, { "epoch": 79.07, "grad_norm": 0.8371859192848206, "learning_rate": 0.0012092592592592593, "loss": 1.1476, "step": 21350 }, { "epoch": 79.11, "grad_norm": 0.8219607472419739, "learning_rate": 0.001208888888888889, "loss": 1.1573, "step": 21360 }, { "epoch": 79.15, "grad_norm": 0.7551726698875427, "learning_rate": 0.0012085185185185185, "loss": 1.1812, "step": 21370 }, { "epoch": 79.19, "grad_norm": 0.774641215801239, "learning_rate": 0.0012081481481481483, "loss": 1.199, "step": 21380 }, { "epoch": 79.22, "grad_norm": 0.8316019773483276, "learning_rate": 0.0012077777777777779, "loss": 1.23, "step": 21390 }, { "epoch": 79.26, "grad_norm": 0.8777797222137451, "learning_rate": 0.0012074074074074075, "loss": 1.2181, "step": 21400 }, { "epoch": 79.3, "grad_norm": 0.8278775215148926, "learning_rate": 0.001207037037037037, "loss": 1.2521, "step": 21410 }, { "epoch": 79.33, "grad_norm": 0.8261604309082031, "learning_rate": 0.0012066666666666669, "loss": 1.2502, "step": 21420 }, { "epoch": 79.37, "grad_norm": 0.784346342086792, "learning_rate": 0.0012062962962962962, "loss": 1.2532, "step": 21430 }, { "epoch": 79.41, "grad_norm": 0.7922409772872925, "learning_rate": 0.0012059259259259258, "loss": 1.2805, "step": 21440 }, { "epoch": 79.44, "grad_norm": 0.8692464232444763, "learning_rate": 0.0012055555555555554, "loss": 1.2817, "step": 21450 }, { "epoch": 79.48, "grad_norm": 0.8120541572570801, "learning_rate": 0.0012051851851851852, "loss": 1.3181, "step": 21460 }, { "epoch": 79.52, "grad_norm": 0.8435699939727783, "learning_rate": 0.0012048148148148148, "loss": 1.3104, "step": 21470 }, { "epoch": 79.56, "grad_norm": 0.8154937028884888, "learning_rate": 0.0012044444444444444, "loss": 1.333, "step": 21480 }, { "epoch": 79.59, "grad_norm": 0.852403461933136, "learning_rate": 0.0012040740740740742, "loss": 1.304, "step": 21490 }, { "epoch": 79.63, "grad_norm": 0.8190922141075134, "learning_rate": 0.0012037037037037038, "loss": 1.3346, "step": 21500 }, { "epoch": 79.67, "grad_norm": 0.8322864770889282, "learning_rate": 0.0012033333333333334, "loss": 1.3163, "step": 21510 }, { "epoch": 79.7, "grad_norm": 0.8163490295410156, "learning_rate": 0.001202962962962963, "loss": 1.342, "step": 21520 }, { "epoch": 79.74, "grad_norm": 0.8012131452560425, "learning_rate": 0.0012025925925925928, "loss": 1.336, "step": 21530 }, { "epoch": 79.78, "grad_norm": 0.8321338295936584, "learning_rate": 0.0012022222222222224, "loss": 1.3515, "step": 21540 }, { "epoch": 79.81, "grad_norm": 0.772287130355835, "learning_rate": 0.001201851851851852, "loss": 1.3789, "step": 21550 }, { "epoch": 79.85, "grad_norm": 0.7578256130218506, "learning_rate": 0.0012014814814814813, "loss": 1.3435, "step": 21560 }, { "epoch": 79.89, "grad_norm": 0.7756651043891907, "learning_rate": 0.0012011111111111111, "loss": 1.3559, "step": 21570 }, { "epoch": 79.93, "grad_norm": 0.7970812916755676, "learning_rate": 0.0012007407407407407, "loss": 1.3538, "step": 21580 }, { "epoch": 79.96, "grad_norm": 0.8378322124481201, "learning_rate": 0.0012003703703703703, "loss": 1.3555, "step": 21590 }, { "epoch": 80.0, "grad_norm": 1.7049368619918823, "learning_rate": 0.0012, "loss": 1.3829, "step": 21600 }, { "epoch": 80.04, "grad_norm": 0.7638210654258728, "learning_rate": 0.0011996296296296297, "loss": 1.1375, "step": 21610 }, { "epoch": 80.07, "grad_norm": 0.7893775701522827, "learning_rate": 0.0011992592592592593, "loss": 1.1598, "step": 21620 }, { "epoch": 80.11, "grad_norm": 0.7928591370582581, "learning_rate": 0.0011988888888888889, "loss": 1.1812, "step": 21630 }, { "epoch": 80.15, "grad_norm": 0.7909486293792725, "learning_rate": 0.0011985185185185185, "loss": 1.156, "step": 21640 }, { "epoch": 80.19, "grad_norm": 0.7979161143302917, "learning_rate": 0.0011981481481481483, "loss": 1.2035, "step": 21650 }, { "epoch": 80.22, "grad_norm": 0.7922539114952087, "learning_rate": 0.0011977777777777779, "loss": 1.2042, "step": 21660 }, { "epoch": 80.26, "grad_norm": 0.7434383630752563, "learning_rate": 0.0011974074074074074, "loss": 1.2278, "step": 21670 }, { "epoch": 80.3, "grad_norm": 0.814395010471344, "learning_rate": 0.0011970370370370373, "loss": 1.2216, "step": 21680 }, { "epoch": 80.33, "grad_norm": 0.77228182554245, "learning_rate": 0.0011966666666666668, "loss": 1.2295, "step": 21690 }, { "epoch": 80.37, "grad_norm": 0.8123186230659485, "learning_rate": 0.0011962962962962962, "loss": 1.2581, "step": 21700 }, { "epoch": 80.41, "grad_norm": 0.7760822176933289, "learning_rate": 0.0011959259259259258, "loss": 1.2721, "step": 21710 }, { "epoch": 80.44, "grad_norm": 0.8242724537849426, "learning_rate": 0.0011955555555555556, "loss": 1.245, "step": 21720 }, { "epoch": 80.48, "grad_norm": 0.823330819606781, "learning_rate": 0.0011951851851851852, "loss": 1.2789, "step": 21730 }, { "epoch": 80.52, "grad_norm": 0.758445143699646, "learning_rate": 0.0011948148148148148, "loss": 1.276, "step": 21740 }, { "epoch": 80.56, "grad_norm": 0.8475410342216492, "learning_rate": 0.0011944444444444444, "loss": 1.3156, "step": 21750 }, { "epoch": 80.59, "grad_norm": 0.8873695731163025, "learning_rate": 0.0011940740740740742, "loss": 1.3206, "step": 21760 }, { "epoch": 80.63, "grad_norm": 0.7427895665168762, "learning_rate": 0.0011937037037037038, "loss": 1.3145, "step": 21770 }, { "epoch": 80.67, "grad_norm": 0.7813780307769775, "learning_rate": 0.0011933333333333334, "loss": 1.3108, "step": 21780 }, { "epoch": 80.7, "grad_norm": 0.8349199891090393, "learning_rate": 0.001192962962962963, "loss": 1.3172, "step": 21790 }, { "epoch": 80.74, "grad_norm": 0.8662471175193787, "learning_rate": 0.0011925925925925927, "loss": 1.3509, "step": 21800 }, { "epoch": 80.78, "grad_norm": 0.7870455980300903, "learning_rate": 0.0011922222222222223, "loss": 1.345, "step": 21810 }, { "epoch": 80.81, "grad_norm": 0.8087645769119263, "learning_rate": 0.001191851851851852, "loss": 1.34, "step": 21820 }, { "epoch": 80.85, "grad_norm": 0.8210803866386414, "learning_rate": 0.0011914814814814815, "loss": 1.3496, "step": 21830 }, { "epoch": 80.89, "grad_norm": 0.7690105438232422, "learning_rate": 0.001191111111111111, "loss": 1.3824, "step": 21840 }, { "epoch": 80.93, "grad_norm": 0.7893791794776917, "learning_rate": 0.0011907407407407407, "loss": 1.3591, "step": 21850 }, { "epoch": 80.96, "grad_norm": 0.7507551312446594, "learning_rate": 0.0011903703703703703, "loss": 1.3562, "step": 21860 }, { "epoch": 81.0, "grad_norm": 1.2659991979599, "learning_rate": 0.0011899999999999999, "loss": 1.3887, "step": 21870 }, { "epoch": 81.04, "grad_norm": 0.7406259179115295, "learning_rate": 0.0011896296296296297, "loss": 1.1216, "step": 21880 }, { "epoch": 81.07, "grad_norm": 0.8164389133453369, "learning_rate": 0.0011892592592592593, "loss": 1.1308, "step": 21890 }, { "epoch": 81.11, "grad_norm": 0.7737013101577759, "learning_rate": 0.0011888888888888889, "loss": 1.1673, "step": 21900 }, { "epoch": 81.15, "grad_norm": 0.7928569912910461, "learning_rate": 0.0011885185185185187, "loss": 1.1678, "step": 21910 }, { "epoch": 81.19, "grad_norm": 0.8306735754013062, "learning_rate": 0.0011881481481481482, "loss": 1.1786, "step": 21920 }, { "epoch": 81.22, "grad_norm": 0.7711634635925293, "learning_rate": 0.0011877777777777778, "loss": 1.2144, "step": 21930 }, { "epoch": 81.26, "grad_norm": 0.8662676811218262, "learning_rate": 0.0011874074074074074, "loss": 1.1809, "step": 21940 }, { "epoch": 81.3, "grad_norm": 0.8031598329544067, "learning_rate": 0.0011870370370370372, "loss": 1.2241, "step": 21950 }, { "epoch": 81.33, "grad_norm": 0.797000527381897, "learning_rate": 0.0011866666666666668, "loss": 1.2302, "step": 21960 }, { "epoch": 81.37, "grad_norm": 0.9229769706726074, "learning_rate": 0.0011862962962962964, "loss": 1.2381, "step": 21970 }, { "epoch": 81.41, "grad_norm": 0.9003340601921082, "learning_rate": 0.0011859259259259258, "loss": 1.2554, "step": 21980 }, { "epoch": 81.44, "grad_norm": 0.798082172870636, "learning_rate": 0.0011855555555555556, "loss": 1.2481, "step": 21990 }, { "epoch": 81.48, "grad_norm": 0.8588927984237671, "learning_rate": 0.0011851851851851852, "loss": 1.2658, "step": 22000 }, { "epoch": 81.52, "grad_norm": 0.846261203289032, "learning_rate": 0.0011848148148148148, "loss": 1.2796, "step": 22010 }, { "epoch": 81.56, "grad_norm": 0.86174076795578, "learning_rate": 0.0011844444444444443, "loss": 1.3243, "step": 22020 }, { "epoch": 81.59, "grad_norm": 0.7928067445755005, "learning_rate": 0.0011840740740740742, "loss": 1.2979, "step": 22030 }, { "epoch": 81.63, "grad_norm": 0.8082354068756104, "learning_rate": 0.0011837037037037037, "loss": 1.311, "step": 22040 }, { "epoch": 81.67, "grad_norm": 0.8235210180282593, "learning_rate": 0.0011833333333333333, "loss": 1.3244, "step": 22050 }, { "epoch": 81.7, "grad_norm": 0.8794702887535095, "learning_rate": 0.001182962962962963, "loss": 1.305, "step": 22060 }, { "epoch": 81.74, "grad_norm": 0.7763599753379822, "learning_rate": 0.0011825925925925927, "loss": 1.3389, "step": 22070 }, { "epoch": 81.78, "grad_norm": 0.8223667740821838, "learning_rate": 0.0011822222222222223, "loss": 1.3309, "step": 22080 }, { "epoch": 81.81, "grad_norm": 0.8330878615379333, "learning_rate": 0.001181851851851852, "loss": 1.3424, "step": 22090 }, { "epoch": 81.85, "grad_norm": 0.8290947079658508, "learning_rate": 0.0011814814814814817, "loss": 1.3654, "step": 22100 }, { "epoch": 81.89, "grad_norm": 0.7992547750473022, "learning_rate": 0.001181111111111111, "loss": 1.3522, "step": 22110 }, { "epoch": 81.93, "grad_norm": 0.795511782169342, "learning_rate": 0.0011807407407407407, "loss": 1.3555, "step": 22120 }, { "epoch": 81.96, "grad_norm": 0.8905317187309265, "learning_rate": 0.0011803703703703703, "loss": 1.3828, "step": 22130 }, { "epoch": 82.0, "grad_norm": 1.376718521118164, "learning_rate": 0.00118, "loss": 1.3494, "step": 22140 }, { "epoch": 82.04, "grad_norm": 0.8007773160934448, "learning_rate": 0.0011796296296296296, "loss": 1.1338, "step": 22150 }, { "epoch": 82.07, "grad_norm": 0.7765668630599976, "learning_rate": 0.0011792592592592592, "loss": 1.1248, "step": 22160 }, { "epoch": 82.11, "grad_norm": 0.8055347800254822, "learning_rate": 0.0011788888888888888, "loss": 1.1597, "step": 22170 }, { "epoch": 82.15, "grad_norm": 0.8667762279510498, "learning_rate": 0.0011785185185185186, "loss": 1.1803, "step": 22180 }, { "epoch": 82.19, "grad_norm": 0.8008133769035339, "learning_rate": 0.0011781481481481482, "loss": 1.1927, "step": 22190 }, { "epoch": 82.22, "grad_norm": 0.8304885625839233, "learning_rate": 0.0011777777777777778, "loss": 1.1807, "step": 22200 }, { "epoch": 82.26, "grad_norm": 0.8007597327232361, "learning_rate": 0.0011774074074074074, "loss": 1.2058, "step": 22210 }, { "epoch": 82.3, "grad_norm": 0.8569217920303345, "learning_rate": 0.0011770370370370372, "loss": 1.2233, "step": 22220 }, { "epoch": 82.33, "grad_norm": 0.8300482034683228, "learning_rate": 0.0011766666666666668, "loss": 1.2384, "step": 22230 }, { "epoch": 82.37, "grad_norm": 0.8271434903144836, "learning_rate": 0.0011762962962962964, "loss": 1.2295, "step": 22240 }, { "epoch": 82.41, "grad_norm": 0.7994027733802795, "learning_rate": 0.0011759259259259257, "loss": 1.235, "step": 22250 }, { "epoch": 82.44, "grad_norm": 0.7732693552970886, "learning_rate": 0.0011755555555555556, "loss": 1.2754, "step": 22260 }, { "epoch": 82.48, "grad_norm": 0.8843140602111816, "learning_rate": 0.0011751851851851851, "loss": 1.2734, "step": 22270 }, { "epoch": 82.52, "grad_norm": 0.7856510281562805, "learning_rate": 0.0011748148148148147, "loss": 1.2753, "step": 22280 }, { "epoch": 82.56, "grad_norm": 0.851887583732605, "learning_rate": 0.0011744444444444443, "loss": 1.2706, "step": 22290 }, { "epoch": 82.59, "grad_norm": 0.8285180926322937, "learning_rate": 0.0011740740740740741, "loss": 1.2877, "step": 22300 }, { "epoch": 82.63, "grad_norm": 0.8394416570663452, "learning_rate": 0.0011737037037037037, "loss": 1.2833, "step": 22310 }, { "epoch": 82.67, "grad_norm": 0.8958646059036255, "learning_rate": 0.0011733333333333333, "loss": 1.311, "step": 22320 }, { "epoch": 82.7, "grad_norm": 0.8369964361190796, "learning_rate": 0.001172962962962963, "loss": 1.3369, "step": 22330 }, { "epoch": 82.74, "grad_norm": 0.8263498544692993, "learning_rate": 0.0011725925925925927, "loss": 1.3226, "step": 22340 }, { "epoch": 82.78, "grad_norm": 0.7990745306015015, "learning_rate": 0.0011722222222222223, "loss": 1.3485, "step": 22350 }, { "epoch": 82.81, "grad_norm": 0.7908133864402771, "learning_rate": 0.0011718518518518519, "loss": 1.3262, "step": 22360 }, { "epoch": 82.85, "grad_norm": 0.8254690170288086, "learning_rate": 0.0011714814814814817, "loss": 1.3466, "step": 22370 }, { "epoch": 82.89, "grad_norm": 0.8180753588676453, "learning_rate": 0.0011711111111111113, "loss": 1.3433, "step": 22380 }, { "epoch": 82.93, "grad_norm": 0.8374766707420349, "learning_rate": 0.0011707407407407406, "loss": 1.3298, "step": 22390 }, { "epoch": 82.96, "grad_norm": 0.8209193348884583, "learning_rate": 0.0011703703703703702, "loss": 1.3422, "step": 22400 }, { "epoch": 83.0, "grad_norm": 1.5476839542388916, "learning_rate": 0.00117, "loss": 1.356, "step": 22410 }, { "epoch": 83.04, "grad_norm": 0.9864056706428528, "learning_rate": 0.0011696296296296296, "loss": 1.1234, "step": 22420 }, { "epoch": 83.07, "grad_norm": 0.847768247127533, "learning_rate": 0.0011692592592592592, "loss": 1.1251, "step": 22430 }, { "epoch": 83.11, "grad_norm": 0.8513081073760986, "learning_rate": 0.0011688888888888888, "loss": 1.132, "step": 22440 }, { "epoch": 83.15, "grad_norm": 0.7776345610618591, "learning_rate": 0.0011685185185185186, "loss": 1.1227, "step": 22450 }, { "epoch": 83.19, "grad_norm": 0.8347729444503784, "learning_rate": 0.0011681481481481482, "loss": 1.178, "step": 22460 }, { "epoch": 83.22, "grad_norm": 0.8397579193115234, "learning_rate": 0.0011677777777777778, "loss": 1.1669, "step": 22470 }, { "epoch": 83.26, "grad_norm": 0.8121944665908813, "learning_rate": 0.0011674074074074074, "loss": 1.1843, "step": 22480 }, { "epoch": 83.3, "grad_norm": 0.8163562417030334, "learning_rate": 0.0011670370370370372, "loss": 1.2016, "step": 22490 }, { "epoch": 83.33, "grad_norm": 0.7758229970932007, "learning_rate": 0.0011666666666666668, "loss": 1.2412, "step": 22500 }, { "epoch": 83.37, "grad_norm": 0.8472750186920166, "learning_rate": 0.0011662962962962964, "loss": 1.2395, "step": 22510 }, { "epoch": 83.41, "grad_norm": 0.802095353603363, "learning_rate": 0.0011659259259259262, "loss": 1.2375, "step": 22520 }, { "epoch": 83.44, "grad_norm": 0.8349648714065552, "learning_rate": 0.0011655555555555555, "loss": 1.2775, "step": 22530 }, { "epoch": 83.48, "grad_norm": 0.8553137183189392, "learning_rate": 0.0011651851851851851, "loss": 1.2516, "step": 22540 }, { "epoch": 83.52, "grad_norm": 0.836848795413971, "learning_rate": 0.0011648148148148147, "loss": 1.2481, "step": 22550 }, { "epoch": 83.56, "grad_norm": 0.7819425463676453, "learning_rate": 0.0011644444444444445, "loss": 1.2914, "step": 22560 }, { "epoch": 83.59, "grad_norm": 0.8485893607139587, "learning_rate": 0.001164074074074074, "loss": 1.3089, "step": 22570 }, { "epoch": 83.63, "grad_norm": 0.8001742362976074, "learning_rate": 0.0011637037037037037, "loss": 1.3022, "step": 22580 }, { "epoch": 83.67, "grad_norm": 0.7983673810958862, "learning_rate": 0.0011633333333333333, "loss": 1.2967, "step": 22590 }, { "epoch": 83.7, "grad_norm": 0.8425639867782593, "learning_rate": 0.001162962962962963, "loss": 1.3126, "step": 22600 }, { "epoch": 83.74, "grad_norm": 0.8058281540870667, "learning_rate": 0.0011625925925925927, "loss": 1.3247, "step": 22610 }, { "epoch": 83.78, "grad_norm": 0.8269002437591553, "learning_rate": 0.0011622222222222223, "loss": 1.3237, "step": 22620 }, { "epoch": 83.81, "grad_norm": 0.8307337164878845, "learning_rate": 0.0011618518518518518, "loss": 1.3169, "step": 22630 }, { "epoch": 83.85, "grad_norm": 0.8474606275558472, "learning_rate": 0.0011614814814814817, "loss": 1.3301, "step": 22640 }, { "epoch": 83.89, "grad_norm": 0.7992864847183228, "learning_rate": 0.0011611111111111112, "loss": 1.3275, "step": 22650 }, { "epoch": 83.93, "grad_norm": 0.9049937129020691, "learning_rate": 0.0011607407407407408, "loss": 1.3411, "step": 22660 }, { "epoch": 83.96, "grad_norm": 0.8497620224952698, "learning_rate": 0.0011603703703703702, "loss": 1.343, "step": 22670 }, { "epoch": 84.0, "grad_norm": 1.48526132106781, "learning_rate": 0.00116, "loss": 1.3499, "step": 22680 }, { "epoch": 84.04, "grad_norm": 0.8892446756362915, "learning_rate": 0.0011596296296296296, "loss": 1.1125, "step": 22690 }, { "epoch": 84.07, "grad_norm": 0.7651263475418091, "learning_rate": 0.0011592592592592592, "loss": 1.0929, "step": 22700 }, { "epoch": 84.11, "grad_norm": 0.8239293098449707, "learning_rate": 0.001158888888888889, "loss": 1.1195, "step": 22710 }, { "epoch": 84.15, "grad_norm": 0.8643993139266968, "learning_rate": 0.0011585185185185186, "loss": 1.1429, "step": 22720 }, { "epoch": 84.19, "grad_norm": 0.806343138217926, "learning_rate": 0.0011581481481481482, "loss": 1.1518, "step": 22730 }, { "epoch": 84.22, "grad_norm": 0.9068303108215332, "learning_rate": 0.0011577777777777778, "loss": 1.1607, "step": 22740 }, { "epoch": 84.26, "grad_norm": 0.8175188899040222, "learning_rate": 0.0011574074074074076, "loss": 1.2107, "step": 22750 }, { "epoch": 84.3, "grad_norm": 0.8616414070129395, "learning_rate": 0.0011570370370370371, "loss": 1.209, "step": 22760 }, { "epoch": 84.33, "grad_norm": 0.8178994655609131, "learning_rate": 0.0011566666666666667, "loss": 1.2123, "step": 22770 }, { "epoch": 84.37, "grad_norm": 0.831458568572998, "learning_rate": 0.0011562962962962963, "loss": 1.221, "step": 22780 }, { "epoch": 84.41, "grad_norm": 0.7806716561317444, "learning_rate": 0.0011559259259259261, "loss": 1.2491, "step": 22790 }, { "epoch": 84.44, "grad_norm": 0.816102147102356, "learning_rate": 0.0011555555555555555, "loss": 1.2511, "step": 22800 }, { "epoch": 84.48, "grad_norm": 0.824440062046051, "learning_rate": 0.001155185185185185, "loss": 1.2438, "step": 22810 }, { "epoch": 84.52, "grad_norm": 0.8962615728378296, "learning_rate": 0.0011548148148148147, "loss": 1.2523, "step": 22820 }, { "epoch": 84.56, "grad_norm": 0.7982916831970215, "learning_rate": 0.0011544444444444445, "loss": 1.2795, "step": 22830 }, { "epoch": 84.59, "grad_norm": 0.9069809913635254, "learning_rate": 0.001154074074074074, "loss": 1.2509, "step": 22840 }, { "epoch": 84.63, "grad_norm": 0.8316929936408997, "learning_rate": 0.0011537037037037037, "loss": 1.3157, "step": 22850 }, { "epoch": 84.67, "grad_norm": 0.8130622506141663, "learning_rate": 0.0011533333333333333, "loss": 1.2828, "step": 22860 }, { "epoch": 84.7, "grad_norm": 0.8644458651542664, "learning_rate": 0.001152962962962963, "loss": 1.3089, "step": 22870 }, { "epoch": 84.74, "grad_norm": 0.8116170763969421, "learning_rate": 0.0011525925925925926, "loss": 1.3101, "step": 22880 }, { "epoch": 84.78, "grad_norm": 0.7835599184036255, "learning_rate": 0.0011522222222222222, "loss": 1.2817, "step": 22890 }, { "epoch": 84.81, "grad_norm": 0.9315503239631653, "learning_rate": 0.0011518518518518518, "loss": 1.3308, "step": 22900 }, { "epoch": 84.85, "grad_norm": 0.823969304561615, "learning_rate": 0.0011514814814814816, "loss": 1.3179, "step": 22910 }, { "epoch": 84.89, "grad_norm": 0.8243674635887146, "learning_rate": 0.0011511111111111112, "loss": 1.3323, "step": 22920 }, { "epoch": 84.93, "grad_norm": 0.8511907458305359, "learning_rate": 0.0011507407407407408, "loss": 1.3621, "step": 22930 }, { "epoch": 84.96, "grad_norm": 0.8778294324874878, "learning_rate": 0.0011503703703703704, "loss": 1.3481, "step": 22940 }, { "epoch": 85.0, "grad_norm": 1.4845128059387207, "learning_rate": 0.00115, "loss": 1.3523, "step": 22950 }, { "epoch": 85.04, "grad_norm": 0.7813113927841187, "learning_rate": 0.0011496296296296296, "loss": 1.1324, "step": 22960 }, { "epoch": 85.07, "grad_norm": 0.881803035736084, "learning_rate": 0.0011492592592592592, "loss": 1.1028, "step": 22970 }, { "epoch": 85.11, "grad_norm": 0.8418541550636292, "learning_rate": 0.001148888888888889, "loss": 1.1163, "step": 22980 }, { "epoch": 85.15, "grad_norm": 0.8543863296508789, "learning_rate": 0.0011485185185185186, "loss": 1.1579, "step": 22990 }, { "epoch": 85.19, "grad_norm": 0.9971126317977905, "learning_rate": 0.0011481481481481481, "loss": 1.1375, "step": 23000 }, { "epoch": 85.22, "grad_norm": 0.90798419713974, "learning_rate": 0.0011477777777777777, "loss": 1.1714, "step": 23010 }, { "epoch": 85.26, "grad_norm": 0.8441275954246521, "learning_rate": 0.0011474074074074075, "loss": 1.1726, "step": 23020 }, { "epoch": 85.3, "grad_norm": 0.8868027925491333, "learning_rate": 0.0011470370370370371, "loss": 1.2162, "step": 23030 }, { "epoch": 85.33, "grad_norm": 0.832936704158783, "learning_rate": 0.0011466666666666667, "loss": 1.2021, "step": 23040 }, { "epoch": 85.37, "grad_norm": 0.8773447871208191, "learning_rate": 0.0011462962962962963, "loss": 1.2186, "step": 23050 }, { "epoch": 85.41, "grad_norm": 0.89570552110672, "learning_rate": 0.001145925925925926, "loss": 1.2462, "step": 23060 }, { "epoch": 85.44, "grad_norm": 0.8472298979759216, "learning_rate": 0.0011455555555555557, "loss": 1.225, "step": 23070 }, { "epoch": 85.48, "grad_norm": 0.8481177687644958, "learning_rate": 0.001145185185185185, "loss": 1.261, "step": 23080 }, { "epoch": 85.52, "grad_norm": 0.8076756596565247, "learning_rate": 0.0011448148148148147, "loss": 1.2411, "step": 23090 }, { "epoch": 85.56, "grad_norm": 0.8581802845001221, "learning_rate": 0.0011444444444444445, "loss": 1.2508, "step": 23100 }, { "epoch": 85.59, "grad_norm": 0.8750701546669006, "learning_rate": 0.001144074074074074, "loss": 1.24, "step": 23110 }, { "epoch": 85.63, "grad_norm": 0.9328594207763672, "learning_rate": 0.0011437037037037036, "loss": 1.2903, "step": 23120 }, { "epoch": 85.67, "grad_norm": 0.850226104259491, "learning_rate": 0.0011433333333333334, "loss": 1.2963, "step": 23130 }, { "epoch": 85.7, "grad_norm": 0.81463223695755, "learning_rate": 0.001142962962962963, "loss": 1.3, "step": 23140 }, { "epoch": 85.74, "grad_norm": 0.7953822016716003, "learning_rate": 0.0011425925925925926, "loss": 1.2952, "step": 23150 }, { "epoch": 85.78, "grad_norm": 0.8868482112884521, "learning_rate": 0.0011422222222222222, "loss": 1.3008, "step": 23160 }, { "epoch": 85.81, "grad_norm": 0.8748762607574463, "learning_rate": 0.001141851851851852, "loss": 1.3137, "step": 23170 }, { "epoch": 85.85, "grad_norm": 0.8665784001350403, "learning_rate": 0.0011414814814814816, "loss": 1.3211, "step": 23180 }, { "epoch": 85.89, "grad_norm": 0.8576427698135376, "learning_rate": 0.0011411111111111112, "loss": 1.3432, "step": 23190 }, { "epoch": 85.93, "grad_norm": 0.8850252032279968, "learning_rate": 0.0011407407407407408, "loss": 1.3025, "step": 23200 }, { "epoch": 85.96, "grad_norm": 0.8282588720321655, "learning_rate": 0.0011403703703703706, "loss": 1.3204, "step": 23210 }, { "epoch": 86.0, "grad_norm": 1.6549211740493774, "learning_rate": 0.00114, "loss": 1.3311, "step": 23220 }, { "epoch": 86.04, "grad_norm": 0.8455803394317627, "learning_rate": 0.0011396296296296295, "loss": 1.1147, "step": 23230 }, { "epoch": 86.07, "grad_norm": 0.8316532969474792, "learning_rate": 0.0011392592592592591, "loss": 1.0876, "step": 23240 }, { "epoch": 86.11, "grad_norm": 0.8999677300453186, "learning_rate": 0.001138888888888889, "loss": 1.1153, "step": 23250 }, { "epoch": 86.15, "grad_norm": 0.8368595242500305, "learning_rate": 0.0011385185185185185, "loss": 1.1155, "step": 23260 }, { "epoch": 86.19, "grad_norm": 0.8152023553848267, "learning_rate": 0.0011381481481481481, "loss": 1.1636, "step": 23270 }, { "epoch": 86.22, "grad_norm": 0.851401150226593, "learning_rate": 0.0011377777777777777, "loss": 1.1503, "step": 23280 }, { "epoch": 86.26, "grad_norm": 0.8852841854095459, "learning_rate": 0.0011374074074074075, "loss": 1.1782, "step": 23290 }, { "epoch": 86.3, "grad_norm": 0.8195506930351257, "learning_rate": 0.001137037037037037, "loss": 1.1948, "step": 23300 }, { "epoch": 86.33, "grad_norm": 0.9015066623687744, "learning_rate": 0.0011366666666666667, "loss": 1.2219, "step": 23310 }, { "epoch": 86.37, "grad_norm": 0.9026146531105042, "learning_rate": 0.0011362962962962965, "loss": 1.2189, "step": 23320 }, { "epoch": 86.41, "grad_norm": 0.8678985238075256, "learning_rate": 0.001135925925925926, "loss": 1.1943, "step": 23330 }, { "epoch": 86.44, "grad_norm": 0.7977427244186401, "learning_rate": 0.0011355555555555557, "loss": 1.2379, "step": 23340 }, { "epoch": 86.48, "grad_norm": 0.8538101315498352, "learning_rate": 0.001135185185185185, "loss": 1.2376, "step": 23350 }, { "epoch": 86.52, "grad_norm": 0.8959237337112427, "learning_rate": 0.0011348148148148148, "loss": 1.2397, "step": 23360 }, { "epoch": 86.56, "grad_norm": 0.7996075749397278, "learning_rate": 0.0011344444444444444, "loss": 1.2342, "step": 23370 }, { "epoch": 86.59, "grad_norm": 0.8509365320205688, "learning_rate": 0.001134074074074074, "loss": 1.2578, "step": 23380 }, { "epoch": 86.63, "grad_norm": 0.8598107099533081, "learning_rate": 0.0011337037037037036, "loss": 1.2738, "step": 23390 }, { "epoch": 86.67, "grad_norm": 0.830723762512207, "learning_rate": 0.0011333333333333334, "loss": 1.2829, "step": 23400 }, { "epoch": 86.7, "grad_norm": 0.8874588012695312, "learning_rate": 0.001132962962962963, "loss": 1.297, "step": 23410 }, { "epoch": 86.74, "grad_norm": 0.8866403698921204, "learning_rate": 0.0011325925925925926, "loss": 1.273, "step": 23420 }, { "epoch": 86.78, "grad_norm": 0.8580409288406372, "learning_rate": 0.0011322222222222222, "loss": 1.2863, "step": 23430 }, { "epoch": 86.81, "grad_norm": 0.8711990714073181, "learning_rate": 0.001131851851851852, "loss": 1.3363, "step": 23440 }, { "epoch": 86.85, "grad_norm": 0.876951277256012, "learning_rate": 0.0011314814814814816, "loss": 1.2952, "step": 23450 }, { "epoch": 86.89, "grad_norm": 0.8068513870239258, "learning_rate": 0.0011311111111111112, "loss": 1.2992, "step": 23460 }, { "epoch": 86.93, "grad_norm": 0.8394823670387268, "learning_rate": 0.0011307407407407408, "loss": 1.3135, "step": 23470 }, { "epoch": 86.96, "grad_norm": 0.8134591579437256, "learning_rate": 0.0011303703703703706, "loss": 1.3362, "step": 23480 }, { "epoch": 87.0, "grad_norm": 1.8034298419952393, "learning_rate": 0.00113, "loss": 1.3274, "step": 23490 }, { "epoch": 87.04, "grad_norm": 0.7800376415252686, "learning_rate": 0.0011296296296296295, "loss": 1.095, "step": 23500 }, { "epoch": 87.07, "grad_norm": 0.7570058703422546, "learning_rate": 0.001129259259259259, "loss": 1.0836, "step": 23510 }, { "epoch": 87.11, "grad_norm": 0.8313126564025879, "learning_rate": 0.001128888888888889, "loss": 1.1297, "step": 23520 }, { "epoch": 87.15, "grad_norm": 0.8045477271080017, "learning_rate": 0.0011285185185185185, "loss": 1.119, "step": 23530 }, { "epoch": 87.19, "grad_norm": 0.8991451859474182, "learning_rate": 0.001128148148148148, "loss": 1.1516, "step": 23540 }, { "epoch": 87.22, "grad_norm": 0.8472118377685547, "learning_rate": 0.001127777777777778, "loss": 1.1534, "step": 23550 }, { "epoch": 87.26, "grad_norm": 0.8414618372917175, "learning_rate": 0.0011274074074074075, "loss": 1.1633, "step": 23560 }, { "epoch": 87.3, "grad_norm": 0.820631206035614, "learning_rate": 0.001127037037037037, "loss": 1.1679, "step": 23570 }, { "epoch": 87.33, "grad_norm": 0.8632519841194153, "learning_rate": 0.0011266666666666667, "loss": 1.2005, "step": 23580 }, { "epoch": 87.37, "grad_norm": 0.896757185459137, "learning_rate": 0.0011262962962962965, "loss": 1.201, "step": 23590 }, { "epoch": 87.41, "grad_norm": 0.815828263759613, "learning_rate": 0.001125925925925926, "loss": 1.236, "step": 23600 }, { "epoch": 87.44, "grad_norm": 0.8946236371994019, "learning_rate": 0.0011255555555555556, "loss": 1.2244, "step": 23610 }, { "epoch": 87.48, "grad_norm": 0.8468908667564392, "learning_rate": 0.0011251851851851852, "loss": 1.237, "step": 23620 }, { "epoch": 87.52, "grad_norm": 0.8856787085533142, "learning_rate": 0.0011248148148148148, "loss": 1.2426, "step": 23630 }, { "epoch": 87.56, "grad_norm": 0.9098236560821533, "learning_rate": 0.0011244444444444444, "loss": 1.2613, "step": 23640 }, { "epoch": 87.59, "grad_norm": 0.8589096665382385, "learning_rate": 0.001124074074074074, "loss": 1.2704, "step": 23650 }, { "epoch": 87.63, "grad_norm": 0.9214912056922913, "learning_rate": 0.0011237037037037036, "loss": 1.243, "step": 23660 }, { "epoch": 87.67, "grad_norm": 0.8244339227676392, "learning_rate": 0.0011233333333333334, "loss": 1.2651, "step": 23670 }, { "epoch": 87.7, "grad_norm": 0.892946183681488, "learning_rate": 0.001122962962962963, "loss": 1.3073, "step": 23680 }, { "epoch": 87.74, "grad_norm": 0.9119607210159302, "learning_rate": 0.0011225925925925926, "loss": 1.2674, "step": 23690 }, { "epoch": 87.78, "grad_norm": 0.7816550731658936, "learning_rate": 0.0011222222222222222, "loss": 1.2668, "step": 23700 }, { "epoch": 87.81, "grad_norm": 0.8281494379043579, "learning_rate": 0.001121851851851852, "loss": 1.2764, "step": 23710 }, { "epoch": 87.85, "grad_norm": 0.8064820170402527, "learning_rate": 0.0011214814814814815, "loss": 1.3038, "step": 23720 }, { "epoch": 87.89, "grad_norm": 0.8691319823265076, "learning_rate": 0.0011211111111111111, "loss": 1.3299, "step": 23730 }, { "epoch": 87.93, "grad_norm": 0.8435444831848145, "learning_rate": 0.001120740740740741, "loss": 1.292, "step": 23740 }, { "epoch": 87.96, "grad_norm": 0.8803039789199829, "learning_rate": 0.0011203703703703705, "loss": 1.3089, "step": 23750 }, { "epoch": 88.0, "grad_norm": 1.5175323486328125, "learning_rate": 0.0011200000000000001, "loss": 1.3126, "step": 23760 }, { "epoch": 88.04, "grad_norm": 0.807668149471283, "learning_rate": 0.0011196296296296295, "loss": 1.0896, "step": 23770 }, { "epoch": 88.07, "grad_norm": 0.8427616953849792, "learning_rate": 0.0011192592592592593, "loss": 1.0739, "step": 23780 }, { "epoch": 88.11, "grad_norm": 0.8259497880935669, "learning_rate": 0.0011188888888888889, "loss": 1.0804, "step": 23790 }, { "epoch": 88.15, "grad_norm": 0.8123213052749634, "learning_rate": 0.0011185185185185185, "loss": 1.1136, "step": 23800 }, { "epoch": 88.19, "grad_norm": 0.8309470415115356, "learning_rate": 0.001118148148148148, "loss": 1.109, "step": 23810 }, { "epoch": 88.22, "grad_norm": 0.7814146876335144, "learning_rate": 0.0011177777777777779, "loss": 1.1466, "step": 23820 }, { "epoch": 88.26, "grad_norm": 0.9524590969085693, "learning_rate": 0.0011174074074074075, "loss": 1.1338, "step": 23830 }, { "epoch": 88.3, "grad_norm": 0.8634507656097412, "learning_rate": 0.001117037037037037, "loss": 1.183, "step": 23840 }, { "epoch": 88.33, "grad_norm": 0.7970640659332275, "learning_rate": 0.0011166666666666666, "loss": 1.2017, "step": 23850 }, { "epoch": 88.37, "grad_norm": 0.8886352777481079, "learning_rate": 0.0011162962962962964, "loss": 1.1993, "step": 23860 }, { "epoch": 88.41, "grad_norm": 0.8125959038734436, "learning_rate": 0.001115925925925926, "loss": 1.1988, "step": 23870 }, { "epoch": 88.44, "grad_norm": 0.8432615995407104, "learning_rate": 0.0011155555555555556, "loss": 1.213, "step": 23880 }, { "epoch": 88.48, "grad_norm": 0.8891323804855347, "learning_rate": 0.0011151851851851852, "loss": 1.2315, "step": 23890 }, { "epoch": 88.52, "grad_norm": 0.8381022214889526, "learning_rate": 0.001114814814814815, "loss": 1.2435, "step": 23900 }, { "epoch": 88.56, "grad_norm": 0.8534558415412903, "learning_rate": 0.0011144444444444444, "loss": 1.2576, "step": 23910 }, { "epoch": 88.59, "grad_norm": 0.8542389273643494, "learning_rate": 0.001114074074074074, "loss": 1.2542, "step": 23920 }, { "epoch": 88.63, "grad_norm": 0.8446402549743652, "learning_rate": 0.0011137037037037036, "loss": 1.2542, "step": 23930 }, { "epoch": 88.67, "grad_norm": 0.9329958558082581, "learning_rate": 0.0011133333333333334, "loss": 1.2625, "step": 23940 }, { "epoch": 88.7, "grad_norm": 0.8419949412345886, "learning_rate": 0.001112962962962963, "loss": 1.2744, "step": 23950 }, { "epoch": 88.74, "grad_norm": 0.8235114812850952, "learning_rate": 0.0011125925925925925, "loss": 1.2628, "step": 23960 }, { "epoch": 88.78, "grad_norm": 0.8560745120048523, "learning_rate": 0.0011122222222222223, "loss": 1.2809, "step": 23970 }, { "epoch": 88.81, "grad_norm": 0.8349394202232361, "learning_rate": 0.001111851851851852, "loss": 1.2943, "step": 23980 }, { "epoch": 88.85, "grad_norm": 0.9164478778839111, "learning_rate": 0.0011114814814814815, "loss": 1.3001, "step": 23990 }, { "epoch": 88.89, "grad_norm": 0.9206733107566833, "learning_rate": 0.0011111111111111111, "loss": 1.2985, "step": 24000 }, { "epoch": 88.93, "grad_norm": 0.8984618186950684, "learning_rate": 0.001110740740740741, "loss": 1.2869, "step": 24010 }, { "epoch": 88.96, "grad_norm": 0.8212258219718933, "learning_rate": 0.0011103703703703705, "loss": 1.3071, "step": 24020 }, { "epoch": 89.0, "grad_norm": 1.393999457359314, "learning_rate": 0.00111, "loss": 1.3113, "step": 24030 }, { "epoch": 89.04, "grad_norm": 0.8700305223464966, "learning_rate": 0.0011096296296296295, "loss": 1.1065, "step": 24040 }, { "epoch": 89.07, "grad_norm": 0.769532322883606, "learning_rate": 0.0011092592592592593, "loss": 1.0946, "step": 24050 }, { "epoch": 89.11, "grad_norm": 0.8377566933631897, "learning_rate": 0.0011088888888888889, "loss": 1.0966, "step": 24060 }, { "epoch": 89.15, "grad_norm": 0.8229069113731384, "learning_rate": 0.0011085185185185184, "loss": 1.1145, "step": 24070 }, { "epoch": 89.19, "grad_norm": 0.8501279354095459, "learning_rate": 0.001108148148148148, "loss": 1.11, "step": 24080 }, { "epoch": 89.22, "grad_norm": 0.8882164359092712, "learning_rate": 0.0011077777777777778, "loss": 1.1364, "step": 24090 }, { "epoch": 89.26, "grad_norm": 0.8216584324836731, "learning_rate": 0.0011074074074074074, "loss": 1.1462, "step": 24100 }, { "epoch": 89.3, "grad_norm": 0.944375216960907, "learning_rate": 0.001107037037037037, "loss": 1.166, "step": 24110 }, { "epoch": 89.33, "grad_norm": 0.9312400221824646, "learning_rate": 0.0011066666666666666, "loss": 1.2047, "step": 24120 }, { "epoch": 89.37, "grad_norm": 0.8870893716812134, "learning_rate": 0.0011062962962962964, "loss": 1.1928, "step": 24130 }, { "epoch": 89.41, "grad_norm": 0.8170095682144165, "learning_rate": 0.001105925925925926, "loss": 1.1968, "step": 24140 }, { "epoch": 89.44, "grad_norm": 0.8027873635292053, "learning_rate": 0.0011055555555555556, "loss": 1.1993, "step": 24150 }, { "epoch": 89.48, "grad_norm": 0.8569116592407227, "learning_rate": 0.0011051851851851854, "loss": 1.2012, "step": 24160 }, { "epoch": 89.52, "grad_norm": 0.8904064893722534, "learning_rate": 0.001104814814814815, "loss": 1.2071, "step": 24170 }, { "epoch": 89.56, "grad_norm": 0.9111688733100891, "learning_rate": 0.0011044444444444444, "loss": 1.2494, "step": 24180 }, { "epoch": 89.59, "grad_norm": 0.8626366853713989, "learning_rate": 0.001104074074074074, "loss": 1.2422, "step": 24190 }, { "epoch": 89.63, "grad_norm": 0.8405423164367676, "learning_rate": 0.0011037037037037037, "loss": 1.2314, "step": 24200 }, { "epoch": 89.67, "grad_norm": 0.9029109477996826, "learning_rate": 0.0011033333333333333, "loss": 1.2691, "step": 24210 }, { "epoch": 89.7, "grad_norm": 0.8996674418449402, "learning_rate": 0.001102962962962963, "loss": 1.2561, "step": 24220 }, { "epoch": 89.74, "grad_norm": 0.9102544784545898, "learning_rate": 0.0011025925925925925, "loss": 1.26, "step": 24230 }, { "epoch": 89.78, "grad_norm": 0.9108865857124329, "learning_rate": 0.0011022222222222223, "loss": 1.275, "step": 24240 }, { "epoch": 89.81, "grad_norm": 0.8576236963272095, "learning_rate": 0.001101851851851852, "loss": 1.2678, "step": 24250 }, { "epoch": 89.85, "grad_norm": 0.8501005172729492, "learning_rate": 0.0011014814814814815, "loss": 1.2909, "step": 24260 }, { "epoch": 89.89, "grad_norm": 0.8619106411933899, "learning_rate": 0.001101111111111111, "loss": 1.2885, "step": 24270 }, { "epoch": 89.93, "grad_norm": 0.8372892141342163, "learning_rate": 0.0011007407407407409, "loss": 1.3103, "step": 24280 }, { "epoch": 89.96, "grad_norm": 0.860112190246582, "learning_rate": 0.0011003703703703705, "loss": 1.2871, "step": 24290 }, { "epoch": 90.0, "grad_norm": 1.5859699249267578, "learning_rate": 0.0011, "loss": 1.3098, "step": 24300 }, { "epoch": 90.04, "grad_norm": 0.8784357309341431, "learning_rate": 0.0010996296296296297, "loss": 1.0776, "step": 24310 }, { "epoch": 90.07, "grad_norm": 0.8911356329917908, "learning_rate": 0.0010992592592592592, "loss": 1.1031, "step": 24320 }, { "epoch": 90.11, "grad_norm": 0.7874282002449036, "learning_rate": 0.0010988888888888888, "loss": 1.1044, "step": 24330 }, { "epoch": 90.15, "grad_norm": 0.8270366787910461, "learning_rate": 0.0010985185185185184, "loss": 1.0937, "step": 24340 }, { "epoch": 90.19, "grad_norm": 0.8908728361129761, "learning_rate": 0.0010981481481481482, "loss": 1.1221, "step": 24350 }, { "epoch": 90.22, "grad_norm": 0.9187735915184021, "learning_rate": 0.0010977777777777778, "loss": 1.141, "step": 24360 }, { "epoch": 90.26, "grad_norm": 0.9003711342811584, "learning_rate": 0.0010974074074074074, "loss": 1.1368, "step": 24370 }, { "epoch": 90.3, "grad_norm": 0.8380781412124634, "learning_rate": 0.001097037037037037, "loss": 1.149, "step": 24380 }, { "epoch": 90.33, "grad_norm": 0.8692864775657654, "learning_rate": 0.0010966666666666668, "loss": 1.1744, "step": 24390 }, { "epoch": 90.37, "grad_norm": 0.8086166977882385, "learning_rate": 0.0010962962962962964, "loss": 1.1666, "step": 24400 }, { "epoch": 90.41, "grad_norm": 0.9610259532928467, "learning_rate": 0.001095925925925926, "loss": 1.1919, "step": 24410 }, { "epoch": 90.44, "grad_norm": 0.9051281809806824, "learning_rate": 0.0010955555555555556, "loss": 1.206, "step": 24420 }, { "epoch": 90.48, "grad_norm": 0.8761212825775146, "learning_rate": 0.0010951851851851854, "loss": 1.1962, "step": 24430 }, { "epoch": 90.52, "grad_norm": 0.8568070530891418, "learning_rate": 0.001094814814814815, "loss": 1.2054, "step": 24440 }, { "epoch": 90.56, "grad_norm": 0.8986058235168457, "learning_rate": 0.0010944444444444445, "loss": 1.1984, "step": 24450 }, { "epoch": 90.59, "grad_norm": 0.8418546319007874, "learning_rate": 0.001094074074074074, "loss": 1.2403, "step": 24460 }, { "epoch": 90.63, "grad_norm": 0.8225176930427551, "learning_rate": 0.0010937037037037037, "loss": 1.2449, "step": 24470 }, { "epoch": 90.67, "grad_norm": 0.9044831991195679, "learning_rate": 0.0010933333333333333, "loss": 1.2398, "step": 24480 }, { "epoch": 90.7, "grad_norm": 0.9028358459472656, "learning_rate": 0.001092962962962963, "loss": 1.2816, "step": 24490 }, { "epoch": 90.74, "grad_norm": 0.8567687273025513, "learning_rate": 0.0010925925925925925, "loss": 1.244, "step": 24500 }, { "epoch": 90.78, "grad_norm": 0.8195727467536926, "learning_rate": 0.0010922222222222223, "loss": 1.2528, "step": 24510 }, { "epoch": 90.81, "grad_norm": 0.9707487225532532, "learning_rate": 0.0010918518518518519, "loss": 1.2705, "step": 24520 }, { "epoch": 90.85, "grad_norm": 0.894263505935669, "learning_rate": 0.0010914814814814815, "loss": 1.2607, "step": 24530 }, { "epoch": 90.89, "grad_norm": 0.8730370402336121, "learning_rate": 0.001091111111111111, "loss": 1.2851, "step": 24540 }, { "epoch": 90.93, "grad_norm": 0.8939180374145508, "learning_rate": 0.0010907407407407409, "loss": 1.2997, "step": 24550 }, { "epoch": 90.96, "grad_norm": 0.8519564867019653, "learning_rate": 0.0010903703703703705, "loss": 1.296, "step": 24560 }, { "epoch": 91.0, "grad_norm": 1.6609879732131958, "learning_rate": 0.00109, "loss": 1.3057, "step": 24570 }, { "epoch": 91.04, "grad_norm": 0.8076825141906738, "learning_rate": 0.0010896296296296298, "loss": 1.0765, "step": 24580 }, { "epoch": 91.07, "grad_norm": 0.844336748123169, "learning_rate": 0.0010892592592592592, "loss": 1.0612, "step": 24590 }, { "epoch": 91.11, "grad_norm": 0.8776166439056396, "learning_rate": 0.0010888888888888888, "loss": 1.0979, "step": 24600 }, { "epoch": 91.15, "grad_norm": 0.8715381622314453, "learning_rate": 0.0010885185185185184, "loss": 1.1177, "step": 24610 }, { "epoch": 91.19, "grad_norm": 0.8251622319221497, "learning_rate": 0.0010881481481481482, "loss": 1.1033, "step": 24620 }, { "epoch": 91.22, "grad_norm": 0.8432281613349915, "learning_rate": 0.0010877777777777778, "loss": 1.1156, "step": 24630 }, { "epoch": 91.26, "grad_norm": 0.8701592683792114, "learning_rate": 0.0010874074074074074, "loss": 1.1421, "step": 24640 }, { "epoch": 91.3, "grad_norm": 0.9401283860206604, "learning_rate": 0.001087037037037037, "loss": 1.1513, "step": 24650 }, { "epoch": 91.33, "grad_norm": 0.8380709290504456, "learning_rate": 0.0010866666666666668, "loss": 1.1442, "step": 24660 }, { "epoch": 91.37, "grad_norm": 0.8541663289070129, "learning_rate": 0.0010862962962962964, "loss": 1.166, "step": 24670 }, { "epoch": 91.41, "grad_norm": 0.869760274887085, "learning_rate": 0.001085925925925926, "loss": 1.1895, "step": 24680 }, { "epoch": 91.44, "grad_norm": 0.8138124346733093, "learning_rate": 0.0010855555555555555, "loss": 1.2083, "step": 24690 }, { "epoch": 91.48, "grad_norm": 0.9111208915710449, "learning_rate": 0.0010851851851851853, "loss": 1.2269, "step": 24700 }, { "epoch": 91.52, "grad_norm": 0.8547767996788025, "learning_rate": 0.001084814814814815, "loss": 1.1998, "step": 24710 }, { "epoch": 91.56, "grad_norm": 0.8412507772445679, "learning_rate": 0.0010844444444444445, "loss": 1.1929, "step": 24720 }, { "epoch": 91.59, "grad_norm": 0.8674251437187195, "learning_rate": 0.0010840740740740739, "loss": 1.2064, "step": 24730 }, { "epoch": 91.63, "grad_norm": 0.9130268692970276, "learning_rate": 0.0010837037037037037, "loss": 1.2473, "step": 24740 }, { "epoch": 91.67, "grad_norm": 0.9256370067596436, "learning_rate": 0.0010833333333333333, "loss": 1.2374, "step": 24750 }, { "epoch": 91.7, "grad_norm": 0.8963067531585693, "learning_rate": 0.0010829629629629629, "loss": 1.2553, "step": 24760 }, { "epoch": 91.74, "grad_norm": 0.8186632394790649, "learning_rate": 0.0010825925925925927, "loss": 1.2617, "step": 24770 }, { "epoch": 91.78, "grad_norm": 0.8365766406059265, "learning_rate": 0.0010822222222222223, "loss": 1.2506, "step": 24780 }, { "epoch": 91.81, "grad_norm": 0.8591589331626892, "learning_rate": 0.0010818518518518519, "loss": 1.2742, "step": 24790 }, { "epoch": 91.85, "grad_norm": 0.8758793473243713, "learning_rate": 0.0010814814814814814, "loss": 1.2634, "step": 24800 }, { "epoch": 91.89, "grad_norm": 0.8638857007026672, "learning_rate": 0.0010811111111111112, "loss": 1.2615, "step": 24810 }, { "epoch": 91.93, "grad_norm": 0.8489501476287842, "learning_rate": 0.0010807407407407408, "loss": 1.2734, "step": 24820 }, { "epoch": 91.96, "grad_norm": 0.8600893616676331, "learning_rate": 0.0010803703703703704, "loss": 1.29, "step": 24830 }, { "epoch": 92.0, "grad_norm": 1.6491703987121582, "learning_rate": 0.00108, "loss": 1.2892, "step": 24840 }, { "epoch": 92.04, "grad_norm": 0.8379200100898743, "learning_rate": 0.0010796296296296298, "loss": 1.048, "step": 24850 }, { "epoch": 92.07, "grad_norm": 0.9475078582763672, "learning_rate": 0.0010792592592592594, "loss": 1.0562, "step": 24860 }, { "epoch": 92.11, "grad_norm": 0.9267643094062805, "learning_rate": 0.0010788888888888888, "loss": 1.0861, "step": 24870 }, { "epoch": 92.15, "grad_norm": 0.8084042072296143, "learning_rate": 0.0010785185185185184, "loss": 1.0767, "step": 24880 }, { "epoch": 92.19, "grad_norm": 0.8792321681976318, "learning_rate": 0.0010781481481481482, "loss": 1.1191, "step": 24890 }, { "epoch": 92.22, "grad_norm": 0.8480071425437927, "learning_rate": 0.0010777777777777778, "loss": 1.1152, "step": 24900 }, { "epoch": 92.26, "grad_norm": 0.8491085767745972, "learning_rate": 0.0010774074074074074, "loss": 1.1557, "step": 24910 }, { "epoch": 92.3, "grad_norm": 0.9552561640739441, "learning_rate": 0.001077037037037037, "loss": 1.131, "step": 24920 }, { "epoch": 92.33, "grad_norm": 0.9046753644943237, "learning_rate": 0.0010766666666666667, "loss": 1.1554, "step": 24930 }, { "epoch": 92.37, "grad_norm": 0.9224265217781067, "learning_rate": 0.0010762962962962963, "loss": 1.1562, "step": 24940 }, { "epoch": 92.41, "grad_norm": 0.8900438547134399, "learning_rate": 0.001075925925925926, "loss": 1.1793, "step": 24950 }, { "epoch": 92.44, "grad_norm": 0.8760910034179688, "learning_rate": 0.0010755555555555557, "loss": 1.1719, "step": 24960 }, { "epoch": 92.48, "grad_norm": 0.9153395891189575, "learning_rate": 0.0010751851851851853, "loss": 1.185, "step": 24970 }, { "epoch": 92.52, "grad_norm": 0.854253888130188, "learning_rate": 0.001074814814814815, "loss": 1.197, "step": 24980 }, { "epoch": 92.56, "grad_norm": 0.8273108601570129, "learning_rate": 0.0010744444444444445, "loss": 1.201, "step": 24990 }, { "epoch": 92.59, "grad_norm": 0.8812609314918518, "learning_rate": 0.0010740740740740743, "loss": 1.2097, "step": 25000 }, { "epoch": 92.63, "grad_norm": 0.8414286971092224, "learning_rate": 0.0010737037037037037, "loss": 1.2252, "step": 25010 }, { "epoch": 92.67, "grad_norm": 0.8828840851783752, "learning_rate": 0.0010733333333333333, "loss": 1.224, "step": 25020 }, { "epoch": 92.7, "grad_norm": 0.8736853003501892, "learning_rate": 0.0010729629629629628, "loss": 1.2576, "step": 25030 }, { "epoch": 92.74, "grad_norm": 0.8438439965248108, "learning_rate": 0.0010725925925925927, "loss": 1.2458, "step": 25040 }, { "epoch": 92.78, "grad_norm": 0.8953974843025208, "learning_rate": 0.0010722222222222222, "loss": 1.2583, "step": 25050 }, { "epoch": 92.81, "grad_norm": 0.8555558323860168, "learning_rate": 0.0010718518518518518, "loss": 1.2383, "step": 25060 }, { "epoch": 92.85, "grad_norm": 0.8639121651649475, "learning_rate": 0.0010714814814814814, "loss": 1.264, "step": 25070 }, { "epoch": 92.89, "grad_norm": 0.8091748952865601, "learning_rate": 0.0010711111111111112, "loss": 1.2607, "step": 25080 }, { "epoch": 92.93, "grad_norm": 0.8656150102615356, "learning_rate": 0.0010707407407407408, "loss": 1.2705, "step": 25090 }, { "epoch": 92.96, "grad_norm": 0.8873560428619385, "learning_rate": 0.0010703703703703704, "loss": 1.3017, "step": 25100 }, { "epoch": 93.0, "grad_norm": 1.5208163261413574, "learning_rate": 0.00107, "loss": 1.2864, "step": 25110 }, { "epoch": 93.04, "grad_norm": 0.8099291324615479, "learning_rate": 0.0010696296296296298, "loss": 1.0601, "step": 25120 }, { "epoch": 93.07, "grad_norm": 0.8671590685844421, "learning_rate": 0.0010692592592592594, "loss": 1.0655, "step": 25130 }, { "epoch": 93.11, "grad_norm": 0.8976333737373352, "learning_rate": 0.001068888888888889, "loss": 1.0471, "step": 25140 }, { "epoch": 93.15, "grad_norm": 0.8625257015228271, "learning_rate": 0.0010685185185185183, "loss": 1.0606, "step": 25150 }, { "epoch": 93.19, "grad_norm": 0.8490420579910278, "learning_rate": 0.0010681481481481481, "loss": 1.0883, "step": 25160 }, { "epoch": 93.22, "grad_norm": 0.9137793779373169, "learning_rate": 0.0010677777777777777, "loss": 1.1154, "step": 25170 }, { "epoch": 93.26, "grad_norm": 0.8413072228431702, "learning_rate": 0.0010674074074074073, "loss": 1.1194, "step": 25180 }, { "epoch": 93.3, "grad_norm": 0.8381540179252625, "learning_rate": 0.0010670370370370371, "loss": 1.1144, "step": 25190 }, { "epoch": 93.33, "grad_norm": 0.8788276314735413, "learning_rate": 0.0010666666666666667, "loss": 1.1448, "step": 25200 }, { "epoch": 93.37, "grad_norm": 0.8643991947174072, "learning_rate": 0.0010662962962962963, "loss": 1.136, "step": 25210 }, { "epoch": 93.41, "grad_norm": 0.8980141878128052, "learning_rate": 0.001065925925925926, "loss": 1.1751, "step": 25220 }, { "epoch": 93.44, "grad_norm": 0.9232836365699768, "learning_rate": 0.0010655555555555557, "loss": 1.1715, "step": 25230 }, { "epoch": 93.48, "grad_norm": 0.9864527583122253, "learning_rate": 0.0010651851851851853, "loss": 1.17, "step": 25240 }, { "epoch": 93.52, "grad_norm": 0.9622303247451782, "learning_rate": 0.0010648148148148149, "loss": 1.2191, "step": 25250 }, { "epoch": 93.56, "grad_norm": 0.9083154797554016, "learning_rate": 0.0010644444444444445, "loss": 1.2056, "step": 25260 }, { "epoch": 93.59, "grad_norm": 0.8920983076095581, "learning_rate": 0.0010640740740740743, "loss": 1.2426, "step": 25270 }, { "epoch": 93.63, "grad_norm": 0.8751955628395081, "learning_rate": 0.0010637037037037036, "loss": 1.2283, "step": 25280 }, { "epoch": 93.67, "grad_norm": 0.9048447608947754, "learning_rate": 0.0010633333333333332, "loss": 1.2235, "step": 25290 }, { "epoch": 93.7, "grad_norm": 0.8830949068069458, "learning_rate": 0.0010629629629629628, "loss": 1.237, "step": 25300 }, { "epoch": 93.74, "grad_norm": 0.9014407992362976, "learning_rate": 0.0010625925925925926, "loss": 1.2593, "step": 25310 }, { "epoch": 93.78, "grad_norm": 0.8940639495849609, "learning_rate": 0.0010622222222222222, "loss": 1.2289, "step": 25320 }, { "epoch": 93.81, "grad_norm": 0.8975448608398438, "learning_rate": 0.0010618518518518518, "loss": 1.2498, "step": 25330 }, { "epoch": 93.85, "grad_norm": 0.8936808705329895, "learning_rate": 0.0010614814814814814, "loss": 1.2903, "step": 25340 }, { "epoch": 93.89, "grad_norm": 0.9473136067390442, "learning_rate": 0.0010611111111111112, "loss": 1.2497, "step": 25350 }, { "epoch": 93.93, "grad_norm": 0.892784595489502, "learning_rate": 0.0010607407407407408, "loss": 1.2611, "step": 25360 }, { "epoch": 93.96, "grad_norm": 0.8802827000617981, "learning_rate": 0.0010603703703703704, "loss": 1.2424, "step": 25370 }, { "epoch": 94.0, "grad_norm": 1.943390130996704, "learning_rate": 0.0010600000000000002, "loss": 1.2733, "step": 25380 }, { "epoch": 94.04, "grad_norm": 0.859840989112854, "learning_rate": 0.0010596296296296298, "loss": 1.0207, "step": 25390 }, { "epoch": 94.07, "grad_norm": 0.8180606961250305, "learning_rate": 0.0010592592592592594, "loss": 1.0581, "step": 25400 }, { "epoch": 94.11, "grad_norm": 0.8518406748771667, "learning_rate": 0.001058888888888889, "loss": 1.095, "step": 25410 }, { "epoch": 94.15, "grad_norm": 0.8576958179473877, "learning_rate": 0.0010585185185185185, "loss": 1.0672, "step": 25420 }, { "epoch": 94.19, "grad_norm": 0.8683270812034607, "learning_rate": 0.0010581481481481481, "loss": 1.0929, "step": 25430 }, { "epoch": 94.22, "grad_norm": 0.8180840611457825, "learning_rate": 0.0010577777777777777, "loss": 1.1359, "step": 25440 }, { "epoch": 94.26, "grad_norm": 0.8479984402656555, "learning_rate": 0.0010574074074074073, "loss": 1.1059, "step": 25450 }, { "epoch": 94.3, "grad_norm": 0.8408753871917725, "learning_rate": 0.001057037037037037, "loss": 1.0997, "step": 25460 }, { "epoch": 94.33, "grad_norm": 0.8361555337905884, "learning_rate": 0.0010566666666666667, "loss": 1.1254, "step": 25470 }, { "epoch": 94.37, "grad_norm": 0.8891112804412842, "learning_rate": 0.0010562962962962963, "loss": 1.1493, "step": 25480 }, { "epoch": 94.41, "grad_norm": 0.9582691788673401, "learning_rate": 0.0010559259259259259, "loss": 1.1596, "step": 25490 }, { "epoch": 94.44, "grad_norm": 0.9022079706192017, "learning_rate": 0.0010555555555555557, "loss": 1.16, "step": 25500 }, { "epoch": 94.48, "grad_norm": 0.8625105619430542, "learning_rate": 0.0010551851851851853, "loss": 1.1839, "step": 25510 }, { "epoch": 94.52, "grad_norm": 0.8581701517105103, "learning_rate": 0.0010548148148148149, "loss": 1.1875, "step": 25520 }, { "epoch": 94.56, "grad_norm": 0.9323828220367432, "learning_rate": 0.0010544444444444444, "loss": 1.2022, "step": 25530 }, { "epoch": 94.59, "grad_norm": 0.8733723163604736, "learning_rate": 0.0010540740740740742, "loss": 1.2044, "step": 25540 }, { "epoch": 94.63, "grad_norm": 0.8635187745094299, "learning_rate": 0.0010537037037037038, "loss": 1.1975, "step": 25550 }, { "epoch": 94.67, "grad_norm": 0.8991508483886719, "learning_rate": 0.0010533333333333332, "loss": 1.1968, "step": 25560 }, { "epoch": 94.7, "grad_norm": 0.9222977757453918, "learning_rate": 0.0010529629629629628, "loss": 1.2285, "step": 25570 }, { "epoch": 94.74, "grad_norm": 0.8603134155273438, "learning_rate": 0.0010525925925925926, "loss": 1.2376, "step": 25580 }, { "epoch": 94.78, "grad_norm": 0.9961068630218506, "learning_rate": 0.0010522222222222222, "loss": 1.2233, "step": 25590 }, { "epoch": 94.81, "grad_norm": 0.8663851022720337, "learning_rate": 0.0010518518518518518, "loss": 1.2645, "step": 25600 }, { "epoch": 94.85, "grad_norm": 0.9062612056732178, "learning_rate": 0.0010514814814814816, "loss": 1.2363, "step": 25610 }, { "epoch": 94.89, "grad_norm": 0.9243030548095703, "learning_rate": 0.0010511111111111112, "loss": 1.2508, "step": 25620 }, { "epoch": 94.93, "grad_norm": 0.8614787459373474, "learning_rate": 0.0010507407407407408, "loss": 1.2322, "step": 25630 }, { "epoch": 94.96, "grad_norm": 0.9229607582092285, "learning_rate": 0.0010503703703703703, "loss": 1.2798, "step": 25640 }, { "epoch": 95.0, "grad_norm": 1.8464992046356201, "learning_rate": 0.0010500000000000002, "loss": 1.2714, "step": 25650 }, { "epoch": 95.04, "grad_norm": 0.7462485432624817, "learning_rate": 0.0010496296296296297, "loss": 1.0297, "step": 25660 }, { "epoch": 95.07, "grad_norm": 0.8666638731956482, "learning_rate": 0.0010492592592592593, "loss": 1.0276, "step": 25670 }, { "epoch": 95.11, "grad_norm": 0.9033090472221375, "learning_rate": 0.001048888888888889, "loss": 1.061, "step": 25680 }, { "epoch": 95.15, "grad_norm": 0.8938956260681152, "learning_rate": 0.0010485185185185187, "loss": 1.077, "step": 25690 }, { "epoch": 95.19, "grad_norm": 0.9139565229415894, "learning_rate": 0.001048148148148148, "loss": 1.0753, "step": 25700 }, { "epoch": 95.22, "grad_norm": 0.8628860712051392, "learning_rate": 0.0010477777777777777, "loss": 1.085, "step": 25710 }, { "epoch": 95.26, "grad_norm": 0.908026397228241, "learning_rate": 0.0010474074074074073, "loss": 1.1087, "step": 25720 }, { "epoch": 95.3, "grad_norm": 0.8440184593200684, "learning_rate": 0.001047037037037037, "loss": 1.1133, "step": 25730 }, { "epoch": 95.33, "grad_norm": 0.9705326557159424, "learning_rate": 0.0010466666666666667, "loss": 1.135, "step": 25740 }, { "epoch": 95.37, "grad_norm": 0.8961520195007324, "learning_rate": 0.0010462962962962963, "loss": 1.161, "step": 25750 }, { "epoch": 95.41, "grad_norm": 0.8462703227996826, "learning_rate": 0.0010459259259259258, "loss": 1.1503, "step": 25760 }, { "epoch": 95.44, "grad_norm": 0.8871561288833618, "learning_rate": 0.0010455555555555556, "loss": 1.1721, "step": 25770 }, { "epoch": 95.48, "grad_norm": 0.9002634882926941, "learning_rate": 0.0010451851851851852, "loss": 1.169, "step": 25780 }, { "epoch": 95.52, "grad_norm": 0.9232921600341797, "learning_rate": 0.0010448148148148148, "loss": 1.1717, "step": 25790 }, { "epoch": 95.56, "grad_norm": 0.8582503795623779, "learning_rate": 0.0010444444444444446, "loss": 1.1827, "step": 25800 }, { "epoch": 95.59, "grad_norm": 0.8893172144889832, "learning_rate": 0.0010440740740740742, "loss": 1.2161, "step": 25810 }, { "epoch": 95.63, "grad_norm": 0.9103512167930603, "learning_rate": 0.0010437037037037038, "loss": 1.2138, "step": 25820 }, { "epoch": 95.67, "grad_norm": 0.8459279537200928, "learning_rate": 0.0010433333333333332, "loss": 1.2184, "step": 25830 }, { "epoch": 95.7, "grad_norm": 0.9495593309402466, "learning_rate": 0.001042962962962963, "loss": 1.1986, "step": 25840 }, { "epoch": 95.74, "grad_norm": 0.8854154348373413, "learning_rate": 0.0010425925925925926, "loss": 1.2264, "step": 25850 }, { "epoch": 95.78, "grad_norm": 0.9533488154411316, "learning_rate": 0.0010422222222222222, "loss": 1.2086, "step": 25860 }, { "epoch": 95.81, "grad_norm": 0.9248517751693726, "learning_rate": 0.0010418518518518517, "loss": 1.2536, "step": 25870 }, { "epoch": 95.85, "grad_norm": 0.8576090931892395, "learning_rate": 0.0010414814814814816, "loss": 1.2604, "step": 25880 }, { "epoch": 95.89, "grad_norm": 0.89030522108078, "learning_rate": 0.0010411111111111111, "loss": 1.2545, "step": 25890 }, { "epoch": 95.93, "grad_norm": 0.9573636651039124, "learning_rate": 0.0010407407407407407, "loss": 1.2565, "step": 25900 }, { "epoch": 95.96, "grad_norm": 0.8615505695343018, "learning_rate": 0.0010403703703703703, "loss": 1.2782, "step": 25910 }, { "epoch": 96.0, "grad_norm": 1.4731565713882446, "learning_rate": 0.0010400000000000001, "loss": 1.2516, "step": 25920 }, { "epoch": 96.04, "grad_norm": 0.8369557857513428, "learning_rate": 0.0010396296296296297, "loss": 1.0401, "step": 25930 }, { "epoch": 96.07, "grad_norm": 0.9119203090667725, "learning_rate": 0.0010392592592592593, "loss": 1.0128, "step": 25940 }, { "epoch": 96.11, "grad_norm": 0.8330380320549011, "learning_rate": 0.0010388888888888889, "loss": 1.0457, "step": 25950 }, { "epoch": 96.15, "grad_norm": 0.901683509349823, "learning_rate": 0.0010385185185185187, "loss": 1.0575, "step": 25960 }, { "epoch": 96.19, "grad_norm": 0.8877632021903992, "learning_rate": 0.001038148148148148, "loss": 1.0487, "step": 25970 }, { "epoch": 96.22, "grad_norm": 0.9112820625305176, "learning_rate": 0.0010377777777777777, "loss": 1.0781, "step": 25980 }, { "epoch": 96.26, "grad_norm": 0.9182432889938354, "learning_rate": 0.0010374074074074075, "loss": 1.1201, "step": 25990 }, { "epoch": 96.3, "grad_norm": 0.9750530123710632, "learning_rate": 0.001037037037037037, "loss": 1.1322, "step": 26000 }, { "epoch": 96.33, "grad_norm": 0.9546353220939636, "learning_rate": 0.0010366666666666666, "loss": 1.124, "step": 26010 }, { "epoch": 96.37, "grad_norm": 0.8506656885147095, "learning_rate": 0.0010362962962962962, "loss": 1.1454, "step": 26020 }, { "epoch": 96.41, "grad_norm": 0.8744280338287354, "learning_rate": 0.001035925925925926, "loss": 1.1397, "step": 26030 }, { "epoch": 96.44, "grad_norm": 0.8825117349624634, "learning_rate": 0.0010355555555555556, "loss": 1.1655, "step": 26040 }, { "epoch": 96.48, "grad_norm": 0.9138257503509521, "learning_rate": 0.0010351851851851852, "loss": 1.1494, "step": 26050 }, { "epoch": 96.52, "grad_norm": 0.91966313123703, "learning_rate": 0.0010348148148148148, "loss": 1.1891, "step": 26060 }, { "epoch": 96.56, "grad_norm": 0.9052321314811707, "learning_rate": 0.0010344444444444446, "loss": 1.1694, "step": 26070 }, { "epoch": 96.59, "grad_norm": 0.87230384349823, "learning_rate": 0.0010340740740740742, "loss": 1.1635, "step": 26080 }, { "epoch": 96.63, "grad_norm": 0.8561148643493652, "learning_rate": 0.0010337037037037038, "loss": 1.1816, "step": 26090 }, { "epoch": 96.67, "grad_norm": 0.9230003952980042, "learning_rate": 0.0010333333333333334, "loss": 1.1944, "step": 26100 }, { "epoch": 96.7, "grad_norm": 0.9017607569694519, "learning_rate": 0.001032962962962963, "loss": 1.1991, "step": 26110 }, { "epoch": 96.74, "grad_norm": 0.9002452492713928, "learning_rate": 0.0010325925925925925, "loss": 1.2403, "step": 26120 }, { "epoch": 96.78, "grad_norm": 0.9941157102584839, "learning_rate": 0.0010322222222222221, "loss": 1.2229, "step": 26130 }, { "epoch": 96.81, "grad_norm": 0.9742600917816162, "learning_rate": 0.0010318518518518517, "loss": 1.2228, "step": 26140 }, { "epoch": 96.85, "grad_norm": 0.9086616635322571, "learning_rate": 0.0010314814814814815, "loss": 1.2287, "step": 26150 }, { "epoch": 96.89, "grad_norm": 0.9041900634765625, "learning_rate": 0.0010311111111111111, "loss": 1.2324, "step": 26160 }, { "epoch": 96.93, "grad_norm": 0.9248058199882507, "learning_rate": 0.0010307407407407407, "loss": 1.269, "step": 26170 }, { "epoch": 96.96, "grad_norm": 0.928729772567749, "learning_rate": 0.0010303703703703703, "loss": 1.2477, "step": 26180 }, { "epoch": 97.0, "grad_norm": 1.5879347324371338, "learning_rate": 0.00103, "loss": 1.2538, "step": 26190 }, { "epoch": 97.04, "grad_norm": 0.9306336045265198, "learning_rate": 0.0010296296296296297, "loss": 1.0265, "step": 26200 }, { "epoch": 97.07, "grad_norm": 0.8678556680679321, "learning_rate": 0.0010292592592592593, "loss": 1.0392, "step": 26210 }, { "epoch": 97.11, "grad_norm": 0.8407970070838928, "learning_rate": 0.001028888888888889, "loss": 1.0319, "step": 26220 }, { "epoch": 97.15, "grad_norm": 0.9070879817008972, "learning_rate": 0.0010285185185185187, "loss": 1.0491, "step": 26230 }, { "epoch": 97.19, "grad_norm": 0.9178571701049805, "learning_rate": 0.0010281481481481483, "loss": 1.0906, "step": 26240 }, { "epoch": 97.22, "grad_norm": 0.8487735986709595, "learning_rate": 0.0010277777777777776, "loss": 1.0685, "step": 26250 }, { "epoch": 97.26, "grad_norm": 0.9132791757583618, "learning_rate": 0.0010274074074074074, "loss": 1.0635, "step": 26260 }, { "epoch": 97.3, "grad_norm": 0.838406503200531, "learning_rate": 0.001027037037037037, "loss": 1.111, "step": 26270 }, { "epoch": 97.33, "grad_norm": 0.9344068169593811, "learning_rate": 0.0010266666666666666, "loss": 1.1066, "step": 26280 }, { "epoch": 97.37, "grad_norm": 0.9382205605506897, "learning_rate": 0.0010262962962962962, "loss": 1.1355, "step": 26290 }, { "epoch": 97.41, "grad_norm": 0.873230516910553, "learning_rate": 0.001025925925925926, "loss": 1.1282, "step": 26300 }, { "epoch": 97.44, "grad_norm": 0.8769469857215881, "learning_rate": 0.0010255555555555556, "loss": 1.1348, "step": 26310 }, { "epoch": 97.48, "grad_norm": 0.9158292412757874, "learning_rate": 0.0010251851851851852, "loss": 1.162, "step": 26320 }, { "epoch": 97.52, "grad_norm": 0.9506550431251526, "learning_rate": 0.0010248148148148148, "loss": 1.1757, "step": 26330 }, { "epoch": 97.56, "grad_norm": 0.8787921071052551, "learning_rate": 0.0010244444444444446, "loss": 1.182, "step": 26340 }, { "epoch": 97.59, "grad_norm": 0.8608382940292358, "learning_rate": 0.0010240740740740742, "loss": 1.1883, "step": 26350 }, { "epoch": 97.63, "grad_norm": 0.8862053751945496, "learning_rate": 0.0010237037037037038, "loss": 1.1919, "step": 26360 }, { "epoch": 97.67, "grad_norm": 0.8187049031257629, "learning_rate": 0.0010233333333333333, "loss": 1.1939, "step": 26370 }, { "epoch": 97.7, "grad_norm": 0.8783975839614868, "learning_rate": 0.0010229629629629631, "loss": 1.1777, "step": 26380 }, { "epoch": 97.74, "grad_norm": 0.8841999173164368, "learning_rate": 0.0010225925925925925, "loss": 1.2191, "step": 26390 }, { "epoch": 97.78, "grad_norm": 0.970224142074585, "learning_rate": 0.0010222222222222221, "loss": 1.2244, "step": 26400 }, { "epoch": 97.81, "grad_norm": 0.8861240148544312, "learning_rate": 0.001021851851851852, "loss": 1.2127, "step": 26410 }, { "epoch": 97.85, "grad_norm": 0.9190117716789246, "learning_rate": 0.0010214814814814815, "loss": 1.2305, "step": 26420 }, { "epoch": 97.89, "grad_norm": 0.8924251794815063, "learning_rate": 0.001021111111111111, "loss": 1.2419, "step": 26430 }, { "epoch": 97.93, "grad_norm": 0.9394065737724304, "learning_rate": 0.0010207407407407407, "loss": 1.2266, "step": 26440 }, { "epoch": 97.96, "grad_norm": 0.9103227257728577, "learning_rate": 0.0010203703703703705, "loss": 1.2484, "step": 26450 }, { "epoch": 98.0, "grad_norm": 1.6072869300842285, "learning_rate": 0.00102, "loss": 1.2523, "step": 26460 }, { "epoch": 98.04, "grad_norm": 0.8078530430793762, "learning_rate": 0.0010196296296296297, "loss": 1.0086, "step": 26470 }, { "epoch": 98.07, "grad_norm": 0.9085745811462402, "learning_rate": 0.0010192592592592593, "loss": 1.0258, "step": 26480 }, { "epoch": 98.11, "grad_norm": 0.8509937524795532, "learning_rate": 0.001018888888888889, "loss": 1.0416, "step": 26490 }, { "epoch": 98.15, "grad_norm": 0.9164654612541199, "learning_rate": 0.0010185185185185186, "loss": 1.0102, "step": 26500 }, { "epoch": 98.19, "grad_norm": 0.8645424842834473, "learning_rate": 0.0010181481481481482, "loss": 1.0746, "step": 26510 }, { "epoch": 98.22, "grad_norm": 0.850975751876831, "learning_rate": 0.0010177777777777776, "loss": 1.0675, "step": 26520 }, { "epoch": 98.26, "grad_norm": 0.922157883644104, "learning_rate": 0.0010174074074074074, "loss": 1.1008, "step": 26530 }, { "epoch": 98.3, "grad_norm": 0.8732514381408691, "learning_rate": 0.001017037037037037, "loss": 1.1038, "step": 26540 }, { "epoch": 98.33, "grad_norm": 0.8597608804702759, "learning_rate": 0.0010166666666666666, "loss": 1.0974, "step": 26550 }, { "epoch": 98.37, "grad_norm": 0.8999730944633484, "learning_rate": 0.0010162962962962962, "loss": 1.0949, "step": 26560 }, { "epoch": 98.41, "grad_norm": 0.8790624141693115, "learning_rate": 0.001015925925925926, "loss": 1.1276, "step": 26570 }, { "epoch": 98.44, "grad_norm": 0.9210389256477356, "learning_rate": 0.0010155555555555556, "loss": 1.1287, "step": 26580 }, { "epoch": 98.48, "grad_norm": 1.0357301235198975, "learning_rate": 0.0010151851851851852, "loss": 1.1505, "step": 26590 }, { "epoch": 98.52, "grad_norm": 0.8954038619995117, "learning_rate": 0.001014814814814815, "loss": 1.156, "step": 26600 }, { "epoch": 98.56, "grad_norm": 0.9502341747283936, "learning_rate": 0.0010144444444444446, "loss": 1.1765, "step": 26610 }, { "epoch": 98.59, "grad_norm": 0.9129772186279297, "learning_rate": 0.0010140740740740741, "loss": 1.1711, "step": 26620 }, { "epoch": 98.63, "grad_norm": 0.9075857996940613, "learning_rate": 0.0010137037037037037, "loss": 1.1851, "step": 26630 }, { "epoch": 98.67, "grad_norm": 0.8680686950683594, "learning_rate": 0.0010133333333333335, "loss": 1.1878, "step": 26640 }, { "epoch": 98.7, "grad_norm": 0.8635008931159973, "learning_rate": 0.0010129629629629631, "loss": 1.2095, "step": 26650 }, { "epoch": 98.74, "grad_norm": 0.871293306350708, "learning_rate": 0.0010125925925925925, "loss": 1.1938, "step": 26660 }, { "epoch": 98.78, "grad_norm": 0.8697920441627502, "learning_rate": 0.001012222222222222, "loss": 1.1918, "step": 26670 }, { "epoch": 98.81, "grad_norm": 0.9331200122833252, "learning_rate": 0.0010118518518518519, "loss": 1.228, "step": 26680 }, { "epoch": 98.85, "grad_norm": 0.9641883969306946, "learning_rate": 0.0010114814814814815, "loss": 1.2206, "step": 26690 }, { "epoch": 98.89, "grad_norm": 0.8545400500297546, "learning_rate": 0.001011111111111111, "loss": 1.2681, "step": 26700 }, { "epoch": 98.93, "grad_norm": 0.9419219493865967, "learning_rate": 0.0010107407407407407, "loss": 1.218, "step": 26710 }, { "epoch": 98.96, "grad_norm": 0.9317721724510193, "learning_rate": 0.0010103703703703705, "loss": 1.2097, "step": 26720 }, { "epoch": 99.0, "grad_norm": 1.9032236337661743, "learning_rate": 0.00101, "loss": 1.2093, "step": 26730 }, { "epoch": 99.04, "grad_norm": 0.873572051525116, "learning_rate": 0.0010096296296296296, "loss": 1.0002, "step": 26740 }, { "epoch": 99.07, "grad_norm": 0.8706053495407104, "learning_rate": 0.0010092592592592592, "loss": 0.9963, "step": 26750 }, { "epoch": 99.11, "grad_norm": 0.881637454032898, "learning_rate": 0.001008888888888889, "loss": 1.0306, "step": 26760 }, { "epoch": 99.15, "grad_norm": 0.9505958557128906, "learning_rate": 0.0010085185185185186, "loss": 1.0336, "step": 26770 }, { "epoch": 99.19, "grad_norm": 0.8608532547950745, "learning_rate": 0.0010081481481481482, "loss": 1.0318, "step": 26780 }, { "epoch": 99.22, "grad_norm": 0.9403489232063293, "learning_rate": 0.001007777777777778, "loss": 1.0933, "step": 26790 }, { "epoch": 99.26, "grad_norm": 0.9129186868667603, "learning_rate": 0.0010074074074074074, "loss": 1.0923, "step": 26800 }, { "epoch": 99.3, "grad_norm": 0.9360917210578918, "learning_rate": 0.001007037037037037, "loss": 1.116, "step": 26810 }, { "epoch": 99.33, "grad_norm": 0.8772351741790771, "learning_rate": 0.0010066666666666666, "loss": 1.1082, "step": 26820 }, { "epoch": 99.37, "grad_norm": 0.8974486589431763, "learning_rate": 0.0010062962962962964, "loss": 1.1225, "step": 26830 }, { "epoch": 99.41, "grad_norm": 0.9104517698287964, "learning_rate": 0.001005925925925926, "loss": 1.1213, "step": 26840 }, { "epoch": 99.44, "grad_norm": 0.9942017793655396, "learning_rate": 0.0010055555555555555, "loss": 1.1117, "step": 26850 }, { "epoch": 99.48, "grad_norm": 0.8957462906837463, "learning_rate": 0.0010051851851851851, "loss": 1.1415, "step": 26860 }, { "epoch": 99.52, "grad_norm": 0.9015626907348633, "learning_rate": 0.001004814814814815, "loss": 1.1527, "step": 26870 }, { "epoch": 99.56, "grad_norm": 0.8942874670028687, "learning_rate": 0.0010044444444444445, "loss": 1.1544, "step": 26880 }, { "epoch": 99.59, "grad_norm": 0.9468600749969482, "learning_rate": 0.0010040740740740741, "loss": 1.145, "step": 26890 }, { "epoch": 99.63, "grad_norm": 0.9078618884086609, "learning_rate": 0.0010037037037037037, "loss": 1.1799, "step": 26900 }, { "epoch": 99.67, "grad_norm": 0.9247041940689087, "learning_rate": 0.0010033333333333335, "loss": 1.1948, "step": 26910 }, { "epoch": 99.7, "grad_norm": 0.9799883961677551, "learning_rate": 0.001002962962962963, "loss": 1.1786, "step": 26920 }, { "epoch": 99.74, "grad_norm": 0.8741432428359985, "learning_rate": 0.0010025925925925927, "loss": 1.1948, "step": 26930 }, { "epoch": 99.78, "grad_norm": 0.9404388666152954, "learning_rate": 0.001002222222222222, "loss": 1.1932, "step": 26940 }, { "epoch": 99.81, "grad_norm": 0.9425485134124756, "learning_rate": 0.0010018518518518519, "loss": 1.2049, "step": 26950 }, { "epoch": 99.85, "grad_norm": 0.9867810010910034, "learning_rate": 0.0010014814814814814, "loss": 1.2078, "step": 26960 }, { "epoch": 99.89, "grad_norm": 0.9349437952041626, "learning_rate": 0.001001111111111111, "loss": 1.2086, "step": 26970 }, { "epoch": 99.93, "grad_norm": 0.8289681673049927, "learning_rate": 0.0010007407407407406, "loss": 1.2475, "step": 26980 }, { "epoch": 99.96, "grad_norm": 0.9117506146430969, "learning_rate": 0.0010003703703703704, "loss": 1.2344, "step": 26990 }, { "epoch": 100.0, "grad_norm": 1.7707023620605469, "learning_rate": 0.001, "loss": 1.2213, "step": 27000 }, { "epoch": 100.04, "grad_norm": 0.8565974235534668, "learning_rate": 0.0009996296296296296, "loss": 0.9891, "step": 27010 }, { "epoch": 100.07, "grad_norm": 0.8615112900733948, "learning_rate": 0.0009992592592592594, "loss": 1.0083, "step": 27020 }, { "epoch": 100.11, "grad_norm": 0.9136719703674316, "learning_rate": 0.000998888888888889, "loss": 0.9975, "step": 27030 }, { "epoch": 100.15, "grad_norm": 0.8897717595100403, "learning_rate": 0.0009985185185185186, "loss": 1.0088, "step": 27040 }, { "epoch": 100.19, "grad_norm": 0.8684051036834717, "learning_rate": 0.0009981481481481482, "loss": 1.0412, "step": 27050 }, { "epoch": 100.22, "grad_norm": 1.019988775253296, "learning_rate": 0.0009977777777777778, "loss": 1.0733, "step": 27060 }, { "epoch": 100.26, "grad_norm": 0.8850982189178467, "learning_rate": 0.0009974074074074074, "loss": 1.0866, "step": 27070 }, { "epoch": 100.3, "grad_norm": 0.8889622688293457, "learning_rate": 0.0009970370370370372, "loss": 1.068, "step": 27080 }, { "epoch": 100.33, "grad_norm": 0.9454230070114136, "learning_rate": 0.0009966666666666668, "loss": 1.0957, "step": 27090 }, { "epoch": 100.37, "grad_norm": 0.9413827657699585, "learning_rate": 0.0009962962962962963, "loss": 1.1336, "step": 27100 }, { "epoch": 100.41, "grad_norm": 0.8245477676391602, "learning_rate": 0.000995925925925926, "loss": 1.1296, "step": 27110 }, { "epoch": 100.44, "grad_norm": 0.9790675640106201, "learning_rate": 0.0009955555555555555, "loss": 1.1404, "step": 27120 }, { "epoch": 100.48, "grad_norm": 0.9440389275550842, "learning_rate": 0.000995185185185185, "loss": 1.1168, "step": 27130 }, { "epoch": 100.52, "grad_norm": 0.9235543012619019, "learning_rate": 0.000994814814814815, "loss": 1.1552, "step": 27140 }, { "epoch": 100.56, "grad_norm": 0.9291632771492004, "learning_rate": 0.0009944444444444445, "loss": 1.156, "step": 27150 }, { "epoch": 100.59, "grad_norm": 0.912441074848175, "learning_rate": 0.000994074074074074, "loss": 1.1397, "step": 27160 }, { "epoch": 100.63, "grad_norm": 0.9047914147377014, "learning_rate": 0.0009937037037037037, "loss": 1.1661, "step": 27170 }, { "epoch": 100.67, "grad_norm": 0.862169623374939, "learning_rate": 0.0009933333333333333, "loss": 1.1737, "step": 27180 }, { "epoch": 100.7, "grad_norm": 1.00532066822052, "learning_rate": 0.0009929629629629629, "loss": 1.172, "step": 27190 }, { "epoch": 100.74, "grad_norm": 0.9798426628112793, "learning_rate": 0.0009925925925925927, "loss": 1.1825, "step": 27200 }, { "epoch": 100.78, "grad_norm": 0.9061884880065918, "learning_rate": 0.0009922222222222222, "loss": 1.1892, "step": 27210 }, { "epoch": 100.81, "grad_norm": 0.9554376006126404, "learning_rate": 0.0009918518518518518, "loss": 1.1927, "step": 27220 }, { "epoch": 100.85, "grad_norm": 0.8725953698158264, "learning_rate": 0.0009914814814814816, "loss": 1.1914, "step": 27230 }, { "epoch": 100.89, "grad_norm": 0.8509664535522461, "learning_rate": 0.0009911111111111112, "loss": 1.1964, "step": 27240 }, { "epoch": 100.93, "grad_norm": 0.9445471167564392, "learning_rate": 0.0009907407407407408, "loss": 1.2348, "step": 27250 }, { "epoch": 100.96, "grad_norm": 0.8974488973617554, "learning_rate": 0.0009903703703703704, "loss": 1.2092, "step": 27260 }, { "epoch": 101.0, "grad_norm": 1.5322643518447876, "learning_rate": 0.00099, "loss": 1.2047, "step": 27270 }, { "epoch": 101.04, "grad_norm": 0.8989846706390381, "learning_rate": 0.0009896296296296296, "loss": 0.989, "step": 27280 }, { "epoch": 101.07, "grad_norm": 0.9590046405792236, "learning_rate": 0.0009892592592592594, "loss": 1.0178, "step": 27290 }, { "epoch": 101.11, "grad_norm": 0.8980274796485901, "learning_rate": 0.000988888888888889, "loss": 1.0279, "step": 27300 }, { "epoch": 101.15, "grad_norm": 0.9266800880432129, "learning_rate": 0.0009885185185185186, "loss": 1.0291, "step": 27310 }, { "epoch": 101.19, "grad_norm": 0.8878993391990662, "learning_rate": 0.0009881481481481482, "loss": 1.0202, "step": 27320 }, { "epoch": 101.22, "grad_norm": 0.9423307776451111, "learning_rate": 0.0009877777777777777, "loss": 1.0479, "step": 27330 }, { "epoch": 101.26, "grad_norm": 0.911369800567627, "learning_rate": 0.0009874074074074073, "loss": 1.0598, "step": 27340 }, { "epoch": 101.3, "grad_norm": 0.842523455619812, "learning_rate": 0.0009870370370370371, "loss": 1.0575, "step": 27350 }, { "epoch": 101.33, "grad_norm": 0.8670434355735779, "learning_rate": 0.0009866666666666667, "loss": 1.086, "step": 27360 }, { "epoch": 101.37, "grad_norm": 0.9488222599029541, "learning_rate": 0.0009862962962962963, "loss": 1.108, "step": 27370 }, { "epoch": 101.41, "grad_norm": 0.935550332069397, "learning_rate": 0.000985925925925926, "loss": 1.0789, "step": 27380 }, { "epoch": 101.44, "grad_norm": 0.930512011051178, "learning_rate": 0.0009855555555555555, "loss": 1.1293, "step": 27390 }, { "epoch": 101.48, "grad_norm": 0.9525163769721985, "learning_rate": 0.000985185185185185, "loss": 1.1127, "step": 27400 }, { "epoch": 101.52, "grad_norm": 0.9982643723487854, "learning_rate": 0.0009848148148148149, "loss": 1.1436, "step": 27410 }, { "epoch": 101.56, "grad_norm": 0.8970288634300232, "learning_rate": 0.0009844444444444445, "loss": 1.11, "step": 27420 }, { "epoch": 101.59, "grad_norm": 0.9926027655601501, "learning_rate": 0.000984074074074074, "loss": 1.1894, "step": 27430 }, { "epoch": 101.63, "grad_norm": 0.9675959348678589, "learning_rate": 0.0009837037037037039, "loss": 1.1559, "step": 27440 }, { "epoch": 101.67, "grad_norm": 0.9283835887908936, "learning_rate": 0.0009833333333333332, "loss": 1.1515, "step": 27450 }, { "epoch": 101.7, "grad_norm": 0.9680076241493225, "learning_rate": 0.000982962962962963, "loss": 1.1891, "step": 27460 }, { "epoch": 101.74, "grad_norm": 0.9401657581329346, "learning_rate": 0.0009825925925925926, "loss": 1.1778, "step": 27470 }, { "epoch": 101.78, "grad_norm": 0.9478588104248047, "learning_rate": 0.0009822222222222222, "loss": 1.1805, "step": 27480 }, { "epoch": 101.81, "grad_norm": 0.8912003636360168, "learning_rate": 0.0009818518518518518, "loss": 1.1929, "step": 27490 }, { "epoch": 101.85, "grad_norm": 0.9238876700401306, "learning_rate": 0.0009814814814814816, "loss": 1.1934, "step": 27500 }, { "epoch": 101.89, "grad_norm": 0.9057316184043884, "learning_rate": 0.0009811111111111112, "loss": 1.1807, "step": 27510 }, { "epoch": 101.93, "grad_norm": 0.8781848549842834, "learning_rate": 0.0009807407407407408, "loss": 1.227, "step": 27520 }, { "epoch": 101.96, "grad_norm": 0.9485657811164856, "learning_rate": 0.0009803703703703704, "loss": 1.227, "step": 27530 }, { "epoch": 102.0, "grad_norm": 1.7745193243026733, "learning_rate": 0.00098, "loss": 1.2126, "step": 27540 }, { "epoch": 102.04, "grad_norm": 0.8811674118041992, "learning_rate": 0.0009796296296296296, "loss": 0.9722, "step": 27550 }, { "epoch": 102.07, "grad_norm": 0.9530802369117737, "learning_rate": 0.0009792592592592594, "loss": 0.9783, "step": 27560 }, { "epoch": 102.11, "grad_norm": 0.8173848986625671, "learning_rate": 0.000978888888888889, "loss": 1.0037, "step": 27570 }, { "epoch": 102.15, "grad_norm": 0.95224529504776, "learning_rate": 0.0009785185185185185, "loss": 1.0521, "step": 27580 }, { "epoch": 102.19, "grad_norm": 0.8620432615280151, "learning_rate": 0.0009781481481481481, "loss": 1.0357, "step": 27590 }, { "epoch": 102.22, "grad_norm": 0.922075092792511, "learning_rate": 0.0009777777777777777, "loss": 1.0585, "step": 27600 }, { "epoch": 102.26, "grad_norm": 0.8372809290885925, "learning_rate": 0.0009774074074074073, "loss": 1.0536, "step": 27610 }, { "epoch": 102.3, "grad_norm": 0.9226149320602417, "learning_rate": 0.0009770370370370371, "loss": 1.0465, "step": 27620 }, { "epoch": 102.33, "grad_norm": 1.0217820405960083, "learning_rate": 0.0009766666666666667, "loss": 1.0734, "step": 27630 }, { "epoch": 102.37, "grad_norm": 0.8591089248657227, "learning_rate": 0.0009762962962962964, "loss": 1.0942, "step": 27640 }, { "epoch": 102.41, "grad_norm": 0.9037002325057983, "learning_rate": 0.000975925925925926, "loss": 1.1171, "step": 27650 }, { "epoch": 102.44, "grad_norm": 0.9767083525657654, "learning_rate": 0.0009755555555555556, "loss": 1.1325, "step": 27660 }, { "epoch": 102.48, "grad_norm": 0.8895253539085388, "learning_rate": 0.0009751851851851852, "loss": 1.1, "step": 27670 }, { "epoch": 102.52, "grad_norm": 0.9272596836090088, "learning_rate": 0.0009748148148148149, "loss": 1.1255, "step": 27680 }, { "epoch": 102.56, "grad_norm": 0.9447619915008545, "learning_rate": 0.0009744444444444444, "loss": 1.1298, "step": 27690 }, { "epoch": 102.59, "grad_norm": 0.8245322704315186, "learning_rate": 0.0009740740740740741, "loss": 1.135, "step": 27700 }, { "epoch": 102.63, "grad_norm": 0.8659734129905701, "learning_rate": 0.0009737037037037037, "loss": 1.1282, "step": 27710 }, { "epoch": 102.67, "grad_norm": 0.8939563035964966, "learning_rate": 0.0009733333333333334, "loss": 1.1602, "step": 27720 }, { "epoch": 102.7, "grad_norm": 0.9657849073410034, "learning_rate": 0.0009729629629629629, "loss": 1.1643, "step": 27730 }, { "epoch": 102.74, "grad_norm": 0.9996271133422852, "learning_rate": 0.0009725925925925926, "loss": 1.1761, "step": 27740 }, { "epoch": 102.78, "grad_norm": 0.8725650310516357, "learning_rate": 0.0009722222222222222, "loss": 1.1291, "step": 27750 }, { "epoch": 102.81, "grad_norm": 0.9499808549880981, "learning_rate": 0.0009718518518518519, "loss": 1.1882, "step": 27760 }, { "epoch": 102.85, "grad_norm": 0.9083239436149597, "learning_rate": 0.0009714814814814815, "loss": 1.1722, "step": 27770 }, { "epoch": 102.89, "grad_norm": 0.9593604803085327, "learning_rate": 0.0009711111111111112, "loss": 1.1936, "step": 27780 }, { "epoch": 102.93, "grad_norm": 0.9239686131477356, "learning_rate": 0.0009707407407407408, "loss": 1.2229, "step": 27790 }, { "epoch": 102.96, "grad_norm": 0.9181411862373352, "learning_rate": 0.0009703703703703704, "loss": 1.1673, "step": 27800 }, { "epoch": 103.0, "grad_norm": 1.701149582862854, "learning_rate": 0.0009699999999999999, "loss": 1.225, "step": 27810 }, { "epoch": 103.04, "grad_norm": 0.8860218524932861, "learning_rate": 0.0009696296296296296, "loss": 0.9796, "step": 27820 }, { "epoch": 103.07, "grad_norm": 0.9513492584228516, "learning_rate": 0.0009692592592592593, "loss": 0.9803, "step": 27830 }, { "epoch": 103.11, "grad_norm": 0.8623229265213013, "learning_rate": 0.0009688888888888889, "loss": 1.0065, "step": 27840 }, { "epoch": 103.15, "grad_norm": 0.9298003911972046, "learning_rate": 0.0009685185185185186, "loss": 1.0145, "step": 27850 }, { "epoch": 103.19, "grad_norm": 0.8768572211265564, "learning_rate": 0.0009681481481481482, "loss": 1.0327, "step": 27860 }, { "epoch": 103.22, "grad_norm": 0.9230377078056335, "learning_rate": 0.0009677777777777778, "loss": 1.0383, "step": 27870 }, { "epoch": 103.26, "grad_norm": 0.8653807640075684, "learning_rate": 0.0009674074074074074, "loss": 1.0602, "step": 27880 }, { "epoch": 103.3, "grad_norm": 0.918269157409668, "learning_rate": 0.0009670370370370371, "loss": 1.0644, "step": 27890 }, { "epoch": 103.33, "grad_norm": 0.940070629119873, "learning_rate": 0.0009666666666666667, "loss": 1.0769, "step": 27900 }, { "epoch": 103.37, "grad_norm": 0.9915195107460022, "learning_rate": 0.0009662962962962964, "loss": 1.0904, "step": 27910 }, { "epoch": 103.41, "grad_norm": 0.9446269869804382, "learning_rate": 0.000965925925925926, "loss": 1.0902, "step": 27920 }, { "epoch": 103.44, "grad_norm": 0.9780631065368652, "learning_rate": 0.0009655555555555555, "loss": 1.1287, "step": 27930 }, { "epoch": 103.48, "grad_norm": 1.0551878213882446, "learning_rate": 0.0009651851851851851, "loss": 1.103, "step": 27940 }, { "epoch": 103.52, "grad_norm": 0.9577930569648743, "learning_rate": 0.0009648148148148148, "loss": 1.0911, "step": 27950 }, { "epoch": 103.56, "grad_norm": 0.8572773933410645, "learning_rate": 0.0009644444444444444, "loss": 1.1244, "step": 27960 }, { "epoch": 103.59, "grad_norm": 0.8974463939666748, "learning_rate": 0.0009640740740740741, "loss": 1.1524, "step": 27970 }, { "epoch": 103.63, "grad_norm": 0.911722719669342, "learning_rate": 0.0009637037037037037, "loss": 1.1478, "step": 27980 }, { "epoch": 103.67, "grad_norm": 0.9285465478897095, "learning_rate": 0.0009633333333333334, "loss": 1.1461, "step": 27990 }, { "epoch": 103.7, "grad_norm": 0.9444763660430908, "learning_rate": 0.0009629629629629629, "loss": 1.1529, "step": 28000 }, { "epoch": 103.74, "grad_norm": 0.897024929523468, "learning_rate": 0.0009625925925925926, "loss": 1.1455, "step": 28010 }, { "epoch": 103.78, "grad_norm": 0.9245248436927795, "learning_rate": 0.0009622222222222222, "loss": 1.172, "step": 28020 }, { "epoch": 103.81, "grad_norm": 0.9002532958984375, "learning_rate": 0.0009618518518518519, "loss": 1.167, "step": 28030 }, { "epoch": 103.85, "grad_norm": 0.938520610332489, "learning_rate": 0.0009614814814814816, "loss": 1.1713, "step": 28040 }, { "epoch": 103.89, "grad_norm": 0.9639129042625427, "learning_rate": 0.0009611111111111112, "loss": 1.181, "step": 28050 }, { "epoch": 103.93, "grad_norm": 0.9609660506248474, "learning_rate": 0.0009607407407407408, "loss": 1.1948, "step": 28060 }, { "epoch": 103.96, "grad_norm": 0.9552947878837585, "learning_rate": 0.0009603703703703703, "loss": 1.2001, "step": 28070 }, { "epoch": 104.0, "grad_norm": 1.452337384223938, "learning_rate": 0.00096, "loss": 1.1939, "step": 28080 }, { "epoch": 104.04, "grad_norm": 0.8623477220535278, "learning_rate": 0.0009596296296296296, "loss": 0.9715, "step": 28090 }, { "epoch": 104.07, "grad_norm": 0.8784454464912415, "learning_rate": 0.0009592592592592593, "loss": 0.9805, "step": 28100 }, { "epoch": 104.11, "grad_norm": 0.939306914806366, "learning_rate": 0.0009588888888888889, "loss": 0.9661, "step": 28110 }, { "epoch": 104.15, "grad_norm": 0.9585001468658447, "learning_rate": 0.0009585185185185186, "loss": 1.0202, "step": 28120 }, { "epoch": 104.19, "grad_norm": 0.9838183522224426, "learning_rate": 0.0009581481481481482, "loss": 1.0172, "step": 28130 }, { "epoch": 104.22, "grad_norm": 0.9469972252845764, "learning_rate": 0.0009577777777777778, "loss": 1.0398, "step": 28140 }, { "epoch": 104.26, "grad_norm": 0.933664083480835, "learning_rate": 0.0009574074074074074, "loss": 1.0533, "step": 28150 }, { "epoch": 104.3, "grad_norm": 0.97636479139328, "learning_rate": 0.0009570370370370371, "loss": 1.0557, "step": 28160 }, { "epoch": 104.33, "grad_norm": 0.8865147829055786, "learning_rate": 0.0009566666666666666, "loss": 1.0592, "step": 28170 }, { "epoch": 104.37, "grad_norm": 0.9205927848815918, "learning_rate": 0.0009562962962962963, "loss": 1.0823, "step": 28180 }, { "epoch": 104.41, "grad_norm": 0.9018583297729492, "learning_rate": 0.0009559259259259259, "loss": 1.0743, "step": 28190 }, { "epoch": 104.44, "grad_norm": 1.000645637512207, "learning_rate": 0.0009555555555555556, "loss": 1.1016, "step": 28200 }, { "epoch": 104.48, "grad_norm": 0.9386842250823975, "learning_rate": 0.0009551851851851851, "loss": 1.0779, "step": 28210 }, { "epoch": 104.52, "grad_norm": 0.8640486001968384, "learning_rate": 0.0009548148148148148, "loss": 1.0997, "step": 28220 }, { "epoch": 104.56, "grad_norm": 0.9148708581924438, "learning_rate": 0.0009544444444444445, "loss": 1.1083, "step": 28230 }, { "epoch": 104.59, "grad_norm": 0.8763410449028015, "learning_rate": 0.0009540740740740741, "loss": 1.1069, "step": 28240 }, { "epoch": 104.63, "grad_norm": 0.9024518728256226, "learning_rate": 0.0009537037037037038, "loss": 1.0925, "step": 28250 }, { "epoch": 104.67, "grad_norm": 1.0220943689346313, "learning_rate": 0.0009533333333333334, "loss": 1.1489, "step": 28260 }, { "epoch": 104.7, "grad_norm": 0.9682163596153259, "learning_rate": 0.0009529629629629631, "loss": 1.1572, "step": 28270 }, { "epoch": 104.74, "grad_norm": 0.9184615612030029, "learning_rate": 0.0009525925925925926, "loss": 1.1555, "step": 28280 }, { "epoch": 104.78, "grad_norm": 0.9440074563026428, "learning_rate": 0.0009522222222222223, "loss": 1.1587, "step": 28290 }, { "epoch": 104.81, "grad_norm": 0.9604818224906921, "learning_rate": 0.0009518518518518518, "loss": 1.16, "step": 28300 }, { "epoch": 104.85, "grad_norm": 0.8982845544815063, "learning_rate": 0.0009514814814814815, "loss": 1.1673, "step": 28310 }, { "epoch": 104.89, "grad_norm": 0.8533664345741272, "learning_rate": 0.0009511111111111111, "loss": 1.1979, "step": 28320 }, { "epoch": 104.93, "grad_norm": 0.9376116394996643, "learning_rate": 0.0009507407407407408, "loss": 1.2014, "step": 28330 }, { "epoch": 104.96, "grad_norm": 0.9965450763702393, "learning_rate": 0.0009503703703703704, "loss": 1.1756, "step": 28340 }, { "epoch": 105.0, "grad_norm": 1.7779862880706787, "learning_rate": 0.00095, "loss": 1.1981, "step": 28350 }, { "epoch": 105.04, "grad_norm": 0.9287167191505432, "learning_rate": 0.0009496296296296296, "loss": 0.9635, "step": 28360 }, { "epoch": 105.07, "grad_norm": 0.8633070588111877, "learning_rate": 0.0009492592592592593, "loss": 0.9735, "step": 28370 }, { "epoch": 105.11, "grad_norm": 0.9253849983215332, "learning_rate": 0.0009488888888888889, "loss": 0.9967, "step": 28380 }, { "epoch": 105.15, "grad_norm": 0.8466000556945801, "learning_rate": 0.0009485185185185186, "loss": 0.9589, "step": 28390 }, { "epoch": 105.19, "grad_norm": 0.8577874302864075, "learning_rate": 0.0009481481481481482, "loss": 0.9805, "step": 28400 }, { "epoch": 105.22, "grad_norm": 0.9374727010726929, "learning_rate": 0.0009477777777777779, "loss": 1.0293, "step": 28410 }, { "epoch": 105.26, "grad_norm": 1.04707932472229, "learning_rate": 0.0009474074074074073, "loss": 1.0373, "step": 28420 }, { "epoch": 105.3, "grad_norm": 0.9615543484687805, "learning_rate": 0.000947037037037037, "loss": 1.0446, "step": 28430 }, { "epoch": 105.33, "grad_norm": 0.9235749244689941, "learning_rate": 0.0009466666666666667, "loss": 1.0288, "step": 28440 }, { "epoch": 105.37, "grad_norm": 0.9541982412338257, "learning_rate": 0.0009462962962962963, "loss": 1.0546, "step": 28450 }, { "epoch": 105.41, "grad_norm": 0.891716718673706, "learning_rate": 0.000945925925925926, "loss": 1.095, "step": 28460 }, { "epoch": 105.44, "grad_norm": 0.9158815145492554, "learning_rate": 0.0009455555555555556, "loss": 1.0881, "step": 28470 }, { "epoch": 105.48, "grad_norm": 1.0235891342163086, "learning_rate": 0.0009451851851851853, "loss": 1.1174, "step": 28480 }, { "epoch": 105.52, "grad_norm": 0.9569323658943176, "learning_rate": 0.0009448148148148148, "loss": 1.1238, "step": 28490 }, { "epoch": 105.56, "grad_norm": 1.079743504524231, "learning_rate": 0.0009444444444444445, "loss": 1.1312, "step": 28500 }, { "epoch": 105.59, "grad_norm": 0.929822564125061, "learning_rate": 0.0009440740740740741, "loss": 1.106, "step": 28510 }, { "epoch": 105.63, "grad_norm": 0.9850989580154419, "learning_rate": 0.0009437037037037038, "loss": 1.1194, "step": 28520 }, { "epoch": 105.67, "grad_norm": 0.919579029083252, "learning_rate": 0.0009433333333333334, "loss": 1.1395, "step": 28530 }, { "epoch": 105.7, "grad_norm": 0.9597387313842773, "learning_rate": 0.000942962962962963, "loss": 1.1527, "step": 28540 }, { "epoch": 105.74, "grad_norm": 0.9359583854675293, "learning_rate": 0.0009425925925925925, "loss": 1.1315, "step": 28550 }, { "epoch": 105.78, "grad_norm": 1.0015721321105957, "learning_rate": 0.0009422222222222222, "loss": 1.1733, "step": 28560 }, { "epoch": 105.81, "grad_norm": 0.9597902894020081, "learning_rate": 0.0009418518518518518, "loss": 1.1285, "step": 28570 }, { "epoch": 105.85, "grad_norm": 0.9335023760795593, "learning_rate": 0.0009414814814814815, "loss": 1.1722, "step": 28580 }, { "epoch": 105.89, "grad_norm": 0.9263126254081726, "learning_rate": 0.0009411111111111111, "loss": 1.164, "step": 28590 }, { "epoch": 105.93, "grad_norm": 0.9743094444274902, "learning_rate": 0.0009407407407407408, "loss": 1.1819, "step": 28600 }, { "epoch": 105.96, "grad_norm": 0.8798518776893616, "learning_rate": 0.0009403703703703704, "loss": 1.1831, "step": 28610 }, { "epoch": 106.0, "grad_norm": 2.3925769329071045, "learning_rate": 0.00094, "loss": 1.1827, "step": 28620 }, { "epoch": 106.04, "grad_norm": 0.9399033188819885, "learning_rate": 0.0009396296296296296, "loss": 0.966, "step": 28630 }, { "epoch": 106.07, "grad_norm": 0.9358739852905273, "learning_rate": 0.0009392592592592593, "loss": 0.9811, "step": 28640 }, { "epoch": 106.11, "grad_norm": 0.9023323059082031, "learning_rate": 0.000938888888888889, "loss": 0.9646, "step": 28650 }, { "epoch": 106.15, "grad_norm": 0.8303003907203674, "learning_rate": 0.0009385185185185185, "loss": 0.9887, "step": 28660 }, { "epoch": 106.19, "grad_norm": 0.8875153064727783, "learning_rate": 0.0009381481481481482, "loss": 0.9814, "step": 28670 }, { "epoch": 106.22, "grad_norm": 0.8730666041374207, "learning_rate": 0.0009377777777777778, "loss": 0.9953, "step": 28680 }, { "epoch": 106.26, "grad_norm": 0.979469895362854, "learning_rate": 0.0009374074074074074, "loss": 1.018, "step": 28690 }, { "epoch": 106.3, "grad_norm": 1.0179818868637085, "learning_rate": 0.000937037037037037, "loss": 1.0454, "step": 28700 }, { "epoch": 106.33, "grad_norm": 1.0186294317245483, "learning_rate": 0.0009366666666666667, "loss": 1.053, "step": 28710 }, { "epoch": 106.37, "grad_norm": 0.8819359540939331, "learning_rate": 0.0009362962962962963, "loss": 1.0552, "step": 28720 }, { "epoch": 106.41, "grad_norm": 0.8499149084091187, "learning_rate": 0.000935925925925926, "loss": 1.077, "step": 28730 }, { "epoch": 106.44, "grad_norm": 0.9276627898216248, "learning_rate": 0.0009355555555555556, "loss": 1.0989, "step": 28740 }, { "epoch": 106.48, "grad_norm": 0.9268879890441895, "learning_rate": 0.0009351851851851853, "loss": 1.0761, "step": 28750 }, { "epoch": 106.52, "grad_norm": 0.9042457342147827, "learning_rate": 0.0009348148148148148, "loss": 1.0841, "step": 28760 }, { "epoch": 106.56, "grad_norm": 0.9719822406768799, "learning_rate": 0.0009344444444444444, "loss": 1.0817, "step": 28770 }, { "epoch": 106.59, "grad_norm": 0.9261794090270996, "learning_rate": 0.000934074074074074, "loss": 1.0902, "step": 28780 }, { "epoch": 106.63, "grad_norm": 0.9322009086608887, "learning_rate": 0.0009337037037037037, "loss": 1.1328, "step": 28790 }, { "epoch": 106.67, "grad_norm": 0.9948487877845764, "learning_rate": 0.0009333333333333333, "loss": 1.1484, "step": 28800 }, { "epoch": 106.7, "grad_norm": 1.0035136938095093, "learning_rate": 0.000932962962962963, "loss": 1.1586, "step": 28810 }, { "epoch": 106.74, "grad_norm": 1.028941035270691, "learning_rate": 0.0009325925925925926, "loss": 1.1316, "step": 28820 }, { "epoch": 106.78, "grad_norm": 0.9474858641624451, "learning_rate": 0.0009322222222222222, "loss": 1.1288, "step": 28830 }, { "epoch": 106.81, "grad_norm": 0.9305292963981628, "learning_rate": 0.0009318518518518518, "loss": 1.119, "step": 28840 }, { "epoch": 106.85, "grad_norm": 0.919070303440094, "learning_rate": 0.0009314814814814815, "loss": 1.1667, "step": 28850 }, { "epoch": 106.89, "grad_norm": 0.9206348657608032, "learning_rate": 0.0009311111111111112, "loss": 1.1655, "step": 28860 }, { "epoch": 106.93, "grad_norm": 1.0312327146530151, "learning_rate": 0.0009307407407407408, "loss": 1.1592, "step": 28870 }, { "epoch": 106.96, "grad_norm": 0.9701687693595886, "learning_rate": 0.0009303703703703705, "loss": 1.1635, "step": 28880 }, { "epoch": 107.0, "grad_norm": 1.8190052509307861, "learning_rate": 0.00093, "loss": 1.178, "step": 28890 }, { "epoch": 107.04, "grad_norm": 0.9251551628112793, "learning_rate": 0.0009296296296296296, "loss": 0.9336, "step": 28900 }, { "epoch": 107.07, "grad_norm": 0.8757014274597168, "learning_rate": 0.0009292592592592592, "loss": 0.9469, "step": 28910 }, { "epoch": 107.11, "grad_norm": 0.9648277163505554, "learning_rate": 0.0009288888888888889, "loss": 0.9738, "step": 28920 }, { "epoch": 107.15, "grad_norm": 0.938714325428009, "learning_rate": 0.0009285185185185185, "loss": 0.9906, "step": 28930 }, { "epoch": 107.19, "grad_norm": 0.9022919535636902, "learning_rate": 0.0009281481481481482, "loss": 1.01, "step": 28940 }, { "epoch": 107.22, "grad_norm": 0.917122483253479, "learning_rate": 0.0009277777777777778, "loss": 1.0086, "step": 28950 }, { "epoch": 107.26, "grad_norm": 0.9072996377944946, "learning_rate": 0.0009274074074074075, "loss": 1.0098, "step": 28960 }, { "epoch": 107.3, "grad_norm": 0.9521521925926208, "learning_rate": 0.000927037037037037, "loss": 1.0383, "step": 28970 }, { "epoch": 107.33, "grad_norm": 0.9218719601631165, "learning_rate": 0.0009266666666666667, "loss": 1.0318, "step": 28980 }, { "epoch": 107.37, "grad_norm": 0.9192262291908264, "learning_rate": 0.0009262962962962963, "loss": 1.0327, "step": 28990 }, { "epoch": 107.41, "grad_norm": 0.9420268535614014, "learning_rate": 0.000925925925925926, "loss": 1.0612, "step": 29000 }, { "epoch": 107.44, "grad_norm": 0.9131075143814087, "learning_rate": 0.0009255555555555555, "loss": 1.0599, "step": 29010 }, { "epoch": 107.48, "grad_norm": 0.9630795121192932, "learning_rate": 0.0009251851851851852, "loss": 1.0725, "step": 29020 }, { "epoch": 107.52, "grad_norm": 0.8927704691886902, "learning_rate": 0.0009248148148148148, "loss": 1.0808, "step": 29030 }, { "epoch": 107.56, "grad_norm": 0.9514608979225159, "learning_rate": 0.0009244444444444444, "loss": 1.098, "step": 29040 }, { "epoch": 107.59, "grad_norm": 0.9280536770820618, "learning_rate": 0.0009240740740740741, "loss": 1.1015, "step": 29050 }, { "epoch": 107.63, "grad_norm": 1.0405588150024414, "learning_rate": 0.0009237037037037037, "loss": 1.1082, "step": 29060 }, { "epoch": 107.67, "grad_norm": 0.8975964188575745, "learning_rate": 0.0009233333333333334, "loss": 1.1108, "step": 29070 }, { "epoch": 107.7, "grad_norm": 1.017134666442871, "learning_rate": 0.000922962962962963, "loss": 1.1371, "step": 29080 }, { "epoch": 107.74, "grad_norm": 0.9074270129203796, "learning_rate": 0.0009225925925925927, "loss": 1.1241, "step": 29090 }, { "epoch": 107.78, "grad_norm": 1.0147840976715088, "learning_rate": 0.0009222222222222223, "loss": 1.1267, "step": 29100 }, { "epoch": 107.81, "grad_norm": 0.9706998467445374, "learning_rate": 0.0009218518518518519, "loss": 1.1578, "step": 29110 }, { "epoch": 107.85, "grad_norm": 0.9495412111282349, "learning_rate": 0.0009214814814814815, "loss": 1.1471, "step": 29120 }, { "epoch": 107.89, "grad_norm": 0.8967522382736206, "learning_rate": 0.0009211111111111112, "loss": 1.167, "step": 29130 }, { "epoch": 107.93, "grad_norm": 0.9381629228591919, "learning_rate": 0.0009207407407407407, "loss": 1.1626, "step": 29140 }, { "epoch": 107.96, "grad_norm": 0.9799057841300964, "learning_rate": 0.0009203703703703704, "loss": 1.1653, "step": 29150 }, { "epoch": 108.0, "grad_norm": 1.7408097982406616, "learning_rate": 0.00092, "loss": 1.1384, "step": 29160 }, { "epoch": 108.04, "grad_norm": 0.8933138251304626, "learning_rate": 0.0009196296296296296, "loss": 0.9233, "step": 29170 }, { "epoch": 108.07, "grad_norm": 0.838168203830719, "learning_rate": 0.0009192592592592592, "loss": 0.9545, "step": 29180 }, { "epoch": 108.11, "grad_norm": 0.9422211050987244, "learning_rate": 0.0009188888888888889, "loss": 0.9547, "step": 29190 }, { "epoch": 108.15, "grad_norm": 0.8847689628601074, "learning_rate": 0.0009185185185185185, "loss": 0.9664, "step": 29200 }, { "epoch": 108.19, "grad_norm": 0.9079718589782715, "learning_rate": 0.0009181481481481482, "loss": 0.9592, "step": 29210 }, { "epoch": 108.22, "grad_norm": 0.9272188544273376, "learning_rate": 0.0009177777777777778, "loss": 0.9965, "step": 29220 }, { "epoch": 108.26, "grad_norm": 0.8760783076286316, "learning_rate": 0.0009174074074074075, "loss": 1.0044, "step": 29230 }, { "epoch": 108.3, "grad_norm": 0.9839217066764832, "learning_rate": 0.000917037037037037, "loss": 1.0261, "step": 29240 }, { "epoch": 108.33, "grad_norm": 0.8970228433609009, "learning_rate": 0.0009166666666666666, "loss": 1.0533, "step": 29250 }, { "epoch": 108.37, "grad_norm": 0.9859808087348938, "learning_rate": 0.0009162962962962963, "loss": 1.0375, "step": 29260 }, { "epoch": 108.41, "grad_norm": 0.9420928955078125, "learning_rate": 0.0009159259259259259, "loss": 1.064, "step": 29270 }, { "epoch": 108.44, "grad_norm": 0.9270564317703247, "learning_rate": 0.0009155555555555556, "loss": 1.0486, "step": 29280 }, { "epoch": 108.48, "grad_norm": 0.9782211184501648, "learning_rate": 0.0009151851851851852, "loss": 1.0586, "step": 29290 }, { "epoch": 108.52, "grad_norm": 0.9622315764427185, "learning_rate": 0.0009148148148148149, "loss": 1.0872, "step": 29300 }, { "epoch": 108.56, "grad_norm": 1.0131900310516357, "learning_rate": 0.0009144444444444444, "loss": 1.0958, "step": 29310 }, { "epoch": 108.59, "grad_norm": 1.0280399322509766, "learning_rate": 0.0009140740740740741, "loss": 1.0959, "step": 29320 }, { "epoch": 108.63, "grad_norm": 0.9791631698608398, "learning_rate": 0.0009137037037037037, "loss": 1.123, "step": 29330 }, { "epoch": 108.67, "grad_norm": 1.0855169296264648, "learning_rate": 0.0009133333333333334, "loss": 1.1025, "step": 29340 }, { "epoch": 108.7, "grad_norm": 0.9479579925537109, "learning_rate": 0.000912962962962963, "loss": 1.1195, "step": 29350 }, { "epoch": 108.74, "grad_norm": 0.9171406626701355, "learning_rate": 0.0009125925925925927, "loss": 1.1071, "step": 29360 }, { "epoch": 108.78, "grad_norm": 0.9832227230072021, "learning_rate": 0.0009122222222222223, "loss": 1.1309, "step": 29370 }, { "epoch": 108.81, "grad_norm": 1.0037956237792969, "learning_rate": 0.0009118518518518518, "loss": 1.1322, "step": 29380 }, { "epoch": 108.85, "grad_norm": 1.0154857635498047, "learning_rate": 0.0009114814814814814, "loss": 1.1363, "step": 29390 }, { "epoch": 108.89, "grad_norm": 0.9449747204780579, "learning_rate": 0.0009111111111111111, "loss": 1.1613, "step": 29400 }, { "epoch": 108.93, "grad_norm": 0.9238153696060181, "learning_rate": 0.0009107407407407407, "loss": 1.1424, "step": 29410 }, { "epoch": 108.96, "grad_norm": 0.9596208930015564, "learning_rate": 0.0009103703703703704, "loss": 1.1441, "step": 29420 }, { "epoch": 109.0, "grad_norm": 1.8561451435089111, "learning_rate": 0.00091, "loss": 1.1483, "step": 29430 }, { "epoch": 109.04, "grad_norm": 0.9528045654296875, "learning_rate": 0.0009096296296296297, "loss": 0.9567, "step": 29440 }, { "epoch": 109.07, "grad_norm": 0.9137498736381531, "learning_rate": 0.0009092592592592592, "loss": 0.9376, "step": 29450 }, { "epoch": 109.11, "grad_norm": 0.8889475464820862, "learning_rate": 0.0009088888888888889, "loss": 0.9589, "step": 29460 }, { "epoch": 109.15, "grad_norm": 0.9228796362876892, "learning_rate": 0.0009085185185185186, "loss": 0.9657, "step": 29470 }, { "epoch": 109.19, "grad_norm": 0.9670528173446655, "learning_rate": 0.0009081481481481482, "loss": 0.9863, "step": 29480 }, { "epoch": 109.22, "grad_norm": 1.0339789390563965, "learning_rate": 0.0009077777777777779, "loss": 0.9995, "step": 29490 }, { "epoch": 109.26, "grad_norm": 0.9455500245094299, "learning_rate": 0.0009074074074074074, "loss": 1.0002, "step": 29500 }, { "epoch": 109.3, "grad_norm": 0.9690384268760681, "learning_rate": 0.0009070370370370371, "loss": 1.0194, "step": 29510 }, { "epoch": 109.33, "grad_norm": 0.9900230765342712, "learning_rate": 0.0009066666666666666, "loss": 1.0351, "step": 29520 }, { "epoch": 109.37, "grad_norm": 1.0021939277648926, "learning_rate": 0.0009062962962962963, "loss": 1.0486, "step": 29530 }, { "epoch": 109.41, "grad_norm": 0.9087638854980469, "learning_rate": 0.0009059259259259259, "loss": 1.0461, "step": 29540 }, { "epoch": 109.44, "grad_norm": 0.8851718306541443, "learning_rate": 0.0009055555555555556, "loss": 1.0577, "step": 29550 }, { "epoch": 109.48, "grad_norm": 0.9451716542243958, "learning_rate": 0.0009051851851851852, "loss": 1.0504, "step": 29560 }, { "epoch": 109.52, "grad_norm": 0.9180352091789246, "learning_rate": 0.0009048148148148149, "loss": 1.0896, "step": 29570 }, { "epoch": 109.56, "grad_norm": 0.9067238569259644, "learning_rate": 0.0009044444444444445, "loss": 1.0742, "step": 29580 }, { "epoch": 109.59, "grad_norm": 0.9143742918968201, "learning_rate": 0.0009040740740740741, "loss": 1.0828, "step": 29590 }, { "epoch": 109.63, "grad_norm": 0.95147305727005, "learning_rate": 0.0009037037037037037, "loss": 1.0835, "step": 29600 }, { "epoch": 109.67, "grad_norm": 0.9197315573692322, "learning_rate": 0.0009033333333333334, "loss": 1.1058, "step": 29610 }, { "epoch": 109.7, "grad_norm": 1.0208014249801636, "learning_rate": 0.0009029629629629629, "loss": 1.1029, "step": 29620 }, { "epoch": 109.74, "grad_norm": 0.9763293266296387, "learning_rate": 0.0009025925925925926, "loss": 1.0799, "step": 29630 }, { "epoch": 109.78, "grad_norm": 0.9239047169685364, "learning_rate": 0.0009022222222222222, "loss": 1.1195, "step": 29640 }, { "epoch": 109.81, "grad_norm": 0.9763393998146057, "learning_rate": 0.0009018518518518519, "loss": 1.1211, "step": 29650 }, { "epoch": 109.85, "grad_norm": 0.9521705508232117, "learning_rate": 0.0009014814814814814, "loss": 1.1382, "step": 29660 }, { "epoch": 109.89, "grad_norm": 0.9943112134933472, "learning_rate": 0.0009011111111111111, "loss": 1.1134, "step": 29670 }, { "epoch": 109.93, "grad_norm": 0.9353739023208618, "learning_rate": 0.0009007407407407408, "loss": 1.1501, "step": 29680 }, { "epoch": 109.96, "grad_norm": 0.9845727682113647, "learning_rate": 0.0009003703703703704, "loss": 1.1209, "step": 29690 }, { "epoch": 110.0, "grad_norm": 1.6984339952468872, "learning_rate": 0.0009000000000000001, "loss": 1.1666, "step": 29700 }, { "epoch": 110.04, "grad_norm": 0.9160362482070923, "learning_rate": 0.0008996296296296297, "loss": 0.9263, "step": 29710 }, { "epoch": 110.07, "grad_norm": 0.908966600894928, "learning_rate": 0.0008992592592592594, "loss": 0.9175, "step": 29720 }, { "epoch": 110.11, "grad_norm": 0.9472144246101379, "learning_rate": 0.0008988888888888888, "loss": 0.935, "step": 29730 }, { "epoch": 110.15, "grad_norm": 0.9157045483589172, "learning_rate": 0.0008985185185185185, "loss": 0.9417, "step": 29740 }, { "epoch": 110.19, "grad_norm": 0.9928398728370667, "learning_rate": 0.0008981481481481481, "loss": 0.9652, "step": 29750 }, { "epoch": 110.22, "grad_norm": 0.9916333556175232, "learning_rate": 0.0008977777777777778, "loss": 1.0001, "step": 29760 }, { "epoch": 110.26, "grad_norm": 0.9936506152153015, "learning_rate": 0.0008974074074074074, "loss": 0.9845, "step": 29770 }, { "epoch": 110.3, "grad_norm": 0.9548264145851135, "learning_rate": 0.0008970370370370371, "loss": 0.9968, "step": 29780 }, { "epoch": 110.33, "grad_norm": 0.9424086213111877, "learning_rate": 0.0008966666666666666, "loss": 0.9978, "step": 29790 }, { "epoch": 110.37, "grad_norm": 0.8982470035552979, "learning_rate": 0.0008962962962962963, "loss": 1.0418, "step": 29800 }, { "epoch": 110.41, "grad_norm": 0.9713009595870972, "learning_rate": 0.0008959259259259259, "loss": 1.051, "step": 29810 }, { "epoch": 110.44, "grad_norm": 0.9629430174827576, "learning_rate": 0.0008955555555555556, "loss": 1.0485, "step": 29820 }, { "epoch": 110.48, "grad_norm": 1.0035597085952759, "learning_rate": 0.0008951851851851852, "loss": 1.0479, "step": 29830 }, { "epoch": 110.52, "grad_norm": 1.0095973014831543, "learning_rate": 0.0008948148148148149, "loss": 1.074, "step": 29840 }, { "epoch": 110.56, "grad_norm": 1.0107605457305908, "learning_rate": 0.0008944444444444445, "loss": 1.0728, "step": 29850 }, { "epoch": 110.59, "grad_norm": 0.9929389357566833, "learning_rate": 0.000894074074074074, "loss": 1.0744, "step": 29860 }, { "epoch": 110.63, "grad_norm": 0.9642514586448669, "learning_rate": 0.0008937037037037037, "loss": 1.1, "step": 29870 }, { "epoch": 110.67, "grad_norm": 0.9914767742156982, "learning_rate": 0.0008933333333333333, "loss": 1.0979, "step": 29880 }, { "epoch": 110.7, "grad_norm": 0.9608526825904846, "learning_rate": 0.000892962962962963, "loss": 1.104, "step": 29890 }, { "epoch": 110.74, "grad_norm": 0.9871772527694702, "learning_rate": 0.0008925925925925926, "loss": 1.1148, "step": 29900 }, { "epoch": 110.78, "grad_norm": 0.9967049956321716, "learning_rate": 0.0008922222222222223, "loss": 1.1284, "step": 29910 }, { "epoch": 110.81, "grad_norm": 1.0077149868011475, "learning_rate": 0.0008918518518518519, "loss": 1.1268, "step": 29920 }, { "epoch": 110.85, "grad_norm": 1.011237382888794, "learning_rate": 0.0008914814814814815, "loss": 1.1051, "step": 29930 }, { "epoch": 110.89, "grad_norm": 0.973310112953186, "learning_rate": 0.0008911111111111111, "loss": 1.1182, "step": 29940 }, { "epoch": 110.93, "grad_norm": 0.9287843108177185, "learning_rate": 0.0008907407407407408, "loss": 1.1358, "step": 29950 }, { "epoch": 110.96, "grad_norm": 0.9778017997741699, "learning_rate": 0.0008903703703703704, "loss": 1.1474, "step": 29960 }, { "epoch": 111.0, "grad_norm": 1.8783273696899414, "learning_rate": 0.0008900000000000001, "loss": 1.127, "step": 29970 }, { "epoch": 111.04, "grad_norm": 0.9425782561302185, "learning_rate": 0.0008896296296296296, "loss": 0.9075, "step": 29980 }, { "epoch": 111.07, "grad_norm": 0.9134290814399719, "learning_rate": 0.0008892592592592593, "loss": 0.934, "step": 29990 }, { "epoch": 111.11, "grad_norm": 0.9246062636375427, "learning_rate": 0.0008888888888888888, "loss": 0.9395, "step": 30000 }, { "epoch": 111.15, "grad_norm": 0.9228861331939697, "learning_rate": 0.0008885185185185185, "loss": 0.9561, "step": 30010 }, { "epoch": 111.19, "grad_norm": 0.9056755304336548, "learning_rate": 0.0008881481481481481, "loss": 0.9332, "step": 30020 }, { "epoch": 111.22, "grad_norm": 0.8870821595191956, "learning_rate": 0.0008877777777777778, "loss": 0.9757, "step": 30030 }, { "epoch": 111.26, "grad_norm": 0.989964485168457, "learning_rate": 0.0008874074074074074, "loss": 0.9914, "step": 30040 }, { "epoch": 111.3, "grad_norm": 0.9611079692840576, "learning_rate": 0.0008870370370370371, "loss": 0.9961, "step": 30050 }, { "epoch": 111.33, "grad_norm": 0.9657167792320251, "learning_rate": 0.0008866666666666667, "loss": 0.9881, "step": 30060 }, { "epoch": 111.37, "grad_norm": 0.9873797297477722, "learning_rate": 0.0008862962962962963, "loss": 1.0194, "step": 30070 }, { "epoch": 111.41, "grad_norm": 0.9685762524604797, "learning_rate": 0.000885925925925926, "loss": 1.0231, "step": 30080 }, { "epoch": 111.44, "grad_norm": 0.9828411340713501, "learning_rate": 0.0008855555555555556, "loss": 1.0538, "step": 30090 }, { "epoch": 111.48, "grad_norm": 0.9536802172660828, "learning_rate": 0.0008851851851851853, "loss": 1.0479, "step": 30100 }, { "epoch": 111.52, "grad_norm": 0.8964811563491821, "learning_rate": 0.0008848148148148148, "loss": 1.0633, "step": 30110 }, { "epoch": 111.56, "grad_norm": 0.9257550239562988, "learning_rate": 0.0008844444444444445, "loss": 1.056, "step": 30120 }, { "epoch": 111.59, "grad_norm": 0.9024585485458374, "learning_rate": 0.0008840740740740741, "loss": 1.0587, "step": 30130 }, { "epoch": 111.63, "grad_norm": 1.0489718914031982, "learning_rate": 0.0008837037037037037, "loss": 1.0819, "step": 30140 }, { "epoch": 111.67, "grad_norm": 0.9898716807365417, "learning_rate": 0.0008833333333333333, "loss": 1.0785, "step": 30150 }, { "epoch": 111.7, "grad_norm": 0.9546767473220825, "learning_rate": 0.000882962962962963, "loss": 1.1157, "step": 30160 }, { "epoch": 111.74, "grad_norm": 0.9728699922561646, "learning_rate": 0.0008825925925925926, "loss": 1.0798, "step": 30170 }, { "epoch": 111.78, "grad_norm": 0.9877276420593262, "learning_rate": 0.0008822222222222223, "loss": 1.1088, "step": 30180 }, { "epoch": 111.81, "grad_norm": 0.9211771488189697, "learning_rate": 0.0008818518518518519, "loss": 1.117, "step": 30190 }, { "epoch": 111.85, "grad_norm": 1.0157147645950317, "learning_rate": 0.0008814814814814816, "loss": 1.1102, "step": 30200 }, { "epoch": 111.89, "grad_norm": 1.092393398284912, "learning_rate": 0.000881111111111111, "loss": 1.1293, "step": 30210 }, { "epoch": 111.93, "grad_norm": 0.9899991750717163, "learning_rate": 0.0008807407407407407, "loss": 1.1227, "step": 30220 }, { "epoch": 111.96, "grad_norm": 0.9688059091567993, "learning_rate": 0.0008803703703703703, "loss": 1.1286, "step": 30230 }, { "epoch": 112.0, "grad_norm": 1.6932052373886108, "learning_rate": 0.00088, "loss": 1.1382, "step": 30240 }, { "epoch": 112.04, "grad_norm": 0.9143195748329163, "learning_rate": 0.0008796296296296296, "loss": 0.9058, "step": 30250 }, { "epoch": 112.07, "grad_norm": 0.990135908126831, "learning_rate": 0.0008792592592592593, "loss": 0.9267, "step": 30260 }, { "epoch": 112.11, "grad_norm": 0.9176962375640869, "learning_rate": 0.000878888888888889, "loss": 0.917, "step": 30270 }, { "epoch": 112.15, "grad_norm": 0.9764485359191895, "learning_rate": 0.0008785185185185185, "loss": 0.9068, "step": 30280 }, { "epoch": 112.19, "grad_norm": 0.9197412133216858, "learning_rate": 0.0008781481481481482, "loss": 0.9584, "step": 30290 }, { "epoch": 112.22, "grad_norm": 1.0627151727676392, "learning_rate": 0.0008777777777777778, "loss": 0.9894, "step": 30300 }, { "epoch": 112.26, "grad_norm": 0.9765280485153198, "learning_rate": 0.0008774074074074075, "loss": 0.9884, "step": 30310 }, { "epoch": 112.3, "grad_norm": 0.9730259776115417, "learning_rate": 0.0008770370370370371, "loss": 0.9733, "step": 30320 }, { "epoch": 112.33, "grad_norm": 0.9327222108840942, "learning_rate": 0.0008766666666666668, "loss": 1.016, "step": 30330 }, { "epoch": 112.37, "grad_norm": 0.9560900330543518, "learning_rate": 0.0008762962962962964, "loss": 1.0147, "step": 30340 }, { "epoch": 112.41, "grad_norm": 0.9207103252410889, "learning_rate": 0.0008759259259259259, "loss": 1.003, "step": 30350 }, { "epoch": 112.44, "grad_norm": 1.0215859413146973, "learning_rate": 0.0008755555555555555, "loss": 1.0212, "step": 30360 }, { "epoch": 112.48, "grad_norm": 0.9427470564842224, "learning_rate": 0.0008751851851851852, "loss": 1.0495, "step": 30370 }, { "epoch": 112.52, "grad_norm": 1.003307819366455, "learning_rate": 0.0008748148148148148, "loss": 1.0227, "step": 30380 }, { "epoch": 112.56, "grad_norm": 1.093425989151001, "learning_rate": 0.0008744444444444445, "loss": 1.0566, "step": 30390 }, { "epoch": 112.59, "grad_norm": 0.9261013269424438, "learning_rate": 0.0008740740740740741, "loss": 1.0736, "step": 30400 }, { "epoch": 112.63, "grad_norm": 0.9942784309387207, "learning_rate": 0.0008737037037037037, "loss": 1.0688, "step": 30410 }, { "epoch": 112.67, "grad_norm": 0.9616612792015076, "learning_rate": 0.0008733333333333333, "loss": 1.0768, "step": 30420 }, { "epoch": 112.7, "grad_norm": 0.9479864239692688, "learning_rate": 0.000872962962962963, "loss": 1.1053, "step": 30430 }, { "epoch": 112.74, "grad_norm": 1.0143781900405884, "learning_rate": 0.0008725925925925926, "loss": 1.0862, "step": 30440 }, { "epoch": 112.78, "grad_norm": 1.0063732862472534, "learning_rate": 0.0008722222222222223, "loss": 1.114, "step": 30450 }, { "epoch": 112.81, "grad_norm": 0.9845717549324036, "learning_rate": 0.0008718518518518518, "loss": 1.079, "step": 30460 }, { "epoch": 112.85, "grad_norm": 0.9684877991676331, "learning_rate": 0.0008714814814814815, "loss": 1.1111, "step": 30470 }, { "epoch": 112.89, "grad_norm": 0.9328610301017761, "learning_rate": 0.000871111111111111, "loss": 1.1305, "step": 30480 }, { "epoch": 112.93, "grad_norm": 0.9969044327735901, "learning_rate": 0.0008707407407407407, "loss": 1.1102, "step": 30490 }, { "epoch": 112.96, "grad_norm": 1.059865117073059, "learning_rate": 0.0008703703703703704, "loss": 1.1202, "step": 30500 }, { "epoch": 113.0, "grad_norm": 2.0708632469177246, "learning_rate": 0.00087, "loss": 1.1286, "step": 30510 }, { "epoch": 113.04, "grad_norm": 0.9408242702484131, "learning_rate": 0.0008696296296296297, "loss": 0.8969, "step": 30520 }, { "epoch": 113.07, "grad_norm": 0.9614538550376892, "learning_rate": 0.0008692592592592593, "loss": 0.9021, "step": 30530 }, { "epoch": 113.11, "grad_norm": 0.9190179109573364, "learning_rate": 0.000868888888888889, "loss": 0.9108, "step": 30540 }, { "epoch": 113.15, "grad_norm": 0.9450747966766357, "learning_rate": 0.0008685185185185185, "loss": 0.9466, "step": 30550 }, { "epoch": 113.19, "grad_norm": 0.9533342123031616, "learning_rate": 0.0008681481481481482, "loss": 0.9656, "step": 30560 }, { "epoch": 113.22, "grad_norm": 0.9292343258857727, "learning_rate": 0.0008677777777777778, "loss": 0.9565, "step": 30570 }, { "epoch": 113.26, "grad_norm": 0.9235038757324219, "learning_rate": 0.0008674074074074074, "loss": 0.9754, "step": 30580 }, { "epoch": 113.3, "grad_norm": 1.0145610570907593, "learning_rate": 0.000867037037037037, "loss": 0.9838, "step": 30590 }, { "epoch": 113.33, "grad_norm": 0.9950644373893738, "learning_rate": 0.0008666666666666667, "loss": 0.9909, "step": 30600 }, { "epoch": 113.37, "grad_norm": 1.0521272420883179, "learning_rate": 0.0008662962962962963, "loss": 0.9892, "step": 30610 }, { "epoch": 113.41, "grad_norm": 0.9194403886795044, "learning_rate": 0.0008659259259259259, "loss": 1.0081, "step": 30620 }, { "epoch": 113.44, "grad_norm": 0.9850278496742249, "learning_rate": 0.0008655555555555555, "loss": 1.0406, "step": 30630 }, { "epoch": 113.48, "grad_norm": 0.971551239490509, "learning_rate": 0.0008651851851851852, "loss": 1.0378, "step": 30640 }, { "epoch": 113.52, "grad_norm": 0.9850895404815674, "learning_rate": 0.0008648148148148148, "loss": 1.0316, "step": 30650 }, { "epoch": 113.56, "grad_norm": 0.9480496048927307, "learning_rate": 0.0008644444444444445, "loss": 1.0412, "step": 30660 }, { "epoch": 113.59, "grad_norm": 1.0179239511489868, "learning_rate": 0.0008640740740740741, "loss": 1.0276, "step": 30670 }, { "epoch": 113.63, "grad_norm": 0.9929103255271912, "learning_rate": 0.0008637037037037038, "loss": 1.0659, "step": 30680 }, { "epoch": 113.67, "grad_norm": 1.006871223449707, "learning_rate": 0.0008633333333333334, "loss": 1.0808, "step": 30690 }, { "epoch": 113.7, "grad_norm": 0.990407407283783, "learning_rate": 0.0008629629629629629, "loss": 1.0747, "step": 30700 }, { "epoch": 113.74, "grad_norm": 1.0033669471740723, "learning_rate": 0.0008625925925925926, "loss": 1.0571, "step": 30710 }, { "epoch": 113.78, "grad_norm": 0.9885395169258118, "learning_rate": 0.0008622222222222222, "loss": 1.0932, "step": 30720 }, { "epoch": 113.81, "grad_norm": 0.9462271928787231, "learning_rate": 0.0008618518518518519, "loss": 1.0848, "step": 30730 }, { "epoch": 113.85, "grad_norm": 0.9825132489204407, "learning_rate": 0.0008614814814814815, "loss": 1.1212, "step": 30740 }, { "epoch": 113.89, "grad_norm": 0.9224004149436951, "learning_rate": 0.0008611111111111112, "loss": 1.1188, "step": 30750 }, { "epoch": 113.93, "grad_norm": 1.0247914791107178, "learning_rate": 0.0008607407407407407, "loss": 1.1017, "step": 30760 }, { "epoch": 113.96, "grad_norm": 1.016513705253601, "learning_rate": 0.0008603703703703704, "loss": 1.1242, "step": 30770 }, { "epoch": 114.0, "grad_norm": 1.9813746213912964, "learning_rate": 0.00086, "loss": 1.1129, "step": 30780 }, { "epoch": 114.04, "grad_norm": 0.9882152676582336, "learning_rate": 0.0008596296296296297, "loss": 0.9152, "step": 30790 }, { "epoch": 114.07, "grad_norm": 0.9574514031410217, "learning_rate": 0.0008592592592592593, "loss": 0.9013, "step": 30800 }, { "epoch": 114.11, "grad_norm": 0.9351339936256409, "learning_rate": 0.000858888888888889, "loss": 0.9075, "step": 30810 }, { "epoch": 114.15, "grad_norm": 0.9363587498664856, "learning_rate": 0.0008585185185185185, "loss": 0.9144, "step": 30820 }, { "epoch": 114.19, "grad_norm": 0.9973554015159607, "learning_rate": 0.0008581481481481481, "loss": 0.9334, "step": 30830 }, { "epoch": 114.22, "grad_norm": 1.1108694076538086, "learning_rate": 0.0008577777777777777, "loss": 0.9683, "step": 30840 }, { "epoch": 114.26, "grad_norm": 1.0352013111114502, "learning_rate": 0.0008574074074074074, "loss": 0.9817, "step": 30850 }, { "epoch": 114.3, "grad_norm": 0.9350374341011047, "learning_rate": 0.000857037037037037, "loss": 0.9652, "step": 30860 }, { "epoch": 114.33, "grad_norm": 0.942579984664917, "learning_rate": 0.0008566666666666667, "loss": 0.9818, "step": 30870 }, { "epoch": 114.37, "grad_norm": 1.0011614561080933, "learning_rate": 0.0008562962962962963, "loss": 0.9991, "step": 30880 }, { "epoch": 114.41, "grad_norm": 0.9837273955345154, "learning_rate": 0.000855925925925926, "loss": 0.9953, "step": 30890 }, { "epoch": 114.44, "grad_norm": 0.9239174127578735, "learning_rate": 0.0008555555555555556, "loss": 1.0251, "step": 30900 }, { "epoch": 114.48, "grad_norm": 0.9788716435432434, "learning_rate": 0.0008551851851851852, "loss": 1.011, "step": 30910 }, { "epoch": 114.52, "grad_norm": 1.0957821607589722, "learning_rate": 0.0008548148148148149, "loss": 1.0122, "step": 30920 }, { "epoch": 114.56, "grad_norm": 1.0088284015655518, "learning_rate": 0.0008544444444444445, "loss": 1.0364, "step": 30930 }, { "epoch": 114.59, "grad_norm": 1.0810710191726685, "learning_rate": 0.0008540740740740742, "loss": 1.0542, "step": 30940 }, { "epoch": 114.63, "grad_norm": 1.0543538331985474, "learning_rate": 0.0008537037037037037, "loss": 1.0374, "step": 30950 }, { "epoch": 114.67, "grad_norm": 1.0714164972305298, "learning_rate": 0.0008533333333333334, "loss": 1.0632, "step": 30960 }, { "epoch": 114.7, "grad_norm": 0.9550819993019104, "learning_rate": 0.0008529629629629629, "loss": 1.0696, "step": 30970 }, { "epoch": 114.74, "grad_norm": 1.0119558572769165, "learning_rate": 0.0008525925925925926, "loss": 1.0603, "step": 30980 }, { "epoch": 114.78, "grad_norm": 0.9348704814910889, "learning_rate": 0.0008522222222222222, "loss": 1.1039, "step": 30990 }, { "epoch": 114.81, "grad_norm": 1.025181770324707, "learning_rate": 0.0008518518518518519, "loss": 1.0769, "step": 31000 }, { "epoch": 114.85, "grad_norm": 1.011375904083252, "learning_rate": 0.0008514814814814815, "loss": 1.077, "step": 31010 }, { "epoch": 114.89, "grad_norm": 0.9699951410293579, "learning_rate": 0.0008511111111111112, "loss": 1.1129, "step": 31020 }, { "epoch": 114.93, "grad_norm": 0.9796174168586731, "learning_rate": 0.0008507407407407407, "loss": 1.1002, "step": 31030 }, { "epoch": 114.96, "grad_norm": 0.9660859704017639, "learning_rate": 0.0008503703703703704, "loss": 1.0966, "step": 31040 }, { "epoch": 115.0, "grad_norm": 1.6068373918533325, "learning_rate": 0.00085, "loss": 1.0997, "step": 31050 }, { "epoch": 115.04, "grad_norm": 0.9463887214660645, "learning_rate": 0.0008496296296296296, "loss": 0.8809, "step": 31060 }, { "epoch": 115.07, "grad_norm": 0.9858732223510742, "learning_rate": 0.0008492592592592592, "loss": 0.8918, "step": 31070 }, { "epoch": 115.11, "grad_norm": 1.081920862197876, "learning_rate": 0.0008488888888888889, "loss": 0.9019, "step": 31080 }, { "epoch": 115.15, "grad_norm": 0.914100170135498, "learning_rate": 0.0008485185185185186, "loss": 0.9338, "step": 31090 }, { "epoch": 115.19, "grad_norm": 0.9406884908676147, "learning_rate": 0.0008481481481481481, "loss": 0.924, "step": 31100 }, { "epoch": 115.22, "grad_norm": 1.0523425340652466, "learning_rate": 0.0008477777777777778, "loss": 0.9472, "step": 31110 }, { "epoch": 115.26, "grad_norm": 1.0571317672729492, "learning_rate": 0.0008474074074074074, "loss": 0.9532, "step": 31120 }, { "epoch": 115.3, "grad_norm": 1.1222143173217773, "learning_rate": 0.0008470370370370371, "loss": 0.973, "step": 31130 }, { "epoch": 115.33, "grad_norm": 1.019164800643921, "learning_rate": 0.0008466666666666667, "loss": 0.9732, "step": 31140 }, { "epoch": 115.37, "grad_norm": 0.9789102077484131, "learning_rate": 0.0008462962962962964, "loss": 0.983, "step": 31150 }, { "epoch": 115.41, "grad_norm": 1.0151702165603638, "learning_rate": 0.000845925925925926, "loss": 1.0162, "step": 31160 }, { "epoch": 115.44, "grad_norm": 1.0261796712875366, "learning_rate": 0.0008455555555555556, "loss": 0.9956, "step": 31170 }, { "epoch": 115.48, "grad_norm": 1.0111148357391357, "learning_rate": 0.0008451851851851851, "loss": 1.0375, "step": 31180 }, { "epoch": 115.52, "grad_norm": 0.9380441904067993, "learning_rate": 0.0008448148148148148, "loss": 1.0244, "step": 31190 }, { "epoch": 115.56, "grad_norm": 0.9684368371963501, "learning_rate": 0.0008444444444444444, "loss": 1.0179, "step": 31200 }, { "epoch": 115.59, "grad_norm": 1.1318254470825195, "learning_rate": 0.0008440740740740741, "loss": 1.0402, "step": 31210 }, { "epoch": 115.63, "grad_norm": 0.9918487071990967, "learning_rate": 0.0008437037037037037, "loss": 1.0473, "step": 31220 }, { "epoch": 115.67, "grad_norm": 1.001694917678833, "learning_rate": 0.0008433333333333334, "loss": 1.0424, "step": 31230 }, { "epoch": 115.7, "grad_norm": 0.9491502046585083, "learning_rate": 0.0008429629629629629, "loss": 1.0667, "step": 31240 }, { "epoch": 115.74, "grad_norm": 1.0236819982528687, "learning_rate": 0.0008425925925925926, "loss": 1.0565, "step": 31250 }, { "epoch": 115.78, "grad_norm": 1.0704491138458252, "learning_rate": 0.0008422222222222222, "loss": 1.0906, "step": 31260 }, { "epoch": 115.81, "grad_norm": 1.1185559034347534, "learning_rate": 0.0008418518518518519, "loss": 1.0476, "step": 31270 }, { "epoch": 115.85, "grad_norm": 1.0499783754348755, "learning_rate": 0.0008414814814814815, "loss": 1.0939, "step": 31280 }, { "epoch": 115.89, "grad_norm": 0.9921557307243347, "learning_rate": 0.0008411111111111112, "loss": 1.0777, "step": 31290 }, { "epoch": 115.93, "grad_norm": 1.0125623941421509, "learning_rate": 0.0008407407407407409, "loss": 1.0939, "step": 31300 }, { "epoch": 115.96, "grad_norm": 0.985906720161438, "learning_rate": 0.0008403703703703703, "loss": 1.0973, "step": 31310 }, { "epoch": 116.0, "grad_norm": 2.2523534297943115, "learning_rate": 0.00084, "loss": 1.0789, "step": 31320 }, { "epoch": 116.04, "grad_norm": 0.9431299567222595, "learning_rate": 0.0008396296296296296, "loss": 0.8982, "step": 31330 }, { "epoch": 116.07, "grad_norm": 0.9881550669670105, "learning_rate": 0.0008392592592592593, "loss": 0.8887, "step": 31340 }, { "epoch": 116.11, "grad_norm": 0.985056459903717, "learning_rate": 0.0008388888888888889, "loss": 0.8935, "step": 31350 }, { "epoch": 116.15, "grad_norm": 0.9958218336105347, "learning_rate": 0.0008385185185185186, "loss": 0.9165, "step": 31360 }, { "epoch": 116.19, "grad_norm": 0.9456029534339905, "learning_rate": 0.0008381481481481482, "loss": 0.922, "step": 31370 }, { "epoch": 116.22, "grad_norm": 1.0164287090301514, "learning_rate": 0.0008377777777777778, "loss": 0.9428, "step": 31380 }, { "epoch": 116.26, "grad_norm": 0.9778671264648438, "learning_rate": 0.0008374074074074074, "loss": 0.9509, "step": 31390 }, { "epoch": 116.3, "grad_norm": 1.0401966571807861, "learning_rate": 0.0008370370370370371, "loss": 0.9524, "step": 31400 }, { "epoch": 116.33, "grad_norm": 0.9463452100753784, "learning_rate": 0.0008366666666666667, "loss": 0.9865, "step": 31410 }, { "epoch": 116.37, "grad_norm": 1.0005884170532227, "learning_rate": 0.0008362962962962964, "loss": 0.9834, "step": 31420 }, { "epoch": 116.41, "grad_norm": 1.0348414182662964, "learning_rate": 0.0008359259259259259, "loss": 0.9834, "step": 31430 }, { "epoch": 116.44, "grad_norm": 1.0169605016708374, "learning_rate": 0.0008355555555555556, "loss": 0.9884, "step": 31440 }, { "epoch": 116.48, "grad_norm": 1.0272061824798584, "learning_rate": 0.0008351851851851851, "loss": 1.0093, "step": 31450 }, { "epoch": 116.52, "grad_norm": 1.0215003490447998, "learning_rate": 0.0008348148148148148, "loss": 1.0236, "step": 31460 }, { "epoch": 116.56, "grad_norm": 0.9531969428062439, "learning_rate": 0.0008344444444444444, "loss": 1.0269, "step": 31470 }, { "epoch": 116.59, "grad_norm": 0.9995023012161255, "learning_rate": 0.0008340740740740741, "loss": 1.0146, "step": 31480 }, { "epoch": 116.63, "grad_norm": 0.9878466129302979, "learning_rate": 0.0008337037037037037, "loss": 1.0294, "step": 31490 }, { "epoch": 116.67, "grad_norm": 0.9585810303688049, "learning_rate": 0.0008333333333333334, "loss": 1.0438, "step": 31500 }, { "epoch": 116.7, "grad_norm": 1.0456923246383667, "learning_rate": 0.0008329629629629631, "loss": 1.0627, "step": 31510 }, { "epoch": 116.74, "grad_norm": 0.9957906603813171, "learning_rate": 0.0008325925925925926, "loss": 1.0584, "step": 31520 }, { "epoch": 116.78, "grad_norm": 1.0360143184661865, "learning_rate": 0.0008322222222222223, "loss": 1.0573, "step": 31530 }, { "epoch": 116.81, "grad_norm": 0.935924768447876, "learning_rate": 0.0008318518518518518, "loss": 1.0462, "step": 31540 }, { "epoch": 116.85, "grad_norm": 1.0759156942367554, "learning_rate": 0.0008314814814814815, "loss": 1.0702, "step": 31550 }, { "epoch": 116.89, "grad_norm": 0.9487184286117554, "learning_rate": 0.0008311111111111111, "loss": 1.0916, "step": 31560 }, { "epoch": 116.93, "grad_norm": 0.9857691526412964, "learning_rate": 0.0008307407407407408, "loss": 1.0904, "step": 31570 }, { "epoch": 116.96, "grad_norm": 0.9965876936912537, "learning_rate": 0.0008303703703703704, "loss": 1.0766, "step": 31580 }, { "epoch": 117.0, "grad_norm": 1.8307380676269531, "learning_rate": 0.00083, "loss": 1.092, "step": 31590 }, { "epoch": 117.04, "grad_norm": 0.9478293061256409, "learning_rate": 0.0008296296296296296, "loss": 0.8833, "step": 31600 }, { "epoch": 117.07, "grad_norm": 0.9572622776031494, "learning_rate": 0.0008292592592592593, "loss": 0.869, "step": 31610 }, { "epoch": 117.11, "grad_norm": 0.9900893568992615, "learning_rate": 0.0008288888888888889, "loss": 0.9014, "step": 31620 }, { "epoch": 117.15, "grad_norm": 0.9924251437187195, "learning_rate": 0.0008285185185185186, "loss": 0.9139, "step": 31630 }, { "epoch": 117.19, "grad_norm": 1.012449026107788, "learning_rate": 0.0008281481481481482, "loss": 0.9283, "step": 31640 }, { "epoch": 117.22, "grad_norm": 0.9699010252952576, "learning_rate": 0.0008277777777777778, "loss": 0.9421, "step": 31650 }, { "epoch": 117.26, "grad_norm": 0.9940907955169678, "learning_rate": 0.0008274074074074073, "loss": 0.9248, "step": 31660 }, { "epoch": 117.3, "grad_norm": 1.0201570987701416, "learning_rate": 0.000827037037037037, "loss": 0.9485, "step": 31670 }, { "epoch": 117.33, "grad_norm": 0.9949468970298767, "learning_rate": 0.0008266666666666666, "loss": 0.9616, "step": 31680 }, { "epoch": 117.37, "grad_norm": 1.0526440143585205, "learning_rate": 0.0008262962962962963, "loss": 0.977, "step": 31690 }, { "epoch": 117.41, "grad_norm": 1.0306730270385742, "learning_rate": 0.0008259259259259259, "loss": 0.9835, "step": 31700 }, { "epoch": 117.44, "grad_norm": 1.047856330871582, "learning_rate": 0.0008255555555555556, "loss": 0.9882, "step": 31710 }, { "epoch": 117.48, "grad_norm": 1.017426609992981, "learning_rate": 0.0008251851851851852, "loss": 1.0126, "step": 31720 }, { "epoch": 117.52, "grad_norm": 1.0520775318145752, "learning_rate": 0.0008248148148148148, "loss": 0.9974, "step": 31730 }, { "epoch": 117.56, "grad_norm": 1.0695947408676147, "learning_rate": 0.0008244444444444445, "loss": 1.0, "step": 31740 }, { "epoch": 117.59, "grad_norm": 1.0761319398880005, "learning_rate": 0.0008240740740740741, "loss": 1.0147, "step": 31750 }, { "epoch": 117.63, "grad_norm": 0.9545421004295349, "learning_rate": 0.0008237037037037038, "loss": 1.0315, "step": 31760 }, { "epoch": 117.67, "grad_norm": 1.005906105041504, "learning_rate": 0.0008233333333333334, "loss": 1.0242, "step": 31770 }, { "epoch": 117.7, "grad_norm": 0.9745824933052063, "learning_rate": 0.0008229629629629631, "loss": 1.0309, "step": 31780 }, { "epoch": 117.74, "grad_norm": 1.0264394283294678, "learning_rate": 0.0008225925925925925, "loss": 1.0684, "step": 31790 }, { "epoch": 117.78, "grad_norm": 1.0310300588607788, "learning_rate": 0.0008222222222222222, "loss": 1.0371, "step": 31800 }, { "epoch": 117.81, "grad_norm": 1.017672061920166, "learning_rate": 0.0008218518518518518, "loss": 1.0531, "step": 31810 }, { "epoch": 117.85, "grad_norm": 1.05331552028656, "learning_rate": 0.0008214814814814815, "loss": 1.0585, "step": 31820 }, { "epoch": 117.89, "grad_norm": 0.970564603805542, "learning_rate": 0.0008211111111111111, "loss": 1.0757, "step": 31830 }, { "epoch": 117.93, "grad_norm": 0.9827378392219543, "learning_rate": 0.0008207407407407408, "loss": 1.0806, "step": 31840 }, { "epoch": 117.96, "grad_norm": 1.0662480592727661, "learning_rate": 0.0008203703703703704, "loss": 1.063, "step": 31850 }, { "epoch": 118.0, "grad_norm": 1.6402631998062134, "learning_rate": 0.00082, "loss": 1.0978, "step": 31860 }, { "epoch": 118.04, "grad_norm": 1.0310686826705933, "learning_rate": 0.0008196296296296296, "loss": 0.8643, "step": 31870 }, { "epoch": 118.07, "grad_norm": 0.9705514907836914, "learning_rate": 0.0008192592592592593, "loss": 0.8804, "step": 31880 }, { "epoch": 118.11, "grad_norm": 1.0193184614181519, "learning_rate": 0.0008188888888888889, "loss": 0.9038, "step": 31890 }, { "epoch": 118.15, "grad_norm": 0.9885033965110779, "learning_rate": 0.0008185185185185186, "loss": 0.9113, "step": 31900 }, { "epoch": 118.19, "grad_norm": 1.0788235664367676, "learning_rate": 0.0008181481481481483, "loss": 0.8971, "step": 31910 }, { "epoch": 118.22, "grad_norm": 1.086966633796692, "learning_rate": 0.0008177777777777778, "loss": 0.9177, "step": 31920 }, { "epoch": 118.26, "grad_norm": 1.0231404304504395, "learning_rate": 0.0008174074074074074, "loss": 0.9102, "step": 31930 }, { "epoch": 118.3, "grad_norm": 0.9642085433006287, "learning_rate": 0.000817037037037037, "loss": 0.9613, "step": 31940 }, { "epoch": 118.33, "grad_norm": 0.9916780591011047, "learning_rate": 0.0008166666666666667, "loss": 0.955, "step": 31950 }, { "epoch": 118.37, "grad_norm": 1.0243855714797974, "learning_rate": 0.0008162962962962963, "loss": 0.9787, "step": 31960 }, { "epoch": 118.41, "grad_norm": 1.0688339471817017, "learning_rate": 0.000815925925925926, "loss": 0.9504, "step": 31970 }, { "epoch": 118.44, "grad_norm": 1.0077600479125977, "learning_rate": 0.0008155555555555556, "loss": 0.972, "step": 31980 }, { "epoch": 118.48, "grad_norm": 1.018144965171814, "learning_rate": 0.0008151851851851853, "loss": 0.9725, "step": 31990 }, { "epoch": 118.52, "grad_norm": 0.9333834052085876, "learning_rate": 0.0008148148148148148, "loss": 1.0076, "step": 32000 }, { "epoch": 118.56, "grad_norm": 0.9722175002098083, "learning_rate": 0.0008144444444444445, "loss": 1.0176, "step": 32010 }, { "epoch": 118.59, "grad_norm": 1.027160882949829, "learning_rate": 0.000814074074074074, "loss": 1.0272, "step": 32020 }, { "epoch": 118.63, "grad_norm": 0.9696611762046814, "learning_rate": 0.0008137037037037037, "loss": 1.0268, "step": 32030 }, { "epoch": 118.67, "grad_norm": 0.994503378868103, "learning_rate": 0.0008133333333333333, "loss": 1.0403, "step": 32040 }, { "epoch": 118.7, "grad_norm": 1.0111019611358643, "learning_rate": 0.000812962962962963, "loss": 1.0298, "step": 32050 }, { "epoch": 118.74, "grad_norm": 0.9856045842170715, "learning_rate": 0.0008125925925925926, "loss": 1.02, "step": 32060 }, { "epoch": 118.78, "grad_norm": 1.0325449705123901, "learning_rate": 0.0008122222222222222, "loss": 1.0528, "step": 32070 }, { "epoch": 118.81, "grad_norm": 1.0795844793319702, "learning_rate": 0.0008118518518518518, "loss": 1.0516, "step": 32080 }, { "epoch": 118.85, "grad_norm": 0.968748152256012, "learning_rate": 0.0008114814814814815, "loss": 1.0355, "step": 32090 }, { "epoch": 118.89, "grad_norm": 1.0011259317398071, "learning_rate": 0.0008111111111111111, "loss": 1.0581, "step": 32100 }, { "epoch": 118.93, "grad_norm": 1.0687892436981201, "learning_rate": 0.0008107407407407408, "loss": 1.0671, "step": 32110 }, { "epoch": 118.96, "grad_norm": 1.000532865524292, "learning_rate": 0.0008103703703703705, "loss": 1.074, "step": 32120 }, { "epoch": 119.0, "grad_norm": 1.6631872653961182, "learning_rate": 0.0008100000000000001, "loss": 1.0843, "step": 32130 }, { "epoch": 119.04, "grad_norm": 0.9295353889465332, "learning_rate": 0.0008096296296296297, "loss": 0.8693, "step": 32140 }, { "epoch": 119.07, "grad_norm": 0.9665623307228088, "learning_rate": 0.0008092592592592592, "loss": 0.8577, "step": 32150 }, { "epoch": 119.11, "grad_norm": 1.050161361694336, "learning_rate": 0.0008088888888888889, "loss": 0.8593, "step": 32160 }, { "epoch": 119.15, "grad_norm": 0.9540622234344482, "learning_rate": 0.0008085185185185185, "loss": 0.8811, "step": 32170 }, { "epoch": 119.19, "grad_norm": 0.9798749685287476, "learning_rate": 0.0008081481481481482, "loss": 0.9113, "step": 32180 }, { "epoch": 119.22, "grad_norm": 0.9423299431800842, "learning_rate": 0.0008077777777777778, "loss": 0.9449, "step": 32190 }, { "epoch": 119.26, "grad_norm": 0.8942784070968628, "learning_rate": 0.0008074074074074075, "loss": 0.9102, "step": 32200 }, { "epoch": 119.3, "grad_norm": 0.9834209084510803, "learning_rate": 0.000807037037037037, "loss": 0.9457, "step": 32210 }, { "epoch": 119.33, "grad_norm": 1.0610178709030151, "learning_rate": 0.0008066666666666667, "loss": 0.9484, "step": 32220 }, { "epoch": 119.37, "grad_norm": 0.979119598865509, "learning_rate": 0.0008062962962962963, "loss": 0.9453, "step": 32230 }, { "epoch": 119.41, "grad_norm": 0.9742558598518372, "learning_rate": 0.000805925925925926, "loss": 0.9711, "step": 32240 }, { "epoch": 119.44, "grad_norm": 0.9968915581703186, "learning_rate": 0.0008055555555555556, "loss": 1.0053, "step": 32250 }, { "epoch": 119.48, "grad_norm": 1.0201388597488403, "learning_rate": 0.0008051851851851853, "loss": 0.9567, "step": 32260 }, { "epoch": 119.52, "grad_norm": 1.1031317710876465, "learning_rate": 0.0008048148148148147, "loss": 0.9918, "step": 32270 }, { "epoch": 119.56, "grad_norm": 0.9733071327209473, "learning_rate": 0.0008044444444444444, "loss": 1.0015, "step": 32280 }, { "epoch": 119.59, "grad_norm": 0.9740049242973328, "learning_rate": 0.000804074074074074, "loss": 1.0096, "step": 32290 }, { "epoch": 119.63, "grad_norm": 1.0093234777450562, "learning_rate": 0.0008037037037037037, "loss": 1.0235, "step": 32300 }, { "epoch": 119.67, "grad_norm": 0.9918221831321716, "learning_rate": 0.0008033333333333333, "loss": 1.0267, "step": 32310 }, { "epoch": 119.7, "grad_norm": 1.023028016090393, "learning_rate": 0.000802962962962963, "loss": 1.0352, "step": 32320 }, { "epoch": 119.74, "grad_norm": 1.0290321111679077, "learning_rate": 0.0008025925925925927, "loss": 1.0427, "step": 32330 }, { "epoch": 119.78, "grad_norm": 0.9687529802322388, "learning_rate": 0.0008022222222222222, "loss": 1.0386, "step": 32340 }, { "epoch": 119.81, "grad_norm": 1.0421777963638306, "learning_rate": 0.0008018518518518519, "loss": 1.0382, "step": 32350 }, { "epoch": 119.85, "grad_norm": 1.0451891422271729, "learning_rate": 0.0008014814814814815, "loss": 1.0428, "step": 32360 }, { "epoch": 119.89, "grad_norm": 1.085588812828064, "learning_rate": 0.0008011111111111112, "loss": 1.0628, "step": 32370 }, { "epoch": 119.93, "grad_norm": 1.0221836566925049, "learning_rate": 0.0008007407407407408, "loss": 1.051, "step": 32380 }, { "epoch": 119.96, "grad_norm": 1.0491361618041992, "learning_rate": 0.0008003703703703704, "loss": 1.0423, "step": 32390 }, { "epoch": 120.0, "grad_norm": 1.8813464641571045, "learning_rate": 0.0008, "loss": 1.0748, "step": 32400 }, { "epoch": 120.04, "grad_norm": 0.8277899026870728, "learning_rate": 0.0007996296296296296, "loss": 0.8565, "step": 32410 }, { "epoch": 120.07, "grad_norm": 0.9462749361991882, "learning_rate": 0.0007992592592592592, "loss": 0.8509, "step": 32420 }, { "epoch": 120.11, "grad_norm": 0.9435334801673889, "learning_rate": 0.0007988888888888889, "loss": 0.8491, "step": 32430 }, { "epoch": 120.15, "grad_norm": 1.0299818515777588, "learning_rate": 0.0007985185185185185, "loss": 0.885, "step": 32440 }, { "epoch": 120.19, "grad_norm": 0.9899550080299377, "learning_rate": 0.0007981481481481482, "loss": 0.8785, "step": 32450 }, { "epoch": 120.22, "grad_norm": 1.0667580366134644, "learning_rate": 0.0007977777777777778, "loss": 0.9158, "step": 32460 }, { "epoch": 120.26, "grad_norm": 0.9750791788101196, "learning_rate": 0.0007974074074074075, "loss": 0.9294, "step": 32470 }, { "epoch": 120.3, "grad_norm": 1.0431256294250488, "learning_rate": 0.000797037037037037, "loss": 0.9302, "step": 32480 }, { "epoch": 120.33, "grad_norm": 1.0735743045806885, "learning_rate": 0.0007966666666666667, "loss": 0.9377, "step": 32490 }, { "epoch": 120.37, "grad_norm": 0.9974359273910522, "learning_rate": 0.0007962962962962962, "loss": 0.9559, "step": 32500 }, { "epoch": 120.41, "grad_norm": 1.0248149633407593, "learning_rate": 0.0007959259259259259, "loss": 0.9594, "step": 32510 }, { "epoch": 120.44, "grad_norm": 0.953029215335846, "learning_rate": 0.0007955555555555555, "loss": 0.9662, "step": 32520 }, { "epoch": 120.48, "grad_norm": 0.9739720821380615, "learning_rate": 0.0007951851851851852, "loss": 0.965, "step": 32530 }, { "epoch": 120.52, "grad_norm": 0.9954352378845215, "learning_rate": 0.0007948148148148149, "loss": 0.9945, "step": 32540 }, { "epoch": 120.56, "grad_norm": 1.062570333480835, "learning_rate": 0.0007944444444444444, "loss": 0.9869, "step": 32550 }, { "epoch": 120.59, "grad_norm": 0.9684034585952759, "learning_rate": 0.0007940740740740741, "loss": 1.019, "step": 32560 }, { "epoch": 120.63, "grad_norm": 0.9902993440628052, "learning_rate": 0.0007937037037037037, "loss": 0.9838, "step": 32570 }, { "epoch": 120.67, "grad_norm": 0.9986051917076111, "learning_rate": 0.0007933333333333334, "loss": 1.0076, "step": 32580 }, { "epoch": 120.7, "grad_norm": 1.0536465644836426, "learning_rate": 0.000792962962962963, "loss": 1.0074, "step": 32590 }, { "epoch": 120.74, "grad_norm": 1.090319275856018, "learning_rate": 0.0007925925925925927, "loss": 1.0298, "step": 32600 }, { "epoch": 120.78, "grad_norm": 0.9828684329986572, "learning_rate": 0.0007922222222222223, "loss": 1.0508, "step": 32610 }, { "epoch": 120.81, "grad_norm": 0.9842421412467957, "learning_rate": 0.0007918518518518519, "loss": 1.0391, "step": 32620 }, { "epoch": 120.85, "grad_norm": 1.0293996334075928, "learning_rate": 0.0007914814814814814, "loss": 1.0497, "step": 32630 }, { "epoch": 120.89, "grad_norm": 1.0321550369262695, "learning_rate": 0.0007911111111111111, "loss": 1.0525, "step": 32640 }, { "epoch": 120.93, "grad_norm": 1.0024299621582031, "learning_rate": 0.0007907407407407407, "loss": 1.0228, "step": 32650 }, { "epoch": 120.96, "grad_norm": 0.9814883470535278, "learning_rate": 0.0007903703703703704, "loss": 1.0406, "step": 32660 }, { "epoch": 121.0, "grad_norm": 1.9922078847885132, "learning_rate": 0.00079, "loss": 1.0307, "step": 32670 }, { "epoch": 121.04, "grad_norm": 1.0609437227249146, "learning_rate": 0.0007896296296296297, "loss": 0.8377, "step": 32680 }, { "epoch": 121.07, "grad_norm": 0.9742987751960754, "learning_rate": 0.0007892592592592592, "loss": 0.8462, "step": 32690 }, { "epoch": 121.11, "grad_norm": 0.940413773059845, "learning_rate": 0.0007888888888888889, "loss": 0.8666, "step": 32700 }, { "epoch": 121.15, "grad_norm": 0.9640865325927734, "learning_rate": 0.0007885185185185185, "loss": 0.8805, "step": 32710 }, { "epoch": 121.19, "grad_norm": 1.014676809310913, "learning_rate": 0.0007881481481481482, "loss": 0.878, "step": 32720 }, { "epoch": 121.22, "grad_norm": 0.9674298167228699, "learning_rate": 0.0007877777777777779, "loss": 0.8901, "step": 32730 }, { "epoch": 121.26, "grad_norm": 1.026246190071106, "learning_rate": 0.0007874074074074075, "loss": 0.9114, "step": 32740 }, { "epoch": 121.3, "grad_norm": 1.008767008781433, "learning_rate": 0.0007870370370370372, "loss": 0.9209, "step": 32750 }, { "epoch": 121.33, "grad_norm": 0.9800826907157898, "learning_rate": 0.0007866666666666666, "loss": 0.9407, "step": 32760 }, { "epoch": 121.37, "grad_norm": 1.0001041889190674, "learning_rate": 0.0007862962962962963, "loss": 0.9482, "step": 32770 }, { "epoch": 121.41, "grad_norm": 0.9887999296188354, "learning_rate": 0.0007859259259259259, "loss": 0.9462, "step": 32780 }, { "epoch": 121.44, "grad_norm": 0.9700247049331665, "learning_rate": 0.0007855555555555556, "loss": 0.9602, "step": 32790 }, { "epoch": 121.48, "grad_norm": 1.0221569538116455, "learning_rate": 0.0007851851851851852, "loss": 0.9525, "step": 32800 }, { "epoch": 121.52, "grad_norm": 1.010459065437317, "learning_rate": 0.0007848148148148149, "loss": 0.9699, "step": 32810 }, { "epoch": 121.56, "grad_norm": 0.9542679786682129, "learning_rate": 0.0007844444444444445, "loss": 0.9942, "step": 32820 }, { "epoch": 121.59, "grad_norm": 1.0301337242126465, "learning_rate": 0.0007840740740740741, "loss": 0.9792, "step": 32830 }, { "epoch": 121.63, "grad_norm": 1.0495370626449585, "learning_rate": 0.0007837037037037037, "loss": 0.9873, "step": 32840 }, { "epoch": 121.67, "grad_norm": 0.9977912902832031, "learning_rate": 0.0007833333333333334, "loss": 0.9885, "step": 32850 }, { "epoch": 121.7, "grad_norm": 1.0673377513885498, "learning_rate": 0.000782962962962963, "loss": 1.006, "step": 32860 }, { "epoch": 121.74, "grad_norm": 0.8986328840255737, "learning_rate": 0.0007825925925925926, "loss": 1.028, "step": 32870 }, { "epoch": 121.78, "grad_norm": 0.9932073354721069, "learning_rate": 0.0007822222222222222, "loss": 1.0151, "step": 32880 }, { "epoch": 121.81, "grad_norm": 0.9659121036529541, "learning_rate": 0.0007818518518518518, "loss": 1.0158, "step": 32890 }, { "epoch": 121.85, "grad_norm": 1.0076465606689453, "learning_rate": 0.0007814814814814814, "loss": 1.0401, "step": 32900 }, { "epoch": 121.89, "grad_norm": 0.9795430898666382, "learning_rate": 0.0007811111111111111, "loss": 1.051, "step": 32910 }, { "epoch": 121.93, "grad_norm": 1.0545165538787842, "learning_rate": 0.0007807407407407407, "loss": 1.0493, "step": 32920 }, { "epoch": 121.96, "grad_norm": 1.021371603012085, "learning_rate": 0.0007803703703703704, "loss": 1.035, "step": 32930 }, { "epoch": 122.0, "grad_norm": 1.738236904144287, "learning_rate": 0.0007800000000000001, "loss": 1.0769, "step": 32940 }, { "epoch": 122.04, "grad_norm": 0.9248874187469482, "learning_rate": 0.0007796296296296297, "loss": 0.8293, "step": 32950 }, { "epoch": 122.07, "grad_norm": 0.9658506512641907, "learning_rate": 0.0007792592592592593, "loss": 0.829, "step": 32960 }, { "epoch": 122.11, "grad_norm": 0.9685617089271545, "learning_rate": 0.0007788888888888889, "loss": 0.8443, "step": 32970 }, { "epoch": 122.15, "grad_norm": 1.0525528192520142, "learning_rate": 0.0007785185185185186, "loss": 0.8754, "step": 32980 }, { "epoch": 122.19, "grad_norm": 0.9952585101127625, "learning_rate": 0.0007781481481481481, "loss": 0.8739, "step": 32990 }, { "epoch": 122.22, "grad_norm": 0.993129312992096, "learning_rate": 0.0007777777777777778, "loss": 0.8789, "step": 33000 }, { "epoch": 122.26, "grad_norm": 0.9832338094711304, "learning_rate": 0.0007774074074074074, "loss": 0.9105, "step": 33010 }, { "epoch": 122.3, "grad_norm": 0.9596456289291382, "learning_rate": 0.0007770370370370371, "loss": 0.9036, "step": 33020 }, { "epoch": 122.33, "grad_norm": 0.9873278141021729, "learning_rate": 0.0007766666666666666, "loss": 0.9134, "step": 33030 }, { "epoch": 122.37, "grad_norm": 0.9594837427139282, "learning_rate": 0.0007762962962962963, "loss": 0.9609, "step": 33040 }, { "epoch": 122.41, "grad_norm": 1.0137587785720825, "learning_rate": 0.0007759259259259259, "loss": 0.9351, "step": 33050 }, { "epoch": 122.44, "grad_norm": 0.9831103086471558, "learning_rate": 0.0007755555555555556, "loss": 0.9798, "step": 33060 }, { "epoch": 122.48, "grad_norm": 1.0451667308807373, "learning_rate": 0.0007751851851851852, "loss": 0.9436, "step": 33070 }, { "epoch": 122.52, "grad_norm": 1.0881359577178955, "learning_rate": 0.0007748148148148149, "loss": 0.9354, "step": 33080 }, { "epoch": 122.56, "grad_norm": 0.9605501890182495, "learning_rate": 0.0007744444444444445, "loss": 0.9794, "step": 33090 }, { "epoch": 122.59, "grad_norm": 1.031227707862854, "learning_rate": 0.000774074074074074, "loss": 0.9752, "step": 33100 }, { "epoch": 122.63, "grad_norm": 0.9974759817123413, "learning_rate": 0.0007737037037037036, "loss": 1.0123, "step": 33110 }, { "epoch": 122.67, "grad_norm": 1.045825481414795, "learning_rate": 0.0007733333333333333, "loss": 0.9922, "step": 33120 }, { "epoch": 122.7, "grad_norm": 1.1240839958190918, "learning_rate": 0.0007729629629629629, "loss": 1.0032, "step": 33130 }, { "epoch": 122.74, "grad_norm": 1.0099389553070068, "learning_rate": 0.0007725925925925926, "loss": 1.0178, "step": 33140 }, { "epoch": 122.78, "grad_norm": 1.0204278230667114, "learning_rate": 0.0007722222222222223, "loss": 1.0352, "step": 33150 }, { "epoch": 122.81, "grad_norm": 0.9793760776519775, "learning_rate": 0.0007718518518518519, "loss": 1.0066, "step": 33160 }, { "epoch": 122.85, "grad_norm": 1.0178688764572144, "learning_rate": 0.0007714814814814815, "loss": 1.0268, "step": 33170 }, { "epoch": 122.89, "grad_norm": 1.049781084060669, "learning_rate": 0.0007711111111111111, "loss": 1.0268, "step": 33180 }, { "epoch": 122.93, "grad_norm": 1.0570471286773682, "learning_rate": 0.0007707407407407408, "loss": 1.0192, "step": 33190 }, { "epoch": 122.96, "grad_norm": 1.07374107837677, "learning_rate": 0.0007703703703703704, "loss": 1.0386, "step": 33200 }, { "epoch": 123.0, "grad_norm": 1.6337435245513916, "learning_rate": 0.0007700000000000001, "loss": 1.0267, "step": 33210 }, { "epoch": 123.04, "grad_norm": 0.9590765833854675, "learning_rate": 0.0007696296296296297, "loss": 0.8187, "step": 33220 }, { "epoch": 123.07, "grad_norm": 0.9379592537879944, "learning_rate": 0.0007692592592592594, "loss": 0.8466, "step": 33230 }, { "epoch": 123.11, "grad_norm": 0.9660229086875916, "learning_rate": 0.0007688888888888888, "loss": 0.8508, "step": 33240 }, { "epoch": 123.15, "grad_norm": 1.0518168210983276, "learning_rate": 0.0007685185185185185, "loss": 0.8775, "step": 33250 }, { "epoch": 123.19, "grad_norm": 0.9530587792396545, "learning_rate": 0.0007681481481481481, "loss": 0.8695, "step": 33260 }, { "epoch": 123.22, "grad_norm": 1.0211882591247559, "learning_rate": 0.0007677777777777778, "loss": 0.8615, "step": 33270 }, { "epoch": 123.26, "grad_norm": 1.0455659627914429, "learning_rate": 0.0007674074074074074, "loss": 0.8946, "step": 33280 }, { "epoch": 123.3, "grad_norm": 1.0107338428497314, "learning_rate": 0.0007670370370370371, "loss": 0.9007, "step": 33290 }, { "epoch": 123.33, "grad_norm": 1.074458122253418, "learning_rate": 0.0007666666666666667, "loss": 0.9061, "step": 33300 }, { "epoch": 123.37, "grad_norm": 1.048465371131897, "learning_rate": 0.0007662962962962963, "loss": 0.9312, "step": 33310 }, { "epoch": 123.41, "grad_norm": 1.0410969257354736, "learning_rate": 0.0007659259259259259, "loss": 0.9258, "step": 33320 }, { "epoch": 123.44, "grad_norm": 1.0864170789718628, "learning_rate": 0.0007655555555555556, "loss": 0.9429, "step": 33330 }, { "epoch": 123.48, "grad_norm": 0.9920828342437744, "learning_rate": 0.0007651851851851852, "loss": 0.941, "step": 33340 }, { "epoch": 123.52, "grad_norm": 1.063452959060669, "learning_rate": 0.0007648148148148148, "loss": 0.9541, "step": 33350 }, { "epoch": 123.56, "grad_norm": 0.9817026853561401, "learning_rate": 0.0007644444444444445, "loss": 0.9551, "step": 33360 }, { "epoch": 123.59, "grad_norm": 0.9735251069068909, "learning_rate": 0.0007640740740740741, "loss": 0.9821, "step": 33370 }, { "epoch": 123.63, "grad_norm": 1.1082037687301636, "learning_rate": 0.0007637037037037037, "loss": 0.9739, "step": 33380 }, { "epoch": 123.67, "grad_norm": 1.0017740726470947, "learning_rate": 0.0007633333333333333, "loss": 1.0015, "step": 33390 }, { "epoch": 123.7, "grad_norm": 1.0714524984359741, "learning_rate": 0.000762962962962963, "loss": 1.0068, "step": 33400 }, { "epoch": 123.74, "grad_norm": 1.0269007682800293, "learning_rate": 0.0007625925925925926, "loss": 0.9995, "step": 33410 }, { "epoch": 123.78, "grad_norm": 0.9959019422531128, "learning_rate": 0.0007622222222222223, "loss": 0.9977, "step": 33420 }, { "epoch": 123.81, "grad_norm": 1.0212533473968506, "learning_rate": 0.0007618518518518519, "loss": 1.0152, "step": 33430 }, { "epoch": 123.85, "grad_norm": 1.0144740343093872, "learning_rate": 0.0007614814814814816, "loss": 1.0319, "step": 33440 }, { "epoch": 123.89, "grad_norm": 1.0200554132461548, "learning_rate": 0.0007611111111111111, "loss": 0.9977, "step": 33450 }, { "epoch": 123.93, "grad_norm": 1.09546959400177, "learning_rate": 0.0007607407407407408, "loss": 1.0041, "step": 33460 }, { "epoch": 123.96, "grad_norm": 1.0374383926391602, "learning_rate": 0.0007603703703703703, "loss": 1.0376, "step": 33470 }, { "epoch": 124.0, "grad_norm": 1.8594433069229126, "learning_rate": 0.00076, "loss": 1.0343, "step": 33480 }, { "epoch": 124.04, "grad_norm": 0.9433583617210388, "learning_rate": 0.0007596296296296296, "loss": 0.8191, "step": 33490 }, { "epoch": 124.07, "grad_norm": 0.9556878805160522, "learning_rate": 0.0007592592592592593, "loss": 0.8319, "step": 33500 }, { "epoch": 124.11, "grad_norm": 1.0316696166992188, "learning_rate": 0.0007588888888888888, "loss": 0.832, "step": 33510 }, { "epoch": 124.15, "grad_norm": 1.0155198574066162, "learning_rate": 0.0007585185185185185, "loss": 0.8381, "step": 33520 }, { "epoch": 124.19, "grad_norm": 0.9573029279708862, "learning_rate": 0.0007581481481481481, "loss": 0.8658, "step": 33530 }, { "epoch": 124.22, "grad_norm": 1.0463179349899292, "learning_rate": 0.0007577777777777778, "loss": 0.8555, "step": 33540 }, { "epoch": 124.26, "grad_norm": 0.9773217439651489, "learning_rate": 0.0007574074074074075, "loss": 0.8756, "step": 33550 }, { "epoch": 124.3, "grad_norm": 1.0205600261688232, "learning_rate": 0.0007570370370370371, "loss": 0.893, "step": 33560 }, { "epoch": 124.33, "grad_norm": 1.0751043558120728, "learning_rate": 0.0007566666666666668, "loss": 0.9006, "step": 33570 }, { "epoch": 124.37, "grad_norm": 1.078392744064331, "learning_rate": 0.0007562962962962963, "loss": 0.9209, "step": 33580 }, { "epoch": 124.41, "grad_norm": 1.0406211614608765, "learning_rate": 0.000755925925925926, "loss": 0.9654, "step": 33590 }, { "epoch": 124.44, "grad_norm": 1.0287216901779175, "learning_rate": 0.0007555555555555555, "loss": 0.942, "step": 33600 }, { "epoch": 124.48, "grad_norm": 1.0419343709945679, "learning_rate": 0.0007551851851851852, "loss": 0.9506, "step": 33610 }, { "epoch": 124.52, "grad_norm": 1.0294045209884644, "learning_rate": 0.0007548148148148148, "loss": 0.9267, "step": 33620 }, { "epoch": 124.56, "grad_norm": 1.050539255142212, "learning_rate": 0.0007544444444444445, "loss": 0.9569, "step": 33630 }, { "epoch": 124.59, "grad_norm": 1.0169801712036133, "learning_rate": 0.0007540740740740741, "loss": 0.964, "step": 33640 }, { "epoch": 124.63, "grad_norm": 1.0593849420547485, "learning_rate": 0.0007537037037037037, "loss": 0.9614, "step": 33650 }, { "epoch": 124.67, "grad_norm": 1.0408095121383667, "learning_rate": 0.0007533333333333333, "loss": 0.9934, "step": 33660 }, { "epoch": 124.7, "grad_norm": 1.070684790611267, "learning_rate": 0.000752962962962963, "loss": 0.9856, "step": 33670 }, { "epoch": 124.74, "grad_norm": 0.9906821250915527, "learning_rate": 0.0007525925925925926, "loss": 0.9733, "step": 33680 }, { "epoch": 124.78, "grad_norm": 1.0587712526321411, "learning_rate": 0.0007522222222222223, "loss": 0.997, "step": 33690 }, { "epoch": 124.81, "grad_norm": 1.0874089002609253, "learning_rate": 0.0007518518518518519, "loss": 1.008, "step": 33700 }, { "epoch": 124.85, "grad_norm": 1.0352959632873535, "learning_rate": 0.0007514814814814816, "loss": 1.0052, "step": 33710 }, { "epoch": 124.89, "grad_norm": 1.0280036926269531, "learning_rate": 0.000751111111111111, "loss": 1.0301, "step": 33720 }, { "epoch": 124.93, "grad_norm": 1.1919608116149902, "learning_rate": 0.0007507407407407407, "loss": 1.0094, "step": 33730 }, { "epoch": 124.96, "grad_norm": 1.0236761569976807, "learning_rate": 0.0007503703703703703, "loss": 1.0026, "step": 33740 }, { "epoch": 125.0, "grad_norm": 1.8308442831039429, "learning_rate": 0.00075, "loss": 1.0376, "step": 33750 }, { "epoch": 125.04, "grad_norm": 1.008743166923523, "learning_rate": 0.0007496296296296297, "loss": 0.8049, "step": 33760 }, { "epoch": 125.07, "grad_norm": 0.9807414412498474, "learning_rate": 0.0007492592592592593, "loss": 0.8093, "step": 33770 }, { "epoch": 125.11, "grad_norm": 1.0613797903060913, "learning_rate": 0.000748888888888889, "loss": 0.8376, "step": 33780 }, { "epoch": 125.15, "grad_norm": 0.9105537533760071, "learning_rate": 0.0007485185185185185, "loss": 0.8249, "step": 33790 }, { "epoch": 125.19, "grad_norm": 0.9461547136306763, "learning_rate": 0.0007481481481481482, "loss": 0.8464, "step": 33800 }, { "epoch": 125.22, "grad_norm": 0.9608639478683472, "learning_rate": 0.0007477777777777778, "loss": 0.8709, "step": 33810 }, { "epoch": 125.26, "grad_norm": 1.0083810091018677, "learning_rate": 0.0007474074074074075, "loss": 0.8711, "step": 33820 }, { "epoch": 125.3, "grad_norm": 1.0758371353149414, "learning_rate": 0.000747037037037037, "loss": 0.874, "step": 33830 }, { "epoch": 125.33, "grad_norm": 1.0421946048736572, "learning_rate": 0.0007466666666666667, "loss": 0.9086, "step": 33840 }, { "epoch": 125.37, "grad_norm": 1.081777572631836, "learning_rate": 0.0007462962962962963, "loss": 0.9225, "step": 33850 }, { "epoch": 125.41, "grad_norm": 1.0842175483703613, "learning_rate": 0.0007459259259259259, "loss": 0.9046, "step": 33860 }, { "epoch": 125.44, "grad_norm": 0.9907249808311462, "learning_rate": 0.0007455555555555555, "loss": 0.9335, "step": 33870 }, { "epoch": 125.48, "grad_norm": 1.116257667541504, "learning_rate": 0.0007451851851851852, "loss": 0.9632, "step": 33880 }, { "epoch": 125.52, "grad_norm": 1.0658822059631348, "learning_rate": 0.0007448148148148148, "loss": 0.9315, "step": 33890 }, { "epoch": 125.56, "grad_norm": 1.0912448167800903, "learning_rate": 0.0007444444444444445, "loss": 0.9481, "step": 33900 }, { "epoch": 125.59, "grad_norm": 1.0179263353347778, "learning_rate": 0.0007440740740740741, "loss": 0.9667, "step": 33910 }, { "epoch": 125.63, "grad_norm": 0.9753648638725281, "learning_rate": 0.0007437037037037038, "loss": 0.986, "step": 33920 }, { "epoch": 125.67, "grad_norm": 1.0527098178863525, "learning_rate": 0.0007433333333333333, "loss": 0.9526, "step": 33930 }, { "epoch": 125.7, "grad_norm": 1.0270941257476807, "learning_rate": 0.000742962962962963, "loss": 0.9615, "step": 33940 }, { "epoch": 125.74, "grad_norm": 1.0542274713516235, "learning_rate": 0.0007425925925925925, "loss": 0.9825, "step": 33950 }, { "epoch": 125.78, "grad_norm": 1.0598552227020264, "learning_rate": 0.0007422222222222222, "loss": 0.9979, "step": 33960 }, { "epoch": 125.81, "grad_norm": 1.0630998611450195, "learning_rate": 0.0007418518518518519, "loss": 0.9916, "step": 33970 }, { "epoch": 125.85, "grad_norm": 1.0409631729125977, "learning_rate": 0.0007414814814814815, "loss": 0.999, "step": 33980 }, { "epoch": 125.89, "grad_norm": 1.0184717178344727, "learning_rate": 0.0007411111111111112, "loss": 1.0104, "step": 33990 }, { "epoch": 125.93, "grad_norm": 1.0719102621078491, "learning_rate": 0.0007407407407407407, "loss": 0.9791, "step": 34000 }, { "epoch": 125.96, "grad_norm": 1.1049607992172241, "learning_rate": 0.0007403703703703704, "loss": 1.0184, "step": 34010 }, { "epoch": 126.0, "grad_norm": 2.2022151947021484, "learning_rate": 0.00074, "loss": 1.0025, "step": 34020 }, { "epoch": 126.04, "grad_norm": 0.9834014773368835, "learning_rate": 0.0007396296296296297, "loss": 0.8303, "step": 34030 }, { "epoch": 126.07, "grad_norm": 0.9932088851928711, "learning_rate": 0.0007392592592592593, "loss": 0.805, "step": 34040 }, { "epoch": 126.11, "grad_norm": 1.0134624242782593, "learning_rate": 0.000738888888888889, "loss": 0.8157, "step": 34050 }, { "epoch": 126.15, "grad_norm": 0.9468620419502258, "learning_rate": 0.0007385185185185185, "loss": 0.8269, "step": 34060 }, { "epoch": 126.19, "grad_norm": 1.043798565864563, "learning_rate": 0.0007381481481481481, "loss": 0.8467, "step": 34070 }, { "epoch": 126.22, "grad_norm": 0.9726648926734924, "learning_rate": 0.0007377777777777777, "loss": 0.8784, "step": 34080 }, { "epoch": 126.26, "grad_norm": 0.96148282289505, "learning_rate": 0.0007374074074074074, "loss": 0.8735, "step": 34090 }, { "epoch": 126.3, "grad_norm": 1.0156971216201782, "learning_rate": 0.000737037037037037, "loss": 0.8906, "step": 34100 }, { "epoch": 126.33, "grad_norm": 1.028420090675354, "learning_rate": 0.0007366666666666667, "loss": 0.8966, "step": 34110 }, { "epoch": 126.37, "grad_norm": 1.0544856786727905, "learning_rate": 0.0007362962962962963, "loss": 0.8922, "step": 34120 }, { "epoch": 126.41, "grad_norm": 1.127189040184021, "learning_rate": 0.0007359259259259259, "loss": 0.9044, "step": 34130 }, { "epoch": 126.44, "grad_norm": 0.9666948914527893, "learning_rate": 0.0007355555555555555, "loss": 0.9138, "step": 34140 }, { "epoch": 126.48, "grad_norm": 1.0822830200195312, "learning_rate": 0.0007351851851851852, "loss": 0.926, "step": 34150 }, { "epoch": 126.52, "grad_norm": 1.027675747871399, "learning_rate": 0.0007348148148148149, "loss": 0.9362, "step": 34160 }, { "epoch": 126.56, "grad_norm": 1.0344058275222778, "learning_rate": 0.0007344444444444445, "loss": 0.948, "step": 34170 }, { "epoch": 126.59, "grad_norm": 0.994404673576355, "learning_rate": 0.0007340740740740742, "loss": 0.9551, "step": 34180 }, { "epoch": 126.63, "grad_norm": 1.0416563749313354, "learning_rate": 0.0007337037037037038, "loss": 0.9298, "step": 34190 }, { "epoch": 126.67, "grad_norm": 0.9902626872062683, "learning_rate": 0.0007333333333333333, "loss": 0.9507, "step": 34200 }, { "epoch": 126.7, "grad_norm": 1.032486915588379, "learning_rate": 0.0007329629629629629, "loss": 0.9601, "step": 34210 }, { "epoch": 126.74, "grad_norm": 0.974919855594635, "learning_rate": 0.0007325925925925926, "loss": 0.9625, "step": 34220 }, { "epoch": 126.78, "grad_norm": 1.0024913549423218, "learning_rate": 0.0007322222222222222, "loss": 0.9646, "step": 34230 }, { "epoch": 126.81, "grad_norm": 1.0604099035263062, "learning_rate": 0.0007318518518518519, "loss": 0.9898, "step": 34240 }, { "epoch": 126.85, "grad_norm": 1.0008738040924072, "learning_rate": 0.0007314814814814815, "loss": 0.9755, "step": 34250 }, { "epoch": 126.89, "grad_norm": 1.154698133468628, "learning_rate": 0.0007311111111111112, "loss": 0.9874, "step": 34260 }, { "epoch": 126.93, "grad_norm": 0.9528394937515259, "learning_rate": 0.0007307407407407407, "loss": 1.0169, "step": 34270 }, { "epoch": 126.96, "grad_norm": 1.0628700256347656, "learning_rate": 0.0007303703703703704, "loss": 1.0054, "step": 34280 }, { "epoch": 127.0, "grad_norm": 2.0968518257141113, "learning_rate": 0.00073, "loss": 1.0023, "step": 34290 }, { "epoch": 127.04, "grad_norm": 1.0144600868225098, "learning_rate": 0.0007296296296296297, "loss": 0.7879, "step": 34300 }, { "epoch": 127.07, "grad_norm": 1.0181728601455688, "learning_rate": 0.0007292592592592592, "loss": 0.8311, "step": 34310 }, { "epoch": 127.11, "grad_norm": 0.9796633720397949, "learning_rate": 0.0007288888888888889, "loss": 0.8241, "step": 34320 }, { "epoch": 127.15, "grad_norm": 1.0185998678207397, "learning_rate": 0.0007285185185185185, "loss": 0.8255, "step": 34330 }, { "epoch": 127.19, "grad_norm": 1.039688229560852, "learning_rate": 0.0007281481481481481, "loss": 0.87, "step": 34340 }, { "epoch": 127.22, "grad_norm": 0.9951873421669006, "learning_rate": 0.0007277777777777777, "loss": 0.8566, "step": 34350 }, { "epoch": 127.26, "grad_norm": 0.9923152923583984, "learning_rate": 0.0007274074074074074, "loss": 0.894, "step": 34360 }, { "epoch": 127.3, "grad_norm": 1.0109975337982178, "learning_rate": 0.0007270370370370371, "loss": 0.847, "step": 34370 }, { "epoch": 127.33, "grad_norm": 1.042435884475708, "learning_rate": 0.0007266666666666667, "loss": 0.893, "step": 34380 }, { "epoch": 127.37, "grad_norm": 1.0469590425491333, "learning_rate": 0.0007262962962962964, "loss": 0.9133, "step": 34390 }, { "epoch": 127.41, "grad_norm": 0.952255129814148, "learning_rate": 0.000725925925925926, "loss": 0.889, "step": 34400 }, { "epoch": 127.44, "grad_norm": 1.045935869216919, "learning_rate": 0.0007255555555555556, "loss": 0.9128, "step": 34410 }, { "epoch": 127.48, "grad_norm": 1.0044687986373901, "learning_rate": 0.0007251851851851852, "loss": 0.8965, "step": 34420 }, { "epoch": 127.52, "grad_norm": 1.0032331943511963, "learning_rate": 0.0007248148148148149, "loss": 0.9286, "step": 34430 }, { "epoch": 127.56, "grad_norm": 1.060608148574829, "learning_rate": 0.0007244444444444444, "loss": 0.9221, "step": 34440 }, { "epoch": 127.59, "grad_norm": 1.0088783502578735, "learning_rate": 0.0007240740740740741, "loss": 0.9412, "step": 34450 }, { "epoch": 127.63, "grad_norm": 1.036992073059082, "learning_rate": 0.0007237037037037037, "loss": 0.9274, "step": 34460 }, { "epoch": 127.67, "grad_norm": 1.015704870223999, "learning_rate": 0.0007233333333333334, "loss": 0.9389, "step": 34470 }, { "epoch": 127.7, "grad_norm": 1.0427706241607666, "learning_rate": 0.0007229629629629629, "loss": 0.9613, "step": 34480 }, { "epoch": 127.74, "grad_norm": 1.0360188484191895, "learning_rate": 0.0007225925925925926, "loss": 0.9712, "step": 34490 }, { "epoch": 127.78, "grad_norm": 1.0515973567962646, "learning_rate": 0.0007222222222222222, "loss": 0.9528, "step": 34500 }, { "epoch": 127.81, "grad_norm": 1.0636948347091675, "learning_rate": 0.0007218518518518519, "loss": 0.9715, "step": 34510 }, { "epoch": 127.85, "grad_norm": 1.0268125534057617, "learning_rate": 0.0007214814814814815, "loss": 0.9949, "step": 34520 }, { "epoch": 127.89, "grad_norm": 1.0717504024505615, "learning_rate": 0.0007211111111111112, "loss": 0.9876, "step": 34530 }, { "epoch": 127.93, "grad_norm": 1.1457937955856323, "learning_rate": 0.0007207407407407408, "loss": 0.9975, "step": 34540 }, { "epoch": 127.96, "grad_norm": 1.0468859672546387, "learning_rate": 0.0007203703703703703, "loss": 0.999, "step": 34550 }, { "epoch": 128.0, "grad_norm": 1.6671861410140991, "learning_rate": 0.0007199999999999999, "loss": 0.9965, "step": 34560 }, { "epoch": 128.04, "grad_norm": 0.9380455017089844, "learning_rate": 0.0007196296296296296, "loss": 0.7957, "step": 34570 }, { "epoch": 128.07, "grad_norm": 1.0969723463058472, "learning_rate": 0.0007192592592592593, "loss": 0.8217, "step": 34580 }, { "epoch": 128.11, "grad_norm": 1.0519886016845703, "learning_rate": 0.0007188888888888889, "loss": 0.8212, "step": 34590 }, { "epoch": 128.15, "grad_norm": 1.0497735738754272, "learning_rate": 0.0007185185185185186, "loss": 0.79, "step": 34600 }, { "epoch": 128.19, "grad_norm": 0.9826639890670776, "learning_rate": 0.0007181481481481482, "loss": 0.8176, "step": 34610 }, { "epoch": 128.22, "grad_norm": 0.9700071811676025, "learning_rate": 0.0007177777777777778, "loss": 0.8425, "step": 34620 }, { "epoch": 128.26, "grad_norm": 1.027782678604126, "learning_rate": 0.0007174074074074074, "loss": 0.8592, "step": 34630 }, { "epoch": 128.3, "grad_norm": 1.0365031957626343, "learning_rate": 0.0007170370370370371, "loss": 0.8615, "step": 34640 }, { "epoch": 128.33, "grad_norm": 1.0428645610809326, "learning_rate": 0.0007166666666666667, "loss": 0.878, "step": 34650 }, { "epoch": 128.37, "grad_norm": 1.0856304168701172, "learning_rate": 0.0007162962962962964, "loss": 0.8684, "step": 34660 }, { "epoch": 128.41, "grad_norm": 1.0055755376815796, "learning_rate": 0.000715925925925926, "loss": 0.8736, "step": 34670 }, { "epoch": 128.44, "grad_norm": 1.0482579469680786, "learning_rate": 0.0007155555555555555, "loss": 0.8844, "step": 34680 }, { "epoch": 128.48, "grad_norm": 1.0333335399627686, "learning_rate": 0.0007151851851851851, "loss": 0.8991, "step": 34690 }, { "epoch": 128.52, "grad_norm": 1.0611454248428345, "learning_rate": 0.0007148148148148148, "loss": 0.9079, "step": 34700 }, { "epoch": 128.56, "grad_norm": 1.0447732210159302, "learning_rate": 0.0007144444444444444, "loss": 0.948, "step": 34710 }, { "epoch": 128.59, "grad_norm": 0.9698881506919861, "learning_rate": 0.0007140740740740741, "loss": 0.935, "step": 34720 }, { "epoch": 128.63, "grad_norm": 1.0098685026168823, "learning_rate": 0.0007137037037037037, "loss": 0.9341, "step": 34730 }, { "epoch": 128.67, "grad_norm": 1.0363351106643677, "learning_rate": 0.0007133333333333334, "loss": 0.9442, "step": 34740 }, { "epoch": 128.7, "grad_norm": 1.1238548755645752, "learning_rate": 0.0007129629629629629, "loss": 0.9577, "step": 34750 }, { "epoch": 128.74, "grad_norm": 1.1039295196533203, "learning_rate": 0.0007125925925925926, "loss": 0.9572, "step": 34760 }, { "epoch": 128.78, "grad_norm": 1.0145635604858398, "learning_rate": 0.0007122222222222222, "loss": 0.9531, "step": 34770 }, { "epoch": 128.81, "grad_norm": 1.1182947158813477, "learning_rate": 0.0007118518518518519, "loss": 0.9555, "step": 34780 }, { "epoch": 128.85, "grad_norm": 1.1297719478607178, "learning_rate": 0.0007114814814814816, "loss": 0.9767, "step": 34790 }, { "epoch": 128.89, "grad_norm": 1.0659856796264648, "learning_rate": 0.0007111111111111111, "loss": 0.9844, "step": 34800 }, { "epoch": 128.93, "grad_norm": 1.1524667739868164, "learning_rate": 0.0007107407407407408, "loss": 1.002, "step": 34810 }, { "epoch": 128.96, "grad_norm": 1.0247752666473389, "learning_rate": 0.0007103703703703703, "loss": 0.9854, "step": 34820 }, { "epoch": 129.0, "grad_norm": 1.7865345478057861, "learning_rate": 0.00071, "loss": 0.994, "step": 34830 }, { "epoch": 129.04, "grad_norm": 0.9138088822364807, "learning_rate": 0.0007096296296296296, "loss": 0.7795, "step": 34840 }, { "epoch": 129.07, "grad_norm": 1.0572221279144287, "learning_rate": 0.0007092592592592593, "loss": 0.7784, "step": 34850 }, { "epoch": 129.11, "grad_norm": 0.9977957010269165, "learning_rate": 0.0007088888888888889, "loss": 0.8109, "step": 34860 }, { "epoch": 129.15, "grad_norm": 0.9493637681007385, "learning_rate": 0.0007085185185185186, "loss": 0.8413, "step": 34870 }, { "epoch": 129.19, "grad_norm": 0.9977490305900574, "learning_rate": 0.0007081481481481482, "loss": 0.8216, "step": 34880 }, { "epoch": 129.22, "grad_norm": 1.0183244943618774, "learning_rate": 0.0007077777777777778, "loss": 0.8501, "step": 34890 }, { "epoch": 129.26, "grad_norm": 0.9969905018806458, "learning_rate": 0.0007074074074074074, "loss": 0.8457, "step": 34900 }, { "epoch": 129.3, "grad_norm": 0.9821087121963501, "learning_rate": 0.000707037037037037, "loss": 0.843, "step": 34910 }, { "epoch": 129.33, "grad_norm": 1.037099838256836, "learning_rate": 0.0007066666666666666, "loss": 0.874, "step": 34920 }, { "epoch": 129.37, "grad_norm": 1.114404559135437, "learning_rate": 0.0007062962962962963, "loss": 0.8413, "step": 34930 }, { "epoch": 129.41, "grad_norm": 1.0478968620300293, "learning_rate": 0.0007059259259259259, "loss": 0.8771, "step": 34940 }, { "epoch": 129.44, "grad_norm": 1.1025892496109009, "learning_rate": 0.0007055555555555556, "loss": 0.8864, "step": 34950 }, { "epoch": 129.48, "grad_norm": 1.0482041835784912, "learning_rate": 0.0007051851851851851, "loss": 0.9107, "step": 34960 }, { "epoch": 129.52, "grad_norm": 1.0418429374694824, "learning_rate": 0.0007048148148148148, "loss": 0.9103, "step": 34970 }, { "epoch": 129.56, "grad_norm": 1.1483935117721558, "learning_rate": 0.0007044444444444445, "loss": 0.9035, "step": 34980 }, { "epoch": 129.59, "grad_norm": 1.089949131011963, "learning_rate": 0.0007040740740740741, "loss": 0.9068, "step": 34990 }, { "epoch": 129.63, "grad_norm": 1.081794261932373, "learning_rate": 0.0007037037037037038, "loss": 0.9361, "step": 35000 }, { "epoch": 129.67, "grad_norm": 1.0955454111099243, "learning_rate": 0.0007033333333333334, "loss": 0.933, "step": 35010 }, { "epoch": 129.7, "grad_norm": 1.129664659500122, "learning_rate": 0.0007029629629629631, "loss": 0.9376, "step": 35020 }, { "epoch": 129.74, "grad_norm": 1.086525559425354, "learning_rate": 0.0007025925925925925, "loss": 0.9453, "step": 35030 }, { "epoch": 129.78, "grad_norm": 0.9380663633346558, "learning_rate": 0.0007022222222222222, "loss": 0.9583, "step": 35040 }, { "epoch": 129.81, "grad_norm": 1.0038551092147827, "learning_rate": 0.0007018518518518518, "loss": 0.9623, "step": 35050 }, { "epoch": 129.85, "grad_norm": 1.0711705684661865, "learning_rate": 0.0007014814814814815, "loss": 0.9572, "step": 35060 }, { "epoch": 129.89, "grad_norm": 1.134537696838379, "learning_rate": 0.0007011111111111111, "loss": 0.9686, "step": 35070 }, { "epoch": 129.93, "grad_norm": 1.0236729383468628, "learning_rate": 0.0007007407407407408, "loss": 0.9866, "step": 35080 }, { "epoch": 129.96, "grad_norm": 1.0966039896011353, "learning_rate": 0.0007003703703703704, "loss": 0.9742, "step": 35090 }, { "epoch": 130.0, "grad_norm": 1.7335938215255737, "learning_rate": 0.0007, "loss": 0.9832, "step": 35100 }, { "epoch": 130.04, "grad_norm": 0.9678661823272705, "learning_rate": 0.0006996296296296296, "loss": 0.7645, "step": 35110 }, { "epoch": 130.07, "grad_norm": 0.9778347015380859, "learning_rate": 0.0006992592592592593, "loss": 0.7927, "step": 35120 }, { "epoch": 130.11, "grad_norm": 1.0727163553237915, "learning_rate": 0.0006988888888888889, "loss": 0.8129, "step": 35130 }, { "epoch": 130.15, "grad_norm": 1.0289082527160645, "learning_rate": 0.0006985185185185186, "loss": 0.8036, "step": 35140 }, { "epoch": 130.19, "grad_norm": 1.00894033908844, "learning_rate": 0.0006981481481481482, "loss": 0.8088, "step": 35150 }, { "epoch": 130.22, "grad_norm": 1.071042776107788, "learning_rate": 0.0006977777777777778, "loss": 0.8325, "step": 35160 }, { "epoch": 130.26, "grad_norm": 1.1067564487457275, "learning_rate": 0.0006974074074074073, "loss": 0.8388, "step": 35170 }, { "epoch": 130.3, "grad_norm": 1.0250170230865479, "learning_rate": 0.000697037037037037, "loss": 0.825, "step": 35180 }, { "epoch": 130.33, "grad_norm": 0.9799749255180359, "learning_rate": 0.0006966666666666667, "loss": 0.8384, "step": 35190 }, { "epoch": 130.37, "grad_norm": 1.0186576843261719, "learning_rate": 0.0006962962962962963, "loss": 0.8747, "step": 35200 }, { "epoch": 130.41, "grad_norm": 1.0256609916687012, "learning_rate": 0.000695925925925926, "loss": 0.8986, "step": 35210 }, { "epoch": 130.44, "grad_norm": 1.1367439031600952, "learning_rate": 0.0006955555555555556, "loss": 0.8918, "step": 35220 }, { "epoch": 130.48, "grad_norm": 1.050878643989563, "learning_rate": 0.0006951851851851853, "loss": 0.8694, "step": 35230 }, { "epoch": 130.52, "grad_norm": 1.0891170501708984, "learning_rate": 0.0006948148148148148, "loss": 0.8837, "step": 35240 }, { "epoch": 130.56, "grad_norm": 1.0573313236236572, "learning_rate": 0.0006944444444444445, "loss": 0.9004, "step": 35250 }, { "epoch": 130.59, "grad_norm": 1.0353542566299438, "learning_rate": 0.0006940740740740741, "loss": 0.922, "step": 35260 }, { "epoch": 130.63, "grad_norm": 1.1184134483337402, "learning_rate": 0.0006937037037037038, "loss": 0.924, "step": 35270 }, { "epoch": 130.67, "grad_norm": 1.0716382265090942, "learning_rate": 0.0006933333333333333, "loss": 0.9273, "step": 35280 }, { "epoch": 130.7, "grad_norm": 1.0608584880828857, "learning_rate": 0.000692962962962963, "loss": 0.9439, "step": 35290 }, { "epoch": 130.74, "grad_norm": 1.103045105934143, "learning_rate": 0.0006925925925925925, "loss": 0.9234, "step": 35300 }, { "epoch": 130.78, "grad_norm": 0.9750839471817017, "learning_rate": 0.0006922222222222222, "loss": 0.926, "step": 35310 }, { "epoch": 130.81, "grad_norm": 1.008241057395935, "learning_rate": 0.0006918518518518518, "loss": 0.9397, "step": 35320 }, { "epoch": 130.85, "grad_norm": 1.0332655906677246, "learning_rate": 0.0006914814814814815, "loss": 0.9596, "step": 35330 }, { "epoch": 130.89, "grad_norm": 1.1135740280151367, "learning_rate": 0.0006911111111111111, "loss": 0.9674, "step": 35340 }, { "epoch": 130.93, "grad_norm": 1.0512356758117676, "learning_rate": 0.0006907407407407408, "loss": 0.963, "step": 35350 }, { "epoch": 130.96, "grad_norm": 1.0693310499191284, "learning_rate": 0.0006903703703703704, "loss": 0.9622, "step": 35360 }, { "epoch": 131.0, "grad_norm": 2.0000877380371094, "learning_rate": 0.00069, "loss": 0.9982, "step": 35370 }, { "epoch": 131.04, "grad_norm": 1.1226000785827637, "learning_rate": 0.0006896296296296296, "loss": 0.7572, "step": 35380 }, { "epoch": 131.07, "grad_norm": 1.0355254411697388, "learning_rate": 0.0006892592592592593, "loss": 0.7667, "step": 35390 }, { "epoch": 131.11, "grad_norm": 0.9900449514389038, "learning_rate": 0.000688888888888889, "loss": 0.7738, "step": 35400 }, { "epoch": 131.15, "grad_norm": 0.9880581498146057, "learning_rate": 0.0006885185185185185, "loss": 0.7815, "step": 35410 }, { "epoch": 131.19, "grad_norm": 1.0864583253860474, "learning_rate": 0.0006881481481481482, "loss": 0.7919, "step": 35420 }, { "epoch": 131.22, "grad_norm": 0.9945310354232788, "learning_rate": 0.0006877777777777778, "loss": 0.8051, "step": 35430 }, { "epoch": 131.26, "grad_norm": 1.0401781797409058, "learning_rate": 0.0006874074074074074, "loss": 0.8267, "step": 35440 }, { "epoch": 131.3, "grad_norm": 1.097060203552246, "learning_rate": 0.000687037037037037, "loss": 0.8195, "step": 35450 }, { "epoch": 131.33, "grad_norm": 1.0526179075241089, "learning_rate": 0.0006866666666666667, "loss": 0.8823, "step": 35460 }, { "epoch": 131.37, "grad_norm": 1.0602213144302368, "learning_rate": 0.0006862962962962963, "loss": 0.8842, "step": 35470 }, { "epoch": 131.41, "grad_norm": 1.0021635293960571, "learning_rate": 0.000685925925925926, "loss": 0.8676, "step": 35480 }, { "epoch": 131.44, "grad_norm": 1.0285704135894775, "learning_rate": 0.0006855555555555556, "loss": 0.882, "step": 35490 }, { "epoch": 131.48, "grad_norm": 1.020806074142456, "learning_rate": 0.0006851851851851853, "loss": 0.924, "step": 35500 }, { "epoch": 131.52, "grad_norm": 1.071965217590332, "learning_rate": 0.0006848148148148147, "loss": 0.8934, "step": 35510 }, { "epoch": 131.56, "grad_norm": 1.0444443225860596, "learning_rate": 0.0006844444444444444, "loss": 0.9157, "step": 35520 }, { "epoch": 131.59, "grad_norm": 1.0531617403030396, "learning_rate": 0.000684074074074074, "loss": 0.8889, "step": 35530 }, { "epoch": 131.63, "grad_norm": 1.088879942893982, "learning_rate": 0.0006837037037037037, "loss": 0.8971, "step": 35540 }, { "epoch": 131.67, "grad_norm": 1.0203979015350342, "learning_rate": 0.0006833333333333333, "loss": 0.9109, "step": 35550 }, { "epoch": 131.7, "grad_norm": 1.16692054271698, "learning_rate": 0.000682962962962963, "loss": 0.8981, "step": 35560 }, { "epoch": 131.74, "grad_norm": 1.067527174949646, "learning_rate": 0.0006825925925925926, "loss": 0.9372, "step": 35570 }, { "epoch": 131.78, "grad_norm": 1.068550944328308, "learning_rate": 0.0006822222222222222, "loss": 0.9395, "step": 35580 }, { "epoch": 131.81, "grad_norm": 1.1175237894058228, "learning_rate": 0.0006818518518518518, "loss": 0.943, "step": 35590 }, { "epoch": 131.85, "grad_norm": 1.0925233364105225, "learning_rate": 0.0006814814814814815, "loss": 0.9499, "step": 35600 }, { "epoch": 131.89, "grad_norm": 1.0584123134613037, "learning_rate": 0.0006811111111111112, "loss": 0.9339, "step": 35610 }, { "epoch": 131.93, "grad_norm": 1.250260591506958, "learning_rate": 0.0006807407407407408, "loss": 0.9787, "step": 35620 }, { "epoch": 131.96, "grad_norm": 1.0699496269226074, "learning_rate": 0.0006803703703703705, "loss": 0.965, "step": 35630 }, { "epoch": 132.0, "grad_norm": 1.947646975517273, "learning_rate": 0.00068, "loss": 0.9718, "step": 35640 }, { "epoch": 132.04, "grad_norm": 1.0208699703216553, "learning_rate": 0.0006796296296296296, "loss": 0.7436, "step": 35650 }, { "epoch": 132.07, "grad_norm": 1.0202369689941406, "learning_rate": 0.0006792592592592592, "loss": 0.7711, "step": 35660 }, { "epoch": 132.11, "grad_norm": 1.000402569770813, "learning_rate": 0.0006788888888888889, "loss": 0.7827, "step": 35670 }, { "epoch": 132.15, "grad_norm": 0.9595361948013306, "learning_rate": 0.0006785185185185185, "loss": 0.7905, "step": 35680 }, { "epoch": 132.19, "grad_norm": 1.0671266317367554, "learning_rate": 0.0006781481481481482, "loss": 0.8049, "step": 35690 }, { "epoch": 132.22, "grad_norm": 1.0414221286773682, "learning_rate": 0.0006777777777777778, "loss": 0.8074, "step": 35700 }, { "epoch": 132.26, "grad_norm": 1.0275877714157104, "learning_rate": 0.0006774074074074075, "loss": 0.8105, "step": 35710 }, { "epoch": 132.3, "grad_norm": 1.0704302787780762, "learning_rate": 0.000677037037037037, "loss": 0.8278, "step": 35720 }, { "epoch": 132.33, "grad_norm": 1.0970571041107178, "learning_rate": 0.0006766666666666667, "loss": 0.8678, "step": 35730 }, { "epoch": 132.37, "grad_norm": 1.033130168914795, "learning_rate": 0.0006762962962962963, "loss": 0.8492, "step": 35740 }, { "epoch": 132.41, "grad_norm": 1.0589123964309692, "learning_rate": 0.000675925925925926, "loss": 0.8804, "step": 35750 }, { "epoch": 132.44, "grad_norm": 1.0511099100112915, "learning_rate": 0.0006755555555555555, "loss": 0.869, "step": 35760 }, { "epoch": 132.48, "grad_norm": 1.077571153640747, "learning_rate": 0.0006751851851851852, "loss": 0.863, "step": 35770 }, { "epoch": 132.52, "grad_norm": 1.062020182609558, "learning_rate": 0.0006748148148148148, "loss": 0.8803, "step": 35780 }, { "epoch": 132.56, "grad_norm": 1.0527825355529785, "learning_rate": 0.0006744444444444444, "loss": 0.9096, "step": 35790 }, { "epoch": 132.59, "grad_norm": 1.0730772018432617, "learning_rate": 0.0006740740740740741, "loss": 0.886, "step": 35800 }, { "epoch": 132.63, "grad_norm": 1.022678017616272, "learning_rate": 0.0006737037037037037, "loss": 0.9041, "step": 35810 }, { "epoch": 132.67, "grad_norm": 1.1468278169631958, "learning_rate": 0.0006733333333333334, "loss": 0.9103, "step": 35820 }, { "epoch": 132.7, "grad_norm": 1.0276800394058228, "learning_rate": 0.000672962962962963, "loss": 0.899, "step": 35830 }, { "epoch": 132.74, "grad_norm": 1.0555604696273804, "learning_rate": 0.0006725925925925927, "loss": 0.9011, "step": 35840 }, { "epoch": 132.78, "grad_norm": 1.079404354095459, "learning_rate": 0.0006722222222222223, "loss": 0.9099, "step": 35850 }, { "epoch": 132.81, "grad_norm": 1.0517371892929077, "learning_rate": 0.0006718518518518519, "loss": 0.9586, "step": 35860 }, { "epoch": 132.85, "grad_norm": 1.044086217880249, "learning_rate": 0.0006714814814814815, "loss": 0.9198, "step": 35870 }, { "epoch": 132.89, "grad_norm": 1.0771526098251343, "learning_rate": 0.0006711111111111111, "loss": 0.953, "step": 35880 }, { "epoch": 132.93, "grad_norm": 1.1051455736160278, "learning_rate": 0.0006707407407407407, "loss": 0.9572, "step": 35890 }, { "epoch": 132.96, "grad_norm": 1.1288875341415405, "learning_rate": 0.0006703703703703704, "loss": 0.9377, "step": 35900 }, { "epoch": 133.0, "grad_norm": 1.9927622079849243, "learning_rate": 0.00067, "loss": 0.9691, "step": 35910 }, { "epoch": 133.04, "grad_norm": 0.9369173049926758, "learning_rate": 0.0006696296296296296, "loss": 0.7519, "step": 35920 }, { "epoch": 133.07, "grad_norm": 0.9838842153549194, "learning_rate": 0.0006692592592592592, "loss": 0.745, "step": 35930 }, { "epoch": 133.11, "grad_norm": 1.00220787525177, "learning_rate": 0.0006688888888888889, "loss": 0.7842, "step": 35940 }, { "epoch": 133.15, "grad_norm": 1.036676287651062, "learning_rate": 0.0006685185185185185, "loss": 0.7638, "step": 35950 }, { "epoch": 133.19, "grad_norm": 1.0177491903305054, "learning_rate": 0.0006681481481481482, "loss": 0.7857, "step": 35960 }, { "epoch": 133.22, "grad_norm": 1.0435868501663208, "learning_rate": 0.0006677777777777778, "loss": 0.799, "step": 35970 }, { "epoch": 133.26, "grad_norm": 1.0663994550704956, "learning_rate": 0.0006674074074074075, "loss": 0.8105, "step": 35980 }, { "epoch": 133.3, "grad_norm": 1.102872610092163, "learning_rate": 0.000667037037037037, "loss": 0.8117, "step": 35990 }, { "epoch": 133.33, "grad_norm": 1.109534502029419, "learning_rate": 0.0006666666666666666, "loss": 0.8529, "step": 36000 }, { "epoch": 133.37, "grad_norm": 1.137069582939148, "learning_rate": 0.0006662962962962963, "loss": 0.8337, "step": 36010 }, { "epoch": 133.41, "grad_norm": 1.0792121887207031, "learning_rate": 0.0006659259259259259, "loss": 0.8385, "step": 36020 }, { "epoch": 133.44, "grad_norm": 1.0801819562911987, "learning_rate": 0.0006655555555555556, "loss": 0.8567, "step": 36030 }, { "epoch": 133.48, "grad_norm": 1.0681891441345215, "learning_rate": 0.0006651851851851852, "loss": 0.8468, "step": 36040 }, { "epoch": 133.52, "grad_norm": 1.0811008214950562, "learning_rate": 0.0006648148148148149, "loss": 0.872, "step": 36050 }, { "epoch": 133.56, "grad_norm": 1.0719008445739746, "learning_rate": 0.0006644444444444444, "loss": 0.8814, "step": 36060 }, { "epoch": 133.59, "grad_norm": 1.0063518285751343, "learning_rate": 0.0006640740740740741, "loss": 0.8777, "step": 36070 }, { "epoch": 133.63, "grad_norm": 1.0991443395614624, "learning_rate": 0.0006637037037037037, "loss": 0.8806, "step": 36080 }, { "epoch": 133.67, "grad_norm": 1.1498854160308838, "learning_rate": 0.0006633333333333334, "loss": 0.9285, "step": 36090 }, { "epoch": 133.7, "grad_norm": 1.061549425125122, "learning_rate": 0.000662962962962963, "loss": 0.9263, "step": 36100 }, { "epoch": 133.74, "grad_norm": 1.0795164108276367, "learning_rate": 0.0006625925925925927, "loss": 0.9244, "step": 36110 }, { "epoch": 133.78, "grad_norm": 1.093467116355896, "learning_rate": 0.0006622222222222222, "loss": 0.9297, "step": 36120 }, { "epoch": 133.81, "grad_norm": 1.2153167724609375, "learning_rate": 0.0006618518518518518, "loss": 0.9381, "step": 36130 }, { "epoch": 133.85, "grad_norm": 1.0366630554199219, "learning_rate": 0.0006614814814814814, "loss": 0.9309, "step": 36140 }, { "epoch": 133.89, "grad_norm": 1.1849980354309082, "learning_rate": 0.0006611111111111111, "loss": 0.9296, "step": 36150 }, { "epoch": 133.93, "grad_norm": 1.1084529161453247, "learning_rate": 0.0006607407407407407, "loss": 0.9504, "step": 36160 }, { "epoch": 133.96, "grad_norm": 1.0794782638549805, "learning_rate": 0.0006603703703703704, "loss": 0.9344, "step": 36170 }, { "epoch": 134.0, "grad_norm": 1.9394818544387817, "learning_rate": 0.00066, "loss": 0.9645, "step": 36180 }, { "epoch": 134.04, "grad_norm": 1.1422300338745117, "learning_rate": 0.0006596296296296297, "loss": 0.7624, "step": 36190 }, { "epoch": 134.07, "grad_norm": 1.0502318143844604, "learning_rate": 0.0006592592592592592, "loss": 0.745, "step": 36200 }, { "epoch": 134.11, "grad_norm": 0.9741052389144897, "learning_rate": 0.0006588888888888889, "loss": 0.7667, "step": 36210 }, { "epoch": 134.15, "grad_norm": 1.0162742137908936, "learning_rate": 0.0006585185185185186, "loss": 0.7539, "step": 36220 }, { "epoch": 134.19, "grad_norm": 1.1134952306747437, "learning_rate": 0.0006581481481481482, "loss": 0.7852, "step": 36230 }, { "epoch": 134.22, "grad_norm": 1.068518877029419, "learning_rate": 0.0006577777777777779, "loss": 0.8275, "step": 36240 }, { "epoch": 134.26, "grad_norm": 1.0402024984359741, "learning_rate": 0.0006574074074074074, "loss": 0.8097, "step": 36250 }, { "epoch": 134.3, "grad_norm": 1.0597301721572876, "learning_rate": 0.0006570370370370371, "loss": 0.8198, "step": 36260 }, { "epoch": 134.33, "grad_norm": 1.1437287330627441, "learning_rate": 0.0006566666666666666, "loss": 0.8308, "step": 36270 }, { "epoch": 134.37, "grad_norm": 1.0142033100128174, "learning_rate": 0.0006562962962962963, "loss": 0.8471, "step": 36280 }, { "epoch": 134.41, "grad_norm": 1.10916268825531, "learning_rate": 0.0006559259259259259, "loss": 0.8532, "step": 36290 }, { "epoch": 134.44, "grad_norm": 1.0391181707382202, "learning_rate": 0.0006555555555555556, "loss": 0.855, "step": 36300 }, { "epoch": 134.48, "grad_norm": 1.0528442859649658, "learning_rate": 0.0006551851851851852, "loss": 0.8665, "step": 36310 }, { "epoch": 134.52, "grad_norm": 1.035757064819336, "learning_rate": 0.0006548148148148149, "loss": 0.8671, "step": 36320 }, { "epoch": 134.56, "grad_norm": 1.064133644104004, "learning_rate": 0.0006544444444444445, "loss": 0.8735, "step": 36330 }, { "epoch": 134.59, "grad_norm": 1.0806165933609009, "learning_rate": 0.0006540740740740741, "loss": 0.8761, "step": 36340 }, { "epoch": 134.63, "grad_norm": 1.0308071374893188, "learning_rate": 0.0006537037037037037, "loss": 0.8762, "step": 36350 }, { "epoch": 134.67, "grad_norm": 1.1964023113250732, "learning_rate": 0.0006533333333333333, "loss": 0.8909, "step": 36360 }, { "epoch": 134.7, "grad_norm": 1.065822958946228, "learning_rate": 0.0006529629629629629, "loss": 0.9089, "step": 36370 }, { "epoch": 134.74, "grad_norm": 1.0590929985046387, "learning_rate": 0.0006525925925925926, "loss": 0.8952, "step": 36380 }, { "epoch": 134.78, "grad_norm": 1.0043385028839111, "learning_rate": 0.0006522222222222222, "loss": 0.8984, "step": 36390 }, { "epoch": 134.81, "grad_norm": 1.0728081464767456, "learning_rate": 0.0006518518518518519, "loss": 0.9091, "step": 36400 }, { "epoch": 134.85, "grad_norm": 1.0429506301879883, "learning_rate": 0.0006514814814814814, "loss": 0.9294, "step": 36410 }, { "epoch": 134.89, "grad_norm": 1.0224993228912354, "learning_rate": 0.0006511111111111111, "loss": 0.9178, "step": 36420 }, { "epoch": 134.93, "grad_norm": 1.0892304182052612, "learning_rate": 0.0006507407407407408, "loss": 0.9257, "step": 36430 }, { "epoch": 134.96, "grad_norm": 1.1427512168884277, "learning_rate": 0.0006503703703703704, "loss": 0.914, "step": 36440 }, { "epoch": 135.0, "grad_norm": 1.7460381984710693, "learning_rate": 0.0006500000000000001, "loss": 0.921, "step": 36450 }, { "epoch": 135.04, "grad_norm": 1.1062042713165283, "learning_rate": 0.0006496296296296297, "loss": 0.7485, "step": 36460 }, { "epoch": 135.07, "grad_norm": 0.9701430201530457, "learning_rate": 0.0006492592592592594, "loss": 0.7171, "step": 36470 }, { "epoch": 135.11, "grad_norm": 0.9689076542854309, "learning_rate": 0.0006488888888888888, "loss": 0.7617, "step": 36480 }, { "epoch": 135.15, "grad_norm": 1.0703787803649902, "learning_rate": 0.0006485185185185185, "loss": 0.7717, "step": 36490 }, { "epoch": 135.19, "grad_norm": 1.0255299806594849, "learning_rate": 0.0006481481481481481, "loss": 0.7933, "step": 36500 }, { "epoch": 135.22, "grad_norm": 1.018778681755066, "learning_rate": 0.0006477777777777778, "loss": 0.8015, "step": 36510 }, { "epoch": 135.26, "grad_norm": 1.0683023929595947, "learning_rate": 0.0006474074074074074, "loss": 0.8079, "step": 36520 }, { "epoch": 135.3, "grad_norm": 1.0992554426193237, "learning_rate": 0.0006470370370370371, "loss": 0.8185, "step": 36530 }, { "epoch": 135.33, "grad_norm": 1.0614908933639526, "learning_rate": 0.0006466666666666666, "loss": 0.8257, "step": 36540 }, { "epoch": 135.37, "grad_norm": 1.0698963403701782, "learning_rate": 0.0006462962962962963, "loss": 0.825, "step": 36550 }, { "epoch": 135.41, "grad_norm": 1.0651202201843262, "learning_rate": 0.0006459259259259259, "loss": 0.8319, "step": 36560 }, { "epoch": 135.44, "grad_norm": 1.0973514318466187, "learning_rate": 0.0006455555555555556, "loss": 0.8525, "step": 36570 }, { "epoch": 135.48, "grad_norm": 1.115716814994812, "learning_rate": 0.0006451851851851852, "loss": 0.8737, "step": 36580 }, { "epoch": 135.52, "grad_norm": 1.034720540046692, "learning_rate": 0.0006448148148148149, "loss": 0.8434, "step": 36590 }, { "epoch": 135.56, "grad_norm": 1.1401606798171997, "learning_rate": 0.0006444444444444444, "loss": 0.8523, "step": 36600 }, { "epoch": 135.59, "grad_norm": 1.0871201753616333, "learning_rate": 0.000644074074074074, "loss": 0.8677, "step": 36610 }, { "epoch": 135.63, "grad_norm": 1.0409997701644897, "learning_rate": 0.0006437037037037037, "loss": 0.8639, "step": 36620 }, { "epoch": 135.67, "grad_norm": 1.1030999422073364, "learning_rate": 0.0006433333333333333, "loss": 0.8791, "step": 36630 }, { "epoch": 135.7, "grad_norm": 1.0416744947433472, "learning_rate": 0.000642962962962963, "loss": 0.8811, "step": 36640 }, { "epoch": 135.74, "grad_norm": 1.1191257238388062, "learning_rate": 0.0006425925925925926, "loss": 0.8853, "step": 36650 }, { "epoch": 135.78, "grad_norm": 1.0980498790740967, "learning_rate": 0.0006422222222222223, "loss": 0.8938, "step": 36660 }, { "epoch": 135.81, "grad_norm": 1.062381625175476, "learning_rate": 0.0006418518518518519, "loss": 0.8988, "step": 36670 }, { "epoch": 135.85, "grad_norm": 1.1369181871414185, "learning_rate": 0.0006414814814814815, "loss": 0.8878, "step": 36680 }, { "epoch": 135.89, "grad_norm": 1.0846474170684814, "learning_rate": 0.0006411111111111111, "loss": 0.9399, "step": 36690 }, { "epoch": 135.93, "grad_norm": 1.0928053855895996, "learning_rate": 0.0006407407407407408, "loss": 0.9173, "step": 36700 }, { "epoch": 135.96, "grad_norm": 1.0402101278305054, "learning_rate": 0.0006403703703703704, "loss": 0.9408, "step": 36710 }, { "epoch": 136.0, "grad_norm": 2.5675206184387207, "learning_rate": 0.00064, "loss": 0.9232, "step": 36720 }, { "epoch": 136.04, "grad_norm": 0.9730777740478516, "learning_rate": 0.0006396296296296296, "loss": 0.7415, "step": 36730 }, { "epoch": 136.07, "grad_norm": 0.9571381211280823, "learning_rate": 0.0006392592592592593, "loss": 0.749, "step": 36740 }, { "epoch": 136.11, "grad_norm": 1.0041247606277466, "learning_rate": 0.0006388888888888888, "loss": 0.7541, "step": 36750 }, { "epoch": 136.15, "grad_norm": 1.0825740098953247, "learning_rate": 0.0006385185185185185, "loss": 0.7562, "step": 36760 }, { "epoch": 136.19, "grad_norm": 0.9825723767280579, "learning_rate": 0.0006381481481481481, "loss": 0.7826, "step": 36770 }, { "epoch": 136.22, "grad_norm": 1.0305911302566528, "learning_rate": 0.0006377777777777778, "loss": 0.7847, "step": 36780 }, { "epoch": 136.26, "grad_norm": 1.0714701414108276, "learning_rate": 0.0006374074074074074, "loss": 0.7714, "step": 36790 }, { "epoch": 136.3, "grad_norm": 1.0887693166732788, "learning_rate": 0.0006370370370370371, "loss": 0.8174, "step": 36800 }, { "epoch": 136.33, "grad_norm": 1.063496470451355, "learning_rate": 0.0006366666666666667, "loss": 0.8019, "step": 36810 }, { "epoch": 136.37, "grad_norm": 1.1221165657043457, "learning_rate": 0.0006362962962962963, "loss": 0.8354, "step": 36820 }, { "epoch": 136.41, "grad_norm": 1.0420963764190674, "learning_rate": 0.000635925925925926, "loss": 0.838, "step": 36830 }, { "epoch": 136.44, "grad_norm": 0.9873172640800476, "learning_rate": 0.0006355555555555555, "loss": 0.8367, "step": 36840 }, { "epoch": 136.48, "grad_norm": 1.1159552335739136, "learning_rate": 0.0006351851851851852, "loss": 0.8414, "step": 36850 }, { "epoch": 136.52, "grad_norm": 1.1081640720367432, "learning_rate": 0.0006348148148148148, "loss": 0.8466, "step": 36860 }, { "epoch": 136.56, "grad_norm": 1.1748673915863037, "learning_rate": 0.0006344444444444445, "loss": 0.8403, "step": 36870 }, { "epoch": 136.59, "grad_norm": 1.0808719396591187, "learning_rate": 0.0006340740740740741, "loss": 0.8689, "step": 36880 }, { "epoch": 136.63, "grad_norm": 1.0514870882034302, "learning_rate": 0.0006337037037037037, "loss": 0.8875, "step": 36890 }, { "epoch": 136.67, "grad_norm": 1.1719036102294922, "learning_rate": 0.0006333333333333333, "loss": 0.8559, "step": 36900 }, { "epoch": 136.7, "grad_norm": 1.1026197671890259, "learning_rate": 0.000632962962962963, "loss": 0.8947, "step": 36910 }, { "epoch": 136.74, "grad_norm": 1.146915316581726, "learning_rate": 0.0006325925925925926, "loss": 0.8931, "step": 36920 }, { "epoch": 136.78, "grad_norm": 1.1203457117080688, "learning_rate": 0.0006322222222222223, "loss": 0.9019, "step": 36930 }, { "epoch": 136.81, "grad_norm": 1.1791220903396606, "learning_rate": 0.0006318518518518519, "loss": 0.9112, "step": 36940 }, { "epoch": 136.85, "grad_norm": 1.0644776821136475, "learning_rate": 0.0006314814814814816, "loss": 0.8946, "step": 36950 }, { "epoch": 136.89, "grad_norm": 1.0956945419311523, "learning_rate": 0.000631111111111111, "loss": 0.9108, "step": 36960 }, { "epoch": 136.93, "grad_norm": 1.064579725265503, "learning_rate": 0.0006307407407407407, "loss": 0.9043, "step": 36970 }, { "epoch": 136.96, "grad_norm": 1.117305874824524, "learning_rate": 0.0006303703703703703, "loss": 0.8842, "step": 36980 }, { "epoch": 137.0, "grad_norm": 2.0160574913024902, "learning_rate": 0.00063, "loss": 0.908, "step": 36990 }, { "epoch": 137.04, "grad_norm": 1.0514990091323853, "learning_rate": 0.0006296296296296296, "loss": 0.7255, "step": 37000 }, { "epoch": 137.07, "grad_norm": 1.0121877193450928, "learning_rate": 0.0006292592592592593, "loss": 0.7161, "step": 37010 }, { "epoch": 137.11, "grad_norm": 1.0525656938552856, "learning_rate": 0.000628888888888889, "loss": 0.7277, "step": 37020 }, { "epoch": 137.15, "grad_norm": 1.0623517036437988, "learning_rate": 0.0006285185185185185, "loss": 0.7355, "step": 37030 }, { "epoch": 137.19, "grad_norm": 1.064374327659607, "learning_rate": 0.0006281481481481482, "loss": 0.7413, "step": 37040 }, { "epoch": 137.22, "grad_norm": 1.085761547088623, "learning_rate": 0.0006277777777777778, "loss": 0.7539, "step": 37050 }, { "epoch": 137.26, "grad_norm": 1.1157646179199219, "learning_rate": 0.0006274074074074075, "loss": 0.7912, "step": 37060 }, { "epoch": 137.3, "grad_norm": 1.0635371208190918, "learning_rate": 0.0006270370370370371, "loss": 0.8024, "step": 37070 }, { "epoch": 137.33, "grad_norm": 1.0670291185379028, "learning_rate": 0.0006266666666666668, "loss": 0.7945, "step": 37080 }, { "epoch": 137.37, "grad_norm": 1.1312631368637085, "learning_rate": 0.0006262962962962963, "loss": 0.8225, "step": 37090 }, { "epoch": 137.41, "grad_norm": 1.0298441648483276, "learning_rate": 0.0006259259259259259, "loss": 0.8254, "step": 37100 }, { "epoch": 137.44, "grad_norm": 1.0292751789093018, "learning_rate": 0.0006255555555555555, "loss": 0.8184, "step": 37110 }, { "epoch": 137.48, "grad_norm": 1.1706461906433105, "learning_rate": 0.0006251851851851852, "loss": 0.8487, "step": 37120 }, { "epoch": 137.52, "grad_norm": 1.0919010639190674, "learning_rate": 0.0006248148148148148, "loss": 0.8532, "step": 37130 }, { "epoch": 137.56, "grad_norm": 1.1142833232879639, "learning_rate": 0.0006244444444444445, "loss": 0.8286, "step": 37140 }, { "epoch": 137.59, "grad_norm": 1.0280237197875977, "learning_rate": 0.0006240740740740741, "loss": 0.8573, "step": 37150 }, { "epoch": 137.63, "grad_norm": 1.1540181636810303, "learning_rate": 0.0006237037037037037, "loss": 0.8634, "step": 37160 }, { "epoch": 137.67, "grad_norm": 1.0904784202575684, "learning_rate": 0.0006233333333333333, "loss": 0.8658, "step": 37170 }, { "epoch": 137.7, "grad_norm": 1.0962737798690796, "learning_rate": 0.000622962962962963, "loss": 0.872, "step": 37180 }, { "epoch": 137.74, "grad_norm": 1.0626097917556763, "learning_rate": 0.0006225925925925926, "loss": 0.8977, "step": 37190 }, { "epoch": 137.78, "grad_norm": 1.079049825668335, "learning_rate": 0.0006222222222222223, "loss": 0.8806, "step": 37200 }, { "epoch": 137.81, "grad_norm": 1.0811468362808228, "learning_rate": 0.0006218518518518518, "loss": 0.8627, "step": 37210 }, { "epoch": 137.85, "grad_norm": 1.1173999309539795, "learning_rate": 0.0006214814814814815, "loss": 0.909, "step": 37220 }, { "epoch": 137.89, "grad_norm": 1.1013028621673584, "learning_rate": 0.000621111111111111, "loss": 0.9217, "step": 37230 }, { "epoch": 137.93, "grad_norm": 1.2114920616149902, "learning_rate": 0.0006207407407407407, "loss": 0.9109, "step": 37240 }, { "epoch": 137.96, "grad_norm": 1.1027235984802246, "learning_rate": 0.0006203703703703704, "loss": 0.9372, "step": 37250 }, { "epoch": 138.0, "grad_norm": 1.8099175691604614, "learning_rate": 0.00062, "loss": 0.9212, "step": 37260 }, { "epoch": 138.04, "grad_norm": 1.0433896780014038, "learning_rate": 0.0006196296296296297, "loss": 0.7166, "step": 37270 }, { "epoch": 138.07, "grad_norm": 0.9847230315208435, "learning_rate": 0.0006192592592592593, "loss": 0.7069, "step": 37280 }, { "epoch": 138.11, "grad_norm": 1.0111587047576904, "learning_rate": 0.000618888888888889, "loss": 0.7178, "step": 37290 }, { "epoch": 138.15, "grad_norm": 1.0253404378890991, "learning_rate": 0.0006185185185185185, "loss": 0.7347, "step": 37300 }, { "epoch": 138.19, "grad_norm": 1.0111534595489502, "learning_rate": 0.0006181481481481482, "loss": 0.7663, "step": 37310 }, { "epoch": 138.22, "grad_norm": 1.0437885522842407, "learning_rate": 0.0006177777777777777, "loss": 0.7502, "step": 37320 }, { "epoch": 138.26, "grad_norm": 1.1091742515563965, "learning_rate": 0.0006174074074074074, "loss": 0.7719, "step": 37330 }, { "epoch": 138.3, "grad_norm": 1.1044790744781494, "learning_rate": 0.000617037037037037, "loss": 0.7994, "step": 37340 }, { "epoch": 138.33, "grad_norm": 1.1266263723373413, "learning_rate": 0.0006166666666666667, "loss": 0.7864, "step": 37350 }, { "epoch": 138.37, "grad_norm": 1.0152026414871216, "learning_rate": 0.0006162962962962963, "loss": 0.8299, "step": 37360 }, { "epoch": 138.41, "grad_norm": 1.0453267097473145, "learning_rate": 0.0006159259259259259, "loss": 0.8212, "step": 37370 }, { "epoch": 138.44, "grad_norm": 1.0323936939239502, "learning_rate": 0.0006155555555555555, "loss": 0.809, "step": 37380 }, { "epoch": 138.48, "grad_norm": 1.1219218969345093, "learning_rate": 0.0006151851851851852, "loss": 0.8379, "step": 37390 }, { "epoch": 138.52, "grad_norm": 1.0550750494003296, "learning_rate": 0.0006148148148148148, "loss": 0.8564, "step": 37400 }, { "epoch": 138.56, "grad_norm": 1.142170786857605, "learning_rate": 0.0006144444444444445, "loss": 0.842, "step": 37410 }, { "epoch": 138.59, "grad_norm": 1.0784144401550293, "learning_rate": 0.0006140740740740741, "loss": 0.8461, "step": 37420 }, { "epoch": 138.63, "grad_norm": 0.992065966129303, "learning_rate": 0.0006137037037037038, "loss": 0.856, "step": 37430 }, { "epoch": 138.67, "grad_norm": 1.1817371845245361, "learning_rate": 0.0006133333333333334, "loss": 0.8397, "step": 37440 }, { "epoch": 138.7, "grad_norm": 1.1475744247436523, "learning_rate": 0.0006129629629629629, "loss": 0.8616, "step": 37450 }, { "epoch": 138.74, "grad_norm": 1.1824603080749512, "learning_rate": 0.0006125925925925926, "loss": 0.8667, "step": 37460 }, { "epoch": 138.78, "grad_norm": 1.0849061012268066, "learning_rate": 0.0006122222222222222, "loss": 0.8609, "step": 37470 }, { "epoch": 138.81, "grad_norm": 1.088152527809143, "learning_rate": 0.0006118518518518519, "loss": 0.8914, "step": 37480 }, { "epoch": 138.85, "grad_norm": 1.1588715314865112, "learning_rate": 0.0006114814814814815, "loss": 0.878, "step": 37490 }, { "epoch": 138.89, "grad_norm": 1.1323331594467163, "learning_rate": 0.0006111111111111112, "loss": 0.8842, "step": 37500 }, { "epoch": 138.93, "grad_norm": 1.0989532470703125, "learning_rate": 0.0006107407407407407, "loss": 0.9039, "step": 37510 }, { "epoch": 138.96, "grad_norm": 1.1324613094329834, "learning_rate": 0.0006103703703703704, "loss": 0.9128, "step": 37520 }, { "epoch": 139.0, "grad_norm": 2.4864327907562256, "learning_rate": 0.00061, "loss": 0.8831, "step": 37530 }, { "epoch": 139.04, "grad_norm": 0.9940552711486816, "learning_rate": 0.0006096296296296297, "loss": 0.7002, "step": 37540 }, { "epoch": 139.07, "grad_norm": 1.0477380752563477, "learning_rate": 0.0006092592592592593, "loss": 0.7302, "step": 37550 }, { "epoch": 139.11, "grad_norm": 1.0235960483551025, "learning_rate": 0.000608888888888889, "loss": 0.7223, "step": 37560 }, { "epoch": 139.15, "grad_norm": 0.9758222103118896, "learning_rate": 0.0006085185185185185, "loss": 0.7183, "step": 37570 }, { "epoch": 139.19, "grad_norm": 1.018068552017212, "learning_rate": 0.0006081481481481481, "loss": 0.7327, "step": 37580 }, { "epoch": 139.22, "grad_norm": 1.0142362117767334, "learning_rate": 0.0006077777777777777, "loss": 0.7574, "step": 37590 }, { "epoch": 139.26, "grad_norm": 1.1294622421264648, "learning_rate": 0.0006074074074074074, "loss": 0.7817, "step": 37600 }, { "epoch": 139.3, "grad_norm": 1.1239715814590454, "learning_rate": 0.000607037037037037, "loss": 0.7688, "step": 37610 }, { "epoch": 139.33, "grad_norm": 1.0944328308105469, "learning_rate": 0.0006066666666666667, "loss": 0.79, "step": 37620 }, { "epoch": 139.37, "grad_norm": 1.0795550346374512, "learning_rate": 0.0006062962962962963, "loss": 0.7937, "step": 37630 }, { "epoch": 139.41, "grad_norm": 1.0517427921295166, "learning_rate": 0.000605925925925926, "loss": 0.8055, "step": 37640 }, { "epoch": 139.44, "grad_norm": 1.1125982999801636, "learning_rate": 0.0006055555555555556, "loss": 0.8028, "step": 37650 }, { "epoch": 139.48, "grad_norm": 1.0875502824783325, "learning_rate": 0.0006051851851851852, "loss": 0.8102, "step": 37660 }, { "epoch": 139.52, "grad_norm": 1.1398983001708984, "learning_rate": 0.0006048148148148149, "loss": 0.8346, "step": 37670 }, { "epoch": 139.56, "grad_norm": 1.16152024269104, "learning_rate": 0.0006044444444444445, "loss": 0.8348, "step": 37680 }, { "epoch": 139.59, "grad_norm": 1.0522875785827637, "learning_rate": 0.0006040740740740741, "loss": 0.83, "step": 37690 }, { "epoch": 139.63, "grad_norm": 1.1236652135849, "learning_rate": 0.0006037037037037037, "loss": 0.8631, "step": 37700 }, { "epoch": 139.67, "grad_norm": 1.0937477350234985, "learning_rate": 0.0006033333333333334, "loss": 0.843, "step": 37710 }, { "epoch": 139.7, "grad_norm": 1.1400094032287598, "learning_rate": 0.0006029629629629629, "loss": 0.857, "step": 37720 }, { "epoch": 139.74, "grad_norm": 1.053101658821106, "learning_rate": 0.0006025925925925926, "loss": 0.8821, "step": 37730 }, { "epoch": 139.78, "grad_norm": 1.1917682886123657, "learning_rate": 0.0006022222222222222, "loss": 0.8669, "step": 37740 }, { "epoch": 139.81, "grad_norm": 1.0519508123397827, "learning_rate": 0.0006018518518518519, "loss": 0.8726, "step": 37750 }, { "epoch": 139.85, "grad_norm": 1.091275691986084, "learning_rate": 0.0006014814814814815, "loss": 0.9034, "step": 37760 }, { "epoch": 139.89, "grad_norm": 1.154052734375, "learning_rate": 0.0006011111111111112, "loss": 0.874, "step": 37770 }, { "epoch": 139.93, "grad_norm": 1.1093419790267944, "learning_rate": 0.0006007407407407407, "loss": 0.87, "step": 37780 }, { "epoch": 139.96, "grad_norm": 1.100222110748291, "learning_rate": 0.0006003703703703704, "loss": 0.8778, "step": 37790 }, { "epoch": 140.0, "grad_norm": 1.992299199104309, "learning_rate": 0.0006, "loss": 0.8879, "step": 37800 }, { "epoch": 140.04, "grad_norm": 0.997854471206665, "learning_rate": 0.0005996296296296296, "loss": 0.6986, "step": 37810 }, { "epoch": 140.07, "grad_norm": 1.0388472080230713, "learning_rate": 0.0005992592592592592, "loss": 0.7165, "step": 37820 }, { "epoch": 140.11, "grad_norm": 1.0725175142288208, "learning_rate": 0.0005988888888888889, "loss": 0.7285, "step": 37830 }, { "epoch": 140.15, "grad_norm": 1.0202758312225342, "learning_rate": 0.0005985185185185186, "loss": 0.74, "step": 37840 }, { "epoch": 140.19, "grad_norm": 1.115369439125061, "learning_rate": 0.0005981481481481481, "loss": 0.7243, "step": 37850 }, { "epoch": 140.22, "grad_norm": 1.033743143081665, "learning_rate": 0.0005977777777777778, "loss": 0.7606, "step": 37860 }, { "epoch": 140.26, "grad_norm": 1.1330405473709106, "learning_rate": 0.0005974074074074074, "loss": 0.746, "step": 37870 }, { "epoch": 140.3, "grad_norm": 1.0477662086486816, "learning_rate": 0.0005970370370370371, "loss": 0.766, "step": 37880 }, { "epoch": 140.33, "grad_norm": 1.0449988842010498, "learning_rate": 0.0005966666666666667, "loss": 0.7704, "step": 37890 }, { "epoch": 140.37, "grad_norm": 1.0974394083023071, "learning_rate": 0.0005962962962962964, "loss": 0.7675, "step": 37900 }, { "epoch": 140.41, "grad_norm": 1.111845850944519, "learning_rate": 0.000595925925925926, "loss": 0.7752, "step": 37910 }, { "epoch": 140.44, "grad_norm": 1.2133774757385254, "learning_rate": 0.0005955555555555556, "loss": 0.8017, "step": 37920 }, { "epoch": 140.48, "grad_norm": 1.1683170795440674, "learning_rate": 0.0005951851851851851, "loss": 0.8009, "step": 37930 }, { "epoch": 140.52, "grad_norm": 1.181700348854065, "learning_rate": 0.0005948148148148148, "loss": 0.8308, "step": 37940 }, { "epoch": 140.56, "grad_norm": 1.0495575666427612, "learning_rate": 0.0005944444444444444, "loss": 0.8417, "step": 37950 }, { "epoch": 140.59, "grad_norm": 1.143425703048706, "learning_rate": 0.0005940740740740741, "loss": 0.8356, "step": 37960 }, { "epoch": 140.63, "grad_norm": 1.1288803815841675, "learning_rate": 0.0005937037037037037, "loss": 0.8432, "step": 37970 }, { "epoch": 140.67, "grad_norm": 1.1121859550476074, "learning_rate": 0.0005933333333333334, "loss": 0.8469, "step": 37980 }, { "epoch": 140.7, "grad_norm": 1.1236166954040527, "learning_rate": 0.0005929629629629629, "loss": 0.8398, "step": 37990 }, { "epoch": 140.74, "grad_norm": 1.1188994646072388, "learning_rate": 0.0005925925925925926, "loss": 0.8561, "step": 38000 }, { "epoch": 140.78, "grad_norm": 1.1710904836654663, "learning_rate": 0.0005922222222222222, "loss": 0.8594, "step": 38010 }, { "epoch": 140.81, "grad_norm": 1.1164005994796753, "learning_rate": 0.0005918518518518519, "loss": 0.8645, "step": 38020 }, { "epoch": 140.85, "grad_norm": 1.1373645067214966, "learning_rate": 0.0005914814814814815, "loss": 0.8587, "step": 38030 }, { "epoch": 140.89, "grad_norm": 1.1462770700454712, "learning_rate": 0.0005911111111111112, "loss": 0.8689, "step": 38040 }, { "epoch": 140.93, "grad_norm": 1.100512981414795, "learning_rate": 0.0005907407407407409, "loss": 0.8783, "step": 38050 }, { "epoch": 140.96, "grad_norm": 1.2024351358413696, "learning_rate": 0.0005903703703703703, "loss": 0.8815, "step": 38060 }, { "epoch": 141.0, "grad_norm": 1.949777364730835, "learning_rate": 0.00059, "loss": 0.8838, "step": 38070 }, { "epoch": 141.04, "grad_norm": 1.0184309482574463, "learning_rate": 0.0005896296296296296, "loss": 0.7022, "step": 38080 }, { "epoch": 141.07, "grad_norm": 1.0396180152893066, "learning_rate": 0.0005892592592592593, "loss": 0.6793, "step": 38090 }, { "epoch": 141.11, "grad_norm": 1.0512199401855469, "learning_rate": 0.0005888888888888889, "loss": 0.7005, "step": 38100 }, { "epoch": 141.15, "grad_norm": 1.046024203300476, "learning_rate": 0.0005885185185185186, "loss": 0.7032, "step": 38110 }, { "epoch": 141.19, "grad_norm": 1.0370922088623047, "learning_rate": 0.0005881481481481482, "loss": 0.721, "step": 38120 }, { "epoch": 141.22, "grad_norm": 1.1544946432113647, "learning_rate": 0.0005877777777777778, "loss": 0.7438, "step": 38130 }, { "epoch": 141.26, "grad_norm": 1.1410951614379883, "learning_rate": 0.0005874074074074074, "loss": 0.7771, "step": 38140 }, { "epoch": 141.3, "grad_norm": 1.0561929941177368, "learning_rate": 0.0005870370370370371, "loss": 0.7694, "step": 38150 }, { "epoch": 141.33, "grad_norm": 1.096622347831726, "learning_rate": 0.0005866666666666667, "loss": 0.7672, "step": 38160 }, { "epoch": 141.37, "grad_norm": 1.0748016834259033, "learning_rate": 0.0005862962962962963, "loss": 0.7778, "step": 38170 }, { "epoch": 141.41, "grad_norm": 1.121734857559204, "learning_rate": 0.0005859259259259259, "loss": 0.8151, "step": 38180 }, { "epoch": 141.44, "grad_norm": 1.2701319456100464, "learning_rate": 0.0005855555555555556, "loss": 0.7857, "step": 38190 }, { "epoch": 141.48, "grad_norm": 1.0674980878829956, "learning_rate": 0.0005851851851851851, "loss": 0.8126, "step": 38200 }, { "epoch": 141.52, "grad_norm": 1.125149130821228, "learning_rate": 0.0005848148148148148, "loss": 0.8048, "step": 38210 }, { "epoch": 141.56, "grad_norm": 1.0857785940170288, "learning_rate": 0.0005844444444444444, "loss": 0.8335, "step": 38220 }, { "epoch": 141.59, "grad_norm": 1.1086872816085815, "learning_rate": 0.0005840740740740741, "loss": 0.8306, "step": 38230 }, { "epoch": 141.63, "grad_norm": 1.1450704336166382, "learning_rate": 0.0005837037037037037, "loss": 0.8205, "step": 38240 }, { "epoch": 141.67, "grad_norm": 1.1717641353607178, "learning_rate": 0.0005833333333333334, "loss": 0.837, "step": 38250 }, { "epoch": 141.7, "grad_norm": 1.1084407567977905, "learning_rate": 0.0005829629629629631, "loss": 0.816, "step": 38260 }, { "epoch": 141.74, "grad_norm": 1.1263971328735352, "learning_rate": 0.0005825925925925926, "loss": 0.8279, "step": 38270 }, { "epoch": 141.78, "grad_norm": 1.140552043914795, "learning_rate": 0.0005822222222222223, "loss": 0.8537, "step": 38280 }, { "epoch": 141.81, "grad_norm": 1.1099085807800293, "learning_rate": 0.0005818518518518518, "loss": 0.8805, "step": 38290 }, { "epoch": 141.85, "grad_norm": 1.0791407823562622, "learning_rate": 0.0005814814814814815, "loss": 0.8542, "step": 38300 }, { "epoch": 141.89, "grad_norm": 1.1818629503250122, "learning_rate": 0.0005811111111111111, "loss": 0.868, "step": 38310 }, { "epoch": 141.93, "grad_norm": 1.0695195198059082, "learning_rate": 0.0005807407407407408, "loss": 0.8643, "step": 38320 }, { "epoch": 141.96, "grad_norm": 1.1952699422836304, "learning_rate": 0.0005803703703703704, "loss": 0.8908, "step": 38330 }, { "epoch": 142.0, "grad_norm": 2.1429734230041504, "learning_rate": 0.00058, "loss": 0.8539, "step": 38340 }, { "epoch": 142.04, "grad_norm": 0.9811264872550964, "learning_rate": 0.0005796296296296296, "loss": 0.6761, "step": 38350 }, { "epoch": 142.07, "grad_norm": 0.9972795844078064, "learning_rate": 0.0005792592592592593, "loss": 0.707, "step": 38360 }, { "epoch": 142.11, "grad_norm": 1.0141464471817017, "learning_rate": 0.0005788888888888889, "loss": 0.6954, "step": 38370 }, { "epoch": 142.15, "grad_norm": 1.0808486938476562, "learning_rate": 0.0005785185185185186, "loss": 0.7068, "step": 38380 }, { "epoch": 142.19, "grad_norm": 1.119739055633545, "learning_rate": 0.0005781481481481482, "loss": 0.7176, "step": 38390 }, { "epoch": 142.22, "grad_norm": 1.1214107275009155, "learning_rate": 0.0005777777777777778, "loss": 0.7453, "step": 38400 }, { "epoch": 142.26, "grad_norm": 1.093930721282959, "learning_rate": 0.0005774074074074073, "loss": 0.7452, "step": 38410 }, { "epoch": 142.3, "grad_norm": 1.102303147315979, "learning_rate": 0.000577037037037037, "loss": 0.7524, "step": 38420 }, { "epoch": 142.33, "grad_norm": 1.0816642045974731, "learning_rate": 0.0005766666666666666, "loss": 0.7636, "step": 38430 }, { "epoch": 142.37, "grad_norm": 1.0792324542999268, "learning_rate": 0.0005762962962962963, "loss": 0.77, "step": 38440 }, { "epoch": 142.41, "grad_norm": 1.1112143993377686, "learning_rate": 0.0005759259259259259, "loss": 0.7736, "step": 38450 }, { "epoch": 142.44, "grad_norm": 1.1527026891708374, "learning_rate": 0.0005755555555555556, "loss": 0.8019, "step": 38460 }, { "epoch": 142.48, "grad_norm": 1.1157525777816772, "learning_rate": 0.0005751851851851852, "loss": 0.8126, "step": 38470 }, { "epoch": 142.52, "grad_norm": 1.1247179508209229, "learning_rate": 0.0005748148148148148, "loss": 0.7943, "step": 38480 }, { "epoch": 142.56, "grad_norm": 1.0445313453674316, "learning_rate": 0.0005744444444444445, "loss": 0.8086, "step": 38490 }, { "epoch": 142.59, "grad_norm": 1.1115723848342896, "learning_rate": 0.0005740740740740741, "loss": 0.8125, "step": 38500 }, { "epoch": 142.63, "grad_norm": 1.0390018224716187, "learning_rate": 0.0005737037037037038, "loss": 0.8189, "step": 38510 }, { "epoch": 142.67, "grad_norm": 1.129560112953186, "learning_rate": 0.0005733333333333334, "loss": 0.8171, "step": 38520 }, { "epoch": 142.7, "grad_norm": 1.1379806995391846, "learning_rate": 0.000572962962962963, "loss": 0.8264, "step": 38530 }, { "epoch": 142.74, "grad_norm": 1.113497257232666, "learning_rate": 0.0005725925925925925, "loss": 0.8442, "step": 38540 }, { "epoch": 142.78, "grad_norm": 1.118584394454956, "learning_rate": 0.0005722222222222222, "loss": 0.8348, "step": 38550 }, { "epoch": 142.81, "grad_norm": 1.0676319599151611, "learning_rate": 0.0005718518518518518, "loss": 0.8437, "step": 38560 }, { "epoch": 142.85, "grad_norm": 1.1327413320541382, "learning_rate": 0.0005714814814814815, "loss": 0.8419, "step": 38570 }, { "epoch": 142.89, "grad_norm": 1.1048108339309692, "learning_rate": 0.0005711111111111111, "loss": 0.8532, "step": 38580 }, { "epoch": 142.93, "grad_norm": 1.1380068063735962, "learning_rate": 0.0005707407407407408, "loss": 0.8534, "step": 38590 }, { "epoch": 142.96, "grad_norm": 1.1003347635269165, "learning_rate": 0.0005703703703703704, "loss": 0.8627, "step": 38600 }, { "epoch": 143.0, "grad_norm": 2.4374372959136963, "learning_rate": 0.00057, "loss": 0.8586, "step": 38610 }, { "epoch": 143.04, "grad_norm": 0.9777768850326538, "learning_rate": 0.0005696296296296296, "loss": 0.6775, "step": 38620 }, { "epoch": 143.07, "grad_norm": 1.0442510843276978, "learning_rate": 0.0005692592592592593, "loss": 0.6821, "step": 38630 }, { "epoch": 143.11, "grad_norm": 1.043049693107605, "learning_rate": 0.0005688888888888889, "loss": 0.6977, "step": 38640 }, { "epoch": 143.15, "grad_norm": 1.0662933588027954, "learning_rate": 0.0005685185185185185, "loss": 0.7017, "step": 38650 }, { "epoch": 143.19, "grad_norm": 1.0610828399658203, "learning_rate": 0.0005681481481481482, "loss": 0.7077, "step": 38660 }, { "epoch": 143.22, "grad_norm": 1.0228286981582642, "learning_rate": 0.0005677777777777778, "loss": 0.7324, "step": 38670 }, { "epoch": 143.26, "grad_norm": 1.0251611471176147, "learning_rate": 0.0005674074074074074, "loss": 0.7398, "step": 38680 }, { "epoch": 143.3, "grad_norm": 1.1510093212127686, "learning_rate": 0.000567037037037037, "loss": 0.7603, "step": 38690 }, { "epoch": 143.33, "grad_norm": 1.1165571212768555, "learning_rate": 0.0005666666666666667, "loss": 0.7761, "step": 38700 }, { "epoch": 143.37, "grad_norm": 1.1153897047042847, "learning_rate": 0.0005662962962962963, "loss": 0.7582, "step": 38710 }, { "epoch": 143.41, "grad_norm": 1.1018493175506592, "learning_rate": 0.000565925925925926, "loss": 0.7589, "step": 38720 }, { "epoch": 143.44, "grad_norm": 1.0931953191757202, "learning_rate": 0.0005655555555555556, "loss": 0.7809, "step": 38730 }, { "epoch": 143.48, "grad_norm": 1.1556371450424194, "learning_rate": 0.0005651851851851853, "loss": 0.8009, "step": 38740 }, { "epoch": 143.52, "grad_norm": 1.1340473890304565, "learning_rate": 0.0005648148148148148, "loss": 0.769, "step": 38750 }, { "epoch": 143.56, "grad_norm": 1.0956592559814453, "learning_rate": 0.0005644444444444445, "loss": 0.806, "step": 38760 }, { "epoch": 143.59, "grad_norm": 1.2477532625198364, "learning_rate": 0.000564074074074074, "loss": 0.8008, "step": 38770 }, { "epoch": 143.63, "grad_norm": 1.0662354230880737, "learning_rate": 0.0005637037037037037, "loss": 0.8115, "step": 38780 }, { "epoch": 143.67, "grad_norm": 1.212112307548523, "learning_rate": 0.0005633333333333333, "loss": 0.806, "step": 38790 }, { "epoch": 143.7, "grad_norm": 1.0893350839614868, "learning_rate": 0.000562962962962963, "loss": 0.8145, "step": 38800 }, { "epoch": 143.74, "grad_norm": 1.0864957571029663, "learning_rate": 0.0005625925925925926, "loss": 0.8426, "step": 38810 }, { "epoch": 143.78, "grad_norm": 1.1166471242904663, "learning_rate": 0.0005622222222222222, "loss": 0.8316, "step": 38820 }, { "epoch": 143.81, "grad_norm": 1.1510049104690552, "learning_rate": 0.0005618518518518518, "loss": 0.8229, "step": 38830 }, { "epoch": 143.85, "grad_norm": 1.0466991662979126, "learning_rate": 0.0005614814814814815, "loss": 0.8368, "step": 38840 }, { "epoch": 143.89, "grad_norm": 1.1131941080093384, "learning_rate": 0.0005611111111111111, "loss": 0.8588, "step": 38850 }, { "epoch": 143.93, "grad_norm": 1.1063518524169922, "learning_rate": 0.0005607407407407408, "loss": 0.8594, "step": 38860 }, { "epoch": 143.96, "grad_norm": 1.011124849319458, "learning_rate": 0.0005603703703703705, "loss": 0.8556, "step": 38870 }, { "epoch": 144.0, "grad_norm": 1.844652533531189, "learning_rate": 0.0005600000000000001, "loss": 0.8584, "step": 38880 }, { "epoch": 144.04, "grad_norm": 1.0818065404891968, "learning_rate": 0.0005596296296296296, "loss": 0.6472, "step": 38890 }, { "epoch": 144.07, "grad_norm": 1.0280637741088867, "learning_rate": 0.0005592592592592592, "loss": 0.6668, "step": 38900 }, { "epoch": 144.11, "grad_norm": 1.105465292930603, "learning_rate": 0.0005588888888888889, "loss": 0.7023, "step": 38910 }, { "epoch": 144.15, "grad_norm": 1.1759517192840576, "learning_rate": 0.0005585185185185185, "loss": 0.6914, "step": 38920 }, { "epoch": 144.19, "grad_norm": 1.0486412048339844, "learning_rate": 0.0005581481481481482, "loss": 0.7048, "step": 38930 }, { "epoch": 144.22, "grad_norm": 1.0130939483642578, "learning_rate": 0.0005577777777777778, "loss": 0.7313, "step": 38940 }, { "epoch": 144.26, "grad_norm": 1.0542813539505005, "learning_rate": 0.0005574074074074075, "loss": 0.7384, "step": 38950 }, { "epoch": 144.3, "grad_norm": 1.1504030227661133, "learning_rate": 0.000557037037037037, "loss": 0.749, "step": 38960 }, { "epoch": 144.33, "grad_norm": 1.2246326208114624, "learning_rate": 0.0005566666666666667, "loss": 0.753, "step": 38970 }, { "epoch": 144.37, "grad_norm": 1.1878856420516968, "learning_rate": 0.0005562962962962963, "loss": 0.7525, "step": 38980 }, { "epoch": 144.41, "grad_norm": 1.0348081588745117, "learning_rate": 0.000555925925925926, "loss": 0.7492, "step": 38990 }, { "epoch": 144.44, "grad_norm": 1.1095950603485107, "learning_rate": 0.0005555555555555556, "loss": 0.7751, "step": 39000 }, { "epoch": 144.48, "grad_norm": 1.143757700920105, "learning_rate": 0.0005551851851851853, "loss": 0.788, "step": 39010 }, { "epoch": 144.52, "grad_norm": 1.1462836265563965, "learning_rate": 0.0005548148148148147, "loss": 0.7998, "step": 39020 }, { "epoch": 144.56, "grad_norm": 1.0627027750015259, "learning_rate": 0.0005544444444444444, "loss": 0.7932, "step": 39030 }, { "epoch": 144.59, "grad_norm": 1.0926839113235474, "learning_rate": 0.000554074074074074, "loss": 0.8, "step": 39040 }, { "epoch": 144.63, "grad_norm": 1.0934412479400635, "learning_rate": 0.0005537037037037037, "loss": 0.7927, "step": 39050 }, { "epoch": 144.67, "grad_norm": 1.0997884273529053, "learning_rate": 0.0005533333333333333, "loss": 0.8074, "step": 39060 }, { "epoch": 144.7, "grad_norm": 1.0809195041656494, "learning_rate": 0.000552962962962963, "loss": 0.816, "step": 39070 }, { "epoch": 144.74, "grad_norm": 1.1381590366363525, "learning_rate": 0.0005525925925925927, "loss": 0.8291, "step": 39080 }, { "epoch": 144.78, "grad_norm": 1.1699448823928833, "learning_rate": 0.0005522222222222222, "loss": 0.8444, "step": 39090 }, { "epoch": 144.81, "grad_norm": 1.0845420360565186, "learning_rate": 0.0005518518518518519, "loss": 0.8203, "step": 39100 }, { "epoch": 144.85, "grad_norm": 1.1182432174682617, "learning_rate": 0.0005514814814814815, "loss": 0.8269, "step": 39110 }, { "epoch": 144.89, "grad_norm": 1.1073358058929443, "learning_rate": 0.0005511111111111112, "loss": 0.8318, "step": 39120 }, { "epoch": 144.93, "grad_norm": 1.1134344339370728, "learning_rate": 0.0005507407407407407, "loss": 0.8389, "step": 39130 }, { "epoch": 144.96, "grad_norm": 1.1211203336715698, "learning_rate": 0.0005503703703703704, "loss": 0.8578, "step": 39140 }, { "epoch": 145.0, "grad_norm": 1.924920678138733, "learning_rate": 0.00055, "loss": 0.8209, "step": 39150 }, { "epoch": 145.04, "grad_norm": 1.0097651481628418, "learning_rate": 0.0005496296296296296, "loss": 0.6572, "step": 39160 }, { "epoch": 145.07, "grad_norm": 1.04509699344635, "learning_rate": 0.0005492592592592592, "loss": 0.6626, "step": 39170 }, { "epoch": 145.11, "grad_norm": 0.9996280670166016, "learning_rate": 0.0005488888888888889, "loss": 0.7079, "step": 39180 }, { "epoch": 145.15, "grad_norm": 1.038508415222168, "learning_rate": 0.0005485185185185185, "loss": 0.6676, "step": 39190 }, { "epoch": 145.19, "grad_norm": 1.0765340328216553, "learning_rate": 0.0005481481481481482, "loss": 0.6973, "step": 39200 }, { "epoch": 145.22, "grad_norm": 1.0047717094421387, "learning_rate": 0.0005477777777777778, "loss": 0.71, "step": 39210 }, { "epoch": 145.26, "grad_norm": 1.1309590339660645, "learning_rate": 0.0005474074074074075, "loss": 0.7334, "step": 39220 }, { "epoch": 145.3, "grad_norm": 1.156137466430664, "learning_rate": 0.000547037037037037, "loss": 0.7345, "step": 39230 }, { "epoch": 145.33, "grad_norm": 1.1593093872070312, "learning_rate": 0.0005466666666666667, "loss": 0.7398, "step": 39240 }, { "epoch": 145.37, "grad_norm": 1.149204969406128, "learning_rate": 0.0005462962962962962, "loss": 0.7511, "step": 39250 }, { "epoch": 145.41, "grad_norm": 1.0335086584091187, "learning_rate": 0.0005459259259259259, "loss": 0.7702, "step": 39260 }, { "epoch": 145.44, "grad_norm": 1.116210699081421, "learning_rate": 0.0005455555555555555, "loss": 0.7714, "step": 39270 }, { "epoch": 145.48, "grad_norm": 1.105791687965393, "learning_rate": 0.0005451851851851852, "loss": 0.7735, "step": 39280 }, { "epoch": 145.52, "grad_norm": 1.0724138021469116, "learning_rate": 0.0005448148148148149, "loss": 0.7731, "step": 39290 }, { "epoch": 145.56, "grad_norm": 1.1065338850021362, "learning_rate": 0.0005444444444444444, "loss": 0.7743, "step": 39300 }, { "epoch": 145.59, "grad_norm": 1.1840715408325195, "learning_rate": 0.0005440740740740741, "loss": 0.798, "step": 39310 }, { "epoch": 145.63, "grad_norm": 1.1680824756622314, "learning_rate": 0.0005437037037037037, "loss": 0.7973, "step": 39320 }, { "epoch": 145.67, "grad_norm": 1.1493988037109375, "learning_rate": 0.0005433333333333334, "loss": 0.7932, "step": 39330 }, { "epoch": 145.7, "grad_norm": 1.14024817943573, "learning_rate": 0.000542962962962963, "loss": 0.8029, "step": 39340 }, { "epoch": 145.74, "grad_norm": 1.1363987922668457, "learning_rate": 0.0005425925925925927, "loss": 0.803, "step": 39350 }, { "epoch": 145.78, "grad_norm": 1.156929612159729, "learning_rate": 0.0005422222222222223, "loss": 0.8108, "step": 39360 }, { "epoch": 145.81, "grad_norm": 1.1919382810592651, "learning_rate": 0.0005418518518518518, "loss": 0.8119, "step": 39370 }, { "epoch": 145.85, "grad_norm": 1.0990703105926514, "learning_rate": 0.0005414814814814814, "loss": 0.834, "step": 39380 }, { "epoch": 145.89, "grad_norm": 1.1778970956802368, "learning_rate": 0.0005411111111111111, "loss": 0.8265, "step": 39390 }, { "epoch": 145.93, "grad_norm": 1.1993062496185303, "learning_rate": 0.0005407407407407407, "loss": 0.8397, "step": 39400 }, { "epoch": 145.96, "grad_norm": 1.1283531188964844, "learning_rate": 0.0005403703703703704, "loss": 0.8346, "step": 39410 }, { "epoch": 146.0, "grad_norm": 2.261765241622925, "learning_rate": 0.00054, "loss": 0.8082, "step": 39420 }, { "epoch": 146.04, "grad_norm": 1.0939857959747314, "learning_rate": 0.0005396296296296297, "loss": 0.6595, "step": 39430 }, { "epoch": 146.07, "grad_norm": 1.005850911140442, "learning_rate": 0.0005392592592592592, "loss": 0.6712, "step": 39440 }, { "epoch": 146.11, "grad_norm": 1.0855950117111206, "learning_rate": 0.0005388888888888889, "loss": 0.6865, "step": 39450 }, { "epoch": 146.15, "grad_norm": 1.111831784248352, "learning_rate": 0.0005385185185185185, "loss": 0.6897, "step": 39460 }, { "epoch": 146.19, "grad_norm": 1.1208680868148804, "learning_rate": 0.0005381481481481482, "loss": 0.6802, "step": 39470 }, { "epoch": 146.22, "grad_norm": 1.0708884000778198, "learning_rate": 0.0005377777777777779, "loss": 0.701, "step": 39480 }, { "epoch": 146.26, "grad_norm": 1.1542518138885498, "learning_rate": 0.0005374074074074075, "loss": 0.7043, "step": 39490 }, { "epoch": 146.3, "grad_norm": 1.1241813898086548, "learning_rate": 0.0005370370370370371, "loss": 0.7316, "step": 39500 }, { "epoch": 146.33, "grad_norm": 1.0299608707427979, "learning_rate": 0.0005366666666666666, "loss": 0.7348, "step": 39510 }, { "epoch": 146.37, "grad_norm": 1.130296230316162, "learning_rate": 0.0005362962962962963, "loss": 0.7505, "step": 39520 }, { "epoch": 146.41, "grad_norm": 1.0225753784179688, "learning_rate": 0.0005359259259259259, "loss": 0.736, "step": 39530 }, { "epoch": 146.44, "grad_norm": 1.0999373197555542, "learning_rate": 0.0005355555555555556, "loss": 0.7673, "step": 39540 }, { "epoch": 146.48, "grad_norm": 1.0895211696624756, "learning_rate": 0.0005351851851851852, "loss": 0.7544, "step": 39550 }, { "epoch": 146.52, "grad_norm": 1.161813735961914, "learning_rate": 0.0005348148148148149, "loss": 0.7711, "step": 39560 }, { "epoch": 146.56, "grad_norm": 1.2281898260116577, "learning_rate": 0.0005344444444444445, "loss": 0.769, "step": 39570 }, { "epoch": 146.59, "grad_norm": 1.1053509712219238, "learning_rate": 0.0005340740740740741, "loss": 0.7886, "step": 39580 }, { "epoch": 146.63, "grad_norm": 1.1369836330413818, "learning_rate": 0.0005337037037037037, "loss": 0.7959, "step": 39590 }, { "epoch": 146.67, "grad_norm": 1.1838639974594116, "learning_rate": 0.0005333333333333334, "loss": 0.7876, "step": 39600 }, { "epoch": 146.7, "grad_norm": 1.1545466184616089, "learning_rate": 0.000532962962962963, "loss": 0.788, "step": 39610 }, { "epoch": 146.74, "grad_norm": 1.1716816425323486, "learning_rate": 0.0005325925925925926, "loss": 0.798, "step": 39620 }, { "epoch": 146.78, "grad_norm": 1.1613768339157104, "learning_rate": 0.0005322222222222222, "loss": 0.8067, "step": 39630 }, { "epoch": 146.81, "grad_norm": 1.1515871286392212, "learning_rate": 0.0005318518518518518, "loss": 0.8183, "step": 39640 }, { "epoch": 146.85, "grad_norm": 1.1251693964004517, "learning_rate": 0.0005314814814814814, "loss": 0.8063, "step": 39650 }, { "epoch": 146.89, "grad_norm": 1.1438850164413452, "learning_rate": 0.0005311111111111111, "loss": 0.7968, "step": 39660 }, { "epoch": 146.93, "grad_norm": 1.132735252380371, "learning_rate": 0.0005307407407407407, "loss": 0.8234, "step": 39670 }, { "epoch": 146.96, "grad_norm": 1.1190030574798584, "learning_rate": 0.0005303703703703704, "loss": 0.8268, "step": 39680 }, { "epoch": 147.0, "grad_norm": 2.033181667327881, "learning_rate": 0.0005300000000000001, "loss": 0.8498, "step": 39690 }, { "epoch": 147.04, "grad_norm": 1.0466933250427246, "learning_rate": 0.0005296296296296297, "loss": 0.6463, "step": 39700 }, { "epoch": 147.07, "grad_norm": 1.0324013233184814, "learning_rate": 0.0005292592592592593, "loss": 0.6445, "step": 39710 }, { "epoch": 147.11, "grad_norm": 1.004692554473877, "learning_rate": 0.0005288888888888889, "loss": 0.6506, "step": 39720 }, { "epoch": 147.15, "grad_norm": 1.0634324550628662, "learning_rate": 0.0005285185185185186, "loss": 0.6859, "step": 39730 }, { "epoch": 147.19, "grad_norm": 1.0275641679763794, "learning_rate": 0.0005281481481481481, "loss": 0.6817, "step": 39740 }, { "epoch": 147.22, "grad_norm": 1.118804931640625, "learning_rate": 0.0005277777777777778, "loss": 0.7118, "step": 39750 }, { "epoch": 147.26, "grad_norm": 1.160846471786499, "learning_rate": 0.0005274074074074074, "loss": 0.7119, "step": 39760 }, { "epoch": 147.3, "grad_norm": 1.1287107467651367, "learning_rate": 0.0005270370370370371, "loss": 0.7383, "step": 39770 }, { "epoch": 147.33, "grad_norm": 1.1108704805374146, "learning_rate": 0.0005266666666666666, "loss": 0.7156, "step": 39780 }, { "epoch": 147.37, "grad_norm": 1.1652884483337402, "learning_rate": 0.0005262962962962963, "loss": 0.7476, "step": 39790 }, { "epoch": 147.41, "grad_norm": 1.122714638710022, "learning_rate": 0.0005259259259259259, "loss": 0.7423, "step": 39800 }, { "epoch": 147.44, "grad_norm": 1.1680740118026733, "learning_rate": 0.0005255555555555556, "loss": 0.743, "step": 39810 }, { "epoch": 147.48, "grad_norm": 1.1342582702636719, "learning_rate": 0.0005251851851851852, "loss": 0.7522, "step": 39820 }, { "epoch": 147.52, "grad_norm": 1.113469123840332, "learning_rate": 0.0005248148148148149, "loss": 0.7405, "step": 39830 }, { "epoch": 147.56, "grad_norm": 1.109640121459961, "learning_rate": 0.0005244444444444445, "loss": 0.7548, "step": 39840 }, { "epoch": 147.59, "grad_norm": 1.059773564338684, "learning_rate": 0.000524074074074074, "loss": 0.7609, "step": 39850 }, { "epoch": 147.63, "grad_norm": 1.2097136974334717, "learning_rate": 0.0005237037037037036, "loss": 0.778, "step": 39860 }, { "epoch": 147.67, "grad_norm": 1.1901289224624634, "learning_rate": 0.0005233333333333333, "loss": 0.774, "step": 39870 }, { "epoch": 147.7, "grad_norm": 1.1514116525650024, "learning_rate": 0.0005229629629629629, "loss": 0.7879, "step": 39880 }, { "epoch": 147.74, "grad_norm": 1.1901254653930664, "learning_rate": 0.0005225925925925926, "loss": 0.7876, "step": 39890 }, { "epoch": 147.78, "grad_norm": 1.1982858180999756, "learning_rate": 0.0005222222222222223, "loss": 0.7948, "step": 39900 }, { "epoch": 147.81, "grad_norm": 1.1372984647750854, "learning_rate": 0.0005218518518518519, "loss": 0.8026, "step": 39910 }, { "epoch": 147.85, "grad_norm": 1.1477608680725098, "learning_rate": 0.0005214814814814815, "loss": 0.7825, "step": 39920 }, { "epoch": 147.89, "grad_norm": 1.1834083795547485, "learning_rate": 0.0005211111111111111, "loss": 0.8171, "step": 39930 }, { "epoch": 147.93, "grad_norm": 1.2041763067245483, "learning_rate": 0.0005207407407407408, "loss": 0.8157, "step": 39940 }, { "epoch": 147.96, "grad_norm": 1.1421923637390137, "learning_rate": 0.0005203703703703704, "loss": 0.8245, "step": 39950 }, { "epoch": 148.0, "grad_norm": 2.4720282554626465, "learning_rate": 0.0005200000000000001, "loss": 0.8358, "step": 39960 }, { "epoch": 148.04, "grad_norm": 1.0653715133666992, "learning_rate": 0.0005196296296296297, "loss": 0.6536, "step": 39970 }, { "epoch": 148.07, "grad_norm": 1.0128833055496216, "learning_rate": 0.0005192592592592593, "loss": 0.6595, "step": 39980 }, { "epoch": 148.11, "grad_norm": 1.0346759557724, "learning_rate": 0.0005188888888888888, "loss": 0.6267, "step": 39990 }, { "epoch": 148.15, "grad_norm": 0.9999263286590576, "learning_rate": 0.0005185185185185185, "loss": 0.6466, "step": 40000 }, { "epoch": 148.19, "grad_norm": 1.1047943830490112, "learning_rate": 0.0005181481481481481, "loss": 0.6875, "step": 40010 }, { "epoch": 148.22, "grad_norm": 1.1172022819519043, "learning_rate": 0.0005177777777777778, "loss": 0.6935, "step": 40020 }, { "epoch": 148.26, "grad_norm": 1.1120051145553589, "learning_rate": 0.0005174074074074074, "loss": 0.7019, "step": 40030 }, { "epoch": 148.3, "grad_norm": 1.0814529657363892, "learning_rate": 0.0005170370370370371, "loss": 0.7039, "step": 40040 }, { "epoch": 148.33, "grad_norm": 1.0708006620407104, "learning_rate": 0.0005166666666666667, "loss": 0.7087, "step": 40050 }, { "epoch": 148.37, "grad_norm": 1.0833553075790405, "learning_rate": 0.0005162962962962963, "loss": 0.7337, "step": 40060 }, { "epoch": 148.41, "grad_norm": 1.0425792932510376, "learning_rate": 0.0005159259259259259, "loss": 0.7404, "step": 40070 }, { "epoch": 148.44, "grad_norm": 1.1267426013946533, "learning_rate": 0.0005155555555555556, "loss": 0.7303, "step": 40080 }, { "epoch": 148.48, "grad_norm": 1.1181142330169678, "learning_rate": 0.0005151851851851851, "loss": 0.7364, "step": 40090 }, { "epoch": 148.52, "grad_norm": 1.0920939445495605, "learning_rate": 0.0005148148148148148, "loss": 0.7446, "step": 40100 }, { "epoch": 148.56, "grad_norm": 1.160544991493225, "learning_rate": 0.0005144444444444445, "loss": 0.7639, "step": 40110 }, { "epoch": 148.59, "grad_norm": 1.1735516786575317, "learning_rate": 0.0005140740740740741, "loss": 0.761, "step": 40120 }, { "epoch": 148.63, "grad_norm": 1.1860623359680176, "learning_rate": 0.0005137037037037037, "loss": 0.7591, "step": 40130 }, { "epoch": 148.67, "grad_norm": 1.2525490522384644, "learning_rate": 0.0005133333333333333, "loss": 0.7968, "step": 40140 }, { "epoch": 148.7, "grad_norm": 1.1567634344100952, "learning_rate": 0.000512962962962963, "loss": 0.773, "step": 40150 }, { "epoch": 148.74, "grad_norm": 1.1326799392700195, "learning_rate": 0.0005125925925925926, "loss": 0.789, "step": 40160 }, { "epoch": 148.78, "grad_norm": 1.1794488430023193, "learning_rate": 0.0005122222222222223, "loss": 0.7952, "step": 40170 }, { "epoch": 148.81, "grad_norm": 1.138416051864624, "learning_rate": 0.0005118518518518519, "loss": 0.7985, "step": 40180 }, { "epoch": 148.85, "grad_norm": 1.2064673900604248, "learning_rate": 0.0005114814814814816, "loss": 0.8087, "step": 40190 }, { "epoch": 148.89, "grad_norm": 1.1867910623550415, "learning_rate": 0.0005111111111111111, "loss": 0.8216, "step": 40200 }, { "epoch": 148.93, "grad_norm": 1.1607937812805176, "learning_rate": 0.0005107407407407408, "loss": 0.7919, "step": 40210 }, { "epoch": 148.96, "grad_norm": 1.1278940439224243, "learning_rate": 0.0005103703703703703, "loss": 0.7983, "step": 40220 }, { "epoch": 149.0, "grad_norm": 1.926332712173462, "learning_rate": 0.00051, "loss": 0.8079, "step": 40230 }, { "epoch": 149.04, "grad_norm": 1.08697509765625, "learning_rate": 0.0005096296296296296, "loss": 0.6352, "step": 40240 }, { "epoch": 149.07, "grad_norm": 0.9892911911010742, "learning_rate": 0.0005092592592592593, "loss": 0.6308, "step": 40250 }, { "epoch": 149.11, "grad_norm": 1.0473777055740356, "learning_rate": 0.0005088888888888888, "loss": 0.6406, "step": 40260 }, { "epoch": 149.15, "grad_norm": 1.0929820537567139, "learning_rate": 0.0005085185185185185, "loss": 0.6733, "step": 40270 }, { "epoch": 149.19, "grad_norm": 1.1148898601531982, "learning_rate": 0.0005081481481481481, "loss": 0.6696, "step": 40280 }, { "epoch": 149.22, "grad_norm": 1.172523856163025, "learning_rate": 0.0005077777777777778, "loss": 0.6854, "step": 40290 }, { "epoch": 149.26, "grad_norm": 1.094595193862915, "learning_rate": 0.0005074074074074075, "loss": 0.6874, "step": 40300 }, { "epoch": 149.3, "grad_norm": 1.0399515628814697, "learning_rate": 0.0005070370370370371, "loss": 0.7074, "step": 40310 }, { "epoch": 149.33, "grad_norm": 1.0524202585220337, "learning_rate": 0.0005066666666666668, "loss": 0.7228, "step": 40320 }, { "epoch": 149.37, "grad_norm": 1.1507049798965454, "learning_rate": 0.0005062962962962962, "loss": 0.7194, "step": 40330 }, { "epoch": 149.41, "grad_norm": 1.099669098854065, "learning_rate": 0.0005059259259259259, "loss": 0.7212, "step": 40340 }, { "epoch": 149.44, "grad_norm": 1.093471646308899, "learning_rate": 0.0005055555555555555, "loss": 0.7277, "step": 40350 }, { "epoch": 149.48, "grad_norm": 1.1976149082183838, "learning_rate": 0.0005051851851851852, "loss": 0.7222, "step": 40360 }, { "epoch": 149.52, "grad_norm": 1.1135153770446777, "learning_rate": 0.0005048148148148148, "loss": 0.7621, "step": 40370 }, { "epoch": 149.56, "grad_norm": 1.1930509805679321, "learning_rate": 0.0005044444444444445, "loss": 0.7326, "step": 40380 }, { "epoch": 149.59, "grad_norm": 1.2204022407531738, "learning_rate": 0.0005040740740740741, "loss": 0.7684, "step": 40390 }, { "epoch": 149.63, "grad_norm": 1.1036962270736694, "learning_rate": 0.0005037037037037037, "loss": 0.7623, "step": 40400 }, { "epoch": 149.67, "grad_norm": 1.1521191596984863, "learning_rate": 0.0005033333333333333, "loss": 0.7744, "step": 40410 }, { "epoch": 149.7, "grad_norm": 1.172692894935608, "learning_rate": 0.000502962962962963, "loss": 0.7797, "step": 40420 }, { "epoch": 149.74, "grad_norm": 1.145432710647583, "learning_rate": 0.0005025925925925926, "loss": 0.7674, "step": 40430 }, { "epoch": 149.78, "grad_norm": 1.1602561473846436, "learning_rate": 0.0005022222222222223, "loss": 0.7686, "step": 40440 }, { "epoch": 149.81, "grad_norm": 1.1843475103378296, "learning_rate": 0.0005018518518518519, "loss": 0.7736, "step": 40450 }, { "epoch": 149.85, "grad_norm": 1.2497076988220215, "learning_rate": 0.0005014814814814815, "loss": 0.7763, "step": 40460 }, { "epoch": 149.89, "grad_norm": 1.1091344356536865, "learning_rate": 0.000501111111111111, "loss": 0.7915, "step": 40470 }, { "epoch": 149.93, "grad_norm": 1.1577190160751343, "learning_rate": 0.0005007407407407407, "loss": 0.799, "step": 40480 }, { "epoch": 149.96, "grad_norm": 1.1448825597763062, "learning_rate": 0.0005003703703703703, "loss": 0.791, "step": 40490 }, { "epoch": 150.0, "grad_norm": 1.9420194625854492, "learning_rate": 0.0005, "loss": 0.8055, "step": 40500 }, { "epoch": 150.04, "grad_norm": 1.0880106687545776, "learning_rate": 0.0004996296296296297, "loss": 0.6164, "step": 40510 }, { "epoch": 150.07, "grad_norm": 1.135968804359436, "learning_rate": 0.0004992592592592593, "loss": 0.6499, "step": 40520 }, { "epoch": 150.11, "grad_norm": 1.111045479774475, "learning_rate": 0.0004988888888888889, "loss": 0.6442, "step": 40530 }, { "epoch": 150.15, "grad_norm": 0.9961147308349609, "learning_rate": 0.0004985185185185186, "loss": 0.6586, "step": 40540 }, { "epoch": 150.19, "grad_norm": 1.0140842199325562, "learning_rate": 0.0004981481481481482, "loss": 0.6714, "step": 40550 }, { "epoch": 150.22, "grad_norm": 1.0996960401535034, "learning_rate": 0.0004977777777777778, "loss": 0.6872, "step": 40560 }, { "epoch": 150.26, "grad_norm": 1.110884428024292, "learning_rate": 0.0004974074074074075, "loss": 0.6758, "step": 40570 }, { "epoch": 150.3, "grad_norm": 1.0551011562347412, "learning_rate": 0.000497037037037037, "loss": 0.695, "step": 40580 }, { "epoch": 150.33, "grad_norm": 1.2043288946151733, "learning_rate": 0.0004966666666666666, "loss": 0.7031, "step": 40590 }, { "epoch": 150.37, "grad_norm": 1.1636004447937012, "learning_rate": 0.0004962962962962963, "loss": 0.7112, "step": 40600 }, { "epoch": 150.41, "grad_norm": 1.0336576700210571, "learning_rate": 0.0004959259259259259, "loss": 0.709, "step": 40610 }, { "epoch": 150.44, "grad_norm": 1.1624946594238281, "learning_rate": 0.0004955555555555556, "loss": 0.7204, "step": 40620 }, { "epoch": 150.48, "grad_norm": 1.1802222728729248, "learning_rate": 0.0004951851851851852, "loss": 0.7153, "step": 40630 }, { "epoch": 150.52, "grad_norm": 1.1276952028274536, "learning_rate": 0.0004948148148148148, "loss": 0.7163, "step": 40640 }, { "epoch": 150.56, "grad_norm": 1.0561846494674683, "learning_rate": 0.0004944444444444445, "loss": 0.7412, "step": 40650 }, { "epoch": 150.59, "grad_norm": 1.2993978261947632, "learning_rate": 0.0004940740740740741, "loss": 0.7463, "step": 40660 }, { "epoch": 150.63, "grad_norm": 1.2309373617172241, "learning_rate": 0.0004937037037037037, "loss": 0.7464, "step": 40670 }, { "epoch": 150.67, "grad_norm": 1.1278002262115479, "learning_rate": 0.0004933333333333334, "loss": 0.7687, "step": 40680 }, { "epoch": 150.7, "grad_norm": 1.2106173038482666, "learning_rate": 0.000492962962962963, "loss": 0.766, "step": 40690 }, { "epoch": 150.74, "grad_norm": 1.2029941082000732, "learning_rate": 0.0004925925925925925, "loss": 0.7765, "step": 40700 }, { "epoch": 150.78, "grad_norm": 1.258554458618164, "learning_rate": 0.0004922222222222222, "loss": 0.7825, "step": 40710 }, { "epoch": 150.81, "grad_norm": 1.215246319770813, "learning_rate": 0.0004918518518518519, "loss": 0.7552, "step": 40720 }, { "epoch": 150.85, "grad_norm": 1.0888087749481201, "learning_rate": 0.0004914814814814815, "loss": 0.7828, "step": 40730 }, { "epoch": 150.89, "grad_norm": 1.1715198755264282, "learning_rate": 0.0004911111111111111, "loss": 0.7868, "step": 40740 }, { "epoch": 150.93, "grad_norm": 1.1972428560256958, "learning_rate": 0.0004907407407407408, "loss": 0.79, "step": 40750 }, { "epoch": 150.96, "grad_norm": 1.139693260192871, "learning_rate": 0.0004903703703703704, "loss": 0.7698, "step": 40760 }, { "epoch": 151.0, "grad_norm": 1.8386411666870117, "learning_rate": 0.00049, "loss": 0.7949, "step": 40770 }, { "epoch": 151.04, "grad_norm": 0.9836550354957581, "learning_rate": 0.0004896296296296297, "loss": 0.6448, "step": 40780 }, { "epoch": 151.07, "grad_norm": 0.9819071292877197, "learning_rate": 0.0004892592592592593, "loss": 0.6159, "step": 40790 }, { "epoch": 151.11, "grad_norm": 1.1223642826080322, "learning_rate": 0.0004888888888888889, "loss": 0.6392, "step": 40800 }, { "epoch": 151.15, "grad_norm": 1.1107615232467651, "learning_rate": 0.0004885185185185186, "loss": 0.6437, "step": 40810 }, { "epoch": 151.19, "grad_norm": 1.05820894241333, "learning_rate": 0.0004881481481481482, "loss": 0.6592, "step": 40820 }, { "epoch": 151.22, "grad_norm": 1.0786594152450562, "learning_rate": 0.0004877777777777778, "loss": 0.6723, "step": 40830 }, { "epoch": 151.26, "grad_norm": 1.0383567810058594, "learning_rate": 0.00048740740740740743, "loss": 0.6724, "step": 40840 }, { "epoch": 151.3, "grad_norm": 1.1029598712921143, "learning_rate": 0.00048703703703703707, "loss": 0.6822, "step": 40850 }, { "epoch": 151.33, "grad_norm": 1.18717360496521, "learning_rate": 0.0004866666666666667, "loss": 0.6843, "step": 40860 }, { "epoch": 151.37, "grad_norm": 1.2326600551605225, "learning_rate": 0.0004862962962962963, "loss": 0.6921, "step": 40870 }, { "epoch": 151.41, "grad_norm": 1.0844796895980835, "learning_rate": 0.00048592592592592595, "loss": 0.696, "step": 40880 }, { "epoch": 151.44, "grad_norm": 1.0959242582321167, "learning_rate": 0.0004855555555555556, "loss": 0.7053, "step": 40890 }, { "epoch": 151.48, "grad_norm": 1.081992268562317, "learning_rate": 0.0004851851851851852, "loss": 0.7171, "step": 40900 }, { "epoch": 151.52, "grad_norm": 1.120227336883545, "learning_rate": 0.0004848148148148148, "loss": 0.742, "step": 40910 }, { "epoch": 151.56, "grad_norm": 1.112587332725525, "learning_rate": 0.00048444444444444446, "loss": 0.7368, "step": 40920 }, { "epoch": 151.59, "grad_norm": 1.1411511898040771, "learning_rate": 0.0004840740740740741, "loss": 0.7487, "step": 40930 }, { "epoch": 151.63, "grad_norm": 1.1513372659683228, "learning_rate": 0.0004837037037037037, "loss": 0.7416, "step": 40940 }, { "epoch": 151.67, "grad_norm": 1.2439723014831543, "learning_rate": 0.00048333333333333334, "loss": 0.7646, "step": 40950 }, { "epoch": 151.7, "grad_norm": 1.1092541217803955, "learning_rate": 0.000482962962962963, "loss": 0.7572, "step": 40960 }, { "epoch": 151.74, "grad_norm": 1.1118839979171753, "learning_rate": 0.00048259259259259257, "loss": 0.7438, "step": 40970 }, { "epoch": 151.78, "grad_norm": 1.1118911504745483, "learning_rate": 0.0004822222222222222, "loss": 0.754, "step": 40980 }, { "epoch": 151.81, "grad_norm": 1.1592152118682861, "learning_rate": 0.00048185185185185185, "loss": 0.7615, "step": 40990 }, { "epoch": 151.85, "grad_norm": 1.1864370107650757, "learning_rate": 0.00048148148148148144, "loss": 0.7654, "step": 41000 }, { "epoch": 151.89, "grad_norm": 1.1990009546279907, "learning_rate": 0.0004811111111111111, "loss": 0.7725, "step": 41010 }, { "epoch": 151.93, "grad_norm": 1.149665117263794, "learning_rate": 0.0004807407407407408, "loss": 0.7701, "step": 41020 }, { "epoch": 151.96, "grad_norm": 1.1219122409820557, "learning_rate": 0.0004803703703703704, "loss": 0.7903, "step": 41030 }, { "epoch": 152.0, "grad_norm": 2.4777867794036865, "learning_rate": 0.00048, "loss": 0.7812, "step": 41040 }, { "epoch": 152.04, "grad_norm": 1.1176631450653076, "learning_rate": 0.00047962962962962965, "loss": 0.601, "step": 41050 }, { "epoch": 152.07, "grad_norm": 1.086923599243164, "learning_rate": 0.0004792592592592593, "loss": 0.6317, "step": 41060 }, { "epoch": 152.11, "grad_norm": 0.9840608835220337, "learning_rate": 0.0004788888888888889, "loss": 0.6132, "step": 41070 }, { "epoch": 152.15, "grad_norm": 0.9909200668334961, "learning_rate": 0.00047851851851851853, "loss": 0.6519, "step": 41080 }, { "epoch": 152.19, "grad_norm": 1.117368459701538, "learning_rate": 0.00047814814814814817, "loss": 0.6605, "step": 41090 }, { "epoch": 152.22, "grad_norm": 1.1555815935134888, "learning_rate": 0.0004777777777777778, "loss": 0.6538, "step": 41100 }, { "epoch": 152.26, "grad_norm": 1.1172096729278564, "learning_rate": 0.0004774074074074074, "loss": 0.6665, "step": 41110 }, { "epoch": 152.3, "grad_norm": 1.1070817708969116, "learning_rate": 0.00047703703703703705, "loss": 0.6885, "step": 41120 }, { "epoch": 152.33, "grad_norm": 1.1454331874847412, "learning_rate": 0.0004766666666666667, "loss": 0.6653, "step": 41130 }, { "epoch": 152.37, "grad_norm": 1.1130858659744263, "learning_rate": 0.0004762962962962963, "loss": 0.6984, "step": 41140 }, { "epoch": 152.41, "grad_norm": 1.106524109840393, "learning_rate": 0.0004759259259259259, "loss": 0.6933, "step": 41150 }, { "epoch": 152.44, "grad_norm": 1.1382125616073608, "learning_rate": 0.00047555555555555556, "loss": 0.6922, "step": 41160 }, { "epoch": 152.48, "grad_norm": 1.1217646598815918, "learning_rate": 0.0004751851851851852, "loss": 0.7216, "step": 41170 }, { "epoch": 152.52, "grad_norm": 1.1261752843856812, "learning_rate": 0.0004748148148148148, "loss": 0.7053, "step": 41180 }, { "epoch": 152.56, "grad_norm": 1.1546183824539185, "learning_rate": 0.00047444444444444444, "loss": 0.7199, "step": 41190 }, { "epoch": 152.59, "grad_norm": 1.0895978212356567, "learning_rate": 0.0004740740740740741, "loss": 0.7476, "step": 41200 }, { "epoch": 152.63, "grad_norm": 1.170520305633545, "learning_rate": 0.00047370370370370367, "loss": 0.7436, "step": 41210 }, { "epoch": 152.67, "grad_norm": 1.1396688222885132, "learning_rate": 0.00047333333333333336, "loss": 0.738, "step": 41220 }, { "epoch": 152.7, "grad_norm": 1.1664234399795532, "learning_rate": 0.000472962962962963, "loss": 0.7358, "step": 41230 }, { "epoch": 152.74, "grad_norm": 1.2222336530685425, "learning_rate": 0.00047259259259259265, "loss": 0.7317, "step": 41240 }, { "epoch": 152.78, "grad_norm": 1.122406244277954, "learning_rate": 0.00047222222222222224, "loss": 0.7431, "step": 41250 }, { "epoch": 152.81, "grad_norm": 1.2545862197875977, "learning_rate": 0.0004718518518518519, "loss": 0.7522, "step": 41260 }, { "epoch": 152.85, "grad_norm": 1.1529812812805176, "learning_rate": 0.0004714814814814815, "loss": 0.7779, "step": 41270 }, { "epoch": 152.89, "grad_norm": 1.2208104133605957, "learning_rate": 0.0004711111111111111, "loss": 0.7697, "step": 41280 }, { "epoch": 152.93, "grad_norm": 1.1053566932678223, "learning_rate": 0.00047074074074074075, "loss": 0.7773, "step": 41290 }, { "epoch": 152.96, "grad_norm": 1.2086186408996582, "learning_rate": 0.0004703703703703704, "loss": 0.7723, "step": 41300 }, { "epoch": 153.0, "grad_norm": 2.352893114089966, "learning_rate": 0.00047, "loss": 0.7823, "step": 41310 }, { "epoch": 153.04, "grad_norm": 1.0613468885421753, "learning_rate": 0.00046962962962962963, "loss": 0.6175, "step": 41320 }, { "epoch": 153.07, "grad_norm": 1.0901178121566772, "learning_rate": 0.00046925925925925927, "loss": 0.6191, "step": 41330 }, { "epoch": 153.11, "grad_norm": 1.1153274774551392, "learning_rate": 0.0004688888888888889, "loss": 0.6235, "step": 41340 }, { "epoch": 153.15, "grad_norm": 1.109421730041504, "learning_rate": 0.0004685185185185185, "loss": 0.6197, "step": 41350 }, { "epoch": 153.19, "grad_norm": 1.1132631301879883, "learning_rate": 0.00046814814814814815, "loss": 0.6406, "step": 41360 }, { "epoch": 153.22, "grad_norm": 1.1249178647994995, "learning_rate": 0.0004677777777777778, "loss": 0.6494, "step": 41370 }, { "epoch": 153.26, "grad_norm": 1.061094045639038, "learning_rate": 0.0004674074074074074, "loss": 0.6581, "step": 41380 }, { "epoch": 153.3, "grad_norm": 1.105556607246399, "learning_rate": 0.000467037037037037, "loss": 0.6691, "step": 41390 }, { "epoch": 153.33, "grad_norm": 1.1356477737426758, "learning_rate": 0.00046666666666666666, "loss": 0.6856, "step": 41400 }, { "epoch": 153.37, "grad_norm": 1.0905135869979858, "learning_rate": 0.0004662962962962963, "loss": 0.6867, "step": 41410 }, { "epoch": 153.41, "grad_norm": 1.1868871450424194, "learning_rate": 0.0004659259259259259, "loss": 0.689, "step": 41420 }, { "epoch": 153.44, "grad_norm": 1.1457031965255737, "learning_rate": 0.0004655555555555556, "loss": 0.6948, "step": 41430 }, { "epoch": 153.48, "grad_norm": 1.1043384075164795, "learning_rate": 0.00046518518518518523, "loss": 0.699, "step": 41440 }, { "epoch": 153.52, "grad_norm": 1.0756326913833618, "learning_rate": 0.0004648148148148148, "loss": 0.7106, "step": 41450 }, { "epoch": 153.56, "grad_norm": 1.1827819347381592, "learning_rate": 0.00046444444444444446, "loss": 0.7403, "step": 41460 }, { "epoch": 153.59, "grad_norm": 1.1916935443878174, "learning_rate": 0.0004640740740740741, "loss": 0.7038, "step": 41470 }, { "epoch": 153.63, "grad_norm": 1.1236070394515991, "learning_rate": 0.00046370370370370375, "loss": 0.7164, "step": 41480 }, { "epoch": 153.67, "grad_norm": 1.1120916604995728, "learning_rate": 0.00046333333333333334, "loss": 0.7496, "step": 41490 }, { "epoch": 153.7, "grad_norm": 1.1568827629089355, "learning_rate": 0.000462962962962963, "loss": 0.7386, "step": 41500 }, { "epoch": 153.74, "grad_norm": 1.1520304679870605, "learning_rate": 0.0004625925925925926, "loss": 0.7354, "step": 41510 }, { "epoch": 153.78, "grad_norm": 1.194348931312561, "learning_rate": 0.0004622222222222222, "loss": 0.7527, "step": 41520 }, { "epoch": 153.81, "grad_norm": 1.0823569297790527, "learning_rate": 0.00046185185185185185, "loss": 0.7386, "step": 41530 }, { "epoch": 153.85, "grad_norm": 1.2366108894348145, "learning_rate": 0.0004614814814814815, "loss": 0.7405, "step": 41540 }, { "epoch": 153.89, "grad_norm": 1.2345499992370605, "learning_rate": 0.00046111111111111114, "loss": 0.7587, "step": 41550 }, { "epoch": 153.93, "grad_norm": 1.1638497114181519, "learning_rate": 0.00046074074074074073, "loss": 0.7585, "step": 41560 }, { "epoch": 153.96, "grad_norm": 1.1882902383804321, "learning_rate": 0.00046037037037037037, "loss": 0.7315, "step": 41570 }, { "epoch": 154.0, "grad_norm": 2.5309550762176514, "learning_rate": 0.00046, "loss": 0.7686, "step": 41580 }, { "epoch": 154.04, "grad_norm": 1.1524628400802612, "learning_rate": 0.0004596296296296296, "loss": 0.5923, "step": 41590 }, { "epoch": 154.07, "grad_norm": 1.1245719194412231, "learning_rate": 0.00045925925925925925, "loss": 0.5912, "step": 41600 }, { "epoch": 154.11, "grad_norm": 1.08280611038208, "learning_rate": 0.0004588888888888889, "loss": 0.6281, "step": 41610 }, { "epoch": 154.15, "grad_norm": 1.1774368286132812, "learning_rate": 0.0004585185185185185, "loss": 0.6129, "step": 41620 }, { "epoch": 154.19, "grad_norm": 1.1452480554580688, "learning_rate": 0.0004581481481481482, "loss": 0.6267, "step": 41630 }, { "epoch": 154.22, "grad_norm": 1.070086121559143, "learning_rate": 0.0004577777777777778, "loss": 0.6509, "step": 41640 }, { "epoch": 154.26, "grad_norm": 1.218158483505249, "learning_rate": 0.00045740740740740746, "loss": 0.6465, "step": 41650 }, { "epoch": 154.3, "grad_norm": 1.0811407566070557, "learning_rate": 0.00045703703703703705, "loss": 0.6516, "step": 41660 }, { "epoch": 154.33, "grad_norm": 1.122278094291687, "learning_rate": 0.0004566666666666667, "loss": 0.6576, "step": 41670 }, { "epoch": 154.37, "grad_norm": 1.0787327289581299, "learning_rate": 0.00045629629629629633, "loss": 0.6562, "step": 41680 }, { "epoch": 154.41, "grad_norm": 1.099676489830017, "learning_rate": 0.0004559259259259259, "loss": 0.6897, "step": 41690 }, { "epoch": 154.44, "grad_norm": 1.1147680282592773, "learning_rate": 0.00045555555555555556, "loss": 0.6945, "step": 41700 }, { "epoch": 154.48, "grad_norm": 1.1603626012802124, "learning_rate": 0.0004551851851851852, "loss": 0.6991, "step": 41710 }, { "epoch": 154.52, "grad_norm": 1.1135655641555786, "learning_rate": 0.00045481481481481485, "loss": 0.7082, "step": 41720 }, { "epoch": 154.56, "grad_norm": 1.1685587167739868, "learning_rate": 0.00045444444444444444, "loss": 0.6951, "step": 41730 }, { "epoch": 154.59, "grad_norm": 1.0956676006317139, "learning_rate": 0.0004540740740740741, "loss": 0.7139, "step": 41740 }, { "epoch": 154.63, "grad_norm": 1.1997812986373901, "learning_rate": 0.0004537037037037037, "loss": 0.7053, "step": 41750 }, { "epoch": 154.67, "grad_norm": 1.2050188779830933, "learning_rate": 0.0004533333333333333, "loss": 0.7286, "step": 41760 }, { "epoch": 154.7, "grad_norm": 1.1398041248321533, "learning_rate": 0.00045296296296296295, "loss": 0.7291, "step": 41770 }, { "epoch": 154.74, "grad_norm": 1.146764874458313, "learning_rate": 0.0004525925925925926, "loss": 0.7308, "step": 41780 }, { "epoch": 154.78, "grad_norm": 1.1888115406036377, "learning_rate": 0.00045222222222222224, "loss": 0.7576, "step": 41790 }, { "epoch": 154.81, "grad_norm": 1.1720606088638306, "learning_rate": 0.00045185185185185183, "loss": 0.7278, "step": 41800 }, { "epoch": 154.85, "grad_norm": 1.167518138885498, "learning_rate": 0.00045148148148148147, "loss": 0.7516, "step": 41810 }, { "epoch": 154.89, "grad_norm": 1.218790054321289, "learning_rate": 0.0004511111111111111, "loss": 0.7533, "step": 41820 }, { "epoch": 154.93, "grad_norm": 1.1420875787734985, "learning_rate": 0.0004507407407407407, "loss": 0.7621, "step": 41830 }, { "epoch": 154.96, "grad_norm": 1.1655449867248535, "learning_rate": 0.0004503703703703704, "loss": 0.7515, "step": 41840 }, { "epoch": 155.0, "grad_norm": 2.171473503112793, "learning_rate": 0.00045000000000000004, "loss": 0.7627, "step": 41850 }, { "epoch": 155.04, "grad_norm": 1.02388334274292, "learning_rate": 0.0004496296296296297, "loss": 0.5863, "step": 41860 }, { "epoch": 155.07, "grad_norm": 1.0087072849273682, "learning_rate": 0.0004492592592592593, "loss": 0.6048, "step": 41870 }, { "epoch": 155.11, "grad_norm": 1.1261789798736572, "learning_rate": 0.0004488888888888889, "loss": 0.624, "step": 41880 }, { "epoch": 155.15, "grad_norm": 1.0319106578826904, "learning_rate": 0.00044851851851851856, "loss": 0.6099, "step": 41890 }, { "epoch": 155.19, "grad_norm": 1.107433557510376, "learning_rate": 0.00044814814814814815, "loss": 0.6352, "step": 41900 }, { "epoch": 155.22, "grad_norm": 1.175877332687378, "learning_rate": 0.0004477777777777778, "loss": 0.6475, "step": 41910 }, { "epoch": 155.26, "grad_norm": 1.1328219175338745, "learning_rate": 0.00044740740740740743, "loss": 0.6498, "step": 41920 }, { "epoch": 155.3, "grad_norm": 1.1170285940170288, "learning_rate": 0.000447037037037037, "loss": 0.6436, "step": 41930 }, { "epoch": 155.33, "grad_norm": 1.050402045249939, "learning_rate": 0.00044666666666666666, "loss": 0.6543, "step": 41940 }, { "epoch": 155.37, "grad_norm": 1.1217918395996094, "learning_rate": 0.0004462962962962963, "loss": 0.6589, "step": 41950 }, { "epoch": 155.41, "grad_norm": 1.080417513847351, "learning_rate": 0.00044592592592592595, "loss": 0.6785, "step": 41960 }, { "epoch": 155.44, "grad_norm": 1.1411579847335815, "learning_rate": 0.00044555555555555554, "loss": 0.6696, "step": 41970 }, { "epoch": 155.48, "grad_norm": 1.161450982093811, "learning_rate": 0.0004451851851851852, "loss": 0.6909, "step": 41980 }, { "epoch": 155.52, "grad_norm": 1.2231916189193726, "learning_rate": 0.0004448148148148148, "loss": 0.6869, "step": 41990 }, { "epoch": 155.56, "grad_norm": 1.163811445236206, "learning_rate": 0.0004444444444444444, "loss": 0.6922, "step": 42000 }, { "epoch": 155.59, "grad_norm": 1.1644994020462036, "learning_rate": 0.00044407407407407405, "loss": 0.696, "step": 42010 }, { "epoch": 155.63, "grad_norm": 1.1336703300476074, "learning_rate": 0.0004437037037037037, "loss": 0.7217, "step": 42020 }, { "epoch": 155.67, "grad_norm": 1.075753927230835, "learning_rate": 0.00044333333333333334, "loss": 0.7201, "step": 42030 }, { "epoch": 155.7, "grad_norm": 1.1686182022094727, "learning_rate": 0.000442962962962963, "loss": 0.722, "step": 42040 }, { "epoch": 155.74, "grad_norm": 1.1514602899551392, "learning_rate": 0.0004425925925925926, "loss": 0.7139, "step": 42050 }, { "epoch": 155.78, "grad_norm": 1.1290898323059082, "learning_rate": 0.00044222222222222227, "loss": 0.7301, "step": 42060 }, { "epoch": 155.81, "grad_norm": 1.2161083221435547, "learning_rate": 0.00044185185185185186, "loss": 0.7197, "step": 42070 }, { "epoch": 155.85, "grad_norm": 1.1686118841171265, "learning_rate": 0.0004414814814814815, "loss": 0.7286, "step": 42080 }, { "epoch": 155.89, "grad_norm": 1.2055273056030273, "learning_rate": 0.00044111111111111114, "loss": 0.7256, "step": 42090 }, { "epoch": 155.93, "grad_norm": 1.1590404510498047, "learning_rate": 0.0004407407407407408, "loss": 0.7513, "step": 42100 }, { "epoch": 155.96, "grad_norm": 1.1125705242156982, "learning_rate": 0.0004403703703703704, "loss": 0.7336, "step": 42110 }, { "epoch": 156.0, "grad_norm": 2.0711772441864014, "learning_rate": 0.00044, "loss": 0.754, "step": 42120 }, { "epoch": 156.04, "grad_norm": 1.1298962831497192, "learning_rate": 0.00043962962962962966, "loss": 0.6014, "step": 42130 }, { "epoch": 156.07, "grad_norm": 1.0285028219223022, "learning_rate": 0.00043925925925925925, "loss": 0.5644, "step": 42140 }, { "epoch": 156.11, "grad_norm": 1.0838044881820679, "learning_rate": 0.0004388888888888889, "loss": 0.6023, "step": 42150 }, { "epoch": 156.15, "grad_norm": 1.0490028858184814, "learning_rate": 0.00043851851851851853, "loss": 0.6113, "step": 42160 }, { "epoch": 156.19, "grad_norm": 1.0957796573638916, "learning_rate": 0.0004381481481481482, "loss": 0.6303, "step": 42170 }, { "epoch": 156.22, "grad_norm": 1.0779811143875122, "learning_rate": 0.00043777777777777776, "loss": 0.6318, "step": 42180 }, { "epoch": 156.26, "grad_norm": 1.145910382270813, "learning_rate": 0.0004374074074074074, "loss": 0.6353, "step": 42190 }, { "epoch": 156.3, "grad_norm": 1.0913562774658203, "learning_rate": 0.00043703703703703705, "loss": 0.6348, "step": 42200 }, { "epoch": 156.33, "grad_norm": 1.3088796138763428, "learning_rate": 0.00043666666666666664, "loss": 0.648, "step": 42210 }, { "epoch": 156.37, "grad_norm": 1.1091128587722778, "learning_rate": 0.0004362962962962963, "loss": 0.6479, "step": 42220 }, { "epoch": 156.41, "grad_norm": 1.184120535850525, "learning_rate": 0.0004359259259259259, "loss": 0.6536, "step": 42230 }, { "epoch": 156.44, "grad_norm": 1.130748987197876, "learning_rate": 0.0004355555555555555, "loss": 0.6752, "step": 42240 }, { "epoch": 156.48, "grad_norm": 1.153894305229187, "learning_rate": 0.0004351851851851852, "loss": 0.669, "step": 42250 }, { "epoch": 156.52, "grad_norm": 1.2015818357467651, "learning_rate": 0.00043481481481481485, "loss": 0.6832, "step": 42260 }, { "epoch": 156.56, "grad_norm": 1.1696724891662598, "learning_rate": 0.0004344444444444445, "loss": 0.6921, "step": 42270 }, { "epoch": 156.59, "grad_norm": 1.0543904304504395, "learning_rate": 0.0004340740740740741, "loss": 0.6832, "step": 42280 }, { "epoch": 156.63, "grad_norm": 1.197202205657959, "learning_rate": 0.0004337037037037037, "loss": 0.7122, "step": 42290 }, { "epoch": 156.67, "grad_norm": 1.171535849571228, "learning_rate": 0.00043333333333333337, "loss": 0.7122, "step": 42300 }, { "epoch": 156.7, "grad_norm": 1.1360234022140503, "learning_rate": 0.00043296296296296296, "loss": 0.725, "step": 42310 }, { "epoch": 156.74, "grad_norm": 1.1630491018295288, "learning_rate": 0.0004325925925925926, "loss": 0.7112, "step": 42320 }, { "epoch": 156.78, "grad_norm": 1.121999979019165, "learning_rate": 0.00043222222222222224, "loss": 0.7346, "step": 42330 }, { "epoch": 156.81, "grad_norm": 1.2397000789642334, "learning_rate": 0.0004318518518518519, "loss": 0.7055, "step": 42340 }, { "epoch": 156.85, "grad_norm": 1.1417386531829834, "learning_rate": 0.00043148148148148147, "loss": 0.7317, "step": 42350 }, { "epoch": 156.89, "grad_norm": 1.1937369108200073, "learning_rate": 0.0004311111111111111, "loss": 0.7176, "step": 42360 }, { "epoch": 156.93, "grad_norm": 1.2735649347305298, "learning_rate": 0.00043074074074074076, "loss": 0.7317, "step": 42370 }, { "epoch": 156.96, "grad_norm": 1.2109440565109253, "learning_rate": 0.00043037037037037035, "loss": 0.7359, "step": 42380 }, { "epoch": 157.0, "grad_norm": 2.1727218627929688, "learning_rate": 0.00043, "loss": 0.7676, "step": 42390 }, { "epoch": 157.04, "grad_norm": 1.0113236904144287, "learning_rate": 0.00042962962962962963, "loss": 0.5726, "step": 42400 }, { "epoch": 157.07, "grad_norm": 1.0725810527801514, "learning_rate": 0.0004292592592592593, "loss": 0.5732, "step": 42410 }, { "epoch": 157.11, "grad_norm": 1.0465131998062134, "learning_rate": 0.00042888888888888886, "loss": 0.5963, "step": 42420 }, { "epoch": 157.15, "grad_norm": 1.0574426651000977, "learning_rate": 0.0004285185185185185, "loss": 0.604, "step": 42430 }, { "epoch": 157.19, "grad_norm": 1.084412932395935, "learning_rate": 0.00042814814814814815, "loss": 0.6123, "step": 42440 }, { "epoch": 157.22, "grad_norm": 1.1177759170532227, "learning_rate": 0.0004277777777777778, "loss": 0.6294, "step": 42450 }, { "epoch": 157.26, "grad_norm": 1.0315555334091187, "learning_rate": 0.00042740740740740743, "loss": 0.6354, "step": 42460 }, { "epoch": 157.3, "grad_norm": 1.1477491855621338, "learning_rate": 0.0004270370370370371, "loss": 0.6207, "step": 42470 }, { "epoch": 157.33, "grad_norm": 1.078210473060608, "learning_rate": 0.0004266666666666667, "loss": 0.6378, "step": 42480 }, { "epoch": 157.37, "grad_norm": 1.1546977758407593, "learning_rate": 0.0004262962962962963, "loss": 0.6455, "step": 42490 }, { "epoch": 157.41, "grad_norm": 1.1062132120132446, "learning_rate": 0.00042592592592592595, "loss": 0.6488, "step": 42500 }, { "epoch": 157.44, "grad_norm": 1.1070458889007568, "learning_rate": 0.0004255555555555556, "loss": 0.6639, "step": 42510 }, { "epoch": 157.48, "grad_norm": 1.1309759616851807, "learning_rate": 0.0004251851851851852, "loss": 0.6768, "step": 42520 }, { "epoch": 157.52, "grad_norm": 1.1529152393341064, "learning_rate": 0.0004248148148148148, "loss": 0.6733, "step": 42530 }, { "epoch": 157.56, "grad_norm": 1.1645874977111816, "learning_rate": 0.00042444444444444447, "loss": 0.6805, "step": 42540 }, { "epoch": 157.59, "grad_norm": 1.1483224630355835, "learning_rate": 0.00042407407407407406, "loss": 0.6802, "step": 42550 }, { "epoch": 157.63, "grad_norm": 1.1695054769515991, "learning_rate": 0.0004237037037037037, "loss": 0.6865, "step": 42560 }, { "epoch": 157.67, "grad_norm": 1.21616792678833, "learning_rate": 0.00042333333333333334, "loss": 0.6992, "step": 42570 }, { "epoch": 157.7, "grad_norm": 1.126473069190979, "learning_rate": 0.000422962962962963, "loss": 0.702, "step": 42580 }, { "epoch": 157.74, "grad_norm": 1.2085374593734741, "learning_rate": 0.00042259259259259257, "loss": 0.7191, "step": 42590 }, { "epoch": 157.78, "grad_norm": 1.2415745258331299, "learning_rate": 0.0004222222222222222, "loss": 0.715, "step": 42600 }, { "epoch": 157.81, "grad_norm": 1.2480597496032715, "learning_rate": 0.00042185185185185186, "loss": 0.7173, "step": 42610 }, { "epoch": 157.85, "grad_norm": 1.1208328008651733, "learning_rate": 0.00042148148148148145, "loss": 0.7093, "step": 42620 }, { "epoch": 157.89, "grad_norm": 1.2241952419281006, "learning_rate": 0.0004211111111111111, "loss": 0.7226, "step": 42630 }, { "epoch": 157.93, "grad_norm": 1.106781005859375, "learning_rate": 0.00042074074074074073, "loss": 0.7443, "step": 42640 }, { "epoch": 157.96, "grad_norm": 1.174251675605774, "learning_rate": 0.00042037037037037043, "loss": 0.722, "step": 42650 }, { "epoch": 158.0, "grad_norm": 2.178400993347168, "learning_rate": 0.00042, "loss": 0.724, "step": 42660 }, { "epoch": 158.04, "grad_norm": 1.0325825214385986, "learning_rate": 0.00041962962962962966, "loss": 0.5514, "step": 42670 }, { "epoch": 158.07, "grad_norm": 1.107414960861206, "learning_rate": 0.0004192592592592593, "loss": 0.599, "step": 42680 }, { "epoch": 158.11, "grad_norm": 1.071731686592102, "learning_rate": 0.0004188888888888889, "loss": 0.5881, "step": 42690 }, { "epoch": 158.15, "grad_norm": 1.1810855865478516, "learning_rate": 0.00041851851851851853, "loss": 0.5949, "step": 42700 }, { "epoch": 158.19, "grad_norm": 1.0901010036468506, "learning_rate": 0.0004181481481481482, "loss": 0.5967, "step": 42710 }, { "epoch": 158.22, "grad_norm": 1.074057936668396, "learning_rate": 0.0004177777777777778, "loss": 0.6035, "step": 42720 }, { "epoch": 158.26, "grad_norm": 1.017867088317871, "learning_rate": 0.0004174074074074074, "loss": 0.6148, "step": 42730 }, { "epoch": 158.3, "grad_norm": 1.0730246305465698, "learning_rate": 0.00041703703703703705, "loss": 0.628, "step": 42740 }, { "epoch": 158.33, "grad_norm": 1.1320347785949707, "learning_rate": 0.0004166666666666667, "loss": 0.6399, "step": 42750 }, { "epoch": 158.37, "grad_norm": 1.1257990598678589, "learning_rate": 0.0004162962962962963, "loss": 0.6465, "step": 42760 }, { "epoch": 158.41, "grad_norm": 1.1283479928970337, "learning_rate": 0.0004159259259259259, "loss": 0.6439, "step": 42770 }, { "epoch": 158.44, "grad_norm": 1.1584910154342651, "learning_rate": 0.00041555555555555557, "loss": 0.6622, "step": 42780 }, { "epoch": 158.48, "grad_norm": 1.1500332355499268, "learning_rate": 0.0004151851851851852, "loss": 0.6598, "step": 42790 }, { "epoch": 158.52, "grad_norm": 1.0922963619232178, "learning_rate": 0.0004148148148148148, "loss": 0.6723, "step": 42800 }, { "epoch": 158.56, "grad_norm": 1.2014588117599487, "learning_rate": 0.00041444444444444444, "loss": 0.6881, "step": 42810 }, { "epoch": 158.59, "grad_norm": 1.1734037399291992, "learning_rate": 0.0004140740740740741, "loss": 0.6841, "step": 42820 }, { "epoch": 158.63, "grad_norm": 1.1617283821105957, "learning_rate": 0.00041370370370370367, "loss": 0.6866, "step": 42830 }, { "epoch": 158.67, "grad_norm": 1.1515743732452393, "learning_rate": 0.0004133333333333333, "loss": 0.6856, "step": 42840 }, { "epoch": 158.7, "grad_norm": 1.1375868320465088, "learning_rate": 0.00041296296296296296, "loss": 0.6761, "step": 42850 }, { "epoch": 158.74, "grad_norm": 1.1889747381210327, "learning_rate": 0.0004125925925925926, "loss": 0.7045, "step": 42860 }, { "epoch": 158.78, "grad_norm": 1.1877621412277222, "learning_rate": 0.00041222222222222224, "loss": 0.6864, "step": 42870 }, { "epoch": 158.81, "grad_norm": 1.2101047039031982, "learning_rate": 0.0004118518518518519, "loss": 0.7121, "step": 42880 }, { "epoch": 158.85, "grad_norm": 1.1501836776733398, "learning_rate": 0.00041148148148148153, "loss": 0.6959, "step": 42890 }, { "epoch": 158.89, "grad_norm": 1.1214256286621094, "learning_rate": 0.0004111111111111111, "loss": 0.7205, "step": 42900 }, { "epoch": 158.93, "grad_norm": 1.2997136116027832, "learning_rate": 0.00041074074074074076, "loss": 0.7005, "step": 42910 }, { "epoch": 158.96, "grad_norm": 1.1824320554733276, "learning_rate": 0.0004103703703703704, "loss": 0.722, "step": 42920 }, { "epoch": 159.0, "grad_norm": 2.101818084716797, "learning_rate": 0.00041, "loss": 0.7108, "step": 42930 }, { "epoch": 159.04, "grad_norm": 1.0299731492996216, "learning_rate": 0.00040962962962962963, "loss": 0.5651, "step": 42940 }, { "epoch": 159.07, "grad_norm": 1.1588903665542603, "learning_rate": 0.0004092592592592593, "loss": 0.5606, "step": 42950 }, { "epoch": 159.11, "grad_norm": 1.090228796005249, "learning_rate": 0.0004088888888888889, "loss": 0.58, "step": 42960 }, { "epoch": 159.15, "grad_norm": 1.138484239578247, "learning_rate": 0.0004085185185185185, "loss": 0.596, "step": 42970 }, { "epoch": 159.19, "grad_norm": 1.0343921184539795, "learning_rate": 0.00040814814814814815, "loss": 0.6039, "step": 42980 }, { "epoch": 159.22, "grad_norm": 1.137735366821289, "learning_rate": 0.0004077777777777778, "loss": 0.6124, "step": 42990 }, { "epoch": 159.26, "grad_norm": 1.0816062688827515, "learning_rate": 0.0004074074074074074, "loss": 0.6041, "step": 43000 }, { "epoch": 159.3, "grad_norm": 1.1814439296722412, "learning_rate": 0.000407037037037037, "loss": 0.6102, "step": 43010 }, { "epoch": 159.33, "grad_norm": 1.1498757600784302, "learning_rate": 0.00040666666666666667, "loss": 0.6114, "step": 43020 }, { "epoch": 159.37, "grad_norm": 1.0893162488937378, "learning_rate": 0.0004062962962962963, "loss": 0.6276, "step": 43030 }, { "epoch": 159.41, "grad_norm": 1.1037179231643677, "learning_rate": 0.0004059259259259259, "loss": 0.6303, "step": 43040 }, { "epoch": 159.44, "grad_norm": 1.0938615798950195, "learning_rate": 0.00040555555555555554, "loss": 0.6602, "step": 43050 }, { "epoch": 159.48, "grad_norm": 1.2383320331573486, "learning_rate": 0.00040518518518518524, "loss": 0.672, "step": 43060 }, { "epoch": 159.52, "grad_norm": 1.1433382034301758, "learning_rate": 0.0004048148148148148, "loss": 0.6534, "step": 43070 }, { "epoch": 159.56, "grad_norm": 1.1631580591201782, "learning_rate": 0.00040444444444444447, "loss": 0.6525, "step": 43080 }, { "epoch": 159.59, "grad_norm": 1.1940419673919678, "learning_rate": 0.0004040740740740741, "loss": 0.6784, "step": 43090 }, { "epoch": 159.63, "grad_norm": 1.2554081678390503, "learning_rate": 0.00040370370370370375, "loss": 0.6611, "step": 43100 }, { "epoch": 159.67, "grad_norm": 1.2060763835906982, "learning_rate": 0.00040333333333333334, "loss": 0.6781, "step": 43110 }, { "epoch": 159.7, "grad_norm": 1.21128511428833, "learning_rate": 0.000402962962962963, "loss": 0.6779, "step": 43120 }, { "epoch": 159.74, "grad_norm": 1.1475038528442383, "learning_rate": 0.00040259259259259263, "loss": 0.7052, "step": 43130 }, { "epoch": 159.78, "grad_norm": 1.1346590518951416, "learning_rate": 0.0004022222222222222, "loss": 0.6993, "step": 43140 }, { "epoch": 159.81, "grad_norm": 1.2075724601745605, "learning_rate": 0.00040185185185185186, "loss": 0.6974, "step": 43150 }, { "epoch": 159.85, "grad_norm": 1.1834874153137207, "learning_rate": 0.0004014814814814815, "loss": 0.6846, "step": 43160 }, { "epoch": 159.89, "grad_norm": 1.2245084047317505, "learning_rate": 0.0004011111111111111, "loss": 0.7141, "step": 43170 }, { "epoch": 159.93, "grad_norm": 1.1702778339385986, "learning_rate": 0.00040074074074074073, "loss": 0.7074, "step": 43180 }, { "epoch": 159.96, "grad_norm": 1.2088655233383179, "learning_rate": 0.0004003703703703704, "loss": 0.7047, "step": 43190 }, { "epoch": 160.0, "grad_norm": 2.2771294116973877, "learning_rate": 0.0004, "loss": 0.7056, "step": 43200 }, { "epoch": 160.04, "grad_norm": 1.0756103992462158, "learning_rate": 0.0003996296296296296, "loss": 0.5502, "step": 43210 }, { "epoch": 160.07, "grad_norm": 1.0512828826904297, "learning_rate": 0.00039925925925925925, "loss": 0.5546, "step": 43220 }, { "epoch": 160.11, "grad_norm": 1.0669021606445312, "learning_rate": 0.0003988888888888889, "loss": 0.5743, "step": 43230 }, { "epoch": 160.15, "grad_norm": 1.011168360710144, "learning_rate": 0.0003985185185185185, "loss": 0.581, "step": 43240 }, { "epoch": 160.19, "grad_norm": 1.0751960277557373, "learning_rate": 0.0003981481481481481, "loss": 0.6031, "step": 43250 }, { "epoch": 160.22, "grad_norm": 1.123462438583374, "learning_rate": 0.00039777777777777777, "loss": 0.5781, "step": 43260 }, { "epoch": 160.26, "grad_norm": 1.0884305238723755, "learning_rate": 0.00039740740740740746, "loss": 0.6239, "step": 43270 }, { "epoch": 160.3, "grad_norm": 1.1230250597000122, "learning_rate": 0.00039703703703703705, "loss": 0.6233, "step": 43280 }, { "epoch": 160.33, "grad_norm": 1.1342320442199707, "learning_rate": 0.0003966666666666667, "loss": 0.6271, "step": 43290 }, { "epoch": 160.37, "grad_norm": 1.0703272819519043, "learning_rate": 0.00039629629629629634, "loss": 0.621, "step": 43300 }, { "epoch": 160.41, "grad_norm": 1.132581353187561, "learning_rate": 0.0003959259259259259, "loss": 0.6373, "step": 43310 }, { "epoch": 160.44, "grad_norm": 1.085752248764038, "learning_rate": 0.00039555555555555557, "loss": 0.6455, "step": 43320 }, { "epoch": 160.48, "grad_norm": 1.1563119888305664, "learning_rate": 0.0003951851851851852, "loss": 0.6423, "step": 43330 }, { "epoch": 160.52, "grad_norm": 1.2023621797561646, "learning_rate": 0.00039481481481481485, "loss": 0.657, "step": 43340 }, { "epoch": 160.56, "grad_norm": 1.1549417972564697, "learning_rate": 0.00039444444444444444, "loss": 0.6601, "step": 43350 }, { "epoch": 160.59, "grad_norm": 1.18620765209198, "learning_rate": 0.0003940740740740741, "loss": 0.6578, "step": 43360 }, { "epoch": 160.63, "grad_norm": 1.1379402875900269, "learning_rate": 0.00039370370370370373, "loss": 0.6663, "step": 43370 }, { "epoch": 160.67, "grad_norm": 1.1374726295471191, "learning_rate": 0.0003933333333333333, "loss": 0.6819, "step": 43380 }, { "epoch": 160.7, "grad_norm": 1.245043158531189, "learning_rate": 0.00039296296296296296, "loss": 0.6728, "step": 43390 }, { "epoch": 160.74, "grad_norm": 1.167038917541504, "learning_rate": 0.0003925925925925926, "loss": 0.668, "step": 43400 }, { "epoch": 160.78, "grad_norm": 1.1261484622955322, "learning_rate": 0.00039222222222222225, "loss": 0.675, "step": 43410 }, { "epoch": 160.81, "grad_norm": 1.1429600715637207, "learning_rate": 0.00039185185185185183, "loss": 0.6842, "step": 43420 }, { "epoch": 160.85, "grad_norm": 1.22481369972229, "learning_rate": 0.0003914814814814815, "loss": 0.6691, "step": 43430 }, { "epoch": 160.89, "grad_norm": 1.1873610019683838, "learning_rate": 0.0003911111111111111, "loss": 0.6811, "step": 43440 }, { "epoch": 160.93, "grad_norm": 1.1495413780212402, "learning_rate": 0.0003907407407407407, "loss": 0.6972, "step": 43450 }, { "epoch": 160.96, "grad_norm": 1.1746470928192139, "learning_rate": 0.00039037037037037035, "loss": 0.6851, "step": 43460 }, { "epoch": 161.0, "grad_norm": 2.1555185317993164, "learning_rate": 0.00039000000000000005, "loss": 0.6896, "step": 43470 }, { "epoch": 161.04, "grad_norm": 1.07627534866333, "learning_rate": 0.00038962962962962964, "loss": 0.5404, "step": 43480 }, { "epoch": 161.07, "grad_norm": 1.0882971286773682, "learning_rate": 0.0003892592592592593, "loss": 0.5519, "step": 43490 }, { "epoch": 161.11, "grad_norm": 1.0781952142715454, "learning_rate": 0.0003888888888888889, "loss": 0.5598, "step": 43500 }, { "epoch": 161.15, "grad_norm": 1.0606975555419922, "learning_rate": 0.00038851851851851856, "loss": 0.5889, "step": 43510 }, { "epoch": 161.19, "grad_norm": 1.0489012002944946, "learning_rate": 0.00038814814814814815, "loss": 0.5871, "step": 43520 }, { "epoch": 161.22, "grad_norm": 1.1675165891647339, "learning_rate": 0.0003877777777777778, "loss": 0.6078, "step": 43530 }, { "epoch": 161.26, "grad_norm": 1.1082695722579956, "learning_rate": 0.00038740740740740744, "loss": 0.5854, "step": 43540 }, { "epoch": 161.3, "grad_norm": 1.103467583656311, "learning_rate": 0.000387037037037037, "loss": 0.5976, "step": 43550 }, { "epoch": 161.33, "grad_norm": 1.0924053192138672, "learning_rate": 0.00038666666666666667, "loss": 0.6114, "step": 43560 }, { "epoch": 161.37, "grad_norm": 1.1562738418579102, "learning_rate": 0.0003862962962962963, "loss": 0.6082, "step": 43570 }, { "epoch": 161.41, "grad_norm": 1.0727171897888184, "learning_rate": 0.00038592592592592595, "loss": 0.6207, "step": 43580 }, { "epoch": 161.44, "grad_norm": 1.0738524198532104, "learning_rate": 0.00038555555555555554, "loss": 0.6413, "step": 43590 }, { "epoch": 161.48, "grad_norm": 1.1282587051391602, "learning_rate": 0.0003851851851851852, "loss": 0.6179, "step": 43600 }, { "epoch": 161.52, "grad_norm": 1.1933233737945557, "learning_rate": 0.00038481481481481483, "loss": 0.6282, "step": 43610 }, { "epoch": 161.56, "grad_norm": 1.1274137496948242, "learning_rate": 0.0003844444444444444, "loss": 0.6456, "step": 43620 }, { "epoch": 161.59, "grad_norm": 1.0575157403945923, "learning_rate": 0.00038407407407407406, "loss": 0.6469, "step": 43630 }, { "epoch": 161.63, "grad_norm": 1.1219662427902222, "learning_rate": 0.0003837037037037037, "loss": 0.6514, "step": 43640 }, { "epoch": 161.67, "grad_norm": 1.160602331161499, "learning_rate": 0.00038333333333333334, "loss": 0.6681, "step": 43650 }, { "epoch": 161.7, "grad_norm": 1.1490904092788696, "learning_rate": 0.00038296296296296293, "loss": 0.6755, "step": 43660 }, { "epoch": 161.74, "grad_norm": 1.1233144998550415, "learning_rate": 0.0003825925925925926, "loss": 0.6718, "step": 43670 }, { "epoch": 161.78, "grad_norm": 1.1319135427474976, "learning_rate": 0.0003822222222222223, "loss": 0.68, "step": 43680 }, { "epoch": 161.81, "grad_norm": 1.164638638496399, "learning_rate": 0.00038185185185185186, "loss": 0.6809, "step": 43690 }, { "epoch": 161.85, "grad_norm": 1.2083297967910767, "learning_rate": 0.0003814814814814815, "loss": 0.6953, "step": 43700 }, { "epoch": 161.89, "grad_norm": 1.1989994049072266, "learning_rate": 0.00038111111111111115, "loss": 0.691, "step": 43710 }, { "epoch": 161.93, "grad_norm": 1.2003412246704102, "learning_rate": 0.0003807407407407408, "loss": 0.6784, "step": 43720 }, { "epoch": 161.96, "grad_norm": 1.177047610282898, "learning_rate": 0.0003803703703703704, "loss": 0.6812, "step": 43730 }, { "epoch": 162.0, "grad_norm": 2.104875326156616, "learning_rate": 0.00038, "loss": 0.6884, "step": 43740 }, { "epoch": 162.04, "grad_norm": 1.0653473138809204, "learning_rate": 0.00037962962962962966, "loss": 0.5542, "step": 43750 }, { "epoch": 162.07, "grad_norm": 1.1091032028198242, "learning_rate": 0.00037925925925925925, "loss": 0.5432, "step": 43760 }, { "epoch": 162.11, "grad_norm": 0.9920121431350708, "learning_rate": 0.0003788888888888889, "loss": 0.5498, "step": 43770 }, { "epoch": 162.15, "grad_norm": 1.0519205331802368, "learning_rate": 0.00037851851851851854, "loss": 0.555, "step": 43780 }, { "epoch": 162.19, "grad_norm": 1.044488787651062, "learning_rate": 0.0003781481481481481, "loss": 0.5752, "step": 43790 }, { "epoch": 162.22, "grad_norm": 1.165174126625061, "learning_rate": 0.00037777777777777777, "loss": 0.5811, "step": 43800 }, { "epoch": 162.26, "grad_norm": 1.1170642375946045, "learning_rate": 0.0003774074074074074, "loss": 0.5996, "step": 43810 }, { "epoch": 162.3, "grad_norm": 1.0534666776657104, "learning_rate": 0.00037703703703703705, "loss": 0.6039, "step": 43820 }, { "epoch": 162.33, "grad_norm": 1.1040012836456299, "learning_rate": 0.00037666666666666664, "loss": 0.6114, "step": 43830 }, { "epoch": 162.37, "grad_norm": 1.1486854553222656, "learning_rate": 0.0003762962962962963, "loss": 0.6191, "step": 43840 }, { "epoch": 162.41, "grad_norm": 1.1414391994476318, "learning_rate": 0.00037592592592592593, "loss": 0.6203, "step": 43850 }, { "epoch": 162.44, "grad_norm": 1.213144063949585, "learning_rate": 0.0003755555555555555, "loss": 0.6278, "step": 43860 }, { "epoch": 162.48, "grad_norm": 1.127974033355713, "learning_rate": 0.00037518518518518516, "loss": 0.6341, "step": 43870 }, { "epoch": 162.52, "grad_norm": 1.1302952766418457, "learning_rate": 0.00037481481481481486, "loss": 0.623, "step": 43880 }, { "epoch": 162.56, "grad_norm": 1.1535022258758545, "learning_rate": 0.0003744444444444445, "loss": 0.66, "step": 43890 }, { "epoch": 162.59, "grad_norm": 1.2633349895477295, "learning_rate": 0.0003740740740740741, "loss": 0.6539, "step": 43900 }, { "epoch": 162.63, "grad_norm": 1.1691466569900513, "learning_rate": 0.00037370370370370373, "loss": 0.6448, "step": 43910 }, { "epoch": 162.67, "grad_norm": 1.20991051197052, "learning_rate": 0.0003733333333333334, "loss": 0.6407, "step": 43920 }, { "epoch": 162.7, "grad_norm": 1.1535216569900513, "learning_rate": 0.00037296296296296296, "loss": 0.6649, "step": 43930 }, { "epoch": 162.74, "grad_norm": 1.1679757833480835, "learning_rate": 0.0003725925925925926, "loss": 0.6485, "step": 43940 }, { "epoch": 162.78, "grad_norm": 1.1296665668487549, "learning_rate": 0.00037222222222222225, "loss": 0.6507, "step": 43950 }, { "epoch": 162.81, "grad_norm": 1.1518938541412354, "learning_rate": 0.0003718518518518519, "loss": 0.66, "step": 43960 }, { "epoch": 162.85, "grad_norm": 1.1524932384490967, "learning_rate": 0.0003714814814814815, "loss": 0.67, "step": 43970 }, { "epoch": 162.89, "grad_norm": 1.1855889558792114, "learning_rate": 0.0003711111111111111, "loss": 0.6697, "step": 43980 }, { "epoch": 162.93, "grad_norm": 1.200710415840149, "learning_rate": 0.00037074074074074076, "loss": 0.6605, "step": 43990 }, { "epoch": 162.96, "grad_norm": 1.20967435836792, "learning_rate": 0.00037037037037037035, "loss": 0.6831, "step": 44000 }, { "epoch": 163.0, "grad_norm": 3.300697088241577, "learning_rate": 0.00037, "loss": 0.663, "step": 44010 }, { "epoch": 163.04, "grad_norm": 1.1124473810195923, "learning_rate": 0.00036962962962962964, "loss": 0.5306, "step": 44020 }, { "epoch": 163.07, "grad_norm": 1.1132569313049316, "learning_rate": 0.0003692592592592592, "loss": 0.5461, "step": 44030 }, { "epoch": 163.11, "grad_norm": 1.1740922927856445, "learning_rate": 0.00036888888888888887, "loss": 0.5393, "step": 44040 }, { "epoch": 163.15, "grad_norm": 1.1028344631195068, "learning_rate": 0.0003685185185185185, "loss": 0.5571, "step": 44050 }, { "epoch": 163.19, "grad_norm": 1.029313564300537, "learning_rate": 0.00036814814814814815, "loss": 0.5373, "step": 44060 }, { "epoch": 163.22, "grad_norm": 1.1128367185592651, "learning_rate": 0.00036777777777777774, "loss": 0.5853, "step": 44070 }, { "epoch": 163.26, "grad_norm": 1.0616388320922852, "learning_rate": 0.00036740740740740744, "loss": 0.5839, "step": 44080 }, { "epoch": 163.3, "grad_norm": 1.1045690774917603, "learning_rate": 0.0003670370370370371, "loss": 0.5866, "step": 44090 }, { "epoch": 163.33, "grad_norm": 1.0463433265686035, "learning_rate": 0.00036666666666666667, "loss": 0.5984, "step": 44100 }, { "epoch": 163.37, "grad_norm": 1.213541865348816, "learning_rate": 0.0003662962962962963, "loss": 0.6033, "step": 44110 }, { "epoch": 163.41, "grad_norm": 1.070902943611145, "learning_rate": 0.00036592592592592596, "loss": 0.6147, "step": 44120 }, { "epoch": 163.44, "grad_norm": 1.0918152332305908, "learning_rate": 0.0003655555555555556, "loss": 0.6183, "step": 44130 }, { "epoch": 163.48, "grad_norm": 1.1264549493789673, "learning_rate": 0.0003651851851851852, "loss": 0.6266, "step": 44140 }, { "epoch": 163.52, "grad_norm": 1.174247145652771, "learning_rate": 0.00036481481481481483, "loss": 0.6421, "step": 44150 }, { "epoch": 163.56, "grad_norm": 1.1570000648498535, "learning_rate": 0.00036444444444444447, "loss": 0.6308, "step": 44160 }, { "epoch": 163.59, "grad_norm": 1.2352962493896484, "learning_rate": 0.00036407407407407406, "loss": 0.6307, "step": 44170 }, { "epoch": 163.63, "grad_norm": 1.1054643392562866, "learning_rate": 0.0003637037037037037, "loss": 0.6464, "step": 44180 }, { "epoch": 163.67, "grad_norm": 1.065085530281067, "learning_rate": 0.00036333333333333335, "loss": 0.6382, "step": 44190 }, { "epoch": 163.7, "grad_norm": 1.1962122917175293, "learning_rate": 0.000362962962962963, "loss": 0.6451, "step": 44200 }, { "epoch": 163.74, "grad_norm": 1.2009409666061401, "learning_rate": 0.0003625925925925926, "loss": 0.6479, "step": 44210 }, { "epoch": 163.78, "grad_norm": 1.1283183097839355, "learning_rate": 0.0003622222222222222, "loss": 0.6485, "step": 44220 }, { "epoch": 163.81, "grad_norm": 1.1767650842666626, "learning_rate": 0.00036185185185185186, "loss": 0.6593, "step": 44230 }, { "epoch": 163.85, "grad_norm": 1.154915690422058, "learning_rate": 0.00036148148148148145, "loss": 0.675, "step": 44240 }, { "epoch": 163.89, "grad_norm": 1.1761237382888794, "learning_rate": 0.0003611111111111111, "loss": 0.662, "step": 44250 }, { "epoch": 163.93, "grad_norm": 1.2740514278411865, "learning_rate": 0.00036074074074074074, "loss": 0.6671, "step": 44260 }, { "epoch": 163.96, "grad_norm": 1.1661285161972046, "learning_rate": 0.0003603703703703704, "loss": 0.6729, "step": 44270 }, { "epoch": 164.0, "grad_norm": 2.0054233074188232, "learning_rate": 0.00035999999999999997, "loss": 0.6686, "step": 44280 }, { "epoch": 164.04, "grad_norm": 0.9928212761878967, "learning_rate": 0.00035962962962962967, "loss": 0.5332, "step": 44290 }, { "epoch": 164.07, "grad_norm": 1.057337760925293, "learning_rate": 0.0003592592592592593, "loss": 0.5347, "step": 44300 }, { "epoch": 164.11, "grad_norm": 1.0809550285339355, "learning_rate": 0.0003588888888888889, "loss": 0.5378, "step": 44310 }, { "epoch": 164.15, "grad_norm": 0.9812051653862, "learning_rate": 0.00035851851851851854, "loss": 0.5619, "step": 44320 }, { "epoch": 164.19, "grad_norm": 1.0844264030456543, "learning_rate": 0.0003581481481481482, "loss": 0.5586, "step": 44330 }, { "epoch": 164.22, "grad_norm": 1.1218096017837524, "learning_rate": 0.00035777777777777777, "loss": 0.5732, "step": 44340 }, { "epoch": 164.26, "grad_norm": 1.1573761701583862, "learning_rate": 0.0003574074074074074, "loss": 0.5704, "step": 44350 }, { "epoch": 164.3, "grad_norm": 1.2225539684295654, "learning_rate": 0.00035703703703703706, "loss": 0.5686, "step": 44360 }, { "epoch": 164.33, "grad_norm": 1.0911756753921509, "learning_rate": 0.0003566666666666667, "loss": 0.5971, "step": 44370 }, { "epoch": 164.37, "grad_norm": 1.0996484756469727, "learning_rate": 0.0003562962962962963, "loss": 0.5912, "step": 44380 }, { "epoch": 164.41, "grad_norm": 1.1536465883255005, "learning_rate": 0.00035592592592592593, "loss": 0.6102, "step": 44390 }, { "epoch": 164.44, "grad_norm": 1.212332844734192, "learning_rate": 0.00035555555555555557, "loss": 0.6031, "step": 44400 }, { "epoch": 164.48, "grad_norm": 1.1705257892608643, "learning_rate": 0.00035518518518518516, "loss": 0.5988, "step": 44410 }, { "epoch": 164.52, "grad_norm": 1.1567353010177612, "learning_rate": 0.0003548148148148148, "loss": 0.6318, "step": 44420 }, { "epoch": 164.56, "grad_norm": 1.1380826234817505, "learning_rate": 0.00035444444444444445, "loss": 0.6225, "step": 44430 }, { "epoch": 164.59, "grad_norm": 1.2010952234268188, "learning_rate": 0.0003540740740740741, "loss": 0.6333, "step": 44440 }, { "epoch": 164.63, "grad_norm": 1.1813087463378906, "learning_rate": 0.0003537037037037037, "loss": 0.6223, "step": 44450 }, { "epoch": 164.67, "grad_norm": 1.2010884284973145, "learning_rate": 0.0003533333333333333, "loss": 0.6246, "step": 44460 }, { "epoch": 164.7, "grad_norm": 1.1103819608688354, "learning_rate": 0.00035296296296296296, "loss": 0.6417, "step": 44470 }, { "epoch": 164.74, "grad_norm": 1.2009899616241455, "learning_rate": 0.00035259259259259255, "loss": 0.6318, "step": 44480 }, { "epoch": 164.78, "grad_norm": 1.2279729843139648, "learning_rate": 0.00035222222222222225, "loss": 0.6437, "step": 44490 }, { "epoch": 164.81, "grad_norm": 1.1632214784622192, "learning_rate": 0.0003518518518518519, "loss": 0.6419, "step": 44500 }, { "epoch": 164.85, "grad_norm": 1.2233213186264038, "learning_rate": 0.00035148148148148153, "loss": 0.6453, "step": 44510 }, { "epoch": 164.89, "grad_norm": 1.2718476057052612, "learning_rate": 0.0003511111111111111, "loss": 0.6696, "step": 44520 }, { "epoch": 164.93, "grad_norm": 1.1639232635498047, "learning_rate": 0.00035074074074074077, "loss": 0.6662, "step": 44530 }, { "epoch": 164.96, "grad_norm": 1.1556235551834106, "learning_rate": 0.0003503703703703704, "loss": 0.6543, "step": 44540 }, { "epoch": 165.0, "grad_norm": 2.22497296333313, "learning_rate": 0.00035, "loss": 0.6562, "step": 44550 }, { "epoch": 165.04, "grad_norm": 0.9978173971176147, "learning_rate": 0.00034962962962962964, "loss": 0.5205, "step": 44560 }, { "epoch": 165.07, "grad_norm": 0.9967636466026306, "learning_rate": 0.0003492592592592593, "loss": 0.5098, "step": 44570 }, { "epoch": 165.11, "grad_norm": 1.1091960668563843, "learning_rate": 0.0003488888888888889, "loss": 0.5431, "step": 44580 }, { "epoch": 165.15, "grad_norm": 1.0127582550048828, "learning_rate": 0.0003485185185185185, "loss": 0.5463, "step": 44590 }, { "epoch": 165.19, "grad_norm": 1.1253612041473389, "learning_rate": 0.00034814814814814816, "loss": 0.5565, "step": 44600 }, { "epoch": 165.22, "grad_norm": 1.1062602996826172, "learning_rate": 0.0003477777777777778, "loss": 0.5653, "step": 44610 }, { "epoch": 165.26, "grad_norm": 1.0730866193771362, "learning_rate": 0.0003474074074074074, "loss": 0.5714, "step": 44620 }, { "epoch": 165.3, "grad_norm": 1.1267579793930054, "learning_rate": 0.00034703703703703703, "loss": 0.5642, "step": 44630 }, { "epoch": 165.33, "grad_norm": 1.1193773746490479, "learning_rate": 0.00034666666666666667, "loss": 0.5742, "step": 44640 }, { "epoch": 165.37, "grad_norm": 1.1450809240341187, "learning_rate": 0.00034629629629629626, "loss": 0.5998, "step": 44650 }, { "epoch": 165.41, "grad_norm": 1.165490746498108, "learning_rate": 0.0003459259259259259, "loss": 0.593, "step": 44660 }, { "epoch": 165.44, "grad_norm": 1.1239349842071533, "learning_rate": 0.00034555555555555555, "loss": 0.5855, "step": 44670 }, { "epoch": 165.48, "grad_norm": 1.146379828453064, "learning_rate": 0.0003451851851851852, "loss": 0.6084, "step": 44680 }, { "epoch": 165.52, "grad_norm": 1.2076387405395508, "learning_rate": 0.0003448148148148148, "loss": 0.6022, "step": 44690 }, { "epoch": 165.56, "grad_norm": 1.1927461624145508, "learning_rate": 0.0003444444444444445, "loss": 0.6044, "step": 44700 }, { "epoch": 165.59, "grad_norm": 1.2381534576416016, "learning_rate": 0.0003440740740740741, "loss": 0.6095, "step": 44710 }, { "epoch": 165.63, "grad_norm": 1.1408498287200928, "learning_rate": 0.0003437037037037037, "loss": 0.6088, "step": 44720 }, { "epoch": 165.67, "grad_norm": 1.1445554494857788, "learning_rate": 0.00034333333333333335, "loss": 0.6393, "step": 44730 }, { "epoch": 165.7, "grad_norm": 1.2025806903839111, "learning_rate": 0.000342962962962963, "loss": 0.6335, "step": 44740 }, { "epoch": 165.74, "grad_norm": 1.2227649688720703, "learning_rate": 0.00034259259259259263, "loss": 0.6401, "step": 44750 }, { "epoch": 165.78, "grad_norm": 1.217417597770691, "learning_rate": 0.0003422222222222222, "loss": 0.6424, "step": 44760 }, { "epoch": 165.81, "grad_norm": 1.2169790267944336, "learning_rate": 0.00034185185185185187, "loss": 0.6449, "step": 44770 }, { "epoch": 165.85, "grad_norm": 1.1809437274932861, "learning_rate": 0.0003414814814814815, "loss": 0.6452, "step": 44780 }, { "epoch": 165.89, "grad_norm": 1.1234064102172852, "learning_rate": 0.0003411111111111111, "loss": 0.6436, "step": 44790 }, { "epoch": 165.93, "grad_norm": 1.191071629524231, "learning_rate": 0.00034074074074074074, "loss": 0.6425, "step": 44800 }, { "epoch": 165.96, "grad_norm": 1.2422044277191162, "learning_rate": 0.0003403703703703704, "loss": 0.6441, "step": 44810 }, { "epoch": 166.0, "grad_norm": 1.9944589138031006, "learning_rate": 0.00034, "loss": 0.6568, "step": 44820 }, { "epoch": 166.04, "grad_norm": 1.0018469095230103, "learning_rate": 0.0003396296296296296, "loss": 0.5225, "step": 44830 }, { "epoch": 166.07, "grad_norm": 1.1185327768325806, "learning_rate": 0.00033925925925925926, "loss": 0.5186, "step": 44840 }, { "epoch": 166.11, "grad_norm": 1.0297755002975464, "learning_rate": 0.0003388888888888889, "loss": 0.5285, "step": 44850 }, { "epoch": 166.15, "grad_norm": 1.054671049118042, "learning_rate": 0.0003385185185185185, "loss": 0.5527, "step": 44860 }, { "epoch": 166.19, "grad_norm": 1.0687144994735718, "learning_rate": 0.00033814814814814813, "loss": 0.5511, "step": 44870 }, { "epoch": 166.22, "grad_norm": 1.1111799478530884, "learning_rate": 0.00033777777777777777, "loss": 0.5365, "step": 44880 }, { "epoch": 166.26, "grad_norm": 1.1225335597991943, "learning_rate": 0.0003374074074074074, "loss": 0.5641, "step": 44890 }, { "epoch": 166.3, "grad_norm": 1.1786049604415894, "learning_rate": 0.00033703703703703706, "loss": 0.5798, "step": 44900 }, { "epoch": 166.33, "grad_norm": 1.1004527807235718, "learning_rate": 0.0003366666666666667, "loss": 0.5602, "step": 44910 }, { "epoch": 166.37, "grad_norm": 1.1575289964675903, "learning_rate": 0.00033629629629629634, "loss": 0.5808, "step": 44920 }, { "epoch": 166.41, "grad_norm": 1.1519818305969238, "learning_rate": 0.00033592592592592593, "loss": 0.581, "step": 44930 }, { "epoch": 166.44, "grad_norm": 1.2028568983078003, "learning_rate": 0.0003355555555555556, "loss": 0.5835, "step": 44940 }, { "epoch": 166.48, "grad_norm": 1.12227201461792, "learning_rate": 0.0003351851851851852, "loss": 0.5909, "step": 44950 }, { "epoch": 166.52, "grad_norm": 1.1458628177642822, "learning_rate": 0.0003348148148148148, "loss": 0.5777, "step": 44960 }, { "epoch": 166.56, "grad_norm": 1.2187118530273438, "learning_rate": 0.00033444444444444445, "loss": 0.6024, "step": 44970 }, { "epoch": 166.59, "grad_norm": 1.142576813697815, "learning_rate": 0.0003340740740740741, "loss": 0.6061, "step": 44980 }, { "epoch": 166.63, "grad_norm": 1.136089563369751, "learning_rate": 0.00033370370370370373, "loss": 0.6222, "step": 44990 }, { "epoch": 166.67, "grad_norm": 1.1738704442977905, "learning_rate": 0.0003333333333333333, "loss": 0.6156, "step": 45000 }, { "epoch": 166.7, "grad_norm": 1.1535799503326416, "learning_rate": 0.00033296296296296296, "loss": 0.6352, "step": 45010 }, { "epoch": 166.74, "grad_norm": 1.1678482294082642, "learning_rate": 0.0003325925925925926, "loss": 0.612, "step": 45020 }, { "epoch": 166.78, "grad_norm": 1.1075305938720703, "learning_rate": 0.0003322222222222222, "loss": 0.6254, "step": 45030 }, { "epoch": 166.81, "grad_norm": 1.2057133913040161, "learning_rate": 0.00033185185185185184, "loss": 0.6143, "step": 45040 }, { "epoch": 166.85, "grad_norm": 1.150559425354004, "learning_rate": 0.0003314814814814815, "loss": 0.6465, "step": 45050 }, { "epoch": 166.89, "grad_norm": 1.2744975090026855, "learning_rate": 0.0003311111111111111, "loss": 0.6407, "step": 45060 }, { "epoch": 166.93, "grad_norm": 1.1772456169128418, "learning_rate": 0.0003307407407407407, "loss": 0.6456, "step": 45070 }, { "epoch": 166.96, "grad_norm": 1.225772500038147, "learning_rate": 0.00033037037037037036, "loss": 0.6467, "step": 45080 }, { "epoch": 167.0, "grad_norm": 2.134972095489502, "learning_rate": 0.00033, "loss": 0.6476, "step": 45090 }, { "epoch": 167.04, "grad_norm": 1.0380076169967651, "learning_rate": 0.0003296296296296296, "loss": 0.503, "step": 45100 }, { "epoch": 167.07, "grad_norm": 0.9915314316749573, "learning_rate": 0.0003292592592592593, "loss": 0.5193, "step": 45110 }, { "epoch": 167.11, "grad_norm": 0.9994888305664062, "learning_rate": 0.0003288888888888889, "loss": 0.515, "step": 45120 }, { "epoch": 167.15, "grad_norm": 1.0514576435089111, "learning_rate": 0.00032851851851851857, "loss": 0.5342, "step": 45130 }, { "epoch": 167.19, "grad_norm": 0.9884275197982788, "learning_rate": 0.00032814814814814816, "loss": 0.5299, "step": 45140 }, { "epoch": 167.22, "grad_norm": 1.1106001138687134, "learning_rate": 0.0003277777777777778, "loss": 0.5489, "step": 45150 }, { "epoch": 167.26, "grad_norm": 1.092725157737732, "learning_rate": 0.00032740740740740744, "loss": 0.5464, "step": 45160 }, { "epoch": 167.3, "grad_norm": 1.180444359779358, "learning_rate": 0.00032703703703703703, "loss": 0.5625, "step": 45170 }, { "epoch": 167.33, "grad_norm": 1.0736066102981567, "learning_rate": 0.0003266666666666667, "loss": 0.5595, "step": 45180 }, { "epoch": 167.37, "grad_norm": 1.1144076585769653, "learning_rate": 0.0003262962962962963, "loss": 0.5551, "step": 45190 }, { "epoch": 167.41, "grad_norm": 1.1394331455230713, "learning_rate": 0.00032592592592592596, "loss": 0.5559, "step": 45200 }, { "epoch": 167.44, "grad_norm": 1.0996838808059692, "learning_rate": 0.00032555555555555555, "loss": 0.5846, "step": 45210 }, { "epoch": 167.48, "grad_norm": 1.1391139030456543, "learning_rate": 0.0003251851851851852, "loss": 0.5875, "step": 45220 }, { "epoch": 167.52, "grad_norm": 1.0955880880355835, "learning_rate": 0.00032481481481481483, "loss": 0.5841, "step": 45230 }, { "epoch": 167.56, "grad_norm": 1.1299819946289062, "learning_rate": 0.0003244444444444444, "loss": 0.6052, "step": 45240 }, { "epoch": 167.59, "grad_norm": 1.1784850358963013, "learning_rate": 0.00032407407407407406, "loss": 0.6128, "step": 45250 }, { "epoch": 167.63, "grad_norm": 1.2381561994552612, "learning_rate": 0.0003237037037037037, "loss": 0.6095, "step": 45260 }, { "epoch": 167.67, "grad_norm": 1.1277177333831787, "learning_rate": 0.0003233333333333333, "loss": 0.6038, "step": 45270 }, { "epoch": 167.7, "grad_norm": 1.1459439992904663, "learning_rate": 0.00032296296296296294, "loss": 0.6173, "step": 45280 }, { "epoch": 167.74, "grad_norm": 1.2963416576385498, "learning_rate": 0.0003225925925925926, "loss": 0.6303, "step": 45290 }, { "epoch": 167.78, "grad_norm": 1.1779850721359253, "learning_rate": 0.0003222222222222222, "loss": 0.6227, "step": 45300 }, { "epoch": 167.81, "grad_norm": 1.140274167060852, "learning_rate": 0.00032185185185185187, "loss": 0.6168, "step": 45310 }, { "epoch": 167.85, "grad_norm": 1.3273096084594727, "learning_rate": 0.0003214814814814815, "loss": 0.6326, "step": 45320 }, { "epoch": 167.89, "grad_norm": 1.1361197233200073, "learning_rate": 0.00032111111111111115, "loss": 0.6225, "step": 45330 }, { "epoch": 167.93, "grad_norm": 1.1830846071243286, "learning_rate": 0.00032074074074074074, "loss": 0.6335, "step": 45340 }, { "epoch": 167.96, "grad_norm": 1.2070780992507935, "learning_rate": 0.0003203703703703704, "loss": 0.6328, "step": 45350 }, { "epoch": 168.0, "grad_norm": 2.6271791458129883, "learning_rate": 0.00032, "loss": 0.6206, "step": 45360 }, { "epoch": 168.04, "grad_norm": 1.031340479850769, "learning_rate": 0.00031962962962962967, "loss": 0.5002, "step": 45370 }, { "epoch": 168.07, "grad_norm": 1.0149120092391968, "learning_rate": 0.00031925925925925926, "loss": 0.5076, "step": 45380 }, { "epoch": 168.11, "grad_norm": 1.0120203495025635, "learning_rate": 0.0003188888888888889, "loss": 0.5156, "step": 45390 }, { "epoch": 168.15, "grad_norm": 1.0357682704925537, "learning_rate": 0.00031851851851851854, "loss": 0.5137, "step": 45400 }, { "epoch": 168.19, "grad_norm": 1.0558855533599854, "learning_rate": 0.00031814814814814813, "loss": 0.523, "step": 45410 }, { "epoch": 168.22, "grad_norm": 1.186493992805481, "learning_rate": 0.0003177777777777778, "loss": 0.5383, "step": 45420 }, { "epoch": 168.26, "grad_norm": 1.0746095180511475, "learning_rate": 0.0003174074074074074, "loss": 0.5434, "step": 45430 }, { "epoch": 168.3, "grad_norm": 1.0704513788223267, "learning_rate": 0.00031703703703703706, "loss": 0.5586, "step": 45440 }, { "epoch": 168.33, "grad_norm": 1.1498819589614868, "learning_rate": 0.00031666666666666665, "loss": 0.5414, "step": 45450 }, { "epoch": 168.37, "grad_norm": 1.046481728553772, "learning_rate": 0.0003162962962962963, "loss": 0.5719, "step": 45460 }, { "epoch": 168.41, "grad_norm": 1.098426342010498, "learning_rate": 0.00031592592592592593, "loss": 0.559, "step": 45470 }, { "epoch": 168.44, "grad_norm": 1.1073126792907715, "learning_rate": 0.0003155555555555555, "loss": 0.5704, "step": 45480 }, { "epoch": 168.48, "grad_norm": 1.1616339683532715, "learning_rate": 0.00031518518518518516, "loss": 0.577, "step": 45490 }, { "epoch": 168.52, "grad_norm": 1.1858264207839966, "learning_rate": 0.0003148148148148148, "loss": 0.5877, "step": 45500 }, { "epoch": 168.56, "grad_norm": 1.1451951265335083, "learning_rate": 0.0003144444444444445, "loss": 0.5846, "step": 45510 }, { "epoch": 168.59, "grad_norm": 1.1581388711929321, "learning_rate": 0.0003140740740740741, "loss": 0.603, "step": 45520 }, { "epoch": 168.63, "grad_norm": 1.1635332107543945, "learning_rate": 0.00031370370370370374, "loss": 0.5953, "step": 45530 }, { "epoch": 168.67, "grad_norm": 1.1322771310806274, "learning_rate": 0.0003133333333333334, "loss": 0.5944, "step": 45540 }, { "epoch": 168.7, "grad_norm": 1.2044880390167236, "learning_rate": 0.00031296296296296297, "loss": 0.6159, "step": 45550 }, { "epoch": 168.74, "grad_norm": 1.1870005130767822, "learning_rate": 0.0003125925925925926, "loss": 0.5962, "step": 45560 }, { "epoch": 168.78, "grad_norm": 1.1659802198410034, "learning_rate": 0.00031222222222222225, "loss": 0.5992, "step": 45570 }, { "epoch": 168.81, "grad_norm": 1.113723635673523, "learning_rate": 0.00031185185185185184, "loss": 0.6133, "step": 45580 }, { "epoch": 168.85, "grad_norm": 1.2763746976852417, "learning_rate": 0.0003114814814814815, "loss": 0.627, "step": 45590 }, { "epoch": 168.89, "grad_norm": 1.1561163663864136, "learning_rate": 0.0003111111111111111, "loss": 0.6268, "step": 45600 }, { "epoch": 168.93, "grad_norm": 1.2296404838562012, "learning_rate": 0.00031074074074074077, "loss": 0.613, "step": 45610 }, { "epoch": 168.96, "grad_norm": 1.1916122436523438, "learning_rate": 0.00031037037037037036, "loss": 0.6283, "step": 45620 }, { "epoch": 169.0, "grad_norm": 2.852231740951538, "learning_rate": 0.00031, "loss": 0.6324, "step": 45630 }, { "epoch": 169.04, "grad_norm": 1.0171003341674805, "learning_rate": 0.00030962962962962964, "loss": 0.4882, "step": 45640 }, { "epoch": 169.07, "grad_norm": 1.0816006660461426, "learning_rate": 0.00030925925925925923, "loss": 0.4913, "step": 45650 }, { "epoch": 169.11, "grad_norm": 1.1123528480529785, "learning_rate": 0.0003088888888888889, "loss": 0.5073, "step": 45660 }, { "epoch": 169.15, "grad_norm": 1.0543406009674072, "learning_rate": 0.0003085185185185185, "loss": 0.5064, "step": 45670 }, { "epoch": 169.19, "grad_norm": 1.0657761096954346, "learning_rate": 0.00030814814814814816, "loss": 0.5045, "step": 45680 }, { "epoch": 169.22, "grad_norm": 1.077353596687317, "learning_rate": 0.00030777777777777775, "loss": 0.534, "step": 45690 }, { "epoch": 169.26, "grad_norm": 1.0854156017303467, "learning_rate": 0.0003074074074074074, "loss": 0.5381, "step": 45700 }, { "epoch": 169.3, "grad_norm": 1.0773741006851196, "learning_rate": 0.00030703703703703703, "loss": 0.5381, "step": 45710 }, { "epoch": 169.33, "grad_norm": 1.239553451538086, "learning_rate": 0.0003066666666666667, "loss": 0.5517, "step": 45720 }, { "epoch": 169.37, "grad_norm": 1.1670728921890259, "learning_rate": 0.0003062962962962963, "loss": 0.5636, "step": 45730 }, { "epoch": 169.41, "grad_norm": 1.0546120405197144, "learning_rate": 0.00030592592592592596, "loss": 0.5617, "step": 45740 }, { "epoch": 169.44, "grad_norm": 1.1616939306259155, "learning_rate": 0.0003055555555555556, "loss": 0.5634, "step": 45750 }, { "epoch": 169.48, "grad_norm": 1.1171976327896118, "learning_rate": 0.0003051851851851852, "loss": 0.5781, "step": 45760 }, { "epoch": 169.52, "grad_norm": 1.0559581518173218, "learning_rate": 0.00030481481481481484, "loss": 0.5842, "step": 45770 }, { "epoch": 169.56, "grad_norm": 1.1623177528381348, "learning_rate": 0.0003044444444444445, "loss": 0.5636, "step": 45780 }, { "epoch": 169.59, "grad_norm": 1.1423313617706299, "learning_rate": 0.00030407407407407407, "loss": 0.5774, "step": 45790 }, { "epoch": 169.63, "grad_norm": 1.2008979320526123, "learning_rate": 0.0003037037037037037, "loss": 0.5891, "step": 45800 }, { "epoch": 169.67, "grad_norm": 1.146252155303955, "learning_rate": 0.00030333333333333335, "loss": 0.6024, "step": 45810 }, { "epoch": 169.7, "grad_norm": 1.2153797149658203, "learning_rate": 0.000302962962962963, "loss": 0.5919, "step": 45820 }, { "epoch": 169.74, "grad_norm": 1.2314988374710083, "learning_rate": 0.0003025925925925926, "loss": 0.5962, "step": 45830 }, { "epoch": 169.78, "grad_norm": 1.1446729898452759, "learning_rate": 0.0003022222222222222, "loss": 0.5979, "step": 45840 }, { "epoch": 169.81, "grad_norm": 1.1876449584960938, "learning_rate": 0.00030185185185185187, "loss": 0.5944, "step": 45850 }, { "epoch": 169.85, "grad_norm": 1.222914695739746, "learning_rate": 0.00030148148148148146, "loss": 0.6196, "step": 45860 }, { "epoch": 169.89, "grad_norm": 1.2583873271942139, "learning_rate": 0.0003011111111111111, "loss": 0.6079, "step": 45870 }, { "epoch": 169.93, "grad_norm": 1.2030463218688965, "learning_rate": 0.00030074074074074074, "loss": 0.6167, "step": 45880 }, { "epoch": 169.96, "grad_norm": 1.1214995384216309, "learning_rate": 0.00030037037037037033, "loss": 0.6141, "step": 45890 }, { "epoch": 170.0, "grad_norm": 2.046583652496338, "learning_rate": 0.0003, "loss": 0.6237, "step": 45900 }, { "epoch": 170.04, "grad_norm": 0.9938816428184509, "learning_rate": 0.0002996296296296296, "loss": 0.4945, "step": 45910 }, { "epoch": 170.07, "grad_norm": 1.0744401216506958, "learning_rate": 0.0002992592592592593, "loss": 0.4909, "step": 45920 }, { "epoch": 170.11, "grad_norm": 1.006912350654602, "learning_rate": 0.0002988888888888889, "loss": 0.5027, "step": 45930 }, { "epoch": 170.15, "grad_norm": 1.0353368520736694, "learning_rate": 0.00029851851851851854, "loss": 0.5027, "step": 45940 }, { "epoch": 170.19, "grad_norm": 1.0443346500396729, "learning_rate": 0.0002981481481481482, "loss": 0.5333, "step": 45950 }, { "epoch": 170.22, "grad_norm": 1.082785964012146, "learning_rate": 0.0002977777777777778, "loss": 0.5197, "step": 45960 }, { "epoch": 170.26, "grad_norm": 1.0719095468521118, "learning_rate": 0.0002974074074074074, "loss": 0.5193, "step": 45970 }, { "epoch": 170.3, "grad_norm": 1.0522955656051636, "learning_rate": 0.00029703703703703706, "loss": 0.5319, "step": 45980 }, { "epoch": 170.33, "grad_norm": 1.0776923894882202, "learning_rate": 0.0002966666666666667, "loss": 0.5204, "step": 45990 }, { "epoch": 170.37, "grad_norm": 1.1122677326202393, "learning_rate": 0.0002962962962962963, "loss": 0.5291, "step": 46000 }, { "epoch": 170.41, "grad_norm": 1.1635634899139404, "learning_rate": 0.00029592592592592594, "loss": 0.5479, "step": 46010 }, { "epoch": 170.44, "grad_norm": 1.2051600217819214, "learning_rate": 0.0002955555555555556, "loss": 0.5419, "step": 46020 }, { "epoch": 170.48, "grad_norm": 1.1886661052703857, "learning_rate": 0.00029518518518518517, "loss": 0.571, "step": 46030 }, { "epoch": 170.52, "grad_norm": 1.1528897285461426, "learning_rate": 0.0002948148148148148, "loss": 0.5575, "step": 46040 }, { "epoch": 170.56, "grad_norm": 1.2057989835739136, "learning_rate": 0.00029444444444444445, "loss": 0.5681, "step": 46050 }, { "epoch": 170.59, "grad_norm": 1.1691752672195435, "learning_rate": 0.0002940740740740741, "loss": 0.573, "step": 46060 }, { "epoch": 170.63, "grad_norm": 1.1771482229232788, "learning_rate": 0.0002937037037037037, "loss": 0.5758, "step": 46070 }, { "epoch": 170.67, "grad_norm": 1.1611257791519165, "learning_rate": 0.0002933333333333333, "loss": 0.5841, "step": 46080 }, { "epoch": 170.7, "grad_norm": 1.1115514039993286, "learning_rate": 0.00029296296296296297, "loss": 0.5776, "step": 46090 }, { "epoch": 170.74, "grad_norm": 1.1738566160202026, "learning_rate": 0.00029259259259259256, "loss": 0.594, "step": 46100 }, { "epoch": 170.78, "grad_norm": 1.2470712661743164, "learning_rate": 0.0002922222222222222, "loss": 0.6051, "step": 46110 }, { "epoch": 170.81, "grad_norm": 1.1863471269607544, "learning_rate": 0.00029185185185185184, "loss": 0.6104, "step": 46120 }, { "epoch": 170.85, "grad_norm": 1.0882984399795532, "learning_rate": 0.00029148148148148154, "loss": 0.605, "step": 46130 }, { "epoch": 170.89, "grad_norm": 1.20223069190979, "learning_rate": 0.00029111111111111113, "loss": 0.5981, "step": 46140 }, { "epoch": 170.93, "grad_norm": 1.1262892484664917, "learning_rate": 0.00029074074074074077, "loss": 0.6117, "step": 46150 }, { "epoch": 170.96, "grad_norm": 1.2084956169128418, "learning_rate": 0.0002903703703703704, "loss": 0.6077, "step": 46160 }, { "epoch": 171.0, "grad_norm": 2.073090076446533, "learning_rate": 0.00029, "loss": 0.6047, "step": 46170 }, { "epoch": 171.04, "grad_norm": 0.9190633296966553, "learning_rate": 0.00028962962962962964, "loss": 0.4772, "step": 46180 }, { "epoch": 171.07, "grad_norm": 1.0562139749526978, "learning_rate": 0.0002892592592592593, "loss": 0.4771, "step": 46190 }, { "epoch": 171.11, "grad_norm": 1.0270583629608154, "learning_rate": 0.0002888888888888889, "loss": 0.4823, "step": 46200 }, { "epoch": 171.15, "grad_norm": 0.9361873865127563, "learning_rate": 0.0002885185185185185, "loss": 0.4957, "step": 46210 }, { "epoch": 171.19, "grad_norm": 1.0008320808410645, "learning_rate": 0.00028814814814814816, "loss": 0.5127, "step": 46220 }, { "epoch": 171.22, "grad_norm": 0.9977531433105469, "learning_rate": 0.0002877777777777778, "loss": 0.497, "step": 46230 }, { "epoch": 171.26, "grad_norm": 1.1229784488677979, "learning_rate": 0.0002874074074074074, "loss": 0.5163, "step": 46240 }, { "epoch": 171.3, "grad_norm": 1.0604517459869385, "learning_rate": 0.00028703703703703703, "loss": 0.5257, "step": 46250 }, { "epoch": 171.33, "grad_norm": 1.1533794403076172, "learning_rate": 0.0002866666666666667, "loss": 0.5284, "step": 46260 }, { "epoch": 171.37, "grad_norm": 1.1037262678146362, "learning_rate": 0.00028629629629629627, "loss": 0.5391, "step": 46270 }, { "epoch": 171.41, "grad_norm": 1.1206879615783691, "learning_rate": 0.0002859259259259259, "loss": 0.5462, "step": 46280 }, { "epoch": 171.44, "grad_norm": 1.1554394960403442, "learning_rate": 0.00028555555555555555, "loss": 0.5422, "step": 46290 }, { "epoch": 171.48, "grad_norm": 1.1456857919692993, "learning_rate": 0.0002851851851851852, "loss": 0.5466, "step": 46300 }, { "epoch": 171.52, "grad_norm": 1.1105420589447021, "learning_rate": 0.0002848148148148148, "loss": 0.5529, "step": 46310 }, { "epoch": 171.56, "grad_norm": 1.1906827688217163, "learning_rate": 0.0002844444444444444, "loss": 0.5663, "step": 46320 }, { "epoch": 171.59, "grad_norm": 1.1303131580352783, "learning_rate": 0.0002840740740740741, "loss": 0.5582, "step": 46330 }, { "epoch": 171.63, "grad_norm": 1.1390284299850464, "learning_rate": 0.0002837037037037037, "loss": 0.5775, "step": 46340 }, { "epoch": 171.67, "grad_norm": 1.1789822578430176, "learning_rate": 0.00028333333333333335, "loss": 0.5834, "step": 46350 }, { "epoch": 171.7, "grad_norm": 1.152759075164795, "learning_rate": 0.000282962962962963, "loss": 0.5897, "step": 46360 }, { "epoch": 171.74, "grad_norm": 1.1558884382247925, "learning_rate": 0.00028259259259259264, "loss": 0.59, "step": 46370 }, { "epoch": 171.78, "grad_norm": 1.1207094192504883, "learning_rate": 0.00028222222222222223, "loss": 0.5746, "step": 46380 }, { "epoch": 171.81, "grad_norm": 1.1491044759750366, "learning_rate": 0.00028185185185185187, "loss": 0.588, "step": 46390 }, { "epoch": 171.85, "grad_norm": 1.1739957332611084, "learning_rate": 0.0002814814814814815, "loss": 0.6055, "step": 46400 }, { "epoch": 171.89, "grad_norm": 1.1879185438156128, "learning_rate": 0.0002811111111111111, "loss": 0.5832, "step": 46410 }, { "epoch": 171.93, "grad_norm": 1.1847784519195557, "learning_rate": 0.00028074074074074074, "loss": 0.6117, "step": 46420 }, { "epoch": 171.96, "grad_norm": 1.1747502088546753, "learning_rate": 0.0002803703703703704, "loss": 0.5978, "step": 46430 }, { "epoch": 172.0, "grad_norm": 2.1897666454315186, "learning_rate": 0.00028000000000000003, "loss": 0.6104, "step": 46440 }, { "epoch": 172.04, "grad_norm": 0.9945102334022522, "learning_rate": 0.0002796296296296296, "loss": 0.479, "step": 46450 }, { "epoch": 172.07, "grad_norm": 0.9526515007019043, "learning_rate": 0.00027925925925925926, "loss": 0.4701, "step": 46460 }, { "epoch": 172.11, "grad_norm": 1.0737017393112183, "learning_rate": 0.0002788888888888889, "loss": 0.4879, "step": 46470 }, { "epoch": 172.15, "grad_norm": 1.036736249923706, "learning_rate": 0.0002785185185185185, "loss": 0.4923, "step": 46480 }, { "epoch": 172.19, "grad_norm": 1.0772678852081299, "learning_rate": 0.00027814814814814813, "loss": 0.5052, "step": 46490 }, { "epoch": 172.22, "grad_norm": 1.1343975067138672, "learning_rate": 0.0002777777777777778, "loss": 0.4975, "step": 46500 }, { "epoch": 172.26, "grad_norm": 1.077694058418274, "learning_rate": 0.00027740740740740737, "loss": 0.507, "step": 46510 }, { "epoch": 172.3, "grad_norm": 1.0724430084228516, "learning_rate": 0.000277037037037037, "loss": 0.5183, "step": 46520 }, { "epoch": 172.33, "grad_norm": 1.1396112442016602, "learning_rate": 0.00027666666666666665, "loss": 0.5216, "step": 46530 }, { "epoch": 172.37, "grad_norm": 1.0889900922775269, "learning_rate": 0.00027629629629629635, "loss": 0.5374, "step": 46540 }, { "epoch": 172.41, "grad_norm": 1.0954948663711548, "learning_rate": 0.00027592592592592594, "loss": 0.5285, "step": 46550 }, { "epoch": 172.44, "grad_norm": 1.1125346422195435, "learning_rate": 0.0002755555555555556, "loss": 0.5407, "step": 46560 }, { "epoch": 172.48, "grad_norm": 1.0506683588027954, "learning_rate": 0.0002751851851851852, "loss": 0.5337, "step": 46570 }, { "epoch": 172.52, "grad_norm": 1.1666901111602783, "learning_rate": 0.0002748148148148148, "loss": 0.5568, "step": 46580 }, { "epoch": 172.56, "grad_norm": 1.1451871395111084, "learning_rate": 0.00027444444444444445, "loss": 0.5502, "step": 46590 }, { "epoch": 172.59, "grad_norm": 1.1013054847717285, "learning_rate": 0.0002740740740740741, "loss": 0.5603, "step": 46600 }, { "epoch": 172.63, "grad_norm": 1.1822503805160522, "learning_rate": 0.00027370370370370374, "loss": 0.5711, "step": 46610 }, { "epoch": 172.67, "grad_norm": 1.215627908706665, "learning_rate": 0.00027333333333333333, "loss": 0.5529, "step": 46620 }, { "epoch": 172.7, "grad_norm": 1.1580448150634766, "learning_rate": 0.00027296296296296297, "loss": 0.5801, "step": 46630 }, { "epoch": 172.74, "grad_norm": 1.1888889074325562, "learning_rate": 0.0002725925925925926, "loss": 0.5762, "step": 46640 }, { "epoch": 172.78, "grad_norm": 1.1299220323562622, "learning_rate": 0.0002722222222222222, "loss": 0.5743, "step": 46650 }, { "epoch": 172.81, "grad_norm": 1.142557978630066, "learning_rate": 0.00027185185185185184, "loss": 0.5833, "step": 46660 }, { "epoch": 172.85, "grad_norm": 1.1916149854660034, "learning_rate": 0.0002714814814814815, "loss": 0.5906, "step": 46670 }, { "epoch": 172.89, "grad_norm": 1.1581250429153442, "learning_rate": 0.00027111111111111113, "loss": 0.5878, "step": 46680 }, { "epoch": 172.93, "grad_norm": 1.2044334411621094, "learning_rate": 0.0002707407407407407, "loss": 0.5708, "step": 46690 }, { "epoch": 172.96, "grad_norm": 1.279520869255066, "learning_rate": 0.00027037037037037036, "loss": 0.597, "step": 46700 }, { "epoch": 173.0, "grad_norm": 2.107897996902466, "learning_rate": 0.00027, "loss": 0.5836, "step": 46710 }, { "epoch": 173.04, "grad_norm": 1.0807936191558838, "learning_rate": 0.0002696296296296296, "loss": 0.4548, "step": 46720 }, { "epoch": 173.07, "grad_norm": 0.9994587302207947, "learning_rate": 0.00026925925925925923, "loss": 0.4707, "step": 46730 }, { "epoch": 173.11, "grad_norm": 1.0478770732879639, "learning_rate": 0.00026888888888888893, "loss": 0.4722, "step": 46740 }, { "epoch": 173.15, "grad_norm": 1.013415813446045, "learning_rate": 0.0002685185185185186, "loss": 0.5006, "step": 46750 }, { "epoch": 173.19, "grad_norm": 1.012677550315857, "learning_rate": 0.00026814814814814816, "loss": 0.4959, "step": 46760 }, { "epoch": 173.22, "grad_norm": 1.0860716104507446, "learning_rate": 0.0002677777777777778, "loss": 0.4915, "step": 46770 }, { "epoch": 173.26, "grad_norm": 1.157912254333496, "learning_rate": 0.00026740740740740745, "loss": 0.5134, "step": 46780 }, { "epoch": 173.3, "grad_norm": 1.1048921346664429, "learning_rate": 0.00026703703703703704, "loss": 0.5078, "step": 46790 }, { "epoch": 173.33, "grad_norm": 1.1350382566452026, "learning_rate": 0.0002666666666666667, "loss": 0.5103, "step": 46800 }, { "epoch": 173.37, "grad_norm": 1.0764796733856201, "learning_rate": 0.0002662962962962963, "loss": 0.5152, "step": 46810 }, { "epoch": 173.41, "grad_norm": 1.0592381954193115, "learning_rate": 0.0002659259259259259, "loss": 0.5038, "step": 46820 }, { "epoch": 173.44, "grad_norm": 1.0957568883895874, "learning_rate": 0.00026555555555555555, "loss": 0.5321, "step": 46830 }, { "epoch": 173.48, "grad_norm": 1.0465227365493774, "learning_rate": 0.0002651851851851852, "loss": 0.5263, "step": 46840 }, { "epoch": 173.52, "grad_norm": 1.106717824935913, "learning_rate": 0.00026481481481481484, "loss": 0.5371, "step": 46850 }, { "epoch": 173.56, "grad_norm": 1.0666769742965698, "learning_rate": 0.00026444444444444443, "loss": 0.5489, "step": 46860 }, { "epoch": 173.59, "grad_norm": 1.1631163358688354, "learning_rate": 0.00026407407407407407, "loss": 0.5543, "step": 46870 }, { "epoch": 173.63, "grad_norm": 1.2045526504516602, "learning_rate": 0.0002637037037037037, "loss": 0.5638, "step": 46880 }, { "epoch": 173.67, "grad_norm": 1.1705162525177002, "learning_rate": 0.0002633333333333333, "loss": 0.5475, "step": 46890 }, { "epoch": 173.7, "grad_norm": 1.1357591152191162, "learning_rate": 0.00026296296296296294, "loss": 0.5737, "step": 46900 }, { "epoch": 173.74, "grad_norm": 1.2002065181732178, "learning_rate": 0.0002625925925925926, "loss": 0.5689, "step": 46910 }, { "epoch": 173.78, "grad_norm": 1.1841059923171997, "learning_rate": 0.00026222222222222223, "loss": 0.5563, "step": 46920 }, { "epoch": 173.81, "grad_norm": 1.2427780628204346, "learning_rate": 0.0002618518518518518, "loss": 0.5768, "step": 46930 }, { "epoch": 173.85, "grad_norm": 1.1587187051773071, "learning_rate": 0.00026148148148148146, "loss": 0.5747, "step": 46940 }, { "epoch": 173.89, "grad_norm": 1.2276320457458496, "learning_rate": 0.00026111111111111116, "loss": 0.5696, "step": 46950 }, { "epoch": 173.93, "grad_norm": 1.2079700231552124, "learning_rate": 0.00026074074074074075, "loss": 0.5909, "step": 46960 }, { "epoch": 173.96, "grad_norm": 1.270209789276123, "learning_rate": 0.0002603703703703704, "loss": 0.5803, "step": 46970 }, { "epoch": 174.0, "grad_norm": 2.246609687805176, "learning_rate": 0.00026000000000000003, "loss": 0.5801, "step": 46980 }, { "epoch": 174.04, "grad_norm": 0.9848663806915283, "learning_rate": 0.0002596296296296297, "loss": 0.4537, "step": 46990 }, { "epoch": 174.07, "grad_norm": 1.026583194732666, "learning_rate": 0.00025925925925925926, "loss": 0.4583, "step": 47000 }, { "epoch": 174.11, "grad_norm": 1.0775405168533325, "learning_rate": 0.0002588888888888889, "loss": 0.4698, "step": 47010 }, { "epoch": 174.15, "grad_norm": 0.9840408563613892, "learning_rate": 0.00025851851851851855, "loss": 0.4818, "step": 47020 }, { "epoch": 174.19, "grad_norm": 0.9654091000556946, "learning_rate": 0.00025814814814814814, "loss": 0.4712, "step": 47030 }, { "epoch": 174.22, "grad_norm": 1.057830810546875, "learning_rate": 0.0002577777777777778, "loss": 0.4991, "step": 47040 }, { "epoch": 174.26, "grad_norm": 1.0927250385284424, "learning_rate": 0.0002574074074074074, "loss": 0.4966, "step": 47050 }, { "epoch": 174.3, "grad_norm": 1.0779294967651367, "learning_rate": 0.00025703703703703706, "loss": 0.5209, "step": 47060 }, { "epoch": 174.33, "grad_norm": 1.0855543613433838, "learning_rate": 0.00025666666666666665, "loss": 0.5019, "step": 47070 }, { "epoch": 174.37, "grad_norm": 1.1064202785491943, "learning_rate": 0.0002562962962962963, "loss": 0.5069, "step": 47080 }, { "epoch": 174.41, "grad_norm": 1.0853049755096436, "learning_rate": 0.00025592592592592594, "loss": 0.5084, "step": 47090 }, { "epoch": 174.44, "grad_norm": 1.0846391916275024, "learning_rate": 0.00025555555555555553, "loss": 0.5129, "step": 47100 }, { "epoch": 174.48, "grad_norm": 1.1598272323608398, "learning_rate": 0.00025518518518518517, "loss": 0.5432, "step": 47110 }, { "epoch": 174.52, "grad_norm": 1.1339960098266602, "learning_rate": 0.0002548148148148148, "loss": 0.5323, "step": 47120 }, { "epoch": 174.56, "grad_norm": 1.1392383575439453, "learning_rate": 0.0002544444444444444, "loss": 0.5429, "step": 47130 }, { "epoch": 174.59, "grad_norm": 1.126197338104248, "learning_rate": 0.00025407407407407404, "loss": 0.5453, "step": 47140 }, { "epoch": 174.63, "grad_norm": 1.1894769668579102, "learning_rate": 0.00025370370370370374, "loss": 0.5416, "step": 47150 }, { "epoch": 174.67, "grad_norm": 1.1315686702728271, "learning_rate": 0.0002533333333333334, "loss": 0.5651, "step": 47160 }, { "epoch": 174.7, "grad_norm": 1.0671674013137817, "learning_rate": 0.00025296296296296297, "loss": 0.5458, "step": 47170 }, { "epoch": 174.74, "grad_norm": 1.1603813171386719, "learning_rate": 0.0002525925925925926, "loss": 0.5408, "step": 47180 }, { "epoch": 174.78, "grad_norm": 1.2179709672927856, "learning_rate": 0.00025222222222222226, "loss": 0.5569, "step": 47190 }, { "epoch": 174.81, "grad_norm": 1.1304550170898438, "learning_rate": 0.00025185185185185185, "loss": 0.5602, "step": 47200 }, { "epoch": 174.85, "grad_norm": 1.1317181587219238, "learning_rate": 0.0002514814814814815, "loss": 0.5668, "step": 47210 }, { "epoch": 174.89, "grad_norm": 1.1473468542099, "learning_rate": 0.00025111111111111113, "loss": 0.5592, "step": 47220 }, { "epoch": 174.93, "grad_norm": 1.1926883459091187, "learning_rate": 0.0002507407407407408, "loss": 0.563, "step": 47230 }, { "epoch": 174.96, "grad_norm": 1.1721458435058594, "learning_rate": 0.00025037037037037036, "loss": 0.5819, "step": 47240 }, { "epoch": 175.0, "grad_norm": 2.136143207550049, "learning_rate": 0.00025, "loss": 0.5768, "step": 47250 }, { "epoch": 175.04, "grad_norm": 0.958111584186554, "learning_rate": 0.00024962962962962965, "loss": 0.4661, "step": 47260 }, { "epoch": 175.07, "grad_norm": 0.9953908324241638, "learning_rate": 0.0002492592592592593, "loss": 0.4594, "step": 47270 }, { "epoch": 175.11, "grad_norm": 0.9846945405006409, "learning_rate": 0.0002488888888888889, "loss": 0.4578, "step": 47280 }, { "epoch": 175.15, "grad_norm": 1.0534905195236206, "learning_rate": 0.0002485185185185185, "loss": 0.4543, "step": 47290 }, { "epoch": 175.19, "grad_norm": 1.0469465255737305, "learning_rate": 0.00024814814814814816, "loss": 0.4752, "step": 47300 }, { "epoch": 175.22, "grad_norm": 1.0580891370773315, "learning_rate": 0.0002477777777777778, "loss": 0.483, "step": 47310 }, { "epoch": 175.26, "grad_norm": 1.050082802772522, "learning_rate": 0.0002474074074074074, "loss": 0.485, "step": 47320 }, { "epoch": 175.3, "grad_norm": 1.0860666036605835, "learning_rate": 0.00024703703703703704, "loss": 0.5029, "step": 47330 }, { "epoch": 175.33, "grad_norm": 1.01170015335083, "learning_rate": 0.0002466666666666667, "loss": 0.5031, "step": 47340 }, { "epoch": 175.37, "grad_norm": 1.034176230430603, "learning_rate": 0.00024629629629629627, "loss": 0.5019, "step": 47350 }, { "epoch": 175.41, "grad_norm": 1.0043485164642334, "learning_rate": 0.00024592592592592597, "loss": 0.5153, "step": 47360 }, { "epoch": 175.44, "grad_norm": 1.0900912284851074, "learning_rate": 0.00024555555555555556, "loss": 0.5226, "step": 47370 }, { "epoch": 175.48, "grad_norm": 1.1493656635284424, "learning_rate": 0.0002451851851851852, "loss": 0.5326, "step": 47380 }, { "epoch": 175.52, "grad_norm": 1.1280782222747803, "learning_rate": 0.00024481481481481484, "loss": 0.52, "step": 47390 }, { "epoch": 175.56, "grad_norm": 1.1246672868728638, "learning_rate": 0.00024444444444444443, "loss": 0.5281, "step": 47400 }, { "epoch": 175.59, "grad_norm": 1.1507177352905273, "learning_rate": 0.0002440740740740741, "loss": 0.5317, "step": 47410 }, { "epoch": 175.63, "grad_norm": 1.1389291286468506, "learning_rate": 0.00024370370370370371, "loss": 0.5342, "step": 47420 }, { "epoch": 175.67, "grad_norm": 1.155777931213379, "learning_rate": 0.00024333333333333336, "loss": 0.5374, "step": 47430 }, { "epoch": 175.7, "grad_norm": 1.1306384801864624, "learning_rate": 0.00024296296296296297, "loss": 0.5269, "step": 47440 }, { "epoch": 175.74, "grad_norm": 1.1292226314544678, "learning_rate": 0.0002425925925925926, "loss": 0.5608, "step": 47450 }, { "epoch": 175.78, "grad_norm": 1.0860081911087036, "learning_rate": 0.00024222222222222223, "loss": 0.5423, "step": 47460 }, { "epoch": 175.81, "grad_norm": 1.207112431526184, "learning_rate": 0.00024185185185185185, "loss": 0.544, "step": 47470 }, { "epoch": 175.85, "grad_norm": 1.0947352647781372, "learning_rate": 0.0002414814814814815, "loss": 0.563, "step": 47480 }, { "epoch": 175.89, "grad_norm": 1.2116717100143433, "learning_rate": 0.0002411111111111111, "loss": 0.562, "step": 47490 }, { "epoch": 175.93, "grad_norm": 1.1679658889770508, "learning_rate": 0.00024074074074074072, "loss": 0.5582, "step": 47500 }, { "epoch": 175.96, "grad_norm": 1.1572608947753906, "learning_rate": 0.0002403703703703704, "loss": 0.5604, "step": 47510 }, { "epoch": 176.0, "grad_norm": 1.975321888923645, "learning_rate": 0.00024, "loss": 0.5608, "step": 47520 }, { "epoch": 176.04, "grad_norm": 0.9298834800720215, "learning_rate": 0.00023962962962962965, "loss": 0.4581, "step": 47530 }, { "epoch": 176.07, "grad_norm": 1.068063497543335, "learning_rate": 0.00023925925925925926, "loss": 0.4384, "step": 47540 }, { "epoch": 176.11, "grad_norm": 1.0253227949142456, "learning_rate": 0.0002388888888888889, "loss": 0.4508, "step": 47550 }, { "epoch": 176.15, "grad_norm": 1.007856011390686, "learning_rate": 0.00023851851851851852, "loss": 0.4536, "step": 47560 }, { "epoch": 176.19, "grad_norm": 1.0197805166244507, "learning_rate": 0.00023814814814814814, "loss": 0.4724, "step": 47570 }, { "epoch": 176.22, "grad_norm": 1.1127703189849854, "learning_rate": 0.00023777777777777778, "loss": 0.4804, "step": 47580 }, { "epoch": 176.26, "grad_norm": 1.1433093547821045, "learning_rate": 0.0002374074074074074, "loss": 0.4765, "step": 47590 }, { "epoch": 176.3, "grad_norm": 1.0717138051986694, "learning_rate": 0.00023703703703703704, "loss": 0.5006, "step": 47600 }, { "epoch": 176.33, "grad_norm": 1.1593323945999146, "learning_rate": 0.00023666666666666668, "loss": 0.4878, "step": 47610 }, { "epoch": 176.37, "grad_norm": 1.0496388673782349, "learning_rate": 0.00023629629629629632, "loss": 0.5008, "step": 47620 }, { "epoch": 176.41, "grad_norm": 1.0645349025726318, "learning_rate": 0.00023592592592592594, "loss": 0.5019, "step": 47630 }, { "epoch": 176.44, "grad_norm": 1.1091699600219727, "learning_rate": 0.00023555555555555556, "loss": 0.5004, "step": 47640 }, { "epoch": 176.48, "grad_norm": 1.156760573387146, "learning_rate": 0.0002351851851851852, "loss": 0.523, "step": 47650 }, { "epoch": 176.52, "grad_norm": 1.110307216644287, "learning_rate": 0.00023481481481481481, "loss": 0.5202, "step": 47660 }, { "epoch": 176.56, "grad_norm": 1.1058788299560547, "learning_rate": 0.00023444444444444446, "loss": 0.5114, "step": 47670 }, { "epoch": 176.59, "grad_norm": 1.1229724884033203, "learning_rate": 0.00023407407407407407, "loss": 0.533, "step": 47680 }, { "epoch": 176.63, "grad_norm": 1.1795262098312378, "learning_rate": 0.0002337037037037037, "loss": 0.5311, "step": 47690 }, { "epoch": 176.67, "grad_norm": 1.178025722503662, "learning_rate": 0.00023333333333333333, "loss": 0.5181, "step": 47700 }, { "epoch": 176.7, "grad_norm": 1.1427292823791504, "learning_rate": 0.00023296296296296295, "loss": 0.5399, "step": 47710 }, { "epoch": 176.74, "grad_norm": 1.0837512016296387, "learning_rate": 0.00023259259259259262, "loss": 0.5336, "step": 47720 }, { "epoch": 176.78, "grad_norm": 1.1414895057678223, "learning_rate": 0.00023222222222222223, "loss": 0.5309, "step": 47730 }, { "epoch": 176.81, "grad_norm": 1.1707152128219604, "learning_rate": 0.00023185185185185187, "loss": 0.5419, "step": 47740 }, { "epoch": 176.85, "grad_norm": 1.2246688604354858, "learning_rate": 0.0002314814814814815, "loss": 0.5507, "step": 47750 }, { "epoch": 176.89, "grad_norm": 1.0815807580947876, "learning_rate": 0.0002311111111111111, "loss": 0.5502, "step": 47760 }, { "epoch": 176.93, "grad_norm": 1.1408597230911255, "learning_rate": 0.00023074074074074075, "loss": 0.5663, "step": 47770 }, { "epoch": 176.96, "grad_norm": 1.1771601438522339, "learning_rate": 0.00023037037037037036, "loss": 0.5381, "step": 47780 }, { "epoch": 177.0, "grad_norm": 2.0750699043273926, "learning_rate": 0.00023, "loss": 0.5629, "step": 47790 }, { "epoch": 177.04, "grad_norm": 0.947895884513855, "learning_rate": 0.00022962962962962962, "loss": 0.4314, "step": 47800 }, { "epoch": 177.07, "grad_norm": 0.9859268069267273, "learning_rate": 0.00022925925925925924, "loss": 0.4374, "step": 47810 }, { "epoch": 177.11, "grad_norm": 0.9737774133682251, "learning_rate": 0.0002288888888888889, "loss": 0.4595, "step": 47820 }, { "epoch": 177.15, "grad_norm": 0.9961501955986023, "learning_rate": 0.00022851851851851852, "loss": 0.451, "step": 47830 }, { "epoch": 177.19, "grad_norm": 0.9911838173866272, "learning_rate": 0.00022814814814814817, "loss": 0.4618, "step": 47840 }, { "epoch": 177.22, "grad_norm": 1.0495216846466064, "learning_rate": 0.00022777777777777778, "loss": 0.4797, "step": 47850 }, { "epoch": 177.26, "grad_norm": 1.0654104948043823, "learning_rate": 0.00022740740740740742, "loss": 0.4764, "step": 47860 }, { "epoch": 177.3, "grad_norm": 1.047002911567688, "learning_rate": 0.00022703703703703704, "loss": 0.4677, "step": 47870 }, { "epoch": 177.33, "grad_norm": 1.1380468606948853, "learning_rate": 0.00022666666666666666, "loss": 0.4924, "step": 47880 }, { "epoch": 177.37, "grad_norm": 1.1353617906570435, "learning_rate": 0.0002262962962962963, "loss": 0.4937, "step": 47890 }, { "epoch": 177.41, "grad_norm": 1.0635641813278198, "learning_rate": 0.00022592592592592591, "loss": 0.4951, "step": 47900 }, { "epoch": 177.44, "grad_norm": 1.0273109674453735, "learning_rate": 0.00022555555555555556, "loss": 0.5008, "step": 47910 }, { "epoch": 177.48, "grad_norm": 1.1229729652404785, "learning_rate": 0.0002251851851851852, "loss": 0.5104, "step": 47920 }, { "epoch": 177.52, "grad_norm": 1.1669470071792603, "learning_rate": 0.00022481481481481484, "loss": 0.4993, "step": 47930 }, { "epoch": 177.56, "grad_norm": 1.099764108657837, "learning_rate": 0.00022444444444444446, "loss": 0.5089, "step": 47940 }, { "epoch": 177.59, "grad_norm": 1.1565130949020386, "learning_rate": 0.00022407407407407407, "loss": 0.5289, "step": 47950 }, { "epoch": 177.63, "grad_norm": 1.0863348245620728, "learning_rate": 0.00022370370370370372, "loss": 0.5218, "step": 47960 }, { "epoch": 177.67, "grad_norm": 1.1537193059921265, "learning_rate": 0.00022333333333333333, "loss": 0.5335, "step": 47970 }, { "epoch": 177.7, "grad_norm": 1.1691408157348633, "learning_rate": 0.00022296296296296297, "loss": 0.5228, "step": 47980 }, { "epoch": 177.74, "grad_norm": 1.1775586605072021, "learning_rate": 0.0002225925925925926, "loss": 0.5109, "step": 47990 }, { "epoch": 177.78, "grad_norm": 1.1857826709747314, "learning_rate": 0.0002222222222222222, "loss": 0.5208, "step": 48000 }, { "epoch": 177.81, "grad_norm": 1.2024893760681152, "learning_rate": 0.00022185185185185185, "loss": 0.5377, "step": 48010 }, { "epoch": 177.85, "grad_norm": 1.0835148096084595, "learning_rate": 0.0002214814814814815, "loss": 0.5361, "step": 48020 }, { "epoch": 177.89, "grad_norm": 1.0845205783843994, "learning_rate": 0.00022111111111111113, "loss": 0.5415, "step": 48030 }, { "epoch": 177.93, "grad_norm": 1.2087277173995972, "learning_rate": 0.00022074074074074075, "loss": 0.5304, "step": 48040 }, { "epoch": 177.96, "grad_norm": 1.1453560590744019, "learning_rate": 0.0002203703703703704, "loss": 0.5516, "step": 48050 }, { "epoch": 178.0, "grad_norm": 2.6465954780578613, "learning_rate": 0.00022, "loss": 0.5592, "step": 48060 }, { "epoch": 178.04, "grad_norm": 1.0190789699554443, "learning_rate": 0.00021962962962962962, "loss": 0.439, "step": 48070 }, { "epoch": 178.07, "grad_norm": 0.9200291633605957, "learning_rate": 0.00021925925925925927, "loss": 0.4346, "step": 48080 }, { "epoch": 178.11, "grad_norm": 0.9328666925430298, "learning_rate": 0.00021888888888888888, "loss": 0.4468, "step": 48090 }, { "epoch": 178.15, "grad_norm": 1.0163941383361816, "learning_rate": 0.00021851851851851852, "loss": 0.463, "step": 48100 }, { "epoch": 178.19, "grad_norm": 0.9540250301361084, "learning_rate": 0.00021814814814814814, "loss": 0.4362, "step": 48110 }, { "epoch": 178.22, "grad_norm": 1.0213724374771118, "learning_rate": 0.00021777777777777776, "loss": 0.4546, "step": 48120 }, { "epoch": 178.26, "grad_norm": 0.9999773502349854, "learning_rate": 0.00021740740740740743, "loss": 0.4761, "step": 48130 }, { "epoch": 178.3, "grad_norm": 0.992883026599884, "learning_rate": 0.00021703703703703704, "loss": 0.4708, "step": 48140 }, { "epoch": 178.33, "grad_norm": 1.0905901193618774, "learning_rate": 0.00021666666666666668, "loss": 0.476, "step": 48150 }, { "epoch": 178.37, "grad_norm": 1.0493611097335815, "learning_rate": 0.0002162962962962963, "loss": 0.4743, "step": 48160 }, { "epoch": 178.41, "grad_norm": 1.0430622100830078, "learning_rate": 0.00021592592592592594, "loss": 0.4763, "step": 48170 }, { "epoch": 178.44, "grad_norm": 1.0460885763168335, "learning_rate": 0.00021555555555555556, "loss": 0.48, "step": 48180 }, { "epoch": 178.48, "grad_norm": 1.0845812559127808, "learning_rate": 0.00021518518518518517, "loss": 0.4926, "step": 48190 }, { "epoch": 178.52, "grad_norm": 1.0286200046539307, "learning_rate": 0.00021481481481481482, "loss": 0.4815, "step": 48200 }, { "epoch": 178.56, "grad_norm": 1.0686702728271484, "learning_rate": 0.00021444444444444443, "loss": 0.5101, "step": 48210 }, { "epoch": 178.59, "grad_norm": 1.112003207206726, "learning_rate": 0.00021407407407407407, "loss": 0.5155, "step": 48220 }, { "epoch": 178.63, "grad_norm": 1.1415358781814575, "learning_rate": 0.00021370370370370372, "loss": 0.5, "step": 48230 }, { "epoch": 178.67, "grad_norm": 1.135293960571289, "learning_rate": 0.00021333333333333336, "loss": 0.5151, "step": 48240 }, { "epoch": 178.7, "grad_norm": 1.0956019163131714, "learning_rate": 0.00021296296296296298, "loss": 0.5123, "step": 48250 }, { "epoch": 178.74, "grad_norm": 1.150425910949707, "learning_rate": 0.0002125925925925926, "loss": 0.5311, "step": 48260 }, { "epoch": 178.78, "grad_norm": 1.1303415298461914, "learning_rate": 0.00021222222222222223, "loss": 0.5246, "step": 48270 }, { "epoch": 178.81, "grad_norm": 1.262482762336731, "learning_rate": 0.00021185185185185185, "loss": 0.5445, "step": 48280 }, { "epoch": 178.85, "grad_norm": 1.236796259880066, "learning_rate": 0.0002114814814814815, "loss": 0.5318, "step": 48290 }, { "epoch": 178.89, "grad_norm": 1.13998281955719, "learning_rate": 0.0002111111111111111, "loss": 0.5236, "step": 48300 }, { "epoch": 178.93, "grad_norm": 1.1601678133010864, "learning_rate": 0.00021074074074074072, "loss": 0.5397, "step": 48310 }, { "epoch": 178.96, "grad_norm": 1.1455624103546143, "learning_rate": 0.00021037037037037037, "loss": 0.5432, "step": 48320 }, { "epoch": 179.0, "grad_norm": 2.160454034805298, "learning_rate": 0.00021, "loss": 0.5366, "step": 48330 }, { "epoch": 179.04, "grad_norm": 1.0072879791259766, "learning_rate": 0.00020962962962962965, "loss": 0.4454, "step": 48340 }, { "epoch": 179.07, "grad_norm": 0.9614301919937134, "learning_rate": 0.00020925925925925927, "loss": 0.4311, "step": 48350 }, { "epoch": 179.11, "grad_norm": 0.9831352233886719, "learning_rate": 0.0002088888888888889, "loss": 0.4359, "step": 48360 }, { "epoch": 179.15, "grad_norm": 0.9975229501724243, "learning_rate": 0.00020851851851851853, "loss": 0.4511, "step": 48370 }, { "epoch": 179.19, "grad_norm": 1.075417399406433, "learning_rate": 0.00020814814814814814, "loss": 0.4466, "step": 48380 }, { "epoch": 179.22, "grad_norm": 0.9628404378890991, "learning_rate": 0.00020777777777777778, "loss": 0.4474, "step": 48390 }, { "epoch": 179.26, "grad_norm": 1.052886724472046, "learning_rate": 0.0002074074074074074, "loss": 0.446, "step": 48400 }, { "epoch": 179.3, "grad_norm": 1.0707443952560425, "learning_rate": 0.00020703703703703704, "loss": 0.4662, "step": 48410 }, { "epoch": 179.33, "grad_norm": 1.0438921451568604, "learning_rate": 0.00020666666666666666, "loss": 0.4695, "step": 48420 }, { "epoch": 179.37, "grad_norm": 1.076917052268982, "learning_rate": 0.0002062962962962963, "loss": 0.4781, "step": 48430 }, { "epoch": 179.41, "grad_norm": 1.067229986190796, "learning_rate": 0.00020592592592592594, "loss": 0.4713, "step": 48440 }, { "epoch": 179.44, "grad_norm": 1.0892492532730103, "learning_rate": 0.00020555555555555556, "loss": 0.4814, "step": 48450 }, { "epoch": 179.48, "grad_norm": 1.1713968515396118, "learning_rate": 0.0002051851851851852, "loss": 0.4865, "step": 48460 }, { "epoch": 179.52, "grad_norm": 1.089644193649292, "learning_rate": 0.00020481481481481482, "loss": 0.4923, "step": 48470 }, { "epoch": 179.56, "grad_norm": 1.1268770694732666, "learning_rate": 0.00020444444444444446, "loss": 0.52, "step": 48480 }, { "epoch": 179.59, "grad_norm": 1.0705839395523071, "learning_rate": 0.00020407407407407408, "loss": 0.5031, "step": 48490 }, { "epoch": 179.63, "grad_norm": 1.0443114042282104, "learning_rate": 0.0002037037037037037, "loss": 0.4888, "step": 48500 }, { "epoch": 179.67, "grad_norm": 1.0626062154769897, "learning_rate": 0.00020333333333333333, "loss": 0.509, "step": 48510 }, { "epoch": 179.7, "grad_norm": 1.1279469728469849, "learning_rate": 0.00020296296296296295, "loss": 0.5049, "step": 48520 }, { "epoch": 179.74, "grad_norm": 1.089354157447815, "learning_rate": 0.00020259259259259262, "loss": 0.4979, "step": 48530 }, { "epoch": 179.78, "grad_norm": 1.1244808435440063, "learning_rate": 0.00020222222222222223, "loss": 0.5134, "step": 48540 }, { "epoch": 179.81, "grad_norm": 1.0787869691848755, "learning_rate": 0.00020185185185185188, "loss": 0.5184, "step": 48550 }, { "epoch": 179.85, "grad_norm": 1.1457653045654297, "learning_rate": 0.0002014814814814815, "loss": 0.5201, "step": 48560 }, { "epoch": 179.89, "grad_norm": 1.1280841827392578, "learning_rate": 0.0002011111111111111, "loss": 0.5205, "step": 48570 }, { "epoch": 179.93, "grad_norm": 1.1580936908721924, "learning_rate": 0.00020074074074074075, "loss": 0.5329, "step": 48580 }, { "epoch": 179.96, "grad_norm": 1.0993684530258179, "learning_rate": 0.00020037037037037037, "loss": 0.5289, "step": 48590 }, { "epoch": 180.0, "grad_norm": 2.2351152896881104, "learning_rate": 0.0002, "loss": 0.5151, "step": 48600 }, { "epoch": 180.04, "grad_norm": 0.9843377470970154, "learning_rate": 0.00019962962962962963, "loss": 0.4304, "step": 48610 }, { "epoch": 180.07, "grad_norm": 0.8959612250328064, "learning_rate": 0.00019925925925925924, "loss": 0.4105, "step": 48620 }, { "epoch": 180.11, "grad_norm": 0.9292177557945251, "learning_rate": 0.00019888888888888888, "loss": 0.4352, "step": 48630 }, { "epoch": 180.15, "grad_norm": 0.9821608662605286, "learning_rate": 0.00019851851851851853, "loss": 0.4408, "step": 48640 }, { "epoch": 180.19, "grad_norm": 1.0155918598175049, "learning_rate": 0.00019814814814814817, "loss": 0.4278, "step": 48650 }, { "epoch": 180.22, "grad_norm": 1.047547698020935, "learning_rate": 0.00019777777777777778, "loss": 0.4486, "step": 48660 }, { "epoch": 180.26, "grad_norm": 1.0473016500473022, "learning_rate": 0.00019740740740740743, "loss": 0.4448, "step": 48670 }, { "epoch": 180.3, "grad_norm": 1.0475410223007202, "learning_rate": 0.00019703703703703704, "loss": 0.4602, "step": 48680 }, { "epoch": 180.33, "grad_norm": 1.0084902048110962, "learning_rate": 0.00019666666666666666, "loss": 0.4469, "step": 48690 }, { "epoch": 180.37, "grad_norm": 1.1199017763137817, "learning_rate": 0.0001962962962962963, "loss": 0.4632, "step": 48700 }, { "epoch": 180.41, "grad_norm": 1.0491690635681152, "learning_rate": 0.00019592592592592592, "loss": 0.4626, "step": 48710 }, { "epoch": 180.44, "grad_norm": 1.0795011520385742, "learning_rate": 0.00019555555555555556, "loss": 0.4826, "step": 48720 }, { "epoch": 180.48, "grad_norm": 1.0450068712234497, "learning_rate": 0.00019518518518518518, "loss": 0.4836, "step": 48730 }, { "epoch": 180.52, "grad_norm": 1.0705045461654663, "learning_rate": 0.00019481481481481482, "loss": 0.479, "step": 48740 }, { "epoch": 180.56, "grad_norm": 1.0864088535308838, "learning_rate": 0.00019444444444444446, "loss": 0.4816, "step": 48750 }, { "epoch": 180.59, "grad_norm": 1.0538465976715088, "learning_rate": 0.00019407407407407408, "loss": 0.4927, "step": 48760 }, { "epoch": 180.63, "grad_norm": 1.0714560747146606, "learning_rate": 0.00019370370370370372, "loss": 0.5075, "step": 48770 }, { "epoch": 180.67, "grad_norm": 1.0492222309112549, "learning_rate": 0.00019333333333333333, "loss": 0.4991, "step": 48780 }, { "epoch": 180.7, "grad_norm": 1.095389485359192, "learning_rate": 0.00019296296296296298, "loss": 0.5102, "step": 48790 }, { "epoch": 180.74, "grad_norm": 1.0268661975860596, "learning_rate": 0.0001925925925925926, "loss": 0.5101, "step": 48800 }, { "epoch": 180.78, "grad_norm": 1.0888630151748657, "learning_rate": 0.0001922222222222222, "loss": 0.5086, "step": 48810 }, { "epoch": 180.81, "grad_norm": 1.1224223375320435, "learning_rate": 0.00019185185185185185, "loss": 0.506, "step": 48820 }, { "epoch": 180.85, "grad_norm": 1.0878612995147705, "learning_rate": 0.00019148148148148147, "loss": 0.5174, "step": 48830 }, { "epoch": 180.89, "grad_norm": 1.1202062368392944, "learning_rate": 0.00019111111111111114, "loss": 0.5192, "step": 48840 }, { "epoch": 180.93, "grad_norm": 1.115593671798706, "learning_rate": 0.00019074074074074075, "loss": 0.5133, "step": 48850 }, { "epoch": 180.96, "grad_norm": 1.1733633279800415, "learning_rate": 0.0001903703703703704, "loss": 0.5132, "step": 48860 }, { "epoch": 181.0, "grad_norm": 2.1127378940582275, "learning_rate": 0.00019, "loss": 0.5097, "step": 48870 }, { "epoch": 181.04, "grad_norm": 0.9501312375068665, "learning_rate": 0.00018962962962962963, "loss": 0.4238, "step": 48880 }, { "epoch": 181.07, "grad_norm": 0.9944826364517212, "learning_rate": 0.00018925925925925927, "loss": 0.421, "step": 48890 }, { "epoch": 181.11, "grad_norm": 0.958714485168457, "learning_rate": 0.00018888888888888888, "loss": 0.4229, "step": 48900 }, { "epoch": 181.15, "grad_norm": 1.00678288936615, "learning_rate": 0.00018851851851851853, "loss": 0.4263, "step": 48910 }, { "epoch": 181.19, "grad_norm": 0.9769282937049866, "learning_rate": 0.00018814814814814814, "loss": 0.443, "step": 48920 }, { "epoch": 181.22, "grad_norm": 1.0257822275161743, "learning_rate": 0.00018777777777777776, "loss": 0.439, "step": 48930 }, { "epoch": 181.26, "grad_norm": 0.9551071524620056, "learning_rate": 0.00018740740740740743, "loss": 0.4545, "step": 48940 }, { "epoch": 181.3, "grad_norm": 0.9713193774223328, "learning_rate": 0.00018703703703703704, "loss": 0.4456, "step": 48950 }, { "epoch": 181.33, "grad_norm": 1.0076040029525757, "learning_rate": 0.0001866666666666667, "loss": 0.4457, "step": 48960 }, { "epoch": 181.37, "grad_norm": 1.1038206815719604, "learning_rate": 0.0001862962962962963, "loss": 0.4587, "step": 48970 }, { "epoch": 181.41, "grad_norm": 1.092207908630371, "learning_rate": 0.00018592592592592594, "loss": 0.4683, "step": 48980 }, { "epoch": 181.44, "grad_norm": 1.0330462455749512, "learning_rate": 0.00018555555555555556, "loss": 0.4657, "step": 48990 }, { "epoch": 181.48, "grad_norm": 1.0785881280899048, "learning_rate": 0.00018518518518518518, "loss": 0.4782, "step": 49000 }, { "epoch": 181.52, "grad_norm": 1.0754797458648682, "learning_rate": 0.00018481481481481482, "loss": 0.4602, "step": 49010 }, { "epoch": 181.56, "grad_norm": 1.1690996885299683, "learning_rate": 0.00018444444444444443, "loss": 0.4749, "step": 49020 }, { "epoch": 181.59, "grad_norm": 1.0903944969177246, "learning_rate": 0.00018407407407407408, "loss": 0.4791, "step": 49030 }, { "epoch": 181.63, "grad_norm": 1.09013032913208, "learning_rate": 0.00018370370370370372, "loss": 0.4915, "step": 49040 }, { "epoch": 181.67, "grad_norm": 1.0747276544570923, "learning_rate": 0.00018333333333333334, "loss": 0.491, "step": 49050 }, { "epoch": 181.7, "grad_norm": 1.0240190029144287, "learning_rate": 0.00018296296296296298, "loss": 0.4837, "step": 49060 }, { "epoch": 181.74, "grad_norm": 1.0852088928222656, "learning_rate": 0.0001825925925925926, "loss": 0.4984, "step": 49070 }, { "epoch": 181.78, "grad_norm": 1.1203248500823975, "learning_rate": 0.00018222222222222224, "loss": 0.5084, "step": 49080 }, { "epoch": 181.81, "grad_norm": 1.1268478631973267, "learning_rate": 0.00018185185185185185, "loss": 0.4992, "step": 49090 }, { "epoch": 181.85, "grad_norm": 1.137208342552185, "learning_rate": 0.0001814814814814815, "loss": 0.4983, "step": 49100 }, { "epoch": 181.89, "grad_norm": 1.1762620210647583, "learning_rate": 0.0001811111111111111, "loss": 0.5113, "step": 49110 }, { "epoch": 181.93, "grad_norm": 1.0956618785858154, "learning_rate": 0.00018074074074074073, "loss": 0.4969, "step": 49120 }, { "epoch": 181.96, "grad_norm": 1.088881015777588, "learning_rate": 0.00018037037037037037, "loss": 0.5037, "step": 49130 }, { "epoch": 182.0, "grad_norm": 2.1335837841033936, "learning_rate": 0.00017999999999999998, "loss": 0.5282, "step": 49140 }, { "epoch": 182.04, "grad_norm": 0.9405384659767151, "learning_rate": 0.00017962962962962965, "loss": 0.4223, "step": 49150 }, { "epoch": 182.07, "grad_norm": 0.9342048168182373, "learning_rate": 0.00017925925925925927, "loss": 0.4154, "step": 49160 }, { "epoch": 182.11, "grad_norm": 0.8926421999931335, "learning_rate": 0.00017888888888888889, "loss": 0.419, "step": 49170 }, { "epoch": 182.15, "grad_norm": 0.9453332424163818, "learning_rate": 0.00017851851851851853, "loss": 0.4238, "step": 49180 }, { "epoch": 182.19, "grad_norm": 0.9516483545303345, "learning_rate": 0.00017814814814814814, "loss": 0.4312, "step": 49190 }, { "epoch": 182.22, "grad_norm": 0.9720765948295593, "learning_rate": 0.00017777777777777779, "loss": 0.4319, "step": 49200 }, { "epoch": 182.26, "grad_norm": 0.9715983867645264, "learning_rate": 0.0001774074074074074, "loss": 0.4378, "step": 49210 }, { "epoch": 182.3, "grad_norm": 1.0771870613098145, "learning_rate": 0.00017703703703703704, "loss": 0.4448, "step": 49220 }, { "epoch": 182.33, "grad_norm": 1.0263774394989014, "learning_rate": 0.00017666666666666666, "loss": 0.4491, "step": 49230 }, { "epoch": 182.37, "grad_norm": 1.0612064599990845, "learning_rate": 0.00017629629629629628, "loss": 0.4352, "step": 49240 }, { "epoch": 182.41, "grad_norm": 1.0565451383590698, "learning_rate": 0.00017592592592592595, "loss": 0.4549, "step": 49250 }, { "epoch": 182.44, "grad_norm": 1.0483824014663696, "learning_rate": 0.00017555555555555556, "loss": 0.4589, "step": 49260 }, { "epoch": 182.48, "grad_norm": 1.0516337156295776, "learning_rate": 0.0001751851851851852, "loss": 0.4741, "step": 49270 }, { "epoch": 182.52, "grad_norm": 1.0335724353790283, "learning_rate": 0.00017481481481481482, "loss": 0.4833, "step": 49280 }, { "epoch": 182.56, "grad_norm": 1.072335124015808, "learning_rate": 0.00017444444444444446, "loss": 0.4679, "step": 49290 }, { "epoch": 182.59, "grad_norm": 1.1325234174728394, "learning_rate": 0.00017407407407407408, "loss": 0.4653, "step": 49300 }, { "epoch": 182.63, "grad_norm": 1.0939109325408936, "learning_rate": 0.0001737037037037037, "loss": 0.4782, "step": 49310 }, { "epoch": 182.67, "grad_norm": 1.039179801940918, "learning_rate": 0.00017333333333333334, "loss": 0.4767, "step": 49320 }, { "epoch": 182.7, "grad_norm": 1.054984450340271, "learning_rate": 0.00017296296296296295, "loss": 0.4917, "step": 49330 }, { "epoch": 182.74, "grad_norm": 1.0941307544708252, "learning_rate": 0.0001725925925925926, "loss": 0.4915, "step": 49340 }, { "epoch": 182.78, "grad_norm": 1.044124960899353, "learning_rate": 0.00017222222222222224, "loss": 0.4686, "step": 49350 }, { "epoch": 182.81, "grad_norm": 1.0401660203933716, "learning_rate": 0.00017185185185185185, "loss": 0.4846, "step": 49360 }, { "epoch": 182.85, "grad_norm": 1.0751421451568604, "learning_rate": 0.0001714814814814815, "loss": 0.4927, "step": 49370 }, { "epoch": 182.89, "grad_norm": 1.1070232391357422, "learning_rate": 0.0001711111111111111, "loss": 0.4959, "step": 49380 }, { "epoch": 182.93, "grad_norm": 1.1466519832611084, "learning_rate": 0.00017074074074074075, "loss": 0.4969, "step": 49390 }, { "epoch": 182.96, "grad_norm": 1.1511880159378052, "learning_rate": 0.00017037037037037037, "loss": 0.5007, "step": 49400 }, { "epoch": 183.0, "grad_norm": 1.9645870923995972, "learning_rate": 0.00017, "loss": 0.509, "step": 49410 }, { "epoch": 183.04, "grad_norm": 0.9212923049926758, "learning_rate": 0.00016962962962962963, "loss": 0.4048, "step": 49420 }, { "epoch": 183.07, "grad_norm": 0.9352565407752991, "learning_rate": 0.00016925925925925924, "loss": 0.4027, "step": 49430 }, { "epoch": 183.11, "grad_norm": 0.9354068040847778, "learning_rate": 0.00016888888888888889, "loss": 0.4185, "step": 49440 }, { "epoch": 183.15, "grad_norm": 1.0052289962768555, "learning_rate": 0.00016851851851851853, "loss": 0.4161, "step": 49450 }, { "epoch": 183.19, "grad_norm": 0.936885416507721, "learning_rate": 0.00016814814814814817, "loss": 0.4236, "step": 49460 }, { "epoch": 183.22, "grad_norm": 1.0176805257797241, "learning_rate": 0.0001677777777777778, "loss": 0.426, "step": 49470 }, { "epoch": 183.26, "grad_norm": 1.0213100910186768, "learning_rate": 0.0001674074074074074, "loss": 0.4317, "step": 49480 }, { "epoch": 183.3, "grad_norm": 0.9696800708770752, "learning_rate": 0.00016703703703703705, "loss": 0.4173, "step": 49490 }, { "epoch": 183.33, "grad_norm": 1.0048062801361084, "learning_rate": 0.00016666666666666666, "loss": 0.433, "step": 49500 }, { "epoch": 183.37, "grad_norm": 1.0107778310775757, "learning_rate": 0.0001662962962962963, "loss": 0.4351, "step": 49510 }, { "epoch": 183.41, "grad_norm": 0.9895990490913391, "learning_rate": 0.00016592592592592592, "loss": 0.4458, "step": 49520 }, { "epoch": 183.44, "grad_norm": 0.9655704498291016, "learning_rate": 0.00016555555555555556, "loss": 0.4602, "step": 49530 }, { "epoch": 183.48, "grad_norm": 1.0920255184173584, "learning_rate": 0.00016518518518518518, "loss": 0.4657, "step": 49540 }, { "epoch": 183.52, "grad_norm": 1.0697599649429321, "learning_rate": 0.0001648148148148148, "loss": 0.4542, "step": 49550 }, { "epoch": 183.56, "grad_norm": 1.09999680519104, "learning_rate": 0.00016444444444444446, "loss": 0.4724, "step": 49560 }, { "epoch": 183.59, "grad_norm": 1.034386157989502, "learning_rate": 0.00016407407407407408, "loss": 0.4664, "step": 49570 }, { "epoch": 183.63, "grad_norm": 1.041436791419983, "learning_rate": 0.00016370370370370372, "loss": 0.4695, "step": 49580 }, { "epoch": 183.67, "grad_norm": 1.0949628353118896, "learning_rate": 0.00016333333333333334, "loss": 0.4737, "step": 49590 }, { "epoch": 183.7, "grad_norm": 1.0208719968795776, "learning_rate": 0.00016296296296296298, "loss": 0.4803, "step": 49600 }, { "epoch": 183.74, "grad_norm": 1.042176604270935, "learning_rate": 0.0001625925925925926, "loss": 0.4867, "step": 49610 }, { "epoch": 183.78, "grad_norm": 1.0463050603866577, "learning_rate": 0.0001622222222222222, "loss": 0.4704, "step": 49620 }, { "epoch": 183.81, "grad_norm": 1.1194572448730469, "learning_rate": 0.00016185185185185185, "loss": 0.4778, "step": 49630 }, { "epoch": 183.85, "grad_norm": 1.06777024269104, "learning_rate": 0.00016148148148148147, "loss": 0.4745, "step": 49640 }, { "epoch": 183.89, "grad_norm": 1.1410287618637085, "learning_rate": 0.0001611111111111111, "loss": 0.496, "step": 49650 }, { "epoch": 183.93, "grad_norm": 1.0613508224487305, "learning_rate": 0.00016074074074074075, "loss": 0.4862, "step": 49660 }, { "epoch": 183.96, "grad_norm": 1.103725552558899, "learning_rate": 0.00016037037037037037, "loss": 0.5019, "step": 49670 }, { "epoch": 184.0, "grad_norm": 2.150892734527588, "learning_rate": 0.00016, "loss": 0.4927, "step": 49680 }, { "epoch": 184.04, "grad_norm": 0.8993950486183167, "learning_rate": 0.00015962962962962963, "loss": 0.3915, "step": 49690 }, { "epoch": 184.07, "grad_norm": 0.9276669025421143, "learning_rate": 0.00015925925925925927, "loss": 0.4048, "step": 49700 }, { "epoch": 184.11, "grad_norm": 0.9294317960739136, "learning_rate": 0.0001588888888888889, "loss": 0.402, "step": 49710 }, { "epoch": 184.15, "grad_norm": 1.004883050918579, "learning_rate": 0.00015851851851851853, "loss": 0.4077, "step": 49720 }, { "epoch": 184.19, "grad_norm": 0.9653359055519104, "learning_rate": 0.00015814814814814815, "loss": 0.4204, "step": 49730 }, { "epoch": 184.22, "grad_norm": 0.9280202984809875, "learning_rate": 0.00015777777777777776, "loss": 0.4257, "step": 49740 }, { "epoch": 184.26, "grad_norm": 0.9918959736824036, "learning_rate": 0.0001574074074074074, "loss": 0.4261, "step": 49750 }, { "epoch": 184.3, "grad_norm": 1.0165759325027466, "learning_rate": 0.00015703703703703705, "loss": 0.4336, "step": 49760 }, { "epoch": 184.33, "grad_norm": 1.0142242908477783, "learning_rate": 0.0001566666666666667, "loss": 0.4342, "step": 49770 }, { "epoch": 184.37, "grad_norm": 1.0278903245925903, "learning_rate": 0.0001562962962962963, "loss": 0.4458, "step": 49780 }, { "epoch": 184.41, "grad_norm": 0.9913373589515686, "learning_rate": 0.00015592592592592592, "loss": 0.4443, "step": 49790 }, { "epoch": 184.44, "grad_norm": 1.059424877166748, "learning_rate": 0.00015555555555555556, "loss": 0.4446, "step": 49800 }, { "epoch": 184.48, "grad_norm": 1.0725332498550415, "learning_rate": 0.00015518518518518518, "loss": 0.4465, "step": 49810 }, { "epoch": 184.52, "grad_norm": 1.0381534099578857, "learning_rate": 0.00015481481481481482, "loss": 0.4481, "step": 49820 }, { "epoch": 184.56, "grad_norm": 0.9897781014442444, "learning_rate": 0.00015444444444444444, "loss": 0.4626, "step": 49830 }, { "epoch": 184.59, "grad_norm": 1.1018785238265991, "learning_rate": 0.00015407407407407408, "loss": 0.4642, "step": 49840 }, { "epoch": 184.63, "grad_norm": 1.0702872276306152, "learning_rate": 0.0001537037037037037, "loss": 0.4719, "step": 49850 }, { "epoch": 184.67, "grad_norm": 1.0390729904174805, "learning_rate": 0.00015333333333333334, "loss": 0.4525, "step": 49860 }, { "epoch": 184.7, "grad_norm": 1.0610655546188354, "learning_rate": 0.00015296296296296298, "loss": 0.4687, "step": 49870 }, { "epoch": 184.74, "grad_norm": 1.009779453277588, "learning_rate": 0.0001525925925925926, "loss": 0.4672, "step": 49880 }, { "epoch": 184.78, "grad_norm": 1.141605019569397, "learning_rate": 0.00015222222222222224, "loss": 0.4612, "step": 49890 }, { "epoch": 184.81, "grad_norm": 1.0932852029800415, "learning_rate": 0.00015185185185185185, "loss": 0.4866, "step": 49900 }, { "epoch": 184.85, "grad_norm": 1.0668141841888428, "learning_rate": 0.0001514814814814815, "loss": 0.4736, "step": 49910 }, { "epoch": 184.89, "grad_norm": 1.0443545579910278, "learning_rate": 0.0001511111111111111, "loss": 0.4682, "step": 49920 }, { "epoch": 184.93, "grad_norm": 1.1250598430633545, "learning_rate": 0.00015074074074074073, "loss": 0.4702, "step": 49930 }, { "epoch": 184.96, "grad_norm": 1.1155190467834473, "learning_rate": 0.00015037037037037037, "loss": 0.4846, "step": 49940 }, { "epoch": 185.0, "grad_norm": 2.0481882095336914, "learning_rate": 0.00015, "loss": 0.4802, "step": 49950 }, { "epoch": 185.04, "grad_norm": 0.8488351702690125, "learning_rate": 0.00014962962962962966, "loss": 0.3989, "step": 49960 }, { "epoch": 185.07, "grad_norm": 0.9094061255455017, "learning_rate": 0.00014925925925925927, "loss": 0.3977, "step": 49970 }, { "epoch": 185.11, "grad_norm": 0.9350351095199585, "learning_rate": 0.0001488888888888889, "loss": 0.4065, "step": 49980 }, { "epoch": 185.15, "grad_norm": 0.938935399055481, "learning_rate": 0.00014851851851851853, "loss": 0.4046, "step": 49990 }, { "epoch": 185.19, "grad_norm": 0.9022473096847534, "learning_rate": 0.00014814814814814815, "loss": 0.4004, "step": 50000 }, { "epoch": 185.22, "grad_norm": 0.9724648594856262, "learning_rate": 0.0001477777777777778, "loss": 0.4207, "step": 50010 }, { "epoch": 185.26, "grad_norm": 0.9791123270988464, "learning_rate": 0.0001474074074074074, "loss": 0.4104, "step": 50020 }, { "epoch": 185.3, "grad_norm": 1.002153992652893, "learning_rate": 0.00014703703703703705, "loss": 0.4349, "step": 50030 }, { "epoch": 185.33, "grad_norm": 0.9777942299842834, "learning_rate": 0.00014666666666666666, "loss": 0.4431, "step": 50040 }, { "epoch": 185.37, "grad_norm": 0.9837242364883423, "learning_rate": 0.00014629629629629628, "loss": 0.4335, "step": 50050 }, { "epoch": 185.41, "grad_norm": 1.0102285146713257, "learning_rate": 0.00014592592592592592, "loss": 0.4342, "step": 50060 }, { "epoch": 185.44, "grad_norm": 0.9805873036384583, "learning_rate": 0.00014555555555555556, "loss": 0.4377, "step": 50070 }, { "epoch": 185.48, "grad_norm": 0.9970133900642395, "learning_rate": 0.0001451851851851852, "loss": 0.4319, "step": 50080 }, { "epoch": 185.52, "grad_norm": 1.07515549659729, "learning_rate": 0.00014481481481481482, "loss": 0.4466, "step": 50090 }, { "epoch": 185.56, "grad_norm": 0.9581437110900879, "learning_rate": 0.00014444444444444444, "loss": 0.4423, "step": 50100 }, { "epoch": 185.59, "grad_norm": 1.0771484375, "learning_rate": 0.00014407407407407408, "loss": 0.4528, "step": 50110 }, { "epoch": 185.63, "grad_norm": 1.1003572940826416, "learning_rate": 0.0001437037037037037, "loss": 0.449, "step": 50120 }, { "epoch": 185.67, "grad_norm": 1.0268014669418335, "learning_rate": 0.00014333333333333334, "loss": 0.4548, "step": 50130 }, { "epoch": 185.7, "grad_norm": 1.0169177055358887, "learning_rate": 0.00014296296296296295, "loss": 0.4519, "step": 50140 }, { "epoch": 185.74, "grad_norm": 1.0599011182785034, "learning_rate": 0.0001425925925925926, "loss": 0.4671, "step": 50150 }, { "epoch": 185.78, "grad_norm": 0.9963357448577881, "learning_rate": 0.0001422222222222222, "loss": 0.4674, "step": 50160 }, { "epoch": 185.81, "grad_norm": 1.004539966583252, "learning_rate": 0.00014185185185185186, "loss": 0.4611, "step": 50170 }, { "epoch": 185.85, "grad_norm": 1.0894609689712524, "learning_rate": 0.0001414814814814815, "loss": 0.461, "step": 50180 }, { "epoch": 185.89, "grad_norm": 1.0321896076202393, "learning_rate": 0.00014111111111111111, "loss": 0.4665, "step": 50190 }, { "epoch": 185.93, "grad_norm": 1.1291207075119019, "learning_rate": 0.00014074074074074076, "loss": 0.4683, "step": 50200 }, { "epoch": 185.96, "grad_norm": 1.114660382270813, "learning_rate": 0.00014037037037037037, "loss": 0.4552, "step": 50210 }, { "epoch": 186.0, "grad_norm": 2.1208958625793457, "learning_rate": 0.00014000000000000001, "loss": 0.4891, "step": 50220 }, { "epoch": 186.04, "grad_norm": 0.8631995916366577, "learning_rate": 0.00013962962962962963, "loss": 0.383, "step": 50230 }, { "epoch": 186.07, "grad_norm": 0.8872973322868347, "learning_rate": 0.00013925925925925925, "loss": 0.3906, "step": 50240 }, { "epoch": 186.11, "grad_norm": 0.8547492027282715, "learning_rate": 0.0001388888888888889, "loss": 0.3952, "step": 50250 }, { "epoch": 186.15, "grad_norm": 0.9142735004425049, "learning_rate": 0.0001385185185185185, "loss": 0.4063, "step": 50260 }, { "epoch": 186.19, "grad_norm": 0.9310815930366516, "learning_rate": 0.00013814814814814817, "loss": 0.4043, "step": 50270 }, { "epoch": 186.22, "grad_norm": 0.9587202668190002, "learning_rate": 0.0001377777777777778, "loss": 0.4163, "step": 50280 }, { "epoch": 186.26, "grad_norm": 0.8949682116508484, "learning_rate": 0.0001374074074074074, "loss": 0.409, "step": 50290 }, { "epoch": 186.3, "grad_norm": 0.9883646368980408, "learning_rate": 0.00013703703703703705, "loss": 0.4019, "step": 50300 }, { "epoch": 186.33, "grad_norm": 1.0231529474258423, "learning_rate": 0.00013666666666666666, "loss": 0.4209, "step": 50310 }, { "epoch": 186.37, "grad_norm": 0.9546399712562561, "learning_rate": 0.0001362962962962963, "loss": 0.4255, "step": 50320 }, { "epoch": 186.41, "grad_norm": 0.9848965406417847, "learning_rate": 0.00013592592592592592, "loss": 0.415, "step": 50330 }, { "epoch": 186.44, "grad_norm": 0.9679991006851196, "learning_rate": 0.00013555555555555556, "loss": 0.428, "step": 50340 }, { "epoch": 186.48, "grad_norm": 1.0232571363449097, "learning_rate": 0.00013518518518518518, "loss": 0.4354, "step": 50350 }, { "epoch": 186.52, "grad_norm": 1.018001675605774, "learning_rate": 0.0001348148148148148, "loss": 0.4345, "step": 50360 }, { "epoch": 186.56, "grad_norm": 1.016631841659546, "learning_rate": 0.00013444444444444447, "loss": 0.4381, "step": 50370 }, { "epoch": 186.59, "grad_norm": 1.0199625492095947, "learning_rate": 0.00013407407407407408, "loss": 0.4364, "step": 50380 }, { "epoch": 186.63, "grad_norm": 1.0431368350982666, "learning_rate": 0.00013370370370370372, "loss": 0.4425, "step": 50390 }, { "epoch": 186.67, "grad_norm": 1.0619491338729858, "learning_rate": 0.00013333333333333334, "loss": 0.4425, "step": 50400 }, { "epoch": 186.7, "grad_norm": 0.9911066293716431, "learning_rate": 0.00013296296296296296, "loss": 0.4477, "step": 50410 }, { "epoch": 186.74, "grad_norm": 1.1205402612686157, "learning_rate": 0.0001325925925925926, "loss": 0.4635, "step": 50420 }, { "epoch": 186.78, "grad_norm": 1.0640531778335571, "learning_rate": 0.00013222222222222221, "loss": 0.4656, "step": 50430 }, { "epoch": 186.81, "grad_norm": 1.092454195022583, "learning_rate": 0.00013185185185185186, "loss": 0.4605, "step": 50440 }, { "epoch": 186.85, "grad_norm": 1.084949254989624, "learning_rate": 0.00013148148148148147, "loss": 0.4588, "step": 50450 }, { "epoch": 186.89, "grad_norm": 1.0481915473937988, "learning_rate": 0.00013111111111111111, "loss": 0.4642, "step": 50460 }, { "epoch": 186.93, "grad_norm": 1.0836292505264282, "learning_rate": 0.00013074074074074073, "loss": 0.4644, "step": 50470 }, { "epoch": 186.96, "grad_norm": 1.1251612901687622, "learning_rate": 0.00013037037037037037, "loss": 0.4697, "step": 50480 }, { "epoch": 187.0, "grad_norm": 1.925938606262207, "learning_rate": 0.00013000000000000002, "loss": 0.4711, "step": 50490 }, { "epoch": 187.04, "grad_norm": 0.8456882834434509, "learning_rate": 0.00012962962962962963, "loss": 0.3921, "step": 50500 }, { "epoch": 187.07, "grad_norm": 0.9210270047187805, "learning_rate": 0.00012925925925925927, "loss": 0.3783, "step": 50510 }, { "epoch": 187.11, "grad_norm": 0.8768045902252197, "learning_rate": 0.0001288888888888889, "loss": 0.3902, "step": 50520 }, { "epoch": 187.15, "grad_norm": 0.8845224380493164, "learning_rate": 0.00012851851851851853, "loss": 0.3985, "step": 50530 }, { "epoch": 187.19, "grad_norm": 0.9048132300376892, "learning_rate": 0.00012814814814814815, "loss": 0.3951, "step": 50540 }, { "epoch": 187.22, "grad_norm": 0.8952040076255798, "learning_rate": 0.00012777777777777776, "loss": 0.4099, "step": 50550 }, { "epoch": 187.26, "grad_norm": 0.962620198726654, "learning_rate": 0.0001274074074074074, "loss": 0.4141, "step": 50560 }, { "epoch": 187.3, "grad_norm": 0.9428215622901917, "learning_rate": 0.00012703703703703702, "loss": 0.4095, "step": 50570 }, { "epoch": 187.33, "grad_norm": 0.9892339110374451, "learning_rate": 0.0001266666666666667, "loss": 0.4171, "step": 50580 }, { "epoch": 187.37, "grad_norm": 1.010561466217041, "learning_rate": 0.0001262962962962963, "loss": 0.4103, "step": 50590 }, { "epoch": 187.41, "grad_norm": 0.9834034442901611, "learning_rate": 0.00012592592592592592, "loss": 0.4231, "step": 50600 }, { "epoch": 187.44, "grad_norm": 0.9447629451751709, "learning_rate": 0.00012555555555555557, "loss": 0.4162, "step": 50610 }, { "epoch": 187.48, "grad_norm": 1.0351061820983887, "learning_rate": 0.00012518518518518518, "loss": 0.4317, "step": 50620 }, { "epoch": 187.52, "grad_norm": 0.973498523235321, "learning_rate": 0.00012481481481481482, "loss": 0.4243, "step": 50630 }, { "epoch": 187.56, "grad_norm": 0.9797120094299316, "learning_rate": 0.00012444444444444444, "loss": 0.4352, "step": 50640 }, { "epoch": 187.59, "grad_norm": 0.987112283706665, "learning_rate": 0.00012407407407407408, "loss": 0.4287, "step": 50650 }, { "epoch": 187.63, "grad_norm": 1.0124151706695557, "learning_rate": 0.0001237037037037037, "loss": 0.4365, "step": 50660 }, { "epoch": 187.67, "grad_norm": 0.9849717617034912, "learning_rate": 0.00012333333333333334, "loss": 0.4454, "step": 50670 }, { "epoch": 187.7, "grad_norm": 1.003054141998291, "learning_rate": 0.00012296296296296298, "loss": 0.4325, "step": 50680 }, { "epoch": 187.74, "grad_norm": 1.047780156135559, "learning_rate": 0.0001225925925925926, "loss": 0.4407, "step": 50690 }, { "epoch": 187.78, "grad_norm": 1.075384259223938, "learning_rate": 0.00012222222222222221, "loss": 0.4403, "step": 50700 }, { "epoch": 187.81, "grad_norm": 1.0416277647018433, "learning_rate": 0.00012185185185185186, "loss": 0.4485, "step": 50710 }, { "epoch": 187.85, "grad_norm": 1.1392602920532227, "learning_rate": 0.00012148148148148149, "loss": 0.4507, "step": 50720 }, { "epoch": 187.89, "grad_norm": 1.0072075128555298, "learning_rate": 0.00012111111111111112, "loss": 0.4673, "step": 50730 }, { "epoch": 187.93, "grad_norm": 1.0627766847610474, "learning_rate": 0.00012074074074074074, "loss": 0.4486, "step": 50740 }, { "epoch": 187.96, "grad_norm": 1.0712461471557617, "learning_rate": 0.00012037037037037036, "loss": 0.4472, "step": 50750 }, { "epoch": 188.0, "grad_norm": 1.8942756652832031, "learning_rate": 0.00012, "loss": 0.4591, "step": 50760 }, { "epoch": 188.04, "grad_norm": 0.862216055393219, "learning_rate": 0.00011962962962962963, "loss": 0.3789, "step": 50770 }, { "epoch": 188.07, "grad_norm": 0.8385865688323975, "learning_rate": 0.00011925925925925926, "loss": 0.3917, "step": 50780 }, { "epoch": 188.11, "grad_norm": 0.9033856391906738, "learning_rate": 0.00011888888888888889, "loss": 0.3856, "step": 50790 }, { "epoch": 188.15, "grad_norm": 0.9251241683959961, "learning_rate": 0.00011851851851851852, "loss": 0.3865, "step": 50800 }, { "epoch": 188.19, "grad_norm": 0.8424283266067505, "learning_rate": 0.00011814814814814816, "loss": 0.3984, "step": 50810 }, { "epoch": 188.22, "grad_norm": 0.9130561351776123, "learning_rate": 0.00011777777777777778, "loss": 0.3929, "step": 50820 }, { "epoch": 188.26, "grad_norm": 0.9251916408538818, "learning_rate": 0.00011740740740740741, "loss": 0.4009, "step": 50830 }, { "epoch": 188.3, "grad_norm": 0.935077965259552, "learning_rate": 0.00011703703703703704, "loss": 0.4057, "step": 50840 }, { "epoch": 188.33, "grad_norm": 0.9387704133987427, "learning_rate": 0.00011666666666666667, "loss": 0.4097, "step": 50850 }, { "epoch": 188.37, "grad_norm": 0.970709502696991, "learning_rate": 0.00011629629629629631, "loss": 0.411, "step": 50860 }, { "epoch": 188.41, "grad_norm": 0.9760639667510986, "learning_rate": 0.00011592592592592594, "loss": 0.4105, "step": 50870 }, { "epoch": 188.44, "grad_norm": 0.9423500895500183, "learning_rate": 0.00011555555555555555, "loss": 0.4136, "step": 50880 }, { "epoch": 188.48, "grad_norm": 0.9758352041244507, "learning_rate": 0.00011518518518518518, "loss": 0.4092, "step": 50890 }, { "epoch": 188.52, "grad_norm": 1.012181043624878, "learning_rate": 0.00011481481481481481, "loss": 0.4238, "step": 50900 }, { "epoch": 188.56, "grad_norm": 1.0230450630187988, "learning_rate": 0.00011444444444444445, "loss": 0.4244, "step": 50910 }, { "epoch": 188.59, "grad_norm": 0.9596276879310608, "learning_rate": 0.00011407407407407408, "loss": 0.4173, "step": 50920 }, { "epoch": 188.63, "grad_norm": 0.9961845278739929, "learning_rate": 0.00011370370370370371, "loss": 0.4259, "step": 50930 }, { "epoch": 188.67, "grad_norm": 1.0274980068206787, "learning_rate": 0.00011333333333333333, "loss": 0.4284, "step": 50940 }, { "epoch": 188.7, "grad_norm": 1.023209571838379, "learning_rate": 0.00011296296296296296, "loss": 0.4429, "step": 50950 }, { "epoch": 188.74, "grad_norm": 1.0073031187057495, "learning_rate": 0.0001125925925925926, "loss": 0.4287, "step": 50960 }, { "epoch": 188.78, "grad_norm": 1.0359915494918823, "learning_rate": 0.00011222222222222223, "loss": 0.4405, "step": 50970 }, { "epoch": 188.81, "grad_norm": 0.9596119523048401, "learning_rate": 0.00011185185185185186, "loss": 0.4434, "step": 50980 }, { "epoch": 188.85, "grad_norm": 0.9704906344413757, "learning_rate": 0.00011148148148148149, "loss": 0.4347, "step": 50990 }, { "epoch": 188.89, "grad_norm": 1.03348970413208, "learning_rate": 0.0001111111111111111, "loss": 0.4527, "step": 51000 }, { "epoch": 188.93, "grad_norm": 1.060300588607788, "learning_rate": 0.00011074074074074075, "loss": 0.4533, "step": 51010 }, { "epoch": 188.96, "grad_norm": 1.0980509519577026, "learning_rate": 0.00011037037037037037, "loss": 0.4441, "step": 51020 }, { "epoch": 189.0, "grad_norm": 2.1591084003448486, "learning_rate": 0.00011, "loss": 0.4329, "step": 51030 }, { "epoch": 189.04, "grad_norm": 0.8770800828933716, "learning_rate": 0.00010962962962962963, "loss": 0.3757, "step": 51040 }, { "epoch": 189.07, "grad_norm": 0.8365693688392639, "learning_rate": 0.00010925925925925926, "loss": 0.3843, "step": 51050 }, { "epoch": 189.11, "grad_norm": 0.8623953461647034, "learning_rate": 0.00010888888888888888, "loss": 0.3778, "step": 51060 }, { "epoch": 189.15, "grad_norm": 0.8259143233299255, "learning_rate": 0.00010851851851851852, "loss": 0.3798, "step": 51070 }, { "epoch": 189.19, "grad_norm": 0.8634757995605469, "learning_rate": 0.00010814814814814815, "loss": 0.3778, "step": 51080 }, { "epoch": 189.22, "grad_norm": 0.8648687601089478, "learning_rate": 0.00010777777777777778, "loss": 0.3899, "step": 51090 }, { "epoch": 189.26, "grad_norm": 0.8897496461868286, "learning_rate": 0.00010740740740740741, "loss": 0.3879, "step": 51100 }, { "epoch": 189.3, "grad_norm": 0.887324333190918, "learning_rate": 0.00010703703703703704, "loss": 0.3915, "step": 51110 }, { "epoch": 189.33, "grad_norm": 0.9308483600616455, "learning_rate": 0.00010666666666666668, "loss": 0.4084, "step": 51120 }, { "epoch": 189.37, "grad_norm": 0.9871852993965149, "learning_rate": 0.0001062962962962963, "loss": 0.4119, "step": 51130 }, { "epoch": 189.41, "grad_norm": 0.9372105002403259, "learning_rate": 0.00010592592592592592, "loss": 0.4044, "step": 51140 }, { "epoch": 189.44, "grad_norm": 0.8799377083778381, "learning_rate": 0.00010555555555555555, "loss": 0.4072, "step": 51150 }, { "epoch": 189.48, "grad_norm": 0.9507384896278381, "learning_rate": 0.00010518518518518518, "loss": 0.4168, "step": 51160 }, { "epoch": 189.52, "grad_norm": 0.9813037514686584, "learning_rate": 0.00010481481481481483, "loss": 0.4126, "step": 51170 }, { "epoch": 189.56, "grad_norm": 0.9744457006454468, "learning_rate": 0.00010444444444444445, "loss": 0.4229, "step": 51180 }, { "epoch": 189.59, "grad_norm": 1.0010546445846558, "learning_rate": 0.00010407407407407407, "loss": 0.4178, "step": 51190 }, { "epoch": 189.63, "grad_norm": 1.0330417156219482, "learning_rate": 0.0001037037037037037, "loss": 0.4212, "step": 51200 }, { "epoch": 189.67, "grad_norm": 0.99884033203125, "learning_rate": 0.00010333333333333333, "loss": 0.4267, "step": 51210 }, { "epoch": 189.7, "grad_norm": 1.0357877016067505, "learning_rate": 0.00010296296296296297, "loss": 0.4199, "step": 51220 }, { "epoch": 189.74, "grad_norm": 0.9598915576934814, "learning_rate": 0.0001025925925925926, "loss": 0.4331, "step": 51230 }, { "epoch": 189.78, "grad_norm": 1.012041449546814, "learning_rate": 0.00010222222222222223, "loss": 0.4225, "step": 51240 }, { "epoch": 189.81, "grad_norm": 0.9845399260520935, "learning_rate": 0.00010185185185185185, "loss": 0.4214, "step": 51250 }, { "epoch": 189.85, "grad_norm": 0.9941114187240601, "learning_rate": 0.00010148148148148147, "loss": 0.4436, "step": 51260 }, { "epoch": 189.89, "grad_norm": 1.0230532884597778, "learning_rate": 0.00010111111111111112, "loss": 0.4325, "step": 51270 }, { "epoch": 189.93, "grad_norm": 1.022606372833252, "learning_rate": 0.00010074074074074075, "loss": 0.4266, "step": 51280 }, { "epoch": 189.96, "grad_norm": 0.9943325519561768, "learning_rate": 0.00010037037037037038, "loss": 0.4408, "step": 51290 }, { "epoch": 190.0, "grad_norm": 1.894622564315796, "learning_rate": 0.0001, "loss": 0.4405, "step": 51300 }, { "epoch": 190.04, "grad_norm": 0.7975081205368042, "learning_rate": 9.962962962962962e-05, "loss": 0.3793, "step": 51310 }, { "epoch": 190.07, "grad_norm": 0.828575611114502, "learning_rate": 9.925925925925926e-05, "loss": 0.3897, "step": 51320 }, { "epoch": 190.11, "grad_norm": 0.8125698566436768, "learning_rate": 9.888888888888889e-05, "loss": 0.3756, "step": 51330 }, { "epoch": 190.15, "grad_norm": 0.9077195525169373, "learning_rate": 9.851851851851852e-05, "loss": 0.3801, "step": 51340 }, { "epoch": 190.19, "grad_norm": 0.8815000057220459, "learning_rate": 9.814814814814815e-05, "loss": 0.3766, "step": 51350 }, { "epoch": 190.22, "grad_norm": 0.8989863395690918, "learning_rate": 9.777777777777778e-05, "loss": 0.3803, "step": 51360 }, { "epoch": 190.26, "grad_norm": 0.8354569673538208, "learning_rate": 9.740740740740741e-05, "loss": 0.3859, "step": 51370 }, { "epoch": 190.3, "grad_norm": 0.9326614141464233, "learning_rate": 9.703703703703704e-05, "loss": 0.3903, "step": 51380 }, { "epoch": 190.33, "grad_norm": 0.9138501286506653, "learning_rate": 9.666666666666667e-05, "loss": 0.3835, "step": 51390 }, { "epoch": 190.37, "grad_norm": 0.9268086552619934, "learning_rate": 9.62962962962963e-05, "loss": 0.3952, "step": 51400 }, { "epoch": 190.41, "grad_norm": 0.8911784887313843, "learning_rate": 9.592592592592593e-05, "loss": 0.3999, "step": 51410 }, { "epoch": 190.44, "grad_norm": 1.0090967416763306, "learning_rate": 9.555555555555557e-05, "loss": 0.4031, "step": 51420 }, { "epoch": 190.48, "grad_norm": 0.9447932243347168, "learning_rate": 9.51851851851852e-05, "loss": 0.4105, "step": 51430 }, { "epoch": 190.52, "grad_norm": 0.9692918658256531, "learning_rate": 9.481481481481481e-05, "loss": 0.3992, "step": 51440 }, { "epoch": 190.56, "grad_norm": 0.9399166703224182, "learning_rate": 9.444444444444444e-05, "loss": 0.4175, "step": 51450 }, { "epoch": 190.59, "grad_norm": 0.9361637830734253, "learning_rate": 9.407407407407407e-05, "loss": 0.4015, "step": 51460 }, { "epoch": 190.63, "grad_norm": 0.9695074558258057, "learning_rate": 9.370370370370371e-05, "loss": 0.4076, "step": 51470 }, { "epoch": 190.67, "grad_norm": 0.957758367061615, "learning_rate": 9.333333333333334e-05, "loss": 0.4195, "step": 51480 }, { "epoch": 190.7, "grad_norm": 0.9265300631523132, "learning_rate": 9.296296296296297e-05, "loss": 0.4211, "step": 51490 }, { "epoch": 190.74, "grad_norm": 0.9656269550323486, "learning_rate": 9.259259259259259e-05, "loss": 0.4063, "step": 51500 }, { "epoch": 190.78, "grad_norm": 1.0088058710098267, "learning_rate": 9.222222222222222e-05, "loss": 0.4265, "step": 51510 }, { "epoch": 190.81, "grad_norm": 0.989260733127594, "learning_rate": 9.185185185185186e-05, "loss": 0.4262, "step": 51520 }, { "epoch": 190.85, "grad_norm": 1.060614824295044, "learning_rate": 9.148148148148149e-05, "loss": 0.4307, "step": 51530 }, { "epoch": 190.89, "grad_norm": 1.0297032594680786, "learning_rate": 9.111111111111112e-05, "loss": 0.4281, "step": 51540 }, { "epoch": 190.93, "grad_norm": 0.989966869354248, "learning_rate": 9.074074074074075e-05, "loss": 0.4226, "step": 51550 }, { "epoch": 190.96, "grad_norm": 1.0322147607803345, "learning_rate": 9.037037037037036e-05, "loss": 0.4361, "step": 51560 }, { "epoch": 191.0, "grad_norm": 1.8464624881744385, "learning_rate": 8.999999999999999e-05, "loss": 0.4243, "step": 51570 }, { "epoch": 191.04, "grad_norm": 0.8357323408126831, "learning_rate": 8.962962962962963e-05, "loss": 0.3711, "step": 51580 }, { "epoch": 191.07, "grad_norm": 0.8935667276382446, "learning_rate": 8.925925925925926e-05, "loss": 0.3692, "step": 51590 }, { "epoch": 191.11, "grad_norm": 0.8498260378837585, "learning_rate": 8.888888888888889e-05, "loss": 0.385, "step": 51600 }, { "epoch": 191.15, "grad_norm": 0.837672770023346, "learning_rate": 8.851851851851852e-05, "loss": 0.3816, "step": 51610 }, { "epoch": 191.19, "grad_norm": 0.8953566551208496, "learning_rate": 8.814814814814814e-05, "loss": 0.3835, "step": 51620 }, { "epoch": 191.22, "grad_norm": 0.8604071736335754, "learning_rate": 8.777777777777778e-05, "loss": 0.3691, "step": 51630 }, { "epoch": 191.26, "grad_norm": 0.9045225381851196, "learning_rate": 8.740740740740741e-05, "loss": 0.3754, "step": 51640 }, { "epoch": 191.3, "grad_norm": 0.9023613333702087, "learning_rate": 8.703703703703704e-05, "loss": 0.3867, "step": 51650 }, { "epoch": 191.33, "grad_norm": 0.8950648307800293, "learning_rate": 8.666666666666667e-05, "loss": 0.3728, "step": 51660 }, { "epoch": 191.37, "grad_norm": 0.9078792929649353, "learning_rate": 8.62962962962963e-05, "loss": 0.3841, "step": 51670 }, { "epoch": 191.41, "grad_norm": 0.8773408532142639, "learning_rate": 8.592592592592593e-05, "loss": 0.4035, "step": 51680 }, { "epoch": 191.44, "grad_norm": 0.9532427191734314, "learning_rate": 8.555555555555556e-05, "loss": 0.3946, "step": 51690 }, { "epoch": 191.48, "grad_norm": 0.9748596549034119, "learning_rate": 8.518518518518518e-05, "loss": 0.3969, "step": 51700 }, { "epoch": 191.52, "grad_norm": 0.913264811038971, "learning_rate": 8.481481481481481e-05, "loss": 0.4019, "step": 51710 }, { "epoch": 191.56, "grad_norm": 0.8992056250572205, "learning_rate": 8.444444444444444e-05, "loss": 0.4035, "step": 51720 }, { "epoch": 191.59, "grad_norm": 0.9542635679244995, "learning_rate": 8.407407407407409e-05, "loss": 0.4045, "step": 51730 }, { "epoch": 191.63, "grad_norm": 1.0166466236114502, "learning_rate": 8.37037037037037e-05, "loss": 0.4131, "step": 51740 }, { "epoch": 191.67, "grad_norm": 0.9533120393753052, "learning_rate": 8.333333333333333e-05, "loss": 0.4115, "step": 51750 }, { "epoch": 191.7, "grad_norm": 1.03888738155365, "learning_rate": 8.296296296296296e-05, "loss": 0.4052, "step": 51760 }, { "epoch": 191.74, "grad_norm": 0.9723010063171387, "learning_rate": 8.259259259259259e-05, "loss": 0.4189, "step": 51770 }, { "epoch": 191.78, "grad_norm": 0.9556922912597656, "learning_rate": 8.222222222222223e-05, "loss": 0.4126, "step": 51780 }, { "epoch": 191.81, "grad_norm": 0.9536474347114563, "learning_rate": 8.185185185185186e-05, "loss": 0.4117, "step": 51790 }, { "epoch": 191.85, "grad_norm": 1.0156913995742798, "learning_rate": 8.148148148148149e-05, "loss": 0.4031, "step": 51800 }, { "epoch": 191.89, "grad_norm": 0.9756938815116882, "learning_rate": 8.11111111111111e-05, "loss": 0.4143, "step": 51810 }, { "epoch": 191.93, "grad_norm": 0.9222599864006042, "learning_rate": 8.074074074074073e-05, "loss": 0.4268, "step": 51820 }, { "epoch": 191.96, "grad_norm": 1.0029851198196411, "learning_rate": 8.037037037037038e-05, "loss": 0.4188, "step": 51830 }, { "epoch": 192.0, "grad_norm": 1.8502293825149536, "learning_rate": 8e-05, "loss": 0.422, "step": 51840 }, { "epoch": 192.04, "grad_norm": 0.8421656489372253, "learning_rate": 7.962962962962964e-05, "loss": 0.3644, "step": 51850 }, { "epoch": 192.07, "grad_norm": 0.7998185157775879, "learning_rate": 7.925925925925926e-05, "loss": 0.368, "step": 51860 }, { "epoch": 192.11, "grad_norm": 0.8072201013565063, "learning_rate": 7.888888888888888e-05, "loss": 0.3671, "step": 51870 }, { "epoch": 192.15, "grad_norm": 0.8365472555160522, "learning_rate": 7.851851851851852e-05, "loss": 0.3645, "step": 51880 }, { "epoch": 192.19, "grad_norm": 0.8714311718940735, "learning_rate": 7.814814814814815e-05, "loss": 0.3776, "step": 51890 }, { "epoch": 192.22, "grad_norm": 0.8516049385070801, "learning_rate": 7.777777777777778e-05, "loss": 0.3858, "step": 51900 }, { "epoch": 192.26, "grad_norm": 0.8652268052101135, "learning_rate": 7.740740740740741e-05, "loss": 0.387, "step": 51910 }, { "epoch": 192.3, "grad_norm": 0.8955600261688232, "learning_rate": 7.703703703703704e-05, "loss": 0.3827, "step": 51920 }, { "epoch": 192.33, "grad_norm": 0.9093692302703857, "learning_rate": 7.666666666666667e-05, "loss": 0.3742, "step": 51930 }, { "epoch": 192.37, "grad_norm": 0.8596659898757935, "learning_rate": 7.62962962962963e-05, "loss": 0.3819, "step": 51940 }, { "epoch": 192.41, "grad_norm": 0.9333339333534241, "learning_rate": 7.592592592592593e-05, "loss": 0.3825, "step": 51950 }, { "epoch": 192.44, "grad_norm": 0.960281252861023, "learning_rate": 7.555555555555556e-05, "loss": 0.3917, "step": 51960 }, { "epoch": 192.48, "grad_norm": 0.905362069606781, "learning_rate": 7.518518518518519e-05, "loss": 0.3956, "step": 51970 }, { "epoch": 192.52, "grad_norm": 0.9406252503395081, "learning_rate": 7.481481481481483e-05, "loss": 0.3814, "step": 51980 }, { "epoch": 192.56, "grad_norm": 0.8892878890037537, "learning_rate": 7.444444444444444e-05, "loss": 0.3888, "step": 51990 }, { "epoch": 192.59, "grad_norm": 0.9101101756095886, "learning_rate": 7.407407407407407e-05, "loss": 0.3983, "step": 52000 }, { "epoch": 192.63, "grad_norm": 1.0517656803131104, "learning_rate": 7.37037037037037e-05, "loss": 0.3925, "step": 52010 }, { "epoch": 192.67, "grad_norm": 0.9784597754478455, "learning_rate": 7.333333333333333e-05, "loss": 0.4013, "step": 52020 }, { "epoch": 192.7, "grad_norm": 0.9348440766334534, "learning_rate": 7.296296296296296e-05, "loss": 0.3899, "step": 52030 }, { "epoch": 192.74, "grad_norm": 0.9543972015380859, "learning_rate": 7.25925925925926e-05, "loss": 0.3974, "step": 52040 }, { "epoch": 192.78, "grad_norm": 0.97087162733078, "learning_rate": 7.222222222222222e-05, "loss": 0.4147, "step": 52050 }, { "epoch": 192.81, "grad_norm": 0.9236013889312744, "learning_rate": 7.185185185185185e-05, "loss": 0.4116, "step": 52060 }, { "epoch": 192.85, "grad_norm": 0.9513102769851685, "learning_rate": 7.148148148148148e-05, "loss": 0.4038, "step": 52070 }, { "epoch": 192.89, "grad_norm": 0.9712694883346558, "learning_rate": 7.11111111111111e-05, "loss": 0.4164, "step": 52080 }, { "epoch": 192.93, "grad_norm": 0.9384704828262329, "learning_rate": 7.074074074074075e-05, "loss": 0.4071, "step": 52090 }, { "epoch": 192.96, "grad_norm": 0.9361010789871216, "learning_rate": 7.037037037037038e-05, "loss": 0.4165, "step": 52100 }, { "epoch": 193.0, "grad_norm": 1.679486632347107, "learning_rate": 7.000000000000001e-05, "loss": 0.4205, "step": 52110 }, { "epoch": 193.04, "grad_norm": 0.812853991985321, "learning_rate": 6.962962962962962e-05, "loss": 0.3496, "step": 52120 }, { "epoch": 193.07, "grad_norm": 0.7948511838912964, "learning_rate": 6.925925925925925e-05, "loss": 0.3662, "step": 52130 }, { "epoch": 193.11, "grad_norm": 0.841355562210083, "learning_rate": 6.88888888888889e-05, "loss": 0.3572, "step": 52140 }, { "epoch": 193.15, "grad_norm": 0.8733651041984558, "learning_rate": 6.851851851851852e-05, "loss": 0.3658, "step": 52150 }, { "epoch": 193.19, "grad_norm": 0.8185422420501709, "learning_rate": 6.814814814814815e-05, "loss": 0.3748, "step": 52160 }, { "epoch": 193.22, "grad_norm": 0.8644742369651794, "learning_rate": 6.777777777777778e-05, "loss": 0.3768, "step": 52170 }, { "epoch": 193.26, "grad_norm": 0.8009930849075317, "learning_rate": 6.74074074074074e-05, "loss": 0.3775, "step": 52180 }, { "epoch": 193.3, "grad_norm": 0.8725280165672302, "learning_rate": 6.703703703703704e-05, "loss": 0.3705, "step": 52190 }, { "epoch": 193.33, "grad_norm": 0.8709507584571838, "learning_rate": 6.666666666666667e-05, "loss": 0.3712, "step": 52200 }, { "epoch": 193.37, "grad_norm": 0.8681923747062683, "learning_rate": 6.62962962962963e-05, "loss": 0.372, "step": 52210 }, { "epoch": 193.41, "grad_norm": 0.8824298977851868, "learning_rate": 6.592592592592593e-05, "loss": 0.3832, "step": 52220 }, { "epoch": 193.44, "grad_norm": 0.8672235608100891, "learning_rate": 6.555555555555556e-05, "loss": 0.3698, "step": 52230 }, { "epoch": 193.48, "grad_norm": 0.8637567162513733, "learning_rate": 6.518518518518519e-05, "loss": 0.3877, "step": 52240 }, { "epoch": 193.52, "grad_norm": 0.8623505234718323, "learning_rate": 6.481481481481482e-05, "loss": 0.3889, "step": 52250 }, { "epoch": 193.56, "grad_norm": 0.904331386089325, "learning_rate": 6.444444444444444e-05, "loss": 0.4042, "step": 52260 }, { "epoch": 193.59, "grad_norm": 0.9281720519065857, "learning_rate": 6.407407407407407e-05, "loss": 0.3965, "step": 52270 }, { "epoch": 193.63, "grad_norm": 0.9167261123657227, "learning_rate": 6.37037037037037e-05, "loss": 0.3921, "step": 52280 }, { "epoch": 193.67, "grad_norm": 0.889362633228302, "learning_rate": 6.333333333333335e-05, "loss": 0.3895, "step": 52290 }, { "epoch": 193.7, "grad_norm": 0.9352006912231445, "learning_rate": 6.296296296296296e-05, "loss": 0.3967, "step": 52300 }, { "epoch": 193.74, "grad_norm": 0.9324561953544617, "learning_rate": 6.259259259259259e-05, "loss": 0.3975, "step": 52310 }, { "epoch": 193.78, "grad_norm": 0.9033541083335876, "learning_rate": 6.222222222222222e-05, "loss": 0.3965, "step": 52320 }, { "epoch": 193.81, "grad_norm": 0.9311366677284241, "learning_rate": 6.185185185185185e-05, "loss": 0.4015, "step": 52330 }, { "epoch": 193.85, "grad_norm": 0.9424006938934326, "learning_rate": 6.148148148148149e-05, "loss": 0.3873, "step": 52340 }, { "epoch": 193.89, "grad_norm": 0.9936563968658447, "learning_rate": 6.111111111111111e-05, "loss": 0.4089, "step": 52350 }, { "epoch": 193.93, "grad_norm": 0.927615225315094, "learning_rate": 6.074074074074074e-05, "loss": 0.3986, "step": 52360 }, { "epoch": 193.96, "grad_norm": 0.9637607932090759, "learning_rate": 6.037037037037037e-05, "loss": 0.4097, "step": 52370 }, { "epoch": 194.0, "grad_norm": 1.747179388999939, "learning_rate": 6e-05, "loss": 0.4011, "step": 52380 }, { "epoch": 194.04, "grad_norm": 0.8385775089263916, "learning_rate": 5.962962962962963e-05, "loss": 0.3581, "step": 52390 }, { "epoch": 194.07, "grad_norm": 0.8298104405403137, "learning_rate": 5.925925925925926e-05, "loss": 0.3515, "step": 52400 }, { "epoch": 194.11, "grad_norm": 0.8504098653793335, "learning_rate": 5.888888888888889e-05, "loss": 0.3593, "step": 52410 }, { "epoch": 194.15, "grad_norm": 0.8432555794715881, "learning_rate": 5.851851851851852e-05, "loss": 0.3615, "step": 52420 }, { "epoch": 194.19, "grad_norm": 0.8588807582855225, "learning_rate": 5.8148148148148154e-05, "loss": 0.3665, "step": 52430 }, { "epoch": 194.22, "grad_norm": 0.8673165440559387, "learning_rate": 5.7777777777777776e-05, "loss": 0.3729, "step": 52440 }, { "epoch": 194.26, "grad_norm": 0.8235089182853699, "learning_rate": 5.7407407407407406e-05, "loss": 0.3636, "step": 52450 }, { "epoch": 194.3, "grad_norm": 0.8563822507858276, "learning_rate": 5.703703703703704e-05, "loss": 0.3736, "step": 52460 }, { "epoch": 194.33, "grad_norm": 0.881270170211792, "learning_rate": 5.6666666666666664e-05, "loss": 0.365, "step": 52470 }, { "epoch": 194.37, "grad_norm": 0.9025720357894897, "learning_rate": 5.62962962962963e-05, "loss": 0.3782, "step": 52480 }, { "epoch": 194.41, "grad_norm": 0.858967661857605, "learning_rate": 5.592592592592593e-05, "loss": 0.3863, "step": 52490 }, { "epoch": 194.44, "grad_norm": 0.8928357362747192, "learning_rate": 5.555555555555555e-05, "loss": 0.3872, "step": 52500 }, { "epoch": 194.48, "grad_norm": 0.8482757806777954, "learning_rate": 5.518518518518519e-05, "loss": 0.3785, "step": 52510 }, { "epoch": 194.52, "grad_norm": 0.930858850479126, "learning_rate": 5.4814814814814817e-05, "loss": 0.3812, "step": 52520 }, { "epoch": 194.56, "grad_norm": 0.8745943307876587, "learning_rate": 5.444444444444444e-05, "loss": 0.3792, "step": 52530 }, { "epoch": 194.59, "grad_norm": 0.9000018239021301, "learning_rate": 5.4074074074074075e-05, "loss": 0.3812, "step": 52540 }, { "epoch": 194.63, "grad_norm": 0.8568631410598755, "learning_rate": 5.3703703703703704e-05, "loss": 0.3838, "step": 52550 }, { "epoch": 194.67, "grad_norm": 0.8851279020309448, "learning_rate": 5.333333333333334e-05, "loss": 0.3793, "step": 52560 }, { "epoch": 194.7, "grad_norm": 0.9476543068885803, "learning_rate": 5.296296296296296e-05, "loss": 0.3962, "step": 52570 }, { "epoch": 194.74, "grad_norm": 0.8866179585456848, "learning_rate": 5.259259259259259e-05, "loss": 0.3878, "step": 52580 }, { "epoch": 194.78, "grad_norm": 0.8752300143241882, "learning_rate": 5.222222222222223e-05, "loss": 0.3951, "step": 52590 }, { "epoch": 194.81, "grad_norm": 0.9218076467514038, "learning_rate": 5.185185185185185e-05, "loss": 0.3894, "step": 52600 }, { "epoch": 194.85, "grad_norm": 0.960334300994873, "learning_rate": 5.1481481481481486e-05, "loss": 0.3855, "step": 52610 }, { "epoch": 194.89, "grad_norm": 0.9215881824493408, "learning_rate": 5.1111111111111115e-05, "loss": 0.3969, "step": 52620 }, { "epoch": 194.93, "grad_norm": 0.9425987005233765, "learning_rate": 5.074074074074074e-05, "loss": 0.389, "step": 52630 }, { "epoch": 194.96, "grad_norm": 0.969688355922699, "learning_rate": 5.037037037037037e-05, "loss": 0.3959, "step": 52640 }, { "epoch": 195.0, "grad_norm": 1.8812493085861206, "learning_rate": 5e-05, "loss": 0.3978, "step": 52650 }, { "epoch": 195.04, "grad_norm": 0.8659247159957886, "learning_rate": 4.962962962962963e-05, "loss": 0.3513, "step": 52660 }, { "epoch": 195.07, "grad_norm": 0.7738781571388245, "learning_rate": 4.925925925925926e-05, "loss": 0.3622, "step": 52670 }, { "epoch": 195.11, "grad_norm": 0.8059188723564148, "learning_rate": 4.888888888888889e-05, "loss": 0.363, "step": 52680 }, { "epoch": 195.15, "grad_norm": 0.7817255258560181, "learning_rate": 4.851851851851852e-05, "loss": 0.359, "step": 52690 }, { "epoch": 195.19, "grad_norm": 0.8181976079940796, "learning_rate": 4.814814814814815e-05, "loss": 0.3601, "step": 52700 }, { "epoch": 195.22, "grad_norm": 0.8579437732696533, "learning_rate": 4.7777777777777784e-05, "loss": 0.3626, "step": 52710 }, { "epoch": 195.26, "grad_norm": 0.8376036882400513, "learning_rate": 4.7407407407407407e-05, "loss": 0.3516, "step": 52720 }, { "epoch": 195.3, "grad_norm": 0.8274129033088684, "learning_rate": 4.7037037037037036e-05, "loss": 0.3735, "step": 52730 }, { "epoch": 195.33, "grad_norm": 0.8801832795143127, "learning_rate": 4.666666666666667e-05, "loss": 0.3638, "step": 52740 }, { "epoch": 195.37, "grad_norm": 0.8347715139389038, "learning_rate": 4.6296296296296294e-05, "loss": 0.3796, "step": 52750 }, { "epoch": 195.41, "grad_norm": 0.8824911117553711, "learning_rate": 4.592592592592593e-05, "loss": 0.3704, "step": 52760 }, { "epoch": 195.44, "grad_norm": 0.8434776067733765, "learning_rate": 4.555555555555556e-05, "loss": 0.3651, "step": 52770 }, { "epoch": 195.48, "grad_norm": 0.8641818761825562, "learning_rate": 4.518518518518518e-05, "loss": 0.3677, "step": 52780 }, { "epoch": 195.52, "grad_norm": 0.909089982509613, "learning_rate": 4.481481481481482e-05, "loss": 0.3743, "step": 52790 }, { "epoch": 195.56, "grad_norm": 0.8464562892913818, "learning_rate": 4.4444444444444447e-05, "loss": 0.3863, "step": 52800 }, { "epoch": 195.59, "grad_norm": 0.9092419147491455, "learning_rate": 4.407407407407407e-05, "loss": 0.3722, "step": 52810 }, { "epoch": 195.63, "grad_norm": 0.8889215588569641, "learning_rate": 4.3703703703703705e-05, "loss": 0.3844, "step": 52820 }, { "epoch": 195.67, "grad_norm": 0.8936458230018616, "learning_rate": 4.3333333333333334e-05, "loss": 0.3887, "step": 52830 }, { "epoch": 195.7, "grad_norm": 0.9093278646469116, "learning_rate": 4.296296296296296e-05, "loss": 0.3724, "step": 52840 }, { "epoch": 195.74, "grad_norm": 0.8708871006965637, "learning_rate": 4.259259259259259e-05, "loss": 0.3799, "step": 52850 }, { "epoch": 195.78, "grad_norm": 0.9522839784622192, "learning_rate": 4.222222222222222e-05, "loss": 0.3799, "step": 52860 }, { "epoch": 195.81, "grad_norm": 0.8670366406440735, "learning_rate": 4.185185185185185e-05, "loss": 0.3832, "step": 52870 }, { "epoch": 195.85, "grad_norm": 0.8927628993988037, "learning_rate": 4.148148148148148e-05, "loss": 0.3794, "step": 52880 }, { "epoch": 195.89, "grad_norm": 0.8557700514793396, "learning_rate": 4.1111111111111116e-05, "loss": 0.3816, "step": 52890 }, { "epoch": 195.93, "grad_norm": 0.8442705869674683, "learning_rate": 4.0740740740740745e-05, "loss": 0.3874, "step": 52900 }, { "epoch": 195.96, "grad_norm": 0.9126988649368286, "learning_rate": 4.037037037037037e-05, "loss": 0.3942, "step": 52910 }, { "epoch": 196.0, "grad_norm": 1.4788200855255127, "learning_rate": 4e-05, "loss": 0.3881, "step": 52920 }, { "epoch": 196.04, "grad_norm": 0.8152386546134949, "learning_rate": 3.962962962962963e-05, "loss": 0.3601, "step": 52930 }, { "epoch": 196.07, "grad_norm": 0.7856703996658325, "learning_rate": 3.925925925925926e-05, "loss": 0.3639, "step": 52940 }, { "epoch": 196.11, "grad_norm": 0.8162696957588196, "learning_rate": 3.888888888888889e-05, "loss": 0.3462, "step": 52950 }, { "epoch": 196.15, "grad_norm": 0.7858487963676453, "learning_rate": 3.851851851851852e-05, "loss": 0.3605, "step": 52960 }, { "epoch": 196.19, "grad_norm": 0.8109729290008545, "learning_rate": 3.814814814814815e-05, "loss": 0.3607, "step": 52970 }, { "epoch": 196.22, "grad_norm": 0.8202926516532898, "learning_rate": 3.777777777777778e-05, "loss": 0.368, "step": 52980 }, { "epoch": 196.26, "grad_norm": 0.7958359122276306, "learning_rate": 3.7407407407407414e-05, "loss": 0.3602, "step": 52990 }, { "epoch": 196.3, "grad_norm": 0.7976880669593811, "learning_rate": 3.7037037037037037e-05, "loss": 0.3616, "step": 53000 }, { "epoch": 196.33, "grad_norm": 0.8153302073478699, "learning_rate": 3.6666666666666666e-05, "loss": 0.3633, "step": 53010 }, { "epoch": 196.37, "grad_norm": 0.8204575777053833, "learning_rate": 3.62962962962963e-05, "loss": 0.3551, "step": 53020 }, { "epoch": 196.41, "grad_norm": 0.8472855687141418, "learning_rate": 3.5925925925925924e-05, "loss": 0.3595, "step": 53030 }, { "epoch": 196.44, "grad_norm": 0.7960186004638672, "learning_rate": 3.555555555555555e-05, "loss": 0.3706, "step": 53040 }, { "epoch": 196.48, "grad_norm": 0.8351572751998901, "learning_rate": 3.518518518518519e-05, "loss": 0.3705, "step": 53050 }, { "epoch": 196.52, "grad_norm": 0.8574457764625549, "learning_rate": 3.481481481481481e-05, "loss": 0.3612, "step": 53060 }, { "epoch": 196.56, "grad_norm": 0.8318818211555481, "learning_rate": 3.444444444444445e-05, "loss": 0.3752, "step": 53070 }, { "epoch": 196.59, "grad_norm": 0.8394012451171875, "learning_rate": 3.4074074074074077e-05, "loss": 0.3672, "step": 53080 }, { "epoch": 196.63, "grad_norm": 0.8550794124603271, "learning_rate": 3.37037037037037e-05, "loss": 0.3623, "step": 53090 }, { "epoch": 196.67, "grad_norm": 0.8827661275863647, "learning_rate": 3.3333333333333335e-05, "loss": 0.371, "step": 53100 }, { "epoch": 196.7, "grad_norm": 0.8500509262084961, "learning_rate": 3.2962962962962964e-05, "loss": 0.3822, "step": 53110 }, { "epoch": 196.74, "grad_norm": 0.8798251748085022, "learning_rate": 3.259259259259259e-05, "loss": 0.3791, "step": 53120 }, { "epoch": 196.78, "grad_norm": 0.8392333388328552, "learning_rate": 3.222222222222222e-05, "loss": 0.368, "step": 53130 }, { "epoch": 196.81, "grad_norm": 0.8432697653770447, "learning_rate": 3.185185185185185e-05, "loss": 0.3705, "step": 53140 }, { "epoch": 196.85, "grad_norm": 0.8685855269432068, "learning_rate": 3.148148148148148e-05, "loss": 0.3806, "step": 53150 }, { "epoch": 196.89, "grad_norm": 0.8553511500358582, "learning_rate": 3.111111111111111e-05, "loss": 0.3804, "step": 53160 }, { "epoch": 196.93, "grad_norm": 0.8809592127799988, "learning_rate": 3.0740740740740746e-05, "loss": 0.379, "step": 53170 }, { "epoch": 196.96, "grad_norm": 0.899813711643219, "learning_rate": 3.037037037037037e-05, "loss": 0.3809, "step": 53180 }, { "epoch": 197.0, "grad_norm": 1.6756294965744019, "learning_rate": 3e-05, "loss": 0.3679, "step": 53190 }, { "epoch": 197.04, "grad_norm": 0.8270112872123718, "learning_rate": 2.962962962962963e-05, "loss": 0.3528, "step": 53200 }, { "epoch": 197.07, "grad_norm": 0.7883965969085693, "learning_rate": 2.925925925925926e-05, "loss": 0.3525, "step": 53210 }, { "epoch": 197.11, "grad_norm": 0.8234068155288696, "learning_rate": 2.8888888888888888e-05, "loss": 0.3484, "step": 53220 }, { "epoch": 197.15, "grad_norm": 0.7610680460929871, "learning_rate": 2.851851851851852e-05, "loss": 0.358, "step": 53230 }, { "epoch": 197.19, "grad_norm": 0.8109197616577148, "learning_rate": 2.814814814814815e-05, "loss": 0.3538, "step": 53240 }, { "epoch": 197.22, "grad_norm": 0.8277642130851746, "learning_rate": 2.7777777777777776e-05, "loss": 0.3544, "step": 53250 }, { "epoch": 197.26, "grad_norm": 0.8044574856758118, "learning_rate": 2.7407407407407408e-05, "loss": 0.3626, "step": 53260 }, { "epoch": 197.3, "grad_norm": 0.7997308373451233, "learning_rate": 2.7037037037037037e-05, "loss": 0.3608, "step": 53270 }, { "epoch": 197.33, "grad_norm": 0.8243195414543152, "learning_rate": 2.666666666666667e-05, "loss": 0.3591, "step": 53280 }, { "epoch": 197.37, "grad_norm": 0.8348344564437866, "learning_rate": 2.6296296296296296e-05, "loss": 0.3582, "step": 53290 }, { "epoch": 197.41, "grad_norm": 0.8257560729980469, "learning_rate": 2.5925925925925925e-05, "loss": 0.3658, "step": 53300 }, { "epoch": 197.44, "grad_norm": 0.8056928515434265, "learning_rate": 2.5555555555555557e-05, "loss": 0.3671, "step": 53310 }, { "epoch": 197.48, "grad_norm": 0.8453297019004822, "learning_rate": 2.5185185185185187e-05, "loss": 0.3618, "step": 53320 }, { "epoch": 197.52, "grad_norm": 0.8153665065765381, "learning_rate": 2.4814814814814816e-05, "loss": 0.3614, "step": 53330 }, { "epoch": 197.56, "grad_norm": 0.8234355449676514, "learning_rate": 2.4444444444444445e-05, "loss": 0.3604, "step": 53340 }, { "epoch": 197.59, "grad_norm": 0.8666888475418091, "learning_rate": 2.4074074074074074e-05, "loss": 0.3733, "step": 53350 }, { "epoch": 197.63, "grad_norm": 0.843656599521637, "learning_rate": 2.3703703703703703e-05, "loss": 0.3698, "step": 53360 }, { "epoch": 197.67, "grad_norm": 0.8797212243080139, "learning_rate": 2.3333333333333336e-05, "loss": 0.3657, "step": 53370 }, { "epoch": 197.7, "grad_norm": 0.8750057816505432, "learning_rate": 2.2962962962962965e-05, "loss": 0.3562, "step": 53380 }, { "epoch": 197.74, "grad_norm": 0.8547198176383972, "learning_rate": 2.259259259259259e-05, "loss": 0.3728, "step": 53390 }, { "epoch": 197.78, "grad_norm": 0.8509737849235535, "learning_rate": 2.2222222222222223e-05, "loss": 0.3611, "step": 53400 }, { "epoch": 197.81, "grad_norm": 0.8830751776695251, "learning_rate": 2.1851851851851852e-05, "loss": 0.3614, "step": 53410 }, { "epoch": 197.85, "grad_norm": 0.8684916496276855, "learning_rate": 2.148148148148148e-05, "loss": 0.3631, "step": 53420 }, { "epoch": 197.89, "grad_norm": 0.8446648120880127, "learning_rate": 2.111111111111111e-05, "loss": 0.3635, "step": 53430 }, { "epoch": 197.93, "grad_norm": 0.8646011352539062, "learning_rate": 2.074074074074074e-05, "loss": 0.3771, "step": 53440 }, { "epoch": 197.96, "grad_norm": 0.8776282072067261, "learning_rate": 2.0370370370370372e-05, "loss": 0.3721, "step": 53450 }, { "epoch": 198.0, "grad_norm": 1.4938170909881592, "learning_rate": 2e-05, "loss": 0.3705, "step": 53460 }, { "epoch": 198.04, "grad_norm": 0.7895182371139526, "learning_rate": 1.962962962962963e-05, "loss": 0.3582, "step": 53470 }, { "epoch": 198.07, "grad_norm": 0.7492555379867554, "learning_rate": 1.925925925925926e-05, "loss": 0.359, "step": 53480 }, { "epoch": 198.11, "grad_norm": 0.7928748726844788, "learning_rate": 1.888888888888889e-05, "loss": 0.3429, "step": 53490 }, { "epoch": 198.15, "grad_norm": 0.7984386682510376, "learning_rate": 1.8518518518518518e-05, "loss": 0.3529, "step": 53500 }, { "epoch": 198.19, "grad_norm": 0.8060497641563416, "learning_rate": 1.814814814814815e-05, "loss": 0.3463, "step": 53510 }, { "epoch": 198.22, "grad_norm": 0.82963627576828, "learning_rate": 1.7777777777777777e-05, "loss": 0.3447, "step": 53520 }, { "epoch": 198.26, "grad_norm": 0.8006116151809692, "learning_rate": 1.7407407407407406e-05, "loss": 0.3503, "step": 53530 }, { "epoch": 198.3, "grad_norm": 0.8024885654449463, "learning_rate": 1.7037037037037038e-05, "loss": 0.3566, "step": 53540 }, { "epoch": 198.33, "grad_norm": 0.8089790344238281, "learning_rate": 1.6666666666666667e-05, "loss": 0.3642, "step": 53550 }, { "epoch": 198.37, "grad_norm": 0.7876638174057007, "learning_rate": 1.6296296296296297e-05, "loss": 0.3563, "step": 53560 }, { "epoch": 198.41, "grad_norm": 0.8106840252876282, "learning_rate": 1.5925925925925926e-05, "loss": 0.3584, "step": 53570 }, { "epoch": 198.44, "grad_norm": 0.7648861408233643, "learning_rate": 1.5555555555555555e-05, "loss": 0.3572, "step": 53580 }, { "epoch": 198.48, "grad_norm": 0.8709505796432495, "learning_rate": 1.5185185185185186e-05, "loss": 0.3603, "step": 53590 }, { "epoch": 198.52, "grad_norm": 0.8210094571113586, "learning_rate": 1.4814814814814815e-05, "loss": 0.352, "step": 53600 }, { "epoch": 198.56, "grad_norm": 0.8383370041847229, "learning_rate": 1.4444444444444444e-05, "loss": 0.3518, "step": 53610 }, { "epoch": 198.59, "grad_norm": 0.8203353881835938, "learning_rate": 1.4074074074074075e-05, "loss": 0.3533, "step": 53620 }, { "epoch": 198.63, "grad_norm": 0.8129621744155884, "learning_rate": 1.3703703703703704e-05, "loss": 0.3533, "step": 53630 }, { "epoch": 198.67, "grad_norm": 0.8171769380569458, "learning_rate": 1.3333333333333335e-05, "loss": 0.3596, "step": 53640 }, { "epoch": 198.7, "grad_norm": 0.8063119053840637, "learning_rate": 1.2962962962962962e-05, "loss": 0.3718, "step": 53650 }, { "epoch": 198.74, "grad_norm": 0.8649743795394897, "learning_rate": 1.2592592592592593e-05, "loss": 0.3586, "step": 53660 }, { "epoch": 198.78, "grad_norm": 0.8293347954750061, "learning_rate": 1.2222222222222222e-05, "loss": 0.3656, "step": 53670 }, { "epoch": 198.81, "grad_norm": 0.8939828872680664, "learning_rate": 1.1851851851851852e-05, "loss": 0.3542, "step": 53680 }, { "epoch": 198.85, "grad_norm": 0.8524936437606812, "learning_rate": 1.1481481481481482e-05, "loss": 0.3573, "step": 53690 }, { "epoch": 198.89, "grad_norm": 0.8106054067611694, "learning_rate": 1.1111111111111112e-05, "loss": 0.3635, "step": 53700 }, { "epoch": 198.93, "grad_norm": 0.7986220717430115, "learning_rate": 1.074074074074074e-05, "loss": 0.3605, "step": 53710 }, { "epoch": 198.96, "grad_norm": 0.8215298056602478, "learning_rate": 1.037037037037037e-05, "loss": 0.3688, "step": 53720 }, { "epoch": 199.0, "grad_norm": 1.5070199966430664, "learning_rate": 1e-05, "loss": 0.3647, "step": 53730 }, { "epoch": 199.04, "grad_norm": 0.8176571130752563, "learning_rate": 9.62962962962963e-06, "loss": 0.3479, "step": 53740 }, { "epoch": 199.07, "grad_norm": 0.7820712327957153, "learning_rate": 9.259259259259259e-06, "loss": 0.3476, "step": 53750 }, { "epoch": 199.11, "grad_norm": 0.7757356762886047, "learning_rate": 8.888888888888888e-06, "loss": 0.3408, "step": 53760 }, { "epoch": 199.15, "grad_norm": 0.8038589954376221, "learning_rate": 8.518518518518519e-06, "loss": 0.3503, "step": 53770 }, { "epoch": 199.19, "grad_norm": 0.8047194480895996, "learning_rate": 8.148148148148148e-06, "loss": 0.3551, "step": 53780 }, { "epoch": 199.22, "grad_norm": 0.7661307454109192, "learning_rate": 7.777777777777777e-06, "loss": 0.3497, "step": 53790 }, { "epoch": 199.26, "grad_norm": 0.7506800889968872, "learning_rate": 7.4074074074074075e-06, "loss": 0.3555, "step": 53800 }, { "epoch": 199.3, "grad_norm": 0.7798722386360168, "learning_rate": 7.0370370370370375e-06, "loss": 0.3453, "step": 53810 }, { "epoch": 199.33, "grad_norm": 0.7841687202453613, "learning_rate": 6.6666666666666675e-06, "loss": 0.3571, "step": 53820 }, { "epoch": 199.37, "grad_norm": 0.817516565322876, "learning_rate": 6.296296296296297e-06, "loss": 0.3488, "step": 53830 }, { "epoch": 199.41, "grad_norm": 0.8079009652137756, "learning_rate": 5.925925925925926e-06, "loss": 0.3561, "step": 53840 }, { "epoch": 199.44, "grad_norm": 0.8136209845542908, "learning_rate": 5.555555555555556e-06, "loss": 0.3479, "step": 53850 }, { "epoch": 199.48, "grad_norm": 0.8167023658752441, "learning_rate": 5.185185185185185e-06, "loss": 0.3494, "step": 53860 }, { "epoch": 199.52, "grad_norm": 0.7994495630264282, "learning_rate": 4.814814814814815e-06, "loss": 0.3535, "step": 53870 }, { "epoch": 199.56, "grad_norm": 0.822665810585022, "learning_rate": 4.444444444444444e-06, "loss": 0.3514, "step": 53880 }, { "epoch": 199.59, "grad_norm": 0.808405339717865, "learning_rate": 4.074074074074074e-06, "loss": 0.3577, "step": 53890 }, { "epoch": 199.63, "grad_norm": 0.7848332524299622, "learning_rate": 3.7037037037037037e-06, "loss": 0.3524, "step": 53900 }, { "epoch": 199.67, "grad_norm": 0.8101984262466431, "learning_rate": 3.3333333333333337e-06, "loss": 0.3549, "step": 53910 }, { "epoch": 199.7, "grad_norm": 0.7706851959228516, "learning_rate": 2.962962962962963e-06, "loss": 0.3526, "step": 53920 }, { "epoch": 199.74, "grad_norm": 0.8184105753898621, "learning_rate": 2.5925925925925925e-06, "loss": 0.3487, "step": 53930 }, { "epoch": 199.78, "grad_norm": 0.8454097509384155, "learning_rate": 2.222222222222222e-06, "loss": 0.3529, "step": 53940 }, { "epoch": 199.81, "grad_norm": 0.7991722226142883, "learning_rate": 1.8518518518518519e-06, "loss": 0.3572, "step": 53950 }, { "epoch": 199.85, "grad_norm": 0.8001537919044495, "learning_rate": 1.4814814814814815e-06, "loss": 0.3614, "step": 53960 }, { "epoch": 199.89, "grad_norm": 0.7946948409080505, "learning_rate": 1.111111111111111e-06, "loss": 0.3427, "step": 53970 }, { "epoch": 199.93, "grad_norm": 0.8384917974472046, "learning_rate": 7.407407407407407e-07, "loss": 0.3575, "step": 53980 }, { "epoch": 199.96, "grad_norm": 0.7977772355079651, "learning_rate": 3.7037037037037036e-07, "loss": 0.3586, "step": 53990 }, { "epoch": 200.0, "grad_norm": 1.766687035560608, "learning_rate": 0.0, "loss": 0.3543, "step": 54000 } ], "logging_steps": 10, "max_steps": 54000, "num_input_tokens_seen": 0, "num_train_epochs": 200, "save_steps": 500, "total_flos": 6.764336504975697e+18, "train_batch_size": 16, "trial_name": null, "trial_params": null }