| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.990024577128813, |
| "eval_steps": 500, |
| "global_step": 648, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.004626283070695388, |
| "grad_norm": 0.6155030131340027, |
| "learning_rate": 9.98456790123457e-06, |
| "loss": 17.5936, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.009252566141390776, |
| "grad_norm": 0.6313583254814148, |
| "learning_rate": 9.969135802469136e-06, |
| "loss": 21.7773, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.013878849212086165, |
| "grad_norm": 0.4570765793323517, |
| "learning_rate": 9.953703703703704e-06, |
| "loss": 17.1664, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.018505132282781552, |
| "grad_norm": 0.3836732506752014, |
| "learning_rate": 9.938271604938273e-06, |
| "loss": 18.3922, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.02313141535347694, |
| "grad_norm": 0.3599857985973358, |
| "learning_rate": 9.92283950617284e-06, |
| "loss": 16.9993, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.02775769842417233, |
| "grad_norm": 0.4342029094696045, |
| "learning_rate": 9.907407407407408e-06, |
| "loss": 17.4883, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.032383981494867715, |
| "grad_norm": 0.40308186411857605, |
| "learning_rate": 9.891975308641975e-06, |
| "loss": 18.2562, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.037010264565563104, |
| "grad_norm": 0.39312881231307983, |
| "learning_rate": 9.876543209876543e-06, |
| "loss": 17.9318, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.04163654763625849, |
| "grad_norm": 0.6749095916748047, |
| "learning_rate": 9.861111111111112e-06, |
| "loss": 18.5244, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.04626283070695388, |
| "grad_norm": 0.3622874617576599, |
| "learning_rate": 9.84567901234568e-06, |
| "loss": 16.4386, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.05088911377764927, |
| "grad_norm": 0.24345389008522034, |
| "learning_rate": 9.830246913580249e-06, |
| "loss": 17.5414, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.05551539684834466, |
| "grad_norm": 0.34016355872154236, |
| "learning_rate": 9.814814814814815e-06, |
| "loss": 16.9405, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.06014167991904005, |
| "grad_norm": 0.25097739696502686, |
| "learning_rate": 9.799382716049384e-06, |
| "loss": 14.4692, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.06476796298973543, |
| "grad_norm": 0.2524467408657074, |
| "learning_rate": 9.783950617283952e-06, |
| "loss": 16.2185, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.06939424606043082, |
| "grad_norm": 0.3705744743347168, |
| "learning_rate": 9.768518518518519e-06, |
| "loss": 16.3573, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.07402052913112621, |
| "grad_norm": 0.2848168909549713, |
| "learning_rate": 9.753086419753087e-06, |
| "loss": 16.331, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.0786468122018216, |
| "grad_norm": 0.300182044506073, |
| "learning_rate": 9.737654320987654e-06, |
| "loss": 15.6623, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.08327309527251699, |
| "grad_norm": 0.36958807706832886, |
| "learning_rate": 9.722222222222223e-06, |
| "loss": 18.5857, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.08789937834321238, |
| "grad_norm": 0.3009611666202545, |
| "learning_rate": 9.706790123456791e-06, |
| "loss": 17.0655, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.09252566141390776, |
| "grad_norm": 0.24450023472309113, |
| "learning_rate": 9.691358024691358e-06, |
| "loss": 15.3945, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.09715194448460315, |
| "grad_norm": 0.3174358010292053, |
| "learning_rate": 9.675925925925926e-06, |
| "loss": 15.3045, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.10177822755529854, |
| "grad_norm": 0.34466007351875305, |
| "learning_rate": 9.660493827160495e-06, |
| "loss": 16.7186, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.10640451062599393, |
| "grad_norm": 0.294209361076355, |
| "learning_rate": 9.645061728395062e-06, |
| "loss": 16.2893, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.11103079369668932, |
| "grad_norm": 0.3418446183204651, |
| "learning_rate": 9.62962962962963e-06, |
| "loss": 16.1827, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.11565707676738471, |
| "grad_norm": 0.2584611177444458, |
| "learning_rate": 9.614197530864198e-06, |
| "loss": 15.7934, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.1202833598380801, |
| "grad_norm": 0.3344869911670685, |
| "learning_rate": 9.598765432098767e-06, |
| "loss": 14.2338, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.12490964290877549, |
| "grad_norm": 0.28123536705970764, |
| "learning_rate": 9.583333333333335e-06, |
| "loss": 16.2229, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.12953592597947086, |
| "grad_norm": 0.22181656956672668, |
| "learning_rate": 9.567901234567902e-06, |
| "loss": 13.7744, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.13416220905016626, |
| "grad_norm": 0.45139142870903015, |
| "learning_rate": 9.55246913580247e-06, |
| "loss": 15.4782, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.13878849212086164, |
| "grad_norm": 0.21577508747577667, |
| "learning_rate": 9.537037037037037e-06, |
| "loss": 15.6752, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.14341477519155704, |
| "grad_norm": 0.29512590169906616, |
| "learning_rate": 9.521604938271606e-06, |
| "loss": 14.7965, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.14804105826225242, |
| "grad_norm": 0.28356751799583435, |
| "learning_rate": 9.506172839506174e-06, |
| "loss": 14.9097, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.15266734133294782, |
| "grad_norm": 0.25781649351119995, |
| "learning_rate": 9.490740740740741e-06, |
| "loss": 14.494, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.1572936244036432, |
| "grad_norm": 0.1967274248600006, |
| "learning_rate": 9.47530864197531e-06, |
| "loss": 15.0075, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.1619199074743386, |
| "grad_norm": 0.42894116044044495, |
| "learning_rate": 9.459876543209878e-06, |
| "loss": 14.8055, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.16654619054503397, |
| "grad_norm": 0.2417590320110321, |
| "learning_rate": 9.444444444444445e-06, |
| "loss": 14.7145, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.17117247361572938, |
| "grad_norm": 0.15463685989379883, |
| "learning_rate": 9.429012345679013e-06, |
| "loss": 14.4273, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.17579875668642475, |
| "grad_norm": 0.21183425188064575, |
| "learning_rate": 9.413580246913581e-06, |
| "loss": 16.6545, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.18042503975712013, |
| "grad_norm": 0.41926464438438416, |
| "learning_rate": 9.398148148148148e-06, |
| "loss": 13.4609, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.18505132282781553, |
| "grad_norm": 0.21585555374622345, |
| "learning_rate": 9.382716049382717e-06, |
| "loss": 14.1515, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.1896776058985109, |
| "grad_norm": 0.28051456809043884, |
| "learning_rate": 9.367283950617285e-06, |
| "loss": 14.0035, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.1943038889692063, |
| "grad_norm": 0.18281345069408417, |
| "learning_rate": 9.351851851851854e-06, |
| "loss": 14.6411, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.19893017203990168, |
| "grad_norm": 0.1747712790966034, |
| "learning_rate": 9.33641975308642e-06, |
| "loss": 14.4986, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.20355645511059708, |
| "grad_norm": 0.2613980174064636, |
| "learning_rate": 9.320987654320989e-06, |
| "loss": 14.2572, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.20818273818129246, |
| "grad_norm": 0.19127604365348816, |
| "learning_rate": 9.305555555555557e-06, |
| "loss": 14.3084, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.21280902125198786, |
| "grad_norm": 0.24545663595199585, |
| "learning_rate": 9.290123456790124e-06, |
| "loss": 13.3381, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.21743530432268324, |
| "grad_norm": 0.25891631841659546, |
| "learning_rate": 9.274691358024692e-06, |
| "loss": 13.1076, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.22206158739337864, |
| "grad_norm": 0.23256707191467285, |
| "learning_rate": 9.25925925925926e-06, |
| "loss": 13.3061, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.22668787046407401, |
| "grad_norm": 0.2427562177181244, |
| "learning_rate": 9.243827160493828e-06, |
| "loss": 16.0646, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.23131415353476942, |
| "grad_norm": 0.22760462760925293, |
| "learning_rate": 9.228395061728396e-06, |
| "loss": 12.937, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.2359404366054648, |
| "grad_norm": 0.27145224809646606, |
| "learning_rate": 9.212962962962963e-06, |
| "loss": 13.8299, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.2405667196761602, |
| "grad_norm": 0.3120661973953247, |
| "learning_rate": 9.197530864197531e-06, |
| "loss": 14.074, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.24519300274685557, |
| "grad_norm": 0.21604514122009277, |
| "learning_rate": 9.1820987654321e-06, |
| "loss": 13.6731, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.24981928581755097, |
| "grad_norm": 0.22970032691955566, |
| "learning_rate": 9.166666666666666e-06, |
| "loss": 14.0123, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.2544455688882464, |
| "grad_norm": 0.24079963564872742, |
| "learning_rate": 9.151234567901235e-06, |
| "loss": 13.5458, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.2590718519589417, |
| "grad_norm": 0.21451319754123688, |
| "learning_rate": 9.135802469135803e-06, |
| "loss": 12.5504, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.2636981350296371, |
| "grad_norm": 0.2508305609226227, |
| "learning_rate": 9.120370370370372e-06, |
| "loss": 14.8642, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.26832441810033253, |
| "grad_norm": 0.24022799730300903, |
| "learning_rate": 9.10493827160494e-06, |
| "loss": 15.1011, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.2729507011710279, |
| "grad_norm": 0.3794403076171875, |
| "learning_rate": 9.089506172839507e-06, |
| "loss": 14.3993, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.2775769842417233, |
| "grad_norm": 0.28563258051872253, |
| "learning_rate": 9.074074074074075e-06, |
| "loss": 14.294, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.2822032673124187, |
| "grad_norm": 0.28061386942863464, |
| "learning_rate": 9.058641975308642e-06, |
| "loss": 13.033, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.2868295503831141, |
| "grad_norm": 0.3351198136806488, |
| "learning_rate": 9.04320987654321e-06, |
| "loss": 14.5075, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.29145583345380943, |
| "grad_norm": 0.23749324679374695, |
| "learning_rate": 9.027777777777779e-06, |
| "loss": 13.7952, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.29608211652450483, |
| "grad_norm": 0.16919538378715515, |
| "learning_rate": 9.012345679012346e-06, |
| "loss": 14.7146, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.30070839959520024, |
| "grad_norm": 0.18566973507404327, |
| "learning_rate": 8.996913580246914e-06, |
| "loss": 13.9967, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.30533468266589564, |
| "grad_norm": 0.25144919753074646, |
| "learning_rate": 8.981481481481483e-06, |
| "loss": 13.0544, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.309960965736591, |
| "grad_norm": 0.1971070021390915, |
| "learning_rate": 8.96604938271605e-06, |
| "loss": 12.7932, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.3145872488072864, |
| "grad_norm": 0.2236046940088272, |
| "learning_rate": 8.950617283950618e-06, |
| "loss": 14.4338, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.3192135318779818, |
| "grad_norm": 0.20062977075576782, |
| "learning_rate": 8.935185185185186e-06, |
| "loss": 13.6965, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.3238398149486772, |
| "grad_norm": 0.3102545738220215, |
| "learning_rate": 8.919753086419753e-06, |
| "loss": 14.174, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.32846609801937254, |
| "grad_norm": 0.282172828912735, |
| "learning_rate": 8.904320987654322e-06, |
| "loss": 14.012, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.33309238109006795, |
| "grad_norm": 0.31374603509902954, |
| "learning_rate": 8.888888888888888e-06, |
| "loss": 12.9434, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.33771866416076335, |
| "grad_norm": 0.24876756966114044, |
| "learning_rate": 8.873456790123458e-06, |
| "loss": 13.0801, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.34234494723145875, |
| "grad_norm": 0.2828700840473175, |
| "learning_rate": 8.858024691358025e-06, |
| "loss": 12.3676, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.3469712303021541, |
| "grad_norm": 0.23395580053329468, |
| "learning_rate": 8.842592592592594e-06, |
| "loss": 12.8463, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.3515975133728495, |
| "grad_norm": 0.26694121956825256, |
| "learning_rate": 8.827160493827162e-06, |
| "loss": 11.5822, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.3562237964435449, |
| "grad_norm": 0.2129782736301422, |
| "learning_rate": 8.811728395061729e-06, |
| "loss": 12.6834, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.36085007951424025, |
| "grad_norm": 0.20400603115558624, |
| "learning_rate": 8.796296296296297e-06, |
| "loss": 14.2065, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.36547636258493565, |
| "grad_norm": 0.45560839772224426, |
| "learning_rate": 8.780864197530866e-06, |
| "loss": 13.1033, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.37010264565563106, |
| "grad_norm": 0.26329922676086426, |
| "learning_rate": 8.765432098765432e-06, |
| "loss": 11.9119, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.37472892872632646, |
| "grad_norm": 0.21817731857299805, |
| "learning_rate": 8.750000000000001e-06, |
| "loss": 12.815, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.3793552117970218, |
| "grad_norm": 0.2743465304374695, |
| "learning_rate": 8.73456790123457e-06, |
| "loss": 12.2468, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.3839814948677172, |
| "grad_norm": 0.2570422291755676, |
| "learning_rate": 8.719135802469136e-06, |
| "loss": 11.9801, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.3886077779384126, |
| "grad_norm": 0.21237581968307495, |
| "learning_rate": 8.703703703703705e-06, |
| "loss": 11.7092, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.393234061009108, |
| "grad_norm": 0.27444881200790405, |
| "learning_rate": 8.688271604938271e-06, |
| "loss": 13.3074, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.39786034407980336, |
| "grad_norm": 0.22574838995933533, |
| "learning_rate": 8.67283950617284e-06, |
| "loss": 13.1573, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.40248662715049877, |
| "grad_norm": 0.252755731344223, |
| "learning_rate": 8.657407407407408e-06, |
| "loss": 12.3472, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.40711291022119417, |
| "grad_norm": 0.2729659378528595, |
| "learning_rate": 8.641975308641975e-06, |
| "loss": 12.8413, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.41173919329188957, |
| "grad_norm": 0.21728309988975525, |
| "learning_rate": 8.626543209876543e-06, |
| "loss": 11.3523, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.4163654763625849, |
| "grad_norm": 0.264457106590271, |
| "learning_rate": 8.611111111111112e-06, |
| "loss": 13.1175, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.4209917594332803, |
| "grad_norm": 0.20264536142349243, |
| "learning_rate": 8.59567901234568e-06, |
| "loss": 12.8442, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.4256180425039757, |
| "grad_norm": 0.3003804385662079, |
| "learning_rate": 8.580246913580249e-06, |
| "loss": 14.5755, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.4302443255746711, |
| "grad_norm": 0.294029176235199, |
| "learning_rate": 8.564814814814816e-06, |
| "loss": 13.5777, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.4348706086453665, |
| "grad_norm": 0.21502816677093506, |
| "learning_rate": 8.549382716049384e-06, |
| "loss": 11.9804, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.4394968917160619, |
| "grad_norm": 0.21898359060287476, |
| "learning_rate": 8.53395061728395e-06, |
| "loss": 13.6382, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.4441231747867573, |
| "grad_norm": 0.20010338723659515, |
| "learning_rate": 8.518518518518519e-06, |
| "loss": 12.4733, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.4487494578574526, |
| "grad_norm": 0.2247803956270218, |
| "learning_rate": 8.503086419753088e-06, |
| "loss": 12.1589, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.45337574092814803, |
| "grad_norm": 0.29326918721199036, |
| "learning_rate": 8.487654320987654e-06, |
| "loss": 12.8177, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.45800202399884343, |
| "grad_norm": 0.18403670191764832, |
| "learning_rate": 8.472222222222223e-06, |
| "loss": 13.8527, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.46262830706953884, |
| "grad_norm": 0.24531710147857666, |
| "learning_rate": 8.456790123456791e-06, |
| "loss": 11.9126, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.4672545901402342, |
| "grad_norm": 0.24183672666549683, |
| "learning_rate": 8.441358024691358e-06, |
| "loss": 12.952, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.4718808732109296, |
| "grad_norm": 0.29563647508621216, |
| "learning_rate": 8.425925925925926e-06, |
| "loss": 14.462, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.476507156281625, |
| "grad_norm": 0.21156498789787292, |
| "learning_rate": 8.410493827160495e-06, |
| "loss": 13.4846, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.4811334393523204, |
| "grad_norm": 0.24183927476406097, |
| "learning_rate": 8.395061728395062e-06, |
| "loss": 11.428, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.48575972242301574, |
| "grad_norm": 0.21241174638271332, |
| "learning_rate": 8.37962962962963e-06, |
| "loss": 13.8068, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.49038600549371114, |
| "grad_norm": 0.2149106115102768, |
| "learning_rate": 8.364197530864199e-06, |
| "loss": 12.7659, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.49501228856440654, |
| "grad_norm": 0.21057994663715363, |
| "learning_rate": 8.348765432098767e-06, |
| "loss": 12.4651, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.49963857163510195, |
| "grad_norm": 0.18793097138404846, |
| "learning_rate": 8.333333333333334e-06, |
| "loss": 11.7854, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.5042648547057973, |
| "grad_norm": 0.1801634281873703, |
| "learning_rate": 8.317901234567902e-06, |
| "loss": 12.9306, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.5088911377764928, |
| "grad_norm": 0.21523930132389069, |
| "learning_rate": 8.30246913580247e-06, |
| "loss": 11.6702, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.5135174208471881, |
| "grad_norm": 0.22931678593158722, |
| "learning_rate": 8.287037037037037e-06, |
| "loss": 11.5856, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.5181437039178834, |
| "grad_norm": 0.23802338540554047, |
| "learning_rate": 8.271604938271606e-06, |
| "loss": 12.7508, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.5227699869885789, |
| "grad_norm": 0.23633763194084167, |
| "learning_rate": 8.256172839506174e-06, |
| "loss": 12.0906, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.5273962700592743, |
| "grad_norm": 0.22553692758083344, |
| "learning_rate": 8.240740740740741e-06, |
| "loss": 12.4982, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.5320225531299696, |
| "grad_norm": 0.2265913337469101, |
| "learning_rate": 8.22530864197531e-06, |
| "loss": 11.4945, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.5366488362006651, |
| "grad_norm": 0.1848449558019638, |
| "learning_rate": 8.209876543209876e-06, |
| "loss": 11.866, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.5412751192713604, |
| "grad_norm": 0.16533511877059937, |
| "learning_rate": 8.194444444444445e-06, |
| "loss": 12.3876, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.5459014023420558, |
| "grad_norm": 0.19875134527683258, |
| "learning_rate": 8.179012345679013e-06, |
| "loss": 12.0186, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.5505276854127512, |
| "grad_norm": 0.20982632040977478, |
| "learning_rate": 8.16358024691358e-06, |
| "loss": 13.1202, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.5551539684834466, |
| "grad_norm": 0.2689734101295471, |
| "learning_rate": 8.148148148148148e-06, |
| "loss": 13.9764, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.559780251554142, |
| "grad_norm": 0.14883318543434143, |
| "learning_rate": 8.132716049382717e-06, |
| "loss": 11.7656, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.5644065346248374, |
| "grad_norm": 0.18115845322608948, |
| "learning_rate": 8.117283950617285e-06, |
| "loss": 11.8902, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.5690328176955327, |
| "grad_norm": 0.23553551733493805, |
| "learning_rate": 8.101851851851854e-06, |
| "loss": 13.0425, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.5736591007662282, |
| "grad_norm": 0.2110109180212021, |
| "learning_rate": 8.08641975308642e-06, |
| "loss": 13.6931, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.5782853838369235, |
| "grad_norm": 0.2748431861400604, |
| "learning_rate": 8.070987654320989e-06, |
| "loss": 12.7852, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.5829116669076189, |
| "grad_norm": 0.1847638040781021, |
| "learning_rate": 8.055555555555557e-06, |
| "loss": 12.9106, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.5875379499783143, |
| "grad_norm": 0.48550117015838623, |
| "learning_rate": 8.040123456790124e-06, |
| "loss": 11.5484, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.5921642330490097, |
| "grad_norm": 0.2579441964626312, |
| "learning_rate": 8.024691358024692e-06, |
| "loss": 11.9785, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.5967905161197051, |
| "grad_norm": 0.19037479162216187, |
| "learning_rate": 8.00925925925926e-06, |
| "loss": 11.0466, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.6014167991904005, |
| "grad_norm": 0.22901126742362976, |
| "learning_rate": 7.993827160493828e-06, |
| "loss": 11.1139, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.6060430822610958, |
| "grad_norm": 0.16406911611557007, |
| "learning_rate": 7.978395061728396e-06, |
| "loss": 10.6613, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.6106693653317913, |
| "grad_norm": 0.23492039740085602, |
| "learning_rate": 7.962962962962963e-06, |
| "loss": 11.3782, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.6152956484024866, |
| "grad_norm": 0.21692270040512085, |
| "learning_rate": 7.947530864197531e-06, |
| "loss": 11.2788, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.619921931473182, |
| "grad_norm": 0.17832407355308533, |
| "learning_rate": 7.9320987654321e-06, |
| "loss": 11.7617, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.6245482145438774, |
| "grad_norm": 0.20935167372226715, |
| "learning_rate": 7.916666666666667e-06, |
| "loss": 10.9103, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.6291744976145728, |
| "grad_norm": 0.199855774641037, |
| "learning_rate": 7.901234567901235e-06, |
| "loss": 11.5339, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.6338007806852681, |
| "grad_norm": 0.1627349704504013, |
| "learning_rate": 7.885802469135803e-06, |
| "loss": 10.3593, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.6384270637559636, |
| "grad_norm": 0.18615403771400452, |
| "learning_rate": 7.870370370370372e-06, |
| "loss": 12.3204, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.6430533468266589, |
| "grad_norm": 0.17866994440555573, |
| "learning_rate": 7.854938271604939e-06, |
| "loss": 11.548, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.6476796298973544, |
| "grad_norm": 0.22644223272800446, |
| "learning_rate": 7.839506172839507e-06, |
| "loss": 12.0289, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.6523059129680497, |
| "grad_norm": 0.18510523438453674, |
| "learning_rate": 7.824074074074076e-06, |
| "loss": 12.1312, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.6569321960387451, |
| "grad_norm": 0.1948799043893814, |
| "learning_rate": 7.808641975308642e-06, |
| "loss": 13.4831, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.6615584791094405, |
| "grad_norm": 0.2084604948759079, |
| "learning_rate": 7.79320987654321e-06, |
| "loss": 11.1289, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.6661847621801359, |
| "grad_norm": 0.194437637925148, |
| "learning_rate": 7.77777777777778e-06, |
| "loss": 11.5961, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.6708110452508312, |
| "grad_norm": 0.2627541124820709, |
| "learning_rate": 7.762345679012346e-06, |
| "loss": 12.0198, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.6754373283215267, |
| "grad_norm": 0.21856093406677246, |
| "learning_rate": 7.746913580246914e-06, |
| "loss": 12.5285, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.680063611392222, |
| "grad_norm": 0.15228690207004547, |
| "learning_rate": 7.731481481481483e-06, |
| "loss": 10.8653, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.6846898944629175, |
| "grad_norm": 0.22951403260231018, |
| "learning_rate": 7.71604938271605e-06, |
| "loss": 10.8751, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.6893161775336129, |
| "grad_norm": 0.25641652941703796, |
| "learning_rate": 7.700617283950618e-06, |
| "loss": 11.0195, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.6939424606043082, |
| "grad_norm": 0.17132331430912018, |
| "learning_rate": 7.685185185185185e-06, |
| "loss": 13.9163, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.6985687436750037, |
| "grad_norm": 0.24479900300502777, |
| "learning_rate": 7.669753086419753e-06, |
| "loss": 11.4487, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.703195026745699, |
| "grad_norm": 0.1990821659564972, |
| "learning_rate": 7.654320987654322e-06, |
| "loss": 10.6267, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.7078213098163944, |
| "grad_norm": 0.17914509773254395, |
| "learning_rate": 7.638888888888888e-06, |
| "loss": 11.5696, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.7124475928870898, |
| "grad_norm": 0.19086718559265137, |
| "learning_rate": 7.623456790123458e-06, |
| "loss": 12.3853, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.7170738759577852, |
| "grad_norm": 0.1417158991098404, |
| "learning_rate": 7.608024691358026e-06, |
| "loss": 10.8805, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.7217001590284805, |
| "grad_norm": 0.32373055815696716, |
| "learning_rate": 7.592592592592594e-06, |
| "loss": 12.586, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.726326442099176, |
| "grad_norm": 0.1995164453983307, |
| "learning_rate": 7.577160493827161e-06, |
| "loss": 11.4977, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.7309527251698713, |
| "grad_norm": 0.16449247300624847, |
| "learning_rate": 7.561728395061729e-06, |
| "loss": 11.8027, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.7355790082405668, |
| "grad_norm": 0.21888476610183716, |
| "learning_rate": 7.546296296296297e-06, |
| "loss": 11.8753, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.7402052913112621, |
| "grad_norm": 0.19433487951755524, |
| "learning_rate": 7.530864197530865e-06, |
| "loss": 11.5252, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.7448315743819575, |
| "grad_norm": 0.19709180295467377, |
| "learning_rate": 7.515432098765433e-06, |
| "loss": 10.7059, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.7494578574526529, |
| "grad_norm": 0.17522279918193817, |
| "learning_rate": 7.500000000000001e-06, |
| "loss": 11.2329, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.7540841405233483, |
| "grad_norm": 0.2038147747516632, |
| "learning_rate": 7.484567901234569e-06, |
| "loss": 11.425, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.7587104235940436, |
| "grad_norm": 0.29561126232147217, |
| "learning_rate": 7.469135802469136e-06, |
| "loss": 13.192, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.7633367066647391, |
| "grad_norm": 0.2295159548521042, |
| "learning_rate": 7.453703703703704e-06, |
| "loss": 10.9714, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.7679629897354344, |
| "grad_norm": 0.1554422378540039, |
| "learning_rate": 7.438271604938272e-06, |
| "loss": 10.5617, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.7725892728061299, |
| "grad_norm": 0.19565965235233307, |
| "learning_rate": 7.42283950617284e-06, |
| "loss": 10.8574, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.7772155558768252, |
| "grad_norm": 0.2645648419857025, |
| "learning_rate": 7.4074074074074075e-06, |
| "loss": 11.4472, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.7818418389475206, |
| "grad_norm": 0.2832958996295929, |
| "learning_rate": 7.391975308641975e-06, |
| "loss": 11.628, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.786468122018216, |
| "grad_norm": 0.2097349464893341, |
| "learning_rate": 7.3765432098765435e-06, |
| "loss": 12.2379, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.7910944050889114, |
| "grad_norm": 0.19168758392333984, |
| "learning_rate": 7.361111111111112e-06, |
| "loss": 10.0489, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.7957206881596067, |
| "grad_norm": 0.22470812499523163, |
| "learning_rate": 7.34567901234568e-06, |
| "loss": 11.8545, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.8003469712303022, |
| "grad_norm": 0.2722543179988861, |
| "learning_rate": 7.330246913580248e-06, |
| "loss": 10.4316, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.8049732543009975, |
| "grad_norm": 0.1686207354068756, |
| "learning_rate": 7.314814814814816e-06, |
| "loss": 12.5917, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.8095995373716929, |
| "grad_norm": 0.2805918753147125, |
| "learning_rate": 7.299382716049383e-06, |
| "loss": 10.6457, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.8142258204423883, |
| "grad_norm": 0.28346166014671326, |
| "learning_rate": 7.283950617283952e-06, |
| "loss": 11.6686, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.8188521035130837, |
| "grad_norm": 0.23147298395633698, |
| "learning_rate": 7.268518518518519e-06, |
| "loss": 10.316, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.8234783865837791, |
| "grad_norm": 0.17895270884037018, |
| "learning_rate": 7.253086419753087e-06, |
| "loss": 10.5762, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.8281046696544745, |
| "grad_norm": 0.2260911464691162, |
| "learning_rate": 7.2376543209876545e-06, |
| "loss": 10.9989, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.8327309527251698, |
| "grad_norm": 0.2163412719964981, |
| "learning_rate": 7.222222222222223e-06, |
| "loss": 10.7737, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.8373572357958653, |
| "grad_norm": 0.21451716125011444, |
| "learning_rate": 7.2067901234567905e-06, |
| "loss": 12.4478, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.8419835188665606, |
| "grad_norm": 0.14181958138942719, |
| "learning_rate": 7.191358024691358e-06, |
| "loss": 11.1375, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.846609801937256, |
| "grad_norm": 0.18365442752838135, |
| "learning_rate": 7.1759259259259266e-06, |
| "loss": 11.3768, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.8512360850079514, |
| "grad_norm": 0.20560483634471893, |
| "learning_rate": 7.160493827160494e-06, |
| "loss": 10.6342, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.8558623680786468, |
| "grad_norm": 0.3396760821342468, |
| "learning_rate": 7.145061728395062e-06, |
| "loss": 12.9821, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.8604886511493423, |
| "grad_norm": 0.18169118463993073, |
| "learning_rate": 7.129629629629629e-06, |
| "loss": 11.067, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.8651149342200376, |
| "grad_norm": 0.18200387060642242, |
| "learning_rate": 7.114197530864199e-06, |
| "loss": 10.6416, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.869741217290733, |
| "grad_norm": 0.23793809115886688, |
| "learning_rate": 7.098765432098766e-06, |
| "loss": 12.6389, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.8743675003614284, |
| "grad_norm": 0.6948350667953491, |
| "learning_rate": 7.083333333333335e-06, |
| "loss": 11.5131, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.8789937834321238, |
| "grad_norm": 0.19556772708892822, |
| "learning_rate": 7.067901234567902e-06, |
| "loss": 12.9335, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.8836200665028191, |
| "grad_norm": 0.21277347207069397, |
| "learning_rate": 7.05246913580247e-06, |
| "loss": 11.533, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.8882463495735146, |
| "grad_norm": 0.13241365551948547, |
| "learning_rate": 7.0370370370370375e-06, |
| "loss": 11.7588, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.8928726326442099, |
| "grad_norm": 0.16065774857997894, |
| "learning_rate": 7.021604938271606e-06, |
| "loss": 10.907, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.8974989157149053, |
| "grad_norm": 0.2425934374332428, |
| "learning_rate": 7.0061728395061735e-06, |
| "loss": 12.0332, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.9021251987856007, |
| "grad_norm": 0.32493147253990173, |
| "learning_rate": 6.990740740740741e-06, |
| "loss": 11.0481, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.9067514818562961, |
| "grad_norm": 0.20006415247917175, |
| "learning_rate": 6.975308641975309e-06, |
| "loss": 10.3825, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.9113777649269915, |
| "grad_norm": 0.14633627235889435, |
| "learning_rate": 6.959876543209877e-06, |
| "loss": 11.3131, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.9160040479976869, |
| "grad_norm": 0.1695437729358673, |
| "learning_rate": 6.944444444444445e-06, |
| "loss": 11.6805, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.9206303310683822, |
| "grad_norm": 0.22382433712482452, |
| "learning_rate": 6.929012345679012e-06, |
| "loss": 12.4138, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.9252566141390777, |
| "grad_norm": 0.18652208149433136, |
| "learning_rate": 6.913580246913581e-06, |
| "loss": 11.1235, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.929882897209773, |
| "grad_norm": 0.21566657721996307, |
| "learning_rate": 6.898148148148148e-06, |
| "loss": 11.1278, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.9345091802804684, |
| "grad_norm": 0.17630067467689514, |
| "learning_rate": 6.882716049382716e-06, |
| "loss": 12.1383, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.9391354633511638, |
| "grad_norm": 0.22753016650676727, |
| "learning_rate": 6.867283950617285e-06, |
| "loss": 12.0065, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.9437617464218592, |
| "grad_norm": 0.12988047301769257, |
| "learning_rate": 6.851851851851853e-06, |
| "loss": 11.9477, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.9483880294925546, |
| "grad_norm": 0.2507847547531128, |
| "learning_rate": 6.8364197530864205e-06, |
| "loss": 11.5818, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.95301431256325, |
| "grad_norm": 0.18030938506126404, |
| "learning_rate": 6.820987654320988e-06, |
| "loss": 13.1558, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.9576405956339453, |
| "grad_norm": 0.21600230038166046, |
| "learning_rate": 6.8055555555555566e-06, |
| "loss": 10.7038, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.9622668787046408, |
| "grad_norm": 0.1989142745733261, |
| "learning_rate": 6.790123456790124e-06, |
| "loss": 11.9244, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.9668931617753361, |
| "grad_norm": 0.2517797648906708, |
| "learning_rate": 6.774691358024692e-06, |
| "loss": 11.0019, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.9715194448460315, |
| "grad_norm": 0.24413681030273438, |
| "learning_rate": 6.75925925925926e-06, |
| "loss": 12.4322, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.9761457279167269, |
| "grad_norm": 0.20667782425880432, |
| "learning_rate": 6.743827160493828e-06, |
| "loss": 10.6134, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.9807720109874223, |
| "grad_norm": 0.17336109280586243, |
| "learning_rate": 6.728395061728395e-06, |
| "loss": 10.4673, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.9853982940581176, |
| "grad_norm": 0.17548175156116486, |
| "learning_rate": 6.712962962962963e-06, |
| "loss": 10.1377, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.9900245771288131, |
| "grad_norm": 0.1717563420534134, |
| "learning_rate": 6.6975308641975314e-06, |
| "loss": 11.5888, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.9946508601995084, |
| "grad_norm": 0.1841057986021042, |
| "learning_rate": 6.682098765432099e-06, |
| "loss": 12.2213, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.9992771432702039, |
| "grad_norm": 0.3000975251197815, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 11.046, |
| "step": 216 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.3000975251197815, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 1.4199, |
| "step": 217 |
| }, |
| { |
| "epoch": 1.0046262830706953, |
| "grad_norm": 0.4262640178203583, |
| "learning_rate": 6.651234567901235e-06, |
| "loss": 10.5265, |
| "step": 218 |
| }, |
| { |
| "epoch": 1.0092525661413907, |
| "grad_norm": 0.21694016456604004, |
| "learning_rate": 6.635802469135803e-06, |
| "loss": 10.4657, |
| "step": 219 |
| }, |
| { |
| "epoch": 1.0138788492120863, |
| "grad_norm": 0.2662133276462555, |
| "learning_rate": 6.620370370370371e-06, |
| "loss": 10.7531, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.0185051322827816, |
| "grad_norm": 0.4052404463291168, |
| "learning_rate": 6.60493827160494e-06, |
| "loss": 11.1598, |
| "step": 221 |
| }, |
| { |
| "epoch": 1.023131415353477, |
| "grad_norm": 0.23900097608566284, |
| "learning_rate": 6.589506172839507e-06, |
| "loss": 11.5257, |
| "step": 222 |
| }, |
| { |
| "epoch": 1.0277576984241723, |
| "grad_norm": 0.22569526731967926, |
| "learning_rate": 6.574074074074075e-06, |
| "loss": 11.5564, |
| "step": 223 |
| }, |
| { |
| "epoch": 1.0323839814948677, |
| "grad_norm": 0.29018473625183105, |
| "learning_rate": 6.558641975308642e-06, |
| "loss": 11.8655, |
| "step": 224 |
| }, |
| { |
| "epoch": 1.037010264565563, |
| "grad_norm": 0.2856806516647339, |
| "learning_rate": 6.543209876543211e-06, |
| "loss": 11.8046, |
| "step": 225 |
| }, |
| { |
| "epoch": 1.0416365476362586, |
| "grad_norm": 0.16691464185714722, |
| "learning_rate": 6.5277777777777784e-06, |
| "loss": 12.3704, |
| "step": 226 |
| }, |
| { |
| "epoch": 1.046262830706954, |
| "grad_norm": 0.24121911823749542, |
| "learning_rate": 6.512345679012346e-06, |
| "loss": 10.7236, |
| "step": 227 |
| }, |
| { |
| "epoch": 1.0508891137776493, |
| "grad_norm": 0.205901101231575, |
| "learning_rate": 6.4969135802469145e-06, |
| "loss": 10.9316, |
| "step": 228 |
| }, |
| { |
| "epoch": 1.0555153968483446, |
| "grad_norm": 0.18375830352306366, |
| "learning_rate": 6.481481481481482e-06, |
| "loss": 9.422, |
| "step": 229 |
| }, |
| { |
| "epoch": 1.06014167991904, |
| "grad_norm": 0.2132750004529953, |
| "learning_rate": 6.46604938271605e-06, |
| "loss": 12.6065, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.0647679629897355, |
| "grad_norm": 0.23490285873413086, |
| "learning_rate": 6.450617283950617e-06, |
| "loss": 11.2282, |
| "step": 231 |
| }, |
| { |
| "epoch": 1.0693942460604309, |
| "grad_norm": 0.2604386806488037, |
| "learning_rate": 6.435185185185186e-06, |
| "loss": 10.8157, |
| "step": 232 |
| }, |
| { |
| "epoch": 1.0740205291311262, |
| "grad_norm": 0.21887636184692383, |
| "learning_rate": 6.419753086419753e-06, |
| "loss": 10.6996, |
| "step": 233 |
| }, |
| { |
| "epoch": 1.0786468122018216, |
| "grad_norm": 0.16946089267730713, |
| "learning_rate": 6.404320987654321e-06, |
| "loss": 10.3805, |
| "step": 234 |
| }, |
| { |
| "epoch": 1.083273095272517, |
| "grad_norm": 0.20204631984233856, |
| "learning_rate": 6.3888888888888885e-06, |
| "loss": 11.4941, |
| "step": 235 |
| }, |
| { |
| "epoch": 1.0878993783432125, |
| "grad_norm": 0.202669158577919, |
| "learning_rate": 6.373456790123458e-06, |
| "loss": 11.7944, |
| "step": 236 |
| }, |
| { |
| "epoch": 1.0925256614139078, |
| "grad_norm": 0.22706535458564758, |
| "learning_rate": 6.358024691358025e-06, |
| "loss": 12.6031, |
| "step": 237 |
| }, |
| { |
| "epoch": 1.0971519444846032, |
| "grad_norm": 0.17785371840000153, |
| "learning_rate": 6.342592592592594e-06, |
| "loss": 10.5881, |
| "step": 238 |
| }, |
| { |
| "epoch": 1.1017782275552985, |
| "grad_norm": 0.17540261149406433, |
| "learning_rate": 6.3271604938271615e-06, |
| "loss": 10.996, |
| "step": 239 |
| }, |
| { |
| "epoch": 1.1064045106259939, |
| "grad_norm": 0.22038735449314117, |
| "learning_rate": 6.311728395061729e-06, |
| "loss": 10.0293, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.1110307936966892, |
| "grad_norm": 0.17491741478443146, |
| "learning_rate": 6.296296296296297e-06, |
| "loss": 10.0996, |
| "step": 241 |
| }, |
| { |
| "epoch": 1.1156570767673848, |
| "grad_norm": 0.2827455699443817, |
| "learning_rate": 6.280864197530865e-06, |
| "loss": 11.0431, |
| "step": 242 |
| }, |
| { |
| "epoch": 1.1202833598380801, |
| "grad_norm": 0.21907830238342285, |
| "learning_rate": 6.265432098765433e-06, |
| "loss": 10.4707, |
| "step": 243 |
| }, |
| { |
| "epoch": 1.1249096429087755, |
| "grad_norm": 0.19320356845855713, |
| "learning_rate": 6.25e-06, |
| "loss": 10.7269, |
| "step": 244 |
| }, |
| { |
| "epoch": 1.1295359259794708, |
| "grad_norm": 0.22684165835380554, |
| "learning_rate": 6.234567901234569e-06, |
| "loss": 11.3143, |
| "step": 245 |
| }, |
| { |
| "epoch": 1.1341622090501662, |
| "grad_norm": 0.21541574597358704, |
| "learning_rate": 6.219135802469136e-06, |
| "loss": 10.397, |
| "step": 246 |
| }, |
| { |
| "epoch": 1.1387884921208618, |
| "grad_norm": 0.21086207032203674, |
| "learning_rate": 6.203703703703704e-06, |
| "loss": 11.0086, |
| "step": 247 |
| }, |
| { |
| "epoch": 1.143414775191557, |
| "grad_norm": 0.26652148365974426, |
| "learning_rate": 6.1882716049382715e-06, |
| "loss": 11.5252, |
| "step": 248 |
| }, |
| { |
| "epoch": 1.1480410582622524, |
| "grad_norm": 0.1995311677455902, |
| "learning_rate": 6.17283950617284e-06, |
| "loss": 9.3802, |
| "step": 249 |
| }, |
| { |
| "epoch": 1.1526673413329478, |
| "grad_norm": 0.19715267419815063, |
| "learning_rate": 6.157407407407408e-06, |
| "loss": 11.6554, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.1572936244036431, |
| "grad_norm": 0.16357477009296417, |
| "learning_rate": 6.141975308641975e-06, |
| "loss": 10.7531, |
| "step": 251 |
| }, |
| { |
| "epoch": 1.1619199074743385, |
| "grad_norm": 0.2201809138059616, |
| "learning_rate": 6.126543209876543e-06, |
| "loss": 11.4639, |
| "step": 252 |
| }, |
| { |
| "epoch": 1.166546190545034, |
| "grad_norm": 0.24760210514068604, |
| "learning_rate": 6.111111111111112e-06, |
| "loss": 10.1224, |
| "step": 253 |
| }, |
| { |
| "epoch": 1.1711724736157294, |
| "grad_norm": 0.18471167981624603, |
| "learning_rate": 6.09567901234568e-06, |
| "loss": 10.8261, |
| "step": 254 |
| }, |
| { |
| "epoch": 1.1757987566864248, |
| "grad_norm": 0.20257946848869324, |
| "learning_rate": 6.080246913580248e-06, |
| "loss": 9.8945, |
| "step": 255 |
| }, |
| { |
| "epoch": 1.18042503975712, |
| "grad_norm": 0.15809153020381927, |
| "learning_rate": 6.064814814814816e-06, |
| "loss": 10.4489, |
| "step": 256 |
| }, |
| { |
| "epoch": 1.1850513228278154, |
| "grad_norm": 0.2175232172012329, |
| "learning_rate": 6.049382716049383e-06, |
| "loss": 10.6301, |
| "step": 257 |
| }, |
| { |
| "epoch": 1.189677605898511, |
| "grad_norm": 0.22652743756771088, |
| "learning_rate": 6.033950617283951e-06, |
| "loss": 9.6189, |
| "step": 258 |
| }, |
| { |
| "epoch": 1.1943038889692064, |
| "grad_norm": 0.21281598508358002, |
| "learning_rate": 6.018518518518519e-06, |
| "loss": 10.8016, |
| "step": 259 |
| }, |
| { |
| "epoch": 1.1989301720399017, |
| "grad_norm": 0.20046480000019073, |
| "learning_rate": 6.003086419753087e-06, |
| "loss": 10.1304, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.203556455110597, |
| "grad_norm": 0.19008351862430573, |
| "learning_rate": 5.9876543209876546e-06, |
| "loss": 10.6718, |
| "step": 261 |
| }, |
| { |
| "epoch": 1.2081827381812924, |
| "grad_norm": 0.2159290611743927, |
| "learning_rate": 5.972222222222222e-06, |
| "loss": 10.1963, |
| "step": 262 |
| }, |
| { |
| "epoch": 1.2128090212519878, |
| "grad_norm": 0.1703040450811386, |
| "learning_rate": 5.956790123456791e-06, |
| "loss": 10.9394, |
| "step": 263 |
| }, |
| { |
| "epoch": 1.2174353043226833, |
| "grad_norm": 0.20243868231773376, |
| "learning_rate": 5.941358024691358e-06, |
| "loss": 9.7358, |
| "step": 264 |
| }, |
| { |
| "epoch": 1.2220615873933787, |
| "grad_norm": 0.21622534096240997, |
| "learning_rate": 5.925925925925926e-06, |
| "loss": 9.2896, |
| "step": 265 |
| }, |
| { |
| "epoch": 1.226687870464074, |
| "grad_norm": 0.19408589601516724, |
| "learning_rate": 5.910493827160494e-06, |
| "loss": 10.0141, |
| "step": 266 |
| }, |
| { |
| "epoch": 1.2313141535347694, |
| "grad_norm": 0.19652226567268372, |
| "learning_rate": 5.895061728395062e-06, |
| "loss": 10.6724, |
| "step": 267 |
| }, |
| { |
| "epoch": 1.2359404366054647, |
| "grad_norm": 0.22394828498363495, |
| "learning_rate": 5.8796296296296295e-06, |
| "loss": 11.7472, |
| "step": 268 |
| }, |
| { |
| "epoch": 1.2405667196761603, |
| "grad_norm": 0.18796682357788086, |
| "learning_rate": 5.864197530864199e-06, |
| "loss": 11.3051, |
| "step": 269 |
| }, |
| { |
| "epoch": 1.2451930027468556, |
| "grad_norm": 0.1897086501121521, |
| "learning_rate": 5.848765432098766e-06, |
| "loss": 9.9475, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.249819285817551, |
| "grad_norm": 0.19912078976631165, |
| "learning_rate": 5.833333333333334e-06, |
| "loss": 10.6907, |
| "step": 271 |
| }, |
| { |
| "epoch": 1.2544455688882463, |
| "grad_norm": 0.25889354944229126, |
| "learning_rate": 5.817901234567902e-06, |
| "loss": 9.9069, |
| "step": 272 |
| }, |
| { |
| "epoch": 1.2590718519589417, |
| "grad_norm": 0.19028806686401367, |
| "learning_rate": 5.80246913580247e-06, |
| "loss": 10.0748, |
| "step": 273 |
| }, |
| { |
| "epoch": 1.263698135029637, |
| "grad_norm": 0.1661495566368103, |
| "learning_rate": 5.787037037037038e-06, |
| "loss": 10.1342, |
| "step": 274 |
| }, |
| { |
| "epoch": 1.2683244181003326, |
| "grad_norm": 0.17706620693206787, |
| "learning_rate": 5.771604938271605e-06, |
| "loss": 12.4703, |
| "step": 275 |
| }, |
| { |
| "epoch": 1.272950701171028, |
| "grad_norm": 0.26582854986190796, |
| "learning_rate": 5.756172839506174e-06, |
| "loss": 12.6464, |
| "step": 276 |
| }, |
| { |
| "epoch": 1.2775769842417233, |
| "grad_norm": 0.178712397813797, |
| "learning_rate": 5.740740740740741e-06, |
| "loss": 9.6011, |
| "step": 277 |
| }, |
| { |
| "epoch": 1.2822032673124186, |
| "grad_norm": 0.18910570442676544, |
| "learning_rate": 5.725308641975309e-06, |
| "loss": 10.8773, |
| "step": 278 |
| }, |
| { |
| "epoch": 1.2868295503831142, |
| "grad_norm": 0.24754610657691956, |
| "learning_rate": 5.7098765432098764e-06, |
| "loss": 10.8981, |
| "step": 279 |
| }, |
| { |
| "epoch": 1.2914558334538095, |
| "grad_norm": 0.14676739275455475, |
| "learning_rate": 5.694444444444445e-06, |
| "loss": 10.1851, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.296082116524505, |
| "grad_norm": 0.21573857963085175, |
| "learning_rate": 5.6790123456790125e-06, |
| "loss": 10.2565, |
| "step": 281 |
| }, |
| { |
| "epoch": 1.3007083995952002, |
| "grad_norm": 0.24850860238075256, |
| "learning_rate": 5.66358024691358e-06, |
| "loss": 9.9282, |
| "step": 282 |
| }, |
| { |
| "epoch": 1.3053346826658956, |
| "grad_norm": 0.180882066488266, |
| "learning_rate": 5.6481481481481485e-06, |
| "loss": 11.6456, |
| "step": 283 |
| }, |
| { |
| "epoch": 1.309960965736591, |
| "grad_norm": 0.19305004179477692, |
| "learning_rate": 5.632716049382716e-06, |
| "loss": 10.0129, |
| "step": 284 |
| }, |
| { |
| "epoch": 1.3145872488072863, |
| "grad_norm": 0.18499980866909027, |
| "learning_rate": 5.617283950617285e-06, |
| "loss": 10.6136, |
| "step": 285 |
| }, |
| { |
| "epoch": 1.3192135318779818, |
| "grad_norm": 0.2523798942565918, |
| "learning_rate": 5.601851851851853e-06, |
| "loss": 9.1629, |
| "step": 286 |
| }, |
| { |
| "epoch": 1.3238398149486772, |
| "grad_norm": 0.19359612464904785, |
| "learning_rate": 5.586419753086421e-06, |
| "loss": 9.9917, |
| "step": 287 |
| }, |
| { |
| "epoch": 1.3284660980193725, |
| "grad_norm": 0.15306031703948975, |
| "learning_rate": 5.570987654320988e-06, |
| "loss": 9.6058, |
| "step": 288 |
| }, |
| { |
| "epoch": 1.333092381090068, |
| "grad_norm": 0.2070600688457489, |
| "learning_rate": 5.555555555555557e-06, |
| "loss": 10.2776, |
| "step": 289 |
| }, |
| { |
| "epoch": 1.3377186641607635, |
| "grad_norm": 0.1619105190038681, |
| "learning_rate": 5.540123456790124e-06, |
| "loss": 9.1465, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.3423449472314588, |
| "grad_norm": 0.16202016174793243, |
| "learning_rate": 5.524691358024692e-06, |
| "loss": 10.0362, |
| "step": 291 |
| }, |
| { |
| "epoch": 1.3469712303021542, |
| "grad_norm": 0.19075003266334534, |
| "learning_rate": 5.5092592592592595e-06, |
| "loss": 10.0189, |
| "step": 292 |
| }, |
| { |
| "epoch": 1.3515975133728495, |
| "grad_norm": 0.16484172642230988, |
| "learning_rate": 5.493827160493828e-06, |
| "loss": 12.3743, |
| "step": 293 |
| }, |
| { |
| "epoch": 1.3562237964435448, |
| "grad_norm": 0.2151799499988556, |
| "learning_rate": 5.4783950617283955e-06, |
| "loss": 9.9846, |
| "step": 294 |
| }, |
| { |
| "epoch": 1.3608500795142402, |
| "grad_norm": 0.20363835990428925, |
| "learning_rate": 5.462962962962963e-06, |
| "loss": 9.671, |
| "step": 295 |
| }, |
| { |
| "epoch": 1.3654763625849355, |
| "grad_norm": 0.15196539461612701, |
| "learning_rate": 5.447530864197531e-06, |
| "loss": 9.97, |
| "step": 296 |
| }, |
| { |
| "epoch": 1.3701026456556311, |
| "grad_norm": 0.17592975497245789, |
| "learning_rate": 5.432098765432099e-06, |
| "loss": 10.8506, |
| "step": 297 |
| }, |
| { |
| "epoch": 1.3747289287263265, |
| "grad_norm": 0.20012719929218292, |
| "learning_rate": 5.416666666666667e-06, |
| "loss": 10.4503, |
| "step": 298 |
| }, |
| { |
| "epoch": 1.3793552117970218, |
| "grad_norm": 0.19312864542007446, |
| "learning_rate": 5.401234567901234e-06, |
| "loss": 9.7727, |
| "step": 299 |
| }, |
| { |
| "epoch": 1.3839814948677172, |
| "grad_norm": 0.14697419106960297, |
| "learning_rate": 5.385802469135803e-06, |
| "loss": 11.9123, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.3886077779384127, |
| "grad_norm": 0.15580600500106812, |
| "learning_rate": 5.370370370370371e-06, |
| "loss": 10.3442, |
| "step": 301 |
| }, |
| { |
| "epoch": 1.393234061009108, |
| "grad_norm": 0.19586268067359924, |
| "learning_rate": 5.354938271604939e-06, |
| "loss": 8.9558, |
| "step": 302 |
| }, |
| { |
| "epoch": 1.3978603440798034, |
| "grad_norm": 0.19394423067569733, |
| "learning_rate": 5.339506172839507e-06, |
| "loss": 10.8312, |
| "step": 303 |
| }, |
| { |
| "epoch": 1.4024866271504988, |
| "grad_norm": 0.21832065284252167, |
| "learning_rate": 5.324074074074075e-06, |
| "loss": 9.9943, |
| "step": 304 |
| }, |
| { |
| "epoch": 1.4071129102211941, |
| "grad_norm": 0.16751375794410706, |
| "learning_rate": 5.3086419753086425e-06, |
| "loss": 11.3979, |
| "step": 305 |
| }, |
| { |
| "epoch": 1.4117391932918895, |
| "grad_norm": 0.1763811856508255, |
| "learning_rate": 5.29320987654321e-06, |
| "loss": 11.175, |
| "step": 306 |
| }, |
| { |
| "epoch": 1.4163654763625848, |
| "grad_norm": 0.1916954666376114, |
| "learning_rate": 5.2777777777777785e-06, |
| "loss": 8.8938, |
| "step": 307 |
| }, |
| { |
| "epoch": 1.4209917594332804, |
| "grad_norm": 0.12443282455205917, |
| "learning_rate": 5.262345679012346e-06, |
| "loss": 9.6263, |
| "step": 308 |
| }, |
| { |
| "epoch": 1.4256180425039757, |
| "grad_norm": 0.18852603435516357, |
| "learning_rate": 5.246913580246914e-06, |
| "loss": 10.6811, |
| "step": 309 |
| }, |
| { |
| "epoch": 1.430244325574671, |
| "grad_norm": 0.1854352355003357, |
| "learning_rate": 5.231481481481482e-06, |
| "loss": 9.9107, |
| "step": 310 |
| }, |
| { |
| "epoch": 1.4348706086453664, |
| "grad_norm": 0.2382035106420517, |
| "learning_rate": 5.21604938271605e-06, |
| "loss": 10.8001, |
| "step": 311 |
| }, |
| { |
| "epoch": 1.439496891716062, |
| "grad_norm": 0.18645620346069336, |
| "learning_rate": 5.200617283950617e-06, |
| "loss": 9.686, |
| "step": 312 |
| }, |
| { |
| "epoch": 1.4441231747867573, |
| "grad_norm": 0.1910058856010437, |
| "learning_rate": 5.185185185185185e-06, |
| "loss": 10.3407, |
| "step": 313 |
| }, |
| { |
| "epoch": 1.4487494578574527, |
| "grad_norm": 0.23428039252758026, |
| "learning_rate": 5.1697530864197534e-06, |
| "loss": 9.2624, |
| "step": 314 |
| }, |
| { |
| "epoch": 1.453375740928148, |
| "grad_norm": 0.17926917970180511, |
| "learning_rate": 5.154320987654321e-06, |
| "loss": 9.5153, |
| "step": 315 |
| }, |
| { |
| "epoch": 1.4580020239988434, |
| "grad_norm": 0.22386913001537323, |
| "learning_rate": 5.138888888888889e-06, |
| "loss": 10.6948, |
| "step": 316 |
| }, |
| { |
| "epoch": 1.4626283070695387, |
| "grad_norm": 0.20423167943954468, |
| "learning_rate": 5.123456790123458e-06, |
| "loss": 10.3966, |
| "step": 317 |
| }, |
| { |
| "epoch": 1.467254590140234, |
| "grad_norm": 0.15411533415317535, |
| "learning_rate": 5.1080246913580255e-06, |
| "loss": 10.2981, |
| "step": 318 |
| }, |
| { |
| "epoch": 1.4718808732109296, |
| "grad_norm": 0.11654549837112427, |
| "learning_rate": 5.092592592592593e-06, |
| "loss": 10.1807, |
| "step": 319 |
| }, |
| { |
| "epoch": 1.476507156281625, |
| "grad_norm": 0.20476582646369934, |
| "learning_rate": 5.0771604938271616e-06, |
| "loss": 9.9201, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.4811334393523203, |
| "grad_norm": 0.19140280783176422, |
| "learning_rate": 5.061728395061729e-06, |
| "loss": 10.3292, |
| "step": 321 |
| }, |
| { |
| "epoch": 1.4857597224230157, |
| "grad_norm": 0.16844482719898224, |
| "learning_rate": 5.046296296296297e-06, |
| "loss": 10.0726, |
| "step": 322 |
| }, |
| { |
| "epoch": 1.4903860054937113, |
| "grad_norm": 0.22298437356948853, |
| "learning_rate": 5.030864197530864e-06, |
| "loss": 10.9447, |
| "step": 323 |
| }, |
| { |
| "epoch": 1.4950122885644066, |
| "grad_norm": 0.1775355190038681, |
| "learning_rate": 5.015432098765433e-06, |
| "loss": 12.1205, |
| "step": 324 |
| }, |
| { |
| "epoch": 1.499638571635102, |
| "grad_norm": 0.198349267244339, |
| "learning_rate": 5e-06, |
| "loss": 9.0277, |
| "step": 325 |
| }, |
| { |
| "epoch": 1.5042648547057973, |
| "grad_norm": 0.16838030517101288, |
| "learning_rate": 4.984567901234568e-06, |
| "loss": 10.3551, |
| "step": 326 |
| }, |
| { |
| "epoch": 1.5088911377764926, |
| "grad_norm": 0.15806855261325836, |
| "learning_rate": 4.9691358024691365e-06, |
| "loss": 10.0048, |
| "step": 327 |
| }, |
| { |
| "epoch": 1.513517420847188, |
| "grad_norm": 0.21915429830551147, |
| "learning_rate": 4.953703703703704e-06, |
| "loss": 10.554, |
| "step": 328 |
| }, |
| { |
| "epoch": 1.5181437039178833, |
| "grad_norm": 0.23262116312980652, |
| "learning_rate": 4.938271604938272e-06, |
| "loss": 10.8198, |
| "step": 329 |
| }, |
| { |
| "epoch": 1.522769986988579, |
| "grad_norm": 0.18927887082099915, |
| "learning_rate": 4.92283950617284e-06, |
| "loss": 10.7398, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.5273962700592743, |
| "grad_norm": 0.138357013463974, |
| "learning_rate": 4.907407407407408e-06, |
| "loss": 9.8479, |
| "step": 331 |
| }, |
| { |
| "epoch": 1.5320225531299696, |
| "grad_norm": 0.20223885774612427, |
| "learning_rate": 4.891975308641976e-06, |
| "loss": 11.1796, |
| "step": 332 |
| }, |
| { |
| "epoch": 1.5366488362006652, |
| "grad_norm": 0.17738379538059235, |
| "learning_rate": 4.876543209876544e-06, |
| "loss": 9.1805, |
| "step": 333 |
| }, |
| { |
| "epoch": 1.5412751192713605, |
| "grad_norm": 0.18442702293395996, |
| "learning_rate": 4.861111111111111e-06, |
| "loss": 8.4714, |
| "step": 334 |
| }, |
| { |
| "epoch": 1.5459014023420559, |
| "grad_norm": 0.19169731438159943, |
| "learning_rate": 4.845679012345679e-06, |
| "loss": 9.1503, |
| "step": 335 |
| }, |
| { |
| "epoch": 1.5505276854127512, |
| "grad_norm": 0.19981014728546143, |
| "learning_rate": 4.830246913580247e-06, |
| "loss": 9.6427, |
| "step": 336 |
| }, |
| { |
| "epoch": 1.5551539684834466, |
| "grad_norm": 0.1916283518075943, |
| "learning_rate": 4.814814814814815e-06, |
| "loss": 8.8991, |
| "step": 337 |
| }, |
| { |
| "epoch": 1.559780251554142, |
| "grad_norm": 0.14049910008907318, |
| "learning_rate": 4.7993827160493834e-06, |
| "loss": 9.2888, |
| "step": 338 |
| }, |
| { |
| "epoch": 1.5644065346248373, |
| "grad_norm": 0.14010216295719147, |
| "learning_rate": 4.783950617283951e-06, |
| "loss": 10.4533, |
| "step": 339 |
| }, |
| { |
| "epoch": 1.5690328176955326, |
| "grad_norm": 0.1919959932565689, |
| "learning_rate": 4.768518518518519e-06, |
| "loss": 10.149, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.5736591007662282, |
| "grad_norm": 0.1598203033208847, |
| "learning_rate": 4.753086419753087e-06, |
| "loss": 11.1529, |
| "step": 341 |
| }, |
| { |
| "epoch": 1.5782853838369235, |
| "grad_norm": 0.19053767621517181, |
| "learning_rate": 4.737654320987655e-06, |
| "loss": 10.5543, |
| "step": 342 |
| }, |
| { |
| "epoch": 1.5829116669076189, |
| "grad_norm": 0.21635524928569794, |
| "learning_rate": 4.722222222222222e-06, |
| "loss": 8.7965, |
| "step": 343 |
| }, |
| { |
| "epoch": 1.5875379499783144, |
| "grad_norm": 0.2088153064250946, |
| "learning_rate": 4.706790123456791e-06, |
| "loss": 9.1113, |
| "step": 344 |
| }, |
| { |
| "epoch": 1.5921642330490098, |
| "grad_norm": 0.1993483006954193, |
| "learning_rate": 4.691358024691358e-06, |
| "loss": 9.7291, |
| "step": 345 |
| }, |
| { |
| "epoch": 1.5967905161197051, |
| "grad_norm": 0.24835029244422913, |
| "learning_rate": 4.675925925925927e-06, |
| "loss": 9.5215, |
| "step": 346 |
| }, |
| { |
| "epoch": 1.6014167991904005, |
| "grad_norm": 0.24942326545715332, |
| "learning_rate": 4.660493827160494e-06, |
| "loss": 9.3023, |
| "step": 347 |
| }, |
| { |
| "epoch": 1.6060430822610958, |
| "grad_norm": 0.18829648196697235, |
| "learning_rate": 4.645061728395062e-06, |
| "loss": 9.6869, |
| "step": 348 |
| }, |
| { |
| "epoch": 1.6106693653317912, |
| "grad_norm": 0.17725762724876404, |
| "learning_rate": 4.62962962962963e-06, |
| "loss": 10.0154, |
| "step": 349 |
| }, |
| { |
| "epoch": 1.6152956484024865, |
| "grad_norm": 0.145020991563797, |
| "learning_rate": 4.614197530864198e-06, |
| "loss": 10.6015, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.6199219314731819, |
| "grad_norm": 0.12719136476516724, |
| "learning_rate": 4.598765432098766e-06, |
| "loss": 9.9244, |
| "step": 351 |
| }, |
| { |
| "epoch": 1.6245482145438774, |
| "grad_norm": 0.1771680861711502, |
| "learning_rate": 4.583333333333333e-06, |
| "loss": 9.4468, |
| "step": 352 |
| }, |
| { |
| "epoch": 1.6291744976145728, |
| "grad_norm": 0.1684349775314331, |
| "learning_rate": 4.567901234567902e-06, |
| "loss": 11.5601, |
| "step": 353 |
| }, |
| { |
| "epoch": 1.6338007806852681, |
| "grad_norm": 0.1284870207309723, |
| "learning_rate": 4.55246913580247e-06, |
| "loss": 9.5658, |
| "step": 354 |
| }, |
| { |
| "epoch": 1.6384270637559637, |
| "grad_norm": 0.15432706475257874, |
| "learning_rate": 4.537037037037038e-06, |
| "loss": 10.2547, |
| "step": 355 |
| }, |
| { |
| "epoch": 1.643053346826659, |
| "grad_norm": 0.19519680738449097, |
| "learning_rate": 4.521604938271605e-06, |
| "loss": 9.2126, |
| "step": 356 |
| }, |
| { |
| "epoch": 1.6476796298973544, |
| "grad_norm": 0.1388140469789505, |
| "learning_rate": 4.506172839506173e-06, |
| "loss": 9.9578, |
| "step": 357 |
| }, |
| { |
| "epoch": 1.6523059129680497, |
| "grad_norm": 0.19714529812335968, |
| "learning_rate": 4.490740740740741e-06, |
| "loss": 11.7677, |
| "step": 358 |
| }, |
| { |
| "epoch": 1.656932196038745, |
| "grad_norm": 0.22792014479637146, |
| "learning_rate": 4.475308641975309e-06, |
| "loss": 9.6164, |
| "step": 359 |
| }, |
| { |
| "epoch": 1.6615584791094404, |
| "grad_norm": 0.1519528478384018, |
| "learning_rate": 4.4598765432098765e-06, |
| "loss": 9.4579, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.6661847621801358, |
| "grad_norm": 0.17547698318958282, |
| "learning_rate": 4.444444444444444e-06, |
| "loss": 8.2953, |
| "step": 361 |
| }, |
| { |
| "epoch": 1.6708110452508311, |
| "grad_norm": 0.2348184585571289, |
| "learning_rate": 4.429012345679013e-06, |
| "loss": 9.8739, |
| "step": 362 |
| }, |
| { |
| "epoch": 1.6754373283215267, |
| "grad_norm": 0.16753868758678436, |
| "learning_rate": 4.413580246913581e-06, |
| "loss": 9.8006, |
| "step": 363 |
| }, |
| { |
| "epoch": 1.680063611392222, |
| "grad_norm": 0.16872891783714294, |
| "learning_rate": 4.398148148148149e-06, |
| "loss": 10.2161, |
| "step": 364 |
| }, |
| { |
| "epoch": 1.6846898944629176, |
| "grad_norm": 0.18189671635627747, |
| "learning_rate": 4.382716049382716e-06, |
| "loss": 9.5235, |
| "step": 365 |
| }, |
| { |
| "epoch": 1.689316177533613, |
| "grad_norm": 0.12333816289901733, |
| "learning_rate": 4.367283950617285e-06, |
| "loss": 10.2188, |
| "step": 366 |
| }, |
| { |
| "epoch": 1.6939424606043083, |
| "grad_norm": 0.22701486945152283, |
| "learning_rate": 4.351851851851852e-06, |
| "loss": 10.7382, |
| "step": 367 |
| }, |
| { |
| "epoch": 1.6985687436750037, |
| "grad_norm": 0.20510244369506836, |
| "learning_rate": 4.33641975308642e-06, |
| "loss": 8.7558, |
| "step": 368 |
| }, |
| { |
| "epoch": 1.703195026745699, |
| "grad_norm": 0.1589343398809433, |
| "learning_rate": 4.3209876543209875e-06, |
| "loss": 9.924, |
| "step": 369 |
| }, |
| { |
| "epoch": 1.7078213098163944, |
| "grad_norm": 0.1433723270893097, |
| "learning_rate": 4.305555555555556e-06, |
| "loss": 11.0008, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.7124475928870897, |
| "grad_norm": 0.17607755959033966, |
| "learning_rate": 4.290123456790124e-06, |
| "loss": 10.0096, |
| "step": 371 |
| }, |
| { |
| "epoch": 1.717073875957785, |
| "grad_norm": 0.16202040016651154, |
| "learning_rate": 4.274691358024692e-06, |
| "loss": 9.9295, |
| "step": 372 |
| }, |
| { |
| "epoch": 1.7217001590284804, |
| "grad_norm": 0.21179550886154175, |
| "learning_rate": 4.2592592592592596e-06, |
| "loss": 8.9856, |
| "step": 373 |
| }, |
| { |
| "epoch": 1.726326442099176, |
| "grad_norm": 0.35698196291923523, |
| "learning_rate": 4.243827160493827e-06, |
| "loss": 9.8793, |
| "step": 374 |
| }, |
| { |
| "epoch": 1.7309527251698713, |
| "grad_norm": 0.18081314861774445, |
| "learning_rate": 4.228395061728396e-06, |
| "loss": 10.8514, |
| "step": 375 |
| }, |
| { |
| "epoch": 1.7355790082405669, |
| "grad_norm": 0.19385862350463867, |
| "learning_rate": 4.212962962962963e-06, |
| "loss": 8.6034, |
| "step": 376 |
| }, |
| { |
| "epoch": 1.7402052913112622, |
| "grad_norm": 0.17623740434646606, |
| "learning_rate": 4.197530864197531e-06, |
| "loss": 10.7457, |
| "step": 377 |
| }, |
| { |
| "epoch": 1.7448315743819576, |
| "grad_norm": 0.1551784873008728, |
| "learning_rate": 4.182098765432099e-06, |
| "loss": 9.2378, |
| "step": 378 |
| }, |
| { |
| "epoch": 1.749457857452653, |
| "grad_norm": 0.15923646092414856, |
| "learning_rate": 4.166666666666667e-06, |
| "loss": 10.0731, |
| "step": 379 |
| }, |
| { |
| "epoch": 1.7540841405233483, |
| "grad_norm": 0.16185562312602997, |
| "learning_rate": 4.151234567901235e-06, |
| "loss": 8.9162, |
| "step": 380 |
| }, |
| { |
| "epoch": 1.7587104235940436, |
| "grad_norm": 0.23534594476222992, |
| "learning_rate": 4.135802469135803e-06, |
| "loss": 9.3805, |
| "step": 381 |
| }, |
| { |
| "epoch": 1.763336706664739, |
| "grad_norm": 0.17162498831748962, |
| "learning_rate": 4.1203703703703705e-06, |
| "loss": 10.0609, |
| "step": 382 |
| }, |
| { |
| "epoch": 1.7679629897354343, |
| "grad_norm": 0.15775223076343536, |
| "learning_rate": 4.104938271604938e-06, |
| "loss": 9.3534, |
| "step": 383 |
| }, |
| { |
| "epoch": 1.7725892728061299, |
| "grad_norm": 0.2251134216785431, |
| "learning_rate": 4.0895061728395066e-06, |
| "loss": 9.746, |
| "step": 384 |
| }, |
| { |
| "epoch": 1.7772155558768252, |
| "grad_norm": 0.19613635540008545, |
| "learning_rate": 4.074074074074074e-06, |
| "loss": 10.42, |
| "step": 385 |
| }, |
| { |
| "epoch": 1.7818418389475206, |
| "grad_norm": 0.16256873309612274, |
| "learning_rate": 4.058641975308643e-06, |
| "loss": 9.4703, |
| "step": 386 |
| }, |
| { |
| "epoch": 1.7864681220182161, |
| "grad_norm": 0.16490335762500763, |
| "learning_rate": 4.04320987654321e-06, |
| "loss": 9.5418, |
| "step": 387 |
| }, |
| { |
| "epoch": 1.7910944050889115, |
| "grad_norm": 0.159736767411232, |
| "learning_rate": 4.027777777777779e-06, |
| "loss": 10.1403, |
| "step": 388 |
| }, |
| { |
| "epoch": 1.7957206881596068, |
| "grad_norm": 0.1494799256324768, |
| "learning_rate": 4.012345679012346e-06, |
| "loss": 9.1583, |
| "step": 389 |
| }, |
| { |
| "epoch": 1.8003469712303022, |
| "grad_norm": 0.19815048575401306, |
| "learning_rate": 3.996913580246914e-06, |
| "loss": 10.2705, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.8049732543009975, |
| "grad_norm": 0.17572833597660065, |
| "learning_rate": 3.9814814814814814e-06, |
| "loss": 9.4608, |
| "step": 391 |
| }, |
| { |
| "epoch": 1.8095995373716929, |
| "grad_norm": 0.23244017362594604, |
| "learning_rate": 3.96604938271605e-06, |
| "loss": 10.6273, |
| "step": 392 |
| }, |
| { |
| "epoch": 1.8142258204423882, |
| "grad_norm": 0.17120610177516937, |
| "learning_rate": 3.9506172839506175e-06, |
| "loss": 10.0444, |
| "step": 393 |
| }, |
| { |
| "epoch": 1.8188521035130836, |
| "grad_norm": 0.27220579981803894, |
| "learning_rate": 3.935185185185186e-06, |
| "loss": 10.64, |
| "step": 394 |
| }, |
| { |
| "epoch": 1.8234783865837791, |
| "grad_norm": 0.1648477017879486, |
| "learning_rate": 3.9197530864197535e-06, |
| "loss": 9.6264, |
| "step": 395 |
| }, |
| { |
| "epoch": 1.8281046696544745, |
| "grad_norm": 0.15771055221557617, |
| "learning_rate": 3.904320987654321e-06, |
| "loss": 10.5564, |
| "step": 396 |
| }, |
| { |
| "epoch": 1.8327309527251698, |
| "grad_norm": 0.21713097393512726, |
| "learning_rate": 3.88888888888889e-06, |
| "loss": 10.6551, |
| "step": 397 |
| }, |
| { |
| "epoch": 1.8373572357958654, |
| "grad_norm": 0.17076119780540466, |
| "learning_rate": 3.873456790123457e-06, |
| "loss": 11.1632, |
| "step": 398 |
| }, |
| { |
| "epoch": 1.8419835188665608, |
| "grad_norm": 0.160685196518898, |
| "learning_rate": 3.858024691358025e-06, |
| "loss": 9.4666, |
| "step": 399 |
| }, |
| { |
| "epoch": 1.846609801937256, |
| "grad_norm": 0.17023253440856934, |
| "learning_rate": 3.842592592592592e-06, |
| "loss": 9.8118, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.8512360850079514, |
| "grad_norm": 0.2186374068260193, |
| "learning_rate": 3.827160493827161e-06, |
| "loss": 10.4903, |
| "step": 401 |
| }, |
| { |
| "epoch": 1.8558623680786468, |
| "grad_norm": 0.17954406142234802, |
| "learning_rate": 3.811728395061729e-06, |
| "loss": 11.3069, |
| "step": 402 |
| }, |
| { |
| "epoch": 1.8604886511493421, |
| "grad_norm": 0.2778474986553192, |
| "learning_rate": 3.796296296296297e-06, |
| "loss": 10.5854, |
| "step": 403 |
| }, |
| { |
| "epoch": 1.8651149342200375, |
| "grad_norm": 0.27708959579467773, |
| "learning_rate": 3.7808641975308645e-06, |
| "loss": 10.2884, |
| "step": 404 |
| }, |
| { |
| "epoch": 1.8697412172907328, |
| "grad_norm": 0.14399868249893188, |
| "learning_rate": 3.7654320987654325e-06, |
| "loss": 10.1108, |
| "step": 405 |
| }, |
| { |
| "epoch": 1.8743675003614284, |
| "grad_norm": 0.19901257753372192, |
| "learning_rate": 3.7500000000000005e-06, |
| "loss": 10.2892, |
| "step": 406 |
| }, |
| { |
| "epoch": 1.8789937834321238, |
| "grad_norm": 0.21875345706939697, |
| "learning_rate": 3.734567901234568e-06, |
| "loss": 9.0521, |
| "step": 407 |
| }, |
| { |
| "epoch": 1.883620066502819, |
| "grad_norm": 0.14913184940814972, |
| "learning_rate": 3.719135802469136e-06, |
| "loss": 10.5905, |
| "step": 408 |
| }, |
| { |
| "epoch": 1.8882463495735147, |
| "grad_norm": 0.22289720177650452, |
| "learning_rate": 3.7037037037037037e-06, |
| "loss": 9.1175, |
| "step": 409 |
| }, |
| { |
| "epoch": 1.89287263264421, |
| "grad_norm": 0.16322648525238037, |
| "learning_rate": 3.6882716049382718e-06, |
| "loss": 10.1569, |
| "step": 410 |
| }, |
| { |
| "epoch": 1.8974989157149054, |
| "grad_norm": 0.13389474153518677, |
| "learning_rate": 3.67283950617284e-06, |
| "loss": 9.7421, |
| "step": 411 |
| }, |
| { |
| "epoch": 1.9021251987856007, |
| "grad_norm": 0.17904877662658691, |
| "learning_rate": 3.657407407407408e-06, |
| "loss": 10.306, |
| "step": 412 |
| }, |
| { |
| "epoch": 1.906751481856296, |
| "grad_norm": 0.22490617632865906, |
| "learning_rate": 3.641975308641976e-06, |
| "loss": 10.0275, |
| "step": 413 |
| }, |
| { |
| "epoch": 1.9113777649269914, |
| "grad_norm": 0.20119769871234894, |
| "learning_rate": 3.6265432098765434e-06, |
| "loss": 9.5942, |
| "step": 414 |
| }, |
| { |
| "epoch": 1.9160040479976868, |
| "grad_norm": 0.156976118683815, |
| "learning_rate": 3.6111111111111115e-06, |
| "loss": 10.3435, |
| "step": 415 |
| }, |
| { |
| "epoch": 1.920630331068382, |
| "grad_norm": 0.13303594291210175, |
| "learning_rate": 3.595679012345679e-06, |
| "loss": 8.8616, |
| "step": 416 |
| }, |
| { |
| "epoch": 1.9252566141390777, |
| "grad_norm": 0.21186932921409607, |
| "learning_rate": 3.580246913580247e-06, |
| "loss": 9.8981, |
| "step": 417 |
| }, |
| { |
| "epoch": 1.929882897209773, |
| "grad_norm": 0.20669791102409363, |
| "learning_rate": 3.5648148148148147e-06, |
| "loss": 10.3689, |
| "step": 418 |
| }, |
| { |
| "epoch": 1.9345091802804684, |
| "grad_norm": 0.20095276832580566, |
| "learning_rate": 3.549382716049383e-06, |
| "loss": 9.136, |
| "step": 419 |
| }, |
| { |
| "epoch": 1.939135463351164, |
| "grad_norm": 0.16348451375961304, |
| "learning_rate": 3.533950617283951e-06, |
| "loss": 10.4419, |
| "step": 420 |
| }, |
| { |
| "epoch": 1.9437617464218593, |
| "grad_norm": 0.16439932584762573, |
| "learning_rate": 3.5185185185185187e-06, |
| "loss": 10.0692, |
| "step": 421 |
| }, |
| { |
| "epoch": 1.9483880294925546, |
| "grad_norm": 0.18562501668930054, |
| "learning_rate": 3.5030864197530868e-06, |
| "loss": 10.4452, |
| "step": 422 |
| }, |
| { |
| "epoch": 1.95301431256325, |
| "grad_norm": 0.15668661892414093, |
| "learning_rate": 3.4876543209876544e-06, |
| "loss": 9.1397, |
| "step": 423 |
| }, |
| { |
| "epoch": 1.9576405956339453, |
| "grad_norm": 0.129827082157135, |
| "learning_rate": 3.4722222222222224e-06, |
| "loss": 8.428, |
| "step": 424 |
| }, |
| { |
| "epoch": 1.9622668787046407, |
| "grad_norm": 0.14903448522090912, |
| "learning_rate": 3.4567901234567904e-06, |
| "loss": 9.0113, |
| "step": 425 |
| }, |
| { |
| "epoch": 1.966893161775336, |
| "grad_norm": 0.2033424973487854, |
| "learning_rate": 3.441358024691358e-06, |
| "loss": 8.6822, |
| "step": 426 |
| }, |
| { |
| "epoch": 1.9715194448460314, |
| "grad_norm": 0.1724868267774582, |
| "learning_rate": 3.4259259259259265e-06, |
| "loss": 9.3924, |
| "step": 427 |
| }, |
| { |
| "epoch": 1.976145727916727, |
| "grad_norm": 0.1788475513458252, |
| "learning_rate": 3.410493827160494e-06, |
| "loss": 7.8849, |
| "step": 428 |
| }, |
| { |
| "epoch": 1.9807720109874223, |
| "grad_norm": 0.13475210964679718, |
| "learning_rate": 3.395061728395062e-06, |
| "loss": 10.321, |
| "step": 429 |
| }, |
| { |
| "epoch": 1.9853982940581176, |
| "grad_norm": 0.15269149839878082, |
| "learning_rate": 3.37962962962963e-06, |
| "loss": 9.5632, |
| "step": 430 |
| }, |
| { |
| "epoch": 1.9900245771288132, |
| "grad_norm": 0.20584136247634888, |
| "learning_rate": 3.3641975308641977e-06, |
| "loss": 9.0406, |
| "step": 431 |
| }, |
| { |
| "epoch": 1.9946508601995085, |
| "grad_norm": 0.1267031729221344, |
| "learning_rate": 3.3487654320987657e-06, |
| "loss": 9.1269, |
| "step": 432 |
| }, |
| { |
| "epoch": 1.999277143270204, |
| "grad_norm": 0.20229946076869965, |
| "learning_rate": 3.3333333333333333e-06, |
| "loss": 9.1644, |
| "step": 433 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.20229946076869965, |
| "learning_rate": 3.3333333333333333e-06, |
| "loss": 1.6753, |
| "step": 434 |
| }, |
| { |
| "epoch": 2.0046262830706953, |
| "grad_norm": 0.19674813747406006, |
| "learning_rate": 3.3179012345679013e-06, |
| "loss": 8.9472, |
| "step": 435 |
| }, |
| { |
| "epoch": 2.0092525661413907, |
| "grad_norm": 0.17176903784275055, |
| "learning_rate": 3.30246913580247e-06, |
| "loss": 8.1618, |
| "step": 436 |
| }, |
| { |
| "epoch": 2.013878849212086, |
| "grad_norm": 0.21751543879508972, |
| "learning_rate": 3.2870370370370374e-06, |
| "loss": 11.298, |
| "step": 437 |
| }, |
| { |
| "epoch": 2.0185051322827814, |
| "grad_norm": 0.15177738666534424, |
| "learning_rate": 3.2716049382716054e-06, |
| "loss": 8.9182, |
| "step": 438 |
| }, |
| { |
| "epoch": 2.0231314153534767, |
| "grad_norm": 0.13782791793346405, |
| "learning_rate": 3.256172839506173e-06, |
| "loss": 9.138, |
| "step": 439 |
| }, |
| { |
| "epoch": 2.0277576984241725, |
| "grad_norm": 0.1871718168258667, |
| "learning_rate": 3.240740740740741e-06, |
| "loss": 9.3935, |
| "step": 440 |
| }, |
| { |
| "epoch": 2.032383981494868, |
| "grad_norm": 0.13550569117069244, |
| "learning_rate": 3.2253086419753086e-06, |
| "loss": 9.4065, |
| "step": 441 |
| }, |
| { |
| "epoch": 2.0370102645655632, |
| "grad_norm": 0.1414715200662613, |
| "learning_rate": 3.2098765432098767e-06, |
| "loss": 10.0772, |
| "step": 442 |
| }, |
| { |
| "epoch": 2.0416365476362586, |
| "grad_norm": 0.16839131712913513, |
| "learning_rate": 3.1944444444444443e-06, |
| "loss": 8.4968, |
| "step": 443 |
| }, |
| { |
| "epoch": 2.046262830706954, |
| "grad_norm": 0.181600421667099, |
| "learning_rate": 3.1790123456790127e-06, |
| "loss": 8.0076, |
| "step": 444 |
| }, |
| { |
| "epoch": 2.0508891137776493, |
| "grad_norm": 0.1106899231672287, |
| "learning_rate": 3.1635802469135807e-06, |
| "loss": 8.5825, |
| "step": 445 |
| }, |
| { |
| "epoch": 2.0555153968483446, |
| "grad_norm": 0.17763055860996246, |
| "learning_rate": 3.1481481481481483e-06, |
| "loss": 9.2981, |
| "step": 446 |
| }, |
| { |
| "epoch": 2.06014167991904, |
| "grad_norm": 0.19595400989055634, |
| "learning_rate": 3.1327160493827164e-06, |
| "loss": 10.5243, |
| "step": 447 |
| }, |
| { |
| "epoch": 2.0647679629897353, |
| "grad_norm": 0.18254916369915009, |
| "learning_rate": 3.1172839506172844e-06, |
| "loss": 10.9755, |
| "step": 448 |
| }, |
| { |
| "epoch": 2.0693942460604307, |
| "grad_norm": 0.18021439015865326, |
| "learning_rate": 3.101851851851852e-06, |
| "loss": 9.3634, |
| "step": 449 |
| }, |
| { |
| "epoch": 2.074020529131126, |
| "grad_norm": 0.20369423925876617, |
| "learning_rate": 3.08641975308642e-06, |
| "loss": 10.1676, |
| "step": 450 |
| }, |
| { |
| "epoch": 2.078646812201822, |
| "grad_norm": 0.1611659973859787, |
| "learning_rate": 3.0709876543209876e-06, |
| "loss": 10.5072, |
| "step": 451 |
| }, |
| { |
| "epoch": 2.083273095272517, |
| "grad_norm": 0.16623584926128387, |
| "learning_rate": 3.055555555555556e-06, |
| "loss": 8.6923, |
| "step": 452 |
| }, |
| { |
| "epoch": 2.0878993783432125, |
| "grad_norm": 0.20521709322929382, |
| "learning_rate": 3.040123456790124e-06, |
| "loss": 10.3398, |
| "step": 453 |
| }, |
| { |
| "epoch": 2.092525661413908, |
| "grad_norm": 0.150822713971138, |
| "learning_rate": 3.0246913580246917e-06, |
| "loss": 8.1734, |
| "step": 454 |
| }, |
| { |
| "epoch": 2.097151944484603, |
| "grad_norm": 0.14610332250595093, |
| "learning_rate": 3.0092592592592597e-06, |
| "loss": 8.0033, |
| "step": 455 |
| }, |
| { |
| "epoch": 2.1017782275552985, |
| "grad_norm": 0.16596634685993195, |
| "learning_rate": 2.9938271604938273e-06, |
| "loss": 9.923, |
| "step": 456 |
| }, |
| { |
| "epoch": 2.106404510625994, |
| "grad_norm": 0.15604960918426514, |
| "learning_rate": 2.9783950617283953e-06, |
| "loss": 9.5128, |
| "step": 457 |
| }, |
| { |
| "epoch": 2.1110307936966892, |
| "grad_norm": 0.1488286554813385, |
| "learning_rate": 2.962962962962963e-06, |
| "loss": 8.704, |
| "step": 458 |
| }, |
| { |
| "epoch": 2.1156570767673846, |
| "grad_norm": 0.1483355164527893, |
| "learning_rate": 2.947530864197531e-06, |
| "loss": 8.4275, |
| "step": 459 |
| }, |
| { |
| "epoch": 2.12028335983808, |
| "grad_norm": 0.2545453906059265, |
| "learning_rate": 2.9320987654320994e-06, |
| "loss": 9.332, |
| "step": 460 |
| }, |
| { |
| "epoch": 2.1249096429087757, |
| "grad_norm": 0.18272359669208527, |
| "learning_rate": 2.916666666666667e-06, |
| "loss": 9.4908, |
| "step": 461 |
| }, |
| { |
| "epoch": 2.129535925979471, |
| "grad_norm": 0.19491931796073914, |
| "learning_rate": 2.901234567901235e-06, |
| "loss": 8.5668, |
| "step": 462 |
| }, |
| { |
| "epoch": 2.1341622090501664, |
| "grad_norm": 0.171804741024971, |
| "learning_rate": 2.8858024691358026e-06, |
| "loss": 8.9085, |
| "step": 463 |
| }, |
| { |
| "epoch": 2.1387884921208618, |
| "grad_norm": 0.22477325797080994, |
| "learning_rate": 2.8703703703703706e-06, |
| "loss": 9.5643, |
| "step": 464 |
| }, |
| { |
| "epoch": 2.143414775191557, |
| "grad_norm": 0.2512359023094177, |
| "learning_rate": 2.8549382716049382e-06, |
| "loss": 10.6236, |
| "step": 465 |
| }, |
| { |
| "epoch": 2.1480410582622524, |
| "grad_norm": 0.15513500571250916, |
| "learning_rate": 2.8395061728395062e-06, |
| "loss": 9.3385, |
| "step": 466 |
| }, |
| { |
| "epoch": 2.152667341332948, |
| "grad_norm": 0.18065772950649261, |
| "learning_rate": 2.8240740740740743e-06, |
| "loss": 9.3942, |
| "step": 467 |
| }, |
| { |
| "epoch": 2.157293624403643, |
| "grad_norm": 0.19600172340869904, |
| "learning_rate": 2.8086419753086423e-06, |
| "loss": 9.0732, |
| "step": 468 |
| }, |
| { |
| "epoch": 2.1619199074743385, |
| "grad_norm": 0.17671746015548706, |
| "learning_rate": 2.7932098765432103e-06, |
| "loss": 10.023, |
| "step": 469 |
| }, |
| { |
| "epoch": 2.166546190545034, |
| "grad_norm": 0.16406480967998505, |
| "learning_rate": 2.7777777777777783e-06, |
| "loss": 9.8105, |
| "step": 470 |
| }, |
| { |
| "epoch": 2.171172473615729, |
| "grad_norm": 0.11806418746709824, |
| "learning_rate": 2.762345679012346e-06, |
| "loss": 8.2566, |
| "step": 471 |
| }, |
| { |
| "epoch": 2.175798756686425, |
| "grad_norm": 0.32428812980651855, |
| "learning_rate": 2.746913580246914e-06, |
| "loss": 9.1191, |
| "step": 472 |
| }, |
| { |
| "epoch": 2.1804250397571203, |
| "grad_norm": 0.22312945127487183, |
| "learning_rate": 2.7314814814814816e-06, |
| "loss": 9.9265, |
| "step": 473 |
| }, |
| { |
| "epoch": 2.1850513228278157, |
| "grad_norm": 0.32287222146987915, |
| "learning_rate": 2.7160493827160496e-06, |
| "loss": 10.6421, |
| "step": 474 |
| }, |
| { |
| "epoch": 2.189677605898511, |
| "grad_norm": 0.15497097373008728, |
| "learning_rate": 2.700617283950617e-06, |
| "loss": 8.9862, |
| "step": 475 |
| }, |
| { |
| "epoch": 2.1943038889692064, |
| "grad_norm": 0.13556216657161713, |
| "learning_rate": 2.6851851851851856e-06, |
| "loss": 8.658, |
| "step": 476 |
| }, |
| { |
| "epoch": 2.1989301720399017, |
| "grad_norm": 0.17666535079479218, |
| "learning_rate": 2.6697530864197536e-06, |
| "loss": 10.2978, |
| "step": 477 |
| }, |
| { |
| "epoch": 2.203556455110597, |
| "grad_norm": 0.16342324018478394, |
| "learning_rate": 2.6543209876543212e-06, |
| "loss": 8.4887, |
| "step": 478 |
| }, |
| { |
| "epoch": 2.2081827381812924, |
| "grad_norm": 0.2034367471933365, |
| "learning_rate": 2.6388888888888893e-06, |
| "loss": 7.8828, |
| "step": 479 |
| }, |
| { |
| "epoch": 2.2128090212519878, |
| "grad_norm": 0.23170992732048035, |
| "learning_rate": 2.623456790123457e-06, |
| "loss": 9.6533, |
| "step": 480 |
| }, |
| { |
| "epoch": 2.217435304322683, |
| "grad_norm": 0.16840022802352905, |
| "learning_rate": 2.608024691358025e-06, |
| "loss": 8.1141, |
| "step": 481 |
| }, |
| { |
| "epoch": 2.2220615873933784, |
| "grad_norm": 0.1619078814983368, |
| "learning_rate": 2.5925925925925925e-06, |
| "loss": 8.6294, |
| "step": 482 |
| }, |
| { |
| "epoch": 2.2266878704640742, |
| "grad_norm": 0.22204962372779846, |
| "learning_rate": 2.5771604938271605e-06, |
| "loss": 10.4391, |
| "step": 483 |
| }, |
| { |
| "epoch": 2.2313141535347696, |
| "grad_norm": 0.1876753270626068, |
| "learning_rate": 2.561728395061729e-06, |
| "loss": 9.9287, |
| "step": 484 |
| }, |
| { |
| "epoch": 2.235940436605465, |
| "grad_norm": 0.18601296842098236, |
| "learning_rate": 2.5462962962962966e-06, |
| "loss": 9.3439, |
| "step": 485 |
| }, |
| { |
| "epoch": 2.2405667196761603, |
| "grad_norm": 0.17974725365638733, |
| "learning_rate": 2.5308641975308646e-06, |
| "loss": 9.6812, |
| "step": 486 |
| }, |
| { |
| "epoch": 2.2451930027468556, |
| "grad_norm": 0.1832571029663086, |
| "learning_rate": 2.515432098765432e-06, |
| "loss": 7.6883, |
| "step": 487 |
| }, |
| { |
| "epoch": 2.249819285817551, |
| "grad_norm": 0.18652378022670746, |
| "learning_rate": 2.5e-06, |
| "loss": 8.9935, |
| "step": 488 |
| }, |
| { |
| "epoch": 2.2544455688882463, |
| "grad_norm": 0.20331954956054688, |
| "learning_rate": 2.4845679012345682e-06, |
| "loss": 9.1709, |
| "step": 489 |
| }, |
| { |
| "epoch": 2.2590718519589417, |
| "grad_norm": 0.15439340472221375, |
| "learning_rate": 2.469135802469136e-06, |
| "loss": 9.0578, |
| "step": 490 |
| }, |
| { |
| "epoch": 2.263698135029637, |
| "grad_norm": 0.16174978017807007, |
| "learning_rate": 2.453703703703704e-06, |
| "loss": 9.2311, |
| "step": 491 |
| }, |
| { |
| "epoch": 2.2683244181003324, |
| "grad_norm": 0.20344924926757812, |
| "learning_rate": 2.438271604938272e-06, |
| "loss": 9.8727, |
| "step": 492 |
| }, |
| { |
| "epoch": 2.2729507011710277, |
| "grad_norm": 0.19355060160160065, |
| "learning_rate": 2.4228395061728395e-06, |
| "loss": 9.3433, |
| "step": 493 |
| }, |
| { |
| "epoch": 2.2775769842417235, |
| "grad_norm": 0.17609727382659912, |
| "learning_rate": 2.4074074074074075e-06, |
| "loss": 8.8289, |
| "step": 494 |
| }, |
| { |
| "epoch": 2.282203267312419, |
| "grad_norm": 0.16068409383296967, |
| "learning_rate": 2.3919753086419755e-06, |
| "loss": 9.1044, |
| "step": 495 |
| }, |
| { |
| "epoch": 2.286829550383114, |
| "grad_norm": 0.15089063346385956, |
| "learning_rate": 2.3765432098765435e-06, |
| "loss": 8.8363, |
| "step": 496 |
| }, |
| { |
| "epoch": 2.2914558334538095, |
| "grad_norm": 0.3163600265979767, |
| "learning_rate": 2.361111111111111e-06, |
| "loss": 9.3244, |
| "step": 497 |
| }, |
| { |
| "epoch": 2.296082116524505, |
| "grad_norm": 0.18847279250621796, |
| "learning_rate": 2.345679012345679e-06, |
| "loss": 9.1338, |
| "step": 498 |
| }, |
| { |
| "epoch": 2.3007083995952002, |
| "grad_norm": 0.23115108907222748, |
| "learning_rate": 2.330246913580247e-06, |
| "loss": 9.6425, |
| "step": 499 |
| }, |
| { |
| "epoch": 2.3053346826658956, |
| "grad_norm": 0.15251374244689941, |
| "learning_rate": 2.314814814814815e-06, |
| "loss": 9.6591, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.309960965736591, |
| "grad_norm": 0.1706661432981491, |
| "learning_rate": 2.299382716049383e-06, |
| "loss": 10.3708, |
| "step": 501 |
| }, |
| { |
| "epoch": 2.3145872488072863, |
| "grad_norm": 0.16492144763469696, |
| "learning_rate": 2.283950617283951e-06, |
| "loss": 8.8977, |
| "step": 502 |
| }, |
| { |
| "epoch": 2.3192135318779816, |
| "grad_norm": 0.15373440086841583, |
| "learning_rate": 2.268518518518519e-06, |
| "loss": 10.9027, |
| "step": 503 |
| }, |
| { |
| "epoch": 2.323839814948677, |
| "grad_norm": 0.18270155787467957, |
| "learning_rate": 2.2530864197530865e-06, |
| "loss": 9.8633, |
| "step": 504 |
| }, |
| { |
| "epoch": 2.3284660980193728, |
| "grad_norm": 0.22537460923194885, |
| "learning_rate": 2.2376543209876545e-06, |
| "loss": 8.6262, |
| "step": 505 |
| }, |
| { |
| "epoch": 2.333092381090068, |
| "grad_norm": 0.1788664162158966, |
| "learning_rate": 2.222222222222222e-06, |
| "loss": 8.3806, |
| "step": 506 |
| }, |
| { |
| "epoch": 2.3377186641607635, |
| "grad_norm": 0.2230851948261261, |
| "learning_rate": 2.2067901234567905e-06, |
| "loss": 9.6656, |
| "step": 507 |
| }, |
| { |
| "epoch": 2.342344947231459, |
| "grad_norm": 0.157254159450531, |
| "learning_rate": 2.191358024691358e-06, |
| "loss": 9.0586, |
| "step": 508 |
| }, |
| { |
| "epoch": 2.346971230302154, |
| "grad_norm": 0.21109546720981598, |
| "learning_rate": 2.175925925925926e-06, |
| "loss": 10.4577, |
| "step": 509 |
| }, |
| { |
| "epoch": 2.3515975133728495, |
| "grad_norm": 0.16909867525100708, |
| "learning_rate": 2.1604938271604937e-06, |
| "loss": 9.5222, |
| "step": 510 |
| }, |
| { |
| "epoch": 2.356223796443545, |
| "grad_norm": 0.1652510166168213, |
| "learning_rate": 2.145061728395062e-06, |
| "loss": 9.5384, |
| "step": 511 |
| }, |
| { |
| "epoch": 2.36085007951424, |
| "grad_norm": 0.14154984056949615, |
| "learning_rate": 2.1296296296296298e-06, |
| "loss": 9.1477, |
| "step": 512 |
| }, |
| { |
| "epoch": 2.3654763625849355, |
| "grad_norm": 0.18620611727237701, |
| "learning_rate": 2.114197530864198e-06, |
| "loss": 9.8372, |
| "step": 513 |
| }, |
| { |
| "epoch": 2.370102645655631, |
| "grad_norm": 0.17792826890945435, |
| "learning_rate": 2.0987654320987654e-06, |
| "loss": 9.4475, |
| "step": 514 |
| }, |
| { |
| "epoch": 2.3747289287263262, |
| "grad_norm": 0.17278362810611725, |
| "learning_rate": 2.0833333333333334e-06, |
| "loss": 8.2241, |
| "step": 515 |
| }, |
| { |
| "epoch": 2.379355211797022, |
| "grad_norm": 0.17132383584976196, |
| "learning_rate": 2.0679012345679015e-06, |
| "loss": 10.9298, |
| "step": 516 |
| }, |
| { |
| "epoch": 2.3839814948677174, |
| "grad_norm": 0.15168847143650055, |
| "learning_rate": 2.052469135802469e-06, |
| "loss": 9.0072, |
| "step": 517 |
| }, |
| { |
| "epoch": 2.3886077779384127, |
| "grad_norm": 0.186196431517601, |
| "learning_rate": 2.037037037037037e-06, |
| "loss": 9.0923, |
| "step": 518 |
| }, |
| { |
| "epoch": 2.393234061009108, |
| "grad_norm": 0.12113353610038757, |
| "learning_rate": 2.021604938271605e-06, |
| "loss": 10.696, |
| "step": 519 |
| }, |
| { |
| "epoch": 2.3978603440798034, |
| "grad_norm": 0.25200334191322327, |
| "learning_rate": 2.006172839506173e-06, |
| "loss": 8.5775, |
| "step": 520 |
| }, |
| { |
| "epoch": 2.4024866271504988, |
| "grad_norm": 0.1717706322669983, |
| "learning_rate": 1.9907407407407407e-06, |
| "loss": 9.8223, |
| "step": 521 |
| }, |
| { |
| "epoch": 2.407112910221194, |
| "grad_norm": 0.21262796223163605, |
| "learning_rate": 1.9753086419753087e-06, |
| "loss": 9.1006, |
| "step": 522 |
| }, |
| { |
| "epoch": 2.4117391932918895, |
| "grad_norm": 0.15240874886512756, |
| "learning_rate": 1.9598765432098768e-06, |
| "loss": 8.8133, |
| "step": 523 |
| }, |
| { |
| "epoch": 2.416365476362585, |
| "grad_norm": 0.20154468715190887, |
| "learning_rate": 1.944444444444445e-06, |
| "loss": 8.9349, |
| "step": 524 |
| }, |
| { |
| "epoch": 2.42099175943328, |
| "grad_norm": 0.2465580701828003, |
| "learning_rate": 1.9290123456790124e-06, |
| "loss": 9.7463, |
| "step": 525 |
| }, |
| { |
| "epoch": 2.4256180425039755, |
| "grad_norm": 0.262588769197464, |
| "learning_rate": 1.9135802469135804e-06, |
| "loss": 8.9882, |
| "step": 526 |
| }, |
| { |
| "epoch": 2.4302443255746713, |
| "grad_norm": 0.1886514276266098, |
| "learning_rate": 1.8981481481481484e-06, |
| "loss": 10.1046, |
| "step": 527 |
| }, |
| { |
| "epoch": 2.4348706086453666, |
| "grad_norm": 0.15457630157470703, |
| "learning_rate": 1.8827160493827162e-06, |
| "loss": 9.1876, |
| "step": 528 |
| }, |
| { |
| "epoch": 2.439496891716062, |
| "grad_norm": 0.18536067008972168, |
| "learning_rate": 1.867283950617284e-06, |
| "loss": 8.9485, |
| "step": 529 |
| }, |
| { |
| "epoch": 2.4441231747867573, |
| "grad_norm": 0.16247576475143433, |
| "learning_rate": 1.8518518518518519e-06, |
| "loss": 9.7982, |
| "step": 530 |
| }, |
| { |
| "epoch": 2.4487494578574527, |
| "grad_norm": 0.13770771026611328, |
| "learning_rate": 1.83641975308642e-06, |
| "loss": 9.9136, |
| "step": 531 |
| }, |
| { |
| "epoch": 2.453375740928148, |
| "grad_norm": 0.1654641479253769, |
| "learning_rate": 1.820987654320988e-06, |
| "loss": 9.1883, |
| "step": 532 |
| }, |
| { |
| "epoch": 2.4580020239988434, |
| "grad_norm": 0.17362841963768005, |
| "learning_rate": 1.8055555555555557e-06, |
| "loss": 9.0925, |
| "step": 533 |
| }, |
| { |
| "epoch": 2.4626283070695387, |
| "grad_norm": 0.15451891720294952, |
| "learning_rate": 1.7901234567901235e-06, |
| "loss": 9.477, |
| "step": 534 |
| }, |
| { |
| "epoch": 2.467254590140234, |
| "grad_norm": 0.14848832786083221, |
| "learning_rate": 1.7746913580246916e-06, |
| "loss": 10.4986, |
| "step": 535 |
| }, |
| { |
| "epoch": 2.4718808732109294, |
| "grad_norm": 0.18573352694511414, |
| "learning_rate": 1.7592592592592594e-06, |
| "loss": 9.2772, |
| "step": 536 |
| }, |
| { |
| "epoch": 2.4765071562816248, |
| "grad_norm": 0.16295011341571808, |
| "learning_rate": 1.7438271604938272e-06, |
| "loss": 10.3985, |
| "step": 537 |
| }, |
| { |
| "epoch": 2.4811334393523206, |
| "grad_norm": 0.1890224814414978, |
| "learning_rate": 1.7283950617283952e-06, |
| "loss": 8.1752, |
| "step": 538 |
| }, |
| { |
| "epoch": 2.485759722423016, |
| "grad_norm": 0.1372375637292862, |
| "learning_rate": 1.7129629629629632e-06, |
| "loss": 8.903, |
| "step": 539 |
| }, |
| { |
| "epoch": 2.4903860054937113, |
| "grad_norm": 0.17448656260967255, |
| "learning_rate": 1.697530864197531e-06, |
| "loss": 8.5784, |
| "step": 540 |
| }, |
| { |
| "epoch": 2.4950122885644066, |
| "grad_norm": 0.1701819896697998, |
| "learning_rate": 1.6820987654320989e-06, |
| "loss": 8.7201, |
| "step": 541 |
| }, |
| { |
| "epoch": 2.499638571635102, |
| "grad_norm": 0.14620746672153473, |
| "learning_rate": 1.6666666666666667e-06, |
| "loss": 9.211, |
| "step": 542 |
| }, |
| { |
| "epoch": 2.5042648547057973, |
| "grad_norm": 0.16162265837192535, |
| "learning_rate": 1.651234567901235e-06, |
| "loss": 9.4963, |
| "step": 543 |
| }, |
| { |
| "epoch": 2.5088911377764926, |
| "grad_norm": 0.15886524319648743, |
| "learning_rate": 1.6358024691358027e-06, |
| "loss": 8.8424, |
| "step": 544 |
| }, |
| { |
| "epoch": 2.513517420847188, |
| "grad_norm": 0.14556364715099335, |
| "learning_rate": 1.6203703703703705e-06, |
| "loss": 7.9769, |
| "step": 545 |
| }, |
| { |
| "epoch": 2.5181437039178833, |
| "grad_norm": 0.21578781306743622, |
| "learning_rate": 1.6049382716049383e-06, |
| "loss": 8.9909, |
| "step": 546 |
| }, |
| { |
| "epoch": 2.522769986988579, |
| "grad_norm": 0.18850085139274597, |
| "learning_rate": 1.5895061728395064e-06, |
| "loss": 9.8237, |
| "step": 547 |
| }, |
| { |
| "epoch": 2.527396270059274, |
| "grad_norm": 0.1415654569864273, |
| "learning_rate": 1.5740740740740742e-06, |
| "loss": 8.9615, |
| "step": 548 |
| }, |
| { |
| "epoch": 2.53202255312997, |
| "grad_norm": 0.17242810130119324, |
| "learning_rate": 1.5586419753086422e-06, |
| "loss": 8.9936, |
| "step": 549 |
| }, |
| { |
| "epoch": 2.536648836200665, |
| "grad_norm": 0.20581774413585663, |
| "learning_rate": 1.54320987654321e-06, |
| "loss": 9.376, |
| "step": 550 |
| }, |
| { |
| "epoch": 2.5412751192713605, |
| "grad_norm": 0.17876863479614258, |
| "learning_rate": 1.527777777777778e-06, |
| "loss": 8.9055, |
| "step": 551 |
| }, |
| { |
| "epoch": 2.545901402342056, |
| "grad_norm": 0.1470121592283249, |
| "learning_rate": 1.5123456790123458e-06, |
| "loss": 9.3424, |
| "step": 552 |
| }, |
| { |
| "epoch": 2.550527685412751, |
| "grad_norm": 0.23757006227970123, |
| "learning_rate": 1.4969135802469136e-06, |
| "loss": 9.7866, |
| "step": 553 |
| }, |
| { |
| "epoch": 2.5551539684834466, |
| "grad_norm": 0.16553503274917603, |
| "learning_rate": 1.4814814814814815e-06, |
| "loss": 8.6868, |
| "step": 554 |
| }, |
| { |
| "epoch": 2.559780251554142, |
| "grad_norm": 0.12771116197109222, |
| "learning_rate": 1.4660493827160497e-06, |
| "loss": 8.9301, |
| "step": 555 |
| }, |
| { |
| "epoch": 2.5644065346248373, |
| "grad_norm": 0.35117506980895996, |
| "learning_rate": 1.4506172839506175e-06, |
| "loss": 9.3037, |
| "step": 556 |
| }, |
| { |
| "epoch": 2.5690328176955326, |
| "grad_norm": 0.1245264932513237, |
| "learning_rate": 1.4351851851851853e-06, |
| "loss": 9.0822, |
| "step": 557 |
| }, |
| { |
| "epoch": 2.5736591007662284, |
| "grad_norm": 0.18772459030151367, |
| "learning_rate": 1.4197530864197531e-06, |
| "loss": 9.146, |
| "step": 558 |
| }, |
| { |
| "epoch": 2.5782853838369233, |
| "grad_norm": 0.18729011714458466, |
| "learning_rate": 1.4043209876543211e-06, |
| "loss": 8.183, |
| "step": 559 |
| }, |
| { |
| "epoch": 2.582911666907619, |
| "grad_norm": 0.12940698862075806, |
| "learning_rate": 1.3888888888888892e-06, |
| "loss": 9.4745, |
| "step": 560 |
| }, |
| { |
| "epoch": 2.5875379499783144, |
| "grad_norm": 0.16963091492652893, |
| "learning_rate": 1.373456790123457e-06, |
| "loss": 8.8193, |
| "step": 561 |
| }, |
| { |
| "epoch": 2.59216423304901, |
| "grad_norm": 0.17672613263130188, |
| "learning_rate": 1.3580246913580248e-06, |
| "loss": 8.2093, |
| "step": 562 |
| }, |
| { |
| "epoch": 2.596790516119705, |
| "grad_norm": 0.24918967485427856, |
| "learning_rate": 1.3425925925925928e-06, |
| "loss": 9.2257, |
| "step": 563 |
| }, |
| { |
| "epoch": 2.6014167991904005, |
| "grad_norm": 0.15989799797534943, |
| "learning_rate": 1.3271604938271606e-06, |
| "loss": 9.08, |
| "step": 564 |
| }, |
| { |
| "epoch": 2.606043082261096, |
| "grad_norm": 0.14997775852680206, |
| "learning_rate": 1.3117283950617284e-06, |
| "loss": 9.8745, |
| "step": 565 |
| }, |
| { |
| "epoch": 2.610669365331791, |
| "grad_norm": 0.17529721558094025, |
| "learning_rate": 1.2962962962962962e-06, |
| "loss": 8.8959, |
| "step": 566 |
| }, |
| { |
| "epoch": 2.6152956484024865, |
| "grad_norm": 0.16525782644748688, |
| "learning_rate": 1.2808641975308645e-06, |
| "loss": 8.2331, |
| "step": 567 |
| }, |
| { |
| "epoch": 2.619921931473182, |
| "grad_norm": 0.14114739000797272, |
| "learning_rate": 1.2654320987654323e-06, |
| "loss": 8.1592, |
| "step": 568 |
| }, |
| { |
| "epoch": 2.6245482145438777, |
| "grad_norm": 0.21027730405330658, |
| "learning_rate": 1.25e-06, |
| "loss": 9.1182, |
| "step": 569 |
| }, |
| { |
| "epoch": 2.6291744976145726, |
| "grad_norm": 0.13383808732032776, |
| "learning_rate": 1.234567901234568e-06, |
| "loss": 8.8642, |
| "step": 570 |
| }, |
| { |
| "epoch": 2.6338007806852684, |
| "grad_norm": 0.15454170107841492, |
| "learning_rate": 1.219135802469136e-06, |
| "loss": 7.1809, |
| "step": 571 |
| }, |
| { |
| "epoch": 2.6384270637559637, |
| "grad_norm": 0.1575554460287094, |
| "learning_rate": 1.2037037037037037e-06, |
| "loss": 8.5097, |
| "step": 572 |
| }, |
| { |
| "epoch": 2.643053346826659, |
| "grad_norm": 0.16009144484996796, |
| "learning_rate": 1.1882716049382718e-06, |
| "loss": 8.8008, |
| "step": 573 |
| }, |
| { |
| "epoch": 2.6476796298973544, |
| "grad_norm": 0.18514862656593323, |
| "learning_rate": 1.1728395061728396e-06, |
| "loss": 8.3351, |
| "step": 574 |
| }, |
| { |
| "epoch": 2.6523059129680497, |
| "grad_norm": 0.16186843812465668, |
| "learning_rate": 1.1574074074074076e-06, |
| "loss": 8.702, |
| "step": 575 |
| }, |
| { |
| "epoch": 2.656932196038745, |
| "grad_norm": 0.19223704934120178, |
| "learning_rate": 1.1419753086419754e-06, |
| "loss": 9.1931, |
| "step": 576 |
| }, |
| { |
| "epoch": 2.6615584791094404, |
| "grad_norm": 0.1866094172000885, |
| "learning_rate": 1.1265432098765432e-06, |
| "loss": 9.5196, |
| "step": 577 |
| }, |
| { |
| "epoch": 2.666184762180136, |
| "grad_norm": 0.2299501746892929, |
| "learning_rate": 1.111111111111111e-06, |
| "loss": 9.4269, |
| "step": 578 |
| }, |
| { |
| "epoch": 2.670811045250831, |
| "grad_norm": 0.19736136496067047, |
| "learning_rate": 1.095679012345679e-06, |
| "loss": 9.3998, |
| "step": 579 |
| }, |
| { |
| "epoch": 2.675437328321527, |
| "grad_norm": 0.16987043619155884, |
| "learning_rate": 1.0802469135802469e-06, |
| "loss": 8.2513, |
| "step": 580 |
| }, |
| { |
| "epoch": 2.680063611392222, |
| "grad_norm": 0.18940360844135284, |
| "learning_rate": 1.0648148148148149e-06, |
| "loss": 7.8771, |
| "step": 581 |
| }, |
| { |
| "epoch": 2.6846898944629176, |
| "grad_norm": 0.16819283366203308, |
| "learning_rate": 1.0493827160493827e-06, |
| "loss": 8.9367, |
| "step": 582 |
| }, |
| { |
| "epoch": 2.689316177533613, |
| "grad_norm": 0.14170341193675995, |
| "learning_rate": 1.0339506172839507e-06, |
| "loss": 7.1555, |
| "step": 583 |
| }, |
| { |
| "epoch": 2.6939424606043083, |
| "grad_norm": 0.16584675014019012, |
| "learning_rate": 1.0185185185185185e-06, |
| "loss": 9.7405, |
| "step": 584 |
| }, |
| { |
| "epoch": 2.6985687436750037, |
| "grad_norm": 0.15231122076511383, |
| "learning_rate": 1.0030864197530866e-06, |
| "loss": 9.8814, |
| "step": 585 |
| }, |
| { |
| "epoch": 2.703195026745699, |
| "grad_norm": 0.14071619510650635, |
| "learning_rate": 9.876543209876544e-07, |
| "loss": 8.0466, |
| "step": 586 |
| }, |
| { |
| "epoch": 2.7078213098163944, |
| "grad_norm": 0.21130667626857758, |
| "learning_rate": 9.722222222222224e-07, |
| "loss": 10.1551, |
| "step": 587 |
| }, |
| { |
| "epoch": 2.7124475928870897, |
| "grad_norm": 0.14736154675483704, |
| "learning_rate": 9.567901234567902e-07, |
| "loss": 7.5648, |
| "step": 588 |
| }, |
| { |
| "epoch": 2.717073875957785, |
| "grad_norm": 0.18979448080062866, |
| "learning_rate": 9.413580246913581e-07, |
| "loss": 9.5863, |
| "step": 589 |
| }, |
| { |
| "epoch": 2.7217001590284804, |
| "grad_norm": 0.14803054928779602, |
| "learning_rate": 9.259259259259259e-07, |
| "loss": 9.2012, |
| "step": 590 |
| }, |
| { |
| "epoch": 2.726326442099176, |
| "grad_norm": 0.24644902348518372, |
| "learning_rate": 9.10493827160494e-07, |
| "loss": 7.9091, |
| "step": 591 |
| }, |
| { |
| "epoch": 2.730952725169871, |
| "grad_norm": 0.18411195278167725, |
| "learning_rate": 8.950617283950618e-07, |
| "loss": 9.0894, |
| "step": 592 |
| }, |
| { |
| "epoch": 2.735579008240567, |
| "grad_norm": 0.21505457162857056, |
| "learning_rate": 8.796296296296297e-07, |
| "loss": 7.4988, |
| "step": 593 |
| }, |
| { |
| "epoch": 2.7402052913112622, |
| "grad_norm": 0.18174538016319275, |
| "learning_rate": 8.641975308641976e-07, |
| "loss": 8.2626, |
| "step": 594 |
| }, |
| { |
| "epoch": 2.7448315743819576, |
| "grad_norm": 0.1982118785381317, |
| "learning_rate": 8.487654320987655e-07, |
| "loss": 8.415, |
| "step": 595 |
| }, |
| { |
| "epoch": 2.749457857452653, |
| "grad_norm": 0.18941333889961243, |
| "learning_rate": 8.333333333333333e-07, |
| "loss": 8.931, |
| "step": 596 |
| }, |
| { |
| "epoch": 2.7540841405233483, |
| "grad_norm": 0.2093527615070343, |
| "learning_rate": 8.179012345679014e-07, |
| "loss": 9.6871, |
| "step": 597 |
| }, |
| { |
| "epoch": 2.7587104235940436, |
| "grad_norm": 0.25371572375297546, |
| "learning_rate": 8.024691358024692e-07, |
| "loss": 9.719, |
| "step": 598 |
| }, |
| { |
| "epoch": 2.763336706664739, |
| "grad_norm": 0.13767634332180023, |
| "learning_rate": 7.870370370370371e-07, |
| "loss": 7.0931, |
| "step": 599 |
| }, |
| { |
| "epoch": 2.7679629897354343, |
| "grad_norm": 0.18252308666706085, |
| "learning_rate": 7.71604938271605e-07, |
| "loss": 9.0813, |
| "step": 600 |
| }, |
| { |
| "epoch": 2.7725892728061297, |
| "grad_norm": 0.13574370741844177, |
| "learning_rate": 7.561728395061729e-07, |
| "loss": 8.8218, |
| "step": 601 |
| }, |
| { |
| "epoch": 2.7772155558768254, |
| "grad_norm": 0.17202846705913544, |
| "learning_rate": 7.407407407407407e-07, |
| "loss": 8.954, |
| "step": 602 |
| }, |
| { |
| "epoch": 2.7818418389475204, |
| "grad_norm": 0.18582294881343842, |
| "learning_rate": 7.253086419753087e-07, |
| "loss": 7.7078, |
| "step": 603 |
| }, |
| { |
| "epoch": 2.786468122018216, |
| "grad_norm": 0.27104949951171875, |
| "learning_rate": 7.098765432098766e-07, |
| "loss": 8.3435, |
| "step": 604 |
| }, |
| { |
| "epoch": 2.7910944050889115, |
| "grad_norm": 0.32489752769470215, |
| "learning_rate": 6.944444444444446e-07, |
| "loss": 8.5035, |
| "step": 605 |
| }, |
| { |
| "epoch": 2.795720688159607, |
| "grad_norm": 0.1454378217458725, |
| "learning_rate": 6.790123456790124e-07, |
| "loss": 8.4067, |
| "step": 606 |
| }, |
| { |
| "epoch": 2.800346971230302, |
| "grad_norm": 0.20940132439136505, |
| "learning_rate": 6.635802469135803e-07, |
| "loss": 9.4978, |
| "step": 607 |
| }, |
| { |
| "epoch": 2.8049732543009975, |
| "grad_norm": 0.17158959805965424, |
| "learning_rate": 6.481481481481481e-07, |
| "loss": 9.3416, |
| "step": 608 |
| }, |
| { |
| "epoch": 2.809599537371693, |
| "grad_norm": 0.17591050267219543, |
| "learning_rate": 6.327160493827161e-07, |
| "loss": 8.7415, |
| "step": 609 |
| }, |
| { |
| "epoch": 2.8142258204423882, |
| "grad_norm": 0.18380632996559143, |
| "learning_rate": 6.17283950617284e-07, |
| "loss": 8.2376, |
| "step": 610 |
| }, |
| { |
| "epoch": 2.8188521035130836, |
| "grad_norm": 0.19443170726299286, |
| "learning_rate": 6.018518518518519e-07, |
| "loss": 7.562, |
| "step": 611 |
| }, |
| { |
| "epoch": 2.823478386583779, |
| "grad_norm": 0.24004080891609192, |
| "learning_rate": 5.864197530864198e-07, |
| "loss": 9.346, |
| "step": 612 |
| }, |
| { |
| "epoch": 2.8281046696544747, |
| "grad_norm": 0.19434408843517303, |
| "learning_rate": 5.709876543209877e-07, |
| "loss": 6.8662, |
| "step": 613 |
| }, |
| { |
| "epoch": 2.8327309527251696, |
| "grad_norm": 0.1413394808769226, |
| "learning_rate": 5.555555555555555e-07, |
| "loss": 7.7949, |
| "step": 614 |
| }, |
| { |
| "epoch": 2.8373572357958654, |
| "grad_norm": 0.21421702206134796, |
| "learning_rate": 5.401234567901234e-07, |
| "loss": 8.6497, |
| "step": 615 |
| }, |
| { |
| "epoch": 2.8419835188665608, |
| "grad_norm": 0.18243171274662018, |
| "learning_rate": 5.246913580246914e-07, |
| "loss": 9.0258, |
| "step": 616 |
| }, |
| { |
| "epoch": 2.846609801937256, |
| "grad_norm": 0.15553410351276398, |
| "learning_rate": 5.092592592592593e-07, |
| "loss": 9.0613, |
| "step": 617 |
| }, |
| { |
| "epoch": 2.8512360850079514, |
| "grad_norm": 0.1925041526556015, |
| "learning_rate": 4.938271604938272e-07, |
| "loss": 8.8279, |
| "step": 618 |
| }, |
| { |
| "epoch": 2.855862368078647, |
| "grad_norm": 0.25465813279151917, |
| "learning_rate": 4.783950617283951e-07, |
| "loss": 8.5684, |
| "step": 619 |
| }, |
| { |
| "epoch": 2.860488651149342, |
| "grad_norm": 0.17943061888217926, |
| "learning_rate": 4.6296296296296297e-07, |
| "loss": 8.9017, |
| "step": 620 |
| }, |
| { |
| "epoch": 2.8651149342200375, |
| "grad_norm": 0.22506891191005707, |
| "learning_rate": 4.475308641975309e-07, |
| "loss": 8.7791, |
| "step": 621 |
| }, |
| { |
| "epoch": 2.869741217290733, |
| "grad_norm": 0.1924242228269577, |
| "learning_rate": 4.320987654320988e-07, |
| "loss": 8.4592, |
| "step": 622 |
| }, |
| { |
| "epoch": 2.874367500361428, |
| "grad_norm": 0.1336522102355957, |
| "learning_rate": 4.1666666666666667e-07, |
| "loss": 9.2921, |
| "step": 623 |
| }, |
| { |
| "epoch": 2.878993783432124, |
| "grad_norm": 0.16631512343883514, |
| "learning_rate": 4.012345679012346e-07, |
| "loss": 8.7771, |
| "step": 624 |
| }, |
| { |
| "epoch": 2.883620066502819, |
| "grad_norm": 0.13000570237636566, |
| "learning_rate": 3.858024691358025e-07, |
| "loss": 7.6173, |
| "step": 625 |
| }, |
| { |
| "epoch": 2.8882463495735147, |
| "grad_norm": 0.16806651651859283, |
| "learning_rate": 3.7037037037037036e-07, |
| "loss": 8.2835, |
| "step": 626 |
| }, |
| { |
| "epoch": 2.89287263264421, |
| "grad_norm": 0.16624124348163605, |
| "learning_rate": 3.549382716049383e-07, |
| "loss": 9.0295, |
| "step": 627 |
| }, |
| { |
| "epoch": 2.8974989157149054, |
| "grad_norm": 0.18300163745880127, |
| "learning_rate": 3.395061728395062e-07, |
| "loss": 9.3673, |
| "step": 628 |
| }, |
| { |
| "epoch": 2.9021251987856007, |
| "grad_norm": 0.1591711938381195, |
| "learning_rate": 3.2407407407407406e-07, |
| "loss": 9.2437, |
| "step": 629 |
| }, |
| { |
| "epoch": 2.906751481856296, |
| "grad_norm": 0.1935225874185562, |
| "learning_rate": 3.08641975308642e-07, |
| "loss": 8.1388, |
| "step": 630 |
| }, |
| { |
| "epoch": 2.9113777649269914, |
| "grad_norm": 0.18095123767852783, |
| "learning_rate": 2.932098765432099e-07, |
| "loss": 8.8998, |
| "step": 631 |
| }, |
| { |
| "epoch": 2.9160040479976868, |
| "grad_norm": 0.15141603350639343, |
| "learning_rate": 2.7777777777777776e-07, |
| "loss": 9.5553, |
| "step": 632 |
| }, |
| { |
| "epoch": 2.920630331068382, |
| "grad_norm": 0.1664353758096695, |
| "learning_rate": 2.623456790123457e-07, |
| "loss": 9.2671, |
| "step": 633 |
| }, |
| { |
| "epoch": 2.9252566141390774, |
| "grad_norm": 0.14757901430130005, |
| "learning_rate": 2.469135802469136e-07, |
| "loss": 9.3753, |
| "step": 634 |
| }, |
| { |
| "epoch": 2.9298828972097732, |
| "grad_norm": 0.12832246720790863, |
| "learning_rate": 2.3148148148148148e-07, |
| "loss": 7.55, |
| "step": 635 |
| }, |
| { |
| "epoch": 2.934509180280468, |
| "grad_norm": 0.14796899259090424, |
| "learning_rate": 2.160493827160494e-07, |
| "loss": 8.9455, |
| "step": 636 |
| }, |
| { |
| "epoch": 2.939135463351164, |
| "grad_norm": 0.16375280916690826, |
| "learning_rate": 2.006172839506173e-07, |
| "loss": 9.1212, |
| "step": 637 |
| }, |
| { |
| "epoch": 2.9437617464218593, |
| "grad_norm": 0.15986521542072296, |
| "learning_rate": 1.8518518518518518e-07, |
| "loss": 8.0966, |
| "step": 638 |
| }, |
| { |
| "epoch": 2.9483880294925546, |
| "grad_norm": 0.14000752568244934, |
| "learning_rate": 1.697530864197531e-07, |
| "loss": 8.1283, |
| "step": 639 |
| }, |
| { |
| "epoch": 2.95301431256325, |
| "grad_norm": 0.10660137236118317, |
| "learning_rate": 1.54320987654321e-07, |
| "loss": 7.7013, |
| "step": 640 |
| }, |
| { |
| "epoch": 2.9576405956339453, |
| "grad_norm": 0.12879547476768494, |
| "learning_rate": 1.3888888888888888e-07, |
| "loss": 8.2472, |
| "step": 641 |
| }, |
| { |
| "epoch": 2.9622668787046407, |
| "grad_norm": 0.24027042090892792, |
| "learning_rate": 1.234567901234568e-07, |
| "loss": 9.3202, |
| "step": 642 |
| }, |
| { |
| "epoch": 2.966893161775336, |
| "grad_norm": 0.18989317119121552, |
| "learning_rate": 1.080246913580247e-07, |
| "loss": 8.6135, |
| "step": 643 |
| }, |
| { |
| "epoch": 2.9715194448460314, |
| "grad_norm": 0.20127813518047333, |
| "learning_rate": 9.259259259259259e-08, |
| "loss": 8.3557, |
| "step": 644 |
| }, |
| { |
| "epoch": 2.9761457279167267, |
| "grad_norm": 0.20816563069820404, |
| "learning_rate": 7.71604938271605e-08, |
| "loss": 9.6997, |
| "step": 645 |
| }, |
| { |
| "epoch": 2.9807720109874225, |
| "grad_norm": 0.2171671837568283, |
| "learning_rate": 6.17283950617284e-08, |
| "loss": 9.6809, |
| "step": 646 |
| }, |
| { |
| "epoch": 2.9853982940581174, |
| "grad_norm": 0.18569743633270264, |
| "learning_rate": 4.6296296296296295e-08, |
| "loss": 8.4721, |
| "step": 647 |
| }, |
| { |
| "epoch": 2.990024577128813, |
| "grad_norm": 0.16961927711963654, |
| "learning_rate": 3.08641975308642e-08, |
| "loss": 9.1963, |
| "step": 648 |
| }, |
| { |
| "epoch": 2.990024577128813, |
| "step": 648, |
| "total_flos": 2.9427853089130414e+18, |
| "train_loss": 10.740490390930647, |
| "train_runtime": 52216.2733, |
| "train_samples_per_second": 1.59, |
| "train_steps_per_second": 0.012 |
| } |
| ], |
| "logging_steps": 1.0, |
| "max_steps": 648, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 50, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2.9427853089130414e+18, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|