| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.36469730123997085, |
| "eval_steps": 500, |
| "global_step": 1000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.00036469730123997083, |
| "grad_norm": 0.2268046736717224, |
| "learning_rate": 4e-05, |
| "loss": 0.6811, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.0007293946024799417, |
| "grad_norm": 0.2326797991991043, |
| "learning_rate": 8e-05, |
| "loss": 0.5105, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.0010940919037199124, |
| "grad_norm": 0.22285917401313782, |
| "learning_rate": 0.00012, |
| "loss": 0.5868, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.0014587892049598833, |
| "grad_norm": 0.19026874005794525, |
| "learning_rate": 0.00016, |
| "loss": 0.4699, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.0018234865061998542, |
| "grad_norm": 0.2471790909767151, |
| "learning_rate": 0.0002, |
| "loss": 0.6535, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.002188183807439825, |
| "grad_norm": 0.24426604807376862, |
| "learning_rate": 0.00019992692729265622, |
| "loss": 0.5943, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.002552881108679796, |
| "grad_norm": 0.22199125587940216, |
| "learning_rate": 0.0001998538545853124, |
| "loss": 0.5346, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.0029175784099197666, |
| "grad_norm": 0.28295931220054626, |
| "learning_rate": 0.00019978078187796858, |
| "loss": 0.5446, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.0032822757111597373, |
| "grad_norm": 0.28008294105529785, |
| "learning_rate": 0.0001997077091706248, |
| "loss": 0.714, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.0036469730123997084, |
| "grad_norm": 0.23789669573307037, |
| "learning_rate": 0.00019963463646328097, |
| "loss": 0.505, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.0040116703136396795, |
| "grad_norm": 0.31461969017982483, |
| "learning_rate": 0.00019956156375593718, |
| "loss": 0.6611, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.00437636761487965, |
| "grad_norm": 0.31836387515068054, |
| "learning_rate": 0.00019948849104859336, |
| "loss": 0.5954, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.004741064916119621, |
| "grad_norm": 0.24425436556339264, |
| "learning_rate": 0.00019941541834124954, |
| "loss": 0.4776, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.005105762217359592, |
| "grad_norm": 0.2626420259475708, |
| "learning_rate": 0.00019934234563390575, |
| "loss": 0.451, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.005470459518599562, |
| "grad_norm": 0.3038848042488098, |
| "learning_rate": 0.00019926927292656196, |
| "loss": 0.7207, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.005835156819839533, |
| "grad_norm": 0.25323376059532166, |
| "learning_rate": 0.0001991962002192181, |
| "loss": 0.4251, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.006199854121079504, |
| "grad_norm": 0.3116416931152344, |
| "learning_rate": 0.00019912312751187432, |
| "loss": 0.724, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.006564551422319475, |
| "grad_norm": 0.28092512488365173, |
| "learning_rate": 0.00019905005480453053, |
| "loss": 0.6591, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.006929248723559446, |
| "grad_norm": 0.3688206076622009, |
| "learning_rate": 0.0001989769820971867, |
| "loss": 0.8551, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.007293946024799417, |
| "grad_norm": 0.2335461676120758, |
| "learning_rate": 0.00019890390938984292, |
| "loss": 0.3711, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.007658643326039387, |
| "grad_norm": 0.28437790274620056, |
| "learning_rate": 0.0001988308366824991, |
| "loss": 0.4625, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.008023340627279359, |
| "grad_norm": 0.31511029601097107, |
| "learning_rate": 0.00019875776397515528, |
| "loss": 0.6102, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.008388037928519328, |
| "grad_norm": 0.2957281768321991, |
| "learning_rate": 0.0001986846912678115, |
| "loss": 0.6009, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.0087527352297593, |
| "grad_norm": 0.26605701446533203, |
| "learning_rate": 0.00019861161856046767, |
| "loss": 0.4492, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.00911743253099927, |
| "grad_norm": 0.2744329273700714, |
| "learning_rate": 0.00019853854585312385, |
| "loss": 0.4166, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.009482129832239242, |
| "grad_norm": 0.3793030083179474, |
| "learning_rate": 0.00019846547314578006, |
| "loss": 0.7185, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.009846827133479213, |
| "grad_norm": 0.3891250789165497, |
| "learning_rate": 0.00019839240043843624, |
| "loss": 0.9231, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.010211524434719184, |
| "grad_norm": 0.4095641076564789, |
| "learning_rate": 0.00019831932773109245, |
| "loss": 0.706, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.010576221735959153, |
| "grad_norm": 0.3578352928161621, |
| "learning_rate": 0.00019824625502374866, |
| "loss": 0.5935, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.010940919037199124, |
| "grad_norm": 0.3827204406261444, |
| "learning_rate": 0.0001981731823164048, |
| "loss": 0.8403, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.011305616338439095, |
| "grad_norm": 0.3178398609161377, |
| "learning_rate": 0.00019810010960906102, |
| "loss": 0.588, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.011670313639679067, |
| "grad_norm": 0.3158668279647827, |
| "learning_rate": 0.00019802703690171723, |
| "loss": 0.5753, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.012035010940919038, |
| "grad_norm": 0.26300448179244995, |
| "learning_rate": 0.0001979539641943734, |
| "loss": 0.573, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.012399708242159009, |
| "grad_norm": 0.2760365605354309, |
| "learning_rate": 0.0001978808914870296, |
| "loss": 0.5887, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.012764405543398978, |
| "grad_norm": 0.34060901403427124, |
| "learning_rate": 0.0001978078187796858, |
| "loss": 0.7347, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.01312910284463895, |
| "grad_norm": 0.2601701319217682, |
| "learning_rate": 0.00019773474607234198, |
| "loss": 0.5684, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.01349380014587892, |
| "grad_norm": 0.33629322052001953, |
| "learning_rate": 0.0001976616733649982, |
| "loss": 0.8124, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.013858497447118891, |
| "grad_norm": 0.36825031042099, |
| "learning_rate": 0.00019758860065765437, |
| "loss": 0.8472, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.014223194748358862, |
| "grad_norm": 0.28288817405700684, |
| "learning_rate": 0.00019751552795031055, |
| "loss": 0.76, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.014587892049598834, |
| "grad_norm": 0.29537615180015564, |
| "learning_rate": 0.00019744245524296676, |
| "loss": 0.7321, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.014952589350838803, |
| "grad_norm": 0.34148740768432617, |
| "learning_rate": 0.00019736938253562297, |
| "loss": 0.8342, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.015317286652078774, |
| "grad_norm": 0.292447566986084, |
| "learning_rate": 0.00019729630982827915, |
| "loss": 0.6134, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.015681983953318747, |
| "grad_norm": 0.2850889265537262, |
| "learning_rate": 0.00019722323712093533, |
| "loss": 0.6998, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.016046681254558718, |
| "grad_norm": 0.3336108326911926, |
| "learning_rate": 0.00019715016441359154, |
| "loss": 0.8365, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.016411378555798686, |
| "grad_norm": 0.34880322217941284, |
| "learning_rate": 0.00019707709170624772, |
| "loss": 0.8137, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.016776075857038657, |
| "grad_norm": 0.27575618028640747, |
| "learning_rate": 0.00019700401899890393, |
| "loss": 0.7041, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.017140773158278628, |
| "grad_norm": 0.32200825214385986, |
| "learning_rate": 0.0001969309462915601, |
| "loss": 0.9865, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.0175054704595186, |
| "grad_norm": 0.31082266569137573, |
| "learning_rate": 0.0001968578735842163, |
| "loss": 0.7872, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.01787016776075857, |
| "grad_norm": 0.3037458658218384, |
| "learning_rate": 0.0001967848008768725, |
| "loss": 1.005, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.01823486506199854, |
| "grad_norm": 0.22701998054981232, |
| "learning_rate": 0.0001967117281695287, |
| "loss": 0.4318, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.018599562363238512, |
| "grad_norm": 0.27476680278778076, |
| "learning_rate": 0.00019663865546218486, |
| "loss": 0.6872, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.018964259664478483, |
| "grad_norm": 0.2562110424041748, |
| "learning_rate": 0.00019656558275484107, |
| "loss": 0.6356, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.019328956965718454, |
| "grad_norm": 0.2805593013763428, |
| "learning_rate": 0.00019649251004749728, |
| "loss": 0.8285, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.019693654266958426, |
| "grad_norm": 0.32811108231544495, |
| "learning_rate": 0.00019641943734015346, |
| "loss": 0.8711, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.020058351568198397, |
| "grad_norm": 0.24847714602947235, |
| "learning_rate": 0.00019634636463280967, |
| "loss": 0.6721, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.020423048869438368, |
| "grad_norm": 0.2888585031032562, |
| "learning_rate": 0.00019627329192546585, |
| "loss": 0.786, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.020787746170678335, |
| "grad_norm": 0.2474086582660675, |
| "learning_rate": 0.00019620021921812203, |
| "loss": 0.6735, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.021152443471918306, |
| "grad_norm": 0.28172221779823303, |
| "learning_rate": 0.00019612714651077824, |
| "loss": 0.7776, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.021517140773158278, |
| "grad_norm": 0.3260613977909088, |
| "learning_rate": 0.00019605407380343442, |
| "loss": 0.8941, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.02188183807439825, |
| "grad_norm": 0.2589282989501953, |
| "learning_rate": 0.0001959810010960906, |
| "loss": 0.6758, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.02224653537563822, |
| "grad_norm": 0.2978575825691223, |
| "learning_rate": 0.0001959079283887468, |
| "loss": 0.7084, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.02261123267687819, |
| "grad_norm": 0.2522169351577759, |
| "learning_rate": 0.000195834855681403, |
| "loss": 0.6634, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.022975929978118162, |
| "grad_norm": 0.3184927999973297, |
| "learning_rate": 0.0001957617829740592, |
| "loss": 0.9409, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.023340627279358133, |
| "grad_norm": 0.24009554088115692, |
| "learning_rate": 0.00019568871026671538, |
| "loss": 0.6867, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.023705324580598104, |
| "grad_norm": 0.2735375761985779, |
| "learning_rate": 0.00019561563755937157, |
| "loss": 0.6533, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.024070021881838075, |
| "grad_norm": 0.2777388095855713, |
| "learning_rate": 0.00019554256485202777, |
| "loss": 0.7517, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.024434719183078046, |
| "grad_norm": 0.271108478307724, |
| "learning_rate": 0.00019546949214468398, |
| "loss": 0.8724, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.024799416484318017, |
| "grad_norm": 0.22648799419403076, |
| "learning_rate": 0.00019539641943734016, |
| "loss": 0.523, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.02516411378555799, |
| "grad_norm": 0.2377820461988449, |
| "learning_rate": 0.00019532334672999634, |
| "loss": 0.5099, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.025528811086797956, |
| "grad_norm": 0.22092792391777039, |
| "learning_rate": 0.00019525027402265255, |
| "loss": 0.5447, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.025893508388037927, |
| "grad_norm": 0.26954007148742676, |
| "learning_rate": 0.00019517720131530873, |
| "loss": 0.6141, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.0262582056892779, |
| "grad_norm": 0.2562531530857086, |
| "learning_rate": 0.00019510412860796494, |
| "loss": 0.7991, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.02662290299051787, |
| "grad_norm": 0.28438082337379456, |
| "learning_rate": 0.00019503105590062112, |
| "loss": 0.9198, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.02698760029175784, |
| "grad_norm": 0.1985676884651184, |
| "learning_rate": 0.0001949579831932773, |
| "loss": 0.4564, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.02735229759299781, |
| "grad_norm": 0.23653608560562134, |
| "learning_rate": 0.00019488491048593351, |
| "loss": 0.6741, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.027716994894237783, |
| "grad_norm": 0.18751463294029236, |
| "learning_rate": 0.00019481183777858972, |
| "loss": 0.378, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.028081692195477754, |
| "grad_norm": 0.2654147148132324, |
| "learning_rate": 0.0001947387650712459, |
| "loss": 0.6615, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.028446389496717725, |
| "grad_norm": 0.2540780007839203, |
| "learning_rate": 0.00019466569236390208, |
| "loss": 0.661, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.028811086797957696, |
| "grad_norm": 0.2665940821170807, |
| "learning_rate": 0.0001945926196565583, |
| "loss": 0.7608, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.029175784099197667, |
| "grad_norm": 0.27822214365005493, |
| "learning_rate": 0.00019451954694921447, |
| "loss": 0.9142, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.02954048140043764, |
| "grad_norm": 0.26205846667289734, |
| "learning_rate": 0.00019444647424187068, |
| "loss": 0.7416, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.029905178701677606, |
| "grad_norm": 0.2633398175239563, |
| "learning_rate": 0.00019437340153452686, |
| "loss": 0.6727, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.030269876002917577, |
| "grad_norm": 0.2297828495502472, |
| "learning_rate": 0.00019430032882718305, |
| "loss": 0.5744, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.030634573304157548, |
| "grad_norm": 0.27555879950523376, |
| "learning_rate": 0.00019422725611983925, |
| "loss": 0.7682, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.03099927060539752, |
| "grad_norm": 0.2495211362838745, |
| "learning_rate": 0.00019415418341249546, |
| "loss": 0.7603, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.031363967906637494, |
| "grad_norm": 0.24564798176288605, |
| "learning_rate": 0.00019408111070515162, |
| "loss": 0.6745, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.03172866520787746, |
| "grad_norm": 0.2123216986656189, |
| "learning_rate": 0.00019400803799780783, |
| "loss": 0.5624, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.032093362509117436, |
| "grad_norm": 0.26791101694107056, |
| "learning_rate": 0.00019393496529046403, |
| "loss": 0.8236, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.032458059810357404, |
| "grad_norm": 0.24294224381446838, |
| "learning_rate": 0.00019386189258312021, |
| "loss": 0.7493, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.03282275711159737, |
| "grad_norm": 0.25676801800727844, |
| "learning_rate": 0.00019378881987577642, |
| "loss": 0.8281, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.033187454412837346, |
| "grad_norm": 0.22630107402801514, |
| "learning_rate": 0.0001937157471684326, |
| "loss": 0.6335, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.03355215171407731, |
| "grad_norm": 0.25208771228790283, |
| "learning_rate": 0.00019364267446108879, |
| "loss": 0.7753, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.03391684901531729, |
| "grad_norm": 0.23288992047309875, |
| "learning_rate": 0.000193569601753745, |
| "loss": 0.683, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.034281546316557256, |
| "grad_norm": 0.2340337187051773, |
| "learning_rate": 0.00019349652904640118, |
| "loss": 0.7111, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.03464624361779723, |
| "grad_norm": 0.24910978972911835, |
| "learning_rate": 0.00019342345633905736, |
| "loss": 0.8183, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.0350109409190372, |
| "grad_norm": 0.23724719882011414, |
| "learning_rate": 0.00019335038363171357, |
| "loss": 0.7408, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.03537563822027717, |
| "grad_norm": 0.2057395875453949, |
| "learning_rate": 0.00019327731092436975, |
| "loss": 0.5099, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.03574033552151714, |
| "grad_norm": 0.22540345788002014, |
| "learning_rate": 0.00019320423821702595, |
| "loss": 0.7391, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.036105032822757115, |
| "grad_norm": 0.2615845799446106, |
| "learning_rate": 0.00019313116550968214, |
| "loss": 0.7005, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.03646973012399708, |
| "grad_norm": 0.19165730476379395, |
| "learning_rate": 0.00019305809280233832, |
| "loss": 0.4355, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.03683442742523705, |
| "grad_norm": 0.22737336158752441, |
| "learning_rate": 0.00019298502009499453, |
| "loss": 0.7579, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.037199124726477024, |
| "grad_norm": 0.21350006759166718, |
| "learning_rate": 0.00019291194738765073, |
| "loss": 0.6608, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.03756382202771699, |
| "grad_norm": 0.24027219414710999, |
| "learning_rate": 0.00019283887468030692, |
| "loss": 0.8235, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.03792851932895697, |
| "grad_norm": 0.21983422338962555, |
| "learning_rate": 0.0001927658019729631, |
| "loss": 0.5319, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.038293216630196934, |
| "grad_norm": 0.26618310809135437, |
| "learning_rate": 0.0001926927292656193, |
| "loss": 0.9259, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.03865791393143691, |
| "grad_norm": 0.24729526042938232, |
| "learning_rate": 0.0001926196565582755, |
| "loss": 0.7653, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.039022611232676876, |
| "grad_norm": 0.2595866024494171, |
| "learning_rate": 0.0001925465838509317, |
| "loss": 0.7662, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.03938730853391685, |
| "grad_norm": 0.16923396289348602, |
| "learning_rate": 0.00019247351114358788, |
| "loss": 0.3119, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.03975200583515682, |
| "grad_norm": 0.2592317759990692, |
| "learning_rate": 0.00019240043843624406, |
| "loss": 0.6767, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.04011670313639679, |
| "grad_norm": 0.2139894813299179, |
| "learning_rate": 0.00019232736572890027, |
| "loss": 0.5058, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.04048140043763676, |
| "grad_norm": 0.2439870685338974, |
| "learning_rate": 0.00019225429302155647, |
| "loss": 0.807, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.040846097738876735, |
| "grad_norm": 0.26212504506111145, |
| "learning_rate": 0.00019218122031421263, |
| "loss": 0.9942, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.0412107950401167, |
| "grad_norm": 0.26018884778022766, |
| "learning_rate": 0.00019210814760686884, |
| "loss": 0.6573, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.04157549234135667, |
| "grad_norm": 0.20158423483371735, |
| "learning_rate": 0.00019203507489952505, |
| "loss": 0.4533, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.041940189642596645, |
| "grad_norm": 0.2270892709493637, |
| "learning_rate": 0.00019196200219218123, |
| "loss": 0.5983, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.04230488694383661, |
| "grad_norm": 0.2140335738658905, |
| "learning_rate": 0.00019188892948483744, |
| "loss": 0.5686, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.04266958424507659, |
| "grad_norm": 0.25761061906814575, |
| "learning_rate": 0.00019181585677749362, |
| "loss": 0.7238, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.043034281546316555, |
| "grad_norm": 0.24370697140693665, |
| "learning_rate": 0.0001917427840701498, |
| "loss": 0.7342, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.04339897884755653, |
| "grad_norm": 0.2632579207420349, |
| "learning_rate": 0.000191669711362806, |
| "loss": 0.8978, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.0437636761487965, |
| "grad_norm": 0.22956986725330353, |
| "learning_rate": 0.00019159663865546221, |
| "loss": 0.6409, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.04412837345003647, |
| "grad_norm": 0.24121011793613434, |
| "learning_rate": 0.00019152356594811837, |
| "loss": 0.7556, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.04449307075127644, |
| "grad_norm": 0.2375144511461258, |
| "learning_rate": 0.00019145049324077458, |
| "loss": 0.7303, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.044857768052516414, |
| "grad_norm": 0.22452694177627563, |
| "learning_rate": 0.00019137742053343079, |
| "loss": 0.6456, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.04522246535375638, |
| "grad_norm": 0.2590137720108032, |
| "learning_rate": 0.00019130434782608697, |
| "loss": 0.7438, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.045587162654996356, |
| "grad_norm": 0.2955920696258545, |
| "learning_rate": 0.00019123127511874318, |
| "loss": 1.0269, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.045951859956236324, |
| "grad_norm": 0.21322080492973328, |
| "learning_rate": 0.00019115820241139936, |
| "loss": 0.6143, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.04631655725747629, |
| "grad_norm": 0.20497886836528778, |
| "learning_rate": 0.00019108512970405554, |
| "loss": 0.5871, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.046681254558716266, |
| "grad_norm": 0.26355209946632385, |
| "learning_rate": 0.00019101205699671175, |
| "loss": 0.8077, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.047045951859956234, |
| "grad_norm": 0.21869653463363647, |
| "learning_rate": 0.00019093898428936793, |
| "loss": 0.6736, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.04741064916119621, |
| "grad_norm": 0.21741290390491486, |
| "learning_rate": 0.0001908659115820241, |
| "loss": 0.597, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.047775346462436176, |
| "grad_norm": 0.2503755986690521, |
| "learning_rate": 0.00019079283887468032, |
| "loss": 0.8038, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.04814004376367615, |
| "grad_norm": 0.2087285965681076, |
| "learning_rate": 0.0001907197661673365, |
| "loss": 0.5409, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.04850474106491612, |
| "grad_norm": 0.2347562611103058, |
| "learning_rate": 0.0001906466934599927, |
| "loss": 0.6741, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.04886943836615609, |
| "grad_norm": 0.23479056358337402, |
| "learning_rate": 0.0001905736207526489, |
| "loss": 0.7257, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.04923413566739606, |
| "grad_norm": 0.2217235416173935, |
| "learning_rate": 0.00019050054804530507, |
| "loss": 0.6539, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.049598832968636035, |
| "grad_norm": 0.23342272639274597, |
| "learning_rate": 0.00019042747533796128, |
| "loss": 0.6716, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.049963530269876, |
| "grad_norm": 0.231741800904274, |
| "learning_rate": 0.00019035440263061749, |
| "loss": 0.6563, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.05032822757111598, |
| "grad_norm": 0.2353263646364212, |
| "learning_rate": 0.00019028132992327367, |
| "loss": 0.7612, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.050692924872355945, |
| "grad_norm": 0.21605569124221802, |
| "learning_rate": 0.00019020825721592985, |
| "loss": 0.6007, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.05105762217359591, |
| "grad_norm": 0.25579389929771423, |
| "learning_rate": 0.00019013518450858606, |
| "loss": 0.8586, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.05142231947483589, |
| "grad_norm": 0.24197165668010712, |
| "learning_rate": 0.00019006211180124224, |
| "loss": 0.7442, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.051787016776075855, |
| "grad_norm": 0.21442811191082, |
| "learning_rate": 0.00018998903909389845, |
| "loss": 0.6356, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.05215171407731583, |
| "grad_norm": 0.23524148762226105, |
| "learning_rate": 0.00018991596638655463, |
| "loss": 0.8438, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.0525164113785558, |
| "grad_norm": 0.1977744698524475, |
| "learning_rate": 0.0001898428936792108, |
| "loss": 0.6249, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.05288110867979577, |
| "grad_norm": 0.22285562753677368, |
| "learning_rate": 0.00018976982097186702, |
| "loss": 0.6603, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.05324580598103574, |
| "grad_norm": 0.223031684756279, |
| "learning_rate": 0.00018969674826452323, |
| "loss": 0.7324, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.053610503282275714, |
| "grad_norm": 0.200596883893013, |
| "learning_rate": 0.00018962367555717938, |
| "loss": 0.546, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.05397520058351568, |
| "grad_norm": 0.23952823877334595, |
| "learning_rate": 0.0001895506028498356, |
| "loss": 0.909, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.054339897884755656, |
| "grad_norm": 0.15994083881378174, |
| "learning_rate": 0.0001894775301424918, |
| "loss": 0.3073, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.05470459518599562, |
| "grad_norm": 0.2209351807832718, |
| "learning_rate": 0.00018940445743514798, |
| "loss": 0.6757, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.05506929248723559, |
| "grad_norm": 0.19749605655670166, |
| "learning_rate": 0.0001893313847278042, |
| "loss": 0.5071, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.055433989788475566, |
| "grad_norm": 0.23921646177768707, |
| "learning_rate": 0.00018925831202046037, |
| "loss": 0.7567, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.05579868708971553, |
| "grad_norm": 0.21612580120563507, |
| "learning_rate": 0.00018918523931311655, |
| "loss": 0.6329, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.05616338439095551, |
| "grad_norm": 0.21304431557655334, |
| "learning_rate": 0.00018911216660577276, |
| "loss": 0.6612, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.056528081692195475, |
| "grad_norm": 0.23509572446346283, |
| "learning_rate": 0.00018903909389842897, |
| "loss": 0.6723, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.05689277899343545, |
| "grad_norm": 0.23066239058971405, |
| "learning_rate": 0.00018896602119108512, |
| "loss": 0.7588, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.05725747629467542, |
| "grad_norm": 0.2865854799747467, |
| "learning_rate": 0.00018889294848374133, |
| "loss": 1.0649, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.05762217359591539, |
| "grad_norm": 0.21136869490146637, |
| "learning_rate": 0.00018881987577639754, |
| "loss": 0.6532, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.05798687089715536, |
| "grad_norm": 0.2841559648513794, |
| "learning_rate": 0.00018874680306905372, |
| "loss": 1.047, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.058351568198395334, |
| "grad_norm": 0.23411215841770172, |
| "learning_rate": 0.0001886737303617099, |
| "loss": 0.6954, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.0587162654996353, |
| "grad_norm": 0.25536251068115234, |
| "learning_rate": 0.0001886006576543661, |
| "loss": 0.8371, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.05908096280087528, |
| "grad_norm": 0.24174682796001434, |
| "learning_rate": 0.0001885275849470223, |
| "loss": 0.6944, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.059445660102115244, |
| "grad_norm": 0.21550963819026947, |
| "learning_rate": 0.0001884545122396785, |
| "loss": 0.6978, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.05981035740335521, |
| "grad_norm": 0.1993122398853302, |
| "learning_rate": 0.00018838143953233468, |
| "loss": 0.4768, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.060175054704595186, |
| "grad_norm": 0.22367194294929504, |
| "learning_rate": 0.00018830836682499086, |
| "loss": 0.6238, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.060539752005835154, |
| "grad_norm": 0.20643705129623413, |
| "learning_rate": 0.00018823529411764707, |
| "loss": 0.598, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.06090444930707513, |
| "grad_norm": 0.21671368181705475, |
| "learning_rate": 0.00018816222141030325, |
| "loss": 0.7071, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.061269146608315096, |
| "grad_norm": 0.19023250043392181, |
| "learning_rate": 0.00018808914870295946, |
| "loss": 0.4923, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.06163384390955507, |
| "grad_norm": 0.26334115862846375, |
| "learning_rate": 0.00018801607599561564, |
| "loss": 0.8478, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.06199854121079504, |
| "grad_norm": 0.2204032838344574, |
| "learning_rate": 0.00018794300328827182, |
| "loss": 0.6593, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.06236323851203501, |
| "grad_norm": 0.20991367101669312, |
| "learning_rate": 0.00018786993058092803, |
| "loss": 0.6712, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.06272793581327499, |
| "grad_norm": 0.2253231555223465, |
| "learning_rate": 0.00018779685787358424, |
| "loss": 0.6589, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.06309263311451495, |
| "grad_norm": 0.1936040222644806, |
| "learning_rate": 0.00018772378516624042, |
| "loss": 0.515, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.06345733041575492, |
| "grad_norm": 0.24025693535804749, |
| "learning_rate": 0.0001876507124588966, |
| "loss": 0.8256, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.0638220277169949, |
| "grad_norm": 0.2465592622756958, |
| "learning_rate": 0.0001875776397515528, |
| "loss": 0.9898, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.06418672501823487, |
| "grad_norm": 0.20442670583724976, |
| "learning_rate": 0.000187504567044209, |
| "loss": 0.6724, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.06455142231947483, |
| "grad_norm": 0.19670630991458893, |
| "learning_rate": 0.0001874314943368652, |
| "loss": 0.6449, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.06491611962071481, |
| "grad_norm": 0.1966039091348648, |
| "learning_rate": 0.00018735842162952138, |
| "loss": 0.787, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.06528081692195478, |
| "grad_norm": 0.21741807460784912, |
| "learning_rate": 0.00018728534892217756, |
| "loss": 0.8073, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.06564551422319474, |
| "grad_norm": 0.20191621780395508, |
| "learning_rate": 0.00018721227621483377, |
| "loss": 0.6297, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.06601021152443472, |
| "grad_norm": 0.21392996609210968, |
| "learning_rate": 0.00018713920350748998, |
| "loss": 0.6611, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.06637490882567469, |
| "grad_norm": 0.21892090141773224, |
| "learning_rate": 0.00018706613080014613, |
| "loss": 0.7913, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.06673960612691467, |
| "grad_norm": 0.240064337849617, |
| "learning_rate": 0.00018699305809280234, |
| "loss": 0.74, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.06710430342815463, |
| "grad_norm": 0.2156010866165161, |
| "learning_rate": 0.00018691998538545855, |
| "loss": 0.6928, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.0674690007293946, |
| "grad_norm": 0.19305868446826935, |
| "learning_rate": 0.00018684691267811473, |
| "loss": 0.6712, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.06783369803063458, |
| "grad_norm": 0.24134770035743713, |
| "learning_rate": 0.00018677383997077094, |
| "loss": 0.7784, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.06819839533187455, |
| "grad_norm": 0.21951039135456085, |
| "learning_rate": 0.00018670076726342712, |
| "loss": 0.7536, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.06856309263311451, |
| "grad_norm": 0.22668084502220154, |
| "learning_rate": 0.0001866276945560833, |
| "loss": 0.7914, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.06892778993435449, |
| "grad_norm": 0.23703083395957947, |
| "learning_rate": 0.0001865546218487395, |
| "loss": 0.8562, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.06929248723559446, |
| "grad_norm": 0.18408019840717316, |
| "learning_rate": 0.00018648154914139572, |
| "loss": 0.5408, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.06965718453683442, |
| "grad_norm": 0.168971985578537, |
| "learning_rate": 0.00018640847643405187, |
| "loss": 0.5145, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.0700218818380744, |
| "grad_norm": 0.2313617318868637, |
| "learning_rate": 0.00018633540372670808, |
| "loss": 0.7116, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.07038657913931437, |
| "grad_norm": 0.22079357504844666, |
| "learning_rate": 0.0001862623310193643, |
| "loss": 0.7453, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.07075127644055434, |
| "grad_norm": 0.20398060977458954, |
| "learning_rate": 0.00018618925831202047, |
| "loss": 0.6914, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.0711159737417943, |
| "grad_norm": 0.21486474573612213, |
| "learning_rate": 0.00018611618560467665, |
| "loss": 0.7252, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.07148067104303428, |
| "grad_norm": 0.21152055263519287, |
| "learning_rate": 0.00018604311289733286, |
| "loss": 0.676, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.07184536834427425, |
| "grad_norm": 0.2242351919412613, |
| "learning_rate": 0.00018597004018998904, |
| "loss": 0.8428, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.07221006564551423, |
| "grad_norm": 0.19831842184066772, |
| "learning_rate": 0.00018589696748264525, |
| "loss": 0.6379, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.07257476294675419, |
| "grad_norm": 0.2057400494813919, |
| "learning_rate": 0.00018582389477530143, |
| "loss": 0.7417, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.07293946024799416, |
| "grad_norm": 0.17779530584812164, |
| "learning_rate": 0.0001857508220679576, |
| "loss": 0.6891, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.07330415754923414, |
| "grad_norm": 0.20354579389095306, |
| "learning_rate": 0.00018567774936061382, |
| "loss": 0.7052, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.0736688548504741, |
| "grad_norm": 0.17558813095092773, |
| "learning_rate": 0.00018560467665327, |
| "loss": 0.6295, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.07403355215171407, |
| "grad_norm": 0.1900980919599533, |
| "learning_rate": 0.0001855316039459262, |
| "loss": 0.6312, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.07439824945295405, |
| "grad_norm": 0.2224852293729782, |
| "learning_rate": 0.0001854585312385824, |
| "loss": 0.7633, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.07476294675419402, |
| "grad_norm": 0.2300332933664322, |
| "learning_rate": 0.00018538545853123857, |
| "loss": 0.8564, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.07512764405543398, |
| "grad_norm": 0.24671219289302826, |
| "learning_rate": 0.00018531238582389478, |
| "loss": 0.9927, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.07549234135667396, |
| "grad_norm": 0.1760961413383484, |
| "learning_rate": 0.000185239313116551, |
| "loss": 0.5943, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.07585703865791393, |
| "grad_norm": 0.25457900762557983, |
| "learning_rate": 0.00018516624040920717, |
| "loss": 1.0929, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.07622173595915391, |
| "grad_norm": 0.1756330132484436, |
| "learning_rate": 0.00018509316770186335, |
| "loss": 0.5191, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.07658643326039387, |
| "grad_norm": 0.22464942932128906, |
| "learning_rate": 0.00018502009499451956, |
| "loss": 0.9786, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.07695113056163384, |
| "grad_norm": 0.17986498773097992, |
| "learning_rate": 0.00018494702228717574, |
| "loss": 0.5692, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.07731582786287382, |
| "grad_norm": 0.20685848593711853, |
| "learning_rate": 0.00018487394957983195, |
| "loss": 0.7599, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.07768052516411379, |
| "grad_norm": 0.19135037064552307, |
| "learning_rate": 0.00018480087687248813, |
| "loss": 0.5762, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.07804522246535375, |
| "grad_norm": 0.18071472644805908, |
| "learning_rate": 0.00018472780416514431, |
| "loss": 0.4915, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.07840991976659373, |
| "grad_norm": 0.2219233512878418, |
| "learning_rate": 0.00018465473145780052, |
| "loss": 0.9144, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.0787746170678337, |
| "grad_norm": 0.20180021226406097, |
| "learning_rate": 0.00018458165875045673, |
| "loss": 0.753, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.07913931436907366, |
| "grad_norm": 0.18585549294948578, |
| "learning_rate": 0.00018450858604311289, |
| "loss": 0.6177, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.07950401167031364, |
| "grad_norm": 0.21946971118450165, |
| "learning_rate": 0.0001844355133357691, |
| "loss": 0.8311, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.07986870897155361, |
| "grad_norm": 0.22317932546138763, |
| "learning_rate": 0.0001843624406284253, |
| "loss": 0.831, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.08023340627279359, |
| "grad_norm": 0.20973660051822662, |
| "learning_rate": 0.00018428936792108148, |
| "loss": 0.718, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.08059810357403355, |
| "grad_norm": 0.19049975275993347, |
| "learning_rate": 0.0001842162952137377, |
| "loss": 0.6143, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.08096280087527352, |
| "grad_norm": 0.22633568942546844, |
| "learning_rate": 0.00018414322250639387, |
| "loss": 0.7479, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.0813274981765135, |
| "grad_norm": 0.22375738620758057, |
| "learning_rate": 0.00018407014979905005, |
| "loss": 0.783, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.08169219547775347, |
| "grad_norm": 0.20708216726779938, |
| "learning_rate": 0.00018399707709170626, |
| "loss": 0.7068, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.08205689277899343, |
| "grad_norm": 0.2057557851076126, |
| "learning_rate": 0.00018392400438436247, |
| "loss": 0.7246, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.0824215900802334, |
| "grad_norm": 0.21535396575927734, |
| "learning_rate": 0.00018385093167701863, |
| "loss": 0.6725, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.08278628738147338, |
| "grad_norm": 0.17876707017421722, |
| "learning_rate": 0.00018377785896967483, |
| "loss": 0.5595, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.08315098468271334, |
| "grad_norm": 0.17822523415088654, |
| "learning_rate": 0.00018370478626233104, |
| "loss": 0.5312, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.08351568198395332, |
| "grad_norm": 0.24496309459209442, |
| "learning_rate": 0.00018363171355498722, |
| "loss": 1.0328, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.08388037928519329, |
| "grad_norm": 0.2105628103017807, |
| "learning_rate": 0.0001835586408476434, |
| "loss": 0.6597, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.08424507658643327, |
| "grad_norm": 0.1954609751701355, |
| "learning_rate": 0.0001834855681402996, |
| "loss": 0.7691, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.08460977388767323, |
| "grad_norm": 0.18443147838115692, |
| "learning_rate": 0.0001834124954329558, |
| "loss": 0.634, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.0849744711889132, |
| "grad_norm": 0.20477977395057678, |
| "learning_rate": 0.000183339422725612, |
| "loss": 0.6758, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.08533916849015317, |
| "grad_norm": 0.21266460418701172, |
| "learning_rate": 0.00018326635001826818, |
| "loss": 0.9446, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.08570386579139315, |
| "grad_norm": 0.2169213891029358, |
| "learning_rate": 0.00018319327731092437, |
| "loss": 0.755, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.08606856309263311, |
| "grad_norm": 0.19145172834396362, |
| "learning_rate": 0.00018312020460358057, |
| "loss": 0.7846, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.08643326039387308, |
| "grad_norm": 0.1787746399641037, |
| "learning_rate": 0.00018304713189623676, |
| "loss": 0.6359, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.08679795769511306, |
| "grad_norm": 0.21680796146392822, |
| "learning_rate": 0.00018297405918889296, |
| "loss": 0.6408, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.08716265499635302, |
| "grad_norm": 0.20149710774421692, |
| "learning_rate": 0.00018290098648154914, |
| "loss": 0.695, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.087527352297593, |
| "grad_norm": 0.20793262124061584, |
| "learning_rate": 0.00018282791377420533, |
| "loss": 0.7159, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.08789204959883297, |
| "grad_norm": 0.21236906945705414, |
| "learning_rate": 0.00018275484106686153, |
| "loss": 0.8924, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.08825674690007294, |
| "grad_norm": 0.17754322290420532, |
| "learning_rate": 0.00018268176835951774, |
| "loss": 0.5931, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.0886214442013129, |
| "grad_norm": 0.21361956000328064, |
| "learning_rate": 0.00018260869565217392, |
| "loss": 0.852, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.08898614150255288, |
| "grad_norm": 0.19755081832408905, |
| "learning_rate": 0.0001825356229448301, |
| "loss": 0.7846, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.08935083880379285, |
| "grad_norm": 0.17052070796489716, |
| "learning_rate": 0.00018246255023748631, |
| "loss": 0.4757, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.08971553610503283, |
| "grad_norm": 0.17633679509162903, |
| "learning_rate": 0.0001823894775301425, |
| "loss": 0.3075, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.09008023340627279, |
| "grad_norm": 0.18541620671749115, |
| "learning_rate": 0.0001823164048227987, |
| "loss": 0.6268, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.09044493070751276, |
| "grad_norm": 0.18999099731445312, |
| "learning_rate": 0.00018224333211545488, |
| "loss": 0.6236, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.09080962800875274, |
| "grad_norm": 0.17926354706287384, |
| "learning_rate": 0.00018217025940811107, |
| "loss": 0.6153, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.09117432530999271, |
| "grad_norm": 0.161162868142128, |
| "learning_rate": 0.00018209718670076727, |
| "loss": 0.436, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.09153902261123267, |
| "grad_norm": 0.17199409008026123, |
| "learning_rate": 0.00018202411399342348, |
| "loss": 0.5008, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.09190371991247265, |
| "grad_norm": 0.20446273684501648, |
| "learning_rate": 0.00018195104128607964, |
| "loss": 0.8099, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.09226841721371262, |
| "grad_norm": 0.22668351233005524, |
| "learning_rate": 0.00018187796857873585, |
| "loss": 0.8039, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.09263311451495258, |
| "grad_norm": 0.20580340921878815, |
| "learning_rate": 0.00018180489587139205, |
| "loss": 0.8551, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.09299781181619256, |
| "grad_norm": 0.20698761940002441, |
| "learning_rate": 0.00018173182316404824, |
| "loss": 0.7958, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.09336250911743253, |
| "grad_norm": 0.18675172328948975, |
| "learning_rate": 0.00018165875045670444, |
| "loss": 0.6952, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.0937272064186725, |
| "grad_norm": 0.21192388236522675, |
| "learning_rate": 0.00018158567774936063, |
| "loss": 0.906, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.09409190371991247, |
| "grad_norm": 0.20454420149326324, |
| "learning_rate": 0.0001815126050420168, |
| "loss": 0.8052, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.09445660102115244, |
| "grad_norm": 0.2223392277956009, |
| "learning_rate": 0.00018143953233467301, |
| "loss": 1.0307, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.09482129832239242, |
| "grad_norm": 0.1847476363182068, |
| "learning_rate": 0.00018136645962732922, |
| "loss": 0.7384, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.09518599562363239, |
| "grad_norm": 0.1858607977628708, |
| "learning_rate": 0.00018129338691998538, |
| "loss": 0.6157, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.09555069292487235, |
| "grad_norm": 0.19958609342575073, |
| "learning_rate": 0.00018122031421264159, |
| "loss": 0.7196, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.09591539022611233, |
| "grad_norm": 0.22875526547431946, |
| "learning_rate": 0.0001811472415052978, |
| "loss": 0.864, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.0962800875273523, |
| "grad_norm": 0.15967297554016113, |
| "learning_rate": 0.00018107416879795398, |
| "loss": 0.5185, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.09664478482859226, |
| "grad_norm": 0.15776892006397247, |
| "learning_rate": 0.00018100109609061016, |
| "loss": 0.5136, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.09700948212983224, |
| "grad_norm": 0.17968744039535522, |
| "learning_rate": 0.00018092802338326637, |
| "loss": 0.6538, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.09737417943107221, |
| "grad_norm": 0.19834838807582855, |
| "learning_rate": 0.00018085495067592255, |
| "loss": 0.7448, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.09773887673231219, |
| "grad_norm": 0.18585754930973053, |
| "learning_rate": 0.00018078187796857875, |
| "loss": 0.6166, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.09810357403355215, |
| "grad_norm": 0.21760611236095428, |
| "learning_rate": 0.00018070880526123494, |
| "loss": 0.9431, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.09846827133479212, |
| "grad_norm": 0.17871254682540894, |
| "learning_rate": 0.00018063573255389112, |
| "loss": 0.5939, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.0988329686360321, |
| "grad_norm": 0.16932818293571472, |
| "learning_rate": 0.00018056265984654733, |
| "loss": 0.5615, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.09919766593727207, |
| "grad_norm": 0.18163429200649261, |
| "learning_rate": 0.0001804895871392035, |
| "loss": 0.6572, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.09956236323851203, |
| "grad_norm": 0.17625723779201508, |
| "learning_rate": 0.00018041651443185972, |
| "loss": 0.5748, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.099927060539752, |
| "grad_norm": 0.19214142858982086, |
| "learning_rate": 0.0001803434417245159, |
| "loss": 0.7589, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.10029175784099198, |
| "grad_norm": 0.19191130995750427, |
| "learning_rate": 0.00018027036901717208, |
| "loss": 0.7201, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.10065645514223195, |
| "grad_norm": 0.18246126174926758, |
| "learning_rate": 0.0001801972963098283, |
| "loss": 0.7919, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.10102115244347191, |
| "grad_norm": 0.17984451353549957, |
| "learning_rate": 0.0001801242236024845, |
| "loss": 0.6746, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.10138584974471189, |
| "grad_norm": 0.18245050311088562, |
| "learning_rate": 0.00018005115089514068, |
| "loss": 0.5672, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.10175054704595186, |
| "grad_norm": 0.1899084746837616, |
| "learning_rate": 0.00017997807818779686, |
| "loss": 0.6676, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.10211524434719182, |
| "grad_norm": 0.21027837693691254, |
| "learning_rate": 0.00017990500548045307, |
| "loss": 0.8676, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.1024799416484318, |
| "grad_norm": 0.178825244307518, |
| "learning_rate": 0.00017983193277310925, |
| "loss": 0.5617, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.10284463894967177, |
| "grad_norm": 0.18551243841648102, |
| "learning_rate": 0.00017975886006576546, |
| "loss": 0.7091, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.10320933625091175, |
| "grad_norm": 0.18075218796730042, |
| "learning_rate": 0.00017968578735842164, |
| "loss": 0.7194, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.10357403355215171, |
| "grad_norm": 0.1333538293838501, |
| "learning_rate": 0.00017961271465107782, |
| "loss": 0.3161, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.10393873085339168, |
| "grad_norm": 0.18160590529441833, |
| "learning_rate": 0.00017953964194373403, |
| "loss": 0.7355, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.10430342815463166, |
| "grad_norm": 0.19034148752689362, |
| "learning_rate": 0.00017946656923639024, |
| "loss": 0.7279, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.10466812545587163, |
| "grad_norm": 0.18591929972171783, |
| "learning_rate": 0.0001793934965290464, |
| "loss": 0.7353, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.1050328227571116, |
| "grad_norm": 0.19063667953014374, |
| "learning_rate": 0.0001793204238217026, |
| "loss": 0.7322, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.10539752005835157, |
| "grad_norm": 0.20110805332660675, |
| "learning_rate": 0.0001792473511143588, |
| "loss": 0.869, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.10576221735959154, |
| "grad_norm": 0.18062824010849, |
| "learning_rate": 0.000179174278407015, |
| "loss": 0.6839, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.1061269146608315, |
| "grad_norm": 0.21643158793449402, |
| "learning_rate": 0.00017910120569967117, |
| "loss": 0.7626, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.10649161196207148, |
| "grad_norm": 0.20745426416397095, |
| "learning_rate": 0.00017902813299232738, |
| "loss": 0.8748, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.10685630926331145, |
| "grad_norm": 0.1999918520450592, |
| "learning_rate": 0.00017895506028498356, |
| "loss": 0.7267, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.10722100656455143, |
| "grad_norm": 0.17721255123615265, |
| "learning_rate": 0.00017888198757763977, |
| "loss": 0.788, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.10758570386579139, |
| "grad_norm": 0.19578172266483307, |
| "learning_rate": 0.00017880891487029598, |
| "loss": 0.6801, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.10795040116703136, |
| "grad_norm": 0.20748168230056763, |
| "learning_rate": 0.00017873584216295213, |
| "loss": 0.7624, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.10831509846827134, |
| "grad_norm": 0.17874249815940857, |
| "learning_rate": 0.00017866276945560834, |
| "loss": 0.7029, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.10867979576951131, |
| "grad_norm": 0.21527762711048126, |
| "learning_rate": 0.00017858969674826455, |
| "loss": 0.9119, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.10904449307075127, |
| "grad_norm": 0.16523544490337372, |
| "learning_rate": 0.00017851662404092073, |
| "loss": 0.406, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.10940919037199125, |
| "grad_norm": 0.18175430595874786, |
| "learning_rate": 0.0001784435513335769, |
| "loss": 0.6711, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.10977388767323122, |
| "grad_norm": 0.19601181149482727, |
| "learning_rate": 0.00017837047862623312, |
| "loss": 0.7117, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.11013858497447118, |
| "grad_norm": 0.1821203976869583, |
| "learning_rate": 0.0001782974059188893, |
| "loss": 0.6163, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.11050328227571116, |
| "grad_norm": 0.1985776126384735, |
| "learning_rate": 0.0001782243332115455, |
| "loss": 0.6357, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.11086797957695113, |
| "grad_norm": 0.18439283967018127, |
| "learning_rate": 0.0001781512605042017, |
| "loss": 0.6007, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.1112326768781911, |
| "grad_norm": 0.19900138676166534, |
| "learning_rate": 0.00017807818779685787, |
| "loss": 0.794, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.11159737417943107, |
| "grad_norm": 0.1659189611673355, |
| "learning_rate": 0.00017800511508951408, |
| "loss": 0.4824, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.11196207148067104, |
| "grad_norm": 0.19788797199726105, |
| "learning_rate": 0.00017793204238217026, |
| "loss": 0.6806, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.11232676878191102, |
| "grad_norm": 0.19467923045158386, |
| "learning_rate": 0.00017785896967482647, |
| "loss": 0.7515, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.11269146608315099, |
| "grad_norm": 0.20231202244758606, |
| "learning_rate": 0.00017778589696748265, |
| "loss": 0.7302, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.11305616338439095, |
| "grad_norm": 0.1800953596830368, |
| "learning_rate": 0.00017771282426013883, |
| "loss": 0.5432, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.11342086068563093, |
| "grad_norm": 0.15572136640548706, |
| "learning_rate": 0.00017763975155279504, |
| "loss": 0.4674, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.1137855579868709, |
| "grad_norm": 0.1948738545179367, |
| "learning_rate": 0.00017756667884545125, |
| "loss": 0.8442, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.11415025528811087, |
| "grad_norm": 0.1789115071296692, |
| "learning_rate": 0.00017749360613810743, |
| "loss": 0.6009, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.11451495258935084, |
| "grad_norm": 0.22927749156951904, |
| "learning_rate": 0.0001774205334307636, |
| "loss": 0.866, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.11487964989059081, |
| "grad_norm": 0.2268659621477127, |
| "learning_rate": 0.00017734746072341982, |
| "loss": 0.918, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.11524434719183078, |
| "grad_norm": 0.14317840337753296, |
| "learning_rate": 0.000177274388016076, |
| "loss": 0.3706, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.11560904449307075, |
| "grad_norm": 0.1896270215511322, |
| "learning_rate": 0.0001772013153087322, |
| "loss": 0.7108, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.11597374179431072, |
| "grad_norm": 0.18470154702663422, |
| "learning_rate": 0.0001771282426013884, |
| "loss": 0.8041, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.1163384390955507, |
| "grad_norm": 0.1888645738363266, |
| "learning_rate": 0.00017705516989404457, |
| "loss": 0.7601, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.11670313639679067, |
| "grad_norm": 0.19202347099781036, |
| "learning_rate": 0.00017698209718670078, |
| "loss": 0.6529, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.11706783369803063, |
| "grad_norm": 0.18936879932880402, |
| "learning_rate": 0.000176909024479357, |
| "loss": 0.768, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.1174325309992706, |
| "grad_norm": 0.18530336022377014, |
| "learning_rate": 0.00017683595177201314, |
| "loss": 0.7997, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.11779722830051058, |
| "grad_norm": 0.1945243924856186, |
| "learning_rate": 0.00017676287906466935, |
| "loss": 0.8864, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.11816192560175055, |
| "grad_norm": 0.16887742280960083, |
| "learning_rate": 0.00017668980635732556, |
| "loss": 0.8014, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.11852662290299051, |
| "grad_norm": 0.20715931057929993, |
| "learning_rate": 0.00017661673364998174, |
| "loss": 0.9641, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.11889132020423049, |
| "grad_norm": 0.18339566886425018, |
| "learning_rate": 0.00017654366094263792, |
| "loss": 0.6723, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.11925601750547046, |
| "grad_norm": 0.1606166958808899, |
| "learning_rate": 0.00017647058823529413, |
| "loss": 0.499, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.11962071480671042, |
| "grad_norm": 0.17019236087799072, |
| "learning_rate": 0.0001763975155279503, |
| "loss": 0.7332, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.1199854121079504, |
| "grad_norm": 0.15148764848709106, |
| "learning_rate": 0.00017632444282060652, |
| "loss": 0.477, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.12035010940919037, |
| "grad_norm": 0.182627871632576, |
| "learning_rate": 0.00017625137011326273, |
| "loss": 0.7443, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.12071480671043035, |
| "grad_norm": 0.18017525970935822, |
| "learning_rate": 0.00017617829740591888, |
| "loss": 0.6439, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.12107950401167031, |
| "grad_norm": 0.1846102774143219, |
| "learning_rate": 0.0001761052246985751, |
| "loss": 0.7023, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.12144420131291028, |
| "grad_norm": 0.19771961867809296, |
| "learning_rate": 0.0001760321519912313, |
| "loss": 0.7772, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.12180889861415026, |
| "grad_norm": 0.18164925277233124, |
| "learning_rate": 0.00017595907928388748, |
| "loss": 0.7348, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.12217359591539023, |
| "grad_norm": 0.182021364569664, |
| "learning_rate": 0.00017588600657654366, |
| "loss": 0.7034, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.12253829321663019, |
| "grad_norm": 0.1674763709306717, |
| "learning_rate": 0.00017581293386919987, |
| "loss": 0.5969, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.12290299051787017, |
| "grad_norm": 0.18998658657073975, |
| "learning_rate": 0.00017573986116185605, |
| "loss": 0.723, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.12326768781911014, |
| "grad_norm": 0.20089052617549896, |
| "learning_rate": 0.00017566678845451226, |
| "loss": 0.8216, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.12363238512035012, |
| "grad_norm": 0.18273848295211792, |
| "learning_rate": 0.00017559371574716844, |
| "loss": 0.6634, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.12399708242159008, |
| "grad_norm": 0.18198589980602264, |
| "learning_rate": 0.00017552064303982462, |
| "loss": 0.7026, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.12436177972283005, |
| "grad_norm": 0.17685921490192413, |
| "learning_rate": 0.00017544757033248083, |
| "loss": 0.5641, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.12472647702407003, |
| "grad_norm": 0.20872358977794647, |
| "learning_rate": 0.000175374497625137, |
| "loss": 0.851, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.12509117432531, |
| "grad_norm": 0.19250471889972687, |
| "learning_rate": 0.00017530142491779322, |
| "loss": 0.6265, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.12545587162654998, |
| "grad_norm": 0.18738161027431488, |
| "learning_rate": 0.0001752283522104494, |
| "loss": 0.7612, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.12582056892778992, |
| "grad_norm": 0.188717782497406, |
| "learning_rate": 0.00017515527950310558, |
| "loss": 0.768, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.1261852662290299, |
| "grad_norm": 0.16923002898693085, |
| "learning_rate": 0.0001750822067957618, |
| "loss": 0.6071, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.12654996353026987, |
| "grad_norm": 0.1732785850763321, |
| "learning_rate": 0.000175009134088418, |
| "loss": 0.5788, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.12691466083150985, |
| "grad_norm": 0.15573865175247192, |
| "learning_rate": 0.00017493606138107418, |
| "loss": 0.5516, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.12727935813274982, |
| "grad_norm": 0.17997129261493683, |
| "learning_rate": 0.00017486298867373036, |
| "loss": 0.5627, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.1276440554339898, |
| "grad_norm": 0.17864130437374115, |
| "learning_rate": 0.00017478991596638657, |
| "loss": 0.6458, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.12800875273522977, |
| "grad_norm": 0.17441441118717194, |
| "learning_rate": 0.00017471684325904275, |
| "loss": 0.5879, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.12837345003646974, |
| "grad_norm": 0.16584378480911255, |
| "learning_rate": 0.00017464377055169896, |
| "loss": 0.571, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.1287381473377097, |
| "grad_norm": 0.16552752256393433, |
| "learning_rate": 0.00017457069784435514, |
| "loss": 0.4731, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.12910284463894967, |
| "grad_norm": 0.19212619960308075, |
| "learning_rate": 0.00017449762513701132, |
| "loss": 0.7846, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.12946754194018964, |
| "grad_norm": 0.15730014443397522, |
| "learning_rate": 0.00017442455242966753, |
| "loss": 0.5544, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.12983223924142961, |
| "grad_norm": 0.16712331771850586, |
| "learning_rate": 0.00017435147972232374, |
| "loss": 0.5702, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.1301969365426696, |
| "grad_norm": 0.19101271033287048, |
| "learning_rate": 0.0001742784070149799, |
| "loss": 0.697, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.13056163384390956, |
| "grad_norm": 0.15471835434436798, |
| "learning_rate": 0.0001742053343076361, |
| "loss": 0.4917, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.13092633114514954, |
| "grad_norm": 0.19589225947856903, |
| "learning_rate": 0.0001741322616002923, |
| "loss": 0.7135, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.13129102844638948, |
| "grad_norm": 0.22023500502109528, |
| "learning_rate": 0.0001740591888929485, |
| "loss": 0.8744, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.13165572574762946, |
| "grad_norm": 0.17470814287662506, |
| "learning_rate": 0.00017398611618560467, |
| "loss": 0.5644, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.13202042304886943, |
| "grad_norm": 0.16508780419826508, |
| "learning_rate": 0.00017391304347826088, |
| "loss": 0.4774, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.1323851203501094, |
| "grad_norm": 0.18133142590522766, |
| "learning_rate": 0.00017383997077091706, |
| "loss": 0.6806, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.13274981765134938, |
| "grad_norm": 0.19952940940856934, |
| "learning_rate": 0.00017376689806357327, |
| "loss": 0.7776, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.13311451495258936, |
| "grad_norm": 0.17585842311382294, |
| "learning_rate": 0.00017369382535622948, |
| "loss": 0.7543, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.13347921225382933, |
| "grad_norm": 0.19367621839046478, |
| "learning_rate": 0.00017362075264888563, |
| "loss": 0.585, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.13384390955506928, |
| "grad_norm": 0.1928108036518097, |
| "learning_rate": 0.00017354767994154184, |
| "loss": 0.6656, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.13420860685630925, |
| "grad_norm": 0.19390811026096344, |
| "learning_rate": 0.00017347460723419805, |
| "loss": 0.713, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.13457330415754923, |
| "grad_norm": 0.19284166395664215, |
| "learning_rate": 0.00017340153452685423, |
| "loss": 0.7414, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.1349380014587892, |
| "grad_norm": 0.16773255169391632, |
| "learning_rate": 0.0001733284618195104, |
| "loss": 0.5786, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.13530269876002918, |
| "grad_norm": 0.15630416572093964, |
| "learning_rate": 0.00017325538911216662, |
| "loss": 0.4774, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.13566739606126915, |
| "grad_norm": 0.1833135038614273, |
| "learning_rate": 0.0001731823164048228, |
| "loss": 0.7178, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.13603209336250913, |
| "grad_norm": 0.19849684834480286, |
| "learning_rate": 0.000173109243697479, |
| "loss": 0.7354, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.1363967906637491, |
| "grad_norm": 0.17072516679763794, |
| "learning_rate": 0.0001730361709901352, |
| "loss": 0.5945, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.13676148796498905, |
| "grad_norm": 0.1784793585538864, |
| "learning_rate": 0.00017296309828279137, |
| "loss": 0.6322, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.13712618526622902, |
| "grad_norm": 0.192670539021492, |
| "learning_rate": 0.00017289002557544758, |
| "loss": 0.7612, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.137490882567469, |
| "grad_norm": 0.1875181794166565, |
| "learning_rate": 0.00017281695286810376, |
| "loss": 0.7439, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.13785557986870897, |
| "grad_norm": 0.1796884834766388, |
| "learning_rate": 0.00017274388016075997, |
| "loss": 0.8694, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.13822027716994895, |
| "grad_norm": 0.20059190690517426, |
| "learning_rate": 0.00017267080745341615, |
| "loss": 0.8027, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.13858497447118892, |
| "grad_norm": 0.19425062835216522, |
| "learning_rate": 0.00017259773474607236, |
| "loss": 0.8948, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.1389496717724289, |
| "grad_norm": 0.20872867107391357, |
| "learning_rate": 0.00017252466203872854, |
| "loss": 0.9552, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.13931436907366884, |
| "grad_norm": 0.1698973923921585, |
| "learning_rate": 0.00017245158933138475, |
| "loss": 0.5604, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.13967906637490882, |
| "grad_norm": 0.19797375798225403, |
| "learning_rate": 0.00017237851662404093, |
| "loss": 0.8872, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.1400437636761488, |
| "grad_norm": 0.16452281177043915, |
| "learning_rate": 0.00017230544391669711, |
| "loss": 0.5724, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.14040846097738877, |
| "grad_norm": 0.17133468389511108, |
| "learning_rate": 0.00017223237120935332, |
| "loss": 0.592, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.14077315827862874, |
| "grad_norm": 0.17517192661762238, |
| "learning_rate": 0.0001721592985020095, |
| "loss": 0.5597, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.14113785557986872, |
| "grad_norm": 0.17406406998634338, |
| "learning_rate": 0.00017208622579466569, |
| "loss": 0.7118, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.1415025528811087, |
| "grad_norm": 0.17395231127738953, |
| "learning_rate": 0.0001720131530873219, |
| "loss": 0.6616, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.14186725018234866, |
| "grad_norm": 0.21478557586669922, |
| "learning_rate": 0.00017194008037997807, |
| "loss": 0.8517, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.1422319474835886, |
| "grad_norm": 0.22842730581760406, |
| "learning_rate": 0.00017186700767263428, |
| "loss": 0.9611, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.14259664478482859, |
| "grad_norm": 0.18962237238883972, |
| "learning_rate": 0.0001717939349652905, |
| "loss": 0.7606, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.14296134208606856, |
| "grad_norm": 0.19581514596939087, |
| "learning_rate": 0.00017172086225794665, |
| "loss": 0.892, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.14332603938730853, |
| "grad_norm": 0.18178801238536835, |
| "learning_rate": 0.00017164778955060285, |
| "loss": 0.7477, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.1436907366885485, |
| "grad_norm": 0.18961721658706665, |
| "learning_rate": 0.00017157471684325906, |
| "loss": 0.771, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.14405543398978848, |
| "grad_norm": 0.18671804666519165, |
| "learning_rate": 0.00017150164413591524, |
| "loss": 0.7383, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.14442013129102846, |
| "grad_norm": 0.17833957076072693, |
| "learning_rate": 0.00017142857142857143, |
| "loss": 0.6363, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.1447848285922684, |
| "grad_norm": 0.18321546912193298, |
| "learning_rate": 0.00017135549872122763, |
| "loss": 0.7068, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.14514952589350838, |
| "grad_norm": 0.20233039557933807, |
| "learning_rate": 0.00017128242601388381, |
| "loss": 0.8699, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.14551422319474835, |
| "grad_norm": 0.1749749630689621, |
| "learning_rate": 0.00017120935330654002, |
| "loss": 0.7355, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.14587892049598833, |
| "grad_norm": 0.14458586275577545, |
| "learning_rate": 0.00017113628059919623, |
| "loss": 0.427, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.1462436177972283, |
| "grad_norm": 0.183969646692276, |
| "learning_rate": 0.00017106320789185239, |
| "loss": 0.6166, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.14660831509846828, |
| "grad_norm": 0.15677833557128906, |
| "learning_rate": 0.0001709901351845086, |
| "loss": 0.4161, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.14697301239970825, |
| "grad_norm": 0.20313799381256104, |
| "learning_rate": 0.0001709170624771648, |
| "loss": 0.8189, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.1473377097009482, |
| "grad_norm": 0.17821913957595825, |
| "learning_rate": 0.00017084398976982098, |
| "loss": 0.6365, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.14770240700218817, |
| "grad_norm": 0.1630881130695343, |
| "learning_rate": 0.00017077091706247717, |
| "loss": 0.5623, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.14806710430342815, |
| "grad_norm": 0.20716772973537445, |
| "learning_rate": 0.00017069784435513337, |
| "loss": 0.9347, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.14843180160466812, |
| "grad_norm": 0.17428599298000336, |
| "learning_rate": 0.00017062477164778956, |
| "loss": 0.6188, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.1487964989059081, |
| "grad_norm": 0.15823887288570404, |
| "learning_rate": 0.00017055169894044576, |
| "loss": 0.5995, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.14916119620714807, |
| "grad_norm": 0.1716741919517517, |
| "learning_rate": 0.00017047862623310194, |
| "loss": 0.6407, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.14952589350838805, |
| "grad_norm": 0.16747424006462097, |
| "learning_rate": 0.00017040555352575813, |
| "loss": 0.5822, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.14989059080962802, |
| "grad_norm": 0.14587102830410004, |
| "learning_rate": 0.00017033248081841433, |
| "loss": 0.431, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.15025528811086797, |
| "grad_norm": 0.1754886656999588, |
| "learning_rate": 0.00017025940811107052, |
| "loss": 0.5515, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.15061998541210794, |
| "grad_norm": 0.20467835664749146, |
| "learning_rate": 0.00017018633540372672, |
| "loss": 0.8614, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.15098468271334792, |
| "grad_norm": 0.2041863650083542, |
| "learning_rate": 0.0001701132626963829, |
| "loss": 0.88, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.1513493800145879, |
| "grad_norm": 0.2153850644826889, |
| "learning_rate": 0.00017004018998903911, |
| "loss": 0.9241, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.15171407731582787, |
| "grad_norm": 0.20441819727420807, |
| "learning_rate": 0.0001699671172816953, |
| "loss": 0.9127, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.15207877461706784, |
| "grad_norm": 0.17975957691669464, |
| "learning_rate": 0.0001698940445743515, |
| "loss": 0.6395, |
| "step": 417 |
| }, |
| { |
| "epoch": 0.15244347191830782, |
| "grad_norm": 0.1702582836151123, |
| "learning_rate": 0.00016982097186700768, |
| "loss": 0.5951, |
| "step": 418 |
| }, |
| { |
| "epoch": 0.15280816921954776, |
| "grad_norm": 0.1543644517660141, |
| "learning_rate": 0.00016974789915966387, |
| "loss": 0.5093, |
| "step": 419 |
| }, |
| { |
| "epoch": 0.15317286652078774, |
| "grad_norm": 0.1747189462184906, |
| "learning_rate": 0.00016967482645232007, |
| "loss": 0.5314, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.1535375638220277, |
| "grad_norm": 0.22516067326068878, |
| "learning_rate": 0.00016960175374497626, |
| "loss": 0.9766, |
| "step": 421 |
| }, |
| { |
| "epoch": 0.1539022611232677, |
| "grad_norm": 0.21535086631774902, |
| "learning_rate": 0.00016952868103763244, |
| "loss": 0.8287, |
| "step": 422 |
| }, |
| { |
| "epoch": 0.15426695842450766, |
| "grad_norm": 0.17004358768463135, |
| "learning_rate": 0.00016945560833028865, |
| "loss": 0.6612, |
| "step": 423 |
| }, |
| { |
| "epoch": 0.15463165572574764, |
| "grad_norm": 0.18636055290699005, |
| "learning_rate": 0.00016938253562294483, |
| "loss": 0.784, |
| "step": 424 |
| }, |
| { |
| "epoch": 0.1549963530269876, |
| "grad_norm": 0.1781081259250641, |
| "learning_rate": 0.00016930946291560104, |
| "loss": 0.5275, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.15536105032822758, |
| "grad_norm": 0.19579234719276428, |
| "learning_rate": 0.00016923639020825724, |
| "loss": 0.7712, |
| "step": 426 |
| }, |
| { |
| "epoch": 0.15572574762946753, |
| "grad_norm": 0.18568329513072968, |
| "learning_rate": 0.0001691633175009134, |
| "loss": 0.7509, |
| "step": 427 |
| }, |
| { |
| "epoch": 0.1560904449307075, |
| "grad_norm": 0.17547035217285156, |
| "learning_rate": 0.0001690902447935696, |
| "loss": 0.711, |
| "step": 428 |
| }, |
| { |
| "epoch": 0.15645514223194748, |
| "grad_norm": 0.19599728286266327, |
| "learning_rate": 0.00016901717208622581, |
| "loss": 0.9256, |
| "step": 429 |
| }, |
| { |
| "epoch": 0.15681983953318746, |
| "grad_norm": 0.18903636932373047, |
| "learning_rate": 0.000168944099378882, |
| "loss": 0.7167, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.15718453683442743, |
| "grad_norm": 0.14108788967132568, |
| "learning_rate": 0.00016887102667153818, |
| "loss": 0.4165, |
| "step": 431 |
| }, |
| { |
| "epoch": 0.1575492341356674, |
| "grad_norm": 0.14891500771045685, |
| "learning_rate": 0.00016879795396419439, |
| "loss": 0.4908, |
| "step": 432 |
| }, |
| { |
| "epoch": 0.15791393143690738, |
| "grad_norm": 0.18601083755493164, |
| "learning_rate": 0.00016872488125685057, |
| "loss": 0.6856, |
| "step": 433 |
| }, |
| { |
| "epoch": 0.15827862873814733, |
| "grad_norm": 0.1775410771369934, |
| "learning_rate": 0.00016865180854950678, |
| "loss": 0.6805, |
| "step": 434 |
| }, |
| { |
| "epoch": 0.1586433260393873, |
| "grad_norm": 0.19800400733947754, |
| "learning_rate": 0.00016857873584216296, |
| "loss": 0.7518, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.15900802334062727, |
| "grad_norm": 0.21951526403427124, |
| "learning_rate": 0.00016850566313481914, |
| "loss": 0.9934, |
| "step": 436 |
| }, |
| { |
| "epoch": 0.15937272064186725, |
| "grad_norm": 0.15916769206523895, |
| "learning_rate": 0.00016843259042747535, |
| "loss": 0.5548, |
| "step": 437 |
| }, |
| { |
| "epoch": 0.15973741794310722, |
| "grad_norm": 0.20369085669517517, |
| "learning_rate": 0.00016835951772013155, |
| "loss": 0.876, |
| "step": 438 |
| }, |
| { |
| "epoch": 0.1601021152443472, |
| "grad_norm": 0.16725899279117584, |
| "learning_rate": 0.00016828644501278774, |
| "loss": 0.6011, |
| "step": 439 |
| }, |
| { |
| "epoch": 0.16046681254558717, |
| "grad_norm": 0.15776625275611877, |
| "learning_rate": 0.00016821337230544392, |
| "loss": 0.4744, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.16083150984682712, |
| "grad_norm": 0.20461910963058472, |
| "learning_rate": 0.00016814029959810013, |
| "loss": 0.8574, |
| "step": 441 |
| }, |
| { |
| "epoch": 0.1611962071480671, |
| "grad_norm": 0.1932179033756256, |
| "learning_rate": 0.0001680672268907563, |
| "loss": 0.7493, |
| "step": 442 |
| }, |
| { |
| "epoch": 0.16156090444930707, |
| "grad_norm": 0.17809215188026428, |
| "learning_rate": 0.00016799415418341252, |
| "loss": 0.6159, |
| "step": 443 |
| }, |
| { |
| "epoch": 0.16192560175054704, |
| "grad_norm": 0.19475284218788147, |
| "learning_rate": 0.0001679210814760687, |
| "loss": 0.7342, |
| "step": 444 |
| }, |
| { |
| "epoch": 0.16229029905178702, |
| "grad_norm": 0.17827290296554565, |
| "learning_rate": 0.00016784800876872488, |
| "loss": 0.7018, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.162654996353027, |
| "grad_norm": 0.18410883843898773, |
| "learning_rate": 0.0001677749360613811, |
| "loss": 0.689, |
| "step": 446 |
| }, |
| { |
| "epoch": 0.16301969365426697, |
| "grad_norm": 0.161324605345726, |
| "learning_rate": 0.00016770186335403727, |
| "loss": 0.5249, |
| "step": 447 |
| }, |
| { |
| "epoch": 0.16338439095550694, |
| "grad_norm": 0.1743507832288742, |
| "learning_rate": 0.00016762879064669348, |
| "loss": 0.7147, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.1637490882567469, |
| "grad_norm": 0.18522591888904572, |
| "learning_rate": 0.00016755571793934966, |
| "loss": 0.6813, |
| "step": 449 |
| }, |
| { |
| "epoch": 0.16411378555798686, |
| "grad_norm": 0.18846698105335236, |
| "learning_rate": 0.00016748264523200587, |
| "loss": 0.8582, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.16447848285922684, |
| "grad_norm": 0.18033850193023682, |
| "learning_rate": 0.00016740957252466205, |
| "loss": 0.7465, |
| "step": 451 |
| }, |
| { |
| "epoch": 0.1648431801604668, |
| "grad_norm": 0.1762772500514984, |
| "learning_rate": 0.00016733649981731826, |
| "loss": 0.6501, |
| "step": 452 |
| }, |
| { |
| "epoch": 0.1652078774617068, |
| "grad_norm": 0.20865213871002197, |
| "learning_rate": 0.00016726342710997444, |
| "loss": 0.5579, |
| "step": 453 |
| }, |
| { |
| "epoch": 0.16557257476294676, |
| "grad_norm": 0.20231905579566956, |
| "learning_rate": 0.00016719035440263062, |
| "loss": 0.8055, |
| "step": 454 |
| }, |
| { |
| "epoch": 0.16593727206418674, |
| "grad_norm": 0.1755821406841278, |
| "learning_rate": 0.00016711728169528683, |
| "loss": 0.6538, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.16630196936542668, |
| "grad_norm": 0.18525098264217377, |
| "learning_rate": 0.000167044208987943, |
| "loss": 0.6979, |
| "step": 456 |
| }, |
| { |
| "epoch": 0.16666666666666666, |
| "grad_norm": 0.18403197824954987, |
| "learning_rate": 0.0001669711362805992, |
| "loss": 0.8222, |
| "step": 457 |
| }, |
| { |
| "epoch": 0.16703136396790663, |
| "grad_norm": 0.1567261666059494, |
| "learning_rate": 0.0001668980635732554, |
| "loss": 0.5104, |
| "step": 458 |
| }, |
| { |
| "epoch": 0.1673960612691466, |
| "grad_norm": 0.1938832849264145, |
| "learning_rate": 0.00016682499086591158, |
| "loss": 0.6972, |
| "step": 459 |
| }, |
| { |
| "epoch": 0.16776075857038658, |
| "grad_norm": 0.19050729274749756, |
| "learning_rate": 0.0001667519181585678, |
| "loss": 0.7917, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.16812545587162656, |
| "grad_norm": 0.22484062612056732, |
| "learning_rate": 0.000166678845451224, |
| "loss": 1.0339, |
| "step": 461 |
| }, |
| { |
| "epoch": 0.16849015317286653, |
| "grad_norm": 0.18188448250293732, |
| "learning_rate": 0.00016660577274388015, |
| "loss": 0.6295, |
| "step": 462 |
| }, |
| { |
| "epoch": 0.1688548504741065, |
| "grad_norm": 0.17250071465969086, |
| "learning_rate": 0.00016653270003653636, |
| "loss": 0.5911, |
| "step": 463 |
| }, |
| { |
| "epoch": 0.16921954777534645, |
| "grad_norm": 0.177398219704628, |
| "learning_rate": 0.00016645962732919257, |
| "loss": 0.6524, |
| "step": 464 |
| }, |
| { |
| "epoch": 0.16958424507658643, |
| "grad_norm": 0.18105994164943695, |
| "learning_rate": 0.00016638655462184875, |
| "loss": 0.6554, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.1699489423778264, |
| "grad_norm": 0.16810593008995056, |
| "learning_rate": 0.00016631348191450493, |
| "loss": 0.5617, |
| "step": 466 |
| }, |
| { |
| "epoch": 0.17031363967906638, |
| "grad_norm": 0.18866056203842163, |
| "learning_rate": 0.00016624040920716114, |
| "loss": 0.848, |
| "step": 467 |
| }, |
| { |
| "epoch": 0.17067833698030635, |
| "grad_norm": 0.18827949464321136, |
| "learning_rate": 0.00016616733649981732, |
| "loss": 0.6413, |
| "step": 468 |
| }, |
| { |
| "epoch": 0.17104303428154632, |
| "grad_norm": 0.16476339101791382, |
| "learning_rate": 0.00016609426379247353, |
| "loss": 0.6205, |
| "step": 469 |
| }, |
| { |
| "epoch": 0.1714077315827863, |
| "grad_norm": 0.1794668436050415, |
| "learning_rate": 0.0001660211910851297, |
| "loss": 0.557, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.17177242888402625, |
| "grad_norm": 0.18404535949230194, |
| "learning_rate": 0.0001659481183777859, |
| "loss": 0.7329, |
| "step": 471 |
| }, |
| { |
| "epoch": 0.17213712618526622, |
| "grad_norm": 0.1756718009710312, |
| "learning_rate": 0.0001658750456704421, |
| "loss": 0.5631, |
| "step": 472 |
| }, |
| { |
| "epoch": 0.1725018234865062, |
| "grad_norm": 0.17083744704723358, |
| "learning_rate": 0.0001658019729630983, |
| "loss": 0.6473, |
| "step": 473 |
| }, |
| { |
| "epoch": 0.17286652078774617, |
| "grad_norm": 0.17385685443878174, |
| "learning_rate": 0.0001657289002557545, |
| "loss": 0.6402, |
| "step": 474 |
| }, |
| { |
| "epoch": 0.17323121808898614, |
| "grad_norm": 0.16864272952079773, |
| "learning_rate": 0.00016565582754841067, |
| "loss": 0.5569, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.17359591539022612, |
| "grad_norm": 0.1955600082874298, |
| "learning_rate": 0.00016558275484106688, |
| "loss": 0.8456, |
| "step": 476 |
| }, |
| { |
| "epoch": 0.1739606126914661, |
| "grad_norm": 0.1845158338546753, |
| "learning_rate": 0.00016550968213372306, |
| "loss": 0.6691, |
| "step": 477 |
| }, |
| { |
| "epoch": 0.17432530999270604, |
| "grad_norm": 0.14487138390541077, |
| "learning_rate": 0.00016543660942637927, |
| "loss": 0.3899, |
| "step": 478 |
| }, |
| { |
| "epoch": 0.17469000729394601, |
| "grad_norm": 0.1681089848279953, |
| "learning_rate": 0.00016536353671903545, |
| "loss": 0.5527, |
| "step": 479 |
| }, |
| { |
| "epoch": 0.175054704595186, |
| "grad_norm": 0.1934751272201538, |
| "learning_rate": 0.00016529046401169163, |
| "loss": 0.764, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.17541940189642596, |
| "grad_norm": 0.17804409563541412, |
| "learning_rate": 0.00016521739130434784, |
| "loss": 0.6075, |
| "step": 481 |
| }, |
| { |
| "epoch": 0.17578409919766594, |
| "grad_norm": 0.17334255576133728, |
| "learning_rate": 0.00016514431859700402, |
| "loss": 0.5402, |
| "step": 482 |
| }, |
| { |
| "epoch": 0.1761487964989059, |
| "grad_norm": 0.166969433426857, |
| "learning_rate": 0.0001650712458896602, |
| "loss": 0.5684, |
| "step": 483 |
| }, |
| { |
| "epoch": 0.1765134938001459, |
| "grad_norm": 0.17517952620983124, |
| "learning_rate": 0.0001649981731823164, |
| "loss": 0.6743, |
| "step": 484 |
| }, |
| { |
| "epoch": 0.17687819110138586, |
| "grad_norm": 0.18539521098136902, |
| "learning_rate": 0.00016492510047497262, |
| "loss": 0.7196, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.1772428884026258, |
| "grad_norm": 0.18387848138809204, |
| "learning_rate": 0.0001648520277676288, |
| "loss": 0.6336, |
| "step": 486 |
| }, |
| { |
| "epoch": 0.17760758570386578, |
| "grad_norm": 0.18015360832214355, |
| "learning_rate": 0.000164778955060285, |
| "loss": 0.7254, |
| "step": 487 |
| }, |
| { |
| "epoch": 0.17797228300510576, |
| "grad_norm": 0.1397644728422165, |
| "learning_rate": 0.0001647058823529412, |
| "loss": 0.3431, |
| "step": 488 |
| }, |
| { |
| "epoch": 0.17833698030634573, |
| "grad_norm": 0.22470806539058685, |
| "learning_rate": 0.00016463280964559737, |
| "loss": 0.9921, |
| "step": 489 |
| }, |
| { |
| "epoch": 0.1787016776075857, |
| "grad_norm": 0.2027674913406372, |
| "learning_rate": 0.00016455973693825358, |
| "loss": 0.8432, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.17906637490882568, |
| "grad_norm": 0.16964022815227509, |
| "learning_rate": 0.00016448666423090976, |
| "loss": 0.659, |
| "step": 491 |
| }, |
| { |
| "epoch": 0.17943107221006566, |
| "grad_norm": 0.18222148716449738, |
| "learning_rate": 0.00016441359152356594, |
| "loss": 0.6367, |
| "step": 492 |
| }, |
| { |
| "epoch": 0.1797957695113056, |
| "grad_norm": 0.19195525348186493, |
| "learning_rate": 0.00016434051881622215, |
| "loss": 0.657, |
| "step": 493 |
| }, |
| { |
| "epoch": 0.18016046681254558, |
| "grad_norm": 0.17382116615772247, |
| "learning_rate": 0.00016426744610887833, |
| "loss": 0.6367, |
| "step": 494 |
| }, |
| { |
| "epoch": 0.18052516411378555, |
| "grad_norm": 0.18641424179077148, |
| "learning_rate": 0.00016419437340153454, |
| "loss": 0.7523, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.18088986141502553, |
| "grad_norm": 0.1832839697599411, |
| "learning_rate": 0.00016412130069419075, |
| "loss": 0.7381, |
| "step": 496 |
| }, |
| { |
| "epoch": 0.1812545587162655, |
| "grad_norm": 0.1541210114955902, |
| "learning_rate": 0.0001640482279868469, |
| "loss": 0.5941, |
| "step": 497 |
| }, |
| { |
| "epoch": 0.18161925601750548, |
| "grad_norm": 0.22230711579322815, |
| "learning_rate": 0.0001639751552795031, |
| "loss": 0.9853, |
| "step": 498 |
| }, |
| { |
| "epoch": 0.18198395331874545, |
| "grad_norm": 0.17912793159484863, |
| "learning_rate": 0.00016390208257215932, |
| "loss": 0.6753, |
| "step": 499 |
| }, |
| { |
| "epoch": 0.18234865061998543, |
| "grad_norm": 0.1971275359392166, |
| "learning_rate": 0.0001638290098648155, |
| "loss": 0.7659, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.18271334792122537, |
| "grad_norm": 0.19420406222343445, |
| "learning_rate": 0.00016375593715747168, |
| "loss": 0.7616, |
| "step": 501 |
| }, |
| { |
| "epoch": 0.18307804522246535, |
| "grad_norm": 0.16759932041168213, |
| "learning_rate": 0.0001636828644501279, |
| "loss": 0.5594, |
| "step": 502 |
| }, |
| { |
| "epoch": 0.18344274252370532, |
| "grad_norm": 0.1994057297706604, |
| "learning_rate": 0.00016360979174278407, |
| "loss": 0.8499, |
| "step": 503 |
| }, |
| { |
| "epoch": 0.1838074398249453, |
| "grad_norm": 0.1942475587129593, |
| "learning_rate": 0.00016353671903544028, |
| "loss": 0.8477, |
| "step": 504 |
| }, |
| { |
| "epoch": 0.18417213712618527, |
| "grad_norm": 0.20013171434402466, |
| "learning_rate": 0.00016346364632809646, |
| "loss": 0.8432, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.18453683442742524, |
| "grad_norm": 0.19940058887004852, |
| "learning_rate": 0.00016339057362075264, |
| "loss": 0.9063, |
| "step": 506 |
| }, |
| { |
| "epoch": 0.18490153172866522, |
| "grad_norm": 0.17371918261051178, |
| "learning_rate": 0.00016331750091340885, |
| "loss": 0.755, |
| "step": 507 |
| }, |
| { |
| "epoch": 0.18526622902990517, |
| "grad_norm": 0.1687958985567093, |
| "learning_rate": 0.00016324442820606506, |
| "loss": 0.6164, |
| "step": 508 |
| }, |
| { |
| "epoch": 0.18563092633114514, |
| "grad_norm": 0.1572989672422409, |
| "learning_rate": 0.00016317135549872124, |
| "loss": 0.5366, |
| "step": 509 |
| }, |
| { |
| "epoch": 0.18599562363238512, |
| "grad_norm": 0.1621757596731186, |
| "learning_rate": 0.00016309828279137742, |
| "loss": 0.572, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.1863603209336251, |
| "grad_norm": 0.16226086020469666, |
| "learning_rate": 0.00016302521008403363, |
| "loss": 0.6112, |
| "step": 511 |
| }, |
| { |
| "epoch": 0.18672501823486506, |
| "grad_norm": 0.19324566423892975, |
| "learning_rate": 0.0001629521373766898, |
| "loss": 0.8287, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.18708971553610504, |
| "grad_norm": 0.18370205163955688, |
| "learning_rate": 0.00016287906466934602, |
| "loss": 0.7856, |
| "step": 513 |
| }, |
| { |
| "epoch": 0.187454412837345, |
| "grad_norm": 0.22035083174705505, |
| "learning_rate": 0.0001628059919620022, |
| "loss": 0.9329, |
| "step": 514 |
| }, |
| { |
| "epoch": 0.187819110138585, |
| "grad_norm": 0.1782984435558319, |
| "learning_rate": 0.00016273291925465838, |
| "loss": 0.8342, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.18818380743982493, |
| "grad_norm": 0.23118692636489868, |
| "learning_rate": 0.0001626598465473146, |
| "loss": 0.9808, |
| "step": 516 |
| }, |
| { |
| "epoch": 0.1885485047410649, |
| "grad_norm": 0.1947745978832245, |
| "learning_rate": 0.00016258677383997077, |
| "loss": 0.8156, |
| "step": 517 |
| }, |
| { |
| "epoch": 0.18891320204230488, |
| "grad_norm": 0.18419331312179565, |
| "learning_rate": 0.00016251370113262695, |
| "loss": 0.7476, |
| "step": 518 |
| }, |
| { |
| "epoch": 0.18927789934354486, |
| "grad_norm": 0.1725061982870102, |
| "learning_rate": 0.00016244062842528316, |
| "loss": 0.646, |
| "step": 519 |
| }, |
| { |
| "epoch": 0.18964259664478483, |
| "grad_norm": 0.17075173556804657, |
| "learning_rate": 0.00016236755571793937, |
| "loss": 0.5723, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.1900072939460248, |
| "grad_norm": 0.17799945175647736, |
| "learning_rate": 0.00016229448301059555, |
| "loss": 0.8289, |
| "step": 521 |
| }, |
| { |
| "epoch": 0.19037199124726478, |
| "grad_norm": 0.18537406623363495, |
| "learning_rate": 0.00016222141030325176, |
| "loss": 0.8107, |
| "step": 522 |
| }, |
| { |
| "epoch": 0.19073668854850473, |
| "grad_norm": 0.20047864317893982, |
| "learning_rate": 0.00016214833759590794, |
| "loss": 0.9138, |
| "step": 523 |
| }, |
| { |
| "epoch": 0.1911013858497447, |
| "grad_norm": 0.20184080302715302, |
| "learning_rate": 0.00016207526488856412, |
| "loss": 0.8424, |
| "step": 524 |
| }, |
| { |
| "epoch": 0.19146608315098468, |
| "grad_norm": 0.2016812413930893, |
| "learning_rate": 0.00016200219218122033, |
| "loss": 0.971, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.19183078045222465, |
| "grad_norm": 0.17220567166805267, |
| "learning_rate": 0.0001619291194738765, |
| "loss": 0.6097, |
| "step": 526 |
| }, |
| { |
| "epoch": 0.19219547775346463, |
| "grad_norm": 0.16681768000125885, |
| "learning_rate": 0.0001618560467665327, |
| "loss": 0.6361, |
| "step": 527 |
| }, |
| { |
| "epoch": 0.1925601750547046, |
| "grad_norm": 0.1976706087589264, |
| "learning_rate": 0.0001617829740591889, |
| "loss": 0.8039, |
| "step": 528 |
| }, |
| { |
| "epoch": 0.19292487235594458, |
| "grad_norm": 0.14784552156925201, |
| "learning_rate": 0.00016170990135184508, |
| "loss": 0.4833, |
| "step": 529 |
| }, |
| { |
| "epoch": 0.19328956965718452, |
| "grad_norm": 0.18679118156433105, |
| "learning_rate": 0.0001616368286445013, |
| "loss": 0.8576, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.1936542669584245, |
| "grad_norm": 0.1557287573814392, |
| "learning_rate": 0.0001615637559371575, |
| "loss": 0.5705, |
| "step": 531 |
| }, |
| { |
| "epoch": 0.19401896425966447, |
| "grad_norm": 0.19311803579330444, |
| "learning_rate": 0.00016149068322981365, |
| "loss": 0.7767, |
| "step": 532 |
| }, |
| { |
| "epoch": 0.19438366156090445, |
| "grad_norm": 0.14209668338298798, |
| "learning_rate": 0.00016141761052246986, |
| "loss": 0.4313, |
| "step": 533 |
| }, |
| { |
| "epoch": 0.19474835886214442, |
| "grad_norm": 0.15875539183616638, |
| "learning_rate": 0.00016134453781512607, |
| "loss": 0.5665, |
| "step": 534 |
| }, |
| { |
| "epoch": 0.1951130561633844, |
| "grad_norm": 0.17052525281906128, |
| "learning_rate": 0.00016127146510778225, |
| "loss": 0.507, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.19547775346462437, |
| "grad_norm": 0.19006414711475372, |
| "learning_rate": 0.00016119839240043843, |
| "loss": 0.6866, |
| "step": 536 |
| }, |
| { |
| "epoch": 0.19584245076586435, |
| "grad_norm": 0.1550920307636261, |
| "learning_rate": 0.00016112531969309464, |
| "loss": 0.537, |
| "step": 537 |
| }, |
| { |
| "epoch": 0.1962071480671043, |
| "grad_norm": 0.17453649640083313, |
| "learning_rate": 0.00016105224698575082, |
| "loss": 0.5476, |
| "step": 538 |
| }, |
| { |
| "epoch": 0.19657184536834427, |
| "grad_norm": 0.1762418895959854, |
| "learning_rate": 0.00016097917427840703, |
| "loss": 0.7063, |
| "step": 539 |
| }, |
| { |
| "epoch": 0.19693654266958424, |
| "grad_norm": 0.1798284500837326, |
| "learning_rate": 0.0001609061015710632, |
| "loss": 0.6109, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.19730123997082422, |
| "grad_norm": 0.18691901862621307, |
| "learning_rate": 0.0001608330288637194, |
| "loss": 0.7341, |
| "step": 541 |
| }, |
| { |
| "epoch": 0.1976659372720642, |
| "grad_norm": 0.21478383243083954, |
| "learning_rate": 0.0001607599561563756, |
| "loss": 0.8744, |
| "step": 542 |
| }, |
| { |
| "epoch": 0.19803063457330417, |
| "grad_norm": 0.17121157050132751, |
| "learning_rate": 0.0001606868834490318, |
| "loss": 0.54, |
| "step": 543 |
| }, |
| { |
| "epoch": 0.19839533187454414, |
| "grad_norm": 0.1967833936214447, |
| "learning_rate": 0.000160613810741688, |
| "loss": 0.5669, |
| "step": 544 |
| }, |
| { |
| "epoch": 0.1987600291757841, |
| "grad_norm": 0.16104774177074432, |
| "learning_rate": 0.00016054073803434417, |
| "loss": 0.4524, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.19912472647702406, |
| "grad_norm": 0.17480875551700592, |
| "learning_rate": 0.00016046766532700038, |
| "loss": 0.6578, |
| "step": 546 |
| }, |
| { |
| "epoch": 0.19948942377826404, |
| "grad_norm": 0.16630177199840546, |
| "learning_rate": 0.00016039459261965656, |
| "loss": 0.6371, |
| "step": 547 |
| }, |
| { |
| "epoch": 0.199854121079504, |
| "grad_norm": 0.18494483828544617, |
| "learning_rate": 0.00016032151991231277, |
| "loss": 0.7218, |
| "step": 548 |
| }, |
| { |
| "epoch": 0.20021881838074398, |
| "grad_norm": 0.17873701453208923, |
| "learning_rate": 0.00016024844720496895, |
| "loss": 0.6248, |
| "step": 549 |
| }, |
| { |
| "epoch": 0.20058351568198396, |
| "grad_norm": 0.16306859254837036, |
| "learning_rate": 0.00016017537449762513, |
| "loss": 0.5553, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.20094821298322393, |
| "grad_norm": 0.1972372829914093, |
| "learning_rate": 0.00016010230179028134, |
| "loss": 0.8172, |
| "step": 551 |
| }, |
| { |
| "epoch": 0.2013129102844639, |
| "grad_norm": 0.21976488828659058, |
| "learning_rate": 0.00016002922908293752, |
| "loss": 1.0276, |
| "step": 552 |
| }, |
| { |
| "epoch": 0.20167760758570386, |
| "grad_norm": 0.16156013309955597, |
| "learning_rate": 0.0001599561563755937, |
| "loss": 0.5599, |
| "step": 553 |
| }, |
| { |
| "epoch": 0.20204230488694383, |
| "grad_norm": 0.18124879896640778, |
| "learning_rate": 0.00015988308366824991, |
| "loss": 0.6038, |
| "step": 554 |
| }, |
| { |
| "epoch": 0.2024070021881838, |
| "grad_norm": 0.16870321333408356, |
| "learning_rate": 0.00015981001096090612, |
| "loss": 0.6956, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.20277169948942378, |
| "grad_norm": 0.18470539152622223, |
| "learning_rate": 0.0001597369382535623, |
| "loss": 0.7895, |
| "step": 556 |
| }, |
| { |
| "epoch": 0.20313639679066375, |
| "grad_norm": 0.1810760498046875, |
| "learning_rate": 0.0001596638655462185, |
| "loss": 0.7416, |
| "step": 557 |
| }, |
| { |
| "epoch": 0.20350109409190373, |
| "grad_norm": 0.19518287479877472, |
| "learning_rate": 0.0001595907928388747, |
| "loss": 0.8724, |
| "step": 558 |
| }, |
| { |
| "epoch": 0.2038657913931437, |
| "grad_norm": 0.16267555952072144, |
| "learning_rate": 0.00015951772013153087, |
| "loss": 0.6208, |
| "step": 559 |
| }, |
| { |
| "epoch": 0.20423048869438365, |
| "grad_norm": 0.19488964974880219, |
| "learning_rate": 0.00015944464742418708, |
| "loss": 0.8636, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.20459518599562362, |
| "grad_norm": 0.1762942373752594, |
| "learning_rate": 0.00015937157471684326, |
| "loss": 0.6351, |
| "step": 561 |
| }, |
| { |
| "epoch": 0.2049598832968636, |
| "grad_norm": 0.1685190051794052, |
| "learning_rate": 0.00015929850200949945, |
| "loss": 0.7862, |
| "step": 562 |
| }, |
| { |
| "epoch": 0.20532458059810357, |
| "grad_norm": 0.17033015191555023, |
| "learning_rate": 0.00015922542930215565, |
| "loss": 0.7065, |
| "step": 563 |
| }, |
| { |
| "epoch": 0.20568927789934355, |
| "grad_norm": 0.1639910638332367, |
| "learning_rate": 0.00015915235659481184, |
| "loss": 0.5806, |
| "step": 564 |
| }, |
| { |
| "epoch": 0.20605397520058352, |
| "grad_norm": 0.16588236391544342, |
| "learning_rate": 0.00015907928388746804, |
| "loss": 0.639, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.2064186725018235, |
| "grad_norm": 0.1781357228755951, |
| "learning_rate": 0.00015900621118012423, |
| "loss": 0.7193, |
| "step": 566 |
| }, |
| { |
| "epoch": 0.20678336980306344, |
| "grad_norm": 0.2006876915693283, |
| "learning_rate": 0.0001589331384727804, |
| "loss": 0.8255, |
| "step": 567 |
| }, |
| { |
| "epoch": 0.20714806710430342, |
| "grad_norm": 0.24458666145801544, |
| "learning_rate": 0.00015886006576543661, |
| "loss": 0.7577, |
| "step": 568 |
| }, |
| { |
| "epoch": 0.2075127644055434, |
| "grad_norm": 0.18870210647583008, |
| "learning_rate": 0.00015878699305809282, |
| "loss": 0.7712, |
| "step": 569 |
| }, |
| { |
| "epoch": 0.20787746170678337, |
| "grad_norm": 0.1800808161497116, |
| "learning_rate": 0.000158713920350749, |
| "loss": 0.5715, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.20824215900802334, |
| "grad_norm": 0.16994856297969818, |
| "learning_rate": 0.00015864084764340519, |
| "loss": 0.6102, |
| "step": 571 |
| }, |
| { |
| "epoch": 0.20860685630926332, |
| "grad_norm": 0.21051564812660217, |
| "learning_rate": 0.0001585677749360614, |
| "loss": 0.9531, |
| "step": 572 |
| }, |
| { |
| "epoch": 0.2089715536105033, |
| "grad_norm": 0.17979969084262848, |
| "learning_rate": 0.00015849470222871758, |
| "loss": 0.6933, |
| "step": 573 |
| }, |
| { |
| "epoch": 0.20933625091174327, |
| "grad_norm": 0.17062553763389587, |
| "learning_rate": 0.00015842162952137378, |
| "loss": 0.5809, |
| "step": 574 |
| }, |
| { |
| "epoch": 0.2097009482129832, |
| "grad_norm": 0.1818009465932846, |
| "learning_rate": 0.00015834855681402997, |
| "loss": 0.6487, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.2100656455142232, |
| "grad_norm": 0.19140908122062683, |
| "learning_rate": 0.00015827548410668615, |
| "loss": 0.7887, |
| "step": 576 |
| }, |
| { |
| "epoch": 0.21043034281546316, |
| "grad_norm": 0.15934517979621887, |
| "learning_rate": 0.00015820241139934236, |
| "loss": 0.5761, |
| "step": 577 |
| }, |
| { |
| "epoch": 0.21079504011670314, |
| "grad_norm": 0.16064375638961792, |
| "learning_rate": 0.00015812933869199856, |
| "loss": 0.5708, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.2111597374179431, |
| "grad_norm": 0.18466386198997498, |
| "learning_rate": 0.00015805626598465474, |
| "loss": 0.7036, |
| "step": 579 |
| }, |
| { |
| "epoch": 0.21152443471918309, |
| "grad_norm": 0.17533251643180847, |
| "learning_rate": 0.00015798319327731093, |
| "loss": 0.801, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.21188913202042306, |
| "grad_norm": 0.1585182249546051, |
| "learning_rate": 0.00015791012056996713, |
| "loss": 0.5439, |
| "step": 581 |
| }, |
| { |
| "epoch": 0.212253829321663, |
| "grad_norm": 0.18301182985305786, |
| "learning_rate": 0.00015783704786262332, |
| "loss": 0.6721, |
| "step": 582 |
| }, |
| { |
| "epoch": 0.21261852662290298, |
| "grad_norm": 0.19599053263664246, |
| "learning_rate": 0.00015776397515527952, |
| "loss": 0.7901, |
| "step": 583 |
| }, |
| { |
| "epoch": 0.21298322392414296, |
| "grad_norm": 0.19529075920581818, |
| "learning_rate": 0.0001576909024479357, |
| "loss": 0.7955, |
| "step": 584 |
| }, |
| { |
| "epoch": 0.21334792122538293, |
| "grad_norm": 0.17965692281723022, |
| "learning_rate": 0.0001576178297405919, |
| "loss": 0.6761, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.2137126185266229, |
| "grad_norm": 0.15989278256893158, |
| "learning_rate": 0.0001575447570332481, |
| "loss": 0.5406, |
| "step": 586 |
| }, |
| { |
| "epoch": 0.21407731582786288, |
| "grad_norm": 0.18571379780769348, |
| "learning_rate": 0.00015747168432590428, |
| "loss": 0.784, |
| "step": 587 |
| }, |
| { |
| "epoch": 0.21444201312910285, |
| "grad_norm": 0.17124933004379272, |
| "learning_rate": 0.00015739861161856046, |
| "loss": 0.6769, |
| "step": 588 |
| }, |
| { |
| "epoch": 0.21480671043034283, |
| "grad_norm": 0.2020951509475708, |
| "learning_rate": 0.00015732553891121667, |
| "loss": 0.9287, |
| "step": 589 |
| }, |
| { |
| "epoch": 0.21517140773158278, |
| "grad_norm": 0.1693318635225296, |
| "learning_rate": 0.00015725246620387287, |
| "loss": 0.6093, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.21553610503282275, |
| "grad_norm": 0.1748688966035843, |
| "learning_rate": 0.00015717939349652906, |
| "loss": 0.6981, |
| "step": 591 |
| }, |
| { |
| "epoch": 0.21590080233406272, |
| "grad_norm": 0.15914174914360046, |
| "learning_rate": 0.00015710632078918526, |
| "loss": 0.5255, |
| "step": 592 |
| }, |
| { |
| "epoch": 0.2162654996353027, |
| "grad_norm": 0.16605016589164734, |
| "learning_rate": 0.00015703324808184145, |
| "loss": 0.613, |
| "step": 593 |
| }, |
| { |
| "epoch": 0.21663019693654267, |
| "grad_norm": 0.13033540546894073, |
| "learning_rate": 0.00015696017537449763, |
| "loss": 0.2626, |
| "step": 594 |
| }, |
| { |
| "epoch": 0.21699489423778265, |
| "grad_norm": 0.17081047594547272, |
| "learning_rate": 0.00015688710266715384, |
| "loss": 0.6266, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.21735959153902262, |
| "grad_norm": 0.18738578259944916, |
| "learning_rate": 0.00015681402995981002, |
| "loss": 0.7505, |
| "step": 596 |
| }, |
| { |
| "epoch": 0.21772428884026257, |
| "grad_norm": 0.18220360577106476, |
| "learning_rate": 0.0001567409572524662, |
| "loss": 0.6359, |
| "step": 597 |
| }, |
| { |
| "epoch": 0.21808898614150254, |
| "grad_norm": 0.12919798493385315, |
| "learning_rate": 0.0001566678845451224, |
| "loss": 0.2671, |
| "step": 598 |
| }, |
| { |
| "epoch": 0.21845368344274252, |
| "grad_norm": 0.19884516298770905, |
| "learning_rate": 0.0001565948118377786, |
| "loss": 0.7276, |
| "step": 599 |
| }, |
| { |
| "epoch": 0.2188183807439825, |
| "grad_norm": 0.21019931137561798, |
| "learning_rate": 0.0001565217391304348, |
| "loss": 0.8272, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.21918307804522247, |
| "grad_norm": 0.18614532053470612, |
| "learning_rate": 0.00015644866642309098, |
| "loss": 0.6676, |
| "step": 601 |
| }, |
| { |
| "epoch": 0.21954777534646244, |
| "grad_norm": 0.1795649528503418, |
| "learning_rate": 0.00015637559371574716, |
| "loss": 0.7105, |
| "step": 602 |
| }, |
| { |
| "epoch": 0.21991247264770242, |
| "grad_norm": 0.2079457938671112, |
| "learning_rate": 0.00015630252100840337, |
| "loss": 1.0025, |
| "step": 603 |
| }, |
| { |
| "epoch": 0.22027716994894236, |
| "grad_norm": 0.16998709738254547, |
| "learning_rate": 0.00015622944830105958, |
| "loss": 0.6432, |
| "step": 604 |
| }, |
| { |
| "epoch": 0.22064186725018234, |
| "grad_norm": 0.20293568074703217, |
| "learning_rate": 0.00015615637559371576, |
| "loss": 0.9977, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.2210065645514223, |
| "grad_norm": 0.1662643551826477, |
| "learning_rate": 0.00015608330288637194, |
| "loss": 0.5779, |
| "step": 606 |
| }, |
| { |
| "epoch": 0.2213712618526623, |
| "grad_norm": 0.21681144833564758, |
| "learning_rate": 0.00015601023017902815, |
| "loss": 0.9133, |
| "step": 607 |
| }, |
| { |
| "epoch": 0.22173595915390226, |
| "grad_norm": 0.1861148327589035, |
| "learning_rate": 0.00015593715747168433, |
| "loss": 0.854, |
| "step": 608 |
| }, |
| { |
| "epoch": 0.22210065645514224, |
| "grad_norm": 0.15234194695949554, |
| "learning_rate": 0.00015586408476434054, |
| "loss": 0.5688, |
| "step": 609 |
| }, |
| { |
| "epoch": 0.2224653537563822, |
| "grad_norm": 0.18444177508354187, |
| "learning_rate": 0.00015579101205699672, |
| "loss": 0.8598, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.22283005105762219, |
| "grad_norm": 0.1650354266166687, |
| "learning_rate": 0.0001557179393496529, |
| "loss": 0.5645, |
| "step": 611 |
| }, |
| { |
| "epoch": 0.22319474835886213, |
| "grad_norm": 0.1783570498228073, |
| "learning_rate": 0.0001556448666423091, |
| "loss": 0.6918, |
| "step": 612 |
| }, |
| { |
| "epoch": 0.2235594456601021, |
| "grad_norm": 0.201633483171463, |
| "learning_rate": 0.00015557179393496532, |
| "loss": 0.9097, |
| "step": 613 |
| }, |
| { |
| "epoch": 0.22392414296134208, |
| "grad_norm": 0.1808750182390213, |
| "learning_rate": 0.00015549872122762147, |
| "loss": 0.5749, |
| "step": 614 |
| }, |
| { |
| "epoch": 0.22428884026258206, |
| "grad_norm": 0.18188488483428955, |
| "learning_rate": 0.00015542564852027768, |
| "loss": 0.7137, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.22465353756382203, |
| "grad_norm": 0.1791074126958847, |
| "learning_rate": 0.0001553525758129339, |
| "loss": 0.8679, |
| "step": 616 |
| }, |
| { |
| "epoch": 0.225018234865062, |
| "grad_norm": 0.18888451159000397, |
| "learning_rate": 0.00015527950310559007, |
| "loss": 0.8296, |
| "step": 617 |
| }, |
| { |
| "epoch": 0.22538293216630198, |
| "grad_norm": 0.1985502392053604, |
| "learning_rate": 0.00015520643039824628, |
| "loss": 0.8388, |
| "step": 618 |
| }, |
| { |
| "epoch": 0.22574762946754193, |
| "grad_norm": 0.20040510594844818, |
| "learning_rate": 0.00015513335769090246, |
| "loss": 0.8381, |
| "step": 619 |
| }, |
| { |
| "epoch": 0.2261123267687819, |
| "grad_norm": 0.20879292488098145, |
| "learning_rate": 0.00015506028498355864, |
| "loss": 0.9998, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.22647702407002188, |
| "grad_norm": 0.18888017535209656, |
| "learning_rate": 0.00015498721227621485, |
| "loss": 0.6915, |
| "step": 621 |
| }, |
| { |
| "epoch": 0.22684172137126185, |
| "grad_norm": 0.19784311950206757, |
| "learning_rate": 0.00015491413956887103, |
| "loss": 1.054, |
| "step": 622 |
| }, |
| { |
| "epoch": 0.22720641867250183, |
| "grad_norm": 0.16005344688892365, |
| "learning_rate": 0.0001548410668615272, |
| "loss": 0.4895, |
| "step": 623 |
| }, |
| { |
| "epoch": 0.2275711159737418, |
| "grad_norm": 0.20644895732402802, |
| "learning_rate": 0.00015476799415418342, |
| "loss": 0.8316, |
| "step": 624 |
| }, |
| { |
| "epoch": 0.22793581327498177, |
| "grad_norm": 0.18263429403305054, |
| "learning_rate": 0.00015469492144683963, |
| "loss": 0.7931, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.22830051057622175, |
| "grad_norm": 0.18260689079761505, |
| "learning_rate": 0.0001546218487394958, |
| "loss": 0.6584, |
| "step": 626 |
| }, |
| { |
| "epoch": 0.2286652078774617, |
| "grad_norm": 0.20142361521720886, |
| "learning_rate": 0.00015454877603215202, |
| "loss": 0.7865, |
| "step": 627 |
| }, |
| { |
| "epoch": 0.22902990517870167, |
| "grad_norm": 0.163302943110466, |
| "learning_rate": 0.0001544757033248082, |
| "loss": 0.6348, |
| "step": 628 |
| }, |
| { |
| "epoch": 0.22939460247994164, |
| "grad_norm": 0.1729688048362732, |
| "learning_rate": 0.00015440263061746438, |
| "loss": 0.6776, |
| "step": 629 |
| }, |
| { |
| "epoch": 0.22975929978118162, |
| "grad_norm": 0.167247012257576, |
| "learning_rate": 0.0001543295579101206, |
| "loss": 0.6274, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.2301239970824216, |
| "grad_norm": 0.17160290479660034, |
| "learning_rate": 0.00015425648520277677, |
| "loss": 0.6194, |
| "step": 631 |
| }, |
| { |
| "epoch": 0.23048869438366157, |
| "grad_norm": 0.17698507010936737, |
| "learning_rate": 0.00015418341249543295, |
| "loss": 0.7014, |
| "step": 632 |
| }, |
| { |
| "epoch": 0.23085339168490154, |
| "grad_norm": 0.1988571584224701, |
| "learning_rate": 0.00015411033978808916, |
| "loss": 0.8195, |
| "step": 633 |
| }, |
| { |
| "epoch": 0.2312180889861415, |
| "grad_norm": 0.18846118450164795, |
| "learning_rate": 0.00015403726708074534, |
| "loss": 0.711, |
| "step": 634 |
| }, |
| { |
| "epoch": 0.23158278628738146, |
| "grad_norm": 0.1780976951122284, |
| "learning_rate": 0.00015396419437340155, |
| "loss": 0.6782, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.23194748358862144, |
| "grad_norm": 0.17137964069843292, |
| "learning_rate": 0.00015389112166605773, |
| "loss": 0.6011, |
| "step": 636 |
| }, |
| { |
| "epoch": 0.2323121808898614, |
| "grad_norm": 0.18975742161273956, |
| "learning_rate": 0.0001538180489587139, |
| "loss": 0.6786, |
| "step": 637 |
| }, |
| { |
| "epoch": 0.2326768781911014, |
| "grad_norm": 0.17563198506832123, |
| "learning_rate": 0.00015374497625137012, |
| "loss": 0.6222, |
| "step": 638 |
| }, |
| { |
| "epoch": 0.23304157549234136, |
| "grad_norm": 0.17805379629135132, |
| "learning_rate": 0.00015367190354402633, |
| "loss": 0.6959, |
| "step": 639 |
| }, |
| { |
| "epoch": 0.23340627279358134, |
| "grad_norm": 0.1799224615097046, |
| "learning_rate": 0.0001535988308366825, |
| "loss": 0.6508, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.2337709700948213, |
| "grad_norm": 0.1822444349527359, |
| "learning_rate": 0.0001535257581293387, |
| "loss": 0.6701, |
| "step": 641 |
| }, |
| { |
| "epoch": 0.23413566739606126, |
| "grad_norm": 0.18402761220932007, |
| "learning_rate": 0.0001534526854219949, |
| "loss": 0.7839, |
| "step": 642 |
| }, |
| { |
| "epoch": 0.23450036469730123, |
| "grad_norm": 0.1818612664937973, |
| "learning_rate": 0.00015337961271465108, |
| "loss": 0.7694, |
| "step": 643 |
| }, |
| { |
| "epoch": 0.2348650619985412, |
| "grad_norm": 0.17730721831321716, |
| "learning_rate": 0.0001533065400073073, |
| "loss": 0.6292, |
| "step": 644 |
| }, |
| { |
| "epoch": 0.23522975929978118, |
| "grad_norm": 0.1790621429681778, |
| "learning_rate": 0.00015323346729996347, |
| "loss": 0.729, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.23559445660102116, |
| "grad_norm": 0.17606894671916962, |
| "learning_rate": 0.00015316039459261965, |
| "loss": 0.6977, |
| "step": 646 |
| }, |
| { |
| "epoch": 0.23595915390226113, |
| "grad_norm": 0.18288281559944153, |
| "learning_rate": 0.00015308732188527586, |
| "loss": 0.7042, |
| "step": 647 |
| }, |
| { |
| "epoch": 0.2363238512035011, |
| "grad_norm": 0.17753863334655762, |
| "learning_rate": 0.00015301424917793207, |
| "loss": 0.7813, |
| "step": 648 |
| }, |
| { |
| "epoch": 0.23668854850474105, |
| "grad_norm": 0.1613297015428543, |
| "learning_rate": 0.00015294117647058822, |
| "loss": 0.5043, |
| "step": 649 |
| }, |
| { |
| "epoch": 0.23705324580598103, |
| "grad_norm": 0.19487784802913666, |
| "learning_rate": 0.00015286810376324443, |
| "loss": 0.8338, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.237417943107221, |
| "grad_norm": 0.17686185240745544, |
| "learning_rate": 0.00015279503105590064, |
| "loss": 0.6406, |
| "step": 651 |
| }, |
| { |
| "epoch": 0.23778264040846098, |
| "grad_norm": 0.18062898516654968, |
| "learning_rate": 0.00015272195834855682, |
| "loss": 0.772, |
| "step": 652 |
| }, |
| { |
| "epoch": 0.23814733770970095, |
| "grad_norm": 0.1765890270471573, |
| "learning_rate": 0.00015264888564121303, |
| "loss": 0.6313, |
| "step": 653 |
| }, |
| { |
| "epoch": 0.23851203501094093, |
| "grad_norm": 0.18338140845298767, |
| "learning_rate": 0.0001525758129338692, |
| "loss": 0.6944, |
| "step": 654 |
| }, |
| { |
| "epoch": 0.2388767323121809, |
| "grad_norm": 0.16114097833633423, |
| "learning_rate": 0.0001525027402265254, |
| "loss": 0.4307, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.23924142961342085, |
| "grad_norm": 0.18049006164073944, |
| "learning_rate": 0.0001524296675191816, |
| "loss": 0.6055, |
| "step": 656 |
| }, |
| { |
| "epoch": 0.23960612691466082, |
| "grad_norm": 0.17693577706813812, |
| "learning_rate": 0.00015235659481183778, |
| "loss": 0.6953, |
| "step": 657 |
| }, |
| { |
| "epoch": 0.2399708242159008, |
| "grad_norm": 0.17377278208732605, |
| "learning_rate": 0.00015228352210449396, |
| "loss": 0.616, |
| "step": 658 |
| }, |
| { |
| "epoch": 0.24033552151714077, |
| "grad_norm": 0.2092607021331787, |
| "learning_rate": 0.00015221044939715017, |
| "loss": 0.9593, |
| "step": 659 |
| }, |
| { |
| "epoch": 0.24070021881838075, |
| "grad_norm": 0.18305638432502747, |
| "learning_rate": 0.00015213737668980638, |
| "loss": 0.7737, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.24106491611962072, |
| "grad_norm": 0.17872866988182068, |
| "learning_rate": 0.00015206430398246256, |
| "loss": 0.6324, |
| "step": 661 |
| }, |
| { |
| "epoch": 0.2414296134208607, |
| "grad_norm": 0.16150373220443726, |
| "learning_rate": 0.00015199123127511874, |
| "loss": 0.4876, |
| "step": 662 |
| }, |
| { |
| "epoch": 0.24179431072210067, |
| "grad_norm": 0.16832824051380157, |
| "learning_rate": 0.00015191815856777495, |
| "loss": 0.5724, |
| "step": 663 |
| }, |
| { |
| "epoch": 0.24215900802334062, |
| "grad_norm": 0.17251615226268768, |
| "learning_rate": 0.00015184508586043113, |
| "loss": 0.6452, |
| "step": 664 |
| }, |
| { |
| "epoch": 0.2425237053245806, |
| "grad_norm": 0.1856648325920105, |
| "learning_rate": 0.00015177201315308734, |
| "loss": 0.6863, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.24288840262582057, |
| "grad_norm": 0.17599360644817352, |
| "learning_rate": 0.00015169894044574352, |
| "loss": 0.573, |
| "step": 666 |
| }, |
| { |
| "epoch": 0.24325309992706054, |
| "grad_norm": 0.1808532327413559, |
| "learning_rate": 0.0001516258677383997, |
| "loss": 0.6037, |
| "step": 667 |
| }, |
| { |
| "epoch": 0.24361779722830051, |
| "grad_norm": 0.18831709027290344, |
| "learning_rate": 0.0001515527950310559, |
| "loss": 0.8105, |
| "step": 668 |
| }, |
| { |
| "epoch": 0.2439824945295405, |
| "grad_norm": 0.1883680522441864, |
| "learning_rate": 0.0001514797223237121, |
| "loss": 0.7836, |
| "step": 669 |
| }, |
| { |
| "epoch": 0.24434719183078046, |
| "grad_norm": 0.18449294567108154, |
| "learning_rate": 0.0001514066496163683, |
| "loss": 0.74, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.2447118891320204, |
| "grad_norm": 0.14520494639873505, |
| "learning_rate": 0.00015133357690902448, |
| "loss": 0.4257, |
| "step": 671 |
| }, |
| { |
| "epoch": 0.24507658643326038, |
| "grad_norm": 0.16999614238739014, |
| "learning_rate": 0.00015126050420168066, |
| "loss": 0.5913, |
| "step": 672 |
| }, |
| { |
| "epoch": 0.24544128373450036, |
| "grad_norm": 0.19084987044334412, |
| "learning_rate": 0.00015118743149433687, |
| "loss": 0.7187, |
| "step": 673 |
| }, |
| { |
| "epoch": 0.24580598103574033, |
| "grad_norm": 0.16345396637916565, |
| "learning_rate": 0.00015111435878699308, |
| "loss": 0.5708, |
| "step": 674 |
| }, |
| { |
| "epoch": 0.2461706783369803, |
| "grad_norm": 0.19265908002853394, |
| "learning_rate": 0.00015104128607964926, |
| "loss": 0.5983, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.24653537563822028, |
| "grad_norm": 0.20187486708164215, |
| "learning_rate": 0.00015096821337230544, |
| "loss": 0.7784, |
| "step": 676 |
| }, |
| { |
| "epoch": 0.24690007293946026, |
| "grad_norm": 0.19124649465084076, |
| "learning_rate": 0.00015089514066496165, |
| "loss": 0.7033, |
| "step": 677 |
| }, |
| { |
| "epoch": 0.24726477024070023, |
| "grad_norm": 0.18420779705047607, |
| "learning_rate": 0.00015082206795761783, |
| "loss": 0.7342, |
| "step": 678 |
| }, |
| { |
| "epoch": 0.24762946754194018, |
| "grad_norm": 0.18304283916950226, |
| "learning_rate": 0.00015074899525027404, |
| "loss": 0.7119, |
| "step": 679 |
| }, |
| { |
| "epoch": 0.24799416484318015, |
| "grad_norm": 0.17313408851623535, |
| "learning_rate": 0.00015067592254293022, |
| "loss": 0.7298, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.24835886214442013, |
| "grad_norm": 0.1871861219406128, |
| "learning_rate": 0.0001506028498355864, |
| "loss": 0.6186, |
| "step": 681 |
| }, |
| { |
| "epoch": 0.2487235594456601, |
| "grad_norm": 0.20435944199562073, |
| "learning_rate": 0.0001505297771282426, |
| "loss": 0.843, |
| "step": 682 |
| }, |
| { |
| "epoch": 0.24908825674690008, |
| "grad_norm": 0.16381756961345673, |
| "learning_rate": 0.00015045670442089882, |
| "loss": 0.5857, |
| "step": 683 |
| }, |
| { |
| "epoch": 0.24945295404814005, |
| "grad_norm": 0.18176475167274475, |
| "learning_rate": 0.00015038363171355497, |
| "loss": 0.7148, |
| "step": 684 |
| }, |
| { |
| "epoch": 0.24981765134938003, |
| "grad_norm": 0.1591276228427887, |
| "learning_rate": 0.00015031055900621118, |
| "loss": 0.6348, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.25018234865062, |
| "grad_norm": 0.16380302608013153, |
| "learning_rate": 0.0001502374862988674, |
| "loss": 0.5438, |
| "step": 686 |
| }, |
| { |
| "epoch": 0.25054704595185995, |
| "grad_norm": 0.16611479222774506, |
| "learning_rate": 0.00015016441359152357, |
| "loss": 0.509, |
| "step": 687 |
| }, |
| { |
| "epoch": 0.25091174325309995, |
| "grad_norm": 0.188828706741333, |
| "learning_rate": 0.00015009134088417978, |
| "loss": 0.7431, |
| "step": 688 |
| }, |
| { |
| "epoch": 0.2512764405543399, |
| "grad_norm": 0.16808100044727325, |
| "learning_rate": 0.00015001826817683596, |
| "loss": 0.5523, |
| "step": 689 |
| }, |
| { |
| "epoch": 0.25164113785557984, |
| "grad_norm": 0.20357143878936768, |
| "learning_rate": 0.00014994519546949214, |
| "loss": 0.7993, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.25200583515681985, |
| "grad_norm": 0.18831704556941986, |
| "learning_rate": 0.00014987212276214835, |
| "loss": 0.7521, |
| "step": 691 |
| }, |
| { |
| "epoch": 0.2523705324580598, |
| "grad_norm": 0.17688477039337158, |
| "learning_rate": 0.00014979905005480453, |
| "loss": 0.6743, |
| "step": 692 |
| }, |
| { |
| "epoch": 0.2527352297592998, |
| "grad_norm": 0.1944332718849182, |
| "learning_rate": 0.00014972597734746071, |
| "loss": 0.869, |
| "step": 693 |
| }, |
| { |
| "epoch": 0.25309992706053974, |
| "grad_norm": 0.1805860847234726, |
| "learning_rate": 0.00014965290464011692, |
| "loss": 0.7884, |
| "step": 694 |
| }, |
| { |
| "epoch": 0.25346462436177974, |
| "grad_norm": 0.149339959025383, |
| "learning_rate": 0.00014957983193277313, |
| "loss": 0.4521, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.2538293216630197, |
| "grad_norm": 0.16970299184322357, |
| "learning_rate": 0.0001495067592254293, |
| "loss": 0.6262, |
| "step": 696 |
| }, |
| { |
| "epoch": 0.25419401896425964, |
| "grad_norm": 0.17406289279460907, |
| "learning_rate": 0.0001494336865180855, |
| "loss": 0.7258, |
| "step": 697 |
| }, |
| { |
| "epoch": 0.25455871626549964, |
| "grad_norm": 0.18558435142040253, |
| "learning_rate": 0.0001493606138107417, |
| "loss": 0.6835, |
| "step": 698 |
| }, |
| { |
| "epoch": 0.2549234135667396, |
| "grad_norm": 0.16771887242794037, |
| "learning_rate": 0.00014928754110339788, |
| "loss": 0.506, |
| "step": 699 |
| }, |
| { |
| "epoch": 0.2552881108679796, |
| "grad_norm": 0.2056199610233307, |
| "learning_rate": 0.0001492144683960541, |
| "loss": 0.8879, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.25565280816921954, |
| "grad_norm": 0.16422101855278015, |
| "learning_rate": 0.00014914139568871027, |
| "loss": 0.4785, |
| "step": 701 |
| }, |
| { |
| "epoch": 0.25601750547045954, |
| "grad_norm": 0.17392300069332123, |
| "learning_rate": 0.00014906832298136645, |
| "loss": 0.667, |
| "step": 702 |
| }, |
| { |
| "epoch": 0.2563822027716995, |
| "grad_norm": 0.18036240339279175, |
| "learning_rate": 0.00014899525027402266, |
| "loss": 0.6646, |
| "step": 703 |
| }, |
| { |
| "epoch": 0.2567469000729395, |
| "grad_norm": 0.19860827922821045, |
| "learning_rate": 0.00014892217756667884, |
| "loss": 0.8341, |
| "step": 704 |
| }, |
| { |
| "epoch": 0.25711159737417943, |
| "grad_norm": 0.15973514318466187, |
| "learning_rate": 0.00014884910485933505, |
| "loss": 0.5357, |
| "step": 705 |
| }, |
| { |
| "epoch": 0.2574762946754194, |
| "grad_norm": 0.17433969676494598, |
| "learning_rate": 0.00014877603215199123, |
| "loss": 0.6615, |
| "step": 706 |
| }, |
| { |
| "epoch": 0.2578409919766594, |
| "grad_norm": 0.17215828597545624, |
| "learning_rate": 0.00014870295944464742, |
| "loss": 0.547, |
| "step": 707 |
| }, |
| { |
| "epoch": 0.25820568927789933, |
| "grad_norm": 0.19459031522274017, |
| "learning_rate": 0.00014862988673730362, |
| "loss": 0.8744, |
| "step": 708 |
| }, |
| { |
| "epoch": 0.25857038657913933, |
| "grad_norm": 0.18972012400627136, |
| "learning_rate": 0.00014855681402995983, |
| "loss": 0.7687, |
| "step": 709 |
| }, |
| { |
| "epoch": 0.2589350838803793, |
| "grad_norm": 0.17814265191555023, |
| "learning_rate": 0.00014848374132261599, |
| "loss": 0.627, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.2592997811816193, |
| "grad_norm": 0.19156868755817413, |
| "learning_rate": 0.0001484106686152722, |
| "loss": 0.882, |
| "step": 711 |
| }, |
| { |
| "epoch": 0.25966447848285923, |
| "grad_norm": 0.17542894184589386, |
| "learning_rate": 0.0001483375959079284, |
| "loss": 0.7863, |
| "step": 712 |
| }, |
| { |
| "epoch": 0.2600291757840992, |
| "grad_norm": 0.15451830625534058, |
| "learning_rate": 0.00014826452320058458, |
| "loss": 0.5987, |
| "step": 713 |
| }, |
| { |
| "epoch": 0.2603938730853392, |
| "grad_norm": 0.17926959693431854, |
| "learning_rate": 0.0001481914504932408, |
| "loss": 0.6238, |
| "step": 714 |
| }, |
| { |
| "epoch": 0.2607585703865791, |
| "grad_norm": 0.15617339313030243, |
| "learning_rate": 0.00014811837778589697, |
| "loss": 0.611, |
| "step": 715 |
| }, |
| { |
| "epoch": 0.2611232676878191, |
| "grad_norm": 0.18667645752429962, |
| "learning_rate": 0.00014804530507855316, |
| "loss": 0.657, |
| "step": 716 |
| }, |
| { |
| "epoch": 0.2614879649890591, |
| "grad_norm": 0.1547141671180725, |
| "learning_rate": 0.00014797223237120936, |
| "loss": 0.4469, |
| "step": 717 |
| }, |
| { |
| "epoch": 0.2618526622902991, |
| "grad_norm": 0.18968114256858826, |
| "learning_rate": 0.00014789915966386557, |
| "loss": 0.8392, |
| "step": 718 |
| }, |
| { |
| "epoch": 0.262217359591539, |
| "grad_norm": 0.1685071587562561, |
| "learning_rate": 0.00014782608695652173, |
| "loss": 0.5476, |
| "step": 719 |
| }, |
| { |
| "epoch": 0.26258205689277897, |
| "grad_norm": 0.18805775046348572, |
| "learning_rate": 0.00014775301424917793, |
| "loss": 0.7899, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.262946754194019, |
| "grad_norm": 0.1985326111316681, |
| "learning_rate": 0.00014767994154183414, |
| "loss": 0.8315, |
| "step": 721 |
| }, |
| { |
| "epoch": 0.2633114514952589, |
| "grad_norm": 0.18817968666553497, |
| "learning_rate": 0.00014760686883449032, |
| "loss": 0.8077, |
| "step": 722 |
| }, |
| { |
| "epoch": 0.2636761487964989, |
| "grad_norm": 0.1707736700773239, |
| "learning_rate": 0.00014753379612714653, |
| "loss": 0.5055, |
| "step": 723 |
| }, |
| { |
| "epoch": 0.26404084609773887, |
| "grad_norm": 0.17432935535907745, |
| "learning_rate": 0.00014746072341980271, |
| "loss": 0.6428, |
| "step": 724 |
| }, |
| { |
| "epoch": 0.26440554339897887, |
| "grad_norm": 0.19772885739803314, |
| "learning_rate": 0.0001473876507124589, |
| "loss": 0.8951, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.2647702407002188, |
| "grad_norm": 0.18301154673099518, |
| "learning_rate": 0.0001473145780051151, |
| "loss": 0.8608, |
| "step": 726 |
| }, |
| { |
| "epoch": 0.26513493800145876, |
| "grad_norm": 0.17923039197921753, |
| "learning_rate": 0.00014724150529777129, |
| "loss": 0.674, |
| "step": 727 |
| }, |
| { |
| "epoch": 0.26549963530269877, |
| "grad_norm": 0.1921747922897339, |
| "learning_rate": 0.00014716843259042747, |
| "loss": 0.7752, |
| "step": 728 |
| }, |
| { |
| "epoch": 0.2658643326039387, |
| "grad_norm": 0.17289309203624725, |
| "learning_rate": 0.00014709535988308367, |
| "loss": 0.7519, |
| "step": 729 |
| }, |
| { |
| "epoch": 0.2662290299051787, |
| "grad_norm": 0.1981101930141449, |
| "learning_rate": 0.00014702228717573988, |
| "loss": 0.8573, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.26659372720641866, |
| "grad_norm": 0.16859839856624603, |
| "learning_rate": 0.00014694921446839606, |
| "loss": 0.6, |
| "step": 731 |
| }, |
| { |
| "epoch": 0.26695842450765866, |
| "grad_norm": 0.15981674194335938, |
| "learning_rate": 0.00014687614176105225, |
| "loss": 0.434, |
| "step": 732 |
| }, |
| { |
| "epoch": 0.2673231218088986, |
| "grad_norm": 0.15087354183197021, |
| "learning_rate": 0.00014680306905370845, |
| "loss": 0.4438, |
| "step": 733 |
| }, |
| { |
| "epoch": 0.26768781911013856, |
| "grad_norm": 0.21009470522403717, |
| "learning_rate": 0.00014672999634636464, |
| "loss": 0.92, |
| "step": 734 |
| }, |
| { |
| "epoch": 0.26805251641137856, |
| "grad_norm": 0.17488998174667358, |
| "learning_rate": 0.00014665692363902084, |
| "loss": 0.6819, |
| "step": 735 |
| }, |
| { |
| "epoch": 0.2684172137126185, |
| "grad_norm": 0.19854167103767395, |
| "learning_rate": 0.00014658385093167703, |
| "loss": 0.832, |
| "step": 736 |
| }, |
| { |
| "epoch": 0.2687819110138585, |
| "grad_norm": 0.17115044593811035, |
| "learning_rate": 0.0001465107782243332, |
| "loss": 0.6293, |
| "step": 737 |
| }, |
| { |
| "epoch": 0.26914660831509846, |
| "grad_norm": 0.14838315546512604, |
| "learning_rate": 0.00014643770551698941, |
| "loss": 0.3817, |
| "step": 738 |
| }, |
| { |
| "epoch": 0.26951130561633846, |
| "grad_norm": 0.1810576319694519, |
| "learning_rate": 0.0001463646328096456, |
| "loss": 0.6395, |
| "step": 739 |
| }, |
| { |
| "epoch": 0.2698760029175784, |
| "grad_norm": 0.2036665976047516, |
| "learning_rate": 0.0001462915601023018, |
| "loss": 0.855, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.2702407002188184, |
| "grad_norm": 0.1664070338010788, |
| "learning_rate": 0.00014621848739495799, |
| "loss": 0.6324, |
| "step": 741 |
| }, |
| { |
| "epoch": 0.27060539752005836, |
| "grad_norm": 0.21777962148189545, |
| "learning_rate": 0.00014614541468761417, |
| "loss": 0.8487, |
| "step": 742 |
| }, |
| { |
| "epoch": 0.2709700948212983, |
| "grad_norm": 0.16543632745742798, |
| "learning_rate": 0.00014607234198027038, |
| "loss": 0.5855, |
| "step": 743 |
| }, |
| { |
| "epoch": 0.2713347921225383, |
| "grad_norm": 0.18875911831855774, |
| "learning_rate": 0.00014599926927292658, |
| "loss": 0.7444, |
| "step": 744 |
| }, |
| { |
| "epoch": 0.27169948942377825, |
| "grad_norm": 0.19937555491924286, |
| "learning_rate": 0.00014592619656558274, |
| "loss": 0.8593, |
| "step": 745 |
| }, |
| { |
| "epoch": 0.27206418672501825, |
| "grad_norm": 0.17685411870479584, |
| "learning_rate": 0.00014585312385823895, |
| "loss": 0.7563, |
| "step": 746 |
| }, |
| { |
| "epoch": 0.2724288840262582, |
| "grad_norm": 0.16490668058395386, |
| "learning_rate": 0.00014578005115089515, |
| "loss": 0.6355, |
| "step": 747 |
| }, |
| { |
| "epoch": 0.2727935813274982, |
| "grad_norm": 0.17879438400268555, |
| "learning_rate": 0.00014570697844355134, |
| "loss": 0.7345, |
| "step": 748 |
| }, |
| { |
| "epoch": 0.27315827862873815, |
| "grad_norm": 0.15404187142848969, |
| "learning_rate": 0.00014563390573620754, |
| "loss": 0.4345, |
| "step": 749 |
| }, |
| { |
| "epoch": 0.2735229759299781, |
| "grad_norm": 0.18319928646087646, |
| "learning_rate": 0.00014556083302886373, |
| "loss": 0.6891, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.2738876732312181, |
| "grad_norm": 0.173502117395401, |
| "learning_rate": 0.0001454877603215199, |
| "loss": 0.6031, |
| "step": 751 |
| }, |
| { |
| "epoch": 0.27425237053245805, |
| "grad_norm": 0.16543245315551758, |
| "learning_rate": 0.00014541468761417612, |
| "loss": 0.5698, |
| "step": 752 |
| }, |
| { |
| "epoch": 0.27461706783369805, |
| "grad_norm": 0.16900089383125305, |
| "learning_rate": 0.00014534161490683232, |
| "loss": 0.7195, |
| "step": 753 |
| }, |
| { |
| "epoch": 0.274981765134938, |
| "grad_norm": 0.1920769065618515, |
| "learning_rate": 0.00014526854219948848, |
| "loss": 0.7794, |
| "step": 754 |
| }, |
| { |
| "epoch": 0.275346462436178, |
| "grad_norm": 0.15764226019382477, |
| "learning_rate": 0.0001451954694921447, |
| "loss": 0.5931, |
| "step": 755 |
| }, |
| { |
| "epoch": 0.27571115973741794, |
| "grad_norm": 0.16820603609085083, |
| "learning_rate": 0.0001451223967848009, |
| "loss": 0.5739, |
| "step": 756 |
| }, |
| { |
| "epoch": 0.2760758570386579, |
| "grad_norm": 0.15538586676120758, |
| "learning_rate": 0.00014504932407745708, |
| "loss": 0.4564, |
| "step": 757 |
| }, |
| { |
| "epoch": 0.2764405543398979, |
| "grad_norm": 0.1703750640153885, |
| "learning_rate": 0.00014497625137011328, |
| "loss": 0.6289, |
| "step": 758 |
| }, |
| { |
| "epoch": 0.27680525164113784, |
| "grad_norm": 0.17181243002414703, |
| "learning_rate": 0.00014490317866276947, |
| "loss": 0.6888, |
| "step": 759 |
| }, |
| { |
| "epoch": 0.27716994894237784, |
| "grad_norm": 0.17971667647361755, |
| "learning_rate": 0.00014483010595542565, |
| "loss": 0.6178, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.2775346462436178, |
| "grad_norm": 0.16959045827388763, |
| "learning_rate": 0.00014475703324808186, |
| "loss": 0.6029, |
| "step": 761 |
| }, |
| { |
| "epoch": 0.2778993435448578, |
| "grad_norm": 0.18184369802474976, |
| "learning_rate": 0.00014468396054073804, |
| "loss": 0.769, |
| "step": 762 |
| }, |
| { |
| "epoch": 0.27826404084609774, |
| "grad_norm": 0.18276362121105194, |
| "learning_rate": 0.00014461088783339422, |
| "loss": 0.6981, |
| "step": 763 |
| }, |
| { |
| "epoch": 0.2786287381473377, |
| "grad_norm": 0.14473621547222137, |
| "learning_rate": 0.00014453781512605043, |
| "loss": 0.4821, |
| "step": 764 |
| }, |
| { |
| "epoch": 0.2789934354485777, |
| "grad_norm": 0.16732734441757202, |
| "learning_rate": 0.00014446474241870664, |
| "loss": 0.5346, |
| "step": 765 |
| }, |
| { |
| "epoch": 0.27935813274981763, |
| "grad_norm": 0.17172355949878693, |
| "learning_rate": 0.00014439166971136282, |
| "loss": 0.6361, |
| "step": 766 |
| }, |
| { |
| "epoch": 0.27972283005105764, |
| "grad_norm": 0.19567203521728516, |
| "learning_rate": 0.000144318597004019, |
| "loss": 0.859, |
| "step": 767 |
| }, |
| { |
| "epoch": 0.2800875273522976, |
| "grad_norm": 0.1898382008075714, |
| "learning_rate": 0.0001442455242966752, |
| "loss": 0.8217, |
| "step": 768 |
| }, |
| { |
| "epoch": 0.2804522246535376, |
| "grad_norm": 0.1976533979177475, |
| "learning_rate": 0.0001441724515893314, |
| "loss": 0.8709, |
| "step": 769 |
| }, |
| { |
| "epoch": 0.28081692195477753, |
| "grad_norm": 0.15758675336837769, |
| "learning_rate": 0.0001440993788819876, |
| "loss": 0.5522, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.2811816192560175, |
| "grad_norm": 0.19185684621334076, |
| "learning_rate": 0.00014402630617464378, |
| "loss": 0.7182, |
| "step": 771 |
| }, |
| { |
| "epoch": 0.2815463165572575, |
| "grad_norm": 0.16831618547439575, |
| "learning_rate": 0.00014395323346729996, |
| "loss": 0.5841, |
| "step": 772 |
| }, |
| { |
| "epoch": 0.28191101385849743, |
| "grad_norm": 0.207386776804924, |
| "learning_rate": 0.00014388016075995617, |
| "loss": 0.8525, |
| "step": 773 |
| }, |
| { |
| "epoch": 0.28227571115973743, |
| "grad_norm": 0.17020228505134583, |
| "learning_rate": 0.00014380708805261235, |
| "loss": 0.5534, |
| "step": 774 |
| }, |
| { |
| "epoch": 0.2826404084609774, |
| "grad_norm": 0.18354672193527222, |
| "learning_rate": 0.00014373401534526856, |
| "loss": 0.7006, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.2830051057622174, |
| "grad_norm": 0.16657932102680206, |
| "learning_rate": 0.00014366094263792474, |
| "loss": 0.6063, |
| "step": 776 |
| }, |
| { |
| "epoch": 0.2833698030634573, |
| "grad_norm": 0.15975604951381683, |
| "learning_rate": 0.00014358786993058092, |
| "loss": 0.4616, |
| "step": 777 |
| }, |
| { |
| "epoch": 0.28373450036469733, |
| "grad_norm": 0.17766812443733215, |
| "learning_rate": 0.00014351479722323713, |
| "loss": 0.7521, |
| "step": 778 |
| }, |
| { |
| "epoch": 0.2840991976659373, |
| "grad_norm": 0.18797791004180908, |
| "learning_rate": 0.00014344172451589334, |
| "loss": 0.7563, |
| "step": 779 |
| }, |
| { |
| "epoch": 0.2844638949671772, |
| "grad_norm": 0.17721839249134064, |
| "learning_rate": 0.0001433686518085495, |
| "loss": 0.8557, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.2848285922684172, |
| "grad_norm": 0.18058152496814728, |
| "learning_rate": 0.0001432955791012057, |
| "loss": 0.6517, |
| "step": 781 |
| }, |
| { |
| "epoch": 0.28519328956965717, |
| "grad_norm": 0.1826111078262329, |
| "learning_rate": 0.0001432225063938619, |
| "loss": 0.6134, |
| "step": 782 |
| }, |
| { |
| "epoch": 0.2855579868708972, |
| "grad_norm": 0.1896212249994278, |
| "learning_rate": 0.0001431494336865181, |
| "loss": 0.7902, |
| "step": 783 |
| }, |
| { |
| "epoch": 0.2859226841721371, |
| "grad_norm": 0.1763574331998825, |
| "learning_rate": 0.0001430763609791743, |
| "loss": 0.6796, |
| "step": 784 |
| }, |
| { |
| "epoch": 0.2862873814733771, |
| "grad_norm": 0.16747136414051056, |
| "learning_rate": 0.00014300328827183048, |
| "loss": 0.5986, |
| "step": 785 |
| }, |
| { |
| "epoch": 0.28665207877461707, |
| "grad_norm": 0.18107502162456512, |
| "learning_rate": 0.00014293021556448666, |
| "loss": 0.6232, |
| "step": 786 |
| }, |
| { |
| "epoch": 0.287016776075857, |
| "grad_norm": 0.17364108562469482, |
| "learning_rate": 0.00014285714285714287, |
| "loss": 0.6362, |
| "step": 787 |
| }, |
| { |
| "epoch": 0.287381473377097, |
| "grad_norm": 0.21109223365783691, |
| "learning_rate": 0.00014278407014979908, |
| "loss": 0.9846, |
| "step": 788 |
| }, |
| { |
| "epoch": 0.28774617067833697, |
| "grad_norm": 0.1875358521938324, |
| "learning_rate": 0.00014271099744245523, |
| "loss": 0.7414, |
| "step": 789 |
| }, |
| { |
| "epoch": 0.28811086797957697, |
| "grad_norm": 0.17905642092227936, |
| "learning_rate": 0.00014263792473511144, |
| "loss": 0.875, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.2884755652808169, |
| "grad_norm": 0.16830816864967346, |
| "learning_rate": 0.00014256485202776765, |
| "loss": 0.6712, |
| "step": 791 |
| }, |
| { |
| "epoch": 0.2888402625820569, |
| "grad_norm": 0.1730622798204422, |
| "learning_rate": 0.00014249177932042383, |
| "loss": 0.5806, |
| "step": 792 |
| }, |
| { |
| "epoch": 0.28920495988329686, |
| "grad_norm": 0.15828505158424377, |
| "learning_rate": 0.00014241870661308, |
| "loss": 0.4814, |
| "step": 793 |
| }, |
| { |
| "epoch": 0.2895696571845368, |
| "grad_norm": 0.16246497631072998, |
| "learning_rate": 0.00014234563390573622, |
| "loss": 0.4964, |
| "step": 794 |
| }, |
| { |
| "epoch": 0.2899343544857768, |
| "grad_norm": 0.1828710287809372, |
| "learning_rate": 0.0001422725611983924, |
| "loss": 0.7139, |
| "step": 795 |
| }, |
| { |
| "epoch": 0.29029905178701676, |
| "grad_norm": 0.15251024067401886, |
| "learning_rate": 0.0001421994884910486, |
| "loss": 0.4463, |
| "step": 796 |
| }, |
| { |
| "epoch": 0.29066374908825676, |
| "grad_norm": 0.17780308425426483, |
| "learning_rate": 0.0001421264157837048, |
| "loss": 0.6125, |
| "step": 797 |
| }, |
| { |
| "epoch": 0.2910284463894967, |
| "grad_norm": 0.19187119603157043, |
| "learning_rate": 0.00014205334307636097, |
| "loss": 0.814, |
| "step": 798 |
| }, |
| { |
| "epoch": 0.2913931436907367, |
| "grad_norm": 0.20423109829425812, |
| "learning_rate": 0.00014198027036901718, |
| "loss": 0.8008, |
| "step": 799 |
| }, |
| { |
| "epoch": 0.29175784099197666, |
| "grad_norm": 0.17211808264255524, |
| "learning_rate": 0.0001419071976616734, |
| "loss": 0.5667, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.2921225382932166, |
| "grad_norm": 0.20684252679347992, |
| "learning_rate": 0.00014183412495432957, |
| "loss": 0.7646, |
| "step": 801 |
| }, |
| { |
| "epoch": 0.2924872355944566, |
| "grad_norm": 0.15838950872421265, |
| "learning_rate": 0.00014176105224698575, |
| "loss": 0.5542, |
| "step": 802 |
| }, |
| { |
| "epoch": 0.29285193289569655, |
| "grad_norm": 0.19353005290031433, |
| "learning_rate": 0.00014168797953964196, |
| "loss": 0.7211, |
| "step": 803 |
| }, |
| { |
| "epoch": 0.29321663019693656, |
| "grad_norm": 0.1527222841978073, |
| "learning_rate": 0.00014161490683229814, |
| "loss": 0.5183, |
| "step": 804 |
| }, |
| { |
| "epoch": 0.2935813274981765, |
| "grad_norm": 0.15716342628002167, |
| "learning_rate": 0.00014154183412495435, |
| "loss": 0.5197, |
| "step": 805 |
| }, |
| { |
| "epoch": 0.2939460247994165, |
| "grad_norm": 0.2023271769285202, |
| "learning_rate": 0.00014146876141761053, |
| "loss": 0.8512, |
| "step": 806 |
| }, |
| { |
| "epoch": 0.29431072210065645, |
| "grad_norm": 0.17139078676700592, |
| "learning_rate": 0.0001413956887102667, |
| "loss": 0.5161, |
| "step": 807 |
| }, |
| { |
| "epoch": 0.2946754194018964, |
| "grad_norm": 0.15848784148693085, |
| "learning_rate": 0.00014132261600292292, |
| "loss": 0.6017, |
| "step": 808 |
| }, |
| { |
| "epoch": 0.2950401167031364, |
| "grad_norm": 0.1857565939426422, |
| "learning_rate": 0.0001412495432955791, |
| "loss": 0.7406, |
| "step": 809 |
| }, |
| { |
| "epoch": 0.29540481400437635, |
| "grad_norm": 0.17538581788539886, |
| "learning_rate": 0.0001411764705882353, |
| "loss": 0.615, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.29576951130561635, |
| "grad_norm": 0.18965482711791992, |
| "learning_rate": 0.0001411033978808915, |
| "loss": 0.7869, |
| "step": 811 |
| }, |
| { |
| "epoch": 0.2961342086068563, |
| "grad_norm": 0.1757965087890625, |
| "learning_rate": 0.00014103032517354767, |
| "loss": 0.7217, |
| "step": 812 |
| }, |
| { |
| "epoch": 0.2964989059080963, |
| "grad_norm": 0.20843826234340668, |
| "learning_rate": 0.00014095725246620388, |
| "loss": 0.9802, |
| "step": 813 |
| }, |
| { |
| "epoch": 0.29686360320933625, |
| "grad_norm": 0.17155392467975616, |
| "learning_rate": 0.0001408841797588601, |
| "loss": 0.67, |
| "step": 814 |
| }, |
| { |
| "epoch": 0.29722830051057625, |
| "grad_norm": 0.1699131578207016, |
| "learning_rate": 0.00014081110705151624, |
| "loss": 0.5432, |
| "step": 815 |
| }, |
| { |
| "epoch": 0.2975929978118162, |
| "grad_norm": 0.2160065770149231, |
| "learning_rate": 0.00014073803434417245, |
| "loss": 0.9022, |
| "step": 816 |
| }, |
| { |
| "epoch": 0.29795769511305614, |
| "grad_norm": 0.17137810587882996, |
| "learning_rate": 0.00014066496163682866, |
| "loss": 0.5853, |
| "step": 817 |
| }, |
| { |
| "epoch": 0.29832239241429614, |
| "grad_norm": 0.18035869300365448, |
| "learning_rate": 0.00014059188892948484, |
| "loss": 0.6308, |
| "step": 818 |
| }, |
| { |
| "epoch": 0.2986870897155361, |
| "grad_norm": 0.15042151510715485, |
| "learning_rate": 0.00014051881622214105, |
| "loss": 0.4335, |
| "step": 819 |
| }, |
| { |
| "epoch": 0.2990517870167761, |
| "grad_norm": 0.19387514889240265, |
| "learning_rate": 0.00014044574351479723, |
| "loss": 0.8104, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.29941648431801604, |
| "grad_norm": 0.1495147943496704, |
| "learning_rate": 0.0001403726708074534, |
| "loss": 0.4551, |
| "step": 821 |
| }, |
| { |
| "epoch": 0.29978118161925604, |
| "grad_norm": 0.1305743157863617, |
| "learning_rate": 0.00014029959810010962, |
| "loss": 0.3016, |
| "step": 822 |
| }, |
| { |
| "epoch": 0.300145878920496, |
| "grad_norm": 0.18238870799541473, |
| "learning_rate": 0.00014022652539276583, |
| "loss": 0.6105, |
| "step": 823 |
| }, |
| { |
| "epoch": 0.30051057622173594, |
| "grad_norm": 0.1913590431213379, |
| "learning_rate": 0.00014015345268542198, |
| "loss": 0.8108, |
| "step": 824 |
| }, |
| { |
| "epoch": 0.30087527352297594, |
| "grad_norm": 0.15131442248821259, |
| "learning_rate": 0.0001400803799780782, |
| "loss": 0.5041, |
| "step": 825 |
| }, |
| { |
| "epoch": 0.3012399708242159, |
| "grad_norm": 0.15665921568870544, |
| "learning_rate": 0.0001400073072707344, |
| "loss": 0.5691, |
| "step": 826 |
| }, |
| { |
| "epoch": 0.3016046681254559, |
| "grad_norm": 0.1675909459590912, |
| "learning_rate": 0.00013993423456339058, |
| "loss": 0.6212, |
| "step": 827 |
| }, |
| { |
| "epoch": 0.30196936542669583, |
| "grad_norm": 0.2064395546913147, |
| "learning_rate": 0.00013986116185604676, |
| "loss": 0.8267, |
| "step": 828 |
| }, |
| { |
| "epoch": 0.30233406272793584, |
| "grad_norm": 0.18026086688041687, |
| "learning_rate": 0.00013978808914870297, |
| "loss": 0.7102, |
| "step": 829 |
| }, |
| { |
| "epoch": 0.3026987600291758, |
| "grad_norm": 0.18619216978549957, |
| "learning_rate": 0.00013971501644135915, |
| "loss": 0.5986, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.30306345733041573, |
| "grad_norm": 0.20318515598773956, |
| "learning_rate": 0.00013964194373401536, |
| "loss": 0.9558, |
| "step": 831 |
| }, |
| { |
| "epoch": 0.30342815463165573, |
| "grad_norm": 0.1473841667175293, |
| "learning_rate": 0.00013956887102667154, |
| "loss": 0.3906, |
| "step": 832 |
| }, |
| { |
| "epoch": 0.3037928519328957, |
| "grad_norm": 0.1872483193874359, |
| "learning_rate": 0.00013949579831932772, |
| "loss": 0.7164, |
| "step": 833 |
| }, |
| { |
| "epoch": 0.3041575492341357, |
| "grad_norm": 0.1955832690000534, |
| "learning_rate": 0.00013942272561198393, |
| "loss": 0.7684, |
| "step": 834 |
| }, |
| { |
| "epoch": 0.30452224653537563, |
| "grad_norm": 0.17416808009147644, |
| "learning_rate": 0.00013934965290464014, |
| "loss": 0.6543, |
| "step": 835 |
| }, |
| { |
| "epoch": 0.30488694383661563, |
| "grad_norm": 0.18365338444709778, |
| "learning_rate": 0.00013927658019729632, |
| "loss": 0.6797, |
| "step": 836 |
| }, |
| { |
| "epoch": 0.3052516411378556, |
| "grad_norm": 0.159471333026886, |
| "learning_rate": 0.0001392035074899525, |
| "loss": 0.5944, |
| "step": 837 |
| }, |
| { |
| "epoch": 0.3056163384390955, |
| "grad_norm": 0.1895028054714203, |
| "learning_rate": 0.0001391304347826087, |
| "loss": 0.7348, |
| "step": 838 |
| }, |
| { |
| "epoch": 0.3059810357403355, |
| "grad_norm": 0.19214889407157898, |
| "learning_rate": 0.0001390573620752649, |
| "loss": 0.8888, |
| "step": 839 |
| }, |
| { |
| "epoch": 0.3063457330415755, |
| "grad_norm": 0.18355461955070496, |
| "learning_rate": 0.0001389842893679211, |
| "loss": 0.7066, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.3067104303428155, |
| "grad_norm": 0.20794498920440674, |
| "learning_rate": 0.00013891121666057728, |
| "loss": 0.7818, |
| "step": 841 |
| }, |
| { |
| "epoch": 0.3070751276440554, |
| "grad_norm": 0.20993681252002716, |
| "learning_rate": 0.00013883814395323346, |
| "loss": 0.8961, |
| "step": 842 |
| }, |
| { |
| "epoch": 0.3074398249452954, |
| "grad_norm": 0.17933285236358643, |
| "learning_rate": 0.00013876507124588967, |
| "loss": 0.7162, |
| "step": 843 |
| }, |
| { |
| "epoch": 0.3078045222465354, |
| "grad_norm": 0.1597253382205963, |
| "learning_rate": 0.00013869199853854585, |
| "loss": 0.4908, |
| "step": 844 |
| }, |
| { |
| "epoch": 0.3081692195477753, |
| "grad_norm": 0.17529501020908356, |
| "learning_rate": 0.00013861892583120206, |
| "loss": 0.5254, |
| "step": 845 |
| }, |
| { |
| "epoch": 0.3085339168490153, |
| "grad_norm": 0.19237765669822693, |
| "learning_rate": 0.00013854585312385824, |
| "loss": 0.7377, |
| "step": 846 |
| }, |
| { |
| "epoch": 0.30889861415025527, |
| "grad_norm": 0.18043018877506256, |
| "learning_rate": 0.00013847278041651442, |
| "loss": 0.6784, |
| "step": 847 |
| }, |
| { |
| "epoch": 0.30926331145149527, |
| "grad_norm": 0.1529788225889206, |
| "learning_rate": 0.00013839970770917063, |
| "loss": 0.5048, |
| "step": 848 |
| }, |
| { |
| "epoch": 0.3096280087527352, |
| "grad_norm": 0.18892444670200348, |
| "learning_rate": 0.00013832663500182684, |
| "loss": 0.7093, |
| "step": 849 |
| }, |
| { |
| "epoch": 0.3099927060539752, |
| "grad_norm": 0.17523162066936493, |
| "learning_rate": 0.000138253562294483, |
| "loss": 0.6947, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.31035740335521517, |
| "grad_norm": 0.17124755680561066, |
| "learning_rate": 0.0001381804895871392, |
| "loss": 0.6009, |
| "step": 851 |
| }, |
| { |
| "epoch": 0.31072210065645517, |
| "grad_norm": 0.19225868582725525, |
| "learning_rate": 0.0001381074168797954, |
| "loss": 0.8058, |
| "step": 852 |
| }, |
| { |
| "epoch": 0.3110867979576951, |
| "grad_norm": 0.19115358591079712, |
| "learning_rate": 0.0001380343441724516, |
| "loss": 0.6689, |
| "step": 853 |
| }, |
| { |
| "epoch": 0.31145149525893506, |
| "grad_norm": 0.15115682780742645, |
| "learning_rate": 0.0001379612714651078, |
| "loss": 0.3348, |
| "step": 854 |
| }, |
| { |
| "epoch": 0.31181619256017507, |
| "grad_norm": 0.18653741478919983, |
| "learning_rate": 0.00013788819875776398, |
| "loss": 0.872, |
| "step": 855 |
| }, |
| { |
| "epoch": 0.312180889861415, |
| "grad_norm": 0.18661542236804962, |
| "learning_rate": 0.00013781512605042016, |
| "loss": 0.9351, |
| "step": 856 |
| }, |
| { |
| "epoch": 0.312545587162655, |
| "grad_norm": 0.1910124570131302, |
| "learning_rate": 0.00013774205334307637, |
| "loss": 0.8575, |
| "step": 857 |
| }, |
| { |
| "epoch": 0.31291028446389496, |
| "grad_norm": 0.18687231838703156, |
| "learning_rate": 0.00013766898063573258, |
| "loss": 0.7134, |
| "step": 858 |
| }, |
| { |
| "epoch": 0.31327498176513496, |
| "grad_norm": 0.17739079892635345, |
| "learning_rate": 0.00013759590792838873, |
| "loss": 0.6473, |
| "step": 859 |
| }, |
| { |
| "epoch": 0.3136396790663749, |
| "grad_norm": 0.16609562933444977, |
| "learning_rate": 0.00013752283522104494, |
| "loss": 0.511, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.31400437636761486, |
| "grad_norm": 0.17272962629795074, |
| "learning_rate": 0.00013744976251370115, |
| "loss": 0.5939, |
| "step": 861 |
| }, |
| { |
| "epoch": 0.31436907366885486, |
| "grad_norm": 0.196741983294487, |
| "learning_rate": 0.00013737668980635733, |
| "loss": 0.6241, |
| "step": 862 |
| }, |
| { |
| "epoch": 0.3147337709700948, |
| "grad_norm": 0.16731449961662292, |
| "learning_rate": 0.00013730361709901351, |
| "loss": 0.6294, |
| "step": 863 |
| }, |
| { |
| "epoch": 0.3150984682713348, |
| "grad_norm": 0.17899174988269806, |
| "learning_rate": 0.00013723054439166972, |
| "loss": 0.724, |
| "step": 864 |
| }, |
| { |
| "epoch": 0.31546316557257476, |
| "grad_norm": 0.18915481865406036, |
| "learning_rate": 0.0001371574716843259, |
| "loss": 0.6508, |
| "step": 865 |
| }, |
| { |
| "epoch": 0.31582786287381476, |
| "grad_norm": 0.17585769295692444, |
| "learning_rate": 0.0001370843989769821, |
| "loss": 0.736, |
| "step": 866 |
| }, |
| { |
| "epoch": 0.3161925601750547, |
| "grad_norm": 0.16982930898666382, |
| "learning_rate": 0.0001370113262696383, |
| "loss": 0.5676, |
| "step": 867 |
| }, |
| { |
| "epoch": 0.31655725747629465, |
| "grad_norm": 0.17551635205745697, |
| "learning_rate": 0.00013693825356229447, |
| "loss": 0.621, |
| "step": 868 |
| }, |
| { |
| "epoch": 0.31692195477753465, |
| "grad_norm": 0.19762268662452698, |
| "learning_rate": 0.00013686518085495068, |
| "loss": 0.872, |
| "step": 869 |
| }, |
| { |
| "epoch": 0.3172866520787746, |
| "grad_norm": 0.16374340653419495, |
| "learning_rate": 0.0001367921081476069, |
| "loss": 0.4612, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.3176513493800146, |
| "grad_norm": 0.20281797647476196, |
| "learning_rate": 0.00013671903544026307, |
| "loss": 0.8791, |
| "step": 871 |
| }, |
| { |
| "epoch": 0.31801604668125455, |
| "grad_norm": 0.1752830445766449, |
| "learning_rate": 0.00013664596273291925, |
| "loss": 0.7158, |
| "step": 872 |
| }, |
| { |
| "epoch": 0.31838074398249455, |
| "grad_norm": 0.1833350956439972, |
| "learning_rate": 0.00013657289002557546, |
| "loss": 0.7307, |
| "step": 873 |
| }, |
| { |
| "epoch": 0.3187454412837345, |
| "grad_norm": 0.158727765083313, |
| "learning_rate": 0.00013649981731823164, |
| "loss": 0.4426, |
| "step": 874 |
| }, |
| { |
| "epoch": 0.31911013858497445, |
| "grad_norm": 0.1881689429283142, |
| "learning_rate": 0.00013642674461088785, |
| "loss": 0.6806, |
| "step": 875 |
| }, |
| { |
| "epoch": 0.31947483588621445, |
| "grad_norm": 0.18638849258422852, |
| "learning_rate": 0.00013635367190354403, |
| "loss": 0.5562, |
| "step": 876 |
| }, |
| { |
| "epoch": 0.3198395331874544, |
| "grad_norm": 0.17578531801700592, |
| "learning_rate": 0.00013628059919620022, |
| "loss": 0.6174, |
| "step": 877 |
| }, |
| { |
| "epoch": 0.3202042304886944, |
| "grad_norm": 0.17400258779525757, |
| "learning_rate": 0.00013620752648885642, |
| "loss": 0.6818, |
| "step": 878 |
| }, |
| { |
| "epoch": 0.32056892778993434, |
| "grad_norm": 0.17320291697978973, |
| "learning_rate": 0.0001361344537815126, |
| "loss": 0.59, |
| "step": 879 |
| }, |
| { |
| "epoch": 0.32093362509117435, |
| "grad_norm": 0.1907418966293335, |
| "learning_rate": 0.0001360613810741688, |
| "loss": 0.8411, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.3212983223924143, |
| "grad_norm": 0.1831212341785431, |
| "learning_rate": 0.000135988308366825, |
| "loss": 0.7538, |
| "step": 881 |
| }, |
| { |
| "epoch": 0.32166301969365424, |
| "grad_norm": 0.16207090020179749, |
| "learning_rate": 0.00013591523565948118, |
| "loss": 0.5343, |
| "step": 882 |
| }, |
| { |
| "epoch": 0.32202771699489424, |
| "grad_norm": 0.16630573570728302, |
| "learning_rate": 0.00013584216295213738, |
| "loss": 0.5422, |
| "step": 883 |
| }, |
| { |
| "epoch": 0.3223924142961342, |
| "grad_norm": 0.16689611971378326, |
| "learning_rate": 0.0001357690902447936, |
| "loss": 0.682, |
| "step": 884 |
| }, |
| { |
| "epoch": 0.3227571115973742, |
| "grad_norm": 0.1957065463066101, |
| "learning_rate": 0.00013569601753744975, |
| "loss": 0.8429, |
| "step": 885 |
| }, |
| { |
| "epoch": 0.32312180889861414, |
| "grad_norm": 0.16777260601520538, |
| "learning_rate": 0.00013562294483010596, |
| "loss": 0.5122, |
| "step": 886 |
| }, |
| { |
| "epoch": 0.32348650619985414, |
| "grad_norm": 0.18913322687149048, |
| "learning_rate": 0.00013554987212276216, |
| "loss": 0.8262, |
| "step": 887 |
| }, |
| { |
| "epoch": 0.3238512035010941, |
| "grad_norm": 0.15186062455177307, |
| "learning_rate": 0.00013547679941541834, |
| "loss": 0.4259, |
| "step": 888 |
| }, |
| { |
| "epoch": 0.3242159008023341, |
| "grad_norm": 0.17760516703128815, |
| "learning_rate": 0.00013540372670807453, |
| "loss": 0.6425, |
| "step": 889 |
| }, |
| { |
| "epoch": 0.32458059810357404, |
| "grad_norm": 0.17440609633922577, |
| "learning_rate": 0.00013533065400073073, |
| "loss": 0.795, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.324945295404814, |
| "grad_norm": 0.1680755466222763, |
| "learning_rate": 0.00013525758129338692, |
| "loss": 0.6218, |
| "step": 891 |
| }, |
| { |
| "epoch": 0.325309992706054, |
| "grad_norm": 0.18546250462532043, |
| "learning_rate": 0.00013518450858604312, |
| "loss": 0.7299, |
| "step": 892 |
| }, |
| { |
| "epoch": 0.32567469000729393, |
| "grad_norm": 0.2105245590209961, |
| "learning_rate": 0.00013511143587869933, |
| "loss": 0.8526, |
| "step": 893 |
| }, |
| { |
| "epoch": 0.32603938730853393, |
| "grad_norm": 0.18736779689788818, |
| "learning_rate": 0.0001350383631713555, |
| "loss": 0.7205, |
| "step": 894 |
| }, |
| { |
| "epoch": 0.3264040846097739, |
| "grad_norm": 0.17326989769935608, |
| "learning_rate": 0.0001349652904640117, |
| "loss": 0.5746, |
| "step": 895 |
| }, |
| { |
| "epoch": 0.3267687819110139, |
| "grad_norm": 0.21187415719032288, |
| "learning_rate": 0.0001348922177566679, |
| "loss": 0.9321, |
| "step": 896 |
| }, |
| { |
| "epoch": 0.32713347921225383, |
| "grad_norm": 0.18079346418380737, |
| "learning_rate": 0.00013481914504932409, |
| "loss": 0.6063, |
| "step": 897 |
| }, |
| { |
| "epoch": 0.3274981765134938, |
| "grad_norm": 0.2063044011592865, |
| "learning_rate": 0.00013474607234198027, |
| "loss": 0.8192, |
| "step": 898 |
| }, |
| { |
| "epoch": 0.3278628738147338, |
| "grad_norm": 0.1921169012784958, |
| "learning_rate": 0.00013467299963463647, |
| "loss": 0.9365, |
| "step": 899 |
| }, |
| { |
| "epoch": 0.3282275711159737, |
| "grad_norm": 0.19340583682060242, |
| "learning_rate": 0.00013459992692729266, |
| "loss": 0.7314, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.32859226841721373, |
| "grad_norm": 0.19069619476795197, |
| "learning_rate": 0.00013452685421994886, |
| "loss": 0.7089, |
| "step": 901 |
| }, |
| { |
| "epoch": 0.3289569657184537, |
| "grad_norm": 0.1918506920337677, |
| "learning_rate": 0.00013445378151260507, |
| "loss": 0.7538, |
| "step": 902 |
| }, |
| { |
| "epoch": 0.3293216630196937, |
| "grad_norm": 0.18830406665802002, |
| "learning_rate": 0.00013438070880526123, |
| "loss": 0.6583, |
| "step": 903 |
| }, |
| { |
| "epoch": 0.3296863603209336, |
| "grad_norm": 0.18953083455562592, |
| "learning_rate": 0.00013430763609791744, |
| "loss": 0.6886, |
| "step": 904 |
| }, |
| { |
| "epoch": 0.33005105762217357, |
| "grad_norm": 0.1635250300168991, |
| "learning_rate": 0.00013423456339057364, |
| "loss": 0.6201, |
| "step": 905 |
| }, |
| { |
| "epoch": 0.3304157549234136, |
| "grad_norm": 0.19504040479660034, |
| "learning_rate": 0.00013416149068322983, |
| "loss": 0.6456, |
| "step": 906 |
| }, |
| { |
| "epoch": 0.3307804522246535, |
| "grad_norm": 0.1988976150751114, |
| "learning_rate": 0.000134088417975886, |
| "loss": 0.7649, |
| "step": 907 |
| }, |
| { |
| "epoch": 0.3311451495258935, |
| "grad_norm": 0.17592492699623108, |
| "learning_rate": 0.00013401534526854221, |
| "loss": 0.6112, |
| "step": 908 |
| }, |
| { |
| "epoch": 0.33150984682713347, |
| "grad_norm": 0.19126677513122559, |
| "learning_rate": 0.0001339422725611984, |
| "loss": 0.7045, |
| "step": 909 |
| }, |
| { |
| "epoch": 0.33187454412837347, |
| "grad_norm": 0.1643197238445282, |
| "learning_rate": 0.0001338691998538546, |
| "loss": 0.4433, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.3322392414296134, |
| "grad_norm": 0.19282923638820648, |
| "learning_rate": 0.00013379612714651079, |
| "loss": 0.7357, |
| "step": 911 |
| }, |
| { |
| "epoch": 0.33260393873085337, |
| "grad_norm": 0.17284700274467468, |
| "learning_rate": 0.00013372305443916697, |
| "loss": 0.8092, |
| "step": 912 |
| }, |
| { |
| "epoch": 0.33296863603209337, |
| "grad_norm": 0.2016737312078476, |
| "learning_rate": 0.00013364998173182318, |
| "loss": 0.8372, |
| "step": 913 |
| }, |
| { |
| "epoch": 0.3333333333333333, |
| "grad_norm": 0.1924041360616684, |
| "learning_rate": 0.00013357690902447936, |
| "loss": 0.7576, |
| "step": 914 |
| }, |
| { |
| "epoch": 0.3336980306345733, |
| "grad_norm": 0.18545666337013245, |
| "learning_rate": 0.00013350383631713557, |
| "loss": 0.7559, |
| "step": 915 |
| }, |
| { |
| "epoch": 0.33406272793581326, |
| "grad_norm": 0.1433466225862503, |
| "learning_rate": 0.00013343076360979175, |
| "loss": 0.4667, |
| "step": 916 |
| }, |
| { |
| "epoch": 0.33442742523705327, |
| "grad_norm": 0.1683466136455536, |
| "learning_rate": 0.00013335769090244793, |
| "loss": 0.5951, |
| "step": 917 |
| }, |
| { |
| "epoch": 0.3347921225382932, |
| "grad_norm": 0.18137226998806, |
| "learning_rate": 0.00013328461819510414, |
| "loss": 0.6964, |
| "step": 918 |
| }, |
| { |
| "epoch": 0.33515681983953316, |
| "grad_norm": 0.19752834737300873, |
| "learning_rate": 0.00013321154548776034, |
| "loss": 0.7344, |
| "step": 919 |
| }, |
| { |
| "epoch": 0.33552151714077316, |
| "grad_norm": 0.17333728075027466, |
| "learning_rate": 0.0001331384727804165, |
| "loss": 0.5658, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.3358862144420131, |
| "grad_norm": 0.17877991497516632, |
| "learning_rate": 0.0001330654000730727, |
| "loss": 0.6542, |
| "step": 921 |
| }, |
| { |
| "epoch": 0.3362509117432531, |
| "grad_norm": 0.15863974392414093, |
| "learning_rate": 0.00013299232736572892, |
| "loss": 0.4601, |
| "step": 922 |
| }, |
| { |
| "epoch": 0.33661560904449306, |
| "grad_norm": 0.20868968963623047, |
| "learning_rate": 0.0001329192546583851, |
| "loss": 0.7883, |
| "step": 923 |
| }, |
| { |
| "epoch": 0.33698030634573306, |
| "grad_norm": 0.20271888375282288, |
| "learning_rate": 0.00013284618195104128, |
| "loss": 0.9605, |
| "step": 924 |
| }, |
| { |
| "epoch": 0.337345003646973, |
| "grad_norm": 0.19558852910995483, |
| "learning_rate": 0.0001327731092436975, |
| "loss": 0.7866, |
| "step": 925 |
| }, |
| { |
| "epoch": 0.337709700948213, |
| "grad_norm": 0.2165563404560089, |
| "learning_rate": 0.00013270003653635367, |
| "loss": 0.9694, |
| "step": 926 |
| }, |
| { |
| "epoch": 0.33807439824945296, |
| "grad_norm": 0.18585020303726196, |
| "learning_rate": 0.00013262696382900988, |
| "loss": 0.7127, |
| "step": 927 |
| }, |
| { |
| "epoch": 0.3384390955506929, |
| "grad_norm": 0.15885639190673828, |
| "learning_rate": 0.00013255389112166608, |
| "loss": 0.4634, |
| "step": 928 |
| }, |
| { |
| "epoch": 0.3388037928519329, |
| "grad_norm": 0.18902234733104706, |
| "learning_rate": 0.00013248081841432224, |
| "loss": 0.7553, |
| "step": 929 |
| }, |
| { |
| "epoch": 0.33916849015317285, |
| "grad_norm": 0.1625453233718872, |
| "learning_rate": 0.00013240774570697845, |
| "loss": 0.6045, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.33953318745441285, |
| "grad_norm": 0.1839369386434555, |
| "learning_rate": 0.00013233467299963466, |
| "loss": 0.6966, |
| "step": 931 |
| }, |
| { |
| "epoch": 0.3398978847556528, |
| "grad_norm": 0.1871074140071869, |
| "learning_rate": 0.00013226160029229084, |
| "loss": 0.7223, |
| "step": 932 |
| }, |
| { |
| "epoch": 0.3402625820568928, |
| "grad_norm": 0.19105811417102814, |
| "learning_rate": 0.00013218852758494702, |
| "loss": 0.8262, |
| "step": 933 |
| }, |
| { |
| "epoch": 0.34062727935813275, |
| "grad_norm": 0.19458365440368652, |
| "learning_rate": 0.00013211545487760323, |
| "loss": 0.6728, |
| "step": 934 |
| }, |
| { |
| "epoch": 0.3409919766593727, |
| "grad_norm": 0.17245818674564362, |
| "learning_rate": 0.0001320423821702594, |
| "loss": 0.6373, |
| "step": 935 |
| }, |
| { |
| "epoch": 0.3413566739606127, |
| "grad_norm": 0.17466460168361664, |
| "learning_rate": 0.00013196930946291562, |
| "loss": 0.6054, |
| "step": 936 |
| }, |
| { |
| "epoch": 0.34172137126185265, |
| "grad_norm": 0.1496109813451767, |
| "learning_rate": 0.0001318962367555718, |
| "loss": 0.5585, |
| "step": 937 |
| }, |
| { |
| "epoch": 0.34208606856309265, |
| "grad_norm": 0.16135789453983307, |
| "learning_rate": 0.00013182316404822798, |
| "loss": 0.4524, |
| "step": 938 |
| }, |
| { |
| "epoch": 0.3424507658643326, |
| "grad_norm": 0.18663141131401062, |
| "learning_rate": 0.0001317500913408842, |
| "loss": 0.6951, |
| "step": 939 |
| }, |
| { |
| "epoch": 0.3428154631655726, |
| "grad_norm": 0.15193338692188263, |
| "learning_rate": 0.0001316770186335404, |
| "loss": 0.5151, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.34318016046681254, |
| "grad_norm": 0.16860604286193848, |
| "learning_rate": 0.00013160394592619658, |
| "loss": 0.5776, |
| "step": 941 |
| }, |
| { |
| "epoch": 0.3435448577680525, |
| "grad_norm": 0.18972420692443848, |
| "learning_rate": 0.00013153087321885276, |
| "loss": 0.7691, |
| "step": 942 |
| }, |
| { |
| "epoch": 0.3439095550692925, |
| "grad_norm": 0.196933776140213, |
| "learning_rate": 0.00013145780051150897, |
| "loss": 0.8165, |
| "step": 943 |
| }, |
| { |
| "epoch": 0.34427425237053244, |
| "grad_norm": 0.19200679659843445, |
| "learning_rate": 0.00013138472780416515, |
| "loss": 0.7874, |
| "step": 944 |
| }, |
| { |
| "epoch": 0.34463894967177244, |
| "grad_norm": 0.1795893758535385, |
| "learning_rate": 0.00013131165509682136, |
| "loss": 0.7507, |
| "step": 945 |
| }, |
| { |
| "epoch": 0.3450036469730124, |
| "grad_norm": 0.18410655856132507, |
| "learning_rate": 0.00013123858238947754, |
| "loss": 0.7354, |
| "step": 946 |
| }, |
| { |
| "epoch": 0.3453683442742524, |
| "grad_norm": 0.1739976406097412, |
| "learning_rate": 0.00013116550968213372, |
| "loss": 0.6932, |
| "step": 947 |
| }, |
| { |
| "epoch": 0.34573304157549234, |
| "grad_norm": 0.14160172641277313, |
| "learning_rate": 0.00013109243697478993, |
| "loss": 0.351, |
| "step": 948 |
| }, |
| { |
| "epoch": 0.3460977388767323, |
| "grad_norm": 0.14415137469768524, |
| "learning_rate": 0.0001310193642674461, |
| "loss": 0.4202, |
| "step": 949 |
| }, |
| { |
| "epoch": 0.3464624361779723, |
| "grad_norm": 0.2061617225408554, |
| "learning_rate": 0.00013094629156010232, |
| "loss": 0.9679, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.34682713347921224, |
| "grad_norm": 0.20319141447544098, |
| "learning_rate": 0.0001308732188527585, |
| "loss": 0.8076, |
| "step": 951 |
| }, |
| { |
| "epoch": 0.34719183078045224, |
| "grad_norm": 0.17571642994880676, |
| "learning_rate": 0.00013080014614541468, |
| "loss": 0.6941, |
| "step": 952 |
| }, |
| { |
| "epoch": 0.3475565280816922, |
| "grad_norm": 0.177334725856781, |
| "learning_rate": 0.0001307270734380709, |
| "loss": 0.7511, |
| "step": 953 |
| }, |
| { |
| "epoch": 0.3479212253829322, |
| "grad_norm": 0.2112066000699997, |
| "learning_rate": 0.0001306540007307271, |
| "loss": 1.0981, |
| "step": 954 |
| }, |
| { |
| "epoch": 0.34828592268417213, |
| "grad_norm": 0.18469132483005524, |
| "learning_rate": 0.00013058092802338325, |
| "loss": 0.8159, |
| "step": 955 |
| }, |
| { |
| "epoch": 0.3486506199854121, |
| "grad_norm": 0.17193461954593658, |
| "learning_rate": 0.00013050785531603946, |
| "loss": 0.6005, |
| "step": 956 |
| }, |
| { |
| "epoch": 0.3490153172866521, |
| "grad_norm": 0.21006590127944946, |
| "learning_rate": 0.00013043478260869567, |
| "loss": 1.0101, |
| "step": 957 |
| }, |
| { |
| "epoch": 0.34938001458789203, |
| "grad_norm": 0.1526053547859192, |
| "learning_rate": 0.00013036170990135185, |
| "loss": 0.5219, |
| "step": 958 |
| }, |
| { |
| "epoch": 0.34974471188913203, |
| "grad_norm": 0.18546460568904877, |
| "learning_rate": 0.00013028863719400803, |
| "loss": 0.7272, |
| "step": 959 |
| }, |
| { |
| "epoch": 0.350109409190372, |
| "grad_norm": 0.1910969465970993, |
| "learning_rate": 0.00013021556448666424, |
| "loss": 0.8131, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.350474106491612, |
| "grad_norm": 0.17536579072475433, |
| "learning_rate": 0.00013014249177932042, |
| "loss": 0.62, |
| "step": 961 |
| }, |
| { |
| "epoch": 0.3508388037928519, |
| "grad_norm": 0.18440979719161987, |
| "learning_rate": 0.00013006941907197663, |
| "loss": 0.7427, |
| "step": 962 |
| }, |
| { |
| "epoch": 0.35120350109409193, |
| "grad_norm": 0.1736113727092743, |
| "learning_rate": 0.00012999634636463284, |
| "loss": 0.6271, |
| "step": 963 |
| }, |
| { |
| "epoch": 0.3515681983953319, |
| "grad_norm": 0.1437050998210907, |
| "learning_rate": 0.000129923273657289, |
| "loss": 0.448, |
| "step": 964 |
| }, |
| { |
| "epoch": 0.3519328956965718, |
| "grad_norm": 0.17444917559623718, |
| "learning_rate": 0.0001298502009499452, |
| "loss": 0.5675, |
| "step": 965 |
| }, |
| { |
| "epoch": 0.3522975929978118, |
| "grad_norm": 0.2057693749666214, |
| "learning_rate": 0.0001297771282426014, |
| "loss": 0.9973, |
| "step": 966 |
| }, |
| { |
| "epoch": 0.3526622902990518, |
| "grad_norm": 0.1640542894601822, |
| "learning_rate": 0.0001297040555352576, |
| "loss": 0.5839, |
| "step": 967 |
| }, |
| { |
| "epoch": 0.3530269876002918, |
| "grad_norm": 0.1888854056596756, |
| "learning_rate": 0.00012963098282791377, |
| "loss": 0.7098, |
| "step": 968 |
| }, |
| { |
| "epoch": 0.3533916849015317, |
| "grad_norm": 0.1694556325674057, |
| "learning_rate": 0.00012955791012056998, |
| "loss": 0.5281, |
| "step": 969 |
| }, |
| { |
| "epoch": 0.3537563822027717, |
| "grad_norm": 0.17687252163887024, |
| "learning_rate": 0.00012948483741322616, |
| "loss": 0.6114, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.35412107950401167, |
| "grad_norm": 0.1951674520969391, |
| "learning_rate": 0.00012941176470588237, |
| "loss": 0.8128, |
| "step": 971 |
| }, |
| { |
| "epoch": 0.3544857768052516, |
| "grad_norm": 0.20023071765899658, |
| "learning_rate": 0.00012933869199853855, |
| "loss": 0.8015, |
| "step": 972 |
| }, |
| { |
| "epoch": 0.3548504741064916, |
| "grad_norm": 0.18741564452648163, |
| "learning_rate": 0.00012926561929119473, |
| "loss": 0.7011, |
| "step": 973 |
| }, |
| { |
| "epoch": 0.35521517140773157, |
| "grad_norm": 0.13944192230701447, |
| "learning_rate": 0.00012919254658385094, |
| "loss": 0.365, |
| "step": 974 |
| }, |
| { |
| "epoch": 0.35557986870897157, |
| "grad_norm": 0.20607557892799377, |
| "learning_rate": 0.00012911947387650715, |
| "loss": 0.89, |
| "step": 975 |
| }, |
| { |
| "epoch": 0.3559445660102115, |
| "grad_norm": 0.2182752937078476, |
| "learning_rate": 0.00012904640116916333, |
| "loss": 0.9494, |
| "step": 976 |
| }, |
| { |
| "epoch": 0.3563092633114515, |
| "grad_norm": 0.18262708187103271, |
| "learning_rate": 0.0001289733284618195, |
| "loss": 0.6899, |
| "step": 977 |
| }, |
| { |
| "epoch": 0.35667396061269147, |
| "grad_norm": 0.18693357706069946, |
| "learning_rate": 0.00012890025575447572, |
| "loss": 0.7838, |
| "step": 978 |
| }, |
| { |
| "epoch": 0.3570386579139314, |
| "grad_norm": 0.19558003544807434, |
| "learning_rate": 0.0001288271830471319, |
| "loss": 0.6937, |
| "step": 979 |
| }, |
| { |
| "epoch": 0.3574033552151714, |
| "grad_norm": 0.1773812621831894, |
| "learning_rate": 0.0001287541103397881, |
| "loss": 0.6103, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.35776805251641136, |
| "grad_norm": 0.19865770637989044, |
| "learning_rate": 0.0001286810376324443, |
| "loss": 0.884, |
| "step": 981 |
| }, |
| { |
| "epoch": 0.35813274981765136, |
| "grad_norm": 0.17361897230148315, |
| "learning_rate": 0.00012860796492510047, |
| "loss": 0.6312, |
| "step": 982 |
| }, |
| { |
| "epoch": 0.3584974471188913, |
| "grad_norm": 0.17946158349514008, |
| "learning_rate": 0.00012853489221775668, |
| "loss": 0.7898, |
| "step": 983 |
| }, |
| { |
| "epoch": 0.3588621444201313, |
| "grad_norm": 0.1990206092596054, |
| "learning_rate": 0.00012846181951041286, |
| "loss": 0.821, |
| "step": 984 |
| }, |
| { |
| "epoch": 0.35922684172137126, |
| "grad_norm": 0.16487091779708862, |
| "learning_rate": 0.00012838874680306904, |
| "loss": 0.6642, |
| "step": 985 |
| }, |
| { |
| "epoch": 0.3595915390226112, |
| "grad_norm": 0.17961853742599487, |
| "learning_rate": 0.00012831567409572525, |
| "loss": 0.8118, |
| "step": 986 |
| }, |
| { |
| "epoch": 0.3599562363238512, |
| "grad_norm": 0.16912826895713806, |
| "learning_rate": 0.00012824260138838143, |
| "loss": 0.6378, |
| "step": 987 |
| }, |
| { |
| "epoch": 0.36032093362509116, |
| "grad_norm": 0.19518031179904938, |
| "learning_rate": 0.00012816952868103764, |
| "loss": 0.9833, |
| "step": 988 |
| }, |
| { |
| "epoch": 0.36068563092633116, |
| "grad_norm": 0.18844519555568695, |
| "learning_rate": 0.00012809645597369385, |
| "loss": 0.7191, |
| "step": 989 |
| }, |
| { |
| "epoch": 0.3610503282275711, |
| "grad_norm": 0.18332988023757935, |
| "learning_rate": 0.00012802338326635, |
| "loss": 0.7232, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.3614150255288111, |
| "grad_norm": 0.18674488365650177, |
| "learning_rate": 0.0001279503105590062, |
| "loss": 0.7617, |
| "step": 991 |
| }, |
| { |
| "epoch": 0.36177972283005105, |
| "grad_norm": 0.15862171351909637, |
| "learning_rate": 0.00012787723785166242, |
| "loss": 0.5411, |
| "step": 992 |
| }, |
| { |
| "epoch": 0.36214442013129106, |
| "grad_norm": 0.15820543467998505, |
| "learning_rate": 0.0001278041651443186, |
| "loss": 0.5198, |
| "step": 993 |
| }, |
| { |
| "epoch": 0.362509117432531, |
| "grad_norm": 0.19230028986930847, |
| "learning_rate": 0.00012773109243697478, |
| "loss": 0.7863, |
| "step": 994 |
| }, |
| { |
| "epoch": 0.36287381473377095, |
| "grad_norm": 0.18754363059997559, |
| "learning_rate": 0.000127658019729631, |
| "loss": 0.7216, |
| "step": 995 |
| }, |
| { |
| "epoch": 0.36323851203501095, |
| "grad_norm": 0.19463679194450378, |
| "learning_rate": 0.00012758494702228717, |
| "loss": 0.8963, |
| "step": 996 |
| }, |
| { |
| "epoch": 0.3636032093362509, |
| "grad_norm": 0.17682290077209473, |
| "learning_rate": 0.00012751187431494338, |
| "loss": 0.7009, |
| "step": 997 |
| }, |
| { |
| "epoch": 0.3639679066374909, |
| "grad_norm": 0.20536212623119354, |
| "learning_rate": 0.0001274388016075996, |
| "loss": 0.9973, |
| "step": 998 |
| }, |
| { |
| "epoch": 0.36433260393873085, |
| "grad_norm": 0.2060231864452362, |
| "learning_rate": 0.00012736572890025574, |
| "loss": 0.8184, |
| "step": 999 |
| }, |
| { |
| "epoch": 0.36469730123997085, |
| "grad_norm": 0.19924308359622955, |
| "learning_rate": 0.00012729265619291195, |
| "loss": 0.7663, |
| "step": 1000 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 2742, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 6.490016210832998e+16, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|