{ "best_metric": null, "best_model_checkpoint": null, "epoch": 150.0, "eval_steps": 500, "global_step": 304500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.25, "grad_norm": 0.13165710866451263, "learning_rate": 3.119868637110016e-05, "loss": 0.8403, "step": 500 }, { "epoch": 0.49, "grad_norm": 0.676755428314209, "learning_rate": 3.114737274220033e-05, "loss": 0.8005, "step": 1000 }, { "epoch": 0.74, "grad_norm": 0.5474051237106323, "learning_rate": 3.109605911330049e-05, "loss": 0.7733, "step": 1500 }, { "epoch": 0.99, "grad_norm": 0.311629056930542, "learning_rate": 3.104474548440066e-05, "loss": 0.7443, "step": 2000 }, { "epoch": 1.23, "grad_norm": 0.3807041645050049, "learning_rate": 3.099343185550082e-05, "loss": 0.7288, "step": 2500 }, { "epoch": 1.48, "grad_norm": 0.38295918703079224, "learning_rate": 3.0942118226600984e-05, "loss": 0.7143, "step": 3000 }, { "epoch": 1.72, "grad_norm": 3.78216814994812, "learning_rate": 3.089080459770115e-05, "loss": 0.7101, "step": 3500 }, { "epoch": 1.97, "grad_norm": 0.34924811124801636, "learning_rate": 3.0839490968801314e-05, "loss": 0.6937, "step": 4000 }, { "epoch": 2.22, "grad_norm": 1.0352131128311157, "learning_rate": 3.078817733990148e-05, "loss": 0.6884, "step": 4500 }, { "epoch": 2.46, "grad_norm": 1.3734943866729736, "learning_rate": 3.0736863711001644e-05, "loss": 0.682, "step": 5000 }, { "epoch": 2.71, "grad_norm": 0.4328332841396332, "learning_rate": 3.0685550082101805e-05, "loss": 0.6766, "step": 5500 }, { "epoch": 2.96, "grad_norm": 0.6840975284576416, "learning_rate": 3.0634236453201974e-05, "loss": 0.6673, "step": 6000 }, { "epoch": 3.2, "grad_norm": 0.43340444564819336, "learning_rate": 3.0582922824302135e-05, "loss": 0.665, "step": 6500 }, { "epoch": 3.45, "grad_norm": 0.46208104491233826, "learning_rate": 3.05316091954023e-05, "loss": 0.657, "step": 7000 }, { "epoch": 3.69, "grad_norm": 0.4085042178630829, "learning_rate": 3.0480295566502465e-05, "loss": 0.6554, "step": 7500 }, { "epoch": 3.94, "grad_norm": 0.9567029476165771, "learning_rate": 3.042898193760263e-05, "loss": 0.6493, "step": 8000 }, { "epoch": 4.19, "grad_norm": 0.449690043926239, "learning_rate": 3.0377668308702792e-05, "loss": 0.6504, "step": 8500 }, { "epoch": 4.43, "grad_norm": 0.5416390299797058, "learning_rate": 3.0326354679802957e-05, "loss": 0.641, "step": 9000 }, { "epoch": 4.68, "grad_norm": 0.48183944821357727, "learning_rate": 3.0275041050903122e-05, "loss": 0.6322, "step": 9500 }, { "epoch": 4.93, "grad_norm": 0.4853132963180542, "learning_rate": 3.0223727422003287e-05, "loss": 0.6303, "step": 10000 }, { "epoch": 5.17, "grad_norm": 0.6569094657897949, "learning_rate": 3.017241379310345e-05, "loss": 0.6226, "step": 10500 }, { "epoch": 5.42, "grad_norm": 0.4652053415775299, "learning_rate": 3.0121100164203614e-05, "loss": 0.6211, "step": 11000 }, { "epoch": 5.67, "grad_norm": 0.49975645542144775, "learning_rate": 3.006978653530378e-05, "loss": 0.6187, "step": 11500 }, { "epoch": 5.91, "grad_norm": 0.39061611890792847, "learning_rate": 3.0018472906403944e-05, "loss": 0.6123, "step": 12000 }, { "epoch": 6.16, "grad_norm": 0.5346322655677795, "learning_rate": 2.996715927750411e-05, "loss": 0.6088, "step": 12500 }, { "epoch": 6.4, "grad_norm": 0.4497845768928528, "learning_rate": 2.991584564860427e-05, "loss": 0.6068, "step": 13000 }, { "epoch": 6.65, "grad_norm": 0.4072265923023224, "learning_rate": 2.9864532019704435e-05, "loss": 0.6011, "step": 13500 }, { "epoch": 6.9, "grad_norm": 0.47786733508110046, "learning_rate": 2.9813218390804597e-05, "loss": 0.6031, "step": 14000 }, { "epoch": 7.14, "grad_norm": 0.47765907645225525, "learning_rate": 2.9761904761904762e-05, "loss": 0.6015, "step": 14500 }, { "epoch": 7.39, "grad_norm": 0.45993849635124207, "learning_rate": 2.9710591133004923e-05, "loss": 0.5956, "step": 15000 }, { "epoch": 7.64, "grad_norm": 0.46012237668037415, "learning_rate": 2.965927750410509e-05, "loss": 0.5958, "step": 15500 }, { "epoch": 7.88, "grad_norm": 0.4168663024902344, "learning_rate": 2.9607963875205253e-05, "loss": 0.5918, "step": 16000 }, { "epoch": 8.13, "grad_norm": 0.452117383480072, "learning_rate": 2.955665024630542e-05, "loss": 0.5871, "step": 16500 }, { "epoch": 8.37, "grad_norm": 0.43112337589263916, "learning_rate": 2.9505336617405583e-05, "loss": 0.5868, "step": 17000 }, { "epoch": 8.62, "grad_norm": 0.44672533869743347, "learning_rate": 2.9454022988505745e-05, "loss": 0.5882, "step": 17500 }, { "epoch": 8.87, "grad_norm": 0.7204737067222595, "learning_rate": 2.940270935960591e-05, "loss": 0.5877, "step": 18000 }, { "epoch": 9.11, "grad_norm": 0.7274026274681091, "learning_rate": 2.9351395730706075e-05, "loss": 0.5867, "step": 18500 }, { "epoch": 9.36, "grad_norm": 0.398488849401474, "learning_rate": 2.930008210180624e-05, "loss": 0.5789, "step": 19000 }, { "epoch": 9.61, "grad_norm": 0.4212868809700012, "learning_rate": 2.9248768472906405e-05, "loss": 0.5795, "step": 19500 }, { "epoch": 9.85, "grad_norm": 0.42726728320121765, "learning_rate": 2.9197454844006567e-05, "loss": 0.5817, "step": 20000 }, { "epoch": 10.1, "grad_norm": 0.4954194724559784, "learning_rate": 2.914614121510673e-05, "loss": 0.5778, "step": 20500 }, { "epoch": 10.34, "grad_norm": 0.46405500173568726, "learning_rate": 2.9094827586206897e-05, "loss": 0.5772, "step": 21000 }, { "epoch": 10.59, "grad_norm": 3.064875602722168, "learning_rate": 2.904351395730706e-05, "loss": 0.5739, "step": 21500 }, { "epoch": 10.84, "grad_norm": 0.41212061047554016, "learning_rate": 2.8992200328407223e-05, "loss": 0.5755, "step": 22000 }, { "epoch": 11.08, "grad_norm": 0.5741709470748901, "learning_rate": 2.8940886699507388e-05, "loss": 0.5726, "step": 22500 }, { "epoch": 11.33, "grad_norm": 0.4510386288166046, "learning_rate": 2.8889573070607553e-05, "loss": 0.5737, "step": 23000 }, { "epoch": 11.58, "grad_norm": 0.44956105947494507, "learning_rate": 2.8838259441707718e-05, "loss": 0.5733, "step": 23500 }, { "epoch": 11.82, "grad_norm": 0.4094774127006531, "learning_rate": 2.8786945812807883e-05, "loss": 0.566, "step": 24000 }, { "epoch": 12.07, "grad_norm": 0.3915603756904602, "learning_rate": 2.8735632183908045e-05, "loss": 0.5621, "step": 24500 }, { "epoch": 12.32, "grad_norm": 0.42128342390060425, "learning_rate": 2.868431855500821e-05, "loss": 0.5647, "step": 25000 }, { "epoch": 12.56, "grad_norm": 0.4155956506729126, "learning_rate": 2.8633004926108375e-05, "loss": 0.5654, "step": 25500 }, { "epoch": 12.81, "grad_norm": 0.43520358204841614, "learning_rate": 2.858169129720854e-05, "loss": 0.5672, "step": 26000 }, { "epoch": 13.05, "grad_norm": 0.42418190836906433, "learning_rate": 2.8530377668308705e-05, "loss": 0.5639, "step": 26500 }, { "epoch": 13.3, "grad_norm": 0.4998001456260681, "learning_rate": 2.8479064039408867e-05, "loss": 0.5619, "step": 27000 }, { "epoch": 13.55, "grad_norm": 0.3863075375556946, "learning_rate": 2.842775041050903e-05, "loss": 0.5588, "step": 27500 }, { "epoch": 13.79, "grad_norm": 0.42726796865463257, "learning_rate": 2.8376436781609196e-05, "loss": 0.5619, "step": 28000 }, { "epoch": 14.04, "grad_norm": 0.5370152592658997, "learning_rate": 2.832512315270936e-05, "loss": 0.5599, "step": 28500 }, { "epoch": 14.29, "grad_norm": 0.5220550894737244, "learning_rate": 2.8273809523809523e-05, "loss": 0.5614, "step": 29000 }, { "epoch": 14.53, "grad_norm": 0.4528503119945526, "learning_rate": 2.8222495894909688e-05, "loss": 0.5544, "step": 29500 }, { "epoch": 14.78, "grad_norm": 0.34627479314804077, "learning_rate": 2.8171182266009853e-05, "loss": 0.5557, "step": 30000 }, { "epoch": 15.02, "grad_norm": 0.4139779210090637, "learning_rate": 2.8119868637110018e-05, "loss": 0.5583, "step": 30500 }, { "epoch": 15.27, "grad_norm": 0.529438316822052, "learning_rate": 2.8068555008210183e-05, "loss": 0.5511, "step": 31000 }, { "epoch": 15.52, "grad_norm": 0.40643787384033203, "learning_rate": 2.8017241379310345e-05, "loss": 0.5533, "step": 31500 }, { "epoch": 15.76, "grad_norm": 0.4177665114402771, "learning_rate": 2.796592775041051e-05, "loss": 0.5554, "step": 32000 }, { "epoch": 16.01, "grad_norm": 0.3691587448120117, "learning_rate": 2.7914614121510675e-05, "loss": 0.551, "step": 32500 }, { "epoch": 16.26, "grad_norm": 0.4278986155986786, "learning_rate": 2.786330049261084e-05, "loss": 0.5495, "step": 33000 }, { "epoch": 16.5, "grad_norm": 0.36237508058547974, "learning_rate": 2.7811986863711e-05, "loss": 0.5516, "step": 33500 }, { "epoch": 16.75, "grad_norm": 0.4242589771747589, "learning_rate": 2.7760673234811166e-05, "loss": 0.5499, "step": 34000 }, { "epoch": 17.0, "grad_norm": 0.35249194502830505, "learning_rate": 2.770935960591133e-05, "loss": 0.5511, "step": 34500 }, { "epoch": 17.24, "grad_norm": 0.42292988300323486, "learning_rate": 2.7658045977011496e-05, "loss": 0.5465, "step": 35000 }, { "epoch": 17.49, "grad_norm": 0.49178698658943176, "learning_rate": 2.760673234811166e-05, "loss": 0.5463, "step": 35500 }, { "epoch": 17.73, "grad_norm": 0.6154900789260864, "learning_rate": 2.7555418719211823e-05, "loss": 0.5459, "step": 36000 }, { "epoch": 17.98, "grad_norm": 0.43449530005455017, "learning_rate": 2.7504105090311988e-05, "loss": 0.5478, "step": 36500 }, { "epoch": 18.23, "grad_norm": 0.3909509479999542, "learning_rate": 2.7452791461412153e-05, "loss": 0.545, "step": 37000 }, { "epoch": 18.47, "grad_norm": 0.3956120014190674, "learning_rate": 2.7401477832512318e-05, "loss": 0.5442, "step": 37500 }, { "epoch": 18.72, "grad_norm": 0.369386225938797, "learning_rate": 2.7350164203612483e-05, "loss": 0.5496, "step": 38000 }, { "epoch": 18.97, "grad_norm": 0.40281441807746887, "learning_rate": 2.7298850574712645e-05, "loss": 0.5415, "step": 38500 }, { "epoch": 19.21, "grad_norm": 0.4241119623184204, "learning_rate": 2.724753694581281e-05, "loss": 0.5453, "step": 39000 }, { "epoch": 19.46, "grad_norm": 0.35802990198135376, "learning_rate": 2.7196223316912975e-05, "loss": 0.5465, "step": 39500 }, { "epoch": 19.7, "grad_norm": 0.3796702027320862, "learning_rate": 2.714490968801314e-05, "loss": 0.5407, "step": 40000 }, { "epoch": 19.95, "grad_norm": 5.289798736572266, "learning_rate": 2.70935960591133e-05, "loss": 0.54, "step": 40500 }, { "epoch": 20.2, "grad_norm": 0.43847599625587463, "learning_rate": 2.7042282430213466e-05, "loss": 0.5437, "step": 41000 }, { "epoch": 20.44, "grad_norm": 0.3674856126308441, "learning_rate": 2.699096880131363e-05, "loss": 0.5394, "step": 41500 }, { "epoch": 20.69, "grad_norm": 0.47424060106277466, "learning_rate": 2.6939655172413793e-05, "loss": 0.5412, "step": 42000 }, { "epoch": 20.94, "grad_norm": 0.40056848526000977, "learning_rate": 2.6888341543513958e-05, "loss": 0.5404, "step": 42500 }, { "epoch": 21.18, "grad_norm": 0.4091791808605194, "learning_rate": 2.683702791461412e-05, "loss": 0.5402, "step": 43000 }, { "epoch": 21.43, "grad_norm": 0.4481904208660126, "learning_rate": 2.6785714285714284e-05, "loss": 0.5358, "step": 43500 }, { "epoch": 21.67, "grad_norm": 0.44363269209861755, "learning_rate": 2.673440065681445e-05, "loss": 0.5378, "step": 44000 }, { "epoch": 21.92, "grad_norm": 0.42436665296554565, "learning_rate": 2.6683087027914614e-05, "loss": 0.541, "step": 44500 }, { "epoch": 22.17, "grad_norm": 0.4019184708595276, "learning_rate": 2.6631773399014776e-05, "loss": 0.5396, "step": 45000 }, { "epoch": 22.41, "grad_norm": 0.4707052409648895, "learning_rate": 2.658045977011494e-05, "loss": 0.5369, "step": 45500 }, { "epoch": 22.66, "grad_norm": 0.4337681233882904, "learning_rate": 2.6529146141215106e-05, "loss": 0.5383, "step": 46000 }, { "epoch": 22.91, "grad_norm": 0.5081506967544556, "learning_rate": 2.647783251231527e-05, "loss": 0.5356, "step": 46500 }, { "epoch": 23.15, "grad_norm": 0.4122790992259979, "learning_rate": 2.6426518883415436e-05, "loss": 0.5386, "step": 47000 }, { "epoch": 23.4, "grad_norm": 0.3679068386554718, "learning_rate": 2.6375205254515598e-05, "loss": 0.5337, "step": 47500 }, { "epoch": 23.65, "grad_norm": 0.4097256660461426, "learning_rate": 2.6323891625615763e-05, "loss": 0.5324, "step": 48000 }, { "epoch": 23.89, "grad_norm": 0.47837790846824646, "learning_rate": 2.6272577996715928e-05, "loss": 0.5374, "step": 48500 }, { "epoch": 24.14, "grad_norm": 0.33660951256752014, "learning_rate": 2.6221264367816093e-05, "loss": 0.5367, "step": 49000 }, { "epoch": 24.38, "grad_norm": 0.40010085701942444, "learning_rate": 2.6169950738916258e-05, "loss": 0.5308, "step": 49500 }, { "epoch": 24.63, "grad_norm": 0.4388476014137268, "learning_rate": 2.611863711001642e-05, "loss": 0.5323, "step": 50000 }, { "epoch": 24.88, "grad_norm": 0.6001474857330322, "learning_rate": 2.6067323481116584e-05, "loss": 0.5373, "step": 50500 }, { "epoch": 25.12, "grad_norm": 0.41148585081100464, "learning_rate": 2.601600985221675e-05, "loss": 0.5296, "step": 51000 }, { "epoch": 25.37, "grad_norm": 0.36909425258636475, "learning_rate": 2.5964696223316914e-05, "loss": 0.5353, "step": 51500 }, { "epoch": 25.62, "grad_norm": 0.4281887710094452, "learning_rate": 2.5913382594417076e-05, "loss": 0.528, "step": 52000 }, { "epoch": 25.86, "grad_norm": 0.37007570266723633, "learning_rate": 2.586206896551724e-05, "loss": 0.5332, "step": 52500 }, { "epoch": 26.11, "grad_norm": 0.44517025351524353, "learning_rate": 2.5810755336617406e-05, "loss": 0.5291, "step": 53000 }, { "epoch": 26.35, "grad_norm": 0.3227912187576294, "learning_rate": 2.575944170771757e-05, "loss": 0.5298, "step": 53500 }, { "epoch": 26.6, "grad_norm": 0.5226913690567017, "learning_rate": 2.5708128078817736e-05, "loss": 0.5269, "step": 54000 }, { "epoch": 26.85, "grad_norm": 0.4505539536476135, "learning_rate": 2.5656814449917897e-05, "loss": 0.5321, "step": 54500 }, { "epoch": 27.09, "grad_norm": 0.33736366033554077, "learning_rate": 2.5605500821018062e-05, "loss": 0.5299, "step": 55000 }, { "epoch": 27.34, "grad_norm": 0.5179808735847473, "learning_rate": 2.5554187192118227e-05, "loss": 0.5301, "step": 55500 }, { "epoch": 27.59, "grad_norm": 0.46012309193611145, "learning_rate": 2.5502873563218392e-05, "loss": 0.5259, "step": 56000 }, { "epoch": 27.83, "grad_norm": 0.3472147583961487, "learning_rate": 2.5451559934318557e-05, "loss": 0.5279, "step": 56500 }, { "epoch": 28.08, "grad_norm": 0.4225232005119324, "learning_rate": 2.540024630541872e-05, "loss": 0.5268, "step": 57000 }, { "epoch": 28.33, "grad_norm": 0.3676544725894928, "learning_rate": 2.5348932676518884e-05, "loss": 0.5285, "step": 57500 }, { "epoch": 28.57, "grad_norm": 0.4334140717983246, "learning_rate": 2.529761904761905e-05, "loss": 0.5306, "step": 58000 }, { "epoch": 28.82, "grad_norm": 0.3348256051540375, "learning_rate": 2.5246305418719214e-05, "loss": 0.5256, "step": 58500 }, { "epoch": 29.06, "grad_norm": 0.3783932030200958, "learning_rate": 2.5194991789819376e-05, "loss": 0.528, "step": 59000 }, { "epoch": 29.31, "grad_norm": 0.3242786228656769, "learning_rate": 2.514367816091954e-05, "loss": 0.5261, "step": 59500 }, { "epoch": 29.56, "grad_norm": 0.36754658818244934, "learning_rate": 2.5092364532019706e-05, "loss": 0.5242, "step": 60000 }, { "epoch": 29.8, "grad_norm": 0.3502654433250427, "learning_rate": 2.504105090311987e-05, "loss": 0.5284, "step": 60500 }, { "epoch": 30.05, "grad_norm": 0.3878660798072815, "learning_rate": 2.4989737274220036e-05, "loss": 0.5235, "step": 61000 }, { "epoch": 30.3, "grad_norm": 0.37782377004623413, "learning_rate": 2.4938423645320197e-05, "loss": 0.5232, "step": 61500 }, { "epoch": 30.54, "grad_norm": 0.49442175030708313, "learning_rate": 2.4887110016420362e-05, "loss": 0.5259, "step": 62000 }, { "epoch": 30.79, "grad_norm": 0.9456015229225159, "learning_rate": 2.4835796387520527e-05, "loss": 0.5261, "step": 62500 }, { "epoch": 31.03, "grad_norm": 0.35271939635276794, "learning_rate": 2.4784482758620692e-05, "loss": 0.5227, "step": 63000 }, { "epoch": 31.28, "grad_norm": 0.3738420903682709, "learning_rate": 2.4733169129720854e-05, "loss": 0.5234, "step": 63500 }, { "epoch": 31.53, "grad_norm": 0.4006379544734955, "learning_rate": 2.468185550082102e-05, "loss": 0.5226, "step": 64000 }, { "epoch": 31.77, "grad_norm": 0.3207830786705017, "learning_rate": 2.4630541871921184e-05, "loss": 0.5237, "step": 64500 }, { "epoch": 32.02, "grad_norm": 0.4237392246723175, "learning_rate": 2.457922824302135e-05, "loss": 0.5209, "step": 65000 }, { "epoch": 32.27, "grad_norm": 0.328134685754776, "learning_rate": 2.4527914614121514e-05, "loss": 0.5278, "step": 65500 }, { "epoch": 32.51, "grad_norm": 0.4005485773086548, "learning_rate": 2.4476600985221675e-05, "loss": 0.5221, "step": 66000 }, { "epoch": 32.76, "grad_norm": 0.4421192705631256, "learning_rate": 2.442528735632184e-05, "loss": 0.5199, "step": 66500 }, { "epoch": 33.0, "grad_norm": 0.38501447439193726, "learning_rate": 2.4373973727422005e-05, "loss": 0.5208, "step": 67000 }, { "epoch": 33.25, "grad_norm": 0.38499945402145386, "learning_rate": 2.432266009852217e-05, "loss": 0.521, "step": 67500 }, { "epoch": 33.5, "grad_norm": 0.31156325340270996, "learning_rate": 2.4271346469622335e-05, "loss": 0.5162, "step": 68000 }, { "epoch": 33.74, "grad_norm": 0.4041023552417755, "learning_rate": 2.4220032840722497e-05, "loss": 0.5199, "step": 68500 }, { "epoch": 33.99, "grad_norm": 0.4398662745952606, "learning_rate": 2.4168719211822662e-05, "loss": 0.5203, "step": 69000 }, { "epoch": 34.24, "grad_norm": 0.4758293330669403, "learning_rate": 2.4117405582922824e-05, "loss": 0.5236, "step": 69500 }, { "epoch": 34.48, "grad_norm": 0.4214634597301483, "learning_rate": 2.406609195402299e-05, "loss": 0.5187, "step": 70000 }, { "epoch": 34.73, "grad_norm": 0.41065362095832825, "learning_rate": 2.401477832512315e-05, "loss": 0.5171, "step": 70500 }, { "epoch": 34.98, "grad_norm": 0.3957192599773407, "learning_rate": 2.3963464696223315e-05, "loss": 0.5203, "step": 71000 }, { "epoch": 35.22, "grad_norm": 0.3957277536392212, "learning_rate": 2.391215106732348e-05, "loss": 0.5205, "step": 71500 }, { "epoch": 35.47, "grad_norm": 0.362846702337265, "learning_rate": 2.3860837438423645e-05, "loss": 0.5206, "step": 72000 }, { "epoch": 35.71, "grad_norm": 0.3435048758983612, "learning_rate": 2.380952380952381e-05, "loss": 0.5179, "step": 72500 }, { "epoch": 35.96, "grad_norm": 0.4027571976184845, "learning_rate": 2.3758210180623972e-05, "loss": 0.5179, "step": 73000 }, { "epoch": 36.21, "grad_norm": 0.40012747049331665, "learning_rate": 2.3706896551724137e-05, "loss": 0.5198, "step": 73500 }, { "epoch": 36.45, "grad_norm": 0.36184847354888916, "learning_rate": 2.3655582922824302e-05, "loss": 0.5171, "step": 74000 }, { "epoch": 36.7, "grad_norm": 0.4480341672897339, "learning_rate": 2.3604269293924467e-05, "loss": 0.5172, "step": 74500 }, { "epoch": 36.95, "grad_norm": 0.5025932788848877, "learning_rate": 2.355295566502463e-05, "loss": 0.5176, "step": 75000 }, { "epoch": 37.19, "grad_norm": 0.4158807694911957, "learning_rate": 2.3501642036124793e-05, "loss": 0.5144, "step": 75500 }, { "epoch": 37.44, "grad_norm": 0.3783656060695648, "learning_rate": 2.345032840722496e-05, "loss": 0.5146, "step": 76000 }, { "epoch": 37.68, "grad_norm": 0.37473171949386597, "learning_rate": 2.3399014778325123e-05, "loss": 0.5189, "step": 76500 }, { "epoch": 37.93, "grad_norm": 0.40515512228012085, "learning_rate": 2.334770114942529e-05, "loss": 0.5155, "step": 77000 }, { "epoch": 38.18, "grad_norm": 0.39813145995140076, "learning_rate": 2.329638752052545e-05, "loss": 0.5165, "step": 77500 }, { "epoch": 38.42, "grad_norm": 0.4116913676261902, "learning_rate": 2.3245073891625615e-05, "loss": 0.5144, "step": 78000 }, { "epoch": 38.67, "grad_norm": 0.37143009901046753, "learning_rate": 2.319376026272578e-05, "loss": 0.5167, "step": 78500 }, { "epoch": 38.92, "grad_norm": 0.38151097297668457, "learning_rate": 2.3142446633825945e-05, "loss": 0.5137, "step": 79000 }, { "epoch": 39.16, "grad_norm": 0.4646279215812683, "learning_rate": 2.309113300492611e-05, "loss": 0.5147, "step": 79500 }, { "epoch": 39.41, "grad_norm": 0.3832203447818756, "learning_rate": 2.303981937602627e-05, "loss": 0.5151, "step": 80000 }, { "epoch": 39.66, "grad_norm": 0.4361591637134552, "learning_rate": 2.2988505747126437e-05, "loss": 0.5141, "step": 80500 }, { "epoch": 39.9, "grad_norm": 0.3762683868408203, "learning_rate": 2.29371921182266e-05, "loss": 0.5142, "step": 81000 }, { "epoch": 40.15, "grad_norm": 0.3509220480918884, "learning_rate": 2.2885878489326767e-05, "loss": 0.5143, "step": 81500 }, { "epoch": 40.39, "grad_norm": 0.3234823942184448, "learning_rate": 2.2834564860426928e-05, "loss": 0.5111, "step": 82000 }, { "epoch": 40.64, "grad_norm": 0.3884125053882599, "learning_rate": 2.2783251231527093e-05, "loss": 0.5093, "step": 82500 }, { "epoch": 40.89, "grad_norm": 0.45307618379592896, "learning_rate": 2.2731937602627258e-05, "loss": 0.5157, "step": 83000 }, { "epoch": 41.13, "grad_norm": 0.4034999907016754, "learning_rate": 2.2680623973727423e-05, "loss": 0.5164, "step": 83500 }, { "epoch": 41.38, "grad_norm": 0.35115841031074524, "learning_rate": 2.2629310344827588e-05, "loss": 0.5151, "step": 84000 }, { "epoch": 41.63, "grad_norm": 0.3815375864505768, "learning_rate": 2.257799671592775e-05, "loss": 0.512, "step": 84500 }, { "epoch": 41.87, "grad_norm": 0.4422209560871124, "learning_rate": 2.2526683087027915e-05, "loss": 0.5088, "step": 85000 }, { "epoch": 42.12, "grad_norm": 0.35160163044929504, "learning_rate": 2.247536945812808e-05, "loss": 0.5123, "step": 85500 }, { "epoch": 42.36, "grad_norm": 0.4204586446285248, "learning_rate": 2.2424055829228245e-05, "loss": 0.5104, "step": 86000 }, { "epoch": 42.61, "grad_norm": 0.2954113185405731, "learning_rate": 2.237274220032841e-05, "loss": 0.5116, "step": 86500 }, { "epoch": 42.86, "grad_norm": 0.40945374965667725, "learning_rate": 2.232142857142857e-05, "loss": 0.5152, "step": 87000 }, { "epoch": 43.1, "grad_norm": 0.3709004819393158, "learning_rate": 2.2270114942528736e-05, "loss": 0.5157, "step": 87500 }, { "epoch": 43.35, "grad_norm": 0.4811795651912689, "learning_rate": 2.22188013136289e-05, "loss": 0.5118, "step": 88000 }, { "epoch": 43.6, "grad_norm": 0.36660343408584595, "learning_rate": 2.2167487684729066e-05, "loss": 0.5099, "step": 88500 }, { "epoch": 43.84, "grad_norm": 0.4243915379047394, "learning_rate": 2.2116174055829228e-05, "loss": 0.512, "step": 89000 }, { "epoch": 44.09, "grad_norm": 0.34263724088668823, "learning_rate": 2.2064860426929393e-05, "loss": 0.5116, "step": 89500 }, { "epoch": 44.33, "grad_norm": 0.374990314245224, "learning_rate": 2.2013546798029558e-05, "loss": 0.5092, "step": 90000 }, { "epoch": 44.58, "grad_norm": 0.33906516432762146, "learning_rate": 2.1962233169129723e-05, "loss": 0.5092, "step": 90500 }, { "epoch": 44.83, "grad_norm": 0.3953360915184021, "learning_rate": 2.1910919540229888e-05, "loss": 0.509, "step": 91000 }, { "epoch": 45.07, "grad_norm": 0.34916388988494873, "learning_rate": 2.185960591133005e-05, "loss": 0.5095, "step": 91500 }, { "epoch": 45.32, "grad_norm": 0.3636263310909271, "learning_rate": 2.1808292282430215e-05, "loss": 0.5091, "step": 92000 }, { "epoch": 45.57, "grad_norm": 0.38837483525276184, "learning_rate": 2.175697865353038e-05, "loss": 0.5077, "step": 92500 }, { "epoch": 45.81, "grad_norm": 0.3651277422904968, "learning_rate": 2.1705665024630545e-05, "loss": 0.5108, "step": 93000 }, { "epoch": 46.06, "grad_norm": 0.5110191702842712, "learning_rate": 2.1654351395730706e-05, "loss": 0.5093, "step": 93500 }, { "epoch": 46.31, "grad_norm": 0.36329400539398193, "learning_rate": 2.160303776683087e-05, "loss": 0.5093, "step": 94000 }, { "epoch": 46.55, "grad_norm": 0.42045408487319946, "learning_rate": 2.1551724137931036e-05, "loss": 0.5073, "step": 94500 }, { "epoch": 46.8, "grad_norm": 0.5139952301979065, "learning_rate": 2.15004105090312e-05, "loss": 0.5053, "step": 95000 }, { "epoch": 47.04, "grad_norm": 0.4146077632904053, "learning_rate": 2.1449096880131366e-05, "loss": 0.5073, "step": 95500 }, { "epoch": 47.29, "grad_norm": 0.5339871048927307, "learning_rate": 2.1397783251231528e-05, "loss": 0.5083, "step": 96000 }, { "epoch": 47.54, "grad_norm": 0.4209000766277313, "learning_rate": 2.1346469622331693e-05, "loss": 0.5098, "step": 96500 }, { "epoch": 47.78, "grad_norm": 0.3803653419017792, "learning_rate": 2.1295155993431854e-05, "loss": 0.5083, "step": 97000 }, { "epoch": 48.03, "grad_norm": 0.35340991616249084, "learning_rate": 2.124384236453202e-05, "loss": 0.5059, "step": 97500 }, { "epoch": 48.28, "grad_norm": 0.3598570227622986, "learning_rate": 2.1192528735632184e-05, "loss": 0.5026, "step": 98000 }, { "epoch": 48.52, "grad_norm": 0.3424580991268158, "learning_rate": 2.1141215106732346e-05, "loss": 0.5092, "step": 98500 }, { "epoch": 48.77, "grad_norm": 0.37450504302978516, "learning_rate": 2.108990147783251e-05, "loss": 0.5052, "step": 99000 }, { "epoch": 49.01, "grad_norm": 0.38776275515556335, "learning_rate": 2.1038587848932676e-05, "loss": 0.5059, "step": 99500 }, { "epoch": 49.26, "grad_norm": 0.4117530882358551, "learning_rate": 2.098727422003284e-05, "loss": 0.5059, "step": 100000 }, { "epoch": 49.51, "grad_norm": 0.43969395756721497, "learning_rate": 2.0935960591133003e-05, "loss": 0.5061, "step": 100500 }, { "epoch": 49.75, "grad_norm": 0.39242610335350037, "learning_rate": 2.0884646962233168e-05, "loss": 0.5056, "step": 101000 }, { "epoch": 50.0, "grad_norm": 0.6274676322937012, "learning_rate": 2.0833333333333333e-05, "loss": 0.5063, "step": 101500 }, { "epoch": 50.25, "grad_norm": 0.5223700404167175, "learning_rate": 2.0782019704433498e-05, "loss": 0.5029, "step": 102000 }, { "epoch": 50.49, "grad_norm": 0.36774346232414246, "learning_rate": 2.0730706075533663e-05, "loss": 0.5048, "step": 102500 }, { "epoch": 50.74, "grad_norm": 0.3601688742637634, "learning_rate": 2.0679392446633824e-05, "loss": 0.5054, "step": 103000 }, { "epoch": 50.99, "grad_norm": 0.35829395055770874, "learning_rate": 2.062807881773399e-05, "loss": 0.5077, "step": 103500 }, { "epoch": 51.23, "grad_norm": 0.394145667552948, "learning_rate": 2.0576765188834154e-05, "loss": 0.5071, "step": 104000 }, { "epoch": 51.48, "grad_norm": 0.39394277334213257, "learning_rate": 2.052545155993432e-05, "loss": 0.503, "step": 104500 }, { "epoch": 51.72, "grad_norm": 0.42796215415000916, "learning_rate": 2.047413793103448e-05, "loss": 0.5015, "step": 105000 }, { "epoch": 51.97, "grad_norm": 0.342929869890213, "learning_rate": 2.0422824302134646e-05, "loss": 0.5041, "step": 105500 }, { "epoch": 52.22, "grad_norm": 0.34425219893455505, "learning_rate": 2.037151067323481e-05, "loss": 0.5048, "step": 106000 }, { "epoch": 52.46, "grad_norm": 0.40054482221603394, "learning_rate": 2.0320197044334976e-05, "loss": 0.505, "step": 106500 }, { "epoch": 52.71, "grad_norm": 0.41854235529899597, "learning_rate": 2.026888341543514e-05, "loss": 0.5028, "step": 107000 }, { "epoch": 52.96, "grad_norm": 0.39581912755966187, "learning_rate": 2.0217569786535302e-05, "loss": 0.5019, "step": 107500 }, { "epoch": 53.2, "grad_norm": 0.3793180286884308, "learning_rate": 2.0166256157635467e-05, "loss": 0.5025, "step": 108000 }, { "epoch": 53.45, "grad_norm": 0.42270803451538086, "learning_rate": 2.0114942528735632e-05, "loss": 0.505, "step": 108500 }, { "epoch": 53.69, "grad_norm": 0.4206148684024811, "learning_rate": 2.0063628899835797e-05, "loss": 0.5032, "step": 109000 }, { "epoch": 53.94, "grad_norm": 0.535724401473999, "learning_rate": 2.0012315270935962e-05, "loss": 0.5041, "step": 109500 }, { "epoch": 54.19, "grad_norm": 0.4594982862472534, "learning_rate": 1.9961001642036124e-05, "loss": 0.5011, "step": 110000 }, { "epoch": 54.43, "grad_norm": 0.38085901737213135, "learning_rate": 1.990968801313629e-05, "loss": 0.5043, "step": 110500 }, { "epoch": 54.68, "grad_norm": 0.4274803102016449, "learning_rate": 1.9858374384236454e-05, "loss": 0.5065, "step": 111000 }, { "epoch": 54.93, "grad_norm": 0.4163910150527954, "learning_rate": 1.980706075533662e-05, "loss": 0.5, "step": 111500 }, { "epoch": 55.17, "grad_norm": 0.42667436599731445, "learning_rate": 1.975574712643678e-05, "loss": 0.5018, "step": 112000 }, { "epoch": 55.42, "grad_norm": 0.3834710717201233, "learning_rate": 1.9704433497536946e-05, "loss": 0.4994, "step": 112500 }, { "epoch": 55.67, "grad_norm": 0.39337608218193054, "learning_rate": 1.965311986863711e-05, "loss": 0.5032, "step": 113000 }, { "epoch": 55.91, "grad_norm": 0.48434558510780334, "learning_rate": 1.9601806239737276e-05, "loss": 0.5046, "step": 113500 }, { "epoch": 56.16, "grad_norm": 0.4098414480686188, "learning_rate": 1.955049261083744e-05, "loss": 0.5032, "step": 114000 }, { "epoch": 56.4, "grad_norm": 0.4482068419456482, "learning_rate": 1.9499178981937602e-05, "loss": 0.501, "step": 114500 }, { "epoch": 56.65, "grad_norm": 0.4341977834701538, "learning_rate": 1.9447865353037767e-05, "loss": 0.5021, "step": 115000 }, { "epoch": 56.9, "grad_norm": 0.40422672033309937, "learning_rate": 1.9396551724137932e-05, "loss": 0.5003, "step": 115500 }, { "epoch": 57.14, "grad_norm": 0.35205212235450745, "learning_rate": 1.9345238095238097e-05, "loss": 0.4969, "step": 116000 }, { "epoch": 57.39, "grad_norm": 0.4301392734050751, "learning_rate": 1.9293924466338262e-05, "loss": 0.4977, "step": 116500 }, { "epoch": 57.64, "grad_norm": 0.3865519165992737, "learning_rate": 1.9242610837438424e-05, "loss": 0.5031, "step": 117000 }, { "epoch": 57.88, "grad_norm": 0.4432295262813568, "learning_rate": 1.919129720853859e-05, "loss": 0.5036, "step": 117500 }, { "epoch": 58.13, "grad_norm": 0.44078487157821655, "learning_rate": 1.9139983579638754e-05, "loss": 0.4985, "step": 118000 }, { "epoch": 58.37, "grad_norm": 0.5300925374031067, "learning_rate": 1.908866995073892e-05, "loss": 0.5014, "step": 118500 }, { "epoch": 58.62, "grad_norm": 0.39328643679618835, "learning_rate": 1.903735632183908e-05, "loss": 0.4988, "step": 119000 }, { "epoch": 58.87, "grad_norm": 0.3447715938091278, "learning_rate": 1.8986042692939245e-05, "loss": 0.5013, "step": 119500 }, { "epoch": 59.11, "grad_norm": 0.4861631989479065, "learning_rate": 1.893472906403941e-05, "loss": 0.4991, "step": 120000 }, { "epoch": 59.36, "grad_norm": 0.3493654131889343, "learning_rate": 1.8883415435139575e-05, "loss": 0.499, "step": 120500 }, { "epoch": 59.61, "grad_norm": 0.4206608235836029, "learning_rate": 1.883210180623974e-05, "loss": 0.499, "step": 121000 }, { "epoch": 59.85, "grad_norm": 0.3103729486465454, "learning_rate": 1.8780788177339902e-05, "loss": 0.4998, "step": 121500 }, { "epoch": 60.1, "grad_norm": 0.3733006417751312, "learning_rate": 1.8729474548440067e-05, "loss": 0.4981, "step": 122000 }, { "epoch": 60.34, "grad_norm": 0.459492564201355, "learning_rate": 1.8678160919540232e-05, "loss": 0.499, "step": 122500 }, { "epoch": 60.59, "grad_norm": 0.35242483019828796, "learning_rate": 1.8626847290640397e-05, "loss": 0.4971, "step": 123000 }, { "epoch": 60.84, "grad_norm": 0.3865768611431122, "learning_rate": 1.857553366174056e-05, "loss": 0.4961, "step": 123500 }, { "epoch": 61.08, "grad_norm": 0.3595926761627197, "learning_rate": 1.8524220032840724e-05, "loss": 0.4987, "step": 124000 }, { "epoch": 61.33, "grad_norm": 0.36663565039634705, "learning_rate": 1.847290640394089e-05, "loss": 0.5008, "step": 124500 }, { "epoch": 61.58, "grad_norm": 0.33786624670028687, "learning_rate": 1.842159277504105e-05, "loss": 0.4975, "step": 125000 }, { "epoch": 61.82, "grad_norm": 0.38265419006347656, "learning_rate": 1.8370279146141215e-05, "loss": 0.4958, "step": 125500 }, { "epoch": 62.07, "grad_norm": 0.353635311126709, "learning_rate": 1.8318965517241377e-05, "loss": 0.4981, "step": 126000 }, { "epoch": 62.32, "grad_norm": 0.35221561789512634, "learning_rate": 1.8267651888341542e-05, "loss": 0.4964, "step": 126500 }, { "epoch": 62.56, "grad_norm": 0.47486743330955505, "learning_rate": 1.8216338259441707e-05, "loss": 0.5008, "step": 127000 }, { "epoch": 62.81, "grad_norm": 0.40547123551368713, "learning_rate": 1.8165024630541872e-05, "loss": 0.4943, "step": 127500 }, { "epoch": 63.05, "grad_norm": 0.3869677484035492, "learning_rate": 1.8113711001642037e-05, "loss": 0.4975, "step": 128000 }, { "epoch": 63.3, "grad_norm": 0.487684428691864, "learning_rate": 1.80623973727422e-05, "loss": 0.4964, "step": 128500 }, { "epoch": 63.55, "grad_norm": 0.4795319139957428, "learning_rate": 1.8011083743842363e-05, "loss": 0.4972, "step": 129000 }, { "epoch": 63.79, "grad_norm": 0.40214163064956665, "learning_rate": 1.795977011494253e-05, "loss": 0.4977, "step": 129500 }, { "epoch": 64.04, "grad_norm": 0.41338789463043213, "learning_rate": 1.7908456486042693e-05, "loss": 0.4969, "step": 130000 }, { "epoch": 64.29, "grad_norm": 0.46818971633911133, "learning_rate": 1.7857142857142855e-05, "loss": 0.4939, "step": 130500 }, { "epoch": 64.53, "grad_norm": 0.5682018399238586, "learning_rate": 1.780582922824302e-05, "loss": 0.4955, "step": 131000 }, { "epoch": 64.78, "grad_norm": 0.3598514497280121, "learning_rate": 1.7754515599343185e-05, "loss": 0.4978, "step": 131500 }, { "epoch": 65.02, "grad_norm": 0.45837077498435974, "learning_rate": 1.770320197044335e-05, "loss": 0.4962, "step": 132000 }, { "epoch": 65.27, "grad_norm": 0.40763700008392334, "learning_rate": 1.7651888341543515e-05, "loss": 0.4972, "step": 132500 }, { "epoch": 65.52, "grad_norm": 0.41413307189941406, "learning_rate": 1.7600574712643677e-05, "loss": 0.4948, "step": 133000 }, { "epoch": 65.76, "grad_norm": 1.8902608156204224, "learning_rate": 1.7549261083743842e-05, "loss": 0.4952, "step": 133500 }, { "epoch": 66.01, "grad_norm": 0.4681415855884552, "learning_rate": 1.7497947454844007e-05, "loss": 0.4949, "step": 134000 }, { "epoch": 66.26, "grad_norm": 0.41859620809555054, "learning_rate": 1.744663382594417e-05, "loss": 0.4927, "step": 134500 }, { "epoch": 66.5, "grad_norm": 0.380066454410553, "learning_rate": 1.7395320197044333e-05, "loss": 0.4984, "step": 135000 }, { "epoch": 66.75, "grad_norm": 0.4095420837402344, "learning_rate": 1.7344006568144498e-05, "loss": 0.4942, "step": 135500 }, { "epoch": 67.0, "grad_norm": 0.3269331753253937, "learning_rate": 1.7292692939244663e-05, "loss": 0.4964, "step": 136000 }, { "epoch": 67.24, "grad_norm": 0.4390511214733124, "learning_rate": 1.7241379310344828e-05, "loss": 0.4931, "step": 136500 }, { "epoch": 67.49, "grad_norm": 0.37324196100234985, "learning_rate": 1.7190065681444993e-05, "loss": 0.4943, "step": 137000 }, { "epoch": 67.73, "grad_norm": 0.37749549746513367, "learning_rate": 1.7138752052545155e-05, "loss": 0.4967, "step": 137500 }, { "epoch": 67.98, "grad_norm": 0.42331600189208984, "learning_rate": 1.708743842364532e-05, "loss": 0.4908, "step": 138000 }, { "epoch": 68.23, "grad_norm": 0.4857657551765442, "learning_rate": 1.7036124794745485e-05, "loss": 0.4933, "step": 138500 }, { "epoch": 68.47, "grad_norm": 0.40747708082199097, "learning_rate": 1.698481116584565e-05, "loss": 0.4921, "step": 139000 }, { "epoch": 68.72, "grad_norm": 0.5494533181190491, "learning_rate": 1.6933497536945815e-05, "loss": 0.4929, "step": 139500 }, { "epoch": 68.97, "grad_norm": 0.44840362668037415, "learning_rate": 1.6882183908045977e-05, "loss": 0.4963, "step": 140000 }, { "epoch": 69.21, "grad_norm": 0.42726320028305054, "learning_rate": 1.683087027914614e-05, "loss": 0.4945, "step": 140500 }, { "epoch": 69.46, "grad_norm": 0.3714928925037384, "learning_rate": 1.6779556650246307e-05, "loss": 0.4946, "step": 141000 }, { "epoch": 69.7, "grad_norm": 0.44776561856269836, "learning_rate": 1.672824302134647e-05, "loss": 0.4932, "step": 141500 }, { "epoch": 69.95, "grad_norm": 0.36242425441741943, "learning_rate": 1.6676929392446633e-05, "loss": 0.4927, "step": 142000 }, { "epoch": 70.2, "grad_norm": 0.40916842222213745, "learning_rate": 1.6625615763546798e-05, "loss": 0.4931, "step": 142500 }, { "epoch": 70.44, "grad_norm": 0.3582199513912201, "learning_rate": 1.6574302134646963e-05, "loss": 0.4909, "step": 143000 }, { "epoch": 70.69, "grad_norm": 0.39829275012016296, "learning_rate": 1.6522988505747128e-05, "loss": 0.49, "step": 143500 }, { "epoch": 70.94, "grad_norm": 0.362525075674057, "learning_rate": 1.6471674876847293e-05, "loss": 0.4933, "step": 144000 }, { "epoch": 71.18, "grad_norm": 0.425618976354599, "learning_rate": 1.6420361247947455e-05, "loss": 0.4968, "step": 144500 }, { "epoch": 71.43, "grad_norm": 0.4285431504249573, "learning_rate": 1.636904761904762e-05, "loss": 0.4922, "step": 145000 }, { "epoch": 71.67, "grad_norm": 0.36322301626205444, "learning_rate": 1.6317733990147785e-05, "loss": 0.491, "step": 145500 }, { "epoch": 71.92, "grad_norm": 0.41439762711524963, "learning_rate": 1.626642036124795e-05, "loss": 0.4906, "step": 146000 }, { "epoch": 72.17, "grad_norm": 0.3911706805229187, "learning_rate": 1.621510673234811e-05, "loss": 0.4888, "step": 146500 }, { "epoch": 72.41, "grad_norm": 0.42668530344963074, "learning_rate": 1.6163793103448276e-05, "loss": 0.4925, "step": 147000 }, { "epoch": 72.66, "grad_norm": 0.4418216943740845, "learning_rate": 1.611247947454844e-05, "loss": 0.4921, "step": 147500 }, { "epoch": 72.91, "grad_norm": 0.5424076318740845, "learning_rate": 1.6061165845648606e-05, "loss": 0.4888, "step": 148000 }, { "epoch": 73.15, "grad_norm": 0.38305142521858215, "learning_rate": 1.600985221674877e-05, "loss": 0.491, "step": 148500 }, { "epoch": 73.4, "grad_norm": 0.4137153923511505, "learning_rate": 1.5958538587848933e-05, "loss": 0.4882, "step": 149000 }, { "epoch": 73.65, "grad_norm": 0.46136030554771423, "learning_rate": 1.5907224958949098e-05, "loss": 0.4907, "step": 149500 }, { "epoch": 73.89, "grad_norm": 0.39286741614341736, "learning_rate": 1.5855911330049263e-05, "loss": 0.4884, "step": 150000 }, { "epoch": 74.14, "grad_norm": 0.4563937783241272, "learning_rate": 1.5804597701149428e-05, "loss": 0.494, "step": 150500 }, { "epoch": 74.38, "grad_norm": 0.4879554510116577, "learning_rate": 1.5753284072249593e-05, "loss": 0.4908, "step": 151000 }, { "epoch": 74.63, "grad_norm": 0.48795485496520996, "learning_rate": 1.5701970443349755e-05, "loss": 0.4949, "step": 151500 }, { "epoch": 74.88, "grad_norm": 0.3959016799926758, "learning_rate": 1.565065681444992e-05, "loss": 0.4868, "step": 152000 }, { "epoch": 75.12, "grad_norm": 0.46233025193214417, "learning_rate": 1.559934318555008e-05, "loss": 0.4851, "step": 152500 }, { "epoch": 75.37, "grad_norm": 0.33422186970710754, "learning_rate": 1.5548029556650246e-05, "loss": 0.488, "step": 153000 }, { "epoch": 75.62, "grad_norm": 0.4503116309642792, "learning_rate": 1.549671592775041e-05, "loss": 0.4885, "step": 153500 }, { "epoch": 75.86, "grad_norm": 0.4860435426235199, "learning_rate": 1.5445402298850576e-05, "loss": 0.4891, "step": 154000 }, { "epoch": 76.11, "grad_norm": 0.40661612153053284, "learning_rate": 1.539408866995074e-05, "loss": 0.4884, "step": 154500 }, { "epoch": 76.35, "grad_norm": 0.4261013865470886, "learning_rate": 1.5342775041050903e-05, "loss": 0.487, "step": 155000 }, { "epoch": 76.6, "grad_norm": 0.4988757073879242, "learning_rate": 1.5291461412151068e-05, "loss": 0.4909, "step": 155500 }, { "epoch": 76.85, "grad_norm": 0.4792279005050659, "learning_rate": 1.5240147783251233e-05, "loss": 0.491, "step": 156000 }, { "epoch": 77.09, "grad_norm": 0.43279606103897095, "learning_rate": 1.5188834154351396e-05, "loss": 0.492, "step": 156500 }, { "epoch": 77.34, "grad_norm": 0.44450482726097107, "learning_rate": 1.5137520525451561e-05, "loss": 0.4855, "step": 157000 }, { "epoch": 77.59, "grad_norm": 0.42101508378982544, "learning_rate": 1.5086206896551724e-05, "loss": 0.4883, "step": 157500 }, { "epoch": 77.83, "grad_norm": 0.48337703943252563, "learning_rate": 1.503489326765189e-05, "loss": 0.4869, "step": 158000 }, { "epoch": 78.08, "grad_norm": 0.4778783321380615, "learning_rate": 1.4983579638752054e-05, "loss": 0.4869, "step": 158500 }, { "epoch": 78.33, "grad_norm": 0.35785460472106934, "learning_rate": 1.4932266009852218e-05, "loss": 0.4926, "step": 159000 }, { "epoch": 78.57, "grad_norm": 0.39978745579719543, "learning_rate": 1.4880952380952381e-05, "loss": 0.4849, "step": 159500 }, { "epoch": 78.82, "grad_norm": 0.36985936760902405, "learning_rate": 1.4829638752052544e-05, "loss": 0.4849, "step": 160000 }, { "epoch": 79.06, "grad_norm": 0.5067600607872009, "learning_rate": 1.477832512315271e-05, "loss": 0.4875, "step": 160500 }, { "epoch": 79.31, "grad_norm": 0.5925462245941162, "learning_rate": 1.4727011494252873e-05, "loss": 0.4895, "step": 161000 }, { "epoch": 79.56, "grad_norm": 0.43907538056373596, "learning_rate": 1.4675697865353038e-05, "loss": 0.4851, "step": 161500 }, { "epoch": 79.8, "grad_norm": 0.39582559466362, "learning_rate": 1.4624384236453203e-05, "loss": 0.485, "step": 162000 }, { "epoch": 80.05, "grad_norm": 0.3945811688899994, "learning_rate": 1.4573070607553366e-05, "loss": 0.49, "step": 162500 }, { "epoch": 80.3, "grad_norm": 0.5007498264312744, "learning_rate": 1.452175697865353e-05, "loss": 0.4871, "step": 163000 }, { "epoch": 80.54, "grad_norm": 0.3767457902431488, "learning_rate": 1.4470443349753694e-05, "loss": 0.4873, "step": 163500 }, { "epoch": 80.79, "grad_norm": 0.584635317325592, "learning_rate": 1.4419129720853859e-05, "loss": 0.4829, "step": 164000 }, { "epoch": 81.03, "grad_norm": 0.41645538806915283, "learning_rate": 1.4367816091954022e-05, "loss": 0.4893, "step": 164500 }, { "epoch": 81.28, "grad_norm": 0.4460589289665222, "learning_rate": 1.4316502463054187e-05, "loss": 0.4867, "step": 165000 }, { "epoch": 81.53, "grad_norm": 0.4328470528125763, "learning_rate": 1.4265188834154352e-05, "loss": 0.4844, "step": 165500 }, { "epoch": 81.77, "grad_norm": 0.41708043217658997, "learning_rate": 1.4213875205254516e-05, "loss": 0.4869, "step": 166000 }, { "epoch": 82.02, "grad_norm": 0.4869779348373413, "learning_rate": 1.416256157635468e-05, "loss": 0.4847, "step": 166500 }, { "epoch": 82.27, "grad_norm": 0.3944786489009857, "learning_rate": 1.4111247947454844e-05, "loss": 0.4856, "step": 167000 }, { "epoch": 82.51, "grad_norm": 0.36754748225212097, "learning_rate": 1.4059934318555009e-05, "loss": 0.4811, "step": 167500 }, { "epoch": 82.76, "grad_norm": 0.4370395839214325, "learning_rate": 1.4008620689655172e-05, "loss": 0.4841, "step": 168000 }, { "epoch": 83.0, "grad_norm": 0.48854583501815796, "learning_rate": 1.3957307060755337e-05, "loss": 0.4866, "step": 168500 }, { "epoch": 83.25, "grad_norm": 0.4292881488800049, "learning_rate": 1.39059934318555e-05, "loss": 0.4831, "step": 169000 }, { "epoch": 83.5, "grad_norm": 0.5421557426452637, "learning_rate": 1.3854679802955666e-05, "loss": 0.4876, "step": 169500 }, { "epoch": 83.74, "grad_norm": 0.4469299614429474, "learning_rate": 1.380336617405583e-05, "loss": 0.4857, "step": 170000 }, { "epoch": 83.99, "grad_norm": 0.4676869213581085, "learning_rate": 1.3752052545155994e-05, "loss": 0.4849, "step": 170500 }, { "epoch": 84.24, "grad_norm": 0.4767548739910126, "learning_rate": 1.3700738916256159e-05, "loss": 0.4858, "step": 171000 }, { "epoch": 84.48, "grad_norm": 0.36193403601646423, "learning_rate": 1.3649425287356322e-05, "loss": 0.4849, "step": 171500 }, { "epoch": 84.73, "grad_norm": 0.4789485037326813, "learning_rate": 1.3598111658456487e-05, "loss": 0.4835, "step": 172000 }, { "epoch": 84.98, "grad_norm": 0.4937196373939514, "learning_rate": 1.354679802955665e-05, "loss": 0.4851, "step": 172500 }, { "epoch": 85.22, "grad_norm": 0.5894652009010315, "learning_rate": 1.3495484400656816e-05, "loss": 0.4827, "step": 173000 }, { "epoch": 85.47, "grad_norm": 0.37527599930763245, "learning_rate": 1.3444170771756979e-05, "loss": 0.4797, "step": 173500 }, { "epoch": 85.71, "grad_norm": 0.4133756160736084, "learning_rate": 1.3392857142857142e-05, "loss": 0.4845, "step": 174000 }, { "epoch": 85.96, "grad_norm": 0.4965701103210449, "learning_rate": 1.3341543513957307e-05, "loss": 0.4849, "step": 174500 }, { "epoch": 86.21, "grad_norm": 0.37571871280670166, "learning_rate": 1.329022988505747e-05, "loss": 0.4846, "step": 175000 }, { "epoch": 86.45, "grad_norm": 0.4651111364364624, "learning_rate": 1.3238916256157635e-05, "loss": 0.4837, "step": 175500 }, { "epoch": 86.7, "grad_norm": 0.43609529733657837, "learning_rate": 1.3187602627257799e-05, "loss": 0.4829, "step": 176000 }, { "epoch": 86.95, "grad_norm": 0.4729571044445038, "learning_rate": 1.3136288998357964e-05, "loss": 0.4846, "step": 176500 }, { "epoch": 87.19, "grad_norm": 0.3406832814216614, "learning_rate": 1.3084975369458129e-05, "loss": 0.4798, "step": 177000 }, { "epoch": 87.44, "grad_norm": 0.40902179479599, "learning_rate": 1.3033661740558292e-05, "loss": 0.4846, "step": 177500 }, { "epoch": 87.68, "grad_norm": 0.44439247250556946, "learning_rate": 1.2982348111658457e-05, "loss": 0.4858, "step": 178000 }, { "epoch": 87.93, "grad_norm": 0.36519864201545715, "learning_rate": 1.293103448275862e-05, "loss": 0.4826, "step": 178500 }, { "epoch": 88.18, "grad_norm": 0.49141621589660645, "learning_rate": 1.2879720853858785e-05, "loss": 0.4843, "step": 179000 }, { "epoch": 88.42, "grad_norm": 0.45474326610565186, "learning_rate": 1.2828407224958949e-05, "loss": 0.4837, "step": 179500 }, { "epoch": 88.67, "grad_norm": 0.522074282169342, "learning_rate": 1.2777093596059114e-05, "loss": 0.4814, "step": 180000 }, { "epoch": 88.92, "grad_norm": 0.48209860920906067, "learning_rate": 1.2725779967159279e-05, "loss": 0.4837, "step": 180500 }, { "epoch": 89.16, "grad_norm": 0.5469162464141846, "learning_rate": 1.2674466338259442e-05, "loss": 0.4809, "step": 181000 }, { "epoch": 89.41, "grad_norm": 0.5114103555679321, "learning_rate": 1.2623152709359607e-05, "loss": 0.4848, "step": 181500 }, { "epoch": 89.66, "grad_norm": 0.3811694085597992, "learning_rate": 1.257183908045977e-05, "loss": 0.4774, "step": 182000 }, { "epoch": 89.9, "grad_norm": 0.4292524456977844, "learning_rate": 1.2520525451559935e-05, "loss": 0.4812, "step": 182500 }, { "epoch": 90.15, "grad_norm": 1.8638391494750977, "learning_rate": 1.2469211822660099e-05, "loss": 0.4837, "step": 183000 }, { "epoch": 90.39, "grad_norm": 0.5892685055732727, "learning_rate": 1.2417898193760264e-05, "loss": 0.4792, "step": 183500 }, { "epoch": 90.64, "grad_norm": 0.41100233793258667, "learning_rate": 1.2366584564860427e-05, "loss": 0.4832, "step": 184000 }, { "epoch": 90.89, "grad_norm": 0.44678598642349243, "learning_rate": 1.2315270935960592e-05, "loss": 0.484, "step": 184500 }, { "epoch": 91.13, "grad_norm": 0.46170106530189514, "learning_rate": 1.2263957307060757e-05, "loss": 0.4826, "step": 185000 }, { "epoch": 91.38, "grad_norm": 0.46171411871910095, "learning_rate": 1.221264367816092e-05, "loss": 0.4819, "step": 185500 }, { "epoch": 91.63, "grad_norm": 0.45027804374694824, "learning_rate": 1.2161330049261085e-05, "loss": 0.4841, "step": 186000 }, { "epoch": 91.87, "grad_norm": 0.407806396484375, "learning_rate": 1.2110016420361248e-05, "loss": 0.4772, "step": 186500 }, { "epoch": 92.12, "grad_norm": 0.4288312792778015, "learning_rate": 1.2058702791461412e-05, "loss": 0.4809, "step": 187000 }, { "epoch": 92.36, "grad_norm": 0.41084253787994385, "learning_rate": 1.2007389162561575e-05, "loss": 0.4806, "step": 187500 }, { "epoch": 92.61, "grad_norm": 0.4116561710834503, "learning_rate": 1.195607553366174e-05, "loss": 0.4829, "step": 188000 }, { "epoch": 92.86, "grad_norm": 0.4914272725582123, "learning_rate": 1.1904761904761905e-05, "loss": 0.4812, "step": 188500 }, { "epoch": 93.1, "grad_norm": 0.4653904139995575, "learning_rate": 1.1853448275862068e-05, "loss": 0.4787, "step": 189000 }, { "epoch": 93.35, "grad_norm": 0.4210495948791504, "learning_rate": 1.1802134646962233e-05, "loss": 0.4773, "step": 189500 }, { "epoch": 93.6, "grad_norm": 0.40525516867637634, "learning_rate": 1.1750821018062397e-05, "loss": 0.4807, "step": 190000 }, { "epoch": 93.84, "grad_norm": 0.3981895446777344, "learning_rate": 1.1699507389162562e-05, "loss": 0.481, "step": 190500 }, { "epoch": 94.09, "grad_norm": 0.846139669418335, "learning_rate": 1.1648193760262725e-05, "loss": 0.4792, "step": 191000 }, { "epoch": 94.33, "grad_norm": 0.42289435863494873, "learning_rate": 1.159688013136289e-05, "loss": 0.4788, "step": 191500 }, { "epoch": 94.58, "grad_norm": 0.3914598226547241, "learning_rate": 1.1545566502463055e-05, "loss": 0.4837, "step": 192000 }, { "epoch": 94.83, "grad_norm": 0.5285059809684753, "learning_rate": 1.1494252873563218e-05, "loss": 0.4799, "step": 192500 }, { "epoch": 95.07, "grad_norm": 0.5075964331626892, "learning_rate": 1.1442939244663383e-05, "loss": 0.4812, "step": 193000 }, { "epoch": 95.32, "grad_norm": 0.43209370970726013, "learning_rate": 1.1391625615763547e-05, "loss": 0.4796, "step": 193500 }, { "epoch": 95.57, "grad_norm": 0.41466042399406433, "learning_rate": 1.1340311986863712e-05, "loss": 0.4775, "step": 194000 }, { "epoch": 95.81, "grad_norm": 0.3891516625881195, "learning_rate": 1.1288998357963875e-05, "loss": 0.4798, "step": 194500 }, { "epoch": 96.06, "grad_norm": 0.3929733633995056, "learning_rate": 1.123768472906404e-05, "loss": 0.4833, "step": 195000 }, { "epoch": 96.31, "grad_norm": 0.6019779443740845, "learning_rate": 1.1186371100164205e-05, "loss": 0.4788, "step": 195500 }, { "epoch": 96.55, "grad_norm": 0.46904659271240234, "learning_rate": 1.1135057471264368e-05, "loss": 0.4775, "step": 196000 }, { "epoch": 96.8, "grad_norm": 0.4163939356803894, "learning_rate": 1.1083743842364533e-05, "loss": 0.4812, "step": 196500 }, { "epoch": 97.04, "grad_norm": 0.4934261441230774, "learning_rate": 1.1032430213464697e-05, "loss": 0.4808, "step": 197000 }, { "epoch": 97.29, "grad_norm": 0.33812370896339417, "learning_rate": 1.0981116584564862e-05, "loss": 0.4773, "step": 197500 }, { "epoch": 97.54, "grad_norm": 0.5475727319717407, "learning_rate": 1.0929802955665025e-05, "loss": 0.4784, "step": 198000 }, { "epoch": 97.78, "grad_norm": 0.4021857976913452, "learning_rate": 1.087848932676519e-05, "loss": 0.4784, "step": 198500 }, { "epoch": 98.03, "grad_norm": 0.5285155773162842, "learning_rate": 1.0827175697865353e-05, "loss": 0.4814, "step": 199000 }, { "epoch": 98.28, "grad_norm": 0.42629796266555786, "learning_rate": 1.0775862068965518e-05, "loss": 0.4825, "step": 199500 }, { "epoch": 98.52, "grad_norm": 0.38368546962738037, "learning_rate": 1.0724548440065683e-05, "loss": 0.4763, "step": 200000 }, { "epoch": 98.77, "grad_norm": 0.46770936250686646, "learning_rate": 1.0673234811165846e-05, "loss": 0.4752, "step": 200500 }, { "epoch": 99.01, "grad_norm": 0.4895331561565399, "learning_rate": 1.062192118226601e-05, "loss": 0.4804, "step": 201000 }, { "epoch": 99.26, "grad_norm": 0.48920783400535583, "learning_rate": 1.0570607553366173e-05, "loss": 0.4795, "step": 201500 }, { "epoch": 99.51, "grad_norm": 0.4836702346801758, "learning_rate": 1.0519293924466338e-05, "loss": 0.4767, "step": 202000 }, { "epoch": 99.75, "grad_norm": 0.3899345397949219, "learning_rate": 1.0467980295566501e-05, "loss": 0.4784, "step": 202500 }, { "epoch": 100.0, "grad_norm": 1.3630714416503906, "learning_rate": 1.0416666666666666e-05, "loss": 0.4802, "step": 203000 }, { "epoch": 100.25, "grad_norm": 0.48866012692451477, "learning_rate": 1.0365353037766831e-05, "loss": 0.4804, "step": 203500 }, { "epoch": 100.49, "grad_norm": 0.4920850694179535, "learning_rate": 1.0314039408866995e-05, "loss": 0.4754, "step": 204000 }, { "epoch": 100.74, "grad_norm": 0.4752250611782074, "learning_rate": 1.026272577996716e-05, "loss": 0.4807, "step": 204500 }, { "epoch": 100.99, "grad_norm": 0.4868377447128296, "learning_rate": 1.0211412151067323e-05, "loss": 0.476, "step": 205000 }, { "epoch": 101.23, "grad_norm": 0.4661599397659302, "learning_rate": 1.0160098522167488e-05, "loss": 0.4773, "step": 205500 }, { "epoch": 101.48, "grad_norm": 0.5039206147193909, "learning_rate": 1.0108784893267651e-05, "loss": 0.4805, "step": 206000 }, { "epoch": 101.72, "grad_norm": 0.4316484332084656, "learning_rate": 1.0057471264367816e-05, "loss": 0.4737, "step": 206500 }, { "epoch": 101.97, "grad_norm": 0.36680227518081665, "learning_rate": 1.0006157635467981e-05, "loss": 0.4753, "step": 207000 }, { "epoch": 102.22, "grad_norm": 0.4917042553424835, "learning_rate": 9.954844006568145e-06, "loss": 0.4771, "step": 207500 }, { "epoch": 102.46, "grad_norm": 0.3863958716392517, "learning_rate": 9.90353037766831e-06, "loss": 0.4782, "step": 208000 }, { "epoch": 102.71, "grad_norm": 0.495717316865921, "learning_rate": 9.852216748768473e-06, "loss": 0.4763, "step": 208500 }, { "epoch": 102.96, "grad_norm": 0.5366889238357544, "learning_rate": 9.800903119868638e-06, "loss": 0.4765, "step": 209000 }, { "epoch": 103.2, "grad_norm": 0.39207398891448975, "learning_rate": 9.749589490968801e-06, "loss": 0.4752, "step": 209500 }, { "epoch": 103.45, "grad_norm": 0.4651632308959961, "learning_rate": 9.698275862068966e-06, "loss": 0.4769, "step": 210000 }, { "epoch": 103.69, "grad_norm": 0.5077354907989502, "learning_rate": 9.646962233169131e-06, "loss": 0.4732, "step": 210500 }, { "epoch": 103.94, "grad_norm": 0.39063769578933716, "learning_rate": 9.595648604269294e-06, "loss": 0.4773, "step": 211000 }, { "epoch": 104.19, "grad_norm": 0.48797911405563354, "learning_rate": 9.54433497536946e-06, "loss": 0.4769, "step": 211500 }, { "epoch": 104.43, "grad_norm": 0.49572136998176575, "learning_rate": 9.493021346469623e-06, "loss": 0.4742, "step": 212000 }, { "epoch": 104.68, "grad_norm": 0.48770076036453247, "learning_rate": 9.441707717569788e-06, "loss": 0.4777, "step": 212500 }, { "epoch": 104.93, "grad_norm": 0.45337972044944763, "learning_rate": 9.390394088669951e-06, "loss": 0.4731, "step": 213000 }, { "epoch": 105.17, "grad_norm": 0.4577206075191498, "learning_rate": 9.339080459770116e-06, "loss": 0.478, "step": 213500 }, { "epoch": 105.42, "grad_norm": 0.4246939718723297, "learning_rate": 9.28776683087028e-06, "loss": 0.4743, "step": 214000 }, { "epoch": 105.67, "grad_norm": 0.510725736618042, "learning_rate": 9.236453201970444e-06, "loss": 0.4772, "step": 214500 }, { "epoch": 105.91, "grad_norm": 0.4228347837924957, "learning_rate": 9.185139573070608e-06, "loss": 0.4745, "step": 215000 }, { "epoch": 106.16, "grad_norm": 0.5115532875061035, "learning_rate": 9.133825944170771e-06, "loss": 0.4753, "step": 215500 }, { "epoch": 106.4, "grad_norm": 0.4812858998775482, "learning_rate": 9.082512315270936e-06, "loss": 0.4753, "step": 216000 }, { "epoch": 106.65, "grad_norm": 0.5218610167503357, "learning_rate": 9.0311986863711e-06, "loss": 0.4786, "step": 216500 }, { "epoch": 106.9, "grad_norm": 0.5687581896781921, "learning_rate": 8.979885057471264e-06, "loss": 0.4733, "step": 217000 }, { "epoch": 107.14, "grad_norm": 0.4318714141845703, "learning_rate": 8.928571428571428e-06, "loss": 0.4756, "step": 217500 }, { "epoch": 107.39, "grad_norm": 0.5956067442893982, "learning_rate": 8.877257799671593e-06, "loss": 0.4755, "step": 218000 }, { "epoch": 107.64, "grad_norm": 0.5590375065803528, "learning_rate": 8.825944170771758e-06, "loss": 0.4759, "step": 218500 }, { "epoch": 107.88, "grad_norm": 0.5045220851898193, "learning_rate": 8.774630541871921e-06, "loss": 0.4734, "step": 219000 }, { "epoch": 108.13, "grad_norm": 0.466327428817749, "learning_rate": 8.723316912972086e-06, "loss": 0.4753, "step": 219500 }, { "epoch": 108.37, "grad_norm": 0.4527340829372406, "learning_rate": 8.672003284072249e-06, "loss": 0.4728, "step": 220000 }, { "epoch": 108.62, "grad_norm": 0.398181676864624, "learning_rate": 8.620689655172414e-06, "loss": 0.4742, "step": 220500 }, { "epoch": 108.87, "grad_norm": 0.5564948320388794, "learning_rate": 8.569376026272577e-06, "loss": 0.4742, "step": 221000 }, { "epoch": 109.11, "grad_norm": 0.4469916522502899, "learning_rate": 8.518062397372742e-06, "loss": 0.4721, "step": 221500 }, { "epoch": 109.36, "grad_norm": 0.45744919776916504, "learning_rate": 8.466748768472907e-06, "loss": 0.4731, "step": 222000 }, { "epoch": 109.61, "grad_norm": 0.5253536105155945, "learning_rate": 8.41543513957307e-06, "loss": 0.4755, "step": 222500 }, { "epoch": 109.85, "grad_norm": 0.5014792680740356, "learning_rate": 8.364121510673236e-06, "loss": 0.4745, "step": 223000 }, { "epoch": 110.1, "grad_norm": 0.44078388810157776, "learning_rate": 8.312807881773399e-06, "loss": 0.4735, "step": 223500 }, { "epoch": 110.34, "grad_norm": 0.5724578499794006, "learning_rate": 8.261494252873564e-06, "loss": 0.4742, "step": 224000 }, { "epoch": 110.59, "grad_norm": 0.5114606022834778, "learning_rate": 8.210180623973727e-06, "loss": 0.4736, "step": 224500 }, { "epoch": 110.84, "grad_norm": 0.5526043176651001, "learning_rate": 8.158866995073892e-06, "loss": 0.4753, "step": 225000 }, { "epoch": 111.08, "grad_norm": 0.36584803462028503, "learning_rate": 8.107553366174056e-06, "loss": 0.4758, "step": 225500 }, { "epoch": 111.33, "grad_norm": 0.5053611397743225, "learning_rate": 8.05623973727422e-06, "loss": 0.4699, "step": 226000 }, { "epoch": 111.58, "grad_norm": 0.47798970341682434, "learning_rate": 8.004926108374386e-06, "loss": 0.4694, "step": 226500 }, { "epoch": 111.82, "grad_norm": 0.42536744475364685, "learning_rate": 7.953612479474549e-06, "loss": 0.4728, "step": 227000 }, { "epoch": 112.07, "grad_norm": 0.451180100440979, "learning_rate": 7.902298850574714e-06, "loss": 0.4788, "step": 227500 }, { "epoch": 112.32, "grad_norm": 0.4156660735607147, "learning_rate": 7.850985221674877e-06, "loss": 0.4726, "step": 228000 }, { "epoch": 112.56, "grad_norm": 0.5824641585350037, "learning_rate": 7.79967159277504e-06, "loss": 0.4713, "step": 228500 }, { "epoch": 112.81, "grad_norm": 0.36896491050720215, "learning_rate": 7.748357963875206e-06, "loss": 0.4742, "step": 229000 }, { "epoch": 113.05, "grad_norm": 0.46299970149993896, "learning_rate": 7.69704433497537e-06, "loss": 0.4724, "step": 229500 }, { "epoch": 113.3, "grad_norm": 0.43692949414253235, "learning_rate": 7.645730706075534e-06, "loss": 0.4757, "step": 230000 }, { "epoch": 113.55, "grad_norm": 0.47571897506713867, "learning_rate": 7.594417077175698e-06, "loss": 0.469, "step": 230500 }, { "epoch": 113.79, "grad_norm": 0.5554032921791077, "learning_rate": 7.543103448275862e-06, "loss": 0.4706, "step": 231000 }, { "epoch": 114.04, "grad_norm": 0.5689848065376282, "learning_rate": 7.491789819376027e-06, "loss": 0.4727, "step": 231500 }, { "epoch": 114.29, "grad_norm": 0.43025562167167664, "learning_rate": 7.4404761904761905e-06, "loss": 0.4741, "step": 232000 }, { "epoch": 114.53, "grad_norm": 0.4013258218765259, "learning_rate": 7.389162561576355e-06, "loss": 0.4756, "step": 232500 }, { "epoch": 114.78, "grad_norm": 0.4342605769634247, "learning_rate": 7.337848932676519e-06, "loss": 0.4682, "step": 233000 }, { "epoch": 115.02, "grad_norm": 0.4698561131954193, "learning_rate": 7.286535303776683e-06, "loss": 0.4734, "step": 233500 }, { "epoch": 115.27, "grad_norm": 0.6582350134849548, "learning_rate": 7.235221674876847e-06, "loss": 0.4752, "step": 234000 }, { "epoch": 115.52, "grad_norm": 0.4564856290817261, "learning_rate": 7.183908045977011e-06, "loss": 0.4716, "step": 234500 }, { "epoch": 115.76, "grad_norm": 0.5398574471473694, "learning_rate": 7.132594417077176e-06, "loss": 0.4695, "step": 235000 }, { "epoch": 116.01, "grad_norm": 0.4755443036556244, "learning_rate": 7.08128078817734e-06, "loss": 0.47, "step": 235500 }, { "epoch": 116.26, "grad_norm": 0.49625080823898315, "learning_rate": 7.0299671592775045e-06, "loss": 0.4724, "step": 236000 }, { "epoch": 116.5, "grad_norm": 0.44892576336860657, "learning_rate": 6.978653530377669e-06, "loss": 0.4716, "step": 236500 }, { "epoch": 116.75, "grad_norm": 0.38705721497535706, "learning_rate": 6.927339901477833e-06, "loss": 0.4727, "step": 237000 }, { "epoch": 117.0, "grad_norm": 0.4698588252067566, "learning_rate": 6.876026272577997e-06, "loss": 0.4709, "step": 237500 }, { "epoch": 117.24, "grad_norm": 0.5400373339653015, "learning_rate": 6.824712643678161e-06, "loss": 0.4708, "step": 238000 }, { "epoch": 117.49, "grad_norm": 0.4378606379032135, "learning_rate": 6.773399014778325e-06, "loss": 0.4755, "step": 238500 }, { "epoch": 117.73, "grad_norm": 0.531797468662262, "learning_rate": 6.7220853858784894e-06, "loss": 0.4668, "step": 239000 }, { "epoch": 117.98, "grad_norm": 0.5268296003341675, "learning_rate": 6.670771756978654e-06, "loss": 0.4714, "step": 239500 }, { "epoch": 118.23, "grad_norm": 0.5087544918060303, "learning_rate": 6.619458128078818e-06, "loss": 0.4736, "step": 240000 }, { "epoch": 118.47, "grad_norm": 0.5163730382919312, "learning_rate": 6.568144499178982e-06, "loss": 0.4687, "step": 240500 }, { "epoch": 118.72, "grad_norm": 0.5498948097229004, "learning_rate": 6.516830870279146e-06, "loss": 0.4703, "step": 241000 }, { "epoch": 118.97, "grad_norm": 0.4399949610233307, "learning_rate": 6.46551724137931e-06, "loss": 0.4682, "step": 241500 }, { "epoch": 119.21, "grad_norm": 0.5559102892875671, "learning_rate": 6.414203612479474e-06, "loss": 0.4705, "step": 242000 }, { "epoch": 119.46, "grad_norm": 0.5814414620399475, "learning_rate": 6.362889983579639e-06, "loss": 0.4719, "step": 242500 }, { "epoch": 119.7, "grad_norm": 0.477924108505249, "learning_rate": 6.3115763546798035e-06, "loss": 0.4681, "step": 243000 }, { "epoch": 119.95, "grad_norm": 0.6359853744506836, "learning_rate": 6.260262725779968e-06, "loss": 0.4724, "step": 243500 }, { "epoch": 120.2, "grad_norm": 0.43428707122802734, "learning_rate": 6.208949096880132e-06, "loss": 0.4705, "step": 244000 }, { "epoch": 120.44, "grad_norm": 0.5314174294471741, "learning_rate": 6.157635467980296e-06, "loss": 0.4671, "step": 244500 }, { "epoch": 120.69, "grad_norm": 0.5369435548782349, "learning_rate": 6.10632183908046e-06, "loss": 0.4719, "step": 245000 }, { "epoch": 120.94, "grad_norm": 0.4553278684616089, "learning_rate": 6.055008210180624e-06, "loss": 0.469, "step": 245500 }, { "epoch": 121.18, "grad_norm": 0.4905393123626709, "learning_rate": 6.0036945812807875e-06, "loss": 0.4698, "step": 246000 }, { "epoch": 121.43, "grad_norm": 0.5438905954360962, "learning_rate": 5.9523809523809525e-06, "loss": 0.4719, "step": 246500 }, { "epoch": 121.67, "grad_norm": 0.5870608687400818, "learning_rate": 5.901067323481117e-06, "loss": 0.4693, "step": 247000 }, { "epoch": 121.92, "grad_norm": 0.5743767619132996, "learning_rate": 5.849753694581281e-06, "loss": 0.4676, "step": 247500 }, { "epoch": 122.17, "grad_norm": 0.5572515726089478, "learning_rate": 5.798440065681445e-06, "loss": 0.4691, "step": 248000 }, { "epoch": 122.41, "grad_norm": 0.5879077315330505, "learning_rate": 5.747126436781609e-06, "loss": 0.4664, "step": 248500 }, { "epoch": 122.66, "grad_norm": 0.48736339807510376, "learning_rate": 5.695812807881773e-06, "loss": 0.469, "step": 249000 }, { "epoch": 122.91, "grad_norm": 0.39413416385650635, "learning_rate": 5.6444991789819375e-06, "loss": 0.4716, "step": 249500 }, { "epoch": 123.15, "grad_norm": 0.5965219140052795, "learning_rate": 5.5931855500821024e-06, "loss": 0.469, "step": 250000 }, { "epoch": 123.4, "grad_norm": 0.48718106746673584, "learning_rate": 5.541871921182267e-06, "loss": 0.4684, "step": 250500 }, { "epoch": 123.65, "grad_norm": 0.5316205620765686, "learning_rate": 5.490558292282431e-06, "loss": 0.4691, "step": 251000 }, { "epoch": 123.89, "grad_norm": 0.5446661114692688, "learning_rate": 5.439244663382595e-06, "loss": 0.4678, "step": 251500 }, { "epoch": 124.14, "grad_norm": 0.41561558842658997, "learning_rate": 5.387931034482759e-06, "loss": 0.4676, "step": 252000 }, { "epoch": 124.38, "grad_norm": 0.4405909776687622, "learning_rate": 5.336617405582923e-06, "loss": 0.4741, "step": 252500 }, { "epoch": 124.63, "grad_norm": 0.5378337502479553, "learning_rate": 5.2853037766830865e-06, "loss": 0.4685, "step": 253000 }, { "epoch": 124.88, "grad_norm": 0.5120199918746948, "learning_rate": 5.233990147783251e-06, "loss": 0.465, "step": 253500 }, { "epoch": 125.12, "grad_norm": 0.5217579007148743, "learning_rate": 5.182676518883416e-06, "loss": 0.4602, "step": 254000 }, { "epoch": 125.37, "grad_norm": 0.4467191696166992, "learning_rate": 5.13136288998358e-06, "loss": 0.4676, "step": 254500 }, { "epoch": 125.62, "grad_norm": 0.5314244031906128, "learning_rate": 5.080049261083744e-06, "loss": 0.4729, "step": 255000 }, { "epoch": 125.86, "grad_norm": 0.5660964250564575, "learning_rate": 5.028735632183908e-06, "loss": 0.4701, "step": 255500 }, { "epoch": 126.11, "grad_norm": 0.6407245993614197, "learning_rate": 4.977422003284072e-06, "loss": 0.4654, "step": 256000 }, { "epoch": 126.35, "grad_norm": 0.6110518574714661, "learning_rate": 4.926108374384236e-06, "loss": 0.4674, "step": 256500 }, { "epoch": 126.6, "grad_norm": 0.5407283306121826, "learning_rate": 4.8747947454844006e-06, "loss": 0.4711, "step": 257000 }, { "epoch": 126.85, "grad_norm": 0.5181688070297241, "learning_rate": 4.8234811165845656e-06, "loss": 0.4705, "step": 257500 }, { "epoch": 127.09, "grad_norm": 0.6265803575515747, "learning_rate": 4.77216748768473e-06, "loss": 0.4688, "step": 258000 }, { "epoch": 127.34, "grad_norm": 0.38801223039627075, "learning_rate": 4.720853858784894e-06, "loss": 0.4673, "step": 258500 }, { "epoch": 127.59, "grad_norm": 0.5875949263572693, "learning_rate": 4.669540229885058e-06, "loss": 0.4643, "step": 259000 }, { "epoch": 127.83, "grad_norm": 0.5323173403739929, "learning_rate": 4.618226600985222e-06, "loss": 0.4697, "step": 259500 }, { "epoch": 128.08, "grad_norm": 0.5532727241516113, "learning_rate": 4.5669129720853855e-06, "loss": 0.4682, "step": 260000 }, { "epoch": 128.33, "grad_norm": 0.5300949811935425, "learning_rate": 4.51559934318555e-06, "loss": 0.4687, "step": 260500 }, { "epoch": 128.57, "grad_norm": 0.7153774499893188, "learning_rate": 4.464285714285714e-06, "loss": 0.4678, "step": 261000 }, { "epoch": 128.82, "grad_norm": 0.5258143544197083, "learning_rate": 4.412972085385879e-06, "loss": 0.4706, "step": 261500 }, { "epoch": 129.06, "grad_norm": 0.5947761535644531, "learning_rate": 4.361658456486043e-06, "loss": 0.4661, "step": 262000 }, { "epoch": 129.31, "grad_norm": 0.5369092226028442, "learning_rate": 4.310344827586207e-06, "loss": 0.4687, "step": 262500 }, { "epoch": 129.56, "grad_norm": 0.5336856245994568, "learning_rate": 4.259031198686371e-06, "loss": 0.4626, "step": 263000 }, { "epoch": 129.8, "grad_norm": 0.5533296465873718, "learning_rate": 4.207717569786535e-06, "loss": 0.4691, "step": 263500 }, { "epoch": 130.05, "grad_norm": 0.579079270362854, "learning_rate": 4.1564039408866995e-06, "loss": 0.4665, "step": 264000 }, { "epoch": 130.3, "grad_norm": 0.6150951385498047, "learning_rate": 4.105090311986864e-06, "loss": 0.4667, "step": 264500 }, { "epoch": 130.54, "grad_norm": 0.5602375864982605, "learning_rate": 4.053776683087028e-06, "loss": 0.4706, "step": 265000 }, { "epoch": 130.79, "grad_norm": 0.58797287940979, "learning_rate": 4.002463054187193e-06, "loss": 0.4664, "step": 265500 }, { "epoch": 131.03, "grad_norm": 0.39354264736175537, "learning_rate": 3.951149425287357e-06, "loss": 0.465, "step": 266000 }, { "epoch": 131.28, "grad_norm": 0.5420950055122375, "learning_rate": 3.89983579638752e-06, "loss": 0.466, "step": 266500 }, { "epoch": 131.53, "grad_norm": 0.5339276790618896, "learning_rate": 3.848522167487685e-06, "loss": 0.4626, "step": 267000 }, { "epoch": 131.77, "grad_norm": 0.6729488372802734, "learning_rate": 3.797208538587849e-06, "loss": 0.4693, "step": 267500 }, { "epoch": 132.02, "grad_norm": 0.5628036260604858, "learning_rate": 3.7458949096880136e-06, "loss": 0.466, "step": 268000 }, { "epoch": 132.27, "grad_norm": 0.5665927529335022, "learning_rate": 3.6945812807881773e-06, "loss": 0.4625, "step": 268500 }, { "epoch": 132.51, "grad_norm": 0.5044068098068237, "learning_rate": 3.6432676518883415e-06, "loss": 0.4644, "step": 269000 }, { "epoch": 132.76, "grad_norm": 0.504945695400238, "learning_rate": 3.5919540229885056e-06, "loss": 0.4695, "step": 269500 }, { "epoch": 133.0, "grad_norm": 0.4701473116874695, "learning_rate": 3.54064039408867e-06, "loss": 0.4652, "step": 270000 }, { "epoch": 133.25, "grad_norm": 0.5132540464401245, "learning_rate": 3.4893267651888343e-06, "loss": 0.4647, "step": 270500 }, { "epoch": 133.5, "grad_norm": 0.652473509311676, "learning_rate": 3.4380131362889985e-06, "loss": 0.4674, "step": 271000 }, { "epoch": 133.74, "grad_norm": 0.5050608515739441, "learning_rate": 3.3866995073891626e-06, "loss": 0.4661, "step": 271500 }, { "epoch": 133.99, "grad_norm": 0.6116757988929749, "learning_rate": 3.335385878489327e-06, "loss": 0.4684, "step": 272000 }, { "epoch": 134.24, "grad_norm": 0.4631555676460266, "learning_rate": 3.284072249589491e-06, "loss": 0.4695, "step": 272500 }, { "epoch": 134.48, "grad_norm": 0.5643542408943176, "learning_rate": 3.232758620689655e-06, "loss": 0.4658, "step": 273000 }, { "epoch": 134.73, "grad_norm": 0.6550360918045044, "learning_rate": 3.1814449917898197e-06, "loss": 0.4696, "step": 273500 }, { "epoch": 134.98, "grad_norm": 0.5142180919647217, "learning_rate": 3.130131362889984e-06, "loss": 0.4624, "step": 274000 }, { "epoch": 135.22, "grad_norm": 0.5686226487159729, "learning_rate": 3.078817733990148e-06, "loss": 0.4638, "step": 274500 }, { "epoch": 135.47, "grad_norm": 0.551745593547821, "learning_rate": 3.027504105090312e-06, "loss": 0.4655, "step": 275000 }, { "epoch": 135.71, "grad_norm": 0.5265413522720337, "learning_rate": 2.9761904761904763e-06, "loss": 0.4649, "step": 275500 }, { "epoch": 135.96, "grad_norm": 0.504638671875, "learning_rate": 2.9248768472906404e-06, "loss": 0.4658, "step": 276000 }, { "epoch": 136.21, "grad_norm": 0.5403110980987549, "learning_rate": 2.8735632183908046e-06, "loss": 0.4659, "step": 276500 }, { "epoch": 136.45, "grad_norm": 0.7049803137779236, "learning_rate": 2.8222495894909687e-06, "loss": 0.4657, "step": 277000 }, { "epoch": 136.7, "grad_norm": 0.46327295899391174, "learning_rate": 2.7709359605911333e-06, "loss": 0.4676, "step": 277500 }, { "epoch": 136.95, "grad_norm": 0.7414257526397705, "learning_rate": 2.7196223316912975e-06, "loss": 0.4673, "step": 278000 }, { "epoch": 137.19, "grad_norm": 0.528343915939331, "learning_rate": 2.6683087027914616e-06, "loss": 0.4659, "step": 278500 }, { "epoch": 137.44, "grad_norm": 0.6523202657699585, "learning_rate": 2.6169950738916253e-06, "loss": 0.4656, "step": 279000 }, { "epoch": 137.68, "grad_norm": 0.496900737285614, "learning_rate": 2.56568144499179e-06, "loss": 0.4659, "step": 279500 }, { "epoch": 137.93, "grad_norm": 0.6677756309509277, "learning_rate": 2.514367816091954e-06, "loss": 0.4644, "step": 280000 }, { "epoch": 138.18, "grad_norm": 0.5693820118904114, "learning_rate": 2.463054187192118e-06, "loss": 0.4648, "step": 280500 }, { "epoch": 138.42, "grad_norm": 0.7365754842758179, "learning_rate": 2.4117405582922828e-06, "loss": 0.4653, "step": 281000 }, { "epoch": 138.67, "grad_norm": 0.7357062697410583, "learning_rate": 2.360426929392447e-06, "loss": 0.4666, "step": 281500 }, { "epoch": 138.92, "grad_norm": 0.6288078427314758, "learning_rate": 2.309113300492611e-06, "loss": 0.4628, "step": 282000 }, { "epoch": 139.16, "grad_norm": 0.5149612426757812, "learning_rate": 2.257799671592775e-06, "loss": 0.4629, "step": 282500 }, { "epoch": 139.41, "grad_norm": 0.572669267654419, "learning_rate": 2.2064860426929394e-06, "loss": 0.4633, "step": 283000 }, { "epoch": 139.66, "grad_norm": 0.5936819314956665, "learning_rate": 2.1551724137931035e-06, "loss": 0.4657, "step": 283500 }, { "epoch": 139.9, "grad_norm": 0.5847501754760742, "learning_rate": 2.1038587848932677e-06, "loss": 0.4664, "step": 284000 }, { "epoch": 140.15, "grad_norm": 0.514391303062439, "learning_rate": 2.052545155993432e-06, "loss": 0.4645, "step": 284500 }, { "epoch": 140.39, "grad_norm": 0.6398583650588989, "learning_rate": 2.0012315270935964e-06, "loss": 0.4658, "step": 285000 }, { "epoch": 140.64, "grad_norm": 0.8860574960708618, "learning_rate": 1.94991789819376e-06, "loss": 0.4646, "step": 285500 }, { "epoch": 140.89, "grad_norm": 0.5785859823226929, "learning_rate": 1.8986042692939245e-06, "loss": 0.4656, "step": 286000 }, { "epoch": 141.13, "grad_norm": 0.5147131681442261, "learning_rate": 1.8472906403940887e-06, "loss": 0.4636, "step": 286500 }, { "epoch": 141.38, "grad_norm": 0.49100789427757263, "learning_rate": 1.7959770114942528e-06, "loss": 0.4641, "step": 287000 }, { "epoch": 141.63, "grad_norm": 0.5071477293968201, "learning_rate": 1.7446633825944172e-06, "loss": 0.4652, "step": 287500 }, { "epoch": 141.87, "grad_norm": 0.5541560649871826, "learning_rate": 1.6933497536945813e-06, "loss": 0.4661, "step": 288000 }, { "epoch": 142.12, "grad_norm": 0.5373649597167969, "learning_rate": 1.6420361247947455e-06, "loss": 0.466, "step": 288500 }, { "epoch": 142.36, "grad_norm": 0.6542106866836548, "learning_rate": 1.5907224958949098e-06, "loss": 0.4596, "step": 289000 }, { "epoch": 142.61, "grad_norm": 0.41288328170776367, "learning_rate": 1.539408866995074e-06, "loss": 0.4633, "step": 289500 }, { "epoch": 142.86, "grad_norm": 0.6059596538543701, "learning_rate": 1.4880952380952381e-06, "loss": 0.4648, "step": 290000 }, { "epoch": 143.1, "grad_norm": 0.5577875375747681, "learning_rate": 1.4367816091954023e-06, "loss": 0.4688, "step": 290500 }, { "epoch": 143.35, "grad_norm": 0.5939833521842957, "learning_rate": 1.3854679802955667e-06, "loss": 0.4653, "step": 291000 }, { "epoch": 143.6, "grad_norm": 0.5009748935699463, "learning_rate": 1.3341543513957308e-06, "loss": 0.4599, "step": 291500 }, { "epoch": 143.84, "grad_norm": 0.6264510154724121, "learning_rate": 1.282840722495895e-06, "loss": 0.4619, "step": 292000 }, { "epoch": 144.09, "grad_norm": 0.4556388258934021, "learning_rate": 1.231527093596059e-06, "loss": 0.4654, "step": 292500 }, { "epoch": 144.33, "grad_norm": 0.5879510045051575, "learning_rate": 1.1802134646962235e-06, "loss": 0.4672, "step": 293000 }, { "epoch": 144.58, "grad_norm": 0.515290379524231, "learning_rate": 1.1288998357963874e-06, "loss": 0.4622, "step": 293500 }, { "epoch": 144.83, "grad_norm": 0.6605761647224426, "learning_rate": 1.0775862068965518e-06, "loss": 0.4621, "step": 294000 }, { "epoch": 145.07, "grad_norm": 0.5039830803871155, "learning_rate": 1.026272577996716e-06, "loss": 0.4619, "step": 294500 }, { "epoch": 145.32, "grad_norm": 0.5800752639770508, "learning_rate": 9.7495894909688e-07, "loss": 0.4631, "step": 295000 }, { "epoch": 145.57, "grad_norm": 0.7062521576881409, "learning_rate": 9.236453201970443e-07, "loss": 0.464, "step": 295500 }, { "epoch": 145.81, "grad_norm": 0.5099909901618958, "learning_rate": 8.723316912972086e-07, "loss": 0.4618, "step": 296000 }, { "epoch": 146.06, "grad_norm": 0.5134597420692444, "learning_rate": 8.210180623973727e-07, "loss": 0.4643, "step": 296500 }, { "epoch": 146.31, "grad_norm": 0.497597336769104, "learning_rate": 7.69704433497537e-07, "loss": 0.4643, "step": 297000 }, { "epoch": 146.55, "grad_norm": 0.613549530506134, "learning_rate": 7.183908045977011e-07, "loss": 0.4598, "step": 297500 }, { "epoch": 146.8, "grad_norm": 0.5238372683525085, "learning_rate": 6.670771756978654e-07, "loss": 0.4653, "step": 298000 }, { "epoch": 147.04, "grad_norm": 0.673534095287323, "learning_rate": 6.157635467980296e-07, "loss": 0.4617, "step": 298500 }, { "epoch": 147.29, "grad_norm": 0.6518653631210327, "learning_rate": 5.644499178981937e-07, "loss": 0.4625, "step": 299000 }, { "epoch": 147.54, "grad_norm": 0.564731776714325, "learning_rate": 5.13136288998358e-07, "loss": 0.4607, "step": 299500 }, { "epoch": 147.78, "grad_norm": 0.5317474603652954, "learning_rate": 4.6182266009852216e-07, "loss": 0.4635, "step": 300000 }, { "epoch": 148.03, "grad_norm": 0.5215335488319397, "learning_rate": 4.1050903119868637e-07, "loss": 0.4642, "step": 300500 }, { "epoch": 148.28, "grad_norm": 0.5172483325004578, "learning_rate": 3.5919540229885057e-07, "loss": 0.4593, "step": 301000 }, { "epoch": 148.52, "grad_norm": 0.49218830466270447, "learning_rate": 3.078817733990148e-07, "loss": 0.466, "step": 301500 }, { "epoch": 148.77, "grad_norm": 0.502910852432251, "learning_rate": 2.56568144499179e-07, "loss": 0.4622, "step": 302000 }, { "epoch": 149.01, "grad_norm": 0.4450142979621887, "learning_rate": 2.0525451559934318e-07, "loss": 0.4653, "step": 302500 }, { "epoch": 149.26, "grad_norm": 0.6324329376220703, "learning_rate": 1.539408866995074e-07, "loss": 0.4606, "step": 303000 }, { "epoch": 149.51, "grad_norm": 0.5597870945930481, "learning_rate": 1.0262725779967159e-07, "loss": 0.4619, "step": 303500 }, { "epoch": 149.75, "grad_norm": 0.6023632884025574, "learning_rate": 5.1313628899835796e-08, "loss": 0.4671, "step": 304000 }, { "epoch": 150.0, "grad_norm": 1.9860941171646118, "learning_rate": 0.0, "loss": 0.4623, "step": 304500 }, { "epoch": 150.0, "step": 304500, "total_flos": 2.4610424381138534e+20, "train_loss": 0.5040761357951047, "train_runtime": 31987.2375, "train_samples_per_second": 76.123, "train_steps_per_second": 9.519 } ], "logging_steps": 500, "max_steps": 304500, "num_input_tokens_seen": 0, "num_train_epochs": 150, "save_steps": 1000000000, "total_flos": 2.4610424381138534e+20, "train_batch_size": 8, "trial_name": null, "trial_params": null }