{ "best_metric": 0.3917270302772522, "best_model_checkpoint": "mikhail-panzo/ceb_b32_le5_s8000/checkpoint-6500", "epoch": 158.41584158415841, "eval_steps": 500, "global_step": 8000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.9900990099009901, "grad_norm": 3.180250644683838, "learning_rate": 2.4500000000000004e-07, "loss": 0.8005, "step": 50 }, { "epoch": 1.9801980198019802, "grad_norm": 3.2645788192749023, "learning_rate": 4.95e-07, "loss": 0.7901, "step": 100 }, { "epoch": 2.9702970297029703, "grad_norm": 2.118293523788452, "learning_rate": 7.450000000000001e-07, "loss": 0.7371, "step": 150 }, { "epoch": 3.9603960396039604, "grad_norm": 1.8010900020599365, "learning_rate": 9.950000000000002e-07, "loss": 0.7469, "step": 200 }, { "epoch": 4.9504950495049505, "grad_norm": 1.6043368577957153, "learning_rate": 1.2450000000000002e-06, "loss": 0.7138, "step": 250 }, { "epoch": 5.9405940594059405, "grad_norm": 2.1476693153381348, "learning_rate": 1.495e-06, "loss": 0.7011, "step": 300 }, { "epoch": 6.930693069306931, "grad_norm": 1.7792506217956543, "learning_rate": 1.745e-06, "loss": 0.6728, "step": 350 }, { "epoch": 7.920792079207921, "grad_norm": 2.7465362548828125, "learning_rate": 1.9950000000000004e-06, "loss": 0.6505, "step": 400 }, { "epoch": 8.910891089108912, "grad_norm": 1.8483725786209106, "learning_rate": 2.245e-06, "loss": 0.617, "step": 450 }, { "epoch": 9.900990099009901, "grad_norm": 2.8589906692504883, "learning_rate": 2.4950000000000003e-06, "loss": 0.5667, "step": 500 }, { "epoch": 9.900990099009901, "eval_loss": 0.47756776213645935, "eval_runtime": 8.0921, "eval_samples_per_second": 22.244, "eval_steps_per_second": 2.842, "step": 500 }, { "epoch": 10.891089108910892, "grad_norm": 2.1117241382598877, "learning_rate": 2.7450000000000004e-06, "loss": 0.5433, "step": 550 }, { "epoch": 11.881188118811881, "grad_norm": 1.311270833015442, "learning_rate": 2.995e-06, "loss": 0.5224, "step": 600 }, { "epoch": 12.871287128712872, "grad_norm": 2.1834664344787598, "learning_rate": 3.2450000000000003e-06, "loss": 0.5134, "step": 650 }, { "epoch": 13.861386138613861, "grad_norm": 1.2540502548217773, "learning_rate": 3.495e-06, "loss": 0.4977, "step": 700 }, { "epoch": 14.851485148514852, "grad_norm": 1.5720808506011963, "learning_rate": 3.745e-06, "loss": 0.4992, "step": 750 }, { "epoch": 15.841584158415841, "grad_norm": 1.3779664039611816, "learning_rate": 3.995000000000001e-06, "loss": 0.5041, "step": 800 }, { "epoch": 16.831683168316832, "grad_norm": 1.2948031425476074, "learning_rate": 4.245e-06, "loss": 0.4998, "step": 850 }, { "epoch": 17.821782178217823, "grad_norm": 1.5043249130249023, "learning_rate": 4.495e-06, "loss": 0.4908, "step": 900 }, { "epoch": 18.81188118811881, "grad_norm": 1.9952532052993774, "learning_rate": 4.745e-06, "loss": 0.4828, "step": 950 }, { "epoch": 19.801980198019802, "grad_norm": 1.1994812488555908, "learning_rate": 4.9950000000000005e-06, "loss": 0.4838, "step": 1000 }, { "epoch": 19.801980198019802, "eval_loss": 0.4320996403694153, "eval_runtime": 7.9832, "eval_samples_per_second": 22.547, "eval_steps_per_second": 2.881, "step": 1000 }, { "epoch": 20.792079207920793, "grad_norm": 2.356138229370117, "learning_rate": 5.245e-06, "loss": 0.4784, "step": 1050 }, { "epoch": 21.782178217821784, "grad_norm": 1.6155955791473389, "learning_rate": 5.495000000000001e-06, "loss": 0.4696, "step": 1100 }, { "epoch": 22.77227722772277, "grad_norm": 1.3177858591079712, "learning_rate": 5.745000000000001e-06, "loss": 0.4694, "step": 1150 }, { "epoch": 23.762376237623762, "grad_norm": 1.5806872844696045, "learning_rate": 5.995000000000001e-06, "loss": 0.4746, "step": 1200 }, { "epoch": 24.752475247524753, "grad_norm": 2.288752794265747, "learning_rate": 6.245000000000001e-06, "loss": 0.4676, "step": 1250 }, { "epoch": 25.742574257425744, "grad_norm": 1.542547583580017, "learning_rate": 6.4950000000000005e-06, "loss": 0.4773, "step": 1300 }, { "epoch": 26.73267326732673, "grad_norm": 1.0209100246429443, "learning_rate": 6.745000000000001e-06, "loss": 0.4683, "step": 1350 }, { "epoch": 27.722772277227723, "grad_norm": 1.2771400213241577, "learning_rate": 6.995000000000001e-06, "loss": 0.4577, "step": 1400 }, { "epoch": 28.712871287128714, "grad_norm": 1.348565936088562, "learning_rate": 7.245000000000001e-06, "loss": 0.4563, "step": 1450 }, { "epoch": 29.702970297029704, "grad_norm": 1.5620756149291992, "learning_rate": 7.495000000000001e-06, "loss": 0.4604, "step": 1500 }, { "epoch": 29.702970297029704, "eval_loss": 0.41570472717285156, "eval_runtime": 7.3746, "eval_samples_per_second": 24.408, "eval_steps_per_second": 3.119, "step": 1500 }, { "epoch": 30.693069306930692, "grad_norm": 1.3051259517669678, "learning_rate": 7.745e-06, "loss": 0.4551, "step": 1550 }, { "epoch": 31.683168316831683, "grad_norm": 1.6832257509231567, "learning_rate": 7.995e-06, "loss": 0.4517, "step": 1600 }, { "epoch": 32.67326732673267, "grad_norm": 2.1702144145965576, "learning_rate": 8.245000000000002e-06, "loss": 0.4546, "step": 1650 }, { "epoch": 33.663366336633665, "grad_norm": 1.5217938423156738, "learning_rate": 8.495e-06, "loss": 0.4549, "step": 1700 }, { "epoch": 34.65346534653465, "grad_norm": 1.8007372617721558, "learning_rate": 8.745000000000002e-06, "loss": 0.4453, "step": 1750 }, { "epoch": 35.64356435643565, "grad_norm": 1.490868091583252, "learning_rate": 8.995000000000001e-06, "loss": 0.4522, "step": 1800 }, { "epoch": 36.633663366336634, "grad_norm": 1.3354467153549194, "learning_rate": 9.245e-06, "loss": 0.4533, "step": 1850 }, { "epoch": 37.62376237623762, "grad_norm": 1.661230206489563, "learning_rate": 9.495000000000001e-06, "loss": 0.4467, "step": 1900 }, { "epoch": 38.613861386138616, "grad_norm": 1.4943360090255737, "learning_rate": 9.745e-06, "loss": 0.4451, "step": 1950 }, { "epoch": 39.603960396039604, "grad_norm": 2.01084566116333, "learning_rate": 9.995000000000002e-06, "loss": 0.4373, "step": 2000 }, { "epoch": 39.603960396039604, "eval_loss": 0.40344154834747314, "eval_runtime": 8.1054, "eval_samples_per_second": 22.207, "eval_steps_per_second": 2.838, "step": 2000 }, { "epoch": 40.59405940594059, "grad_norm": 2.1528162956237793, "learning_rate": 9.918333333333335e-06, "loss": 0.4446, "step": 2050 }, { "epoch": 41.584158415841586, "grad_norm": 1.692417025566101, "learning_rate": 9.835000000000002e-06, "loss": 0.4363, "step": 2100 }, { "epoch": 42.57425742574257, "grad_norm": 1.190176248550415, "learning_rate": 9.751666666666667e-06, "loss": 0.4425, "step": 2150 }, { "epoch": 43.56435643564357, "grad_norm": 1.027122974395752, "learning_rate": 9.668333333333334e-06, "loss": 0.4318, "step": 2200 }, { "epoch": 44.554455445544555, "grad_norm": 1.3497254848480225, "learning_rate": 9.585e-06, "loss": 0.4377, "step": 2250 }, { "epoch": 45.54455445544554, "grad_norm": 1.483748197555542, "learning_rate": 9.501666666666667e-06, "loss": 0.4358, "step": 2300 }, { "epoch": 46.53465346534654, "grad_norm": 1.2555333375930786, "learning_rate": 9.418333333333334e-06, "loss": 0.4397, "step": 2350 }, { "epoch": 47.524752475247524, "grad_norm": 1.345694899559021, "learning_rate": 9.335000000000001e-06, "loss": 0.4341, "step": 2400 }, { "epoch": 48.51485148514851, "grad_norm": 1.3507237434387207, "learning_rate": 9.251666666666668e-06, "loss": 0.4363, "step": 2450 }, { "epoch": 49.504950495049506, "grad_norm": 2.3417470455169678, "learning_rate": 9.168333333333333e-06, "loss": 0.4359, "step": 2500 }, { "epoch": 49.504950495049506, "eval_loss": 0.40056198835372925, "eval_runtime": 7.9369, "eval_samples_per_second": 22.679, "eval_steps_per_second": 2.898, "step": 2500 }, { "epoch": 50.495049504950494, "grad_norm": 1.1603953838348389, "learning_rate": 9.085e-06, "loss": 0.4316, "step": 2550 }, { "epoch": 51.48514851485149, "grad_norm": 1.2483686208724976, "learning_rate": 9.001666666666667e-06, "loss": 0.4264, "step": 2600 }, { "epoch": 52.475247524752476, "grad_norm": 1.3609150648117065, "learning_rate": 8.918333333333334e-06, "loss": 0.4375, "step": 2650 }, { "epoch": 53.46534653465346, "grad_norm": 1.287948727607727, "learning_rate": 8.835000000000001e-06, "loss": 0.4296, "step": 2700 }, { "epoch": 54.45544554455446, "grad_norm": 1.1371320486068726, "learning_rate": 8.751666666666668e-06, "loss": 0.4293, "step": 2750 }, { "epoch": 55.445544554455445, "grad_norm": 1.2330107688903809, "learning_rate": 8.668333333333335e-06, "loss": 0.4211, "step": 2800 }, { "epoch": 56.43564356435643, "grad_norm": 1.494728446006775, "learning_rate": 8.585000000000002e-06, "loss": 0.4265, "step": 2850 }, { "epoch": 57.42574257425743, "grad_norm": 1.195456862449646, "learning_rate": 8.501666666666667e-06, "loss": 0.4255, "step": 2900 }, { "epoch": 58.415841584158414, "grad_norm": 2.2788732051849365, "learning_rate": 8.418333333333334e-06, "loss": 0.4262, "step": 2950 }, { "epoch": 59.40594059405941, "grad_norm": 1.3329004049301147, "learning_rate": 8.335e-06, "loss": 0.4236, "step": 3000 }, { "epoch": 59.40594059405941, "eval_loss": 0.39749667048454285, "eval_runtime": 7.652, "eval_samples_per_second": 23.523, "eval_steps_per_second": 3.006, "step": 3000 }, { "epoch": 60.396039603960396, "grad_norm": 1.0736905336380005, "learning_rate": 8.251666666666668e-06, "loss": 0.4324, "step": 3050 }, { "epoch": 61.386138613861384, "grad_norm": 1.4117296934127808, "learning_rate": 8.168333333333334e-06, "loss": 0.421, "step": 3100 }, { "epoch": 62.37623762376238, "grad_norm": 1.2299913167953491, "learning_rate": 8.085000000000001e-06, "loss": 0.426, "step": 3150 }, { "epoch": 63.366336633663366, "grad_norm": 1.780414342880249, "learning_rate": 8.001666666666668e-06, "loss": 0.4217, "step": 3200 }, { "epoch": 64.35643564356435, "grad_norm": 0.953036367893219, "learning_rate": 7.918333333333333e-06, "loss": 0.4275, "step": 3250 }, { "epoch": 65.34653465346534, "grad_norm": 1.4100078344345093, "learning_rate": 7.835e-06, "loss": 0.4207, "step": 3300 }, { "epoch": 66.33663366336634, "grad_norm": 1.2050296068191528, "learning_rate": 7.751666666666667e-06, "loss": 0.4194, "step": 3350 }, { "epoch": 67.32673267326733, "grad_norm": 1.2712194919586182, "learning_rate": 7.668333333333334e-06, "loss": 0.4185, "step": 3400 }, { "epoch": 68.31683168316832, "grad_norm": 1.4407145977020264, "learning_rate": 7.585e-06, "loss": 0.42, "step": 3450 }, { "epoch": 69.3069306930693, "grad_norm": 0.9915832281112671, "learning_rate": 7.501666666666667e-06, "loss": 0.4196, "step": 3500 }, { "epoch": 69.3069306930693, "eval_loss": 0.3955594301223755, "eval_runtime": 8.0095, "eval_samples_per_second": 22.473, "eval_steps_per_second": 2.872, "step": 3500 }, { "epoch": 70.29702970297029, "grad_norm": 1.399967074394226, "learning_rate": 7.418333333333334e-06, "loss": 0.413, "step": 3550 }, { "epoch": 71.2871287128713, "grad_norm": 1.4354674816131592, "learning_rate": 7.335000000000001e-06, "loss": 0.4166, "step": 3600 }, { "epoch": 72.27722772277228, "grad_norm": 1.5903120040893555, "learning_rate": 7.251666666666667e-06, "loss": 0.4206, "step": 3650 }, { "epoch": 73.26732673267327, "grad_norm": 1.184046983718872, "learning_rate": 7.168333333333334e-06, "loss": 0.4179, "step": 3700 }, { "epoch": 74.25742574257426, "grad_norm": 1.1453702449798584, "learning_rate": 7.085000000000001e-06, "loss": 0.4145, "step": 3750 }, { "epoch": 75.24752475247524, "grad_norm": 2.3300318717956543, "learning_rate": 7.001666666666668e-06, "loss": 0.4255, "step": 3800 }, { "epoch": 76.23762376237623, "grad_norm": 1.671020746231079, "learning_rate": 6.918333333333334e-06, "loss": 0.4213, "step": 3850 }, { "epoch": 77.22772277227723, "grad_norm": 1.1744346618652344, "learning_rate": 6.835000000000001e-06, "loss": 0.4181, "step": 3900 }, { "epoch": 78.21782178217822, "grad_norm": 1.3221676349639893, "learning_rate": 6.7516666666666675e-06, "loss": 0.4163, "step": 3950 }, { "epoch": 79.20792079207921, "grad_norm": 1.0609053373336792, "learning_rate": 6.668333333333334e-06, "loss": 0.4183, "step": 4000 }, { "epoch": 79.20792079207921, "eval_loss": 0.39377516508102417, "eval_runtime": 7.9303, "eval_samples_per_second": 22.698, "eval_steps_per_second": 2.9, "step": 4000 }, { "epoch": 80.1980198019802, "grad_norm": 1.0227816104888916, "learning_rate": 6.5850000000000005e-06, "loss": 0.4164, "step": 4050 }, { "epoch": 81.18811881188118, "grad_norm": 1.0951721668243408, "learning_rate": 6.501666666666667e-06, "loss": 0.4093, "step": 4100 }, { "epoch": 82.17821782178218, "grad_norm": 1.0732530355453491, "learning_rate": 6.418333333333334e-06, "loss": 0.4161, "step": 4150 }, { "epoch": 83.16831683168317, "grad_norm": 1.13483464717865, "learning_rate": 6.336666666666667e-06, "loss": 0.4113, "step": 4200 }, { "epoch": 84.15841584158416, "grad_norm": 0.9911147356033325, "learning_rate": 6.253333333333333e-06, "loss": 0.4123, "step": 4250 }, { "epoch": 85.14851485148515, "grad_norm": 1.4925992488861084, "learning_rate": 6.17e-06, "loss": 0.4122, "step": 4300 }, { "epoch": 86.13861386138613, "grad_norm": 1.4414916038513184, "learning_rate": 6.086666666666667e-06, "loss": 0.4103, "step": 4350 }, { "epoch": 87.12871287128714, "grad_norm": 1.1858835220336914, "learning_rate": 6.003333333333334e-06, "loss": 0.4137, "step": 4400 }, { "epoch": 88.11881188118812, "grad_norm": 1.104232907295227, "learning_rate": 5.92e-06, "loss": 0.4096, "step": 4450 }, { "epoch": 89.10891089108911, "grad_norm": 1.7566862106323242, "learning_rate": 5.836666666666667e-06, "loss": 0.4148, "step": 4500 }, { "epoch": 89.10891089108911, "eval_loss": 0.3940875828266144, "eval_runtime": 7.934, "eval_samples_per_second": 22.687, "eval_steps_per_second": 2.899, "step": 4500 }, { "epoch": 90.0990099009901, "grad_norm": 1.0610746145248413, "learning_rate": 5.753333333333334e-06, "loss": 0.414, "step": 4550 }, { "epoch": 91.08910891089108, "grad_norm": 1.1602790355682373, "learning_rate": 5.67e-06, "loss": 0.415, "step": 4600 }, { "epoch": 92.07920792079207, "grad_norm": 1.1488662958145142, "learning_rate": 5.586666666666667e-06, "loss": 0.4162, "step": 4650 }, { "epoch": 93.06930693069307, "grad_norm": 1.279823899269104, "learning_rate": 5.503333333333334e-06, "loss": 0.4086, "step": 4700 }, { "epoch": 94.05940594059406, "grad_norm": 1.4046456813812256, "learning_rate": 5.420000000000001e-06, "loss": 0.4094, "step": 4750 }, { "epoch": 95.04950495049505, "grad_norm": 1.0302339792251587, "learning_rate": 5.336666666666667e-06, "loss": 0.4126, "step": 4800 }, { "epoch": 96.03960396039604, "grad_norm": 1.2423168420791626, "learning_rate": 5.2533333333333336e-06, "loss": 0.4115, "step": 4850 }, { "epoch": 97.02970297029702, "grad_norm": 1.0046170949935913, "learning_rate": 5.1700000000000005e-06, "loss": 0.4027, "step": 4900 }, { "epoch": 98.01980198019803, "grad_norm": 1.3599224090576172, "learning_rate": 5.086666666666667e-06, "loss": 0.4124, "step": 4950 }, { "epoch": 99.00990099009901, "grad_norm": 1.408510446548462, "learning_rate": 5.0033333333333334e-06, "loss": 0.4034, "step": 5000 }, { "epoch": 99.00990099009901, "eval_loss": 0.3930128216743469, "eval_runtime": 7.3645, "eval_samples_per_second": 24.442, "eval_steps_per_second": 3.123, "step": 5000 }, { "epoch": 100.0, "grad_norm": 1.2026501893997192, "learning_rate": 4.92e-06, "loss": 0.4097, "step": 5050 }, { "epoch": 100.99009900990099, "grad_norm": 1.1910372972488403, "learning_rate": 4.836666666666667e-06, "loss": 0.4126, "step": 5100 }, { "epoch": 101.98019801980197, "grad_norm": 1.084506630897522, "learning_rate": 4.753333333333333e-06, "loss": 0.4075, "step": 5150 }, { "epoch": 102.97029702970298, "grad_norm": 1.115693211555481, "learning_rate": 4.670000000000001e-06, "loss": 0.4069, "step": 5200 }, { "epoch": 103.96039603960396, "grad_norm": 1.1751248836517334, "learning_rate": 4.586666666666667e-06, "loss": 0.4098, "step": 5250 }, { "epoch": 104.95049504950495, "grad_norm": 1.0256706476211548, "learning_rate": 4.503333333333333e-06, "loss": 0.4044, "step": 5300 }, { "epoch": 105.94059405940594, "grad_norm": 1.200976848602295, "learning_rate": 4.42e-06, "loss": 0.4064, "step": 5350 }, { "epoch": 106.93069306930693, "grad_norm": 1.0933268070220947, "learning_rate": 4.336666666666667e-06, "loss": 0.4073, "step": 5400 }, { "epoch": 107.92079207920793, "grad_norm": 1.3021538257598877, "learning_rate": 4.253333333333334e-06, "loss": 0.4118, "step": 5450 }, { "epoch": 108.91089108910892, "grad_norm": 1.7753965854644775, "learning_rate": 4.17e-06, "loss": 0.4137, "step": 5500 }, { "epoch": 108.91089108910892, "eval_loss": 0.3955301344394684, "eval_runtime": 7.5702, "eval_samples_per_second": 23.777, "eval_steps_per_second": 3.038, "step": 5500 }, { "epoch": 109.9009900990099, "grad_norm": 1.4470231533050537, "learning_rate": 4.086666666666667e-06, "loss": 0.4053, "step": 5550 }, { "epoch": 110.89108910891089, "grad_norm": 1.7531636953353882, "learning_rate": 4.003333333333334e-06, "loss": 0.4061, "step": 5600 }, { "epoch": 111.88118811881188, "grad_norm": 0.9751603603363037, "learning_rate": 3.920000000000001e-06, "loss": 0.4086, "step": 5650 }, { "epoch": 112.87128712871286, "grad_norm": 1.1000280380249023, "learning_rate": 3.836666666666667e-06, "loss": 0.4031, "step": 5700 }, { "epoch": 113.86138613861387, "grad_norm": 1.2386143207550049, "learning_rate": 3.753333333333334e-06, "loss": 0.4059, "step": 5750 }, { "epoch": 114.85148514851485, "grad_norm": 1.4081382751464844, "learning_rate": 3.6700000000000004e-06, "loss": 0.4084, "step": 5800 }, { "epoch": 115.84158415841584, "grad_norm": 1.3724069595336914, "learning_rate": 3.5866666666666673e-06, "loss": 0.4046, "step": 5850 }, { "epoch": 116.83168316831683, "grad_norm": 1.188818097114563, "learning_rate": 3.5033333333333334e-06, "loss": 0.4109, "step": 5900 }, { "epoch": 117.82178217821782, "grad_norm": 1.3779815435409546, "learning_rate": 3.4200000000000007e-06, "loss": 0.4042, "step": 5950 }, { "epoch": 118.81188118811882, "grad_norm": 1.3800166845321655, "learning_rate": 3.3366666666666668e-06, "loss": 0.4094, "step": 6000 }, { "epoch": 118.81188118811882, "eval_loss": 0.39243730902671814, "eval_runtime": 7.911, "eval_samples_per_second": 22.753, "eval_steps_per_second": 2.907, "step": 6000 }, { "epoch": 119.8019801980198, "grad_norm": 1.2765828371047974, "learning_rate": 3.2533333333333332e-06, "loss": 0.4022, "step": 6050 }, { "epoch": 120.79207920792079, "grad_norm": 1.3544330596923828, "learning_rate": 3.17e-06, "loss": 0.4062, "step": 6100 }, { "epoch": 121.78217821782178, "grad_norm": 1.9719560146331787, "learning_rate": 3.0866666666666666e-06, "loss": 0.4002, "step": 6150 }, { "epoch": 122.77227722772277, "grad_norm": 1.2542638778686523, "learning_rate": 3.0033333333333335e-06, "loss": 0.4037, "step": 6200 }, { "epoch": 123.76237623762377, "grad_norm": 0.9950863122940063, "learning_rate": 2.92e-06, "loss": 0.4034, "step": 6250 }, { "epoch": 124.75247524752476, "grad_norm": 1.2220611572265625, "learning_rate": 2.836666666666667e-06, "loss": 0.405, "step": 6300 }, { "epoch": 125.74257425742574, "grad_norm": 1.3050748109817505, "learning_rate": 2.7533333333333334e-06, "loss": 0.4019, "step": 6350 }, { "epoch": 126.73267326732673, "grad_norm": 1.2218639850616455, "learning_rate": 2.6700000000000003e-06, "loss": 0.4077, "step": 6400 }, { "epoch": 127.72277227722772, "grad_norm": 1.0454591512680054, "learning_rate": 2.5866666666666667e-06, "loss": 0.4009, "step": 6450 }, { "epoch": 128.7128712871287, "grad_norm": 1.211366057395935, "learning_rate": 2.505e-06, "loss": 0.4112, "step": 6500 }, { "epoch": 128.7128712871287, "eval_loss": 0.3917270302772522, "eval_runtime": 7.8896, "eval_samples_per_second": 22.815, "eval_steps_per_second": 2.915, "step": 6500 }, { "epoch": 129.7029702970297, "grad_norm": 1.1961348056793213, "learning_rate": 2.421666666666667e-06, "loss": 0.4009, "step": 6550 }, { "epoch": 130.69306930693068, "grad_norm": 1.6463865041732788, "learning_rate": 2.3383333333333335e-06, "loss": 0.403, "step": 6600 }, { "epoch": 131.68316831683168, "grad_norm": 1.1030319929122925, "learning_rate": 2.2550000000000004e-06, "loss": 0.4009, "step": 6650 }, { "epoch": 132.67326732673268, "grad_norm": 1.0525782108306885, "learning_rate": 2.171666666666667e-06, "loss": 0.4038, "step": 6700 }, { "epoch": 133.66336633663366, "grad_norm": 1.0361270904541016, "learning_rate": 2.088333333333334e-06, "loss": 0.4046, "step": 6750 }, { "epoch": 134.65346534653466, "grad_norm": 1.3003863096237183, "learning_rate": 2.0050000000000003e-06, "loss": 0.404, "step": 6800 }, { "epoch": 135.64356435643563, "grad_norm": 1.6495325565338135, "learning_rate": 1.9216666666666668e-06, "loss": 0.401, "step": 6850 }, { "epoch": 136.63366336633663, "grad_norm": 1.070357084274292, "learning_rate": 1.8383333333333334e-06, "loss": 0.4033, "step": 6900 }, { "epoch": 137.62376237623764, "grad_norm": 0.9179391860961914, "learning_rate": 1.7550000000000001e-06, "loss": 0.4018, "step": 6950 }, { "epoch": 138.6138613861386, "grad_norm": 1.398165225982666, "learning_rate": 1.6716666666666666e-06, "loss": 0.4041, "step": 7000 }, { "epoch": 138.6138613861386, "eval_loss": 0.39232540130615234, "eval_runtime": 7.5562, "eval_samples_per_second": 23.822, "eval_steps_per_second": 3.044, "step": 7000 }, { "epoch": 139.6039603960396, "grad_norm": 1.1762559413909912, "learning_rate": 1.5900000000000002e-06, "loss": 0.4117, "step": 7050 }, { "epoch": 140.59405940594058, "grad_norm": 1.515325665473938, "learning_rate": 1.506666666666667e-06, "loss": 0.4014, "step": 7100 }, { "epoch": 141.58415841584159, "grad_norm": 1.0724811553955078, "learning_rate": 1.4233333333333336e-06, "loss": 0.4028, "step": 7150 }, { "epoch": 142.5742574257426, "grad_norm": 1.0034635066986084, "learning_rate": 1.34e-06, "loss": 0.3977, "step": 7200 }, { "epoch": 143.56435643564356, "grad_norm": 1.1974977254867554, "learning_rate": 1.2566666666666668e-06, "loss": 0.4013, "step": 7250 }, { "epoch": 144.55445544554456, "grad_norm": 1.167351484298706, "learning_rate": 1.1733333333333335e-06, "loss": 0.3992, "step": 7300 }, { "epoch": 145.54455445544554, "grad_norm": 1.1566020250320435, "learning_rate": 1.0900000000000002e-06, "loss": 0.4, "step": 7350 }, { "epoch": 146.53465346534654, "grad_norm": 0.8720223307609558, "learning_rate": 1.0066666666666668e-06, "loss": 0.4006, "step": 7400 }, { "epoch": 147.52475247524754, "grad_norm": 1.1982648372650146, "learning_rate": 9.233333333333334e-07, "loss": 0.3999, "step": 7450 }, { "epoch": 148.5148514851485, "grad_norm": 1.1742349863052368, "learning_rate": 8.400000000000001e-07, "loss": 0.3989, "step": 7500 }, { "epoch": 148.5148514851485, "eval_loss": 0.3927270770072937, "eval_runtime": 8.9179, "eval_samples_per_second": 20.184, "eval_steps_per_second": 2.579, "step": 7500 }, { "epoch": 149.5049504950495, "grad_norm": 1.1068042516708374, "learning_rate": 7.566666666666667e-07, "loss": 0.4029, "step": 7550 }, { "epoch": 150.4950495049505, "grad_norm": 1.2785730361938477, "learning_rate": 6.733333333333334e-07, "loss": 0.4001, "step": 7600 }, { "epoch": 151.4851485148515, "grad_norm": 1.0063281059265137, "learning_rate": 5.900000000000001e-07, "loss": 0.3977, "step": 7650 }, { "epoch": 152.47524752475246, "grad_norm": 1.1545789241790771, "learning_rate": 5.066666666666667e-07, "loss": 0.4015, "step": 7700 }, { "epoch": 153.46534653465346, "grad_norm": 1.0624966621398926, "learning_rate": 4.233333333333334e-07, "loss": 0.4044, "step": 7750 }, { "epoch": 154.45544554455446, "grad_norm": 1.3898247480392456, "learning_rate": 3.4000000000000003e-07, "loss": 0.4005, "step": 7800 }, { "epoch": 155.44554455445544, "grad_norm": 1.2618342638015747, "learning_rate": 2.566666666666667e-07, "loss": 0.4025, "step": 7850 }, { "epoch": 156.43564356435644, "grad_norm": 1.0798155069351196, "learning_rate": 1.7333333333333335e-07, "loss": 0.4015, "step": 7900 }, { "epoch": 157.4257425742574, "grad_norm": 1.0550168752670288, "learning_rate": 9e-08, "loss": 0.3952, "step": 7950 }, { "epoch": 158.41584158415841, "grad_norm": 1.0714497566223145, "learning_rate": 6.666666666666667e-09, "loss": 0.3989, "step": 8000 }, { "epoch": 158.41584158415841, "eval_loss": 0.39282524585723877, "eval_runtime": 8.2483, "eval_samples_per_second": 21.823, "eval_steps_per_second": 2.788, "step": 8000 } ], "logging_steps": 50, "max_steps": 8000, "num_input_tokens_seen": 0, "num_train_epochs": 160, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.322114324975938e+16, "train_batch_size": 16, "trial_name": null, "trial_params": null }