{ "best_metric": 0.40008437633514404, "best_model_checkpoint": "mikhail-panzo/zlm-ceb_b64_le5_s8000/checkpoint-6000", "epoch": 313.72549019607845, "eval_steps": 500, "global_step": 8000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 1.9607843137254903, "grad_norm": 1.6111680269241333, "learning_rate": 2.5e-06, "loss": 0.821, "step": 50 }, { "epoch": 3.9215686274509802, "grad_norm": 1.9582017660140991, "learning_rate": 5e-06, "loss": 0.673, "step": 100 }, { "epoch": 5.882352941176471, "grad_norm": 0.9988609552383423, "learning_rate": 7.5e-06, "loss": 0.5518, "step": 150 }, { "epoch": 7.8431372549019605, "grad_norm": 1.0111173391342163, "learning_rate": 1e-05, "loss": 0.5172, "step": 200 }, { "epoch": 9.803921568627452, "grad_norm": 0.9361169338226318, "learning_rate": 1.25e-05, "loss": 0.4999, "step": 250 }, { "epoch": 11.764705882352942, "grad_norm": 0.8657887578010559, "learning_rate": 1.5e-05, "loss": 0.4911, "step": 300 }, { "epoch": 13.72549019607843, "grad_norm": 0.8616572618484497, "learning_rate": 1.75e-05, "loss": 0.4757, "step": 350 }, { "epoch": 15.686274509803921, "grad_norm": 1.100374698638916, "learning_rate": 2e-05, "loss": 0.47, "step": 400 }, { "epoch": 17.647058823529413, "grad_norm": 1.3860764503479004, "learning_rate": 2.25e-05, "loss": 0.4671, "step": 450 }, { "epoch": 19.607843137254903, "grad_norm": 1.2481627464294434, "learning_rate": 2.5e-05, "loss": 0.4626, "step": 500 }, { "epoch": 19.607843137254903, "eval_loss": 0.42633864283561707, "eval_runtime": 6.9188, "eval_samples_per_second": 26.016, "eval_steps_per_second": 3.324, "step": 500 }, { "epoch": 21.568627450980394, "grad_norm": 1.079897165298462, "learning_rate": 2.7500000000000004e-05, "loss": 0.4593, "step": 550 }, { "epoch": 23.529411764705884, "grad_norm": 0.9019867777824402, "learning_rate": 3e-05, "loss": 0.4474, "step": 600 }, { "epoch": 25.49019607843137, "grad_norm": 1.900229811668396, "learning_rate": 3.2500000000000004e-05, "loss": 0.4439, "step": 650 }, { "epoch": 27.45098039215686, "grad_norm": 1.3894062042236328, "learning_rate": 3.5e-05, "loss": 0.4397, "step": 700 }, { "epoch": 29.41176470588235, "grad_norm": 2.318305253982544, "learning_rate": 3.7500000000000003e-05, "loss": 0.4376, "step": 750 }, { "epoch": 31.372549019607842, "grad_norm": 0.9606600999832153, "learning_rate": 4e-05, "loss": 0.4394, "step": 800 }, { "epoch": 33.333333333333336, "grad_norm": 1.4766446352005005, "learning_rate": 4.25e-05, "loss": 0.4376, "step": 850 }, { "epoch": 35.294117647058826, "grad_norm": 1.1099992990493774, "learning_rate": 4.5e-05, "loss": 0.4333, "step": 900 }, { "epoch": 37.254901960784316, "grad_norm": 1.1501879692077637, "learning_rate": 4.75e-05, "loss": 0.4281, "step": 950 }, { "epoch": 39.21568627450981, "grad_norm": 0.9793707728385925, "learning_rate": 5e-05, "loss": 0.4288, "step": 1000 }, { "epoch": 39.21568627450981, "eval_loss": 0.4076782763004303, "eval_runtime": 6.5279, "eval_samples_per_second": 27.574, "eval_steps_per_second": 3.523, "step": 1000 }, { "epoch": 41.1764705882353, "grad_norm": 1.812929630279541, "learning_rate": 5.25e-05, "loss": 0.4224, "step": 1050 }, { "epoch": 43.13725490196079, "grad_norm": 1.066340446472168, "learning_rate": 5.500000000000001e-05, "loss": 0.4231, "step": 1100 }, { "epoch": 45.09803921568628, "grad_norm": 2.465958595275879, "learning_rate": 5.745e-05, "loss": 0.4261, "step": 1150 }, { "epoch": 47.05882352941177, "grad_norm": 1.3723503351211548, "learning_rate": 5.995000000000001e-05, "loss": 0.416, "step": 1200 }, { "epoch": 49.01960784313726, "grad_norm": 1.6065536737442017, "learning_rate": 6.245000000000001e-05, "loss": 0.416, "step": 1250 }, { "epoch": 50.98039215686274, "grad_norm": 3.8591885566711426, "learning_rate": 6.494999999999999e-05, "loss": 0.4162, "step": 1300 }, { "epoch": 52.94117647058823, "grad_norm": 2.415721893310547, "learning_rate": 6.745e-05, "loss": 0.4121, "step": 1350 }, { "epoch": 54.90196078431372, "grad_norm": 1.969329833984375, "learning_rate": 6.995e-05, "loss": 0.4104, "step": 1400 }, { "epoch": 56.86274509803921, "grad_norm": 3.190781593322754, "learning_rate": 7.245000000000001e-05, "loss": 0.4055, "step": 1450 }, { "epoch": 58.8235294117647, "grad_norm": 1.4585760831832886, "learning_rate": 7.495e-05, "loss": 0.4109, "step": 1500 }, { "epoch": 58.8235294117647, "eval_loss": 0.4012959897518158, "eval_runtime": 6.3508, "eval_samples_per_second": 28.343, "eval_steps_per_second": 3.622, "step": 1500 }, { "epoch": 60.78431372549019, "grad_norm": 4.141477584838867, "learning_rate": 7.745e-05, "loss": 0.4117, "step": 1550 }, { "epoch": 62.745098039215684, "grad_norm": 5.334190845489502, "learning_rate": 7.995e-05, "loss": 0.413, "step": 1600 }, { "epoch": 64.70588235294117, "grad_norm": 3.1039328575134277, "learning_rate": 8.245e-05, "loss": 0.405, "step": 1650 }, { "epoch": 66.66666666666667, "grad_norm": 2.6190948486328125, "learning_rate": 8.495e-05, "loss": 0.4098, "step": 1700 }, { "epoch": 68.62745098039215, "grad_norm": 1.8761863708496094, "learning_rate": 8.745000000000001e-05, "loss": 0.4035, "step": 1750 }, { "epoch": 70.58823529411765, "grad_norm": 0.9362647533416748, "learning_rate": 8.995e-05, "loss": 0.4253, "step": 1800 }, { "epoch": 72.54901960784314, "grad_norm": 1.6635262966156006, "learning_rate": 9.245e-05, "loss": 0.4295, "step": 1850 }, { "epoch": 74.50980392156863, "grad_norm": 1.3353509902954102, "learning_rate": 9.495e-05, "loss": 0.4078, "step": 1900 }, { "epoch": 76.47058823529412, "grad_norm": 8.328239440917969, "learning_rate": 9.745000000000001e-05, "loss": 0.4051, "step": 1950 }, { "epoch": 78.43137254901961, "grad_norm": 1.1230006217956543, "learning_rate": 9.995e-05, "loss": 0.3978, "step": 2000 }, { "epoch": 78.43137254901961, "eval_loss": 0.40350666642189026, "eval_runtime": 6.5593, "eval_samples_per_second": 27.442, "eval_steps_per_second": 3.506, "step": 2000 }, { "epoch": 80.3921568627451, "grad_norm": 3.8058626651763916, "learning_rate": 9.918333333333334e-05, "loss": 0.3997, "step": 2050 }, { "epoch": 82.3529411764706, "grad_norm": 2.170177936553955, "learning_rate": 9.835e-05, "loss": 0.3988, "step": 2100 }, { "epoch": 84.31372549019608, "grad_norm": 0.6957865357398987, "learning_rate": 9.751666666666666e-05, "loss": 0.3938, "step": 2150 }, { "epoch": 86.27450980392157, "grad_norm": 1.1778301000595093, "learning_rate": 9.668333333333334e-05, "loss": 0.4008, "step": 2200 }, { "epoch": 88.23529411764706, "grad_norm": 1.1749225854873657, "learning_rate": 9.585000000000001e-05, "loss": 0.406, "step": 2250 }, { "epoch": 90.19607843137256, "grad_norm": 1.1235101222991943, "learning_rate": 9.501666666666668e-05, "loss": 0.3906, "step": 2300 }, { "epoch": 92.15686274509804, "grad_norm": 5.676591873168945, "learning_rate": 9.418333333333334e-05, "loss": 0.3954, "step": 2350 }, { "epoch": 94.11764705882354, "grad_norm": 1.4196738004684448, "learning_rate": 9.335e-05, "loss": 0.4018, "step": 2400 }, { "epoch": 96.07843137254902, "grad_norm": 2.848329544067383, "learning_rate": 9.251666666666667e-05, "loss": 0.387, "step": 2450 }, { "epoch": 98.03921568627452, "grad_norm": 1.7598848342895508, "learning_rate": 9.168333333333333e-05, "loss": 0.3898, "step": 2500 }, { "epoch": 98.03921568627452, "eval_loss": 0.40130773186683655, "eval_runtime": 6.4746, "eval_samples_per_second": 27.801, "eval_steps_per_second": 3.552, "step": 2500 }, { "epoch": 100.0, "grad_norm": 5.990013599395752, "learning_rate": 9.085e-05, "loss": 0.3982, "step": 2550 }, { "epoch": 101.96078431372548, "grad_norm": 2.984307050704956, "learning_rate": 9.001666666666667e-05, "loss": 0.3889, "step": 2600 }, { "epoch": 103.92156862745098, "grad_norm": 1.976651668548584, "learning_rate": 8.918333333333334e-05, "loss": 0.382, "step": 2650 }, { "epoch": 105.88235294117646, "grad_norm": 1.1698123216629028, "learning_rate": 8.834999999999999e-05, "loss": 0.3877, "step": 2700 }, { "epoch": 107.84313725490196, "grad_norm": 1.8311396837234497, "learning_rate": 8.751666666666668e-05, "loss": 0.3848, "step": 2750 }, { "epoch": 109.80392156862744, "grad_norm": 3.1962406635284424, "learning_rate": 8.668333333333334e-05, "loss": 0.384, "step": 2800 }, { "epoch": 111.76470588235294, "grad_norm": 1.4986674785614014, "learning_rate": 8.585000000000001e-05, "loss": 0.3737, "step": 2850 }, { "epoch": 113.72549019607843, "grad_norm": 4.846043109893799, "learning_rate": 8.501666666666667e-05, "loss": 0.3787, "step": 2900 }, { "epoch": 115.68627450980392, "grad_norm": 2.5160129070281982, "learning_rate": 8.418333333333334e-05, "loss": 0.3707, "step": 2950 }, { "epoch": 117.6470588235294, "grad_norm": 1.209688663482666, "learning_rate": 8.335e-05, "loss": 0.373, "step": 3000 }, { "epoch": 117.6470588235294, "eval_loss": 0.40103283524513245, "eval_runtime": 6.6876, "eval_samples_per_second": 26.916, "eval_steps_per_second": 3.439, "step": 3000 }, { "epoch": 119.6078431372549, "grad_norm": 1.4071707725524902, "learning_rate": 8.251666666666668e-05, "loss": 0.3731, "step": 3050 }, { "epoch": 121.56862745098039, "grad_norm": 2.3906798362731934, "learning_rate": 8.168333333333333e-05, "loss": 0.3705, "step": 3100 }, { "epoch": 123.52941176470588, "grad_norm": 2.833566188812256, "learning_rate": 8.085e-05, "loss": 0.3754, "step": 3150 }, { "epoch": 125.49019607843137, "grad_norm": 1.7871581315994263, "learning_rate": 8.001666666666667e-05, "loss": 0.3734, "step": 3200 }, { "epoch": 127.45098039215686, "grad_norm": 1.0529006719589233, "learning_rate": 7.918333333333334e-05, "loss": 0.3792, "step": 3250 }, { "epoch": 129.41176470588235, "grad_norm": 2.939656972885132, "learning_rate": 7.835000000000001e-05, "loss": 0.3693, "step": 3300 }, { "epoch": 131.37254901960785, "grad_norm": 0.9791694283485413, "learning_rate": 7.751666666666668e-05, "loss": 0.366, "step": 3350 }, { "epoch": 133.33333333333334, "grad_norm": 1.6775403022766113, "learning_rate": 7.668333333333335e-05, "loss": 0.3709, "step": 3400 }, { "epoch": 135.2941176470588, "grad_norm": 1.0679121017456055, "learning_rate": 7.585e-05, "loss": 0.3677, "step": 3450 }, { "epoch": 137.2549019607843, "grad_norm": 2.068532705307007, "learning_rate": 7.501666666666667e-05, "loss": 0.3644, "step": 3500 }, { "epoch": 137.2549019607843, "eval_loss": 0.4005117118358612, "eval_runtime": 6.7669, "eval_samples_per_second": 26.6, "eval_steps_per_second": 3.399, "step": 3500 }, { "epoch": 139.2156862745098, "grad_norm": 0.9116536378860474, "learning_rate": 7.418333333333334e-05, "loss": 0.365, "step": 3550 }, { "epoch": 141.1764705882353, "grad_norm": 1.6524943113327026, "learning_rate": 7.335000000000001e-05, "loss": 0.362, "step": 3600 }, { "epoch": 143.13725490196077, "grad_norm": 0.9939906597137451, "learning_rate": 7.251666666666666e-05, "loss": 0.3641, "step": 3650 }, { "epoch": 145.09803921568627, "grad_norm": 1.7879794836044312, "learning_rate": 7.168333333333333e-05, "loss": 0.3649, "step": 3700 }, { "epoch": 147.05882352941177, "grad_norm": 2.4735560417175293, "learning_rate": 7.085e-05, "loss": 0.3591, "step": 3750 }, { "epoch": 149.01960784313727, "grad_norm": 1.4222075939178467, "learning_rate": 7.001666666666667e-05, "loss": 0.3605, "step": 3800 }, { "epoch": 150.98039215686273, "grad_norm": 2.095998525619507, "learning_rate": 6.918333333333334e-05, "loss": 0.3597, "step": 3850 }, { "epoch": 152.94117647058823, "grad_norm": 3.8910038471221924, "learning_rate": 6.835000000000001e-05, "loss": 0.3605, "step": 3900 }, { "epoch": 154.90196078431373, "grad_norm": 0.7127563953399658, "learning_rate": 6.751666666666668e-05, "loss": 0.3629, "step": 3950 }, { "epoch": 156.86274509803923, "grad_norm": 1.66582190990448, "learning_rate": 6.668333333333333e-05, "loss": 0.3569, "step": 4000 }, { "epoch": 156.86274509803923, "eval_loss": 0.4029478132724762, "eval_runtime": 6.4438, "eval_samples_per_second": 27.934, "eval_steps_per_second": 3.569, "step": 4000 }, { "epoch": 158.8235294117647, "grad_norm": 1.4073445796966553, "learning_rate": 6.585e-05, "loss": 0.3648, "step": 4050 }, { "epoch": 160.7843137254902, "grad_norm": 1.070461392402649, "learning_rate": 6.501666666666667e-05, "loss": 0.3569, "step": 4100 }, { "epoch": 162.7450980392157, "grad_norm": 2.141096353530884, "learning_rate": 6.418333333333334e-05, "loss": 0.3607, "step": 4150 }, { "epoch": 164.7058823529412, "grad_norm": 1.1931374073028564, "learning_rate": 6.335e-05, "loss": 0.3575, "step": 4200 }, { "epoch": 166.66666666666666, "grad_norm": 1.0879640579223633, "learning_rate": 6.251666666666666e-05, "loss": 0.356, "step": 4250 }, { "epoch": 168.62745098039215, "grad_norm": 1.5755019187927246, "learning_rate": 6.168333333333333e-05, "loss": 0.362, "step": 4300 }, { "epoch": 170.58823529411765, "grad_norm": 9.463594436645508, "learning_rate": 6.085000000000001e-05, "loss": 0.3612, "step": 4350 }, { "epoch": 172.54901960784315, "grad_norm": 5.174690246582031, "learning_rate": 6.0016666666666664e-05, "loss": 0.3537, "step": 4400 }, { "epoch": 174.50980392156862, "grad_norm": 1.342671513557434, "learning_rate": 5.918333333333333e-05, "loss": 0.3521, "step": 4450 }, { "epoch": 176.47058823529412, "grad_norm": 1.7866413593292236, "learning_rate": 5.835e-05, "loss": 0.3515, "step": 4500 }, { "epoch": 176.47058823529412, "eval_loss": 0.4038984179496765, "eval_runtime": 6.8698, "eval_samples_per_second": 26.202, "eval_steps_per_second": 3.348, "step": 4500 }, { "epoch": 178.4313725490196, "grad_norm": 2.6986000537872314, "learning_rate": 5.751666666666667e-05, "loss": 0.3539, "step": 4550 }, { "epoch": 180.3921568627451, "grad_norm": 1.255347728729248, "learning_rate": 5.668333333333333e-05, "loss": 0.3578, "step": 4600 }, { "epoch": 182.35294117647058, "grad_norm": 2.3221983909606934, "learning_rate": 5.585e-05, "loss": 0.3594, "step": 4650 }, { "epoch": 184.31372549019608, "grad_norm": 1.0801488161087036, "learning_rate": 5.501666666666667e-05, "loss": 0.3481, "step": 4700 }, { "epoch": 186.27450980392157, "grad_norm": 0.7307039499282837, "learning_rate": 5.4183333333333334e-05, "loss": 0.3486, "step": 4750 }, { "epoch": 188.23529411764707, "grad_norm": 0.9629709720611572, "learning_rate": 5.335e-05, "loss": 0.3465, "step": 4800 }, { "epoch": 190.19607843137254, "grad_norm": 0.5935477614402771, "learning_rate": 5.251666666666667e-05, "loss": 0.3481, "step": 4850 }, { "epoch": 192.15686274509804, "grad_norm": 1.965649127960205, "learning_rate": 5.168333333333334e-05, "loss": 0.3468, "step": 4900 }, { "epoch": 194.11764705882354, "grad_norm": 0.9675062894821167, "learning_rate": 5.0849999999999996e-05, "loss": 0.3529, "step": 4950 }, { "epoch": 196.07843137254903, "grad_norm": 0.7859313488006592, "learning_rate": 5.0016666666666665e-05, "loss": 0.3443, "step": 5000 }, { "epoch": 196.07843137254903, "eval_loss": 0.4005332589149475, "eval_runtime": 6.4088, "eval_samples_per_second": 28.086, "eval_steps_per_second": 3.589, "step": 5000 }, { "epoch": 198.0392156862745, "grad_norm": 0.9622226357460022, "learning_rate": 4.9183333333333334e-05, "loss": 0.349, "step": 5050 }, { "epoch": 200.0, "grad_norm": 0.7713702321052551, "learning_rate": 4.835e-05, "loss": 0.3446, "step": 5100 }, { "epoch": 201.9607843137255, "grad_norm": 1.2544292211532593, "learning_rate": 4.751666666666667e-05, "loss": 0.3443, "step": 5150 }, { "epoch": 203.92156862745097, "grad_norm": 0.7886931300163269, "learning_rate": 4.6683333333333334e-05, "loss": 0.3424, "step": 5200 }, { "epoch": 205.88235294117646, "grad_norm": 1.3799712657928467, "learning_rate": 4.585e-05, "loss": 0.348, "step": 5250 }, { "epoch": 207.84313725490196, "grad_norm": 0.9509637355804443, "learning_rate": 4.5016666666666665e-05, "loss": 0.3447, "step": 5300 }, { "epoch": 209.80392156862746, "grad_norm": 1.3633594512939453, "learning_rate": 4.4183333333333334e-05, "loss": 0.343, "step": 5350 }, { "epoch": 211.76470588235293, "grad_norm": 3.379939317703247, "learning_rate": 4.335e-05, "loss": 0.3451, "step": 5400 }, { "epoch": 213.72549019607843, "grad_norm": 0.7113842368125916, "learning_rate": 4.251666666666667e-05, "loss": 0.3429, "step": 5450 }, { "epoch": 215.68627450980392, "grad_norm": 0.8408060073852539, "learning_rate": 4.1683333333333335e-05, "loss": 0.3469, "step": 5500 }, { "epoch": 215.68627450980392, "eval_loss": 0.40183132886886597, "eval_runtime": 6.5032, "eval_samples_per_second": 27.679, "eval_steps_per_second": 3.537, "step": 5500 }, { "epoch": 217.64705882352942, "grad_norm": 1.2636085748672485, "learning_rate": 4.085e-05, "loss": 0.3371, "step": 5550 }, { "epoch": 219.6078431372549, "grad_norm": 2.5460524559020996, "learning_rate": 4.0016666666666666e-05, "loss": 0.3449, "step": 5600 }, { "epoch": 221.5686274509804, "grad_norm": 0.6922531723976135, "learning_rate": 3.9183333333333335e-05, "loss": 0.3432, "step": 5650 }, { "epoch": 223.52941176470588, "grad_norm": 0.5262002944946289, "learning_rate": 3.8350000000000004e-05, "loss": 0.3408, "step": 5700 }, { "epoch": 225.49019607843138, "grad_norm": 0.9883546233177185, "learning_rate": 3.7516666666666666e-05, "loss": 0.3399, "step": 5750 }, { "epoch": 227.45098039215685, "grad_norm": 0.7871970534324646, "learning_rate": 3.6683333333333335e-05, "loss": 0.3417, "step": 5800 }, { "epoch": 229.41176470588235, "grad_norm": 0.7385707497596741, "learning_rate": 3.585e-05, "loss": 0.3399, "step": 5850 }, { "epoch": 231.37254901960785, "grad_norm": 0.5304046869277954, "learning_rate": 3.501666666666667e-05, "loss": 0.3388, "step": 5900 }, { "epoch": 233.33333333333334, "grad_norm": 0.5726514458656311, "learning_rate": 3.4183333333333335e-05, "loss": 0.3361, "step": 5950 }, { "epoch": 235.2941176470588, "grad_norm": 0.49944621324539185, "learning_rate": 3.3350000000000004e-05, "loss": 0.3427, "step": 6000 }, { "epoch": 235.2941176470588, "eval_loss": 0.40008437633514404, "eval_runtime": 6.512, "eval_samples_per_second": 27.641, "eval_steps_per_second": 3.532, "step": 6000 }, { "epoch": 237.2549019607843, "grad_norm": 1.1355714797973633, "learning_rate": 3.2516666666666666e-05, "loss": 0.3377, "step": 6050 }, { "epoch": 239.2156862745098, "grad_norm": 0.7759362459182739, "learning_rate": 3.1683333333333335e-05, "loss": 0.339, "step": 6100 }, { "epoch": 241.1764705882353, "grad_norm": 1.1163146495819092, "learning_rate": 3.0850000000000004e-05, "loss": 0.337, "step": 6150 }, { "epoch": 243.13725490196077, "grad_norm": 0.8132624626159668, "learning_rate": 3.001666666666667e-05, "loss": 0.3378, "step": 6200 }, { "epoch": 245.09803921568627, "grad_norm": 0.7331820130348206, "learning_rate": 2.9183333333333336e-05, "loss": 0.3323, "step": 6250 }, { "epoch": 247.05882352941177, "grad_norm": 0.9192725419998169, "learning_rate": 2.8349999999999998e-05, "loss": 0.3382, "step": 6300 }, { "epoch": 249.01960784313727, "grad_norm": 3.9876034259796143, "learning_rate": 2.7516666666666667e-05, "loss": 0.3384, "step": 6350 }, { "epoch": 250.98039215686273, "grad_norm": 0.6646643280982971, "learning_rate": 2.6683333333333333e-05, "loss": 0.3365, "step": 6400 }, { "epoch": 252.94117647058823, "grad_norm": 0.5878006815910339, "learning_rate": 2.585e-05, "loss": 0.3403, "step": 6450 }, { "epoch": 254.90196078431373, "grad_norm": 0.8833298087120056, "learning_rate": 2.5016666666666667e-05, "loss": 0.3401, "step": 6500 }, { "epoch": 254.90196078431373, "eval_loss": 0.4041617810726166, "eval_runtime": 6.702, "eval_samples_per_second": 26.858, "eval_steps_per_second": 3.432, "step": 6500 }, { "epoch": 256.8627450980392, "grad_norm": 0.751977801322937, "learning_rate": 2.4183333333333336e-05, "loss": 0.3319, "step": 6550 }, { "epoch": 258.8235294117647, "grad_norm": 1.289437174797058, "learning_rate": 2.3350000000000002e-05, "loss": 0.3382, "step": 6600 }, { "epoch": 260.7843137254902, "grad_norm": 0.5142917633056641, "learning_rate": 2.2516666666666667e-05, "loss": 0.334, "step": 6650 }, { "epoch": 262.7450980392157, "grad_norm": 0.4598856568336487, "learning_rate": 2.1683333333333333e-05, "loss": 0.338, "step": 6700 }, { "epoch": 264.70588235294116, "grad_norm": 0.47424769401550293, "learning_rate": 2.085e-05, "loss": 0.3369, "step": 6750 }, { "epoch": 266.6666666666667, "grad_norm": 0.6651924848556519, "learning_rate": 2.0016666666666668e-05, "loss": 0.3379, "step": 6800 }, { "epoch": 268.62745098039215, "grad_norm": 0.6406331062316895, "learning_rate": 1.9183333333333333e-05, "loss": 0.3409, "step": 6850 }, { "epoch": 270.5882352941176, "grad_norm": 0.8162440061569214, "learning_rate": 1.8350000000000002e-05, "loss": 0.3336, "step": 6900 }, { "epoch": 272.54901960784315, "grad_norm": 0.5905545353889465, "learning_rate": 1.7516666666666668e-05, "loss": 0.3324, "step": 6950 }, { "epoch": 274.5098039215686, "grad_norm": 0.7422587871551514, "learning_rate": 1.6700000000000003e-05, "loss": 0.3419, "step": 7000 }, { "epoch": 274.5098039215686, "eval_loss": 0.40541204810142517, "eval_runtime": 6.8059, "eval_samples_per_second": 26.447, "eval_steps_per_second": 3.379, "step": 7000 }, { "epoch": 276.47058823529414, "grad_norm": 0.5827329158782959, "learning_rate": 1.586666666666667e-05, "loss": 0.3372, "step": 7050 }, { "epoch": 278.4313725490196, "grad_norm": 0.6354573369026184, "learning_rate": 1.5033333333333336e-05, "loss": 0.3316, "step": 7100 }, { "epoch": 280.3921568627451, "grad_norm": 0.47346118092536926, "learning_rate": 1.42e-05, "loss": 0.335, "step": 7150 }, { "epoch": 282.3529411764706, "grad_norm": 0.6653403043746948, "learning_rate": 1.3366666666666667e-05, "loss": 0.333, "step": 7200 }, { "epoch": 284.3137254901961, "grad_norm": 0.44320473074913025, "learning_rate": 1.2533333333333332e-05, "loss": 0.3322, "step": 7250 }, { "epoch": 286.27450980392155, "grad_norm": 0.40339425206184387, "learning_rate": 1.1700000000000001e-05, "loss": 0.3286, "step": 7300 }, { "epoch": 288.2352941176471, "grad_norm": 0.6550925374031067, "learning_rate": 1.0866666666666667e-05, "loss": 0.3351, "step": 7350 }, { "epoch": 290.19607843137254, "grad_norm": 0.5210171341896057, "learning_rate": 1.0033333333333333e-05, "loss": 0.3309, "step": 7400 }, { "epoch": 292.15686274509807, "grad_norm": 0.4037918448448181, "learning_rate": 9.2e-06, "loss": 0.3291, "step": 7450 }, { "epoch": 294.11764705882354, "grad_norm": 0.7901822328567505, "learning_rate": 8.366666666666667e-06, "loss": 0.3318, "step": 7500 }, { "epoch": 294.11764705882354, "eval_loss": 0.405687540769577, "eval_runtime": 6.5967, "eval_samples_per_second": 27.286, "eval_steps_per_second": 3.487, "step": 7500 }, { "epoch": 296.078431372549, "grad_norm": 0.6064783930778503, "learning_rate": 7.533333333333334e-06, "loss": 0.3314, "step": 7550 }, { "epoch": 298.03921568627453, "grad_norm": 0.5984176397323608, "learning_rate": 6.700000000000001e-06, "loss": 0.3312, "step": 7600 }, { "epoch": 300.0, "grad_norm": 0.6286676526069641, "learning_rate": 5.866666666666667e-06, "loss": 0.3286, "step": 7650 }, { "epoch": 301.96078431372547, "grad_norm": 0.45633211731910706, "learning_rate": 5.033333333333334e-06, "loss": 0.3285, "step": 7700 }, { "epoch": 303.921568627451, "grad_norm": 0.41933074593544006, "learning_rate": 4.2000000000000004e-06, "loss": 0.3272, "step": 7750 }, { "epoch": 305.88235294117646, "grad_norm": 0.4385930597782135, "learning_rate": 3.3666666666666665e-06, "loss": 0.3297, "step": 7800 }, { "epoch": 307.84313725490193, "grad_norm": 0.49151408672332764, "learning_rate": 2.5333333333333334e-06, "loss": 0.3315, "step": 7850 }, { "epoch": 309.80392156862746, "grad_norm": 0.5050336718559265, "learning_rate": 1.7000000000000002e-06, "loss": 0.331, "step": 7900 }, { "epoch": 311.7647058823529, "grad_norm": 0.3783471882343292, "learning_rate": 8.666666666666667e-07, "loss": 0.3269, "step": 7950 }, { "epoch": 313.72549019607845, "grad_norm": 0.4698627293109894, "learning_rate": 3.3333333333333334e-08, "loss": 0.3312, "step": 8000 }, { "epoch": 313.72549019607845, "eval_loss": 0.4050500988960266, "eval_runtime": 6.4964, "eval_samples_per_second": 27.708, "eval_steps_per_second": 3.54, "step": 8000 } ], "logging_steps": 50, "max_steps": 8000, "num_input_tokens_seen": 0, "num_train_epochs": 320, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.70355911863679e+16, "train_batch_size": 32, "trial_name": null, "trial_params": null }