| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.25, |
| "eval_steps": 500, |
| "global_step": 21420, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0010504201680672268, |
| "grad_norm": 22.625, |
| "learning_rate": 6.000000000000001e-08, |
| "loss": 4.9287, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.0021008403361344537, |
| "grad_norm": 93.0, |
| "learning_rate": 1.2666666666666666e-07, |
| "loss": 4.58, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.0031512605042016808, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.9333333333333337e-07, |
| "loss": 4.8248, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.004201680672268907, |
| "grad_norm": 8.1875, |
| "learning_rate": 2.6e-07, |
| "loss": 5.5906, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.005252100840336135, |
| "grad_norm": 9.0625, |
| "learning_rate": 3.266666666666667e-07, |
| "loss": 5.144, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.0063025210084033615, |
| "grad_norm": 8.5, |
| "learning_rate": 3.9333333333333336e-07, |
| "loss": 5.3149, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.007352941176470588, |
| "grad_norm": 368.0, |
| "learning_rate": 4.6000000000000004e-07, |
| "loss": 5.0195, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.008403361344537815, |
| "grad_norm": 292.0, |
| "learning_rate": 5.266666666666667e-07, |
| "loss": 4.943, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.009453781512605041, |
| "grad_norm": 8.625, |
| "learning_rate": 5.933333333333334e-07, |
| "loss": 5.0446, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.01050420168067227, |
| "grad_norm": 1456.0, |
| "learning_rate": 6.6e-07, |
| "loss": 5.1314, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.011554621848739496, |
| "grad_norm": 13.0625, |
| "learning_rate": 7.266666666666668e-07, |
| "loss": 4.6218, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.012605042016806723, |
| "grad_norm": 249.0, |
| "learning_rate": 7.933333333333335e-07, |
| "loss": 5.4531, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.01365546218487395, |
| "grad_norm": 10.25, |
| "learning_rate": 8.6e-07, |
| "loss": 5.4136, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.014705882352941176, |
| "grad_norm": 13.1875, |
| "learning_rate": 9.266666666666667e-07, |
| "loss": 5.5019, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.015756302521008403, |
| "grad_norm": 26.25, |
| "learning_rate": 9.933333333333333e-07, |
| "loss": 4.987, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.01680672268907563, |
| "grad_norm": 11.0625, |
| "learning_rate": 1.06e-06, |
| "loss": 5.0357, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.017857142857142856, |
| "grad_norm": 8.3125, |
| "learning_rate": 1.1266666666666667e-06, |
| "loss": 5.3842, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.018907563025210083, |
| "grad_norm": 8.125, |
| "learning_rate": 1.1933333333333335e-06, |
| "loss": 5.0255, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.01995798319327731, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.26e-06, |
| "loss": 4.9036, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.02100840336134454, |
| "grad_norm": 11.1875, |
| "learning_rate": 1.3266666666666667e-06, |
| "loss": 5.2209, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.022058823529411766, |
| "grad_norm": 9.5, |
| "learning_rate": 1.3933333333333335e-06, |
| "loss": 4.6235, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.023109243697478993, |
| "grad_norm": 17.25, |
| "learning_rate": 1.46e-06, |
| "loss": 5.4189, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.02415966386554622, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.526666666666667e-06, |
| "loss": 5.3731, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.025210084033613446, |
| "grad_norm": 17.25, |
| "learning_rate": 1.5933333333333335e-06, |
| "loss": 5.1251, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.026260504201680673, |
| "grad_norm": 9.8125, |
| "learning_rate": 1.6600000000000002e-06, |
| "loss": 4.694, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.0273109243697479, |
| "grad_norm": 16.5, |
| "learning_rate": 1.7266666666666667e-06, |
| "loss": 5.1736, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.028361344537815126, |
| "grad_norm": 16.875, |
| "learning_rate": 1.7933333333333337e-06, |
| "loss": 5.0188, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.029411764705882353, |
| "grad_norm": 20.375, |
| "learning_rate": 1.8600000000000002e-06, |
| "loss": 4.5767, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.03046218487394958, |
| "grad_norm": 35.0, |
| "learning_rate": 1.926666666666667e-06, |
| "loss": 4.9608, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.031512605042016806, |
| "grad_norm": 17.125, |
| "learning_rate": 1.9933333333333334e-06, |
| "loss": 5.0574, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.032563025210084036, |
| "grad_norm": 13.375, |
| "learning_rate": 2.06e-06, |
| "loss": 5.245, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.03361344537815126, |
| "grad_norm": 19.25, |
| "learning_rate": 2.126666666666667e-06, |
| "loss": 5.2969, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.03466386554621849, |
| "grad_norm": 10.3125, |
| "learning_rate": 2.1933333333333332e-06, |
| "loss": 4.9717, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.03571428571428571, |
| "grad_norm": 18.375, |
| "learning_rate": 2.2600000000000004e-06, |
| "loss": 4.6562, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.03676470588235294, |
| "grad_norm": 11.5625, |
| "learning_rate": 2.3266666666666667e-06, |
| "loss": 4.7494, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.037815126050420166, |
| "grad_norm": 17.0, |
| "learning_rate": 2.3933333333333334e-06, |
| "loss": 5.1655, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.038865546218487396, |
| "grad_norm": 14.4375, |
| "learning_rate": 2.46e-06, |
| "loss": 5.3803, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.03991596638655462, |
| "grad_norm": 20.625, |
| "learning_rate": 2.526666666666667e-06, |
| "loss": 4.8406, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.04096638655462185, |
| "grad_norm": 27.25, |
| "learning_rate": 2.5933333333333336e-06, |
| "loss": 4.5881, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.04201680672268908, |
| "grad_norm": 408.0, |
| "learning_rate": 2.6600000000000004e-06, |
| "loss": 5.4696, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.0430672268907563, |
| "grad_norm": 15.125, |
| "learning_rate": 2.726666666666667e-06, |
| "loss": 5.1682, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.04411764705882353, |
| "grad_norm": 14.1875, |
| "learning_rate": 2.7933333333333334e-06, |
| "loss": 4.4683, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.045168067226890755, |
| "grad_norm": 32.5, |
| "learning_rate": 2.86e-06, |
| "loss": 5.496, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.046218487394957986, |
| "grad_norm": 23.375, |
| "learning_rate": 2.9266666666666673e-06, |
| "loss": 4.9379, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.04726890756302521, |
| "grad_norm": 14.0625, |
| "learning_rate": 2.9933333333333336e-06, |
| "loss": 5.0653, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.04831932773109244, |
| "grad_norm": 26.625, |
| "learning_rate": 3.0600000000000003e-06, |
| "loss": 4.5886, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.04936974789915966, |
| "grad_norm": 24.125, |
| "learning_rate": 3.1266666666666667e-06, |
| "loss": 5.2361, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.05042016806722689, |
| "grad_norm": 31.5, |
| "learning_rate": 3.193333333333334e-06, |
| "loss": 4.8335, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.051470588235294115, |
| "grad_norm": 13.5625, |
| "learning_rate": 3.2600000000000006e-06, |
| "loss": 4.671, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.052521008403361345, |
| "grad_norm": 22.25, |
| "learning_rate": 3.326666666666667e-06, |
| "loss": 4.9443, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.05357142857142857, |
| "grad_norm": 19.375, |
| "learning_rate": 3.3933333333333336e-06, |
| "loss": 4.8105, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.0546218487394958, |
| "grad_norm": 15.5, |
| "learning_rate": 3.46e-06, |
| "loss": 5.0683, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.05567226890756303, |
| "grad_norm": 28.125, |
| "learning_rate": 3.526666666666667e-06, |
| "loss": 4.3196, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.05672268907563025, |
| "grad_norm": 18.25, |
| "learning_rate": 3.593333333333334e-06, |
| "loss": 4.8723, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.05777310924369748, |
| "grad_norm": 29.25, |
| "learning_rate": 3.66e-06, |
| "loss": 4.8467, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.058823529411764705, |
| "grad_norm": 19.25, |
| "learning_rate": 3.726666666666667e-06, |
| "loss": 4.4188, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.059873949579831935, |
| "grad_norm": 20.75, |
| "learning_rate": 3.793333333333334e-06, |
| "loss": 4.999, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.06092436974789916, |
| "grad_norm": 19.25, |
| "learning_rate": 3.86e-06, |
| "loss": 5.0784, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.06197478991596639, |
| "grad_norm": 25.0, |
| "learning_rate": 3.926666666666667e-06, |
| "loss": 4.2984, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.06302521008403361, |
| "grad_norm": 21.875, |
| "learning_rate": 3.993333333333334e-06, |
| "loss": 4.6611, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.06407563025210083, |
| "grad_norm": 21.75, |
| "learning_rate": 4.060000000000001e-06, |
| "loss": 4.5503, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.06512605042016807, |
| "grad_norm": 31.75, |
| "learning_rate": 4.126666666666667e-06, |
| "loss": 4.2121, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.0661764705882353, |
| "grad_norm": 23.375, |
| "learning_rate": 4.1933333333333336e-06, |
| "loss": 4.3586, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.06722689075630252, |
| "grad_norm": 27.0, |
| "learning_rate": 4.26e-06, |
| "loss": 4.3319, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.06827731092436974, |
| "grad_norm": 108.5, |
| "learning_rate": 4.326666666666667e-06, |
| "loss": 4.4466, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.06932773109243698, |
| "grad_norm": 39.75, |
| "learning_rate": 4.393333333333334e-06, |
| "loss": 4.1232, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.0703781512605042, |
| "grad_norm": 30.75, |
| "learning_rate": 4.4600000000000005e-06, |
| "loss": 4.2004, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.07142857142857142, |
| "grad_norm": 31.875, |
| "learning_rate": 4.526666666666667e-06, |
| "loss": 3.6652, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.07247899159663866, |
| "grad_norm": 30.625, |
| "learning_rate": 4.593333333333333e-06, |
| "loss": 4.6072, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.07352941176470588, |
| "grad_norm": 28.25, |
| "learning_rate": 4.66e-06, |
| "loss": 3.615, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.07457983193277311, |
| "grad_norm": 44.5, |
| "learning_rate": 4.7266666666666674e-06, |
| "loss": 3.7423, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.07563025210084033, |
| "grad_norm": 51.0, |
| "learning_rate": 4.793333333333334e-06, |
| "loss": 3.2943, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.07668067226890757, |
| "grad_norm": 37.25, |
| "learning_rate": 4.86e-06, |
| "loss": 2.8386, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.07773109243697479, |
| "grad_norm": 34.75, |
| "learning_rate": 4.926666666666667e-06, |
| "loss": 2.6554, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.07878151260504201, |
| "grad_norm": 25.75, |
| "learning_rate": 4.9933333333333335e-06, |
| "loss": 3.8282, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.07983193277310924, |
| "grad_norm": 94.0, |
| "learning_rate": 5.060000000000001e-06, |
| "loss": 3.9544, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.08088235294117647, |
| "grad_norm": 36.75, |
| "learning_rate": 5.126666666666668e-06, |
| "loss": 3.3941, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.0819327731092437, |
| "grad_norm": 20.5, |
| "learning_rate": 5.193333333333333e-06, |
| "loss": 2.5309, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.08298319327731092, |
| "grad_norm": 11.125, |
| "learning_rate": 5.2600000000000005e-06, |
| "loss": 2.4891, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.08403361344537816, |
| "grad_norm": 65.0, |
| "learning_rate": 5.326666666666667e-06, |
| "loss": 2.4388, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.08508403361344538, |
| "grad_norm": 21.0, |
| "learning_rate": 5.393333333333334e-06, |
| "loss": 2.3751, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.0861344537815126, |
| "grad_norm": 132.0, |
| "learning_rate": 5.460000000000001e-06, |
| "loss": 3.6548, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.08718487394957983, |
| "grad_norm": 38.0, |
| "learning_rate": 5.5266666666666666e-06, |
| "loss": 2.9053, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.08823529411764706, |
| "grad_norm": 27.375, |
| "learning_rate": 5.593333333333334e-06, |
| "loss": 2.8639, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.08928571428571429, |
| "grad_norm": 98.5, |
| "learning_rate": 5.66e-06, |
| "loss": 2.9273, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.09033613445378151, |
| "grad_norm": 27.875, |
| "learning_rate": 5.726666666666667e-06, |
| "loss": 2.5684, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.09138655462184873, |
| "grad_norm": 16.875, |
| "learning_rate": 5.793333333333334e-06, |
| "loss": 2.7857, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.09243697478991597, |
| "grad_norm": 13.1875, |
| "learning_rate": 5.86e-06, |
| "loss": 2.4058, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.0934873949579832, |
| "grad_norm": 17.0, |
| "learning_rate": 5.926666666666667e-06, |
| "loss": 2.7086, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.09453781512605042, |
| "grad_norm": 15.8125, |
| "learning_rate": 5.993333333333334e-06, |
| "loss": 1.9958, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.09558823529411764, |
| "grad_norm": 25.25, |
| "learning_rate": 6.0600000000000004e-06, |
| "loss": 2.5835, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.09663865546218488, |
| "grad_norm": 22.25, |
| "learning_rate": 6.126666666666668e-06, |
| "loss": 2.6418, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.0976890756302521, |
| "grad_norm": 41.75, |
| "learning_rate": 6.193333333333333e-06, |
| "loss": 2.225, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.09873949579831932, |
| "grad_norm": 38.75, |
| "learning_rate": 6.26e-06, |
| "loss": 2.4876, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.09978991596638656, |
| "grad_norm": 18.375, |
| "learning_rate": 6.326666666666667e-06, |
| "loss": 2.6567, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.10084033613445378, |
| "grad_norm": 35.25, |
| "learning_rate": 6.393333333333334e-06, |
| "loss": 2.9652, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.10189075630252101, |
| "grad_norm": 38.25, |
| "learning_rate": 6.460000000000001e-06, |
| "loss": 2.9891, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.10294117647058823, |
| "grad_norm": 12.1875, |
| "learning_rate": 6.526666666666666e-06, |
| "loss": 1.9093, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.10399159663865547, |
| "grad_norm": 37.0, |
| "learning_rate": 6.5933333333333335e-06, |
| "loss": 2.344, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.10504201680672269, |
| "grad_norm": 36.75, |
| "learning_rate": 6.660000000000001e-06, |
| "loss": 2.3952, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.10609243697478991, |
| "grad_norm": 416.0, |
| "learning_rate": 6.726666666666667e-06, |
| "loss": 1.9161, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.10714285714285714, |
| "grad_norm": 39.25, |
| "learning_rate": 6.793333333333334e-06, |
| "loss": 2.9488, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.10819327731092437, |
| "grad_norm": 7.53125, |
| "learning_rate": 6.860000000000001e-06, |
| "loss": 1.9871, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.1092436974789916, |
| "grad_norm": 18.25, |
| "learning_rate": 6.926666666666667e-06, |
| "loss": 2.176, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.11029411764705882, |
| "grad_norm": 24.0, |
| "learning_rate": 6.993333333333334e-06, |
| "loss": 2.589, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.11134453781512606, |
| "grad_norm": 21.75, |
| "learning_rate": 7.06e-06, |
| "loss": 2.5755, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.11239495798319328, |
| "grad_norm": 17.0, |
| "learning_rate": 7.126666666666667e-06, |
| "loss": 1.8312, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.1134453781512605, |
| "grad_norm": 25.875, |
| "learning_rate": 7.1933333333333345e-06, |
| "loss": 1.9363, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.11449579831932773, |
| "grad_norm": 13.8125, |
| "learning_rate": 7.260000000000001e-06, |
| "loss": 2.6568, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.11554621848739496, |
| "grad_norm": 94.0, |
| "learning_rate": 7.326666666666667e-06, |
| "loss": 2.3152, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.11659663865546219, |
| "grad_norm": 8.875, |
| "learning_rate": 7.393333333333333e-06, |
| "loss": 1.8939, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.11764705882352941, |
| "grad_norm": 29.625, |
| "learning_rate": 7.4600000000000006e-06, |
| "loss": 1.865, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.11869747899159663, |
| "grad_norm": 16.625, |
| "learning_rate": 7.526666666666668e-06, |
| "loss": 2.1626, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.11974789915966387, |
| "grad_norm": 286.0, |
| "learning_rate": 7.593333333333334e-06, |
| "loss": 2.6556, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.1207983193277311, |
| "grad_norm": 25.625, |
| "learning_rate": 7.660000000000001e-06, |
| "loss": 1.9427, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.12184873949579832, |
| "grad_norm": 30.875, |
| "learning_rate": 7.726666666666667e-06, |
| "loss": 1.8481, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.12289915966386554, |
| "grad_norm": 19.0, |
| "learning_rate": 7.793333333333334e-06, |
| "loss": 1.8739, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.12394957983193278, |
| "grad_norm": 22.375, |
| "learning_rate": 7.860000000000001e-06, |
| "loss": 1.7951, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.125, |
| "grad_norm": 215.0, |
| "learning_rate": 7.926666666666666e-06, |
| "loss": 1.5977, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.12605042016806722, |
| "grad_norm": 22.375, |
| "learning_rate": 7.993333333333334e-06, |
| "loss": 1.2865, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.12710084033613445, |
| "grad_norm": 18.25, |
| "learning_rate": 8.06e-06, |
| "loss": 1.9623, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.12815126050420167, |
| "grad_norm": 18.25, |
| "learning_rate": 8.126666666666668e-06, |
| "loss": 2.3094, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.12920168067226892, |
| "grad_norm": 118.5, |
| "learning_rate": 8.193333333333335e-06, |
| "loss": 2.0879, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.13025210084033614, |
| "grad_norm": 7.8125, |
| "learning_rate": 8.26e-06, |
| "loss": 1.5947, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.13130252100840337, |
| "grad_norm": 22.5, |
| "learning_rate": 8.326666666666668e-06, |
| "loss": 2.245, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.1323529411764706, |
| "grad_norm": 14.5, |
| "learning_rate": 8.393333333333335e-06, |
| "loss": 1.8151, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.1334033613445378, |
| "grad_norm": 18.5, |
| "learning_rate": 8.46e-06, |
| "loss": 1.4881, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.13445378151260504, |
| "grad_norm": 33.25, |
| "learning_rate": 8.526666666666667e-06, |
| "loss": 1.9902, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.13550420168067226, |
| "grad_norm": 26.25, |
| "learning_rate": 8.593333333333333e-06, |
| "loss": 2.0831, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.13655462184873948, |
| "grad_norm": 21.25, |
| "learning_rate": 8.66e-06, |
| "loss": 1.6901, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.13760504201680673, |
| "grad_norm": 16.75, |
| "learning_rate": 8.726666666666667e-06, |
| "loss": 1.9426, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.13865546218487396, |
| "grad_norm": 18.0, |
| "learning_rate": 8.793333333333334e-06, |
| "loss": 1.9525, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.13970588235294118, |
| "grad_norm": 8.5, |
| "learning_rate": 8.860000000000002e-06, |
| "loss": 1.9608, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.1407563025210084, |
| "grad_norm": 17.25, |
| "learning_rate": 8.926666666666669e-06, |
| "loss": 1.9182, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.14180672268907563, |
| "grad_norm": 15.625, |
| "learning_rate": 8.993333333333334e-06, |
| "loss": 1.974, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.14285714285714285, |
| "grad_norm": 7.96875, |
| "learning_rate": 9.060000000000001e-06, |
| "loss": 1.5636, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.14390756302521007, |
| "grad_norm": 14.3125, |
| "learning_rate": 9.126666666666667e-06, |
| "loss": 1.9022, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.14495798319327732, |
| "grad_norm": 22.0, |
| "learning_rate": 9.193333333333334e-06, |
| "loss": 2.12, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.14600840336134455, |
| "grad_norm": 8.5625, |
| "learning_rate": 9.260000000000001e-06, |
| "loss": 1.6334, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.14705882352941177, |
| "grad_norm": 24.625, |
| "learning_rate": 9.326666666666667e-06, |
| "loss": 1.3727, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.148109243697479, |
| "grad_norm": 14.75, |
| "learning_rate": 9.393333333333334e-06, |
| "loss": 1.8891, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.14915966386554622, |
| "grad_norm": 10.875, |
| "learning_rate": 9.460000000000001e-06, |
| "loss": 1.5872, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.15021008403361344, |
| "grad_norm": 29.625, |
| "learning_rate": 9.526666666666668e-06, |
| "loss": 1.3031, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.15126050420168066, |
| "grad_norm": 9.8125, |
| "learning_rate": 9.593333333333335e-06, |
| "loss": 1.407, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.15231092436974789, |
| "grad_norm": 24.875, |
| "learning_rate": 9.66e-06, |
| "loss": 2.4515, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.15336134453781514, |
| "grad_norm": 4.59375, |
| "learning_rate": 9.726666666666668e-06, |
| "loss": 1.3885, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.15441176470588236, |
| "grad_norm": 13.625, |
| "learning_rate": 9.793333333333333e-06, |
| "loss": 1.9857, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.15546218487394958, |
| "grad_norm": 16.375, |
| "learning_rate": 9.86e-06, |
| "loss": 2.2004, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.1565126050420168, |
| "grad_norm": 13.4375, |
| "learning_rate": 9.926666666666668e-06, |
| "loss": 1.7066, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.15756302521008403, |
| "grad_norm": 27.625, |
| "learning_rate": 9.993333333333333e-06, |
| "loss": 1.9221, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.15861344537815125, |
| "grad_norm": 16.0, |
| "learning_rate": 1.006e-05, |
| "loss": 1.8841, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.15966386554621848, |
| "grad_norm": 15.1875, |
| "learning_rate": 1.0126666666666667e-05, |
| "loss": 1.9627, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.16071428571428573, |
| "grad_norm": 25.625, |
| "learning_rate": 1.0193333333333335e-05, |
| "loss": 2.4313, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.16176470588235295, |
| "grad_norm": 31.125, |
| "learning_rate": 1.0260000000000002e-05, |
| "loss": 1.8476, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.16281512605042017, |
| "grad_norm": 19.375, |
| "learning_rate": 1.0326666666666667e-05, |
| "loss": 1.4604, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.1638655462184874, |
| "grad_norm": 18.25, |
| "learning_rate": 1.0393333333333336e-05, |
| "loss": 2.0726, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.16491596638655462, |
| "grad_norm": 5.40625, |
| "learning_rate": 1.0460000000000001e-05, |
| "loss": 1.6203, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.16596638655462184, |
| "grad_norm": 8.6875, |
| "learning_rate": 1.0526666666666667e-05, |
| "loss": 1.5707, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.16701680672268907, |
| "grad_norm": 17.125, |
| "learning_rate": 1.0593333333333332e-05, |
| "loss": 1.8378, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.16806722689075632, |
| "grad_norm": 10.75, |
| "learning_rate": 1.0660000000000001e-05, |
| "loss": 1.675, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.16911764705882354, |
| "grad_norm": 20.0, |
| "learning_rate": 1.0726666666666667e-05, |
| "loss": 1.98, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.17016806722689076, |
| "grad_norm": 12.75, |
| "learning_rate": 1.0793333333333334e-05, |
| "loss": 1.4521, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.17121848739495799, |
| "grad_norm": 22.125, |
| "learning_rate": 1.0860000000000001e-05, |
| "loss": 2.1697, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.1722689075630252, |
| "grad_norm": 21.75, |
| "learning_rate": 1.0926666666666668e-05, |
| "loss": 2.4472, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.17331932773109243, |
| "grad_norm": 700.0, |
| "learning_rate": 1.0993333333333334e-05, |
| "loss": 1.9322, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.17436974789915966, |
| "grad_norm": 16.375, |
| "learning_rate": 1.1060000000000003e-05, |
| "loss": 2.1861, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.17542016806722688, |
| "grad_norm": 25.625, |
| "learning_rate": 1.1126666666666668e-05, |
| "loss": 1.8695, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.17647058823529413, |
| "grad_norm": 7.3125, |
| "learning_rate": 1.1193333333333333e-05, |
| "loss": 1.7726, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.17752100840336135, |
| "grad_norm": 25.25, |
| "learning_rate": 1.126e-05, |
| "loss": 2.1717, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.17857142857142858, |
| "grad_norm": 15.6875, |
| "learning_rate": 1.1326666666666668e-05, |
| "loss": 1.6906, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.1796218487394958, |
| "grad_norm": 24.5, |
| "learning_rate": 1.1393333333333333e-05, |
| "loss": 1.5808, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.18067226890756302, |
| "grad_norm": 6.6875, |
| "learning_rate": 1.146e-05, |
| "loss": 1.2771, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.18172268907563024, |
| "grad_norm": 18.0, |
| "learning_rate": 1.1526666666666668e-05, |
| "loss": 2.0885, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.18277310924369747, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.1593333333333335e-05, |
| "loss": 1.1955, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.18382352941176472, |
| "grad_norm": 16.625, |
| "learning_rate": 1.166e-05, |
| "loss": 2.1007, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.18487394957983194, |
| "grad_norm": 24.375, |
| "learning_rate": 1.1726666666666669e-05, |
| "loss": 2.0714, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.18592436974789917, |
| "grad_norm": 19.625, |
| "learning_rate": 1.1793333333333334e-05, |
| "loss": 1.6826, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.1869747899159664, |
| "grad_norm": 7.875, |
| "learning_rate": 1.186e-05, |
| "loss": 1.3679, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.1880252100840336, |
| "grad_norm": 13.625, |
| "learning_rate": 1.1926666666666669e-05, |
| "loss": 1.4829, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.18907563025210083, |
| "grad_norm": 25.25, |
| "learning_rate": 1.1993333333333334e-05, |
| "loss": 1.6078, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.19012605042016806, |
| "grad_norm": 292.0, |
| "learning_rate": 1.2060000000000001e-05, |
| "loss": 1.6539, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.19117647058823528, |
| "grad_norm": 23.5, |
| "learning_rate": 1.2126666666666667e-05, |
| "loss": 1.6501, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.19222689075630253, |
| "grad_norm": 15.375, |
| "learning_rate": 1.2193333333333334e-05, |
| "loss": 1.3654, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.19327731092436976, |
| "grad_norm": 18.25, |
| "learning_rate": 1.2260000000000001e-05, |
| "loss": 1.5726, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.19432773109243698, |
| "grad_norm": 19.625, |
| "learning_rate": 1.2326666666666667e-05, |
| "loss": 1.7464, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.1953781512605042, |
| "grad_norm": 8.5625, |
| "learning_rate": 1.2393333333333336e-05, |
| "loss": 1.4811, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.19642857142857142, |
| "grad_norm": 42.0, |
| "learning_rate": 1.2460000000000001e-05, |
| "loss": 1.2926, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.19747899159663865, |
| "grad_norm": 19.25, |
| "learning_rate": 1.2526666666666666e-05, |
| "loss": 1.2072, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.19852941176470587, |
| "grad_norm": 13.0625, |
| "learning_rate": 1.2593333333333335e-05, |
| "loss": 1.6926, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.19957983193277312, |
| "grad_norm": 193.0, |
| "learning_rate": 1.266e-05, |
| "loss": 1.8289, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.20063025210084034, |
| "grad_norm": 31.875, |
| "learning_rate": 1.2726666666666668e-05, |
| "loss": 1.2674, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.20168067226890757, |
| "grad_norm": 20.125, |
| "learning_rate": 1.2793333333333335e-05, |
| "loss": 1.8923, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.2027310924369748, |
| "grad_norm": 5.8125, |
| "learning_rate": 1.286e-05, |
| "loss": 1.7089, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.20378151260504201, |
| "grad_norm": 21.75, |
| "learning_rate": 1.2926666666666668e-05, |
| "loss": 1.486, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.20483193277310924, |
| "grad_norm": 13.6875, |
| "learning_rate": 1.2993333333333333e-05, |
| "loss": 2.2097, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.20588235294117646, |
| "grad_norm": 28.25, |
| "learning_rate": 1.3060000000000002e-05, |
| "loss": 1.7722, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.20693277310924368, |
| "grad_norm": 13.375, |
| "learning_rate": 1.3126666666666667e-05, |
| "loss": 1.7385, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.20798319327731093, |
| "grad_norm": 6.40625, |
| "learning_rate": 1.3193333333333333e-05, |
| "loss": 1.3068, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.20903361344537816, |
| "grad_norm": 43.25, |
| "learning_rate": 1.3260000000000002e-05, |
| "loss": 1.9285, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.21008403361344538, |
| "grad_norm": 15.0, |
| "learning_rate": 1.3326666666666667e-05, |
| "loss": 1.4772, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.2111344537815126, |
| "grad_norm": 5.90625, |
| "learning_rate": 1.3393333333333334e-05, |
| "loss": 1.6393, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.21218487394957983, |
| "grad_norm": 19.125, |
| "learning_rate": 1.3460000000000002e-05, |
| "loss": 1.3677, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.21323529411764705, |
| "grad_norm": 14.6875, |
| "learning_rate": 1.3526666666666669e-05, |
| "loss": 1.8989, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.21428571428571427, |
| "grad_norm": 14.25, |
| "learning_rate": 1.3593333333333334e-05, |
| "loss": 1.5803, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.21533613445378152, |
| "grad_norm": 15.5625, |
| "learning_rate": 1.3660000000000001e-05, |
| "loss": 1.5648, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.21638655462184875, |
| "grad_norm": 6.46875, |
| "learning_rate": 1.3726666666666669e-05, |
| "loss": 1.5311, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.21743697478991597, |
| "grad_norm": 18.0, |
| "learning_rate": 1.3793333333333334e-05, |
| "loss": 1.5656, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.2184873949579832, |
| "grad_norm": 17.875, |
| "learning_rate": 1.386e-05, |
| "loss": 1.6681, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.21953781512605042, |
| "grad_norm": 13.0, |
| "learning_rate": 1.3926666666666668e-05, |
| "loss": 1.629, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.22058823529411764, |
| "grad_norm": 13.8125, |
| "learning_rate": 1.3993333333333334e-05, |
| "loss": 1.6206, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.22163865546218486, |
| "grad_norm": 11.375, |
| "learning_rate": 1.4060000000000001e-05, |
| "loss": 1.5289, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.22268907563025211, |
| "grad_norm": 19.875, |
| "learning_rate": 1.4126666666666668e-05, |
| "loss": 1.991, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.22373949579831934, |
| "grad_norm": 16.25, |
| "learning_rate": 1.4193333333333335e-05, |
| "loss": 1.9697, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.22478991596638656, |
| "grad_norm": 7.90625, |
| "learning_rate": 1.426e-05, |
| "loss": 1.8469, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.22584033613445378, |
| "grad_norm": 6.28125, |
| "learning_rate": 1.432666666666667e-05, |
| "loss": 1.3915, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.226890756302521, |
| "grad_norm": 6.25, |
| "learning_rate": 1.4393333333333335e-05, |
| "loss": 1.2475, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.22794117647058823, |
| "grad_norm": 7.75, |
| "learning_rate": 1.446e-05, |
| "loss": 1.8652, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.22899159663865545, |
| "grad_norm": 18.125, |
| "learning_rate": 1.452666666666667e-05, |
| "loss": 1.898, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.23004201680672268, |
| "grad_norm": 14.8125, |
| "learning_rate": 1.4593333333333335e-05, |
| "loss": 1.4146, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.23109243697478993, |
| "grad_norm": 13.8125, |
| "learning_rate": 1.466e-05, |
| "loss": 1.761, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.23214285714285715, |
| "grad_norm": 15.0, |
| "learning_rate": 1.4726666666666667e-05, |
| "loss": 1.5954, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.23319327731092437, |
| "grad_norm": 12.8125, |
| "learning_rate": 1.4793333333333335e-05, |
| "loss": 1.7692, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.2342436974789916, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.4860000000000002e-05, |
| "loss": 1.6226, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.23529411764705882, |
| "grad_norm": 15.5, |
| "learning_rate": 1.4926666666666667e-05, |
| "loss": 1.5973, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.23634453781512604, |
| "grad_norm": 14.875, |
| "learning_rate": 1.4993333333333336e-05, |
| "loss": 1.7924, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.23739495798319327, |
| "grad_norm": 12.875, |
| "learning_rate": 1.5060000000000001e-05, |
| "loss": 1.9794, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.23844537815126052, |
| "grad_norm": 18.625, |
| "learning_rate": 1.5126666666666667e-05, |
| "loss": 1.0344, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.23949579831932774, |
| "grad_norm": 25.0, |
| "learning_rate": 1.5193333333333336e-05, |
| "loss": 1.8067, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.24054621848739496, |
| "grad_norm": 13.375, |
| "learning_rate": 1.5260000000000003e-05, |
| "loss": 1.9954, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.2415966386554622, |
| "grad_norm": 15.5, |
| "learning_rate": 1.5326666666666667e-05, |
| "loss": 1.7709, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.2426470588235294, |
| "grad_norm": 7.75, |
| "learning_rate": 1.5393333333333337e-05, |
| "loss": 1.6276, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.24369747899159663, |
| "grad_norm": 15.4375, |
| "learning_rate": 1.546e-05, |
| "loss": 1.7505, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.24474789915966386, |
| "grad_norm": 14.0625, |
| "learning_rate": 1.5526666666666668e-05, |
| "loss": 1.3429, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.24579831932773108, |
| "grad_norm": 16.0, |
| "learning_rate": 1.5593333333333332e-05, |
| "loss": 1.2232, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.24684873949579833, |
| "grad_norm": 13.375, |
| "learning_rate": 1.5660000000000003e-05, |
| "loss": 1.6442, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.24789915966386555, |
| "grad_norm": 10.9375, |
| "learning_rate": 1.5726666666666666e-05, |
| "loss": 1.6963, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.24894957983193278, |
| "grad_norm": 13.5, |
| "learning_rate": 1.5793333333333333e-05, |
| "loss": 1.6566, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 18.0, |
| "learning_rate": 1.586e-05, |
| "loss": 1.7326, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.2510504201680672, |
| "grad_norm": 12.3125, |
| "learning_rate": 1.5926666666666668e-05, |
| "loss": 1.672, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.25210084033613445, |
| "grad_norm": 17.875, |
| "learning_rate": 1.5993333333333335e-05, |
| "loss": 0.9257, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.25315126050420167, |
| "grad_norm": 17.625, |
| "learning_rate": 1.6060000000000002e-05, |
| "loss": 1.9682, |
| "step": 2410 |
| }, |
| { |
| "epoch": 0.2542016806722689, |
| "grad_norm": 9.375, |
| "learning_rate": 1.612666666666667e-05, |
| "loss": 1.6902, |
| "step": 2420 |
| }, |
| { |
| "epoch": 0.2552521008403361, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.6193333333333333e-05, |
| "loss": 1.423, |
| "step": 2430 |
| }, |
| { |
| "epoch": 0.25630252100840334, |
| "grad_norm": 11.4375, |
| "learning_rate": 1.626e-05, |
| "loss": 1.2585, |
| "step": 2440 |
| }, |
| { |
| "epoch": 0.25735294117647056, |
| "grad_norm": 6.65625, |
| "learning_rate": 1.6326666666666667e-05, |
| "loss": 1.127, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.25840336134453784, |
| "grad_norm": 16.25, |
| "learning_rate": 1.6393333333333334e-05, |
| "loss": 1.9252, |
| "step": 2460 |
| }, |
| { |
| "epoch": 0.25945378151260506, |
| "grad_norm": 24.5, |
| "learning_rate": 1.646e-05, |
| "loss": 1.3567, |
| "step": 2470 |
| }, |
| { |
| "epoch": 0.2605042016806723, |
| "grad_norm": 6.5, |
| "learning_rate": 1.652666666666667e-05, |
| "loss": 1.4298, |
| "step": 2480 |
| }, |
| { |
| "epoch": 0.2615546218487395, |
| "grad_norm": 31.5, |
| "learning_rate": 1.6593333333333336e-05, |
| "loss": 1.4108, |
| "step": 2490 |
| }, |
| { |
| "epoch": 0.26260504201680673, |
| "grad_norm": 20.125, |
| "learning_rate": 1.666e-05, |
| "loss": 2.1598, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.26365546218487396, |
| "grad_norm": 25.25, |
| "learning_rate": 1.672666666666667e-05, |
| "loss": 1.3126, |
| "step": 2510 |
| }, |
| { |
| "epoch": 0.2647058823529412, |
| "grad_norm": 13.375, |
| "learning_rate": 1.6793333333333334e-05, |
| "loss": 1.4508, |
| "step": 2520 |
| }, |
| { |
| "epoch": 0.2657563025210084, |
| "grad_norm": 14.75, |
| "learning_rate": 1.686e-05, |
| "loss": 1.2838, |
| "step": 2530 |
| }, |
| { |
| "epoch": 0.2668067226890756, |
| "grad_norm": 26.875, |
| "learning_rate": 1.692666666666667e-05, |
| "loss": 1.3981, |
| "step": 2540 |
| }, |
| { |
| "epoch": 0.26785714285714285, |
| "grad_norm": 14.75, |
| "learning_rate": 1.6993333333333336e-05, |
| "loss": 1.3673, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.2689075630252101, |
| "grad_norm": 20.5, |
| "learning_rate": 1.7060000000000003e-05, |
| "loss": 1.6032, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.2699579831932773, |
| "grad_norm": 16.875, |
| "learning_rate": 1.7126666666666666e-05, |
| "loss": 1.7339, |
| "step": 2570 |
| }, |
| { |
| "epoch": 0.2710084033613445, |
| "grad_norm": 16.25, |
| "learning_rate": 1.7193333333333334e-05, |
| "loss": 1.9788, |
| "step": 2580 |
| }, |
| { |
| "epoch": 0.27205882352941174, |
| "grad_norm": 19.5, |
| "learning_rate": 1.726e-05, |
| "loss": 1.6745, |
| "step": 2590 |
| }, |
| { |
| "epoch": 0.27310924369747897, |
| "grad_norm": 13.5, |
| "learning_rate": 1.7326666666666668e-05, |
| "loss": 1.1851, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.27415966386554624, |
| "grad_norm": 20.125, |
| "learning_rate": 1.7393333333333335e-05, |
| "loss": 1.515, |
| "step": 2610 |
| }, |
| { |
| "epoch": 0.27521008403361347, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.7460000000000002e-05, |
| "loss": 1.4682, |
| "step": 2620 |
| }, |
| { |
| "epoch": 0.2762605042016807, |
| "grad_norm": 21.75, |
| "learning_rate": 1.7526666666666666e-05, |
| "loss": 1.7853, |
| "step": 2630 |
| }, |
| { |
| "epoch": 0.2773109243697479, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.7593333333333337e-05, |
| "loss": 1.7173, |
| "step": 2640 |
| }, |
| { |
| "epoch": 0.27836134453781514, |
| "grad_norm": 11.75, |
| "learning_rate": 1.766e-05, |
| "loss": 1.3274, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.27941176470588236, |
| "grad_norm": 27.125, |
| "learning_rate": 1.7726666666666667e-05, |
| "loss": 2.1297, |
| "step": 2660 |
| }, |
| { |
| "epoch": 0.2804621848739496, |
| "grad_norm": 7.09375, |
| "learning_rate": 1.7793333333333335e-05, |
| "loss": 1.3744, |
| "step": 2670 |
| }, |
| { |
| "epoch": 0.2815126050420168, |
| "grad_norm": 11.0625, |
| "learning_rate": 1.7860000000000002e-05, |
| "loss": 1.5248, |
| "step": 2680 |
| }, |
| { |
| "epoch": 0.28256302521008403, |
| "grad_norm": 18.5, |
| "learning_rate": 1.792666666666667e-05, |
| "loss": 1.6294, |
| "step": 2690 |
| }, |
| { |
| "epoch": 0.28361344537815125, |
| "grad_norm": 24.125, |
| "learning_rate": 1.7993333333333333e-05, |
| "loss": 1.737, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.2846638655462185, |
| "grad_norm": 16.25, |
| "learning_rate": 1.8060000000000003e-05, |
| "loss": 1.3969, |
| "step": 2710 |
| }, |
| { |
| "epoch": 0.2857142857142857, |
| "grad_norm": 5.8125, |
| "learning_rate": 1.8126666666666667e-05, |
| "loss": 1.911, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.2867647058823529, |
| "grad_norm": 15.5, |
| "learning_rate": 1.8193333333333334e-05, |
| "loss": 1.5679, |
| "step": 2730 |
| }, |
| { |
| "epoch": 0.28781512605042014, |
| "grad_norm": 14.0625, |
| "learning_rate": 1.826e-05, |
| "loss": 1.7564, |
| "step": 2740 |
| }, |
| { |
| "epoch": 0.28886554621848737, |
| "grad_norm": 12.375, |
| "learning_rate": 1.832666666666667e-05, |
| "loss": 1.4142, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.28991596638655465, |
| "grad_norm": 19.125, |
| "learning_rate": 1.8393333333333336e-05, |
| "loss": 2.1592, |
| "step": 2760 |
| }, |
| { |
| "epoch": 0.29096638655462187, |
| "grad_norm": 24.875, |
| "learning_rate": 1.8460000000000003e-05, |
| "loss": 1.7029, |
| "step": 2770 |
| }, |
| { |
| "epoch": 0.2920168067226891, |
| "grad_norm": 24.0, |
| "learning_rate": 1.852666666666667e-05, |
| "loss": 1.3118, |
| "step": 2780 |
| }, |
| { |
| "epoch": 0.2930672268907563, |
| "grad_norm": 29.25, |
| "learning_rate": 1.8593333333333334e-05, |
| "loss": 1.8215, |
| "step": 2790 |
| }, |
| { |
| "epoch": 0.29411764705882354, |
| "grad_norm": 6.4375, |
| "learning_rate": 1.866e-05, |
| "loss": 1.5881, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.29516806722689076, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.8726666666666668e-05, |
| "loss": 1.5423, |
| "step": 2810 |
| }, |
| { |
| "epoch": 0.296218487394958, |
| "grad_norm": 8.25, |
| "learning_rate": 1.8793333333333335e-05, |
| "loss": 1.5408, |
| "step": 2820 |
| }, |
| { |
| "epoch": 0.2972689075630252, |
| "grad_norm": 18.125, |
| "learning_rate": 1.886e-05, |
| "loss": 1.3339, |
| "step": 2830 |
| }, |
| { |
| "epoch": 0.29831932773109243, |
| "grad_norm": 12.25, |
| "learning_rate": 1.892666666666667e-05, |
| "loss": 1.3236, |
| "step": 2840 |
| }, |
| { |
| "epoch": 0.29936974789915966, |
| "grad_norm": 32.5, |
| "learning_rate": 1.8993333333333333e-05, |
| "loss": 2.0409, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.3004201680672269, |
| "grad_norm": 20.125, |
| "learning_rate": 1.906e-05, |
| "loss": 1.5573, |
| "step": 2860 |
| }, |
| { |
| "epoch": 0.3014705882352941, |
| "grad_norm": 12.125, |
| "learning_rate": 1.9126666666666668e-05, |
| "loss": 1.3649, |
| "step": 2870 |
| }, |
| { |
| "epoch": 0.3025210084033613, |
| "grad_norm": 13.125, |
| "learning_rate": 1.9193333333333335e-05, |
| "loss": 1.6068, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.30357142857142855, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.9260000000000002e-05, |
| "loss": 1.1925, |
| "step": 2890 |
| }, |
| { |
| "epoch": 0.30462184873949577, |
| "grad_norm": 19.875, |
| "learning_rate": 1.932666666666667e-05, |
| "loss": 1.8134, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.30567226890756305, |
| "grad_norm": 18.375, |
| "learning_rate": 1.9393333333333336e-05, |
| "loss": 1.5501, |
| "step": 2910 |
| }, |
| { |
| "epoch": 0.3067226890756303, |
| "grad_norm": 9.75, |
| "learning_rate": 1.946e-05, |
| "loss": 1.1981, |
| "step": 2920 |
| }, |
| { |
| "epoch": 0.3077731092436975, |
| "grad_norm": 23.0, |
| "learning_rate": 1.952666666666667e-05, |
| "loss": 1.3437, |
| "step": 2930 |
| }, |
| { |
| "epoch": 0.3088235294117647, |
| "grad_norm": 23.75, |
| "learning_rate": 1.9593333333333334e-05, |
| "loss": 1.3747, |
| "step": 2940 |
| }, |
| { |
| "epoch": 0.30987394957983194, |
| "grad_norm": 18.875, |
| "learning_rate": 1.966e-05, |
| "loss": 1.4291, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.31092436974789917, |
| "grad_norm": 10.5, |
| "learning_rate": 1.972666666666667e-05, |
| "loss": 1.3904, |
| "step": 2960 |
| }, |
| { |
| "epoch": 0.3119747899159664, |
| "grad_norm": 8.75, |
| "learning_rate": 1.9793333333333336e-05, |
| "loss": 1.3657, |
| "step": 2970 |
| }, |
| { |
| "epoch": 0.3130252100840336, |
| "grad_norm": 16.875, |
| "learning_rate": 1.9860000000000003e-05, |
| "loss": 1.5698, |
| "step": 2980 |
| }, |
| { |
| "epoch": 0.31407563025210083, |
| "grad_norm": 20.125, |
| "learning_rate": 1.9926666666666667e-05, |
| "loss": 1.4325, |
| "step": 2990 |
| }, |
| { |
| "epoch": 0.31512605042016806, |
| "grad_norm": 13.75, |
| "learning_rate": 1.9993333333333337e-05, |
| "loss": 1.4939, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.3161764705882353, |
| "grad_norm": 5.15625, |
| "learning_rate": 1.9992957746478873e-05, |
| "loss": 1.5617, |
| "step": 3010 |
| }, |
| { |
| "epoch": 0.3172268907563025, |
| "grad_norm": 12.25, |
| "learning_rate": 1.998513302034429e-05, |
| "loss": 1.813, |
| "step": 3020 |
| }, |
| { |
| "epoch": 0.3182773109243697, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.9977308294209706e-05, |
| "loss": 1.3044, |
| "step": 3030 |
| }, |
| { |
| "epoch": 0.31932773109243695, |
| "grad_norm": 15.5, |
| "learning_rate": 1.9969483568075117e-05, |
| "loss": 1.8378, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.32037815126050423, |
| "grad_norm": 16.375, |
| "learning_rate": 1.9961658841940535e-05, |
| "loss": 1.6258, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.32142857142857145, |
| "grad_norm": 11.375, |
| "learning_rate": 1.9953834115805946e-05, |
| "loss": 1.2288, |
| "step": 3060 |
| }, |
| { |
| "epoch": 0.3224789915966387, |
| "grad_norm": 14.0, |
| "learning_rate": 1.9946009389671364e-05, |
| "loss": 1.6408, |
| "step": 3070 |
| }, |
| { |
| "epoch": 0.3235294117647059, |
| "grad_norm": 18.0, |
| "learning_rate": 1.9938184663536776e-05, |
| "loss": 1.5973, |
| "step": 3080 |
| }, |
| { |
| "epoch": 0.3245798319327731, |
| "grad_norm": 12.625, |
| "learning_rate": 1.9930359937402194e-05, |
| "loss": 1.5506, |
| "step": 3090 |
| }, |
| { |
| "epoch": 0.32563025210084034, |
| "grad_norm": 14.75, |
| "learning_rate": 1.9922535211267605e-05, |
| "loss": 1.5711, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.32668067226890757, |
| "grad_norm": 5.71875, |
| "learning_rate": 1.9914710485133023e-05, |
| "loss": 1.2216, |
| "step": 3110 |
| }, |
| { |
| "epoch": 0.3277310924369748, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.9906885758998438e-05, |
| "loss": 2.0402, |
| "step": 3120 |
| }, |
| { |
| "epoch": 0.328781512605042, |
| "grad_norm": 14.75, |
| "learning_rate": 1.9899061032863852e-05, |
| "loss": 1.7379, |
| "step": 3130 |
| }, |
| { |
| "epoch": 0.32983193277310924, |
| "grad_norm": 16.125, |
| "learning_rate": 1.9891236306729267e-05, |
| "loss": 1.4854, |
| "step": 3140 |
| }, |
| { |
| "epoch": 0.33088235294117646, |
| "grad_norm": 20.75, |
| "learning_rate": 1.988341158059468e-05, |
| "loss": 1.5693, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.3319327731092437, |
| "grad_norm": 4.21875, |
| "learning_rate": 1.9875586854460096e-05, |
| "loss": 1.3189, |
| "step": 3160 |
| }, |
| { |
| "epoch": 0.3329831932773109, |
| "grad_norm": 21.25, |
| "learning_rate": 1.986776212832551e-05, |
| "loss": 1.4291, |
| "step": 3170 |
| }, |
| { |
| "epoch": 0.33403361344537813, |
| "grad_norm": 9.5, |
| "learning_rate": 1.9859937402190926e-05, |
| "loss": 1.9245, |
| "step": 3180 |
| }, |
| { |
| "epoch": 0.33508403361344535, |
| "grad_norm": 16.375, |
| "learning_rate": 1.985211267605634e-05, |
| "loss": 1.8689, |
| "step": 3190 |
| }, |
| { |
| "epoch": 0.33613445378151263, |
| "grad_norm": 27.125, |
| "learning_rate": 1.9844287949921755e-05, |
| "loss": 1.8231, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.33718487394957986, |
| "grad_norm": 20.125, |
| "learning_rate": 1.983646322378717e-05, |
| "loss": 1.8372, |
| "step": 3210 |
| }, |
| { |
| "epoch": 0.3382352941176471, |
| "grad_norm": 15.25, |
| "learning_rate": 1.9828638497652584e-05, |
| "loss": 1.4144, |
| "step": 3220 |
| }, |
| { |
| "epoch": 0.3392857142857143, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.9820813771518e-05, |
| "loss": 1.3104, |
| "step": 3230 |
| }, |
| { |
| "epoch": 0.3403361344537815, |
| "grad_norm": 18.375, |
| "learning_rate": 1.9812989045383414e-05, |
| "loss": 1.7318, |
| "step": 3240 |
| }, |
| { |
| "epoch": 0.34138655462184875, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.9805164319248828e-05, |
| "loss": 0.8846, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.34243697478991597, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.9797339593114243e-05, |
| "loss": 1.5413, |
| "step": 3260 |
| }, |
| { |
| "epoch": 0.3434873949579832, |
| "grad_norm": 18.0, |
| "learning_rate": 1.9789514866979657e-05, |
| "loss": 1.5614, |
| "step": 3270 |
| }, |
| { |
| "epoch": 0.3445378151260504, |
| "grad_norm": 29.125, |
| "learning_rate": 1.9781690140845072e-05, |
| "loss": 1.4372, |
| "step": 3280 |
| }, |
| { |
| "epoch": 0.34558823529411764, |
| "grad_norm": 17.0, |
| "learning_rate": 1.9773865414710487e-05, |
| "loss": 1.4374, |
| "step": 3290 |
| }, |
| { |
| "epoch": 0.34663865546218486, |
| "grad_norm": 12.25, |
| "learning_rate": 1.97660406885759e-05, |
| "loss": 1.9225, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.3476890756302521, |
| "grad_norm": 6.375, |
| "learning_rate": 1.9758215962441316e-05, |
| "loss": 1.3581, |
| "step": 3310 |
| }, |
| { |
| "epoch": 0.3487394957983193, |
| "grad_norm": 13.6875, |
| "learning_rate": 1.975039123630673e-05, |
| "loss": 1.4966, |
| "step": 3320 |
| }, |
| { |
| "epoch": 0.34978991596638653, |
| "grad_norm": 12.625, |
| "learning_rate": 1.9742566510172145e-05, |
| "loss": 1.8483, |
| "step": 3330 |
| }, |
| { |
| "epoch": 0.35084033613445376, |
| "grad_norm": 19.375, |
| "learning_rate": 1.973474178403756e-05, |
| "loss": 1.5848, |
| "step": 3340 |
| }, |
| { |
| "epoch": 0.35189075630252103, |
| "grad_norm": 12.875, |
| "learning_rate": 1.9726917057902975e-05, |
| "loss": 1.0251, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.35294117647058826, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.971909233176839e-05, |
| "loss": 1.5157, |
| "step": 3360 |
| }, |
| { |
| "epoch": 0.3539915966386555, |
| "grad_norm": 18.125, |
| "learning_rate": 1.9711267605633804e-05, |
| "loss": 1.6588, |
| "step": 3370 |
| }, |
| { |
| "epoch": 0.3550420168067227, |
| "grad_norm": 20.0, |
| "learning_rate": 1.970344287949922e-05, |
| "loss": 1.3472, |
| "step": 3380 |
| }, |
| { |
| "epoch": 0.3560924369747899, |
| "grad_norm": 4.5, |
| "learning_rate": 1.9695618153364633e-05, |
| "loss": 1.5791, |
| "step": 3390 |
| }, |
| { |
| "epoch": 0.35714285714285715, |
| "grad_norm": 15.0625, |
| "learning_rate": 1.9687793427230048e-05, |
| "loss": 1.769, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.3581932773109244, |
| "grad_norm": 5.09375, |
| "learning_rate": 1.9679968701095463e-05, |
| "loss": 1.1975, |
| "step": 3410 |
| }, |
| { |
| "epoch": 0.3592436974789916, |
| "grad_norm": 11.9375, |
| "learning_rate": 1.9672143974960877e-05, |
| "loss": 1.3957, |
| "step": 3420 |
| }, |
| { |
| "epoch": 0.3602941176470588, |
| "grad_norm": 15.0, |
| "learning_rate": 1.9664319248826292e-05, |
| "loss": 1.4442, |
| "step": 3430 |
| }, |
| { |
| "epoch": 0.36134453781512604, |
| "grad_norm": 13.0, |
| "learning_rate": 1.9656494522691707e-05, |
| "loss": 1.2509, |
| "step": 3440 |
| }, |
| { |
| "epoch": 0.36239495798319327, |
| "grad_norm": 20.125, |
| "learning_rate": 1.964866979655712e-05, |
| "loss": 1.4355, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.3634453781512605, |
| "grad_norm": 22.0, |
| "learning_rate": 1.9640845070422536e-05, |
| "loss": 1.2378, |
| "step": 3460 |
| }, |
| { |
| "epoch": 0.3644957983193277, |
| "grad_norm": 27.0, |
| "learning_rate": 1.963302034428795e-05, |
| "loss": 1.9886, |
| "step": 3470 |
| }, |
| { |
| "epoch": 0.36554621848739494, |
| "grad_norm": 11.4375, |
| "learning_rate": 1.9625195618153365e-05, |
| "loss": 1.3546, |
| "step": 3480 |
| }, |
| { |
| "epoch": 0.36659663865546216, |
| "grad_norm": 24.625, |
| "learning_rate": 1.961737089201878e-05, |
| "loss": 1.4929, |
| "step": 3490 |
| }, |
| { |
| "epoch": 0.36764705882352944, |
| "grad_norm": 16.875, |
| "learning_rate": 1.9609546165884195e-05, |
| "loss": 1.4343, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.36869747899159666, |
| "grad_norm": 18.0, |
| "learning_rate": 1.960172143974961e-05, |
| "loss": 1.6022, |
| "step": 3510 |
| }, |
| { |
| "epoch": 0.3697478991596639, |
| "grad_norm": 12.75, |
| "learning_rate": 1.9593896713615027e-05, |
| "loss": 1.7385, |
| "step": 3520 |
| }, |
| { |
| "epoch": 0.3707983193277311, |
| "grad_norm": 14.8125, |
| "learning_rate": 1.958607198748044e-05, |
| "loss": 1.4431, |
| "step": 3530 |
| }, |
| { |
| "epoch": 0.37184873949579833, |
| "grad_norm": 17.75, |
| "learning_rate": 1.9578247261345857e-05, |
| "loss": 1.7195, |
| "step": 3540 |
| }, |
| { |
| "epoch": 0.37289915966386555, |
| "grad_norm": 27.25, |
| "learning_rate": 1.9570422535211268e-05, |
| "loss": 0.941, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.3739495798319328, |
| "grad_norm": 25.875, |
| "learning_rate": 1.9562597809076686e-05, |
| "loss": 1.0686, |
| "step": 3560 |
| }, |
| { |
| "epoch": 0.375, |
| "grad_norm": 22.375, |
| "learning_rate": 1.9554773082942097e-05, |
| "loss": 1.5486, |
| "step": 3570 |
| }, |
| { |
| "epoch": 0.3760504201680672, |
| "grad_norm": 19.0, |
| "learning_rate": 1.9546948356807515e-05, |
| "loss": 1.5345, |
| "step": 3580 |
| }, |
| { |
| "epoch": 0.37710084033613445, |
| "grad_norm": 12.625, |
| "learning_rate": 1.9539123630672926e-05, |
| "loss": 1.3936, |
| "step": 3590 |
| }, |
| { |
| "epoch": 0.37815126050420167, |
| "grad_norm": 20.25, |
| "learning_rate": 1.9531298904538344e-05, |
| "loss": 1.3183, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.3792016806722689, |
| "grad_norm": 15.8125, |
| "learning_rate": 1.952347417840376e-05, |
| "loss": 1.793, |
| "step": 3610 |
| }, |
| { |
| "epoch": 0.3802521008403361, |
| "grad_norm": 18.375, |
| "learning_rate": 1.951564945226917e-05, |
| "loss": 1.3554, |
| "step": 3620 |
| }, |
| { |
| "epoch": 0.38130252100840334, |
| "grad_norm": 6.375, |
| "learning_rate": 1.950782472613459e-05, |
| "loss": 1.4536, |
| "step": 3630 |
| }, |
| { |
| "epoch": 0.38235294117647056, |
| "grad_norm": 15.3125, |
| "learning_rate": 1.95e-05, |
| "loss": 1.2886, |
| "step": 3640 |
| }, |
| { |
| "epoch": 0.38340336134453784, |
| "grad_norm": 5.59375, |
| "learning_rate": 1.9492175273865418e-05, |
| "loss": 1.3608, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.38445378151260506, |
| "grad_norm": 16.125, |
| "learning_rate": 1.948435054773083e-05, |
| "loss": 1.4178, |
| "step": 3660 |
| }, |
| { |
| "epoch": 0.3855042016806723, |
| "grad_norm": 17.125, |
| "learning_rate": 1.9476525821596247e-05, |
| "loss": 1.1901, |
| "step": 3670 |
| }, |
| { |
| "epoch": 0.3865546218487395, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.9468701095461658e-05, |
| "loss": 1.7493, |
| "step": 3680 |
| }, |
| { |
| "epoch": 0.38760504201680673, |
| "grad_norm": 19.375, |
| "learning_rate": 1.9460876369327076e-05, |
| "loss": 1.5029, |
| "step": 3690 |
| }, |
| { |
| "epoch": 0.38865546218487396, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.945305164319249e-05, |
| "loss": 1.0809, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.3897058823529412, |
| "grad_norm": 20.125, |
| "learning_rate": 1.9445226917057906e-05, |
| "loss": 1.4481, |
| "step": 3710 |
| }, |
| { |
| "epoch": 0.3907563025210084, |
| "grad_norm": 19.875, |
| "learning_rate": 1.943740219092332e-05, |
| "loss": 1.2546, |
| "step": 3720 |
| }, |
| { |
| "epoch": 0.3918067226890756, |
| "grad_norm": 19.625, |
| "learning_rate": 1.9429577464788735e-05, |
| "loss": 1.382, |
| "step": 3730 |
| }, |
| { |
| "epoch": 0.39285714285714285, |
| "grad_norm": 9.125, |
| "learning_rate": 1.942175273865415e-05, |
| "loss": 1.262, |
| "step": 3740 |
| }, |
| { |
| "epoch": 0.3939075630252101, |
| "grad_norm": 15.75, |
| "learning_rate": 1.9413928012519564e-05, |
| "loss": 1.6274, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.3949579831932773, |
| "grad_norm": 13.6875, |
| "learning_rate": 1.940610328638498e-05, |
| "loss": 1.6542, |
| "step": 3760 |
| }, |
| { |
| "epoch": 0.3960084033613445, |
| "grad_norm": 22.75, |
| "learning_rate": 1.9398278560250394e-05, |
| "loss": 1.6889, |
| "step": 3770 |
| }, |
| { |
| "epoch": 0.39705882352941174, |
| "grad_norm": 26.125, |
| "learning_rate": 1.9390453834115808e-05, |
| "loss": 1.3694, |
| "step": 3780 |
| }, |
| { |
| "epoch": 0.39810924369747897, |
| "grad_norm": 15.6875, |
| "learning_rate": 1.9382629107981223e-05, |
| "loss": 1.5103, |
| "step": 3790 |
| }, |
| { |
| "epoch": 0.39915966386554624, |
| "grad_norm": 5.96875, |
| "learning_rate": 1.9374804381846637e-05, |
| "loss": 1.1785, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.40021008403361347, |
| "grad_norm": 19.625, |
| "learning_rate": 1.9366979655712052e-05, |
| "loss": 1.7284, |
| "step": 3810 |
| }, |
| { |
| "epoch": 0.4012605042016807, |
| "grad_norm": 13.5, |
| "learning_rate": 1.9359154929577467e-05, |
| "loss": 1.4694, |
| "step": 3820 |
| }, |
| { |
| "epoch": 0.4023109243697479, |
| "grad_norm": 6.875, |
| "learning_rate": 1.935133020344288e-05, |
| "loss": 1.051, |
| "step": 3830 |
| }, |
| { |
| "epoch": 0.40336134453781514, |
| "grad_norm": 7.375, |
| "learning_rate": 1.9343505477308296e-05, |
| "loss": 1.3056, |
| "step": 3840 |
| }, |
| { |
| "epoch": 0.40441176470588236, |
| "grad_norm": 18.5, |
| "learning_rate": 1.933568075117371e-05, |
| "loss": 1.433, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.4054621848739496, |
| "grad_norm": 17.375, |
| "learning_rate": 1.9327856025039125e-05, |
| "loss": 1.4449, |
| "step": 3860 |
| }, |
| { |
| "epoch": 0.4065126050420168, |
| "grad_norm": 17.25, |
| "learning_rate": 1.932003129890454e-05, |
| "loss": 1.4828, |
| "step": 3870 |
| }, |
| { |
| "epoch": 0.40756302521008403, |
| "grad_norm": 12.125, |
| "learning_rate": 1.9312206572769955e-05, |
| "loss": 1.561, |
| "step": 3880 |
| }, |
| { |
| "epoch": 0.40861344537815125, |
| "grad_norm": 13.4375, |
| "learning_rate": 1.930438184663537e-05, |
| "loss": 1.2401, |
| "step": 3890 |
| }, |
| { |
| "epoch": 0.4096638655462185, |
| "grad_norm": 22.375, |
| "learning_rate": 1.9296557120500784e-05, |
| "loss": 1.4516, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.4107142857142857, |
| "grad_norm": 6.625, |
| "learning_rate": 1.92887323943662e-05, |
| "loss": 1.2575, |
| "step": 3910 |
| }, |
| { |
| "epoch": 0.4117647058823529, |
| "grad_norm": 13.5625, |
| "learning_rate": 1.9280907668231613e-05, |
| "loss": 1.4083, |
| "step": 3920 |
| }, |
| { |
| "epoch": 0.41281512605042014, |
| "grad_norm": 52.0, |
| "learning_rate": 1.9273082942097028e-05, |
| "loss": 1.8745, |
| "step": 3930 |
| }, |
| { |
| "epoch": 0.41386554621848737, |
| "grad_norm": 13.375, |
| "learning_rate": 1.9265258215962443e-05, |
| "loss": 1.4336, |
| "step": 3940 |
| }, |
| { |
| "epoch": 0.41491596638655465, |
| "grad_norm": 13.375, |
| "learning_rate": 1.9257433489827857e-05, |
| "loss": 1.3154, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.41596638655462187, |
| "grad_norm": 16.125, |
| "learning_rate": 1.9249608763693272e-05, |
| "loss": 1.3826, |
| "step": 3960 |
| }, |
| { |
| "epoch": 0.4170168067226891, |
| "grad_norm": 5.65625, |
| "learning_rate": 1.9241784037558687e-05, |
| "loss": 0.9214, |
| "step": 3970 |
| }, |
| { |
| "epoch": 0.4180672268907563, |
| "grad_norm": 18.75, |
| "learning_rate": 1.92339593114241e-05, |
| "loss": 1.4686, |
| "step": 3980 |
| }, |
| { |
| "epoch": 0.41911764705882354, |
| "grad_norm": 12.8125, |
| "learning_rate": 1.9226134585289516e-05, |
| "loss": 1.1541, |
| "step": 3990 |
| }, |
| { |
| "epoch": 0.42016806722689076, |
| "grad_norm": 9.625, |
| "learning_rate": 1.921830985915493e-05, |
| "loss": 1.3521, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.421218487394958, |
| "grad_norm": 14.9375, |
| "learning_rate": 1.9210485133020345e-05, |
| "loss": 1.5706, |
| "step": 4010 |
| }, |
| { |
| "epoch": 0.4222689075630252, |
| "grad_norm": 15.75, |
| "learning_rate": 1.920266040688576e-05, |
| "loss": 1.816, |
| "step": 4020 |
| }, |
| { |
| "epoch": 0.42331932773109243, |
| "grad_norm": 6.40625, |
| "learning_rate": 1.9194835680751174e-05, |
| "loss": 1.3072, |
| "step": 4030 |
| }, |
| { |
| "epoch": 0.42436974789915966, |
| "grad_norm": 15.0, |
| "learning_rate": 1.918701095461659e-05, |
| "loss": 1.0601, |
| "step": 4040 |
| }, |
| { |
| "epoch": 0.4254201680672269, |
| "grad_norm": 12.75, |
| "learning_rate": 1.9179186228482004e-05, |
| "loss": 1.655, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.4264705882352941, |
| "grad_norm": 16.75, |
| "learning_rate": 1.917136150234742e-05, |
| "loss": 1.0496, |
| "step": 4060 |
| }, |
| { |
| "epoch": 0.4275210084033613, |
| "grad_norm": 9.375, |
| "learning_rate": 1.9163536776212833e-05, |
| "loss": 1.2718, |
| "step": 4070 |
| }, |
| { |
| "epoch": 0.42857142857142855, |
| "grad_norm": 13.5, |
| "learning_rate": 1.9155712050078248e-05, |
| "loss": 1.6943, |
| "step": 4080 |
| }, |
| { |
| "epoch": 0.42962184873949577, |
| "grad_norm": 19.625, |
| "learning_rate": 1.9147887323943662e-05, |
| "loss": 1.5612, |
| "step": 4090 |
| }, |
| { |
| "epoch": 0.43067226890756305, |
| "grad_norm": 8.5, |
| "learning_rate": 1.914006259780908e-05, |
| "loss": 1.3209, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.4317226890756303, |
| "grad_norm": 16.5, |
| "learning_rate": 1.9132237871674492e-05, |
| "loss": 1.032, |
| "step": 4110 |
| }, |
| { |
| "epoch": 0.4327731092436975, |
| "grad_norm": 6.28125, |
| "learning_rate": 1.912441314553991e-05, |
| "loss": 0.8253, |
| "step": 4120 |
| }, |
| { |
| "epoch": 0.4338235294117647, |
| "grad_norm": 6.0625, |
| "learning_rate": 1.911658841940532e-05, |
| "loss": 1.4287, |
| "step": 4130 |
| }, |
| { |
| "epoch": 0.43487394957983194, |
| "grad_norm": 28.625, |
| "learning_rate": 1.910876369327074e-05, |
| "loss": 1.7322, |
| "step": 4140 |
| }, |
| { |
| "epoch": 0.43592436974789917, |
| "grad_norm": 19.375, |
| "learning_rate": 1.910093896713615e-05, |
| "loss": 1.7938, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.4369747899159664, |
| "grad_norm": 18.5, |
| "learning_rate": 1.909311424100157e-05, |
| "loss": 1.452, |
| "step": 4160 |
| }, |
| { |
| "epoch": 0.4380252100840336, |
| "grad_norm": 18.375, |
| "learning_rate": 1.908528951486698e-05, |
| "loss": 1.5159, |
| "step": 4170 |
| }, |
| { |
| "epoch": 0.43907563025210083, |
| "grad_norm": 14.5, |
| "learning_rate": 1.9077464788732398e-05, |
| "loss": 1.1355, |
| "step": 4180 |
| }, |
| { |
| "epoch": 0.44012605042016806, |
| "grad_norm": 28.375, |
| "learning_rate": 1.906964006259781e-05, |
| "loss": 1.6948, |
| "step": 4190 |
| }, |
| { |
| "epoch": 0.4411764705882353, |
| "grad_norm": 11.5, |
| "learning_rate": 1.9061815336463224e-05, |
| "loss": 1.709, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.4422268907563025, |
| "grad_norm": 4.28125, |
| "learning_rate": 1.905399061032864e-05, |
| "loss": 1.4682, |
| "step": 4210 |
| }, |
| { |
| "epoch": 0.4432773109243697, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.9046165884194053e-05, |
| "loss": 0.7947, |
| "step": 4220 |
| }, |
| { |
| "epoch": 0.44432773109243695, |
| "grad_norm": 22.875, |
| "learning_rate": 1.903834115805947e-05, |
| "loss": 1.4962, |
| "step": 4230 |
| }, |
| { |
| "epoch": 0.44537815126050423, |
| "grad_norm": 11.625, |
| "learning_rate": 1.9030516431924882e-05, |
| "loss": 1.0637, |
| "step": 4240 |
| }, |
| { |
| "epoch": 0.44642857142857145, |
| "grad_norm": 13.25, |
| "learning_rate": 1.90226917057903e-05, |
| "loss": 1.3702, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.4474789915966387, |
| "grad_norm": 16.75, |
| "learning_rate": 1.901486697965571e-05, |
| "loss": 1.5466, |
| "step": 4260 |
| }, |
| { |
| "epoch": 0.4485294117647059, |
| "grad_norm": 4.03125, |
| "learning_rate": 1.900704225352113e-05, |
| "loss": 0.9244, |
| "step": 4270 |
| }, |
| { |
| "epoch": 0.4495798319327731, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.899921752738654e-05, |
| "loss": 1.0728, |
| "step": 4280 |
| }, |
| { |
| "epoch": 0.45063025210084034, |
| "grad_norm": 52.0, |
| "learning_rate": 1.899139280125196e-05, |
| "loss": 1.7117, |
| "step": 4290 |
| }, |
| { |
| "epoch": 0.45168067226890757, |
| "grad_norm": 20.375, |
| "learning_rate": 1.8983568075117374e-05, |
| "loss": 1.5765, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.4527310924369748, |
| "grad_norm": 16.0, |
| "learning_rate": 1.8975743348982788e-05, |
| "loss": 1.5938, |
| "step": 4310 |
| }, |
| { |
| "epoch": 0.453781512605042, |
| "grad_norm": 16.25, |
| "learning_rate": 1.8967918622848203e-05, |
| "loss": 1.9375, |
| "step": 4320 |
| }, |
| { |
| "epoch": 0.45483193277310924, |
| "grad_norm": 3.8125, |
| "learning_rate": 1.8960093896713617e-05, |
| "loss": 1.0625, |
| "step": 4330 |
| }, |
| { |
| "epoch": 0.45588235294117646, |
| "grad_norm": 12.625, |
| "learning_rate": 1.8952269170579032e-05, |
| "loss": 1.3167, |
| "step": 4340 |
| }, |
| { |
| "epoch": 0.4569327731092437, |
| "grad_norm": 12.8125, |
| "learning_rate": 1.8944444444444447e-05, |
| "loss": 1.2067, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.4579831932773109, |
| "grad_norm": 4.46875, |
| "learning_rate": 1.893661971830986e-05, |
| "loss": 1.3717, |
| "step": 4360 |
| }, |
| { |
| "epoch": 0.45903361344537813, |
| "grad_norm": 22.625, |
| "learning_rate": 1.8928794992175276e-05, |
| "loss": 1.4667, |
| "step": 4370 |
| }, |
| { |
| "epoch": 0.46008403361344535, |
| "grad_norm": 13.75, |
| "learning_rate": 1.892097026604069e-05, |
| "loss": 1.6269, |
| "step": 4380 |
| }, |
| { |
| "epoch": 0.46113445378151263, |
| "grad_norm": 21.25, |
| "learning_rate": 1.8913145539906105e-05, |
| "loss": 1.2358, |
| "step": 4390 |
| }, |
| { |
| "epoch": 0.46218487394957986, |
| "grad_norm": 12.0, |
| "learning_rate": 1.890532081377152e-05, |
| "loss": 1.0913, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.4632352941176471, |
| "grad_norm": 17.5, |
| "learning_rate": 1.8897496087636935e-05, |
| "loss": 1.2854, |
| "step": 4410 |
| }, |
| { |
| "epoch": 0.4642857142857143, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.888967136150235e-05, |
| "loss": 1.2785, |
| "step": 4420 |
| }, |
| { |
| "epoch": 0.4653361344537815, |
| "grad_norm": 11.625, |
| "learning_rate": 1.8881846635367764e-05, |
| "loss": 1.3992, |
| "step": 4430 |
| }, |
| { |
| "epoch": 0.46638655462184875, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.887402190923318e-05, |
| "loss": 1.5738, |
| "step": 4440 |
| }, |
| { |
| "epoch": 0.46743697478991597, |
| "grad_norm": 13.0625, |
| "learning_rate": 1.8866197183098593e-05, |
| "loss": 1.1751, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.4684873949579832, |
| "grad_norm": 12.8125, |
| "learning_rate": 1.8858372456964008e-05, |
| "loss": 1.0964, |
| "step": 4460 |
| }, |
| { |
| "epoch": 0.4695378151260504, |
| "grad_norm": 17.75, |
| "learning_rate": 1.8850547730829423e-05, |
| "loss": 1.6314, |
| "step": 4470 |
| }, |
| { |
| "epoch": 0.47058823529411764, |
| "grad_norm": 13.5, |
| "learning_rate": 1.8842723004694837e-05, |
| "loss": 1.5794, |
| "step": 4480 |
| }, |
| { |
| "epoch": 0.47163865546218486, |
| "grad_norm": 13.9375, |
| "learning_rate": 1.8834898278560252e-05, |
| "loss": 1.7204, |
| "step": 4490 |
| }, |
| { |
| "epoch": 0.4726890756302521, |
| "grad_norm": 16.625, |
| "learning_rate": 1.8827073552425667e-05, |
| "loss": 1.4432, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.4737394957983193, |
| "grad_norm": 12.75, |
| "learning_rate": 1.881924882629108e-05, |
| "loss": 1.3706, |
| "step": 4510 |
| }, |
| { |
| "epoch": 0.47478991596638653, |
| "grad_norm": 13.125, |
| "learning_rate": 1.8811424100156496e-05, |
| "loss": 1.2891, |
| "step": 4520 |
| }, |
| { |
| "epoch": 0.47584033613445376, |
| "grad_norm": 13.8125, |
| "learning_rate": 1.880359937402191e-05, |
| "loss": 1.4449, |
| "step": 4530 |
| }, |
| { |
| "epoch": 0.47689075630252103, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.8795774647887325e-05, |
| "loss": 1.6297, |
| "step": 4540 |
| }, |
| { |
| "epoch": 0.47794117647058826, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.878794992175274e-05, |
| "loss": 1.6155, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.4789915966386555, |
| "grad_norm": 12.25, |
| "learning_rate": 1.8780125195618154e-05, |
| "loss": 1.5584, |
| "step": 4560 |
| }, |
| { |
| "epoch": 0.4800420168067227, |
| "grad_norm": 16.0, |
| "learning_rate": 1.877230046948357e-05, |
| "loss": 1.3156, |
| "step": 4570 |
| }, |
| { |
| "epoch": 0.4810924369747899, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.8764475743348984e-05, |
| "loss": 1.5137, |
| "step": 4580 |
| }, |
| { |
| "epoch": 0.48214285714285715, |
| "grad_norm": 17.75, |
| "learning_rate": 1.87566510172144e-05, |
| "loss": 1.4913, |
| "step": 4590 |
| }, |
| { |
| "epoch": 0.4831932773109244, |
| "grad_norm": 11.5, |
| "learning_rate": 1.8748826291079813e-05, |
| "loss": 0.9244, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.4842436974789916, |
| "grad_norm": 14.125, |
| "learning_rate": 1.8741001564945228e-05, |
| "loss": 2.0681, |
| "step": 4610 |
| }, |
| { |
| "epoch": 0.4852941176470588, |
| "grad_norm": 17.0, |
| "learning_rate": 1.8733176838810642e-05, |
| "loss": 1.6281, |
| "step": 4620 |
| }, |
| { |
| "epoch": 0.48634453781512604, |
| "grad_norm": 21.0, |
| "learning_rate": 1.8725352112676057e-05, |
| "loss": 1.6307, |
| "step": 4630 |
| }, |
| { |
| "epoch": 0.48739495798319327, |
| "grad_norm": 12.75, |
| "learning_rate": 1.8717527386541472e-05, |
| "loss": 1.4405, |
| "step": 4640 |
| }, |
| { |
| "epoch": 0.4884453781512605, |
| "grad_norm": 12.75, |
| "learning_rate": 1.8709702660406886e-05, |
| "loss": 1.3397, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.4894957983193277, |
| "grad_norm": 14.5, |
| "learning_rate": 1.87018779342723e-05, |
| "loss": 1.521, |
| "step": 4660 |
| }, |
| { |
| "epoch": 0.49054621848739494, |
| "grad_norm": 15.375, |
| "learning_rate": 1.8694053208137716e-05, |
| "loss": 1.4432, |
| "step": 4670 |
| }, |
| { |
| "epoch": 0.49159663865546216, |
| "grad_norm": 15.8125, |
| "learning_rate": 1.868622848200313e-05, |
| "loss": 1.7055, |
| "step": 4680 |
| }, |
| { |
| "epoch": 0.49264705882352944, |
| "grad_norm": 14.0, |
| "learning_rate": 1.8678403755868545e-05, |
| "loss": 1.5279, |
| "step": 4690 |
| }, |
| { |
| "epoch": 0.49369747899159666, |
| "grad_norm": 11.625, |
| "learning_rate": 1.8670579029733963e-05, |
| "loss": 1.4292, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.4947478991596639, |
| "grad_norm": 9.8125, |
| "learning_rate": 1.8662754303599374e-05, |
| "loss": 1.2928, |
| "step": 4710 |
| }, |
| { |
| "epoch": 0.4957983193277311, |
| "grad_norm": 9.375, |
| "learning_rate": 1.8654929577464792e-05, |
| "loss": 1.3798, |
| "step": 4720 |
| }, |
| { |
| "epoch": 0.49684873949579833, |
| "grad_norm": 14.375, |
| "learning_rate": 1.8647104851330204e-05, |
| "loss": 1.4781, |
| "step": 4730 |
| }, |
| { |
| "epoch": 0.49789915966386555, |
| "grad_norm": 18.0, |
| "learning_rate": 1.863928012519562e-05, |
| "loss": 1.6945, |
| "step": 4740 |
| }, |
| { |
| "epoch": 0.4989495798319328, |
| "grad_norm": 12.375, |
| "learning_rate": 1.8631455399061033e-05, |
| "loss": 1.632, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.862363067292645e-05, |
| "loss": 1.2287, |
| "step": 4760 |
| }, |
| { |
| "epoch": 0.5010504201680672, |
| "grad_norm": 10.625, |
| "learning_rate": 1.8615805946791862e-05, |
| "loss": 1.4539, |
| "step": 4770 |
| }, |
| { |
| "epoch": 0.5021008403361344, |
| "grad_norm": 16.25, |
| "learning_rate": 1.8607981220657277e-05, |
| "loss": 1.6725, |
| "step": 4780 |
| }, |
| { |
| "epoch": 0.5031512605042017, |
| "grad_norm": 12.3125, |
| "learning_rate": 1.8600156494522695e-05, |
| "loss": 1.7122, |
| "step": 4790 |
| }, |
| { |
| "epoch": 0.5042016806722689, |
| "grad_norm": 13.6875, |
| "learning_rate": 1.8592331768388106e-05, |
| "loss": 1.8102, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.5052521008403361, |
| "grad_norm": 12.625, |
| "learning_rate": 1.8584507042253524e-05, |
| "loss": 1.2716, |
| "step": 4810 |
| }, |
| { |
| "epoch": 0.5063025210084033, |
| "grad_norm": 18.75, |
| "learning_rate": 1.8576682316118935e-05, |
| "loss": 1.52, |
| "step": 4820 |
| }, |
| { |
| "epoch": 0.5073529411764706, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.8568857589984354e-05, |
| "loss": 1.4568, |
| "step": 4830 |
| }, |
| { |
| "epoch": 0.5084033613445378, |
| "grad_norm": 11.75, |
| "learning_rate": 1.8561032863849765e-05, |
| "loss": 1.6998, |
| "step": 4840 |
| }, |
| { |
| "epoch": 0.509453781512605, |
| "grad_norm": 13.625, |
| "learning_rate": 1.8553208137715183e-05, |
| "loss": 1.4233, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.5105042016806722, |
| "grad_norm": 20.625, |
| "learning_rate": 1.8545383411580594e-05, |
| "loss": 1.4319, |
| "step": 4860 |
| }, |
| { |
| "epoch": 0.5115546218487395, |
| "grad_norm": 22.875, |
| "learning_rate": 1.8537558685446012e-05, |
| "loss": 1.4499, |
| "step": 4870 |
| }, |
| { |
| "epoch": 0.5126050420168067, |
| "grad_norm": 7.0625, |
| "learning_rate": 1.8529733959311427e-05, |
| "loss": 1.3792, |
| "step": 4880 |
| }, |
| { |
| "epoch": 0.5136554621848739, |
| "grad_norm": 6.9375, |
| "learning_rate": 1.852190923317684e-05, |
| "loss": 1.0939, |
| "step": 4890 |
| }, |
| { |
| "epoch": 0.5147058823529411, |
| "grad_norm": 6.125, |
| "learning_rate": 1.8514084507042256e-05, |
| "loss": 1.5026, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.5157563025210085, |
| "grad_norm": 14.75, |
| "learning_rate": 1.850625978090767e-05, |
| "loss": 1.472, |
| "step": 4910 |
| }, |
| { |
| "epoch": 0.5168067226890757, |
| "grad_norm": 14.375, |
| "learning_rate": 1.8498435054773085e-05, |
| "loss": 1.2507, |
| "step": 4920 |
| }, |
| { |
| "epoch": 0.5178571428571429, |
| "grad_norm": 3.921875, |
| "learning_rate": 1.84906103286385e-05, |
| "loss": 1.6288, |
| "step": 4930 |
| }, |
| { |
| "epoch": 0.5189075630252101, |
| "grad_norm": 18.5, |
| "learning_rate": 1.8482785602503915e-05, |
| "loss": 1.7701, |
| "step": 4940 |
| }, |
| { |
| "epoch": 0.5199579831932774, |
| "grad_norm": 15.5625, |
| "learning_rate": 1.847496087636933e-05, |
| "loss": 1.5399, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.5210084033613446, |
| "grad_norm": 17.5, |
| "learning_rate": 1.8467136150234744e-05, |
| "loss": 1.5214, |
| "step": 4960 |
| }, |
| { |
| "epoch": 0.5220588235294118, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.845931142410016e-05, |
| "loss": 1.6711, |
| "step": 4970 |
| }, |
| { |
| "epoch": 0.523109243697479, |
| "grad_norm": 15.5, |
| "learning_rate": 1.8451486697965573e-05, |
| "loss": 1.5255, |
| "step": 4980 |
| }, |
| { |
| "epoch": 0.5241596638655462, |
| "grad_norm": 23.75, |
| "learning_rate": 1.8443661971830988e-05, |
| "loss": 1.1936, |
| "step": 4990 |
| }, |
| { |
| "epoch": 0.5252100840336135, |
| "grad_norm": 11.25, |
| "learning_rate": 1.8435837245696403e-05, |
| "loss": 1.6067, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.5262605042016807, |
| "grad_norm": 9.75, |
| "learning_rate": 1.8428012519561817e-05, |
| "loss": 1.6151, |
| "step": 5010 |
| }, |
| { |
| "epoch": 0.5273109243697479, |
| "grad_norm": 23.625, |
| "learning_rate": 1.8420187793427232e-05, |
| "loss": 1.7412, |
| "step": 5020 |
| }, |
| { |
| "epoch": 0.5283613445378151, |
| "grad_norm": 14.625, |
| "learning_rate": 1.8412363067292647e-05, |
| "loss": 1.2956, |
| "step": 5030 |
| }, |
| { |
| "epoch": 0.5294117647058824, |
| "grad_norm": 14.0625, |
| "learning_rate": 1.840453834115806e-05, |
| "loss": 1.5333, |
| "step": 5040 |
| }, |
| { |
| "epoch": 0.5304621848739496, |
| "grad_norm": 12.375, |
| "learning_rate": 1.8396713615023476e-05, |
| "loss": 1.3507, |
| "step": 5050 |
| }, |
| { |
| "epoch": 0.5315126050420168, |
| "grad_norm": 17.375, |
| "learning_rate": 1.838888888888889e-05, |
| "loss": 1.2024, |
| "step": 5060 |
| }, |
| { |
| "epoch": 0.532563025210084, |
| "grad_norm": 4.6875, |
| "learning_rate": 1.8381064162754305e-05, |
| "loss": 1.542, |
| "step": 5070 |
| }, |
| { |
| "epoch": 0.5336134453781513, |
| "grad_norm": 24.875, |
| "learning_rate": 1.837323943661972e-05, |
| "loss": 1.3459, |
| "step": 5080 |
| }, |
| { |
| "epoch": 0.5346638655462185, |
| "grad_norm": 20.25, |
| "learning_rate": 1.8365414710485134e-05, |
| "loss": 1.4547, |
| "step": 5090 |
| }, |
| { |
| "epoch": 0.5357142857142857, |
| "grad_norm": 8.75, |
| "learning_rate": 1.835758998435055e-05, |
| "loss": 1.3747, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.5367647058823529, |
| "grad_norm": 4.90625, |
| "learning_rate": 1.8349765258215964e-05, |
| "loss": 1.449, |
| "step": 5110 |
| }, |
| { |
| "epoch": 0.5378151260504201, |
| "grad_norm": 5.34375, |
| "learning_rate": 1.834194053208138e-05, |
| "loss": 1.4577, |
| "step": 5120 |
| }, |
| { |
| "epoch": 0.5388655462184874, |
| "grad_norm": 14.375, |
| "learning_rate": 1.8334115805946793e-05, |
| "loss": 1.2082, |
| "step": 5130 |
| }, |
| { |
| "epoch": 0.5399159663865546, |
| "grad_norm": 16.375, |
| "learning_rate": 1.8326291079812208e-05, |
| "loss": 1.3699, |
| "step": 5140 |
| }, |
| { |
| "epoch": 0.5409663865546218, |
| "grad_norm": 60.75, |
| "learning_rate": 1.8318466353677622e-05, |
| "loss": 1.5371, |
| "step": 5150 |
| }, |
| { |
| "epoch": 0.542016806722689, |
| "grad_norm": 13.1875, |
| "learning_rate": 1.8310641627543037e-05, |
| "loss": 1.1132, |
| "step": 5160 |
| }, |
| { |
| "epoch": 0.5430672268907563, |
| "grad_norm": 16.625, |
| "learning_rate": 1.8302816901408452e-05, |
| "loss": 1.3545, |
| "step": 5170 |
| }, |
| { |
| "epoch": 0.5441176470588235, |
| "grad_norm": 5.0625, |
| "learning_rate": 1.8294992175273866e-05, |
| "loss": 1.3821, |
| "step": 5180 |
| }, |
| { |
| "epoch": 0.5451680672268907, |
| "grad_norm": 14.875, |
| "learning_rate": 1.828716744913928e-05, |
| "loss": 1.4026, |
| "step": 5190 |
| }, |
| { |
| "epoch": 0.5462184873949579, |
| "grad_norm": 21.0, |
| "learning_rate": 1.8279342723004696e-05, |
| "loss": 1.3391, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.5472689075630253, |
| "grad_norm": 16.375, |
| "learning_rate": 1.827151799687011e-05, |
| "loss": 1.603, |
| "step": 5210 |
| }, |
| { |
| "epoch": 0.5483193277310925, |
| "grad_norm": 7.75, |
| "learning_rate": 1.8263693270735525e-05, |
| "loss": 1.4944, |
| "step": 5220 |
| }, |
| { |
| "epoch": 0.5493697478991597, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.825586854460094e-05, |
| "loss": 1.4024, |
| "step": 5230 |
| }, |
| { |
| "epoch": 0.5504201680672269, |
| "grad_norm": 15.3125, |
| "learning_rate": 1.8248043818466354e-05, |
| "loss": 1.0601, |
| "step": 5240 |
| }, |
| { |
| "epoch": 0.5514705882352942, |
| "grad_norm": 9.8125, |
| "learning_rate": 1.824021909233177e-05, |
| "loss": 1.6795, |
| "step": 5250 |
| }, |
| { |
| "epoch": 0.5525210084033614, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.8232394366197184e-05, |
| "loss": 1.3527, |
| "step": 5260 |
| }, |
| { |
| "epoch": 0.5535714285714286, |
| "grad_norm": 4.3125, |
| "learning_rate": 1.8224569640062598e-05, |
| "loss": 1.422, |
| "step": 5270 |
| }, |
| { |
| "epoch": 0.5546218487394958, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.8216744913928016e-05, |
| "loss": 1.3717, |
| "step": 5280 |
| }, |
| { |
| "epoch": 0.555672268907563, |
| "grad_norm": 12.875, |
| "learning_rate": 1.8208920187793428e-05, |
| "loss": 1.3349, |
| "step": 5290 |
| }, |
| { |
| "epoch": 0.5567226890756303, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.8201095461658846e-05, |
| "loss": 1.3049, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.5577731092436975, |
| "grad_norm": 15.3125, |
| "learning_rate": 1.8193270735524257e-05, |
| "loss": 1.6232, |
| "step": 5310 |
| }, |
| { |
| "epoch": 0.5588235294117647, |
| "grad_norm": 16.125, |
| "learning_rate": 1.8185446009389675e-05, |
| "loss": 1.2363, |
| "step": 5320 |
| }, |
| { |
| "epoch": 0.5598739495798319, |
| "grad_norm": 12.875, |
| "learning_rate": 1.8177621283255086e-05, |
| "loss": 1.3058, |
| "step": 5330 |
| }, |
| { |
| "epoch": 0.5609243697478992, |
| "grad_norm": 13.25, |
| "learning_rate": 1.8169796557120504e-05, |
| "loss": 1.4614, |
| "step": 5340 |
| }, |
| { |
| "epoch": 0.5619747899159664, |
| "grad_norm": 13.5625, |
| "learning_rate": 1.8161971830985915e-05, |
| "loss": 1.5089, |
| "step": 5350 |
| }, |
| { |
| "epoch": 0.5630252100840336, |
| "grad_norm": 10.375, |
| "learning_rate": 1.8154147104851333e-05, |
| "loss": 1.0185, |
| "step": 5360 |
| }, |
| { |
| "epoch": 0.5640756302521008, |
| "grad_norm": 22.5, |
| "learning_rate": 1.8146322378716745e-05, |
| "loss": 1.0896, |
| "step": 5370 |
| }, |
| { |
| "epoch": 0.5651260504201681, |
| "grad_norm": 26.375, |
| "learning_rate": 1.813849765258216e-05, |
| "loss": 1.449, |
| "step": 5380 |
| }, |
| { |
| "epoch": 0.5661764705882353, |
| "grad_norm": 15.0625, |
| "learning_rate": 1.8130672926447577e-05, |
| "loss": 0.989, |
| "step": 5390 |
| }, |
| { |
| "epoch": 0.5672268907563025, |
| "grad_norm": 5.34375, |
| "learning_rate": 1.812284820031299e-05, |
| "loss": 1.3221, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.5682773109243697, |
| "grad_norm": 19.625, |
| "learning_rate": 1.8115023474178407e-05, |
| "loss": 1.1908, |
| "step": 5410 |
| }, |
| { |
| "epoch": 0.569327731092437, |
| "grad_norm": 17.875, |
| "learning_rate": 1.8107198748043818e-05, |
| "loss": 1.6788, |
| "step": 5420 |
| }, |
| { |
| "epoch": 0.5703781512605042, |
| "grad_norm": 12.875, |
| "learning_rate": 1.8099374021909236e-05, |
| "loss": 1.5798, |
| "step": 5430 |
| }, |
| { |
| "epoch": 0.5714285714285714, |
| "grad_norm": 47.25, |
| "learning_rate": 1.8091549295774647e-05, |
| "loss": 1.3783, |
| "step": 5440 |
| }, |
| { |
| "epoch": 0.5724789915966386, |
| "grad_norm": 5.5, |
| "learning_rate": 1.8083724569640065e-05, |
| "loss": 1.4343, |
| "step": 5450 |
| }, |
| { |
| "epoch": 0.5735294117647058, |
| "grad_norm": 15.25, |
| "learning_rate": 1.8075899843505477e-05, |
| "loss": 1.3189, |
| "step": 5460 |
| }, |
| { |
| "epoch": 0.5745798319327731, |
| "grad_norm": 17.5, |
| "learning_rate": 1.8068075117370895e-05, |
| "loss": 1.821, |
| "step": 5470 |
| }, |
| { |
| "epoch": 0.5756302521008403, |
| "grad_norm": 9.8125, |
| "learning_rate": 1.806025039123631e-05, |
| "loss": 1.3433, |
| "step": 5480 |
| }, |
| { |
| "epoch": 0.5766806722689075, |
| "grad_norm": 9.1875, |
| "learning_rate": 1.8052425665101724e-05, |
| "loss": 1.647, |
| "step": 5490 |
| }, |
| { |
| "epoch": 0.5777310924369747, |
| "grad_norm": 6.53125, |
| "learning_rate": 1.804460093896714e-05, |
| "loss": 1.1225, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.5787815126050421, |
| "grad_norm": 21.625, |
| "learning_rate": 1.8036776212832553e-05, |
| "loss": 1.2338, |
| "step": 5510 |
| }, |
| { |
| "epoch": 0.5798319327731093, |
| "grad_norm": 15.9375, |
| "learning_rate": 1.8028951486697968e-05, |
| "loss": 1.333, |
| "step": 5520 |
| }, |
| { |
| "epoch": 0.5808823529411765, |
| "grad_norm": 5.0, |
| "learning_rate": 1.8021126760563383e-05, |
| "loss": 1.1024, |
| "step": 5530 |
| }, |
| { |
| "epoch": 0.5819327731092437, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.8013302034428797e-05, |
| "loss": 1.2991, |
| "step": 5540 |
| }, |
| { |
| "epoch": 0.582983193277311, |
| "grad_norm": 13.25, |
| "learning_rate": 1.8005477308294212e-05, |
| "loss": 1.7033, |
| "step": 5550 |
| }, |
| { |
| "epoch": 0.5840336134453782, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.7997652582159627e-05, |
| "loss": 1.3945, |
| "step": 5560 |
| }, |
| { |
| "epoch": 0.5850840336134454, |
| "grad_norm": 5.21875, |
| "learning_rate": 1.798982785602504e-05, |
| "loss": 1.3108, |
| "step": 5570 |
| }, |
| { |
| "epoch": 0.5861344537815126, |
| "grad_norm": 7.71875, |
| "learning_rate": 1.7982003129890456e-05, |
| "loss": 1.3495, |
| "step": 5580 |
| }, |
| { |
| "epoch": 0.5871848739495799, |
| "grad_norm": 17.875, |
| "learning_rate": 1.797417840375587e-05, |
| "loss": 1.2221, |
| "step": 5590 |
| }, |
| { |
| "epoch": 0.5882352941176471, |
| "grad_norm": 7.96875, |
| "learning_rate": 1.7966353677621285e-05, |
| "loss": 1.3192, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.5892857142857143, |
| "grad_norm": 16.875, |
| "learning_rate": 1.79585289514867e-05, |
| "loss": 1.7956, |
| "step": 5610 |
| }, |
| { |
| "epoch": 0.5903361344537815, |
| "grad_norm": 15.0, |
| "learning_rate": 1.7950704225352114e-05, |
| "loss": 1.57, |
| "step": 5620 |
| }, |
| { |
| "epoch": 0.5913865546218487, |
| "grad_norm": 15.625, |
| "learning_rate": 1.794287949921753e-05, |
| "loss": 1.216, |
| "step": 5630 |
| }, |
| { |
| "epoch": 0.592436974789916, |
| "grad_norm": 17.375, |
| "learning_rate": 1.7935054773082944e-05, |
| "loss": 1.39, |
| "step": 5640 |
| }, |
| { |
| "epoch": 0.5934873949579832, |
| "grad_norm": 4.25, |
| "learning_rate": 1.792723004694836e-05, |
| "loss": 1.1288, |
| "step": 5650 |
| }, |
| { |
| "epoch": 0.5945378151260504, |
| "grad_norm": 17.375, |
| "learning_rate": 1.7919405320813773e-05, |
| "loss": 1.3017, |
| "step": 5660 |
| }, |
| { |
| "epoch": 0.5955882352941176, |
| "grad_norm": 10.125, |
| "learning_rate": 1.7911580594679188e-05, |
| "loss": 1.2518, |
| "step": 5670 |
| }, |
| { |
| "epoch": 0.5966386554621849, |
| "grad_norm": 13.0625, |
| "learning_rate": 1.7903755868544602e-05, |
| "loss": 1.3653, |
| "step": 5680 |
| }, |
| { |
| "epoch": 0.5976890756302521, |
| "grad_norm": 13.0, |
| "learning_rate": 1.7895931142410017e-05, |
| "loss": 1.2755, |
| "step": 5690 |
| }, |
| { |
| "epoch": 0.5987394957983193, |
| "grad_norm": 8.8125, |
| "learning_rate": 1.788810641627543e-05, |
| "loss": 1.4182, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.5997899159663865, |
| "grad_norm": 19.375, |
| "learning_rate": 1.7880281690140846e-05, |
| "loss": 1.4829, |
| "step": 5710 |
| }, |
| { |
| "epoch": 0.6008403361344538, |
| "grad_norm": 5.4375, |
| "learning_rate": 1.787245696400626e-05, |
| "loss": 1.4376, |
| "step": 5720 |
| }, |
| { |
| "epoch": 0.601890756302521, |
| "grad_norm": 14.3125, |
| "learning_rate": 1.7864632237871676e-05, |
| "loss": 1.6832, |
| "step": 5730 |
| }, |
| { |
| "epoch": 0.6029411764705882, |
| "grad_norm": 16.5, |
| "learning_rate": 1.785680751173709e-05, |
| "loss": 1.8146, |
| "step": 5740 |
| }, |
| { |
| "epoch": 0.6039915966386554, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.7848982785602505e-05, |
| "loss": 1.3456, |
| "step": 5750 |
| }, |
| { |
| "epoch": 0.6050420168067226, |
| "grad_norm": 18.875, |
| "learning_rate": 1.784115805946792e-05, |
| "loss": 1.2098, |
| "step": 5760 |
| }, |
| { |
| "epoch": 0.6060924369747899, |
| "grad_norm": 13.0, |
| "learning_rate": 1.7833333333333334e-05, |
| "loss": 1.2479, |
| "step": 5770 |
| }, |
| { |
| "epoch": 0.6071428571428571, |
| "grad_norm": 11.125, |
| "learning_rate": 1.782550860719875e-05, |
| "loss": 1.1653, |
| "step": 5780 |
| }, |
| { |
| "epoch": 0.6081932773109243, |
| "grad_norm": 13.25, |
| "learning_rate": 1.7817683881064164e-05, |
| "loss": 1.8153, |
| "step": 5790 |
| }, |
| { |
| "epoch": 0.6092436974789915, |
| "grad_norm": 14.375, |
| "learning_rate": 1.7809859154929578e-05, |
| "loss": 1.6105, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.6102941176470589, |
| "grad_norm": 27.5, |
| "learning_rate": 1.7802034428794993e-05, |
| "loss": 1.4028, |
| "step": 5810 |
| }, |
| { |
| "epoch": 0.6113445378151261, |
| "grad_norm": 4.125, |
| "learning_rate": 1.7794209702660408e-05, |
| "loss": 1.2498, |
| "step": 5820 |
| }, |
| { |
| "epoch": 0.6123949579831933, |
| "grad_norm": 11.625, |
| "learning_rate": 1.7786384976525822e-05, |
| "loss": 0.9192, |
| "step": 5830 |
| }, |
| { |
| "epoch": 0.6134453781512605, |
| "grad_norm": 18.75, |
| "learning_rate": 1.7778560250391237e-05, |
| "loss": 1.1645, |
| "step": 5840 |
| }, |
| { |
| "epoch": 0.6144957983193278, |
| "grad_norm": 15.3125, |
| "learning_rate": 1.777073552425665e-05, |
| "loss": 1.9024, |
| "step": 5850 |
| }, |
| { |
| "epoch": 0.615546218487395, |
| "grad_norm": 15.25, |
| "learning_rate": 1.7762910798122066e-05, |
| "loss": 1.5987, |
| "step": 5860 |
| }, |
| { |
| "epoch": 0.6165966386554622, |
| "grad_norm": 13.5625, |
| "learning_rate": 1.775508607198748e-05, |
| "loss": 1.3094, |
| "step": 5870 |
| }, |
| { |
| "epoch": 0.6176470588235294, |
| "grad_norm": 3.625, |
| "learning_rate": 1.77472613458529e-05, |
| "loss": 1.7913, |
| "step": 5880 |
| }, |
| { |
| "epoch": 0.6186974789915967, |
| "grad_norm": 17.0, |
| "learning_rate": 1.773943661971831e-05, |
| "loss": 1.3136, |
| "step": 5890 |
| }, |
| { |
| "epoch": 0.6197478991596639, |
| "grad_norm": 20.625, |
| "learning_rate": 1.7731611893583728e-05, |
| "loss": 1.5556, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.6207983193277311, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.772378716744914e-05, |
| "loss": 0.7191, |
| "step": 5910 |
| }, |
| { |
| "epoch": 0.6218487394957983, |
| "grad_norm": 13.5, |
| "learning_rate": 1.7715962441314557e-05, |
| "loss": 1.0429, |
| "step": 5920 |
| }, |
| { |
| "epoch": 0.6228991596638656, |
| "grad_norm": 13.5625, |
| "learning_rate": 1.770813771517997e-05, |
| "loss": 1.7025, |
| "step": 5930 |
| }, |
| { |
| "epoch": 0.6239495798319328, |
| "grad_norm": 19.75, |
| "learning_rate": 1.7700312989045387e-05, |
| "loss": 1.5732, |
| "step": 5940 |
| }, |
| { |
| "epoch": 0.625, |
| "grad_norm": 15.8125, |
| "learning_rate": 1.7692488262910798e-05, |
| "loss": 1.5835, |
| "step": 5950 |
| }, |
| { |
| "epoch": 0.6260504201680672, |
| "grad_norm": 15.4375, |
| "learning_rate": 1.7684663536776213e-05, |
| "loss": 1.7755, |
| "step": 5960 |
| }, |
| { |
| "epoch": 0.6271008403361344, |
| "grad_norm": 17.75, |
| "learning_rate": 1.767683881064163e-05, |
| "loss": 1.5965, |
| "step": 5970 |
| }, |
| { |
| "epoch": 0.6281512605042017, |
| "grad_norm": 4.5625, |
| "learning_rate": 1.7669014084507042e-05, |
| "loss": 1.0207, |
| "step": 5980 |
| }, |
| { |
| "epoch": 0.6292016806722689, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.766118935837246e-05, |
| "loss": 1.4202, |
| "step": 5990 |
| }, |
| { |
| "epoch": 0.6302521008403361, |
| "grad_norm": 14.5, |
| "learning_rate": 1.765336463223787e-05, |
| "loss": 1.0423, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.6313025210084033, |
| "grad_norm": 15.3125, |
| "learning_rate": 1.764553990610329e-05, |
| "loss": 2.0845, |
| "step": 6010 |
| }, |
| { |
| "epoch": 0.6323529411764706, |
| "grad_norm": 4.8125, |
| "learning_rate": 1.76377151799687e-05, |
| "loss": 1.4338, |
| "step": 6020 |
| }, |
| { |
| "epoch": 0.6334033613445378, |
| "grad_norm": 14.6875, |
| "learning_rate": 1.762989045383412e-05, |
| "loss": 1.3871, |
| "step": 6030 |
| }, |
| { |
| "epoch": 0.634453781512605, |
| "grad_norm": 14.25, |
| "learning_rate": 1.762206572769953e-05, |
| "loss": 1.7668, |
| "step": 6040 |
| }, |
| { |
| "epoch": 0.6355042016806722, |
| "grad_norm": 18.25, |
| "learning_rate": 1.7614241001564948e-05, |
| "loss": 1.7259, |
| "step": 6050 |
| }, |
| { |
| "epoch": 0.6365546218487395, |
| "grad_norm": 23.5, |
| "learning_rate": 1.7606416275430363e-05, |
| "loss": 0.9416, |
| "step": 6060 |
| }, |
| { |
| "epoch": 0.6376050420168067, |
| "grad_norm": 15.375, |
| "learning_rate": 1.7598591549295777e-05, |
| "loss": 1.9263, |
| "step": 6070 |
| }, |
| { |
| "epoch": 0.6386554621848739, |
| "grad_norm": 11.625, |
| "learning_rate": 1.7590766823161192e-05, |
| "loss": 1.0191, |
| "step": 6080 |
| }, |
| { |
| "epoch": 0.6397058823529411, |
| "grad_norm": 15.9375, |
| "learning_rate": 1.7582942097026607e-05, |
| "loss": 1.8691, |
| "step": 6090 |
| }, |
| { |
| "epoch": 0.6407563025210085, |
| "grad_norm": 4.90625, |
| "learning_rate": 1.757511737089202e-05, |
| "loss": 1.4114, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.6418067226890757, |
| "grad_norm": 24.0, |
| "learning_rate": 1.7567292644757436e-05, |
| "loss": 1.3372, |
| "step": 6110 |
| }, |
| { |
| "epoch": 0.6428571428571429, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.755946791862285e-05, |
| "loss": 1.2139, |
| "step": 6120 |
| }, |
| { |
| "epoch": 0.6439075630252101, |
| "grad_norm": 11.625, |
| "learning_rate": 1.7551643192488265e-05, |
| "loss": 1.0369, |
| "step": 6130 |
| }, |
| { |
| "epoch": 0.6449579831932774, |
| "grad_norm": 6.65625, |
| "learning_rate": 1.754381846635368e-05, |
| "loss": 1.2482, |
| "step": 6140 |
| }, |
| { |
| "epoch": 0.6460084033613446, |
| "grad_norm": 13.1875, |
| "learning_rate": 1.753599374021909e-05, |
| "loss": 1.4508, |
| "step": 6150 |
| }, |
| { |
| "epoch": 0.6470588235294118, |
| "grad_norm": 23.125, |
| "learning_rate": 1.752816901408451e-05, |
| "loss": 1.4622, |
| "step": 6160 |
| }, |
| { |
| "epoch": 0.648109243697479, |
| "grad_norm": 5.875, |
| "learning_rate": 1.7520344287949924e-05, |
| "loss": 1.6703, |
| "step": 6170 |
| }, |
| { |
| "epoch": 0.6491596638655462, |
| "grad_norm": 10.5, |
| "learning_rate": 1.751251956181534e-05, |
| "loss": 1.0821, |
| "step": 6180 |
| }, |
| { |
| "epoch": 0.6502100840336135, |
| "grad_norm": 6.65625, |
| "learning_rate": 1.7504694835680753e-05, |
| "loss": 1.4789, |
| "step": 6190 |
| }, |
| { |
| "epoch": 0.6512605042016807, |
| "grad_norm": 13.5625, |
| "learning_rate": 1.7496870109546168e-05, |
| "loss": 1.2124, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.6523109243697479, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.7489045383411582e-05, |
| "loss": 1.7802, |
| "step": 6210 |
| }, |
| { |
| "epoch": 0.6533613445378151, |
| "grad_norm": 16.5, |
| "learning_rate": 1.7481220657276997e-05, |
| "loss": 1.644, |
| "step": 6220 |
| }, |
| { |
| "epoch": 0.6544117647058824, |
| "grad_norm": 28.5, |
| "learning_rate": 1.747339593114241e-05, |
| "loss": 1.4378, |
| "step": 6230 |
| }, |
| { |
| "epoch": 0.6554621848739496, |
| "grad_norm": 3.40625, |
| "learning_rate": 1.7465571205007826e-05, |
| "loss": 1.4979, |
| "step": 6240 |
| }, |
| { |
| "epoch": 0.6565126050420168, |
| "grad_norm": 17.375, |
| "learning_rate": 1.745774647887324e-05, |
| "loss": 1.7841, |
| "step": 6250 |
| }, |
| { |
| "epoch": 0.657563025210084, |
| "grad_norm": 10.3125, |
| "learning_rate": 1.7449921752738656e-05, |
| "loss": 1.1316, |
| "step": 6260 |
| }, |
| { |
| "epoch": 0.6586134453781513, |
| "grad_norm": 5.375, |
| "learning_rate": 1.744209702660407e-05, |
| "loss": 1.0144, |
| "step": 6270 |
| }, |
| { |
| "epoch": 0.6596638655462185, |
| "grad_norm": 13.5, |
| "learning_rate": 1.7434272300469485e-05, |
| "loss": 1.1497, |
| "step": 6280 |
| }, |
| { |
| "epoch": 0.6607142857142857, |
| "grad_norm": 12.5, |
| "learning_rate": 1.74264475743349e-05, |
| "loss": 1.2047, |
| "step": 6290 |
| }, |
| { |
| "epoch": 0.6617647058823529, |
| "grad_norm": 16.25, |
| "learning_rate": 1.7418622848200314e-05, |
| "loss": 1.4126, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.6628151260504201, |
| "grad_norm": 15.1875, |
| "learning_rate": 1.741079812206573e-05, |
| "loss": 1.5036, |
| "step": 6310 |
| }, |
| { |
| "epoch": 0.6638655462184874, |
| "grad_norm": 12.5, |
| "learning_rate": 1.7402973395931144e-05, |
| "loss": 1.7585, |
| "step": 6320 |
| }, |
| { |
| "epoch": 0.6649159663865546, |
| "grad_norm": 13.0625, |
| "learning_rate": 1.7395148669796558e-05, |
| "loss": 1.359, |
| "step": 6330 |
| }, |
| { |
| "epoch": 0.6659663865546218, |
| "grad_norm": 16.75, |
| "learning_rate": 1.7387323943661973e-05, |
| "loss": 1.2548, |
| "step": 6340 |
| }, |
| { |
| "epoch": 0.667016806722689, |
| "grad_norm": 15.625, |
| "learning_rate": 1.7379499217527388e-05, |
| "loss": 1.4111, |
| "step": 6350 |
| }, |
| { |
| "epoch": 0.6680672268907563, |
| "grad_norm": 9.9375, |
| "learning_rate": 1.7371674491392802e-05, |
| "loss": 1.1517, |
| "step": 6360 |
| }, |
| { |
| "epoch": 0.6691176470588235, |
| "grad_norm": 13.6875, |
| "learning_rate": 1.7363849765258217e-05, |
| "loss": 1.2131, |
| "step": 6370 |
| }, |
| { |
| "epoch": 0.6701680672268907, |
| "grad_norm": 6.75, |
| "learning_rate": 1.735602503912363e-05, |
| "loss": 1.3011, |
| "step": 6380 |
| }, |
| { |
| "epoch": 0.6712184873949579, |
| "grad_norm": 14.8125, |
| "learning_rate": 1.7348200312989046e-05, |
| "loss": 1.2044, |
| "step": 6390 |
| }, |
| { |
| "epoch": 0.6722689075630253, |
| "grad_norm": 3.75, |
| "learning_rate": 1.734037558685446e-05, |
| "loss": 1.1038, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.6733193277310925, |
| "grad_norm": 11.3125, |
| "learning_rate": 1.7332550860719875e-05, |
| "loss": 1.6009, |
| "step": 6410 |
| }, |
| { |
| "epoch": 0.6743697478991597, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.732472613458529e-05, |
| "loss": 1.4377, |
| "step": 6420 |
| }, |
| { |
| "epoch": 0.6754201680672269, |
| "grad_norm": 20.0, |
| "learning_rate": 1.7316901408450705e-05, |
| "loss": 1.5643, |
| "step": 6430 |
| }, |
| { |
| "epoch": 0.6764705882352942, |
| "grad_norm": 5.59375, |
| "learning_rate": 1.730907668231612e-05, |
| "loss": 1.5667, |
| "step": 6440 |
| }, |
| { |
| "epoch": 0.6775210084033614, |
| "grad_norm": 12.25, |
| "learning_rate": 1.7301251956181534e-05, |
| "loss": 1.1483, |
| "step": 6450 |
| }, |
| { |
| "epoch": 0.6785714285714286, |
| "grad_norm": 12.375, |
| "learning_rate": 1.7293427230046952e-05, |
| "loss": 1.3863, |
| "step": 6460 |
| }, |
| { |
| "epoch": 0.6796218487394958, |
| "grad_norm": 11.25, |
| "learning_rate": 1.7285602503912363e-05, |
| "loss": 1.1266, |
| "step": 6470 |
| }, |
| { |
| "epoch": 0.680672268907563, |
| "grad_norm": 17.25, |
| "learning_rate": 1.727777777777778e-05, |
| "loss": 1.649, |
| "step": 6480 |
| }, |
| { |
| "epoch": 0.6817226890756303, |
| "grad_norm": 16.25, |
| "learning_rate": 1.7269953051643193e-05, |
| "loss": 1.582, |
| "step": 6490 |
| }, |
| { |
| "epoch": 0.6827731092436975, |
| "grad_norm": 29.0, |
| "learning_rate": 1.726212832550861e-05, |
| "loss": 1.6732, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.6838235294117647, |
| "grad_norm": 10.25, |
| "learning_rate": 1.7254303599374022e-05, |
| "loss": 1.4154, |
| "step": 6510 |
| }, |
| { |
| "epoch": 0.6848739495798319, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.724647887323944e-05, |
| "loss": 1.5973, |
| "step": 6520 |
| }, |
| { |
| "epoch": 0.6859243697478992, |
| "grad_norm": 13.4375, |
| "learning_rate": 1.723865414710485e-05, |
| "loss": 1.9317, |
| "step": 6530 |
| }, |
| { |
| "epoch": 0.6869747899159664, |
| "grad_norm": 11.5625, |
| "learning_rate": 1.7230829420970266e-05, |
| "loss": 1.6537, |
| "step": 6540 |
| }, |
| { |
| "epoch": 0.6880252100840336, |
| "grad_norm": 10.0625, |
| "learning_rate": 1.7223004694835684e-05, |
| "loss": 1.3775, |
| "step": 6550 |
| }, |
| { |
| "epoch": 0.6890756302521008, |
| "grad_norm": 11.9375, |
| "learning_rate": 1.7215179968701095e-05, |
| "loss": 1.455, |
| "step": 6560 |
| }, |
| { |
| "epoch": 0.6901260504201681, |
| "grad_norm": 21.0, |
| "learning_rate": 1.7207355242566513e-05, |
| "loss": 1.2449, |
| "step": 6570 |
| }, |
| { |
| "epoch": 0.6911764705882353, |
| "grad_norm": 16.5, |
| "learning_rate": 1.7199530516431925e-05, |
| "loss": 1.5573, |
| "step": 6580 |
| }, |
| { |
| "epoch": 0.6922268907563025, |
| "grad_norm": 11.25, |
| "learning_rate": 1.7191705790297343e-05, |
| "loss": 1.4817, |
| "step": 6590 |
| }, |
| { |
| "epoch": 0.6932773109243697, |
| "grad_norm": 13.5, |
| "learning_rate": 1.7183881064162754e-05, |
| "loss": 1.333, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.694327731092437, |
| "grad_norm": 17.75, |
| "learning_rate": 1.7176056338028172e-05, |
| "loss": 1.5649, |
| "step": 6610 |
| }, |
| { |
| "epoch": 0.6953781512605042, |
| "grad_norm": 4.75, |
| "learning_rate": 1.7168231611893583e-05, |
| "loss": 0.8342, |
| "step": 6620 |
| }, |
| { |
| "epoch": 0.6964285714285714, |
| "grad_norm": 16.0, |
| "learning_rate": 1.7160406885759e-05, |
| "loss": 1.4425, |
| "step": 6630 |
| }, |
| { |
| "epoch": 0.6974789915966386, |
| "grad_norm": 11.9375, |
| "learning_rate": 1.7152582159624412e-05, |
| "loss": 1.434, |
| "step": 6640 |
| }, |
| { |
| "epoch": 0.6985294117647058, |
| "grad_norm": 11.625, |
| "learning_rate": 1.714475743348983e-05, |
| "loss": 1.3348, |
| "step": 6650 |
| }, |
| { |
| "epoch": 0.6995798319327731, |
| "grad_norm": 19.875, |
| "learning_rate": 1.7136932707355245e-05, |
| "loss": 1.5359, |
| "step": 6660 |
| }, |
| { |
| "epoch": 0.7006302521008403, |
| "grad_norm": 10.125, |
| "learning_rate": 1.712910798122066e-05, |
| "loss": 1.5929, |
| "step": 6670 |
| }, |
| { |
| "epoch": 0.7016806722689075, |
| "grad_norm": 8.5, |
| "learning_rate": 1.7121283255086074e-05, |
| "loss": 1.4151, |
| "step": 6680 |
| }, |
| { |
| "epoch": 0.7027310924369747, |
| "grad_norm": 6.78125, |
| "learning_rate": 1.711345852895149e-05, |
| "loss": 1.0809, |
| "step": 6690 |
| }, |
| { |
| "epoch": 0.7037815126050421, |
| "grad_norm": 11.4375, |
| "learning_rate": 1.7105633802816904e-05, |
| "loss": 1.3458, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.7048319327731093, |
| "grad_norm": 11.4375, |
| "learning_rate": 1.709780907668232e-05, |
| "loss": 1.5675, |
| "step": 6710 |
| }, |
| { |
| "epoch": 0.7058823529411765, |
| "grad_norm": 14.9375, |
| "learning_rate": 1.7089984350547733e-05, |
| "loss": 1.9173, |
| "step": 6720 |
| }, |
| { |
| "epoch": 0.7069327731092437, |
| "grad_norm": 10.4375, |
| "learning_rate": 1.7082159624413148e-05, |
| "loss": 1.032, |
| "step": 6730 |
| }, |
| { |
| "epoch": 0.707983193277311, |
| "grad_norm": 18.5, |
| "learning_rate": 1.7074334898278562e-05, |
| "loss": 1.7023, |
| "step": 6740 |
| }, |
| { |
| "epoch": 0.7090336134453782, |
| "grad_norm": 15.9375, |
| "learning_rate": 1.7066510172143977e-05, |
| "loss": 1.3879, |
| "step": 6750 |
| }, |
| { |
| "epoch": 0.7100840336134454, |
| "grad_norm": 6.46875, |
| "learning_rate": 1.705868544600939e-05, |
| "loss": 1.8452, |
| "step": 6760 |
| }, |
| { |
| "epoch": 0.7111344537815126, |
| "grad_norm": 15.0, |
| "learning_rate": 1.7050860719874806e-05, |
| "loss": 1.4104, |
| "step": 6770 |
| }, |
| { |
| "epoch": 0.7121848739495799, |
| "grad_norm": 16.625, |
| "learning_rate": 1.704303599374022e-05, |
| "loss": 1.3641, |
| "step": 6780 |
| }, |
| { |
| "epoch": 0.7132352941176471, |
| "grad_norm": 17.25, |
| "learning_rate": 1.7035211267605636e-05, |
| "loss": 1.4804, |
| "step": 6790 |
| }, |
| { |
| "epoch": 0.7142857142857143, |
| "grad_norm": 27.375, |
| "learning_rate": 1.702738654147105e-05, |
| "loss": 1.9552, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.7153361344537815, |
| "grad_norm": 4.21875, |
| "learning_rate": 1.7019561815336465e-05, |
| "loss": 1.0423, |
| "step": 6810 |
| }, |
| { |
| "epoch": 0.7163865546218487, |
| "grad_norm": 10.9375, |
| "learning_rate": 1.701173708920188e-05, |
| "loss": 1.4615, |
| "step": 6820 |
| }, |
| { |
| "epoch": 0.717436974789916, |
| "grad_norm": 16.5, |
| "learning_rate": 1.7003912363067294e-05, |
| "loss": 1.3719, |
| "step": 6830 |
| }, |
| { |
| "epoch": 0.7184873949579832, |
| "grad_norm": 13.8125, |
| "learning_rate": 1.699608763693271e-05, |
| "loss": 2.0062, |
| "step": 6840 |
| }, |
| { |
| "epoch": 0.7195378151260504, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.6988262910798124e-05, |
| "loss": 1.6726, |
| "step": 6850 |
| }, |
| { |
| "epoch": 0.7205882352941176, |
| "grad_norm": 11.0, |
| "learning_rate": 1.6980438184663538e-05, |
| "loss": 1.4726, |
| "step": 6860 |
| }, |
| { |
| "epoch": 0.7216386554621849, |
| "grad_norm": 12.625, |
| "learning_rate": 1.6972613458528953e-05, |
| "loss": 1.3042, |
| "step": 6870 |
| }, |
| { |
| "epoch": 0.7226890756302521, |
| "grad_norm": 11.75, |
| "learning_rate": 1.6964788732394367e-05, |
| "loss": 1.5685, |
| "step": 6880 |
| }, |
| { |
| "epoch": 0.7237394957983193, |
| "grad_norm": 5.8125, |
| "learning_rate": 1.6956964006259782e-05, |
| "loss": 1.4803, |
| "step": 6890 |
| }, |
| { |
| "epoch": 0.7247899159663865, |
| "grad_norm": 14.1875, |
| "learning_rate": 1.6949139280125197e-05, |
| "loss": 1.1679, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.7258403361344538, |
| "grad_norm": 14.875, |
| "learning_rate": 1.694131455399061e-05, |
| "loss": 1.4983, |
| "step": 6910 |
| }, |
| { |
| "epoch": 0.726890756302521, |
| "grad_norm": 8.8125, |
| "learning_rate": 1.6933489827856026e-05, |
| "loss": 1.2823, |
| "step": 6920 |
| }, |
| { |
| "epoch": 0.7279411764705882, |
| "grad_norm": 9.375, |
| "learning_rate": 1.692566510172144e-05, |
| "loss": 1.3766, |
| "step": 6930 |
| }, |
| { |
| "epoch": 0.7289915966386554, |
| "grad_norm": 13.875, |
| "learning_rate": 1.6917840375586855e-05, |
| "loss": 1.4109, |
| "step": 6940 |
| }, |
| { |
| "epoch": 0.7300420168067226, |
| "grad_norm": 13.375, |
| "learning_rate": 1.691001564945227e-05, |
| "loss": 1.5843, |
| "step": 6950 |
| }, |
| { |
| "epoch": 0.7310924369747899, |
| "grad_norm": 13.1875, |
| "learning_rate": 1.6902190923317685e-05, |
| "loss": 1.2523, |
| "step": 6960 |
| }, |
| { |
| "epoch": 0.7321428571428571, |
| "grad_norm": 24.125, |
| "learning_rate": 1.68943661971831e-05, |
| "loss": 1.1301, |
| "step": 6970 |
| }, |
| { |
| "epoch": 0.7331932773109243, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.6886541471048514e-05, |
| "loss": 1.4351, |
| "step": 6980 |
| }, |
| { |
| "epoch": 0.7342436974789915, |
| "grad_norm": 15.6875, |
| "learning_rate": 1.687871674491393e-05, |
| "loss": 1.276, |
| "step": 6990 |
| }, |
| { |
| "epoch": 0.7352941176470589, |
| "grad_norm": 9.625, |
| "learning_rate": 1.6870892018779343e-05, |
| "loss": 1.2223, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.7363445378151261, |
| "grad_norm": 5.21875, |
| "learning_rate": 1.6863067292644758e-05, |
| "loss": 0.9195, |
| "step": 7010 |
| }, |
| { |
| "epoch": 0.7373949579831933, |
| "grad_norm": 5.4375, |
| "learning_rate": 1.6855242566510173e-05, |
| "loss": 1.5059, |
| "step": 7020 |
| }, |
| { |
| "epoch": 0.7384453781512605, |
| "grad_norm": 17.0, |
| "learning_rate": 1.6847417840375587e-05, |
| "loss": 1.2798, |
| "step": 7030 |
| }, |
| { |
| "epoch": 0.7394957983193278, |
| "grad_norm": 11.4375, |
| "learning_rate": 1.6839593114241002e-05, |
| "loss": 1.3882, |
| "step": 7040 |
| }, |
| { |
| "epoch": 0.740546218487395, |
| "grad_norm": 8.6875, |
| "learning_rate": 1.6831768388106417e-05, |
| "loss": 1.4359, |
| "step": 7050 |
| }, |
| { |
| "epoch": 0.7415966386554622, |
| "grad_norm": 19.5, |
| "learning_rate": 1.6823943661971835e-05, |
| "loss": 1.1035, |
| "step": 7060 |
| }, |
| { |
| "epoch": 0.7426470588235294, |
| "grad_norm": 21.875, |
| "learning_rate": 1.6816118935837246e-05, |
| "loss": 1.1844, |
| "step": 7070 |
| }, |
| { |
| "epoch": 0.7436974789915967, |
| "grad_norm": 13.5625, |
| "learning_rate": 1.6808294209702664e-05, |
| "loss": 1.0824, |
| "step": 7080 |
| }, |
| { |
| "epoch": 0.7447478991596639, |
| "grad_norm": 11.875, |
| "learning_rate": 1.6800469483568075e-05, |
| "loss": 1.2525, |
| "step": 7090 |
| }, |
| { |
| "epoch": 0.7457983193277311, |
| "grad_norm": 39.25, |
| "learning_rate": 1.6792644757433493e-05, |
| "loss": 1.0738, |
| "step": 7100 |
| }, |
| { |
| "epoch": 0.7468487394957983, |
| "grad_norm": 13.125, |
| "learning_rate": 1.6784820031298905e-05, |
| "loss": 1.4299, |
| "step": 7110 |
| }, |
| { |
| "epoch": 0.7478991596638656, |
| "grad_norm": 18.5, |
| "learning_rate": 1.677699530516432e-05, |
| "loss": 1.5707, |
| "step": 7120 |
| }, |
| { |
| "epoch": 0.7489495798319328, |
| "grad_norm": 18.125, |
| "learning_rate": 1.6769170579029734e-05, |
| "loss": 1.5046, |
| "step": 7130 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 21.5, |
| "learning_rate": 1.676134585289515e-05, |
| "loss": 0.9223, |
| "step": 7140 |
| }, |
| { |
| "epoch": 0.7510504201680672, |
| "grad_norm": 11.75, |
| "learning_rate": 1.6753521126760567e-05, |
| "loss": 1.3039, |
| "step": 7150 |
| }, |
| { |
| "epoch": 0.7521008403361344, |
| "grad_norm": 15.6875, |
| "learning_rate": 1.6745696400625978e-05, |
| "loss": 0.8591, |
| "step": 7160 |
| }, |
| { |
| "epoch": 0.7531512605042017, |
| "grad_norm": 10.375, |
| "learning_rate": 1.6737871674491396e-05, |
| "loss": 1.2332, |
| "step": 7170 |
| }, |
| { |
| "epoch": 0.7542016806722689, |
| "grad_norm": 10.3125, |
| "learning_rate": 1.6730046948356807e-05, |
| "loss": 1.2879, |
| "step": 7180 |
| }, |
| { |
| "epoch": 0.7552521008403361, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.6722222222222225e-05, |
| "loss": 1.484, |
| "step": 7190 |
| }, |
| { |
| "epoch": 0.7563025210084033, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.6714397496087636e-05, |
| "loss": 1.0901, |
| "step": 7200 |
| }, |
| { |
| "epoch": 0.7573529411764706, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.6706572769953054e-05, |
| "loss": 1.1962, |
| "step": 7210 |
| }, |
| { |
| "epoch": 0.7584033613445378, |
| "grad_norm": 4.96875, |
| "learning_rate": 1.6698748043818466e-05, |
| "loss": 0.949, |
| "step": 7220 |
| }, |
| { |
| "epoch": 0.759453781512605, |
| "grad_norm": 18.0, |
| "learning_rate": 1.6690923317683884e-05, |
| "loss": 1.31, |
| "step": 7230 |
| }, |
| { |
| "epoch": 0.7605042016806722, |
| "grad_norm": 4.28125, |
| "learning_rate": 1.66830985915493e-05, |
| "loss": 1.1472, |
| "step": 7240 |
| }, |
| { |
| "epoch": 0.7615546218487395, |
| "grad_norm": 12.3125, |
| "learning_rate": 1.6675273865414713e-05, |
| "loss": 1.94, |
| "step": 7250 |
| }, |
| { |
| "epoch": 0.7626050420168067, |
| "grad_norm": 13.75, |
| "learning_rate": 1.6667449139280128e-05, |
| "loss": 1.1797, |
| "step": 7260 |
| }, |
| { |
| "epoch": 0.7636554621848739, |
| "grad_norm": 14.8125, |
| "learning_rate": 1.6659624413145542e-05, |
| "loss": 1.0651, |
| "step": 7270 |
| }, |
| { |
| "epoch": 0.7647058823529411, |
| "grad_norm": 19.25, |
| "learning_rate": 1.6651799687010957e-05, |
| "loss": 1.6434, |
| "step": 7280 |
| }, |
| { |
| "epoch": 0.7657563025210085, |
| "grad_norm": 4.625, |
| "learning_rate": 1.664397496087637e-05, |
| "loss": 1.4819, |
| "step": 7290 |
| }, |
| { |
| "epoch": 0.7668067226890757, |
| "grad_norm": 5.96875, |
| "learning_rate": 1.6636150234741786e-05, |
| "loss": 1.1262, |
| "step": 7300 |
| }, |
| { |
| "epoch": 0.7678571428571429, |
| "grad_norm": 14.0625, |
| "learning_rate": 1.66283255086072e-05, |
| "loss": 1.3299, |
| "step": 7310 |
| }, |
| { |
| "epoch": 0.7689075630252101, |
| "grad_norm": 17.125, |
| "learning_rate": 1.6620500782472616e-05, |
| "loss": 1.4443, |
| "step": 7320 |
| }, |
| { |
| "epoch": 0.7699579831932774, |
| "grad_norm": 18.125, |
| "learning_rate": 1.6612676056338027e-05, |
| "loss": 1.3762, |
| "step": 7330 |
| }, |
| { |
| "epoch": 0.7710084033613446, |
| "grad_norm": 14.25, |
| "learning_rate": 1.6604851330203445e-05, |
| "loss": 1.3657, |
| "step": 7340 |
| }, |
| { |
| "epoch": 0.7720588235294118, |
| "grad_norm": 14.5625, |
| "learning_rate": 1.659702660406886e-05, |
| "loss": 1.2761, |
| "step": 7350 |
| }, |
| { |
| "epoch": 0.773109243697479, |
| "grad_norm": 9.5, |
| "learning_rate": 1.6589201877934274e-05, |
| "loss": 1.4405, |
| "step": 7360 |
| }, |
| { |
| "epoch": 0.7741596638655462, |
| "grad_norm": 9.9375, |
| "learning_rate": 1.658137715179969e-05, |
| "loss": 0.9356, |
| "step": 7370 |
| }, |
| { |
| "epoch": 0.7752100840336135, |
| "grad_norm": 12.375, |
| "learning_rate": 1.6573552425665104e-05, |
| "loss": 1.2672, |
| "step": 7380 |
| }, |
| { |
| "epoch": 0.7762605042016807, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.6565727699530518e-05, |
| "loss": 1.3292, |
| "step": 7390 |
| }, |
| { |
| "epoch": 0.7773109243697479, |
| "grad_norm": 17.875, |
| "learning_rate": 1.6557902973395933e-05, |
| "loss": 1.2816, |
| "step": 7400 |
| }, |
| { |
| "epoch": 0.7783613445378151, |
| "grad_norm": 14.375, |
| "learning_rate": 1.6550078247261347e-05, |
| "loss": 1.2162, |
| "step": 7410 |
| }, |
| { |
| "epoch": 0.7794117647058824, |
| "grad_norm": 20.0, |
| "learning_rate": 1.6542253521126762e-05, |
| "loss": 1.1232, |
| "step": 7420 |
| }, |
| { |
| "epoch": 0.7804621848739496, |
| "grad_norm": 12.625, |
| "learning_rate": 1.6534428794992177e-05, |
| "loss": 1.711, |
| "step": 7430 |
| }, |
| { |
| "epoch": 0.7815126050420168, |
| "grad_norm": 13.375, |
| "learning_rate": 1.652660406885759e-05, |
| "loss": 1.4417, |
| "step": 7440 |
| }, |
| { |
| "epoch": 0.782563025210084, |
| "grad_norm": 18.0, |
| "learning_rate": 1.6518779342723006e-05, |
| "loss": 1.5316, |
| "step": 7450 |
| }, |
| { |
| "epoch": 0.7836134453781513, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.651095461658842e-05, |
| "loss": 1.7166, |
| "step": 7460 |
| }, |
| { |
| "epoch": 0.7846638655462185, |
| "grad_norm": 21.625, |
| "learning_rate": 1.6503129890453835e-05, |
| "loss": 1.2015, |
| "step": 7470 |
| }, |
| { |
| "epoch": 0.7857142857142857, |
| "grad_norm": 22.5, |
| "learning_rate": 1.649530516431925e-05, |
| "loss": 1.4377, |
| "step": 7480 |
| }, |
| { |
| "epoch": 0.7867647058823529, |
| "grad_norm": 16.125, |
| "learning_rate": 1.6487480438184665e-05, |
| "loss": 1.5353, |
| "step": 7490 |
| }, |
| { |
| "epoch": 0.7878151260504201, |
| "grad_norm": 19.75, |
| "learning_rate": 1.647965571205008e-05, |
| "loss": 1.4953, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.7888655462184874, |
| "grad_norm": 16.125, |
| "learning_rate": 1.6471830985915494e-05, |
| "loss": 1.1865, |
| "step": 7510 |
| }, |
| { |
| "epoch": 0.7899159663865546, |
| "grad_norm": 7.125, |
| "learning_rate": 1.646400625978091e-05, |
| "loss": 0.8322, |
| "step": 7520 |
| }, |
| { |
| "epoch": 0.7909663865546218, |
| "grad_norm": 11.625, |
| "learning_rate": 1.6456181533646323e-05, |
| "loss": 1.4808, |
| "step": 7530 |
| }, |
| { |
| "epoch": 0.792016806722689, |
| "grad_norm": 14.875, |
| "learning_rate": 1.6448356807511738e-05, |
| "loss": 1.7362, |
| "step": 7540 |
| }, |
| { |
| "epoch": 0.7930672268907563, |
| "grad_norm": 25.75, |
| "learning_rate": 1.6440532081377153e-05, |
| "loss": 1.4707, |
| "step": 7550 |
| }, |
| { |
| "epoch": 0.7941176470588235, |
| "grad_norm": 15.8125, |
| "learning_rate": 1.6432707355242567e-05, |
| "loss": 1.3344, |
| "step": 7560 |
| }, |
| { |
| "epoch": 0.7951680672268907, |
| "grad_norm": 6.5625, |
| "learning_rate": 1.6424882629107982e-05, |
| "loss": 1.3918, |
| "step": 7570 |
| }, |
| { |
| "epoch": 0.7962184873949579, |
| "grad_norm": 10.0, |
| "learning_rate": 1.6417057902973397e-05, |
| "loss": 1.3308, |
| "step": 7580 |
| }, |
| { |
| "epoch": 0.7972689075630253, |
| "grad_norm": 26.625, |
| "learning_rate": 1.640923317683881e-05, |
| "loss": 1.3687, |
| "step": 7590 |
| }, |
| { |
| "epoch": 0.7983193277310925, |
| "grad_norm": 13.6875, |
| "learning_rate": 1.6401408450704226e-05, |
| "loss": 1.3443, |
| "step": 7600 |
| }, |
| { |
| "epoch": 0.7993697478991597, |
| "grad_norm": 4.5, |
| "learning_rate": 1.639358372456964e-05, |
| "loss": 0.9478, |
| "step": 7610 |
| }, |
| { |
| "epoch": 0.8004201680672269, |
| "grad_norm": 19.75, |
| "learning_rate": 1.6385758998435055e-05, |
| "loss": 0.9845, |
| "step": 7620 |
| }, |
| { |
| "epoch": 0.8014705882352942, |
| "grad_norm": 6.375, |
| "learning_rate": 1.637793427230047e-05, |
| "loss": 1.1467, |
| "step": 7630 |
| }, |
| { |
| "epoch": 0.8025210084033614, |
| "grad_norm": 18.5, |
| "learning_rate": 1.6370109546165888e-05, |
| "loss": 1.2671, |
| "step": 7640 |
| }, |
| { |
| "epoch": 0.8035714285714286, |
| "grad_norm": 12.8125, |
| "learning_rate": 1.63622848200313e-05, |
| "loss": 1.454, |
| "step": 7650 |
| }, |
| { |
| "epoch": 0.8046218487394958, |
| "grad_norm": 13.875, |
| "learning_rate": 1.6354460093896717e-05, |
| "loss": 1.8004, |
| "step": 7660 |
| }, |
| { |
| "epoch": 0.805672268907563, |
| "grad_norm": 10.9375, |
| "learning_rate": 1.634663536776213e-05, |
| "loss": 1.1873, |
| "step": 7670 |
| }, |
| { |
| "epoch": 0.8067226890756303, |
| "grad_norm": 17.625, |
| "learning_rate": 1.6338810641627547e-05, |
| "loss": 1.7385, |
| "step": 7680 |
| }, |
| { |
| "epoch": 0.8077731092436975, |
| "grad_norm": 13.0, |
| "learning_rate": 1.6330985915492958e-05, |
| "loss": 1.1763, |
| "step": 7690 |
| }, |
| { |
| "epoch": 0.8088235294117647, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.6323161189358376e-05, |
| "loss": 1.1858, |
| "step": 7700 |
| }, |
| { |
| "epoch": 0.8098739495798319, |
| "grad_norm": 21.75, |
| "learning_rate": 1.6315336463223787e-05, |
| "loss": 1.6386, |
| "step": 7710 |
| }, |
| { |
| "epoch": 0.8109243697478992, |
| "grad_norm": 26.5, |
| "learning_rate": 1.6307511737089202e-05, |
| "loss": 1.2271, |
| "step": 7720 |
| }, |
| { |
| "epoch": 0.8119747899159664, |
| "grad_norm": 14.0, |
| "learning_rate": 1.629968701095462e-05, |
| "loss": 1.4109, |
| "step": 7730 |
| }, |
| { |
| "epoch": 0.8130252100840336, |
| "grad_norm": 16.25, |
| "learning_rate": 1.629186228482003e-05, |
| "loss": 1.5447, |
| "step": 7740 |
| }, |
| { |
| "epoch": 0.8140756302521008, |
| "grad_norm": 13.75, |
| "learning_rate": 1.628403755868545e-05, |
| "loss": 1.4124, |
| "step": 7750 |
| }, |
| { |
| "epoch": 0.8151260504201681, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.627621283255086e-05, |
| "loss": 1.3217, |
| "step": 7760 |
| }, |
| { |
| "epoch": 0.8161764705882353, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.626838810641628e-05, |
| "loss": 1.1625, |
| "step": 7770 |
| }, |
| { |
| "epoch": 0.8172268907563025, |
| "grad_norm": 14.375, |
| "learning_rate": 1.626056338028169e-05, |
| "loss": 1.7034, |
| "step": 7780 |
| }, |
| { |
| "epoch": 0.8182773109243697, |
| "grad_norm": 3.375, |
| "learning_rate": 1.6252738654147108e-05, |
| "loss": 1.1285, |
| "step": 7790 |
| }, |
| { |
| "epoch": 0.819327731092437, |
| "grad_norm": 5.21875, |
| "learning_rate": 1.624491392801252e-05, |
| "loss": 1.305, |
| "step": 7800 |
| }, |
| { |
| "epoch": 0.8203781512605042, |
| "grad_norm": 14.125, |
| "learning_rate": 1.6237089201877937e-05, |
| "loss": 1.1519, |
| "step": 7810 |
| }, |
| { |
| "epoch": 0.8214285714285714, |
| "grad_norm": 11.125, |
| "learning_rate": 1.6229264475743348e-05, |
| "loss": 1.2225, |
| "step": 7820 |
| }, |
| { |
| "epoch": 0.8224789915966386, |
| "grad_norm": 18.625, |
| "learning_rate": 1.6221439749608766e-05, |
| "loss": 1.2256, |
| "step": 7830 |
| }, |
| { |
| "epoch": 0.8235294117647058, |
| "grad_norm": 14.1875, |
| "learning_rate": 1.621361502347418e-05, |
| "loss": 1.3576, |
| "step": 7840 |
| }, |
| { |
| "epoch": 0.8245798319327731, |
| "grad_norm": 11.5, |
| "learning_rate": 1.6205790297339596e-05, |
| "loss": 1.0199, |
| "step": 7850 |
| }, |
| { |
| "epoch": 0.8256302521008403, |
| "grad_norm": 16.375, |
| "learning_rate": 1.619796557120501e-05, |
| "loss": 1.4019, |
| "step": 7860 |
| }, |
| { |
| "epoch": 0.8266806722689075, |
| "grad_norm": 39.75, |
| "learning_rate": 1.6190140845070425e-05, |
| "loss": 1.0616, |
| "step": 7870 |
| }, |
| { |
| "epoch": 0.8277310924369747, |
| "grad_norm": 20.0, |
| "learning_rate": 1.618231611893584e-05, |
| "loss": 1.3819, |
| "step": 7880 |
| }, |
| { |
| "epoch": 0.8287815126050421, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.6174491392801254e-05, |
| "loss": 1.2698, |
| "step": 7890 |
| }, |
| { |
| "epoch": 0.8298319327731093, |
| "grad_norm": 15.0, |
| "learning_rate": 1.616666666666667e-05, |
| "loss": 1.4022, |
| "step": 7900 |
| }, |
| { |
| "epoch": 0.8308823529411765, |
| "grad_norm": 13.4375, |
| "learning_rate": 1.615884194053208e-05, |
| "loss": 1.3826, |
| "step": 7910 |
| }, |
| { |
| "epoch": 0.8319327731092437, |
| "grad_norm": 6.1875, |
| "learning_rate": 1.6151017214397498e-05, |
| "loss": 1.6889, |
| "step": 7920 |
| }, |
| { |
| "epoch": 0.832983193277311, |
| "grad_norm": 11.625, |
| "learning_rate": 1.6143192488262913e-05, |
| "loss": 1.072, |
| "step": 7930 |
| }, |
| { |
| "epoch": 0.8340336134453782, |
| "grad_norm": 11.625, |
| "learning_rate": 1.6135367762128327e-05, |
| "loss": 1.6721, |
| "step": 7940 |
| }, |
| { |
| "epoch": 0.8350840336134454, |
| "grad_norm": 28.625, |
| "learning_rate": 1.6127543035993742e-05, |
| "loss": 1.4931, |
| "step": 7950 |
| }, |
| { |
| "epoch": 0.8361344537815126, |
| "grad_norm": 9.125, |
| "learning_rate": 1.6119718309859157e-05, |
| "loss": 1.4431, |
| "step": 7960 |
| }, |
| { |
| "epoch": 0.8371848739495799, |
| "grad_norm": 4.1875, |
| "learning_rate": 1.611189358372457e-05, |
| "loss": 1.3817, |
| "step": 7970 |
| }, |
| { |
| "epoch": 0.8382352941176471, |
| "grad_norm": 16.75, |
| "learning_rate": 1.6104068857589986e-05, |
| "loss": 1.1158, |
| "step": 7980 |
| }, |
| { |
| "epoch": 0.8392857142857143, |
| "grad_norm": 3.90625, |
| "learning_rate": 1.60962441314554e-05, |
| "loss": 1.2636, |
| "step": 7990 |
| }, |
| { |
| "epoch": 0.8403361344537815, |
| "grad_norm": 13.25, |
| "learning_rate": 1.6088419405320815e-05, |
| "loss": 1.5959, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.8413865546218487, |
| "grad_norm": 37.5, |
| "learning_rate": 1.608059467918623e-05, |
| "loss": 1.061, |
| "step": 8010 |
| }, |
| { |
| "epoch": 0.842436974789916, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.6072769953051645e-05, |
| "loss": 1.6377, |
| "step": 8020 |
| }, |
| { |
| "epoch": 0.8434873949579832, |
| "grad_norm": 16.5, |
| "learning_rate": 1.606494522691706e-05, |
| "loss": 1.2668, |
| "step": 8030 |
| }, |
| { |
| "epoch": 0.8445378151260504, |
| "grad_norm": 17.125, |
| "learning_rate": 1.6057120500782474e-05, |
| "loss": 1.4443, |
| "step": 8040 |
| }, |
| { |
| "epoch": 0.8455882352941176, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.604929577464789e-05, |
| "loss": 1.3508, |
| "step": 8050 |
| }, |
| { |
| "epoch": 0.8466386554621849, |
| "grad_norm": 12.875, |
| "learning_rate": 1.6041471048513303e-05, |
| "loss": 1.1882, |
| "step": 8060 |
| }, |
| { |
| "epoch": 0.8476890756302521, |
| "grad_norm": 7.03125, |
| "learning_rate": 1.6033646322378718e-05, |
| "loss": 0.9828, |
| "step": 8070 |
| }, |
| { |
| "epoch": 0.8487394957983193, |
| "grad_norm": 19.5, |
| "learning_rate": 1.6025821596244133e-05, |
| "loss": 1.7828, |
| "step": 8080 |
| }, |
| { |
| "epoch": 0.8497899159663865, |
| "grad_norm": 18.0, |
| "learning_rate": 1.6017996870109547e-05, |
| "loss": 1.5972, |
| "step": 8090 |
| }, |
| { |
| "epoch": 0.8508403361344538, |
| "grad_norm": 21.0, |
| "learning_rate": 1.6010172143974962e-05, |
| "loss": 1.6549, |
| "step": 8100 |
| }, |
| { |
| "epoch": 0.851890756302521, |
| "grad_norm": 3.453125, |
| "learning_rate": 1.6002347417840377e-05, |
| "loss": 1.634, |
| "step": 8110 |
| }, |
| { |
| "epoch": 0.8529411764705882, |
| "grad_norm": 13.875, |
| "learning_rate": 1.599452269170579e-05, |
| "loss": 1.0437, |
| "step": 8120 |
| }, |
| { |
| "epoch": 0.8539915966386554, |
| "grad_norm": 3.71875, |
| "learning_rate": 1.5986697965571206e-05, |
| "loss": 1.4985, |
| "step": 8130 |
| }, |
| { |
| "epoch": 0.8550420168067226, |
| "grad_norm": 22.75, |
| "learning_rate": 1.597887323943662e-05, |
| "loss": 1.6172, |
| "step": 8140 |
| }, |
| { |
| "epoch": 0.8560924369747899, |
| "grad_norm": 20.875, |
| "learning_rate": 1.5971048513302035e-05, |
| "loss": 1.484, |
| "step": 8150 |
| }, |
| { |
| "epoch": 0.8571428571428571, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.596322378716745e-05, |
| "loss": 1.1782, |
| "step": 8160 |
| }, |
| { |
| "epoch": 0.8581932773109243, |
| "grad_norm": 11.375, |
| "learning_rate": 1.5955399061032864e-05, |
| "loss": 1.5498, |
| "step": 8170 |
| }, |
| { |
| "epoch": 0.8592436974789915, |
| "grad_norm": 16.0, |
| "learning_rate": 1.594757433489828e-05, |
| "loss": 1.3166, |
| "step": 8180 |
| }, |
| { |
| "epoch": 0.8602941176470589, |
| "grad_norm": 15.25, |
| "learning_rate": 1.5939749608763694e-05, |
| "loss": 1.3896, |
| "step": 8190 |
| }, |
| { |
| "epoch": 0.8613445378151261, |
| "grad_norm": 13.1875, |
| "learning_rate": 1.593192488262911e-05, |
| "loss": 1.238, |
| "step": 8200 |
| }, |
| { |
| "epoch": 0.8623949579831933, |
| "grad_norm": 13.0, |
| "learning_rate": 1.5924100156494523e-05, |
| "loss": 1.4749, |
| "step": 8210 |
| }, |
| { |
| "epoch": 0.8634453781512605, |
| "grad_norm": 18.875, |
| "learning_rate": 1.5916275430359938e-05, |
| "loss": 1.3273, |
| "step": 8220 |
| }, |
| { |
| "epoch": 0.8644957983193278, |
| "grad_norm": 13.9375, |
| "learning_rate": 1.5908450704225352e-05, |
| "loss": 1.7615, |
| "step": 8230 |
| }, |
| { |
| "epoch": 0.865546218487395, |
| "grad_norm": 3.140625, |
| "learning_rate": 1.590062597809077e-05, |
| "loss": 0.9801, |
| "step": 8240 |
| }, |
| { |
| "epoch": 0.8665966386554622, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.5892801251956182e-05, |
| "loss": 1.1506, |
| "step": 8250 |
| }, |
| { |
| "epoch": 0.8676470588235294, |
| "grad_norm": 20.25, |
| "learning_rate": 1.58849765258216e-05, |
| "loss": 1.5939, |
| "step": 8260 |
| }, |
| { |
| "epoch": 0.8686974789915967, |
| "grad_norm": 7.875, |
| "learning_rate": 1.587715179968701e-05, |
| "loss": 1.2275, |
| "step": 8270 |
| }, |
| { |
| "epoch": 0.8697478991596639, |
| "grad_norm": 47.25, |
| "learning_rate": 1.586932707355243e-05, |
| "loss": 1.5948, |
| "step": 8280 |
| }, |
| { |
| "epoch": 0.8707983193277311, |
| "grad_norm": 12.5, |
| "learning_rate": 1.586150234741784e-05, |
| "loss": 1.5074, |
| "step": 8290 |
| }, |
| { |
| "epoch": 0.8718487394957983, |
| "grad_norm": 14.9375, |
| "learning_rate": 1.5853677621283255e-05, |
| "loss": 1.1553, |
| "step": 8300 |
| }, |
| { |
| "epoch": 0.8728991596638656, |
| "grad_norm": 5.0625, |
| "learning_rate": 1.584585289514867e-05, |
| "loss": 1.4229, |
| "step": 8310 |
| }, |
| { |
| "epoch": 0.8739495798319328, |
| "grad_norm": 11.1875, |
| "learning_rate": 1.5838028169014084e-05, |
| "loss": 1.1038, |
| "step": 8320 |
| }, |
| { |
| "epoch": 0.875, |
| "grad_norm": 8.5625, |
| "learning_rate": 1.5830203442879502e-05, |
| "loss": 1.5116, |
| "step": 8330 |
| }, |
| { |
| "epoch": 0.8760504201680672, |
| "grad_norm": 10.25, |
| "learning_rate": 1.5822378716744914e-05, |
| "loss": 1.5387, |
| "step": 8340 |
| }, |
| { |
| "epoch": 0.8771008403361344, |
| "grad_norm": 14.0625, |
| "learning_rate": 1.581455399061033e-05, |
| "loss": 1.3233, |
| "step": 8350 |
| }, |
| { |
| "epoch": 0.8781512605042017, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.5806729264475743e-05, |
| "loss": 1.1451, |
| "step": 8360 |
| }, |
| { |
| "epoch": 0.8792016806722689, |
| "grad_norm": 19.25, |
| "learning_rate": 1.579890453834116e-05, |
| "loss": 1.4902, |
| "step": 8370 |
| }, |
| { |
| "epoch": 0.8802521008403361, |
| "grad_norm": 11.3125, |
| "learning_rate": 1.5791079812206572e-05, |
| "loss": 1.3482, |
| "step": 8380 |
| }, |
| { |
| "epoch": 0.8813025210084033, |
| "grad_norm": 26.25, |
| "learning_rate": 1.578325508607199e-05, |
| "loss": 1.251, |
| "step": 8390 |
| }, |
| { |
| "epoch": 0.8823529411764706, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.57754303599374e-05, |
| "loss": 1.2147, |
| "step": 8400 |
| }, |
| { |
| "epoch": 0.8834033613445378, |
| "grad_norm": 5.0625, |
| "learning_rate": 1.576760563380282e-05, |
| "loss": 1.4074, |
| "step": 8410 |
| }, |
| { |
| "epoch": 0.884453781512605, |
| "grad_norm": 4.40625, |
| "learning_rate": 1.5759780907668234e-05, |
| "loss": 1.0801, |
| "step": 8420 |
| }, |
| { |
| "epoch": 0.8855042016806722, |
| "grad_norm": 5.34375, |
| "learning_rate": 1.575195618153365e-05, |
| "loss": 1.0651, |
| "step": 8430 |
| }, |
| { |
| "epoch": 0.8865546218487395, |
| "grad_norm": 12.375, |
| "learning_rate": 1.5744131455399064e-05, |
| "loss": 1.6375, |
| "step": 8440 |
| }, |
| { |
| "epoch": 0.8876050420168067, |
| "grad_norm": 27.75, |
| "learning_rate": 1.5736306729264478e-05, |
| "loss": 1.3611, |
| "step": 8450 |
| }, |
| { |
| "epoch": 0.8886554621848739, |
| "grad_norm": 9.875, |
| "learning_rate": 1.5728482003129893e-05, |
| "loss": 1.0472, |
| "step": 8460 |
| }, |
| { |
| "epoch": 0.8897058823529411, |
| "grad_norm": 16.125, |
| "learning_rate": 1.5720657276995307e-05, |
| "loss": 1.3257, |
| "step": 8470 |
| }, |
| { |
| "epoch": 0.8907563025210085, |
| "grad_norm": 16.375, |
| "learning_rate": 1.5712832550860722e-05, |
| "loss": 0.9886, |
| "step": 8480 |
| }, |
| { |
| "epoch": 0.8918067226890757, |
| "grad_norm": 5.1875, |
| "learning_rate": 1.5705007824726133e-05, |
| "loss": 0.8322, |
| "step": 8490 |
| }, |
| { |
| "epoch": 0.8928571428571429, |
| "grad_norm": 17.75, |
| "learning_rate": 1.569718309859155e-05, |
| "loss": 1.3795, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.8939075630252101, |
| "grad_norm": 17.5, |
| "learning_rate": 1.5689358372456963e-05, |
| "loss": 1.6337, |
| "step": 8510 |
| }, |
| { |
| "epoch": 0.8949579831932774, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.568153364632238e-05, |
| "loss": 1.7422, |
| "step": 8520 |
| }, |
| { |
| "epoch": 0.8960084033613446, |
| "grad_norm": 16.75, |
| "learning_rate": 1.5673708920187795e-05, |
| "loss": 1.7619, |
| "step": 8530 |
| }, |
| { |
| "epoch": 0.8970588235294118, |
| "grad_norm": 13.0625, |
| "learning_rate": 1.566588419405321e-05, |
| "loss": 1.4839, |
| "step": 8540 |
| }, |
| { |
| "epoch": 0.898109243697479, |
| "grad_norm": 16.625, |
| "learning_rate": 1.5658059467918625e-05, |
| "loss": 1.3422, |
| "step": 8550 |
| }, |
| { |
| "epoch": 0.8991596638655462, |
| "grad_norm": 12.875, |
| "learning_rate": 1.565023474178404e-05, |
| "loss": 1.5977, |
| "step": 8560 |
| }, |
| { |
| "epoch": 0.9002100840336135, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.5642410015649454e-05, |
| "loss": 1.3554, |
| "step": 8570 |
| }, |
| { |
| "epoch": 0.9012605042016807, |
| "grad_norm": 16.75, |
| "learning_rate": 1.563458528951487e-05, |
| "loss": 1.1915, |
| "step": 8580 |
| }, |
| { |
| "epoch": 0.9023109243697479, |
| "grad_norm": 12.0, |
| "learning_rate": 1.5626760563380283e-05, |
| "loss": 1.3909, |
| "step": 8590 |
| }, |
| { |
| "epoch": 0.9033613445378151, |
| "grad_norm": 12.75, |
| "learning_rate": 1.5618935837245698e-05, |
| "loss": 1.2245, |
| "step": 8600 |
| }, |
| { |
| "epoch": 0.9044117647058824, |
| "grad_norm": 22.375, |
| "learning_rate": 1.5611111111111113e-05, |
| "loss": 1.3636, |
| "step": 8610 |
| }, |
| { |
| "epoch": 0.9054621848739496, |
| "grad_norm": 11.5625, |
| "learning_rate": 1.5603286384976527e-05, |
| "loss": 1.2219, |
| "step": 8620 |
| }, |
| { |
| "epoch": 0.9065126050420168, |
| "grad_norm": 15.875, |
| "learning_rate": 1.5595461658841942e-05, |
| "loss": 1.4637, |
| "step": 8630 |
| }, |
| { |
| "epoch": 0.907563025210084, |
| "grad_norm": 8.125, |
| "learning_rate": 1.5587636932707357e-05, |
| "loss": 1.3819, |
| "step": 8640 |
| }, |
| { |
| "epoch": 0.9086134453781513, |
| "grad_norm": 20.75, |
| "learning_rate": 1.557981220657277e-05, |
| "loss": 1.5721, |
| "step": 8650 |
| }, |
| { |
| "epoch": 0.9096638655462185, |
| "grad_norm": 3.890625, |
| "learning_rate": 1.5571987480438186e-05, |
| "loss": 1.3215, |
| "step": 8660 |
| }, |
| { |
| "epoch": 0.9107142857142857, |
| "grad_norm": 13.375, |
| "learning_rate": 1.55641627543036e-05, |
| "loss": 1.4286, |
| "step": 8670 |
| }, |
| { |
| "epoch": 0.9117647058823529, |
| "grad_norm": 12.5, |
| "learning_rate": 1.5556338028169015e-05, |
| "loss": 1.0645, |
| "step": 8680 |
| }, |
| { |
| "epoch": 0.9128151260504201, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.554851330203443e-05, |
| "loss": 1.1888, |
| "step": 8690 |
| }, |
| { |
| "epoch": 0.9138655462184874, |
| "grad_norm": 14.5, |
| "learning_rate": 1.5540688575899844e-05, |
| "loss": 1.6634, |
| "step": 8700 |
| }, |
| { |
| "epoch": 0.9149159663865546, |
| "grad_norm": 18.25, |
| "learning_rate": 1.553286384976526e-05, |
| "loss": 1.4643, |
| "step": 8710 |
| }, |
| { |
| "epoch": 0.9159663865546218, |
| "grad_norm": 4.8125, |
| "learning_rate": 1.5525039123630674e-05, |
| "loss": 1.4156, |
| "step": 8720 |
| }, |
| { |
| "epoch": 0.917016806722689, |
| "grad_norm": 11.75, |
| "learning_rate": 1.551721439749609e-05, |
| "loss": 1.5662, |
| "step": 8730 |
| }, |
| { |
| "epoch": 0.9180672268907563, |
| "grad_norm": 19.5, |
| "learning_rate": 1.5509389671361503e-05, |
| "loss": 1.1295, |
| "step": 8740 |
| }, |
| { |
| "epoch": 0.9191176470588235, |
| "grad_norm": 9.25, |
| "learning_rate": 1.5501564945226918e-05, |
| "loss": 1.422, |
| "step": 8750 |
| }, |
| { |
| "epoch": 0.9201680672268907, |
| "grad_norm": 10.4375, |
| "learning_rate": 1.5493740219092332e-05, |
| "loss": 0.6702, |
| "step": 8760 |
| }, |
| { |
| "epoch": 0.9212184873949579, |
| "grad_norm": 15.6875, |
| "learning_rate": 1.5485915492957747e-05, |
| "loss": 1.3917, |
| "step": 8770 |
| }, |
| { |
| "epoch": 0.9222689075630253, |
| "grad_norm": 11.625, |
| "learning_rate": 1.5478090766823162e-05, |
| "loss": 1.3655, |
| "step": 8780 |
| }, |
| { |
| "epoch": 0.9233193277310925, |
| "grad_norm": 15.25, |
| "learning_rate": 1.5470266040688576e-05, |
| "loss": 1.3897, |
| "step": 8790 |
| }, |
| { |
| "epoch": 0.9243697478991597, |
| "grad_norm": 20.75, |
| "learning_rate": 1.546244131455399e-05, |
| "loss": 1.2722, |
| "step": 8800 |
| }, |
| { |
| "epoch": 0.9254201680672269, |
| "grad_norm": 13.875, |
| "learning_rate": 1.5454616588419406e-05, |
| "loss": 1.2772, |
| "step": 8810 |
| }, |
| { |
| "epoch": 0.9264705882352942, |
| "grad_norm": 33.75, |
| "learning_rate": 1.5446791862284824e-05, |
| "loss": 1.3379, |
| "step": 8820 |
| }, |
| { |
| "epoch": 0.9275210084033614, |
| "grad_norm": 14.8125, |
| "learning_rate": 1.5438967136150235e-05, |
| "loss": 1.0489, |
| "step": 8830 |
| }, |
| { |
| "epoch": 0.9285714285714286, |
| "grad_norm": 13.375, |
| "learning_rate": 1.5431142410015653e-05, |
| "loss": 0.8453, |
| "step": 8840 |
| }, |
| { |
| "epoch": 0.9296218487394958, |
| "grad_norm": 11.5, |
| "learning_rate": 1.5423317683881064e-05, |
| "loss": 1.3735, |
| "step": 8850 |
| }, |
| { |
| "epoch": 0.930672268907563, |
| "grad_norm": 18.25, |
| "learning_rate": 1.5415492957746482e-05, |
| "loss": 1.5835, |
| "step": 8860 |
| }, |
| { |
| "epoch": 0.9317226890756303, |
| "grad_norm": 10.8125, |
| "learning_rate": 1.5407668231611894e-05, |
| "loss": 1.5527, |
| "step": 8870 |
| }, |
| { |
| "epoch": 0.9327731092436975, |
| "grad_norm": 3.96875, |
| "learning_rate": 1.5399843505477308e-05, |
| "loss": 1.0899, |
| "step": 8880 |
| }, |
| { |
| "epoch": 0.9338235294117647, |
| "grad_norm": 15.8125, |
| "learning_rate": 1.5392018779342723e-05, |
| "loss": 1.9109, |
| "step": 8890 |
| }, |
| { |
| "epoch": 0.9348739495798319, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.5384194053208138e-05, |
| "loss": 1.3079, |
| "step": 8900 |
| }, |
| { |
| "epoch": 0.9359243697478992, |
| "grad_norm": 11.625, |
| "learning_rate": 1.5376369327073556e-05, |
| "loss": 1.4786, |
| "step": 8910 |
| }, |
| { |
| "epoch": 0.9369747899159664, |
| "grad_norm": 16.25, |
| "learning_rate": 1.5368544600938967e-05, |
| "loss": 1.5381, |
| "step": 8920 |
| }, |
| { |
| "epoch": 0.9380252100840336, |
| "grad_norm": 9.375, |
| "learning_rate": 1.5360719874804385e-05, |
| "loss": 1.1872, |
| "step": 8930 |
| }, |
| { |
| "epoch": 0.9390756302521008, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.5352895148669796e-05, |
| "loss": 1.785, |
| "step": 8940 |
| }, |
| { |
| "epoch": 0.9401260504201681, |
| "grad_norm": 4.5, |
| "learning_rate": 1.5345070422535214e-05, |
| "loss": 1.0459, |
| "step": 8950 |
| }, |
| { |
| "epoch": 0.9411764705882353, |
| "grad_norm": 25.125, |
| "learning_rate": 1.5337245696400625e-05, |
| "loss": 1.0509, |
| "step": 8960 |
| }, |
| { |
| "epoch": 0.9422268907563025, |
| "grad_norm": 18.25, |
| "learning_rate": 1.5329420970266043e-05, |
| "loss": 1.6948, |
| "step": 8970 |
| }, |
| { |
| "epoch": 0.9432773109243697, |
| "grad_norm": 21.875, |
| "learning_rate": 1.5321596244131455e-05, |
| "loss": 0.8567, |
| "step": 8980 |
| }, |
| { |
| "epoch": 0.944327731092437, |
| "grad_norm": 20.75, |
| "learning_rate": 1.5313771517996873e-05, |
| "loss": 1.8198, |
| "step": 8990 |
| }, |
| { |
| "epoch": 0.9453781512605042, |
| "grad_norm": 17.5, |
| "learning_rate": 1.5305946791862284e-05, |
| "loss": 1.6832, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.9464285714285714, |
| "grad_norm": 19.875, |
| "learning_rate": 1.5298122065727702e-05, |
| "loss": 1.2455, |
| "step": 9010 |
| }, |
| { |
| "epoch": 0.9474789915966386, |
| "grad_norm": 11.875, |
| "learning_rate": 1.5290297339593117e-05, |
| "loss": 1.3775, |
| "step": 9020 |
| }, |
| { |
| "epoch": 0.9485294117647058, |
| "grad_norm": 35.5, |
| "learning_rate": 1.528247261345853e-05, |
| "loss": 1.7105, |
| "step": 9030 |
| }, |
| { |
| "epoch": 0.9495798319327731, |
| "grad_norm": 9.0625, |
| "learning_rate": 1.5274647887323946e-05, |
| "loss": 0.8712, |
| "step": 9040 |
| }, |
| { |
| "epoch": 0.9506302521008403, |
| "grad_norm": 4.125, |
| "learning_rate": 1.526682316118936e-05, |
| "loss": 1.1452, |
| "step": 9050 |
| }, |
| { |
| "epoch": 0.9516806722689075, |
| "grad_norm": 10.125, |
| "learning_rate": 1.5258998435054775e-05, |
| "loss": 1.3654, |
| "step": 9060 |
| }, |
| { |
| "epoch": 0.9527310924369747, |
| "grad_norm": 29.5, |
| "learning_rate": 1.5251173708920188e-05, |
| "loss": 1.4345, |
| "step": 9070 |
| }, |
| { |
| "epoch": 0.9537815126050421, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.5243348982785605e-05, |
| "loss": 1.1807, |
| "step": 9080 |
| }, |
| { |
| "epoch": 0.9548319327731093, |
| "grad_norm": 14.3125, |
| "learning_rate": 1.5235524256651018e-05, |
| "loss": 1.409, |
| "step": 9090 |
| }, |
| { |
| "epoch": 0.9558823529411765, |
| "grad_norm": 13.9375, |
| "learning_rate": 1.5227699530516434e-05, |
| "loss": 1.2369, |
| "step": 9100 |
| }, |
| { |
| "epoch": 0.9569327731092437, |
| "grad_norm": 14.375, |
| "learning_rate": 1.5219874804381849e-05, |
| "loss": 1.1645, |
| "step": 9110 |
| }, |
| { |
| "epoch": 0.957983193277311, |
| "grad_norm": 15.8125, |
| "learning_rate": 1.5212050078247262e-05, |
| "loss": 1.4289, |
| "step": 9120 |
| }, |
| { |
| "epoch": 0.9590336134453782, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.5204225352112678e-05, |
| "loss": 1.4257, |
| "step": 9130 |
| }, |
| { |
| "epoch": 0.9600840336134454, |
| "grad_norm": 19.625, |
| "learning_rate": 1.5196400625978091e-05, |
| "loss": 1.1819, |
| "step": 9140 |
| }, |
| { |
| "epoch": 0.9611344537815126, |
| "grad_norm": 17.5, |
| "learning_rate": 1.5188575899843507e-05, |
| "loss": 1.4177, |
| "step": 9150 |
| }, |
| { |
| "epoch": 0.9621848739495799, |
| "grad_norm": 9.125, |
| "learning_rate": 1.518075117370892e-05, |
| "loss": 1.0235, |
| "step": 9160 |
| }, |
| { |
| "epoch": 0.9632352941176471, |
| "grad_norm": 13.125, |
| "learning_rate": 1.5172926447574337e-05, |
| "loss": 1.4327, |
| "step": 9170 |
| }, |
| { |
| "epoch": 0.9642857142857143, |
| "grad_norm": 6.46875, |
| "learning_rate": 1.516510172143975e-05, |
| "loss": 1.326, |
| "step": 9180 |
| }, |
| { |
| "epoch": 0.9653361344537815, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.5157276995305166e-05, |
| "loss": 1.7656, |
| "step": 9190 |
| }, |
| { |
| "epoch": 0.9663865546218487, |
| "grad_norm": 9.1875, |
| "learning_rate": 1.5149452269170579e-05, |
| "loss": 1.368, |
| "step": 9200 |
| }, |
| { |
| "epoch": 0.967436974789916, |
| "grad_norm": 18.25, |
| "learning_rate": 1.5141627543035995e-05, |
| "loss": 1.7504, |
| "step": 9210 |
| }, |
| { |
| "epoch": 0.9684873949579832, |
| "grad_norm": 12.0, |
| "learning_rate": 1.513380281690141e-05, |
| "loss": 1.2685, |
| "step": 9220 |
| }, |
| { |
| "epoch": 0.9695378151260504, |
| "grad_norm": 4.9375, |
| "learning_rate": 1.5125978090766824e-05, |
| "loss": 1.1237, |
| "step": 9230 |
| }, |
| { |
| "epoch": 0.9705882352941176, |
| "grad_norm": 4.3125, |
| "learning_rate": 1.5118153364632239e-05, |
| "loss": 1.0553, |
| "step": 9240 |
| }, |
| { |
| "epoch": 0.9716386554621849, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.5110328638497654e-05, |
| "loss": 1.795, |
| "step": 9250 |
| }, |
| { |
| "epoch": 0.9726890756302521, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.5102503912363068e-05, |
| "loss": 1.0514, |
| "step": 9260 |
| }, |
| { |
| "epoch": 0.9737394957983193, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.5094679186228483e-05, |
| "loss": 1.6683, |
| "step": 9270 |
| }, |
| { |
| "epoch": 0.9747899159663865, |
| "grad_norm": 24.125, |
| "learning_rate": 1.5086854460093898e-05, |
| "loss": 1.3332, |
| "step": 9280 |
| }, |
| { |
| "epoch": 0.9758403361344538, |
| "grad_norm": 11.9375, |
| "learning_rate": 1.5079029733959312e-05, |
| "loss": 1.1958, |
| "step": 9290 |
| }, |
| { |
| "epoch": 0.976890756302521, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.5071205007824727e-05, |
| "loss": 1.4438, |
| "step": 9300 |
| }, |
| { |
| "epoch": 0.9779411764705882, |
| "grad_norm": 17.25, |
| "learning_rate": 1.5063380281690143e-05, |
| "loss": 1.1032, |
| "step": 9310 |
| }, |
| { |
| "epoch": 0.9789915966386554, |
| "grad_norm": 18.125, |
| "learning_rate": 1.5055555555555556e-05, |
| "loss": 1.5746, |
| "step": 9320 |
| }, |
| { |
| "epoch": 0.9800420168067226, |
| "grad_norm": 16.5, |
| "learning_rate": 1.5047730829420973e-05, |
| "loss": 1.1873, |
| "step": 9330 |
| }, |
| { |
| "epoch": 0.9810924369747899, |
| "grad_norm": 5.15625, |
| "learning_rate": 1.5039906103286386e-05, |
| "loss": 1.0214, |
| "step": 9340 |
| }, |
| { |
| "epoch": 0.9821428571428571, |
| "grad_norm": 12.0, |
| "learning_rate": 1.5032081377151802e-05, |
| "loss": 1.3897, |
| "step": 9350 |
| }, |
| { |
| "epoch": 0.9831932773109243, |
| "grad_norm": 17.25, |
| "learning_rate": 1.5024256651017215e-05, |
| "loss": 1.7093, |
| "step": 9360 |
| }, |
| { |
| "epoch": 0.9842436974789915, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.5016431924882631e-05, |
| "loss": 1.1554, |
| "step": 9370 |
| }, |
| { |
| "epoch": 0.9852941176470589, |
| "grad_norm": 10.6875, |
| "learning_rate": 1.5008607198748044e-05, |
| "loss": 1.2735, |
| "step": 9380 |
| }, |
| { |
| "epoch": 0.9863445378151261, |
| "grad_norm": 15.375, |
| "learning_rate": 1.500078247261346e-05, |
| "loss": 1.3506, |
| "step": 9390 |
| }, |
| { |
| "epoch": 0.9873949579831933, |
| "grad_norm": 4.125, |
| "learning_rate": 1.4992957746478874e-05, |
| "loss": 1.1523, |
| "step": 9400 |
| }, |
| { |
| "epoch": 0.9884453781512605, |
| "grad_norm": 17.875, |
| "learning_rate": 1.498513302034429e-05, |
| "loss": 1.7688, |
| "step": 9410 |
| }, |
| { |
| "epoch": 0.9894957983193278, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.4977308294209705e-05, |
| "loss": 1.3584, |
| "step": 9420 |
| }, |
| { |
| "epoch": 0.990546218487395, |
| "grad_norm": 5.0, |
| "learning_rate": 1.4969483568075118e-05, |
| "loss": 0.7708, |
| "step": 9430 |
| }, |
| { |
| "epoch": 0.9915966386554622, |
| "grad_norm": 16.375, |
| "learning_rate": 1.4961658841940534e-05, |
| "loss": 1.4436, |
| "step": 9440 |
| }, |
| { |
| "epoch": 0.9926470588235294, |
| "grad_norm": 4.5625, |
| "learning_rate": 1.4953834115805947e-05, |
| "loss": 1.2742, |
| "step": 9450 |
| }, |
| { |
| "epoch": 0.9936974789915967, |
| "grad_norm": 10.875, |
| "learning_rate": 1.4946009389671363e-05, |
| "loss": 1.735, |
| "step": 9460 |
| }, |
| { |
| "epoch": 0.9947478991596639, |
| "grad_norm": 16.875, |
| "learning_rate": 1.4938184663536776e-05, |
| "loss": 1.4445, |
| "step": 9470 |
| }, |
| { |
| "epoch": 0.9957983193277311, |
| "grad_norm": 17.625, |
| "learning_rate": 1.4930359937402192e-05, |
| "loss": 1.5068, |
| "step": 9480 |
| }, |
| { |
| "epoch": 0.9968487394957983, |
| "grad_norm": 21.375, |
| "learning_rate": 1.4922535211267605e-05, |
| "loss": 1.1406, |
| "step": 9490 |
| }, |
| { |
| "epoch": 0.9978991596638656, |
| "grad_norm": 8.5, |
| "learning_rate": 1.4914710485133022e-05, |
| "loss": 1.3998, |
| "step": 9500 |
| }, |
| { |
| "epoch": 0.9989495798319328, |
| "grad_norm": 4.6875, |
| "learning_rate": 1.4906885758998436e-05, |
| "loss": 1.3126, |
| "step": 9510 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 14.25, |
| "learning_rate": 1.4899061032863851e-05, |
| "loss": 1.0987, |
| "step": 9520 |
| }, |
| { |
| "epoch": 1.0010504201680672, |
| "grad_norm": 35.25, |
| "learning_rate": 1.4891236306729266e-05, |
| "loss": 1.3145, |
| "step": 9530 |
| }, |
| { |
| "epoch": 1.0021008403361344, |
| "grad_norm": 28.375, |
| "learning_rate": 1.488341158059468e-05, |
| "loss": 1.5549, |
| "step": 9540 |
| }, |
| { |
| "epoch": 1.0031512605042017, |
| "grad_norm": 12.875, |
| "learning_rate": 1.4875586854460095e-05, |
| "loss": 1.0721, |
| "step": 9550 |
| }, |
| { |
| "epoch": 1.004201680672269, |
| "grad_norm": 27.25, |
| "learning_rate": 1.486776212832551e-05, |
| "loss": 0.9465, |
| "step": 9560 |
| }, |
| { |
| "epoch": 1.0052521008403361, |
| "grad_norm": 5.46875, |
| "learning_rate": 1.4859937402190924e-05, |
| "loss": 1.2033, |
| "step": 9570 |
| }, |
| { |
| "epoch": 1.0063025210084033, |
| "grad_norm": 16.125, |
| "learning_rate": 1.4852112676056339e-05, |
| "loss": 1.5769, |
| "step": 9580 |
| }, |
| { |
| "epoch": 1.0073529411764706, |
| "grad_norm": 17.0, |
| "learning_rate": 1.4844287949921754e-05, |
| "loss": 1.5204, |
| "step": 9590 |
| }, |
| { |
| "epoch": 1.0084033613445378, |
| "grad_norm": 16.625, |
| "learning_rate": 1.483646322378717e-05, |
| "loss": 1.3765, |
| "step": 9600 |
| }, |
| { |
| "epoch": 1.009453781512605, |
| "grad_norm": 19.25, |
| "learning_rate": 1.4828638497652583e-05, |
| "loss": 1.3035, |
| "step": 9610 |
| }, |
| { |
| "epoch": 1.0105042016806722, |
| "grad_norm": 2.046875, |
| "learning_rate": 1.4820813771518e-05, |
| "loss": 1.3045, |
| "step": 9620 |
| }, |
| { |
| "epoch": 1.0115546218487395, |
| "grad_norm": 15.4375, |
| "learning_rate": 1.4812989045383412e-05, |
| "loss": 1.4246, |
| "step": 9630 |
| }, |
| { |
| "epoch": 1.0126050420168067, |
| "grad_norm": 16.5, |
| "learning_rate": 1.4805164319248829e-05, |
| "loss": 1.1389, |
| "step": 9640 |
| }, |
| { |
| "epoch": 1.013655462184874, |
| "grad_norm": 12.5, |
| "learning_rate": 1.4797339593114242e-05, |
| "loss": 1.1888, |
| "step": 9650 |
| }, |
| { |
| "epoch": 1.0147058823529411, |
| "grad_norm": 23.0, |
| "learning_rate": 1.4789514866979658e-05, |
| "loss": 1.6063, |
| "step": 9660 |
| }, |
| { |
| "epoch": 1.0157563025210083, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.4781690140845071e-05, |
| "loss": 1.4559, |
| "step": 9670 |
| }, |
| { |
| "epoch": 1.0168067226890756, |
| "grad_norm": 16.25, |
| "learning_rate": 1.4773865414710487e-05, |
| "loss": 1.1611, |
| "step": 9680 |
| }, |
| { |
| "epoch": 1.0178571428571428, |
| "grad_norm": 23.5, |
| "learning_rate": 1.47660406885759e-05, |
| "loss": 1.3645, |
| "step": 9690 |
| }, |
| { |
| "epoch": 1.01890756302521, |
| "grad_norm": 11.375, |
| "learning_rate": 1.4758215962441317e-05, |
| "loss": 1.2812, |
| "step": 9700 |
| }, |
| { |
| "epoch": 1.0199579831932772, |
| "grad_norm": 11.625, |
| "learning_rate": 1.4750391236306731e-05, |
| "loss": 1.5295, |
| "step": 9710 |
| }, |
| { |
| "epoch": 1.0210084033613445, |
| "grad_norm": 7.96875, |
| "learning_rate": 1.4742566510172144e-05, |
| "loss": 1.1815, |
| "step": 9720 |
| }, |
| { |
| "epoch": 1.0220588235294117, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.473474178403756e-05, |
| "loss": 1.4964, |
| "step": 9730 |
| }, |
| { |
| "epoch": 1.023109243697479, |
| "grad_norm": 19.25, |
| "learning_rate": 1.4726917057902973e-05, |
| "loss": 1.3701, |
| "step": 9740 |
| }, |
| { |
| "epoch": 1.0241596638655461, |
| "grad_norm": 9.875, |
| "learning_rate": 1.471909233176839e-05, |
| "loss": 1.1485, |
| "step": 9750 |
| }, |
| { |
| "epoch": 1.0252100840336134, |
| "grad_norm": 18.5, |
| "learning_rate": 1.4711267605633803e-05, |
| "loss": 1.5789, |
| "step": 9760 |
| }, |
| { |
| "epoch": 1.0262605042016806, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.4703442879499219e-05, |
| "loss": 1.6354, |
| "step": 9770 |
| }, |
| { |
| "epoch": 1.0273109243697478, |
| "grad_norm": 15.375, |
| "learning_rate": 1.4695618153364632e-05, |
| "loss": 1.3682, |
| "step": 9780 |
| }, |
| { |
| "epoch": 1.028361344537815, |
| "grad_norm": 4.28125, |
| "learning_rate": 1.4687793427230048e-05, |
| "loss": 0.7953, |
| "step": 9790 |
| }, |
| { |
| "epoch": 1.0294117647058822, |
| "grad_norm": 13.75, |
| "learning_rate": 1.4679968701095463e-05, |
| "loss": 0.9399, |
| "step": 9800 |
| }, |
| { |
| "epoch": 1.0304621848739495, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.4672143974960878e-05, |
| "loss": 1.0148, |
| "step": 9810 |
| }, |
| { |
| "epoch": 1.0315126050420167, |
| "grad_norm": 5.25, |
| "learning_rate": 1.4664319248826292e-05, |
| "loss": 1.4619, |
| "step": 9820 |
| }, |
| { |
| "epoch": 1.0325630252100841, |
| "grad_norm": 18.625, |
| "learning_rate": 1.4656494522691707e-05, |
| "loss": 1.3168, |
| "step": 9830 |
| }, |
| { |
| "epoch": 1.0336134453781514, |
| "grad_norm": 5.25, |
| "learning_rate": 1.4648669796557122e-05, |
| "loss": 1.1152, |
| "step": 9840 |
| }, |
| { |
| "epoch": 1.0346638655462186, |
| "grad_norm": 12.0, |
| "learning_rate": 1.4640845070422536e-05, |
| "loss": 1.0424, |
| "step": 9850 |
| }, |
| { |
| "epoch": 1.0357142857142858, |
| "grad_norm": 14.9375, |
| "learning_rate": 1.4633020344287951e-05, |
| "loss": 1.2628, |
| "step": 9860 |
| }, |
| { |
| "epoch": 1.036764705882353, |
| "grad_norm": 14.0, |
| "learning_rate": 1.4625195618153366e-05, |
| "loss": 1.5196, |
| "step": 9870 |
| }, |
| { |
| "epoch": 1.0378151260504203, |
| "grad_norm": 12.5, |
| "learning_rate": 1.461737089201878e-05, |
| "loss": 1.02, |
| "step": 9880 |
| }, |
| { |
| "epoch": 1.0388655462184875, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.4609546165884195e-05, |
| "loss": 1.7924, |
| "step": 9890 |
| }, |
| { |
| "epoch": 1.0399159663865547, |
| "grad_norm": 18.875, |
| "learning_rate": 1.460172143974961e-05, |
| "loss": 1.303, |
| "step": 9900 |
| }, |
| { |
| "epoch": 1.040966386554622, |
| "grad_norm": 18.125, |
| "learning_rate": 1.4593896713615026e-05, |
| "loss": 1.0753, |
| "step": 9910 |
| }, |
| { |
| "epoch": 1.0420168067226891, |
| "grad_norm": 12.5, |
| "learning_rate": 1.4586071987480439e-05, |
| "loss": 1.2265, |
| "step": 9920 |
| }, |
| { |
| "epoch": 1.0430672268907564, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.4578247261345855e-05, |
| "loss": 1.6092, |
| "step": 9930 |
| }, |
| { |
| "epoch": 1.0441176470588236, |
| "grad_norm": 21.5, |
| "learning_rate": 1.4570422535211268e-05, |
| "loss": 1.2881, |
| "step": 9940 |
| }, |
| { |
| "epoch": 1.0451680672268908, |
| "grad_norm": 18.375, |
| "learning_rate": 1.4562597809076685e-05, |
| "loss": 1.4533, |
| "step": 9950 |
| }, |
| { |
| "epoch": 1.046218487394958, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.4554773082942098e-05, |
| "loss": 1.0597, |
| "step": 9960 |
| }, |
| { |
| "epoch": 1.0472689075630253, |
| "grad_norm": 4.03125, |
| "learning_rate": 1.4546948356807514e-05, |
| "loss": 1.3047, |
| "step": 9970 |
| }, |
| { |
| "epoch": 1.0483193277310925, |
| "grad_norm": 13.5, |
| "learning_rate": 1.4539123630672927e-05, |
| "loss": 1.4317, |
| "step": 9980 |
| }, |
| { |
| "epoch": 1.0493697478991597, |
| "grad_norm": 21.625, |
| "learning_rate": 1.4531298904538343e-05, |
| "loss": 1.5113, |
| "step": 9990 |
| }, |
| { |
| "epoch": 1.050420168067227, |
| "grad_norm": 14.6875, |
| "learning_rate": 1.4523474178403758e-05, |
| "loss": 1.5173, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.0514705882352942, |
| "grad_norm": 15.8125, |
| "learning_rate": 1.451564945226917e-05, |
| "loss": 1.4162, |
| "step": 10010 |
| }, |
| { |
| "epoch": 1.0525210084033614, |
| "grad_norm": 17.75, |
| "learning_rate": 1.4507824726134587e-05, |
| "loss": 1.7786, |
| "step": 10020 |
| }, |
| { |
| "epoch": 1.0535714285714286, |
| "grad_norm": 354.0, |
| "learning_rate": 1.45e-05, |
| "loss": 1.0287, |
| "step": 10030 |
| }, |
| { |
| "epoch": 1.0546218487394958, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.4492175273865416e-05, |
| "loss": 1.2295, |
| "step": 10040 |
| }, |
| { |
| "epoch": 1.055672268907563, |
| "grad_norm": 11.5625, |
| "learning_rate": 1.448435054773083e-05, |
| "loss": 1.391, |
| "step": 10050 |
| }, |
| { |
| "epoch": 1.0567226890756303, |
| "grad_norm": 15.125, |
| "learning_rate": 1.4476525821596246e-05, |
| "loss": 1.3713, |
| "step": 10060 |
| }, |
| { |
| "epoch": 1.0577731092436975, |
| "grad_norm": 16.75, |
| "learning_rate": 1.4468701095461659e-05, |
| "loss": 1.4097, |
| "step": 10070 |
| }, |
| { |
| "epoch": 1.0588235294117647, |
| "grad_norm": 11.375, |
| "learning_rate": 1.4460876369327075e-05, |
| "loss": 1.2367, |
| "step": 10080 |
| }, |
| { |
| "epoch": 1.059873949579832, |
| "grad_norm": 13.875, |
| "learning_rate": 1.4453051643192491e-05, |
| "loss": 1.8857, |
| "step": 10090 |
| }, |
| { |
| "epoch": 1.0609243697478992, |
| "grad_norm": 3.96875, |
| "learning_rate": 1.4445226917057904e-05, |
| "loss": 1.5483, |
| "step": 10100 |
| }, |
| { |
| "epoch": 1.0619747899159664, |
| "grad_norm": 5.71875, |
| "learning_rate": 1.4437402190923319e-05, |
| "loss": 1.0275, |
| "step": 10110 |
| }, |
| { |
| "epoch": 1.0630252100840336, |
| "grad_norm": 16.5, |
| "learning_rate": 1.4429577464788734e-05, |
| "loss": 1.4544, |
| "step": 10120 |
| }, |
| { |
| "epoch": 1.0640756302521008, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.4421752738654148e-05, |
| "loss": 1.4458, |
| "step": 10130 |
| }, |
| { |
| "epoch": 1.065126050420168, |
| "grad_norm": 16.25, |
| "learning_rate": 1.4413928012519563e-05, |
| "loss": 0.9895, |
| "step": 10140 |
| }, |
| { |
| "epoch": 1.0661764705882353, |
| "grad_norm": 16.625, |
| "learning_rate": 1.4406103286384978e-05, |
| "loss": 1.7776, |
| "step": 10150 |
| }, |
| { |
| "epoch": 1.0672268907563025, |
| "grad_norm": 24.0, |
| "learning_rate": 1.4398278560250392e-05, |
| "loss": 1.0594, |
| "step": 10160 |
| }, |
| { |
| "epoch": 1.0682773109243697, |
| "grad_norm": 12.25, |
| "learning_rate": 1.4390453834115807e-05, |
| "loss": 1.393, |
| "step": 10170 |
| }, |
| { |
| "epoch": 1.069327731092437, |
| "grad_norm": 12.875, |
| "learning_rate": 1.4382629107981222e-05, |
| "loss": 1.5787, |
| "step": 10180 |
| }, |
| { |
| "epoch": 1.0703781512605042, |
| "grad_norm": 13.625, |
| "learning_rate": 1.4374804381846636e-05, |
| "loss": 1.2195, |
| "step": 10190 |
| }, |
| { |
| "epoch": 1.0714285714285714, |
| "grad_norm": 6.34375, |
| "learning_rate": 1.4366979655712053e-05, |
| "loss": 1.3009, |
| "step": 10200 |
| }, |
| { |
| "epoch": 1.0724789915966386, |
| "grad_norm": 19.875, |
| "learning_rate": 1.4359154929577466e-05, |
| "loss": 1.6442, |
| "step": 10210 |
| }, |
| { |
| "epoch": 1.0735294117647058, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.4351330203442882e-05, |
| "loss": 1.4135, |
| "step": 10220 |
| }, |
| { |
| "epoch": 1.074579831932773, |
| "grad_norm": 5.125, |
| "learning_rate": 1.4343505477308295e-05, |
| "loss": 1.3497, |
| "step": 10230 |
| }, |
| { |
| "epoch": 1.0756302521008403, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.4335680751173711e-05, |
| "loss": 1.3815, |
| "step": 10240 |
| }, |
| { |
| "epoch": 1.0766806722689075, |
| "grad_norm": 17.875, |
| "learning_rate": 1.4327856025039124e-05, |
| "loss": 1.3823, |
| "step": 10250 |
| }, |
| { |
| "epoch": 1.0777310924369747, |
| "grad_norm": 19.0, |
| "learning_rate": 1.432003129890454e-05, |
| "loss": 1.7767, |
| "step": 10260 |
| }, |
| { |
| "epoch": 1.078781512605042, |
| "grad_norm": 20.75, |
| "learning_rate": 1.4312206572769953e-05, |
| "loss": 1.2912, |
| "step": 10270 |
| }, |
| { |
| "epoch": 1.0798319327731092, |
| "grad_norm": 12.625, |
| "learning_rate": 1.430438184663537e-05, |
| "loss": 1.0742, |
| "step": 10280 |
| }, |
| { |
| "epoch": 1.0808823529411764, |
| "grad_norm": 19.625, |
| "learning_rate": 1.4296557120500784e-05, |
| "loss": 0.905, |
| "step": 10290 |
| }, |
| { |
| "epoch": 1.0819327731092436, |
| "grad_norm": 4.78125, |
| "learning_rate": 1.4288732394366197e-05, |
| "loss": 0.8441, |
| "step": 10300 |
| }, |
| { |
| "epoch": 1.0829831932773109, |
| "grad_norm": 16.375, |
| "learning_rate": 1.4280907668231614e-05, |
| "loss": 1.4784, |
| "step": 10310 |
| }, |
| { |
| "epoch": 1.084033613445378, |
| "grad_norm": 9.625, |
| "learning_rate": 1.4273082942097027e-05, |
| "loss": 1.3086, |
| "step": 10320 |
| }, |
| { |
| "epoch": 1.0850840336134453, |
| "grad_norm": 12.0, |
| "learning_rate": 1.4265258215962443e-05, |
| "loss": 1.2392, |
| "step": 10330 |
| }, |
| { |
| "epoch": 1.0861344537815125, |
| "grad_norm": 24.75, |
| "learning_rate": 1.4257433489827856e-05, |
| "loss": 1.7351, |
| "step": 10340 |
| }, |
| { |
| "epoch": 1.0871848739495797, |
| "grad_norm": 16.25, |
| "learning_rate": 1.4249608763693272e-05, |
| "loss": 1.3572, |
| "step": 10350 |
| }, |
| { |
| "epoch": 1.088235294117647, |
| "grad_norm": 23.625, |
| "learning_rate": 1.4241784037558685e-05, |
| "loss": 1.2551, |
| "step": 10360 |
| }, |
| { |
| "epoch": 1.0892857142857142, |
| "grad_norm": 10.3125, |
| "learning_rate": 1.4233959311424102e-05, |
| "loss": 1.2559, |
| "step": 10370 |
| }, |
| { |
| "epoch": 1.0903361344537814, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.4226134585289515e-05, |
| "loss": 0.9075, |
| "step": 10380 |
| }, |
| { |
| "epoch": 1.0913865546218486, |
| "grad_norm": 18.75, |
| "learning_rate": 1.4218309859154931e-05, |
| "loss": 1.1107, |
| "step": 10390 |
| }, |
| { |
| "epoch": 1.092436974789916, |
| "grad_norm": 11.75, |
| "learning_rate": 1.4210485133020346e-05, |
| "loss": 1.5606, |
| "step": 10400 |
| }, |
| { |
| "epoch": 1.0934873949579833, |
| "grad_norm": 4.40625, |
| "learning_rate": 1.420266040688576e-05, |
| "loss": 1.2448, |
| "step": 10410 |
| }, |
| { |
| "epoch": 1.0945378151260505, |
| "grad_norm": 24.375, |
| "learning_rate": 1.4194835680751175e-05, |
| "loss": 1.3813, |
| "step": 10420 |
| }, |
| { |
| "epoch": 1.0955882352941178, |
| "grad_norm": 11.1875, |
| "learning_rate": 1.418701095461659e-05, |
| "loss": 1.2514, |
| "step": 10430 |
| }, |
| { |
| "epoch": 1.096638655462185, |
| "grad_norm": 12.5, |
| "learning_rate": 1.4179186228482004e-05, |
| "loss": 1.2028, |
| "step": 10440 |
| }, |
| { |
| "epoch": 1.0976890756302522, |
| "grad_norm": 17.5, |
| "learning_rate": 1.4171361502347419e-05, |
| "loss": 1.5367, |
| "step": 10450 |
| }, |
| { |
| "epoch": 1.0987394957983194, |
| "grad_norm": 15.5625, |
| "learning_rate": 1.4163536776212834e-05, |
| "loss": 1.3679, |
| "step": 10460 |
| }, |
| { |
| "epoch": 1.0997899159663866, |
| "grad_norm": 17.625, |
| "learning_rate": 1.4155712050078248e-05, |
| "loss": 1.7243, |
| "step": 10470 |
| }, |
| { |
| "epoch": 1.1008403361344539, |
| "grad_norm": 11.9375, |
| "learning_rate": 1.4147887323943663e-05, |
| "loss": 1.1804, |
| "step": 10480 |
| }, |
| { |
| "epoch": 1.101890756302521, |
| "grad_norm": 3.125, |
| "learning_rate": 1.414006259780908e-05, |
| "loss": 1.5059, |
| "step": 10490 |
| }, |
| { |
| "epoch": 1.1029411764705883, |
| "grad_norm": 9.125, |
| "learning_rate": 1.4132237871674492e-05, |
| "loss": 1.171, |
| "step": 10500 |
| }, |
| { |
| "epoch": 1.1039915966386555, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.4124413145539909e-05, |
| "loss": 1.5, |
| "step": 10510 |
| }, |
| { |
| "epoch": 1.1050420168067228, |
| "grad_norm": 11.9375, |
| "learning_rate": 1.4116588419405321e-05, |
| "loss": 1.8615, |
| "step": 10520 |
| }, |
| { |
| "epoch": 1.10609243697479, |
| "grad_norm": 11.5625, |
| "learning_rate": 1.4108763693270738e-05, |
| "loss": 1.547, |
| "step": 10530 |
| }, |
| { |
| "epoch": 1.1071428571428572, |
| "grad_norm": 18.625, |
| "learning_rate": 1.410093896713615e-05, |
| "loss": 1.4725, |
| "step": 10540 |
| }, |
| { |
| "epoch": 1.1081932773109244, |
| "grad_norm": 35.25, |
| "learning_rate": 1.4093114241001567e-05, |
| "loss": 1.4691, |
| "step": 10550 |
| }, |
| { |
| "epoch": 1.1092436974789917, |
| "grad_norm": 14.5625, |
| "learning_rate": 1.408528951486698e-05, |
| "loss": 1.2334, |
| "step": 10560 |
| }, |
| { |
| "epoch": 1.1102941176470589, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.4077464788732396e-05, |
| "loss": 1.5374, |
| "step": 10570 |
| }, |
| { |
| "epoch": 1.111344537815126, |
| "grad_norm": 13.75, |
| "learning_rate": 1.406964006259781e-05, |
| "loss": 1.4166, |
| "step": 10580 |
| }, |
| { |
| "epoch": 1.1123949579831933, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.4061815336463224e-05, |
| "loss": 1.1928, |
| "step": 10590 |
| }, |
| { |
| "epoch": 1.1134453781512605, |
| "grad_norm": 13.8125, |
| "learning_rate": 1.405399061032864e-05, |
| "loss": 1.2728, |
| "step": 10600 |
| }, |
| { |
| "epoch": 1.1144957983193278, |
| "grad_norm": 17.25, |
| "learning_rate": 1.4046165884194053e-05, |
| "loss": 1.3545, |
| "step": 10610 |
| }, |
| { |
| "epoch": 1.115546218487395, |
| "grad_norm": 20.0, |
| "learning_rate": 1.403834115805947e-05, |
| "loss": 1.6394, |
| "step": 10620 |
| }, |
| { |
| "epoch": 1.1165966386554622, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.4030516431924883e-05, |
| "loss": 1.6101, |
| "step": 10630 |
| }, |
| { |
| "epoch": 1.1176470588235294, |
| "grad_norm": 10.5625, |
| "learning_rate": 1.4022691705790299e-05, |
| "loss": 0.965, |
| "step": 10640 |
| }, |
| { |
| "epoch": 1.1186974789915967, |
| "grad_norm": 13.375, |
| "learning_rate": 1.4014866979655712e-05, |
| "loss": 1.4676, |
| "step": 10650 |
| }, |
| { |
| "epoch": 1.1197478991596639, |
| "grad_norm": 11.5625, |
| "learning_rate": 1.4007042253521128e-05, |
| "loss": 1.4875, |
| "step": 10660 |
| }, |
| { |
| "epoch": 1.120798319327731, |
| "grad_norm": 18.875, |
| "learning_rate": 1.3999217527386541e-05, |
| "loss": 1.5261, |
| "step": 10670 |
| }, |
| { |
| "epoch": 1.1218487394957983, |
| "grad_norm": 9.5, |
| "learning_rate": 1.3991392801251958e-05, |
| "loss": 1.3436, |
| "step": 10680 |
| }, |
| { |
| "epoch": 1.1228991596638656, |
| "grad_norm": 18.375, |
| "learning_rate": 1.3983568075117372e-05, |
| "loss": 1.3652, |
| "step": 10690 |
| }, |
| { |
| "epoch": 1.1239495798319328, |
| "grad_norm": 15.8125, |
| "learning_rate": 1.3975743348982787e-05, |
| "loss": 1.5565, |
| "step": 10700 |
| }, |
| { |
| "epoch": 1.125, |
| "grad_norm": 20.375, |
| "learning_rate": 1.3967918622848202e-05, |
| "loss": 1.3983, |
| "step": 10710 |
| }, |
| { |
| "epoch": 1.1260504201680672, |
| "grad_norm": 20.625, |
| "learning_rate": 1.3960093896713616e-05, |
| "loss": 1.2691, |
| "step": 10720 |
| }, |
| { |
| "epoch": 1.1271008403361344, |
| "grad_norm": 7.59375, |
| "learning_rate": 1.3952269170579031e-05, |
| "loss": 1.1509, |
| "step": 10730 |
| }, |
| { |
| "epoch": 1.1281512605042017, |
| "grad_norm": 12.875, |
| "learning_rate": 1.3944444444444446e-05, |
| "loss": 0.952, |
| "step": 10740 |
| }, |
| { |
| "epoch": 1.129201680672269, |
| "grad_norm": 14.375, |
| "learning_rate": 1.393661971830986e-05, |
| "loss": 1.3555, |
| "step": 10750 |
| }, |
| { |
| "epoch": 1.1302521008403361, |
| "grad_norm": 12.0625, |
| "learning_rate": 1.3928794992175275e-05, |
| "loss": 1.2439, |
| "step": 10760 |
| }, |
| { |
| "epoch": 1.1313025210084033, |
| "grad_norm": 11.875, |
| "learning_rate": 1.392097026604069e-05, |
| "loss": 1.1128, |
| "step": 10770 |
| }, |
| { |
| "epoch": 1.1323529411764706, |
| "grad_norm": 15.5625, |
| "learning_rate": 1.3913145539906106e-05, |
| "loss": 1.4089, |
| "step": 10780 |
| }, |
| { |
| "epoch": 1.1334033613445378, |
| "grad_norm": 16.0, |
| "learning_rate": 1.3905320813771519e-05, |
| "loss": 1.3108, |
| "step": 10790 |
| }, |
| { |
| "epoch": 1.134453781512605, |
| "grad_norm": 12.625, |
| "learning_rate": 1.3897496087636935e-05, |
| "loss": 0.923, |
| "step": 10800 |
| }, |
| { |
| "epoch": 1.1355042016806722, |
| "grad_norm": 5.6875, |
| "learning_rate": 1.3889671361502348e-05, |
| "loss": 0.8651, |
| "step": 10810 |
| }, |
| { |
| "epoch": 1.1365546218487395, |
| "grad_norm": 7.46875, |
| "learning_rate": 1.3881846635367764e-05, |
| "loss": 0.9996, |
| "step": 10820 |
| }, |
| { |
| "epoch": 1.1376050420168067, |
| "grad_norm": 15.25, |
| "learning_rate": 1.3874021909233177e-05, |
| "loss": 1.188, |
| "step": 10830 |
| }, |
| { |
| "epoch": 1.138655462184874, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.3866197183098594e-05, |
| "loss": 1.2221, |
| "step": 10840 |
| }, |
| { |
| "epoch": 1.1397058823529411, |
| "grad_norm": 18.75, |
| "learning_rate": 1.3858372456964007e-05, |
| "loss": 1.3794, |
| "step": 10850 |
| }, |
| { |
| "epoch": 1.1407563025210083, |
| "grad_norm": 16.875, |
| "learning_rate": 1.3850547730829423e-05, |
| "loss": 1.4104, |
| "step": 10860 |
| }, |
| { |
| "epoch": 1.1418067226890756, |
| "grad_norm": 27.25, |
| "learning_rate": 1.3842723004694836e-05, |
| "loss": 1.3409, |
| "step": 10870 |
| }, |
| { |
| "epoch": 1.1428571428571428, |
| "grad_norm": 23.875, |
| "learning_rate": 1.383489827856025e-05, |
| "loss": 1.4767, |
| "step": 10880 |
| }, |
| { |
| "epoch": 1.14390756302521, |
| "grad_norm": 5.46875, |
| "learning_rate": 1.3827073552425667e-05, |
| "loss": 0.9714, |
| "step": 10890 |
| }, |
| { |
| "epoch": 1.1449579831932772, |
| "grad_norm": 5.59375, |
| "learning_rate": 1.381924882629108e-05, |
| "loss": 1.3208, |
| "step": 10900 |
| }, |
| { |
| "epoch": 1.1460084033613445, |
| "grad_norm": 15.5, |
| "learning_rate": 1.3811424100156496e-05, |
| "loss": 1.2431, |
| "step": 10910 |
| }, |
| { |
| "epoch": 1.1470588235294117, |
| "grad_norm": 9.0625, |
| "learning_rate": 1.380359937402191e-05, |
| "loss": 1.254, |
| "step": 10920 |
| }, |
| { |
| "epoch": 1.148109243697479, |
| "grad_norm": 16.75, |
| "learning_rate": 1.3795774647887326e-05, |
| "loss": 1.6899, |
| "step": 10930 |
| }, |
| { |
| "epoch": 1.1491596638655461, |
| "grad_norm": 15.1875, |
| "learning_rate": 1.3787949921752739e-05, |
| "loss": 1.1452, |
| "step": 10940 |
| }, |
| { |
| "epoch": 1.1502100840336134, |
| "grad_norm": 17.75, |
| "learning_rate": 1.3780125195618155e-05, |
| "loss": 1.8214, |
| "step": 10950 |
| }, |
| { |
| "epoch": 1.1512605042016806, |
| "grad_norm": 4.71875, |
| "learning_rate": 1.3772300469483568e-05, |
| "loss": 0.724, |
| "step": 10960 |
| }, |
| { |
| "epoch": 1.1523109243697478, |
| "grad_norm": 23.375, |
| "learning_rate": 1.3764475743348984e-05, |
| "loss": 1.2687, |
| "step": 10970 |
| }, |
| { |
| "epoch": 1.153361344537815, |
| "grad_norm": 13.0, |
| "learning_rate": 1.3756651017214399e-05, |
| "loss": 1.2736, |
| "step": 10980 |
| }, |
| { |
| "epoch": 1.1544117647058822, |
| "grad_norm": 4.40625, |
| "learning_rate": 1.3748826291079814e-05, |
| "loss": 1.5121, |
| "step": 10990 |
| }, |
| { |
| "epoch": 1.1554621848739495, |
| "grad_norm": 16.375, |
| "learning_rate": 1.3741001564945228e-05, |
| "loss": 1.3454, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.1565126050420167, |
| "grad_norm": 17.0, |
| "learning_rate": 1.3733176838810643e-05, |
| "loss": 1.5489, |
| "step": 11010 |
| }, |
| { |
| "epoch": 1.157563025210084, |
| "grad_norm": 15.9375, |
| "learning_rate": 1.3725352112676057e-05, |
| "loss": 1.0976, |
| "step": 11020 |
| }, |
| { |
| "epoch": 1.1586134453781511, |
| "grad_norm": 34.25, |
| "learning_rate": 1.3717527386541472e-05, |
| "loss": 1.1138, |
| "step": 11030 |
| }, |
| { |
| "epoch": 1.1596638655462184, |
| "grad_norm": 16.75, |
| "learning_rate": 1.3709702660406887e-05, |
| "loss": 1.4818, |
| "step": 11040 |
| }, |
| { |
| "epoch": 1.1607142857142858, |
| "grad_norm": 16.25, |
| "learning_rate": 1.3701877934272301e-05, |
| "loss": 1.3685, |
| "step": 11050 |
| }, |
| { |
| "epoch": 1.161764705882353, |
| "grad_norm": 9.25, |
| "learning_rate": 1.3694053208137716e-05, |
| "loss": 1.2225, |
| "step": 11060 |
| }, |
| { |
| "epoch": 1.1628151260504203, |
| "grad_norm": 12.0, |
| "learning_rate": 1.368622848200313e-05, |
| "loss": 1.2856, |
| "step": 11070 |
| }, |
| { |
| "epoch": 1.1638655462184875, |
| "grad_norm": 7.28125, |
| "learning_rate": 1.3678403755868545e-05, |
| "loss": 1.1306, |
| "step": 11080 |
| }, |
| { |
| "epoch": 1.1649159663865547, |
| "grad_norm": 9.3125, |
| "learning_rate": 1.3670579029733962e-05, |
| "loss": 1.333, |
| "step": 11090 |
| }, |
| { |
| "epoch": 1.165966386554622, |
| "grad_norm": 9.75, |
| "learning_rate": 1.3662754303599375e-05, |
| "loss": 1.265, |
| "step": 11100 |
| }, |
| { |
| "epoch": 1.1670168067226891, |
| "grad_norm": 19.625, |
| "learning_rate": 1.3654929577464791e-05, |
| "loss": 1.5931, |
| "step": 11110 |
| }, |
| { |
| "epoch": 1.1680672268907564, |
| "grad_norm": 4.3125, |
| "learning_rate": 1.3647104851330204e-05, |
| "loss": 1.6028, |
| "step": 11120 |
| }, |
| { |
| "epoch": 1.1691176470588236, |
| "grad_norm": 13.1875, |
| "learning_rate": 1.363928012519562e-05, |
| "loss": 1.2405, |
| "step": 11130 |
| }, |
| { |
| "epoch": 1.1701680672268908, |
| "grad_norm": 12.8125, |
| "learning_rate": 1.3631455399061033e-05, |
| "loss": 1.6825, |
| "step": 11140 |
| }, |
| { |
| "epoch": 1.171218487394958, |
| "grad_norm": 18.125, |
| "learning_rate": 1.362363067292645e-05, |
| "loss": 1.2953, |
| "step": 11150 |
| }, |
| { |
| "epoch": 1.1722689075630253, |
| "grad_norm": 14.625, |
| "learning_rate": 1.3615805946791863e-05, |
| "loss": 1.254, |
| "step": 11160 |
| }, |
| { |
| "epoch": 1.1733193277310925, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.3607981220657277e-05, |
| "loss": 0.9121, |
| "step": 11170 |
| }, |
| { |
| "epoch": 1.1743697478991597, |
| "grad_norm": 15.875, |
| "learning_rate": 1.3600156494522694e-05, |
| "loss": 1.2803, |
| "step": 11180 |
| }, |
| { |
| "epoch": 1.175420168067227, |
| "grad_norm": 20.0, |
| "learning_rate": 1.3592331768388107e-05, |
| "loss": 1.0452, |
| "step": 11190 |
| }, |
| { |
| "epoch": 1.1764705882352942, |
| "grad_norm": 17.875, |
| "learning_rate": 1.3584507042253523e-05, |
| "loss": 1.6264, |
| "step": 11200 |
| }, |
| { |
| "epoch": 1.1775210084033614, |
| "grad_norm": 36.25, |
| "learning_rate": 1.3576682316118936e-05, |
| "loss": 1.2355, |
| "step": 11210 |
| }, |
| { |
| "epoch": 1.1785714285714286, |
| "grad_norm": 11.375, |
| "learning_rate": 1.3568857589984352e-05, |
| "loss": 1.7128, |
| "step": 11220 |
| }, |
| { |
| "epoch": 1.1796218487394958, |
| "grad_norm": 9.875, |
| "learning_rate": 1.3561032863849765e-05, |
| "loss": 1.0711, |
| "step": 11230 |
| }, |
| { |
| "epoch": 1.180672268907563, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.3553208137715182e-05, |
| "loss": 1.1493, |
| "step": 11240 |
| }, |
| { |
| "epoch": 1.1817226890756303, |
| "grad_norm": 14.6875, |
| "learning_rate": 1.3545383411580595e-05, |
| "loss": 1.3251, |
| "step": 11250 |
| }, |
| { |
| "epoch": 1.1827731092436975, |
| "grad_norm": 21.625, |
| "learning_rate": 1.3537558685446011e-05, |
| "loss": 1.8988, |
| "step": 11260 |
| }, |
| { |
| "epoch": 1.1838235294117647, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.3529733959311426e-05, |
| "loss": 1.6502, |
| "step": 11270 |
| }, |
| { |
| "epoch": 1.184873949579832, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.352190923317684e-05, |
| "loss": 1.2318, |
| "step": 11280 |
| }, |
| { |
| "epoch": 1.1859243697478992, |
| "grad_norm": 12.125, |
| "learning_rate": 1.3514084507042255e-05, |
| "loss": 1.6075, |
| "step": 11290 |
| }, |
| { |
| "epoch": 1.1869747899159664, |
| "grad_norm": 14.5, |
| "learning_rate": 1.350625978090767e-05, |
| "loss": 1.43, |
| "step": 11300 |
| }, |
| { |
| "epoch": 1.1880252100840336, |
| "grad_norm": 20.875, |
| "learning_rate": 1.3498435054773084e-05, |
| "loss": 1.5842, |
| "step": 11310 |
| }, |
| { |
| "epoch": 1.1890756302521008, |
| "grad_norm": 14.375, |
| "learning_rate": 1.3490610328638499e-05, |
| "loss": 1.573, |
| "step": 11320 |
| }, |
| { |
| "epoch": 1.190126050420168, |
| "grad_norm": 11.25, |
| "learning_rate": 1.3482785602503913e-05, |
| "loss": 1.5654, |
| "step": 11330 |
| }, |
| { |
| "epoch": 1.1911764705882353, |
| "grad_norm": 18.125, |
| "learning_rate": 1.3474960876369328e-05, |
| "loss": 2.1257, |
| "step": 11340 |
| }, |
| { |
| "epoch": 1.1922268907563025, |
| "grad_norm": 16.875, |
| "learning_rate": 1.3467136150234743e-05, |
| "loss": 1.4576, |
| "step": 11350 |
| }, |
| { |
| "epoch": 1.1932773109243697, |
| "grad_norm": 8.0625, |
| "learning_rate": 1.3459311424100157e-05, |
| "loss": 1.3279, |
| "step": 11360 |
| }, |
| { |
| "epoch": 1.194327731092437, |
| "grad_norm": 189.0, |
| "learning_rate": 1.3451486697965572e-05, |
| "loss": 1.4203, |
| "step": 11370 |
| }, |
| { |
| "epoch": 1.1953781512605042, |
| "grad_norm": 21.375, |
| "learning_rate": 1.3443661971830988e-05, |
| "loss": 1.4392, |
| "step": 11380 |
| }, |
| { |
| "epoch": 1.1964285714285714, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.3435837245696401e-05, |
| "loss": 1.6435, |
| "step": 11390 |
| }, |
| { |
| "epoch": 1.1974789915966386, |
| "grad_norm": 22.125, |
| "learning_rate": 1.3428012519561818e-05, |
| "loss": 1.1396, |
| "step": 11400 |
| }, |
| { |
| "epoch": 1.1985294117647058, |
| "grad_norm": 13.125, |
| "learning_rate": 1.342018779342723e-05, |
| "loss": 0.9979, |
| "step": 11410 |
| }, |
| { |
| "epoch": 1.199579831932773, |
| "grad_norm": 14.9375, |
| "learning_rate": 1.3412363067292647e-05, |
| "loss": 1.2108, |
| "step": 11420 |
| }, |
| { |
| "epoch": 1.2006302521008403, |
| "grad_norm": 19.625, |
| "learning_rate": 1.340453834115806e-05, |
| "loss": 0.9337, |
| "step": 11430 |
| }, |
| { |
| "epoch": 1.2016806722689075, |
| "grad_norm": 14.25, |
| "learning_rate": 1.3396713615023476e-05, |
| "loss": 1.4902, |
| "step": 11440 |
| }, |
| { |
| "epoch": 1.2027310924369747, |
| "grad_norm": 16.875, |
| "learning_rate": 1.338888888888889e-05, |
| "loss": 1.4991, |
| "step": 11450 |
| }, |
| { |
| "epoch": 1.203781512605042, |
| "grad_norm": 18.5, |
| "learning_rate": 1.3381064162754304e-05, |
| "loss": 1.2124, |
| "step": 11460 |
| }, |
| { |
| "epoch": 1.2048319327731092, |
| "grad_norm": 15.5625, |
| "learning_rate": 1.337323943661972e-05, |
| "loss": 1.4139, |
| "step": 11470 |
| }, |
| { |
| "epoch": 1.2058823529411764, |
| "grad_norm": 14.0, |
| "learning_rate": 1.3365414710485133e-05, |
| "loss": 1.3933, |
| "step": 11480 |
| }, |
| { |
| "epoch": 1.2069327731092436, |
| "grad_norm": 4.9375, |
| "learning_rate": 1.335758998435055e-05, |
| "loss": 1.5434, |
| "step": 11490 |
| }, |
| { |
| "epoch": 1.2079831932773109, |
| "grad_norm": 16.25, |
| "learning_rate": 1.3349765258215963e-05, |
| "loss": 1.5367, |
| "step": 11500 |
| }, |
| { |
| "epoch": 1.209033613445378, |
| "grad_norm": 9.25, |
| "learning_rate": 1.3341940532081379e-05, |
| "loss": 1.3468, |
| "step": 11510 |
| }, |
| { |
| "epoch": 1.2100840336134453, |
| "grad_norm": 15.875, |
| "learning_rate": 1.3334115805946792e-05, |
| "loss": 1.4425, |
| "step": 11520 |
| }, |
| { |
| "epoch": 1.2111344537815125, |
| "grad_norm": 18.625, |
| "learning_rate": 1.3326291079812208e-05, |
| "loss": 1.7911, |
| "step": 11530 |
| }, |
| { |
| "epoch": 1.2121848739495797, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.3318466353677621e-05, |
| "loss": 1.1501, |
| "step": 11540 |
| }, |
| { |
| "epoch": 1.213235294117647, |
| "grad_norm": 13.1875, |
| "learning_rate": 1.3310641627543037e-05, |
| "loss": 1.2397, |
| "step": 11550 |
| }, |
| { |
| "epoch": 1.2142857142857142, |
| "grad_norm": 14.25, |
| "learning_rate": 1.330281690140845e-05, |
| "loss": 1.5209, |
| "step": 11560 |
| }, |
| { |
| "epoch": 1.2153361344537816, |
| "grad_norm": 3.5, |
| "learning_rate": 1.3294992175273867e-05, |
| "loss": 1.6226, |
| "step": 11570 |
| }, |
| { |
| "epoch": 1.2163865546218489, |
| "grad_norm": 17.25, |
| "learning_rate": 1.3287167449139281e-05, |
| "loss": 1.2548, |
| "step": 11580 |
| }, |
| { |
| "epoch": 1.217436974789916, |
| "grad_norm": 15.9375, |
| "learning_rate": 1.3279342723004696e-05, |
| "loss": 1.5391, |
| "step": 11590 |
| }, |
| { |
| "epoch": 1.2184873949579833, |
| "grad_norm": 17.25, |
| "learning_rate": 1.327151799687011e-05, |
| "loss": 1.0214, |
| "step": 11600 |
| }, |
| { |
| "epoch": 1.2195378151260505, |
| "grad_norm": 17.625, |
| "learning_rate": 1.3263693270735525e-05, |
| "loss": 1.2153, |
| "step": 11610 |
| }, |
| { |
| "epoch": 1.2205882352941178, |
| "grad_norm": 9.1875, |
| "learning_rate": 1.325586854460094e-05, |
| "loss": 0.8848, |
| "step": 11620 |
| }, |
| { |
| "epoch": 1.221638655462185, |
| "grad_norm": 11.875, |
| "learning_rate": 1.3248043818466355e-05, |
| "loss": 1.4852, |
| "step": 11630 |
| }, |
| { |
| "epoch": 1.2226890756302522, |
| "grad_norm": 5.1875, |
| "learning_rate": 1.324021909233177e-05, |
| "loss": 1.2781, |
| "step": 11640 |
| }, |
| { |
| "epoch": 1.2237394957983194, |
| "grad_norm": 13.5, |
| "learning_rate": 1.3232394366197184e-05, |
| "loss": 1.4085, |
| "step": 11650 |
| }, |
| { |
| "epoch": 1.2247899159663866, |
| "grad_norm": 12.875, |
| "learning_rate": 1.3224569640062599e-05, |
| "loss": 0.9212, |
| "step": 11660 |
| }, |
| { |
| "epoch": 1.2258403361344539, |
| "grad_norm": 9.8125, |
| "learning_rate": 1.3216744913928015e-05, |
| "loss": 1.5026, |
| "step": 11670 |
| }, |
| { |
| "epoch": 1.226890756302521, |
| "grad_norm": 15.1875, |
| "learning_rate": 1.3208920187793428e-05, |
| "loss": 1.153, |
| "step": 11680 |
| }, |
| { |
| "epoch": 1.2279411764705883, |
| "grad_norm": 5.78125, |
| "learning_rate": 1.3201095461658844e-05, |
| "loss": 1.5746, |
| "step": 11690 |
| }, |
| { |
| "epoch": 1.2289915966386555, |
| "grad_norm": 5.5625, |
| "learning_rate": 1.3193270735524257e-05, |
| "loss": 1.3192, |
| "step": 11700 |
| }, |
| { |
| "epoch": 1.2300420168067228, |
| "grad_norm": 12.5, |
| "learning_rate": 1.3185446009389674e-05, |
| "loss": 1.6797, |
| "step": 11710 |
| }, |
| { |
| "epoch": 1.23109243697479, |
| "grad_norm": 5.21875, |
| "learning_rate": 1.3177621283255087e-05, |
| "loss": 1.0005, |
| "step": 11720 |
| }, |
| { |
| "epoch": 1.2321428571428572, |
| "grad_norm": 11.25, |
| "learning_rate": 1.3169796557120503e-05, |
| "loss": 1.5104, |
| "step": 11730 |
| }, |
| { |
| "epoch": 1.2331932773109244, |
| "grad_norm": 8.8125, |
| "learning_rate": 1.3161971830985916e-05, |
| "loss": 1.4712, |
| "step": 11740 |
| }, |
| { |
| "epoch": 1.2342436974789917, |
| "grad_norm": 6.84375, |
| "learning_rate": 1.3154147104851332e-05, |
| "loss": 1.3192, |
| "step": 11750 |
| }, |
| { |
| "epoch": 1.2352941176470589, |
| "grad_norm": 11.875, |
| "learning_rate": 1.3146322378716745e-05, |
| "loss": 1.0407, |
| "step": 11760 |
| }, |
| { |
| "epoch": 1.236344537815126, |
| "grad_norm": 5.46875, |
| "learning_rate": 1.313849765258216e-05, |
| "loss": 1.7191, |
| "step": 11770 |
| }, |
| { |
| "epoch": 1.2373949579831933, |
| "grad_norm": 17.375, |
| "learning_rate": 1.3130672926447576e-05, |
| "loss": 1.6019, |
| "step": 11780 |
| }, |
| { |
| "epoch": 1.2384453781512605, |
| "grad_norm": 9.75, |
| "learning_rate": 1.3122848200312989e-05, |
| "loss": 1.4871, |
| "step": 11790 |
| }, |
| { |
| "epoch": 1.2394957983193278, |
| "grad_norm": 7.8125, |
| "learning_rate": 1.3115023474178405e-05, |
| "loss": 1.3843, |
| "step": 11800 |
| }, |
| { |
| "epoch": 1.240546218487395, |
| "grad_norm": 7.28125, |
| "learning_rate": 1.3107198748043818e-05, |
| "loss": 1.1443, |
| "step": 11810 |
| }, |
| { |
| "epoch": 1.2415966386554622, |
| "grad_norm": 13.75, |
| "learning_rate": 1.3099374021909235e-05, |
| "loss": 1.1077, |
| "step": 11820 |
| }, |
| { |
| "epoch": 1.2426470588235294, |
| "grad_norm": 17.125, |
| "learning_rate": 1.3091549295774648e-05, |
| "loss": 1.1865, |
| "step": 11830 |
| }, |
| { |
| "epoch": 1.2436974789915967, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.3083724569640064e-05, |
| "loss": 1.5642, |
| "step": 11840 |
| }, |
| { |
| "epoch": 1.2447478991596639, |
| "grad_norm": 9.6875, |
| "learning_rate": 1.3075899843505477e-05, |
| "loss": 1.1188, |
| "step": 11850 |
| }, |
| { |
| "epoch": 1.245798319327731, |
| "grad_norm": 16.125, |
| "learning_rate": 1.3068075117370893e-05, |
| "loss": 2.0191, |
| "step": 11860 |
| }, |
| { |
| "epoch": 1.2468487394957983, |
| "grad_norm": 12.25, |
| "learning_rate": 1.3060250391236308e-05, |
| "loss": 1.4672, |
| "step": 11870 |
| }, |
| { |
| "epoch": 1.2478991596638656, |
| "grad_norm": 5.28125, |
| "learning_rate": 1.3052425665101723e-05, |
| "loss": 1.1893, |
| "step": 11880 |
| }, |
| { |
| "epoch": 1.2489495798319328, |
| "grad_norm": 9.0625, |
| "learning_rate": 1.3044600938967137e-05, |
| "loss": 1.175, |
| "step": 11890 |
| }, |
| { |
| "epoch": 1.25, |
| "grad_norm": 13.125, |
| "learning_rate": 1.3036776212832552e-05, |
| "loss": 1.8234, |
| "step": 11900 |
| }, |
| { |
| "epoch": 1.2510504201680672, |
| "grad_norm": 20.875, |
| "learning_rate": 1.3028951486697967e-05, |
| "loss": 1.6298, |
| "step": 11910 |
| }, |
| { |
| "epoch": 1.2521008403361344, |
| "grad_norm": 14.0625, |
| "learning_rate": 1.3021126760563381e-05, |
| "loss": 1.3639, |
| "step": 11920 |
| }, |
| { |
| "epoch": 1.2531512605042017, |
| "grad_norm": 24.375, |
| "learning_rate": 1.3013302034428796e-05, |
| "loss": 1.449, |
| "step": 11930 |
| }, |
| { |
| "epoch": 1.254201680672269, |
| "grad_norm": 11.3125, |
| "learning_rate": 1.300547730829421e-05, |
| "loss": 1.1688, |
| "step": 11940 |
| }, |
| { |
| "epoch": 1.2552521008403361, |
| "grad_norm": 13.25, |
| "learning_rate": 1.2997652582159625e-05, |
| "loss": 1.6437, |
| "step": 11950 |
| }, |
| { |
| "epoch": 1.2563025210084033, |
| "grad_norm": 13.4375, |
| "learning_rate": 1.2989827856025042e-05, |
| "loss": 1.0691, |
| "step": 11960 |
| }, |
| { |
| "epoch": 1.2573529411764706, |
| "grad_norm": 24.125, |
| "learning_rate": 1.2982003129890455e-05, |
| "loss": 1.7317, |
| "step": 11970 |
| }, |
| { |
| "epoch": 1.2584033613445378, |
| "grad_norm": 19.125, |
| "learning_rate": 1.2974178403755871e-05, |
| "loss": 1.1363, |
| "step": 11980 |
| }, |
| { |
| "epoch": 1.259453781512605, |
| "grad_norm": 14.1875, |
| "learning_rate": 1.2966353677621284e-05, |
| "loss": 1.362, |
| "step": 11990 |
| }, |
| { |
| "epoch": 1.2605042016806722, |
| "grad_norm": 11.375, |
| "learning_rate": 1.29585289514867e-05, |
| "loss": 1.6147, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.2615546218487395, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.2950704225352113e-05, |
| "loss": 1.3879, |
| "step": 12010 |
| }, |
| { |
| "epoch": 1.2626050420168067, |
| "grad_norm": 11.9375, |
| "learning_rate": 1.294287949921753e-05, |
| "loss": 1.3297, |
| "step": 12020 |
| }, |
| { |
| "epoch": 1.263655462184874, |
| "grad_norm": 11.5, |
| "learning_rate": 1.2935054773082943e-05, |
| "loss": 1.1665, |
| "step": 12030 |
| }, |
| { |
| "epoch": 1.2647058823529411, |
| "grad_norm": 18.125, |
| "learning_rate": 1.2927230046948359e-05, |
| "loss": 1.4025, |
| "step": 12040 |
| }, |
| { |
| "epoch": 1.2657563025210083, |
| "grad_norm": 25.0, |
| "learning_rate": 1.2919405320813772e-05, |
| "loss": 1.2586, |
| "step": 12050 |
| }, |
| { |
| "epoch": 1.2668067226890756, |
| "grad_norm": 12.5, |
| "learning_rate": 1.2911580594679186e-05, |
| "loss": 1.7389, |
| "step": 12060 |
| }, |
| { |
| "epoch": 1.2678571428571428, |
| "grad_norm": 19.25, |
| "learning_rate": 1.2903755868544603e-05, |
| "loss": 1.1541, |
| "step": 12070 |
| }, |
| { |
| "epoch": 1.26890756302521, |
| "grad_norm": 7.625, |
| "learning_rate": 1.2895931142410016e-05, |
| "loss": 1.1763, |
| "step": 12080 |
| }, |
| { |
| "epoch": 1.2699579831932772, |
| "grad_norm": 23.0, |
| "learning_rate": 1.2888106416275432e-05, |
| "loss": 1.2426, |
| "step": 12090 |
| }, |
| { |
| "epoch": 1.2710084033613445, |
| "grad_norm": 4.84375, |
| "learning_rate": 1.2880281690140845e-05, |
| "loss": 1.3512, |
| "step": 12100 |
| }, |
| { |
| "epoch": 1.2720588235294117, |
| "grad_norm": 15.5, |
| "learning_rate": 1.2872456964006261e-05, |
| "loss": 1.1114, |
| "step": 12110 |
| }, |
| { |
| "epoch": 1.273109243697479, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.2864632237871674e-05, |
| "loss": 1.3452, |
| "step": 12120 |
| }, |
| { |
| "epoch": 1.2741596638655461, |
| "grad_norm": 21.125, |
| "learning_rate": 1.285680751173709e-05, |
| "loss": 1.4839, |
| "step": 12130 |
| }, |
| { |
| "epoch": 1.2752100840336134, |
| "grad_norm": 26.0, |
| "learning_rate": 1.2848982785602504e-05, |
| "loss": 1.4127, |
| "step": 12140 |
| }, |
| { |
| "epoch": 1.2762605042016806, |
| "grad_norm": 13.375, |
| "learning_rate": 1.284115805946792e-05, |
| "loss": 0.9342, |
| "step": 12150 |
| }, |
| { |
| "epoch": 1.2773109243697478, |
| "grad_norm": 13.125, |
| "learning_rate": 1.2833333333333335e-05, |
| "loss": 1.1793, |
| "step": 12160 |
| }, |
| { |
| "epoch": 1.278361344537815, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.282550860719875e-05, |
| "loss": 1.3344, |
| "step": 12170 |
| }, |
| { |
| "epoch": 1.2794117647058822, |
| "grad_norm": 21.25, |
| "learning_rate": 1.2817683881064164e-05, |
| "loss": 0.8956, |
| "step": 12180 |
| }, |
| { |
| "epoch": 1.2804621848739495, |
| "grad_norm": 15.5625, |
| "learning_rate": 1.2809859154929579e-05, |
| "loss": 1.3549, |
| "step": 12190 |
| }, |
| { |
| "epoch": 1.2815126050420167, |
| "grad_norm": 17.875, |
| "learning_rate": 1.2802034428794993e-05, |
| "loss": 1.5818, |
| "step": 12200 |
| }, |
| { |
| "epoch": 1.282563025210084, |
| "grad_norm": 5.65625, |
| "learning_rate": 1.2794209702660408e-05, |
| "loss": 1.0335, |
| "step": 12210 |
| }, |
| { |
| "epoch": 1.2836134453781511, |
| "grad_norm": 3.921875, |
| "learning_rate": 1.2786384976525823e-05, |
| "loss": 1.2507, |
| "step": 12220 |
| }, |
| { |
| "epoch": 1.2846638655462184, |
| "grad_norm": 27.875, |
| "learning_rate": 1.2778560250391237e-05, |
| "loss": 1.9042, |
| "step": 12230 |
| }, |
| { |
| "epoch": 1.2857142857142856, |
| "grad_norm": 3.921875, |
| "learning_rate": 1.2770735524256652e-05, |
| "loss": 1.6034, |
| "step": 12240 |
| }, |
| { |
| "epoch": 1.2867647058823528, |
| "grad_norm": 4.6875, |
| "learning_rate": 1.2762910798122065e-05, |
| "loss": 1.2122, |
| "step": 12250 |
| }, |
| { |
| "epoch": 1.28781512605042, |
| "grad_norm": 8.75, |
| "learning_rate": 1.2755086071987481e-05, |
| "loss": 1.5361, |
| "step": 12260 |
| }, |
| { |
| "epoch": 1.2888655462184873, |
| "grad_norm": 25.25, |
| "learning_rate": 1.2747261345852898e-05, |
| "loss": 1.5994, |
| "step": 12270 |
| }, |
| { |
| "epoch": 1.2899159663865547, |
| "grad_norm": 20.125, |
| "learning_rate": 1.273943661971831e-05, |
| "loss": 1.5681, |
| "step": 12280 |
| }, |
| { |
| "epoch": 1.290966386554622, |
| "grad_norm": 13.4375, |
| "learning_rate": 1.2731611893583727e-05, |
| "loss": 1.2839, |
| "step": 12290 |
| }, |
| { |
| "epoch": 1.2920168067226891, |
| "grad_norm": 11.9375, |
| "learning_rate": 1.272378716744914e-05, |
| "loss": 1.1758, |
| "step": 12300 |
| }, |
| { |
| "epoch": 1.2930672268907564, |
| "grad_norm": 4.6875, |
| "learning_rate": 1.2715962441314556e-05, |
| "loss": 1.2308, |
| "step": 12310 |
| }, |
| { |
| "epoch": 1.2941176470588236, |
| "grad_norm": 27.125, |
| "learning_rate": 1.2708137715179969e-05, |
| "loss": 1.236, |
| "step": 12320 |
| }, |
| { |
| "epoch": 1.2951680672268908, |
| "grad_norm": 11.875, |
| "learning_rate": 1.2700312989045385e-05, |
| "loss": 1.0556, |
| "step": 12330 |
| }, |
| { |
| "epoch": 1.296218487394958, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.2692488262910798e-05, |
| "loss": 1.2603, |
| "step": 12340 |
| }, |
| { |
| "epoch": 1.2972689075630253, |
| "grad_norm": 9.875, |
| "learning_rate": 1.2684663536776213e-05, |
| "loss": 0.8483, |
| "step": 12350 |
| }, |
| { |
| "epoch": 1.2983193277310925, |
| "grad_norm": 15.875, |
| "learning_rate": 1.267683881064163e-05, |
| "loss": 1.514, |
| "step": 12360 |
| }, |
| { |
| "epoch": 1.2993697478991597, |
| "grad_norm": 9.0, |
| "learning_rate": 1.2669014084507042e-05, |
| "loss": 1.1939, |
| "step": 12370 |
| }, |
| { |
| "epoch": 1.300420168067227, |
| "grad_norm": 15.75, |
| "learning_rate": 1.2661189358372459e-05, |
| "loss": 1.5688, |
| "step": 12380 |
| }, |
| { |
| "epoch": 1.3014705882352942, |
| "grad_norm": 19.5, |
| "learning_rate": 1.2653364632237872e-05, |
| "loss": 1.3059, |
| "step": 12390 |
| }, |
| { |
| "epoch": 1.3025210084033614, |
| "grad_norm": 12.375, |
| "learning_rate": 1.2645539906103288e-05, |
| "loss": 0.9362, |
| "step": 12400 |
| }, |
| { |
| "epoch": 1.3035714285714286, |
| "grad_norm": 9.625, |
| "learning_rate": 1.2637715179968701e-05, |
| "loss": 1.2489, |
| "step": 12410 |
| }, |
| { |
| "epoch": 1.3046218487394958, |
| "grad_norm": 10.8125, |
| "learning_rate": 1.2629890453834117e-05, |
| "loss": 1.4674, |
| "step": 12420 |
| }, |
| { |
| "epoch": 1.305672268907563, |
| "grad_norm": 11.75, |
| "learning_rate": 1.262206572769953e-05, |
| "loss": 1.2837, |
| "step": 12430 |
| }, |
| { |
| "epoch": 1.3067226890756303, |
| "grad_norm": 4.46875, |
| "learning_rate": 1.2614241001564947e-05, |
| "loss": 1.1496, |
| "step": 12440 |
| }, |
| { |
| "epoch": 1.3077731092436975, |
| "grad_norm": 5.65625, |
| "learning_rate": 1.2606416275430361e-05, |
| "loss": 1.2523, |
| "step": 12450 |
| }, |
| { |
| "epoch": 1.3088235294117647, |
| "grad_norm": 14.875, |
| "learning_rate": 1.2598591549295776e-05, |
| "loss": 1.1827, |
| "step": 12460 |
| }, |
| { |
| "epoch": 1.309873949579832, |
| "grad_norm": 6.375, |
| "learning_rate": 1.259076682316119e-05, |
| "loss": 1.3326, |
| "step": 12470 |
| }, |
| { |
| "epoch": 1.3109243697478992, |
| "grad_norm": 14.0, |
| "learning_rate": 1.2582942097026605e-05, |
| "loss": 1.5072, |
| "step": 12480 |
| }, |
| { |
| "epoch": 1.3119747899159664, |
| "grad_norm": 17.5, |
| "learning_rate": 1.257511737089202e-05, |
| "loss": 1.0364, |
| "step": 12490 |
| }, |
| { |
| "epoch": 1.3130252100840336, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.2567292644757435e-05, |
| "loss": 1.4135, |
| "step": 12500 |
| }, |
| { |
| "epoch": 1.3140756302521008, |
| "grad_norm": 12.875, |
| "learning_rate": 1.255946791862285e-05, |
| "loss": 1.4885, |
| "step": 12510 |
| }, |
| { |
| "epoch": 1.315126050420168, |
| "grad_norm": 12.0, |
| "learning_rate": 1.2551643192488264e-05, |
| "loss": 1.3126, |
| "step": 12520 |
| }, |
| { |
| "epoch": 1.3161764705882353, |
| "grad_norm": 17.875, |
| "learning_rate": 1.2543818466353679e-05, |
| "loss": 1.2657, |
| "step": 12530 |
| }, |
| { |
| "epoch": 1.3172268907563025, |
| "grad_norm": 17.5, |
| "learning_rate": 1.2535993740219092e-05, |
| "loss": 1.389, |
| "step": 12540 |
| }, |
| { |
| "epoch": 1.3182773109243697, |
| "grad_norm": 17.125, |
| "learning_rate": 1.2528169014084508e-05, |
| "loss": 1.4456, |
| "step": 12550 |
| }, |
| { |
| "epoch": 1.319327731092437, |
| "grad_norm": 15.375, |
| "learning_rate": 1.2520344287949924e-05, |
| "loss": 1.2432, |
| "step": 12560 |
| }, |
| { |
| "epoch": 1.3203781512605042, |
| "grad_norm": 13.4375, |
| "learning_rate": 1.2512519561815337e-05, |
| "loss": 0.82, |
| "step": 12570 |
| }, |
| { |
| "epoch": 1.3214285714285714, |
| "grad_norm": 15.5625, |
| "learning_rate": 1.2504694835680754e-05, |
| "loss": 1.4444, |
| "step": 12580 |
| }, |
| { |
| "epoch": 1.3224789915966386, |
| "grad_norm": 12.625, |
| "learning_rate": 1.2496870109546166e-05, |
| "loss": 1.4157, |
| "step": 12590 |
| }, |
| { |
| "epoch": 1.3235294117647058, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.2489045383411583e-05, |
| "loss": 1.3183, |
| "step": 12600 |
| }, |
| { |
| "epoch": 1.324579831932773, |
| "grad_norm": 18.875, |
| "learning_rate": 1.2481220657276996e-05, |
| "loss": 1.5178, |
| "step": 12610 |
| }, |
| { |
| "epoch": 1.3256302521008403, |
| "grad_norm": 12.3125, |
| "learning_rate": 1.2473395931142412e-05, |
| "loss": 1.2979, |
| "step": 12620 |
| }, |
| { |
| "epoch": 1.3266806722689075, |
| "grad_norm": 11.75, |
| "learning_rate": 1.2465571205007825e-05, |
| "loss": 1.5851, |
| "step": 12630 |
| }, |
| { |
| "epoch": 1.3277310924369747, |
| "grad_norm": 6.15625, |
| "learning_rate": 1.245774647887324e-05, |
| "loss": 1.2124, |
| "step": 12640 |
| }, |
| { |
| "epoch": 1.328781512605042, |
| "grad_norm": 13.1875, |
| "learning_rate": 1.2449921752738656e-05, |
| "loss": 1.0869, |
| "step": 12650 |
| }, |
| { |
| "epoch": 1.3298319327731092, |
| "grad_norm": 50.5, |
| "learning_rate": 1.2442097026604069e-05, |
| "loss": 1.2768, |
| "step": 12660 |
| }, |
| { |
| "epoch": 1.3308823529411764, |
| "grad_norm": 18.125, |
| "learning_rate": 1.2434272300469485e-05, |
| "loss": 1.8448, |
| "step": 12670 |
| }, |
| { |
| "epoch": 1.3319327731092436, |
| "grad_norm": 16.25, |
| "learning_rate": 1.2426447574334898e-05, |
| "loss": 1.1748, |
| "step": 12680 |
| }, |
| { |
| "epoch": 1.3329831932773109, |
| "grad_norm": 14.875, |
| "learning_rate": 1.2418622848200315e-05, |
| "loss": 1.5983, |
| "step": 12690 |
| }, |
| { |
| "epoch": 1.334033613445378, |
| "grad_norm": 9.3125, |
| "learning_rate": 1.2410798122065728e-05, |
| "loss": 1.3208, |
| "step": 12700 |
| }, |
| { |
| "epoch": 1.3350840336134453, |
| "grad_norm": 12.625, |
| "learning_rate": 1.2402973395931144e-05, |
| "loss": 0.8592, |
| "step": 12710 |
| }, |
| { |
| "epoch": 1.3361344537815127, |
| "grad_norm": 5.15625, |
| "learning_rate": 1.2395148669796557e-05, |
| "loss": 1.002, |
| "step": 12720 |
| }, |
| { |
| "epoch": 1.33718487394958, |
| "grad_norm": 13.0625, |
| "learning_rate": 1.2387323943661973e-05, |
| "loss": 1.1317, |
| "step": 12730 |
| }, |
| { |
| "epoch": 1.3382352941176472, |
| "grad_norm": 20.375, |
| "learning_rate": 1.2379499217527386e-05, |
| "loss": 1.7459, |
| "step": 12740 |
| }, |
| { |
| "epoch": 1.3392857142857144, |
| "grad_norm": 7.5625, |
| "learning_rate": 1.2371674491392803e-05, |
| "loss": 1.2175, |
| "step": 12750 |
| }, |
| { |
| "epoch": 1.3403361344537816, |
| "grad_norm": 14.625, |
| "learning_rate": 1.2363849765258217e-05, |
| "loss": 1.3963, |
| "step": 12760 |
| }, |
| { |
| "epoch": 1.3413865546218489, |
| "grad_norm": 13.875, |
| "learning_rate": 1.2356025039123632e-05, |
| "loss": 1.1123, |
| "step": 12770 |
| }, |
| { |
| "epoch": 1.342436974789916, |
| "grad_norm": 9.0, |
| "learning_rate": 1.2348200312989047e-05, |
| "loss": 1.2574, |
| "step": 12780 |
| }, |
| { |
| "epoch": 1.3434873949579833, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.2340375586854461e-05, |
| "loss": 1.5698, |
| "step": 12790 |
| }, |
| { |
| "epoch": 1.3445378151260505, |
| "grad_norm": 12.375, |
| "learning_rate": 1.2332550860719876e-05, |
| "loss": 1.0556, |
| "step": 12800 |
| }, |
| { |
| "epoch": 1.3455882352941178, |
| "grad_norm": 2.953125, |
| "learning_rate": 1.232472613458529e-05, |
| "loss": 0.9861, |
| "step": 12810 |
| }, |
| { |
| "epoch": 1.346638655462185, |
| "grad_norm": 18.375, |
| "learning_rate": 1.2316901408450705e-05, |
| "loss": 1.039, |
| "step": 12820 |
| }, |
| { |
| "epoch": 1.3476890756302522, |
| "grad_norm": 18.625, |
| "learning_rate": 1.2309076682316118e-05, |
| "loss": 1.6725, |
| "step": 12830 |
| }, |
| { |
| "epoch": 1.3487394957983194, |
| "grad_norm": 12.3125, |
| "learning_rate": 1.2301251956181534e-05, |
| "loss": 0.8868, |
| "step": 12840 |
| }, |
| { |
| "epoch": 1.3497899159663866, |
| "grad_norm": 15.5, |
| "learning_rate": 1.229342723004695e-05, |
| "loss": 1.8068, |
| "step": 12850 |
| }, |
| { |
| "epoch": 1.3508403361344539, |
| "grad_norm": 10.1875, |
| "learning_rate": 1.2285602503912364e-05, |
| "loss": 1.4605, |
| "step": 12860 |
| }, |
| { |
| "epoch": 1.351890756302521, |
| "grad_norm": 2.21875, |
| "learning_rate": 1.227777777777778e-05, |
| "loss": 1.1635, |
| "step": 12870 |
| }, |
| { |
| "epoch": 1.3529411764705883, |
| "grad_norm": 12.3125, |
| "learning_rate": 1.2269953051643193e-05, |
| "loss": 1.4335, |
| "step": 12880 |
| }, |
| { |
| "epoch": 1.3539915966386555, |
| "grad_norm": 14.3125, |
| "learning_rate": 1.226212832550861e-05, |
| "loss": 1.5138, |
| "step": 12890 |
| }, |
| { |
| "epoch": 1.3550420168067228, |
| "grad_norm": 14.125, |
| "learning_rate": 1.2254303599374022e-05, |
| "loss": 1.5198, |
| "step": 12900 |
| }, |
| { |
| "epoch": 1.35609243697479, |
| "grad_norm": 5.3125, |
| "learning_rate": 1.2246478873239439e-05, |
| "loss": 1.2806, |
| "step": 12910 |
| }, |
| { |
| "epoch": 1.3571428571428572, |
| "grad_norm": 12.25, |
| "learning_rate": 1.2238654147104852e-05, |
| "loss": 1.7758, |
| "step": 12920 |
| }, |
| { |
| "epoch": 1.3581932773109244, |
| "grad_norm": 7.53125, |
| "learning_rate": 1.2230829420970266e-05, |
| "loss": 1.541, |
| "step": 12930 |
| }, |
| { |
| "epoch": 1.3592436974789917, |
| "grad_norm": 10.0625, |
| "learning_rate": 1.2223004694835683e-05, |
| "loss": 1.1173, |
| "step": 12940 |
| }, |
| { |
| "epoch": 1.3602941176470589, |
| "grad_norm": 11.625, |
| "learning_rate": 1.2215179968701096e-05, |
| "loss": 1.6626, |
| "step": 12950 |
| }, |
| { |
| "epoch": 1.361344537815126, |
| "grad_norm": 10.625, |
| "learning_rate": 1.2207355242566512e-05, |
| "loss": 1.4362, |
| "step": 12960 |
| }, |
| { |
| "epoch": 1.3623949579831933, |
| "grad_norm": 14.1875, |
| "learning_rate": 1.2199530516431925e-05, |
| "loss": 1.4961, |
| "step": 12970 |
| }, |
| { |
| "epoch": 1.3634453781512605, |
| "grad_norm": 12.25, |
| "learning_rate": 1.2191705790297341e-05, |
| "loss": 1.1132, |
| "step": 12980 |
| }, |
| { |
| "epoch": 1.3644957983193278, |
| "grad_norm": 5.53125, |
| "learning_rate": 1.2183881064162754e-05, |
| "loss": 1.3196, |
| "step": 12990 |
| }, |
| { |
| "epoch": 1.365546218487395, |
| "grad_norm": 5.34375, |
| "learning_rate": 1.217605633802817e-05, |
| "loss": 1.4392, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.3665966386554622, |
| "grad_norm": 3.9375, |
| "learning_rate": 1.2168231611893584e-05, |
| "loss": 1.4112, |
| "step": 13010 |
| }, |
| { |
| "epoch": 1.3676470588235294, |
| "grad_norm": 13.9375, |
| "learning_rate": 1.2160406885759e-05, |
| "loss": 1.3188, |
| "step": 13020 |
| }, |
| { |
| "epoch": 1.3686974789915967, |
| "grad_norm": 14.75, |
| "learning_rate": 1.2152582159624413e-05, |
| "loss": 1.1736, |
| "step": 13030 |
| }, |
| { |
| "epoch": 1.3697478991596639, |
| "grad_norm": 16.625, |
| "learning_rate": 1.214475743348983e-05, |
| "loss": 1.6647, |
| "step": 13040 |
| }, |
| { |
| "epoch": 1.370798319327731, |
| "grad_norm": 8.375, |
| "learning_rate": 1.2136932707355244e-05, |
| "loss": 1.325, |
| "step": 13050 |
| }, |
| { |
| "epoch": 1.3718487394957983, |
| "grad_norm": 18.125, |
| "learning_rate": 1.2129107981220659e-05, |
| "loss": 1.6822, |
| "step": 13060 |
| }, |
| { |
| "epoch": 1.3728991596638656, |
| "grad_norm": 12.375, |
| "learning_rate": 1.2121283255086073e-05, |
| "loss": 1.1331, |
| "step": 13070 |
| }, |
| { |
| "epoch": 1.3739495798319328, |
| "grad_norm": 4.375, |
| "learning_rate": 1.2113458528951488e-05, |
| "loss": 1.4808, |
| "step": 13080 |
| }, |
| { |
| "epoch": 1.375, |
| "grad_norm": 15.75, |
| "learning_rate": 1.2105633802816902e-05, |
| "loss": 1.4181, |
| "step": 13090 |
| }, |
| { |
| "epoch": 1.3760504201680672, |
| "grad_norm": 22.0, |
| "learning_rate": 1.2097809076682317e-05, |
| "loss": 1.7831, |
| "step": 13100 |
| }, |
| { |
| "epoch": 1.3771008403361344, |
| "grad_norm": 11.4375, |
| "learning_rate": 1.2089984350547732e-05, |
| "loss": 1.4553, |
| "step": 13110 |
| }, |
| { |
| "epoch": 1.3781512605042017, |
| "grad_norm": 13.625, |
| "learning_rate": 1.2082159624413146e-05, |
| "loss": 1.625, |
| "step": 13120 |
| }, |
| { |
| "epoch": 1.379201680672269, |
| "grad_norm": 16.125, |
| "learning_rate": 1.2074334898278561e-05, |
| "loss": 1.2871, |
| "step": 13130 |
| }, |
| { |
| "epoch": 1.3802521008403361, |
| "grad_norm": 15.625, |
| "learning_rate": 1.2066510172143977e-05, |
| "loss": 1.0753, |
| "step": 13140 |
| }, |
| { |
| "epoch": 1.3813025210084033, |
| "grad_norm": 13.875, |
| "learning_rate": 1.205868544600939e-05, |
| "loss": 1.3164, |
| "step": 13150 |
| }, |
| { |
| "epoch": 1.3823529411764706, |
| "grad_norm": 18.5, |
| "learning_rate": 1.2050860719874807e-05, |
| "loss": 1.3868, |
| "step": 13160 |
| }, |
| { |
| "epoch": 1.3834033613445378, |
| "grad_norm": 13.5625, |
| "learning_rate": 1.204303599374022e-05, |
| "loss": 1.3437, |
| "step": 13170 |
| }, |
| { |
| "epoch": 1.384453781512605, |
| "grad_norm": 13.0, |
| "learning_rate": 1.2035211267605636e-05, |
| "loss": 1.4609, |
| "step": 13180 |
| }, |
| { |
| "epoch": 1.3855042016806722, |
| "grad_norm": 7.65625, |
| "learning_rate": 1.2027386541471049e-05, |
| "loss": 1.4758, |
| "step": 13190 |
| }, |
| { |
| "epoch": 1.3865546218487395, |
| "grad_norm": 14.3125, |
| "learning_rate": 1.2019561815336465e-05, |
| "loss": 1.1268, |
| "step": 13200 |
| }, |
| { |
| "epoch": 1.3876050420168067, |
| "grad_norm": 4.65625, |
| "learning_rate": 1.2011737089201878e-05, |
| "loss": 1.0771, |
| "step": 13210 |
| }, |
| { |
| "epoch": 1.388655462184874, |
| "grad_norm": 17.125, |
| "learning_rate": 1.2003912363067293e-05, |
| "loss": 1.1369, |
| "step": 13220 |
| }, |
| { |
| "epoch": 1.3897058823529411, |
| "grad_norm": 17.5, |
| "learning_rate": 1.1996087636932708e-05, |
| "loss": 1.4544, |
| "step": 13230 |
| }, |
| { |
| "epoch": 1.3907563025210083, |
| "grad_norm": 18.875, |
| "learning_rate": 1.1988262910798122e-05, |
| "loss": 1.1155, |
| "step": 13240 |
| }, |
| { |
| "epoch": 1.3918067226890756, |
| "grad_norm": 12.8125, |
| "learning_rate": 1.1980438184663539e-05, |
| "loss": 0.973, |
| "step": 13250 |
| }, |
| { |
| "epoch": 1.3928571428571428, |
| "grad_norm": 19.75, |
| "learning_rate": 1.1972613458528952e-05, |
| "loss": 1.501, |
| "step": 13260 |
| }, |
| { |
| "epoch": 1.39390756302521, |
| "grad_norm": 12.625, |
| "learning_rate": 1.1964788732394368e-05, |
| "loss": 1.0422, |
| "step": 13270 |
| }, |
| { |
| "epoch": 1.3949579831932772, |
| "grad_norm": 19.625, |
| "learning_rate": 1.1956964006259781e-05, |
| "loss": 1.7695, |
| "step": 13280 |
| }, |
| { |
| "epoch": 1.3960084033613445, |
| "grad_norm": 9.9375, |
| "learning_rate": 1.1949139280125197e-05, |
| "loss": 0.7722, |
| "step": 13290 |
| }, |
| { |
| "epoch": 1.3970588235294117, |
| "grad_norm": 11.25, |
| "learning_rate": 1.194131455399061e-05, |
| "loss": 0.8371, |
| "step": 13300 |
| }, |
| { |
| "epoch": 1.398109243697479, |
| "grad_norm": 11.25, |
| "learning_rate": 1.1933489827856027e-05, |
| "loss": 1.4412, |
| "step": 13310 |
| }, |
| { |
| "epoch": 1.3991596638655461, |
| "grad_norm": 10.25, |
| "learning_rate": 1.192566510172144e-05, |
| "loss": 1.3016, |
| "step": 13320 |
| }, |
| { |
| "epoch": 1.4002100840336134, |
| "grad_norm": 12.3125, |
| "learning_rate": 1.1917840375586856e-05, |
| "loss": 1.5842, |
| "step": 13330 |
| }, |
| { |
| "epoch": 1.4012605042016806, |
| "grad_norm": 16.625, |
| "learning_rate": 1.191001564945227e-05, |
| "loss": 0.8069, |
| "step": 13340 |
| }, |
| { |
| "epoch": 1.4023109243697478, |
| "grad_norm": 13.75, |
| "learning_rate": 1.1902190923317685e-05, |
| "loss": 1.2223, |
| "step": 13350 |
| }, |
| { |
| "epoch": 1.403361344537815, |
| "grad_norm": 16.0, |
| "learning_rate": 1.18943661971831e-05, |
| "loss": 1.4568, |
| "step": 13360 |
| }, |
| { |
| "epoch": 1.4044117647058822, |
| "grad_norm": 5.625, |
| "learning_rate": 1.1886541471048514e-05, |
| "loss": 1.0528, |
| "step": 13370 |
| }, |
| { |
| "epoch": 1.4054621848739495, |
| "grad_norm": 11.875, |
| "learning_rate": 1.1878716744913929e-05, |
| "loss": 1.503, |
| "step": 13380 |
| }, |
| { |
| "epoch": 1.4065126050420167, |
| "grad_norm": 13.0625, |
| "learning_rate": 1.1870892018779344e-05, |
| "loss": 1.5821, |
| "step": 13390 |
| }, |
| { |
| "epoch": 1.407563025210084, |
| "grad_norm": 13.25, |
| "learning_rate": 1.1863067292644758e-05, |
| "loss": 1.2337, |
| "step": 13400 |
| }, |
| { |
| "epoch": 1.4086134453781511, |
| "grad_norm": 15.1875, |
| "learning_rate": 1.1855242566510173e-05, |
| "loss": 1.6133, |
| "step": 13410 |
| }, |
| { |
| "epoch": 1.4096638655462184, |
| "grad_norm": 15.875, |
| "learning_rate": 1.1847417840375588e-05, |
| "loss": 1.632, |
| "step": 13420 |
| }, |
| { |
| "epoch": 1.4107142857142856, |
| "grad_norm": 14.0625, |
| "learning_rate": 1.1839593114241e-05, |
| "loss": 1.0302, |
| "step": 13430 |
| }, |
| { |
| "epoch": 1.4117647058823528, |
| "grad_norm": 11.8125, |
| "learning_rate": 1.1831768388106417e-05, |
| "loss": 1.6045, |
| "step": 13440 |
| }, |
| { |
| "epoch": 1.41281512605042, |
| "grad_norm": 13.25, |
| "learning_rate": 1.1823943661971833e-05, |
| "loss": 1.2232, |
| "step": 13450 |
| }, |
| { |
| "epoch": 1.4138655462184873, |
| "grad_norm": 16.375, |
| "learning_rate": 1.1816118935837246e-05, |
| "loss": 1.2389, |
| "step": 13460 |
| }, |
| { |
| "epoch": 1.4149159663865547, |
| "grad_norm": 4.65625, |
| "learning_rate": 1.1808294209702663e-05, |
| "loss": 1.2232, |
| "step": 13470 |
| }, |
| { |
| "epoch": 1.415966386554622, |
| "grad_norm": 3.671875, |
| "learning_rate": 1.1800469483568076e-05, |
| "loss": 1.2644, |
| "step": 13480 |
| }, |
| { |
| "epoch": 1.4170168067226891, |
| "grad_norm": 11.3125, |
| "learning_rate": 1.1792644757433492e-05, |
| "loss": 1.3765, |
| "step": 13490 |
| }, |
| { |
| "epoch": 1.4180672268907564, |
| "grad_norm": 11.5625, |
| "learning_rate": 1.1784820031298905e-05, |
| "loss": 1.1586, |
| "step": 13500 |
| }, |
| { |
| "epoch": 1.4191176470588236, |
| "grad_norm": 13.625, |
| "learning_rate": 1.177699530516432e-05, |
| "loss": 0.9487, |
| "step": 13510 |
| }, |
| { |
| "epoch": 1.4201680672268908, |
| "grad_norm": 25.125, |
| "learning_rate": 1.1769170579029734e-05, |
| "loss": 0.9073, |
| "step": 13520 |
| }, |
| { |
| "epoch": 1.421218487394958, |
| "grad_norm": 20.875, |
| "learning_rate": 1.1761345852895149e-05, |
| "loss": 1.4868, |
| "step": 13530 |
| }, |
| { |
| "epoch": 1.4222689075630253, |
| "grad_norm": 22.25, |
| "learning_rate": 1.1753521126760565e-05, |
| "loss": 1.6157, |
| "step": 13540 |
| }, |
| { |
| "epoch": 1.4233193277310925, |
| "grad_norm": 18.0, |
| "learning_rate": 1.1745696400625978e-05, |
| "loss": 1.4386, |
| "step": 13550 |
| }, |
| { |
| "epoch": 1.4243697478991597, |
| "grad_norm": 15.25, |
| "learning_rate": 1.1737871674491395e-05, |
| "loss": 1.4988, |
| "step": 13560 |
| }, |
| { |
| "epoch": 1.425420168067227, |
| "grad_norm": 17.125, |
| "learning_rate": 1.1730046948356808e-05, |
| "loss": 0.8417, |
| "step": 13570 |
| }, |
| { |
| "epoch": 1.4264705882352942, |
| "grad_norm": 18.125, |
| "learning_rate": 1.1722222222222224e-05, |
| "loss": 1.6105, |
| "step": 13580 |
| }, |
| { |
| "epoch": 1.4275210084033614, |
| "grad_norm": 13.5, |
| "learning_rate": 1.1714397496087637e-05, |
| "loss": 1.5684, |
| "step": 13590 |
| }, |
| { |
| "epoch": 1.4285714285714286, |
| "grad_norm": 6.84375, |
| "learning_rate": 1.1706572769953053e-05, |
| "loss": 1.2026, |
| "step": 13600 |
| }, |
| { |
| "epoch": 1.4296218487394958, |
| "grad_norm": 16.125, |
| "learning_rate": 1.1698748043818466e-05, |
| "loss": 1.4824, |
| "step": 13610 |
| }, |
| { |
| "epoch": 1.430672268907563, |
| "grad_norm": 4.59375, |
| "learning_rate": 1.1690923317683882e-05, |
| "loss": 1.2055, |
| "step": 13620 |
| }, |
| { |
| "epoch": 1.4317226890756303, |
| "grad_norm": 14.0, |
| "learning_rate": 1.1683098591549297e-05, |
| "loss": 1.2775, |
| "step": 13630 |
| }, |
| { |
| "epoch": 1.4327731092436975, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.1675273865414712e-05, |
| "loss": 1.0015, |
| "step": 13640 |
| }, |
| { |
| "epoch": 1.4338235294117647, |
| "grad_norm": 16.25, |
| "learning_rate": 1.1667449139280126e-05, |
| "loss": 1.6765, |
| "step": 13650 |
| }, |
| { |
| "epoch": 1.434873949579832, |
| "grad_norm": 13.8125, |
| "learning_rate": 1.1659624413145541e-05, |
| "loss": 1.5888, |
| "step": 13660 |
| }, |
| { |
| "epoch": 1.4359243697478992, |
| "grad_norm": 14.5, |
| "learning_rate": 1.1651799687010956e-05, |
| "loss": 1.382, |
| "step": 13670 |
| }, |
| { |
| "epoch": 1.4369747899159664, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.164397496087637e-05, |
| "loss": 1.1734, |
| "step": 13680 |
| }, |
| { |
| "epoch": 1.4380252100840336, |
| "grad_norm": 13.375, |
| "learning_rate": 1.1636150234741785e-05, |
| "loss": 1.091, |
| "step": 13690 |
| }, |
| { |
| "epoch": 1.4390756302521008, |
| "grad_norm": 8.3125, |
| "learning_rate": 1.16283255086072e-05, |
| "loss": 1.7296, |
| "step": 13700 |
| }, |
| { |
| "epoch": 1.440126050420168, |
| "grad_norm": 4.59375, |
| "learning_rate": 1.1620500782472614e-05, |
| "loss": 0.7291, |
| "step": 13710 |
| }, |
| { |
| "epoch": 1.4411764705882353, |
| "grad_norm": 17.5, |
| "learning_rate": 1.1612676056338027e-05, |
| "loss": 1.3262, |
| "step": 13720 |
| }, |
| { |
| "epoch": 1.4422268907563025, |
| "grad_norm": 11.5625, |
| "learning_rate": 1.1604851330203444e-05, |
| "loss": 1.1583, |
| "step": 13730 |
| }, |
| { |
| "epoch": 1.4432773109243697, |
| "grad_norm": 20.75, |
| "learning_rate": 1.159702660406886e-05, |
| "loss": 1.6833, |
| "step": 13740 |
| }, |
| { |
| "epoch": 1.444327731092437, |
| "grad_norm": 20.0, |
| "learning_rate": 1.1589201877934273e-05, |
| "loss": 1.3845, |
| "step": 13750 |
| }, |
| { |
| "epoch": 1.4453781512605042, |
| "grad_norm": 12.5, |
| "learning_rate": 1.158137715179969e-05, |
| "loss": 0.8729, |
| "step": 13760 |
| }, |
| { |
| "epoch": 1.4464285714285714, |
| "grad_norm": 18.5, |
| "learning_rate": 1.1573552425665102e-05, |
| "loss": 1.3549, |
| "step": 13770 |
| }, |
| { |
| "epoch": 1.4474789915966386, |
| "grad_norm": 16.625, |
| "learning_rate": 1.1565727699530519e-05, |
| "loss": 1.3161, |
| "step": 13780 |
| }, |
| { |
| "epoch": 1.4485294117647058, |
| "grad_norm": 22.25, |
| "learning_rate": 1.1557902973395932e-05, |
| "loss": 1.431, |
| "step": 13790 |
| }, |
| { |
| "epoch": 1.449579831932773, |
| "grad_norm": 5.5, |
| "learning_rate": 1.1550078247261346e-05, |
| "loss": 1.1607, |
| "step": 13800 |
| }, |
| { |
| "epoch": 1.4506302521008403, |
| "grad_norm": 13.625, |
| "learning_rate": 1.1542253521126761e-05, |
| "loss": 1.1485, |
| "step": 13810 |
| }, |
| { |
| "epoch": 1.4516806722689075, |
| "grad_norm": 13.9375, |
| "learning_rate": 1.1534428794992176e-05, |
| "loss": 1.3236, |
| "step": 13820 |
| }, |
| { |
| "epoch": 1.4527310924369747, |
| "grad_norm": 15.5625, |
| "learning_rate": 1.1526604068857592e-05, |
| "loss": 1.0652, |
| "step": 13830 |
| }, |
| { |
| "epoch": 1.453781512605042, |
| "grad_norm": 15.0, |
| "learning_rate": 1.1518779342723005e-05, |
| "loss": 1.6018, |
| "step": 13840 |
| }, |
| { |
| "epoch": 1.4548319327731092, |
| "grad_norm": 11.375, |
| "learning_rate": 1.1510954616588421e-05, |
| "loss": 0.9, |
| "step": 13850 |
| }, |
| { |
| "epoch": 1.4558823529411764, |
| "grad_norm": 4.375, |
| "learning_rate": 1.1503129890453834e-05, |
| "loss": 0.8587, |
| "step": 13860 |
| }, |
| { |
| "epoch": 1.4569327731092436, |
| "grad_norm": 13.125, |
| "learning_rate": 1.149530516431925e-05, |
| "loss": 1.1876, |
| "step": 13870 |
| }, |
| { |
| "epoch": 1.4579831932773109, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.1487480438184663e-05, |
| "loss": 1.221, |
| "step": 13880 |
| }, |
| { |
| "epoch": 1.459033613445378, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.147965571205008e-05, |
| "loss": 1.2262, |
| "step": 13890 |
| }, |
| { |
| "epoch": 1.4600840336134453, |
| "grad_norm": 13.625, |
| "learning_rate": 1.1471830985915493e-05, |
| "loss": 1.3682, |
| "step": 13900 |
| }, |
| { |
| "epoch": 1.4611344537815127, |
| "grad_norm": 15.6875, |
| "learning_rate": 1.1464006259780909e-05, |
| "loss": 1.432, |
| "step": 13910 |
| }, |
| { |
| "epoch": 1.46218487394958, |
| "grad_norm": 39.75, |
| "learning_rate": 1.1456181533646322e-05, |
| "loss": 1.1011, |
| "step": 13920 |
| }, |
| { |
| "epoch": 1.4632352941176472, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.1448356807511738e-05, |
| "loss": 0.961, |
| "step": 13930 |
| }, |
| { |
| "epoch": 1.4642857142857144, |
| "grad_norm": 12.375, |
| "learning_rate": 1.1440532081377153e-05, |
| "loss": 0.9199, |
| "step": 13940 |
| }, |
| { |
| "epoch": 1.4653361344537816, |
| "grad_norm": 15.5, |
| "learning_rate": 1.1432707355242568e-05, |
| "loss": 1.3411, |
| "step": 13950 |
| }, |
| { |
| "epoch": 1.4663865546218489, |
| "grad_norm": 14.8125, |
| "learning_rate": 1.1424882629107982e-05, |
| "loss": 1.3674, |
| "step": 13960 |
| }, |
| { |
| "epoch": 1.467436974789916, |
| "grad_norm": 6.53125, |
| "learning_rate": 1.1417057902973397e-05, |
| "loss": 1.1447, |
| "step": 13970 |
| }, |
| { |
| "epoch": 1.4684873949579833, |
| "grad_norm": 14.875, |
| "learning_rate": 1.1409233176838812e-05, |
| "loss": 1.4201, |
| "step": 13980 |
| }, |
| { |
| "epoch": 1.4695378151260505, |
| "grad_norm": 3.953125, |
| "learning_rate": 1.1401408450704226e-05, |
| "loss": 1.3417, |
| "step": 13990 |
| }, |
| { |
| "epoch": 1.4705882352941178, |
| "grad_norm": 8.5625, |
| "learning_rate": 1.1393583724569641e-05, |
| "loss": 0.9248, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.471638655462185, |
| "grad_norm": 19.875, |
| "learning_rate": 1.1385758998435054e-05, |
| "loss": 1.7956, |
| "step": 14010 |
| }, |
| { |
| "epoch": 1.4726890756302522, |
| "grad_norm": 18.75, |
| "learning_rate": 1.137793427230047e-05, |
| "loss": 1.2636, |
| "step": 14020 |
| }, |
| { |
| "epoch": 1.4737394957983194, |
| "grad_norm": 4.78125, |
| "learning_rate": 1.1370109546165887e-05, |
| "loss": 1.4248, |
| "step": 14030 |
| }, |
| { |
| "epoch": 1.4747899159663866, |
| "grad_norm": 4.25, |
| "learning_rate": 1.13622848200313e-05, |
| "loss": 1.1417, |
| "step": 14040 |
| }, |
| { |
| "epoch": 1.4758403361344539, |
| "grad_norm": 23.25, |
| "learning_rate": 1.1354460093896716e-05, |
| "loss": 1.0123, |
| "step": 14050 |
| }, |
| { |
| "epoch": 1.476890756302521, |
| "grad_norm": 4.34375, |
| "learning_rate": 1.1346635367762129e-05, |
| "loss": 1.5176, |
| "step": 14060 |
| }, |
| { |
| "epoch": 1.4779411764705883, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.1338810641627545e-05, |
| "loss": 1.6103, |
| "step": 14070 |
| }, |
| { |
| "epoch": 1.4789915966386555, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.1330985915492958e-05, |
| "loss": 2.1391, |
| "step": 14080 |
| }, |
| { |
| "epoch": 1.4800420168067228, |
| "grad_norm": 4.625, |
| "learning_rate": 1.1323161189358375e-05, |
| "loss": 1.3055, |
| "step": 14090 |
| }, |
| { |
| "epoch": 1.48109243697479, |
| "grad_norm": 12.4375, |
| "learning_rate": 1.1315336463223788e-05, |
| "loss": 1.306, |
| "step": 14100 |
| }, |
| { |
| "epoch": 1.4821428571428572, |
| "grad_norm": 16.0, |
| "learning_rate": 1.1307511737089202e-05, |
| "loss": 1.2485, |
| "step": 14110 |
| }, |
| { |
| "epoch": 1.4831932773109244, |
| "grad_norm": 22.0, |
| "learning_rate": 1.1299687010954619e-05, |
| "loss": 1.2959, |
| "step": 14120 |
| }, |
| { |
| "epoch": 1.4842436974789917, |
| "grad_norm": 18.125, |
| "learning_rate": 1.1291862284820031e-05, |
| "loss": 1.419, |
| "step": 14130 |
| }, |
| { |
| "epoch": 1.4852941176470589, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.1284037558685448e-05, |
| "loss": 1.2413, |
| "step": 14140 |
| }, |
| { |
| "epoch": 1.486344537815126, |
| "grad_norm": 13.75, |
| "learning_rate": 1.127621283255086e-05, |
| "loss": 1.1521, |
| "step": 14150 |
| }, |
| { |
| "epoch": 1.4873949579831933, |
| "grad_norm": 6.71875, |
| "learning_rate": 1.1268388106416277e-05, |
| "loss": 1.0156, |
| "step": 14160 |
| }, |
| { |
| "epoch": 1.4884453781512605, |
| "grad_norm": 5.40625, |
| "learning_rate": 1.126056338028169e-05, |
| "loss": 1.4527, |
| "step": 14170 |
| }, |
| { |
| "epoch": 1.4894957983193278, |
| "grad_norm": 14.0, |
| "learning_rate": 1.1252738654147106e-05, |
| "loss": 1.3498, |
| "step": 14180 |
| }, |
| { |
| "epoch": 1.490546218487395, |
| "grad_norm": 4.1875, |
| "learning_rate": 1.124491392801252e-05, |
| "loss": 1.7654, |
| "step": 14190 |
| }, |
| { |
| "epoch": 1.4915966386554622, |
| "grad_norm": 13.4375, |
| "learning_rate": 1.1237089201877936e-05, |
| "loss": 1.0088, |
| "step": 14200 |
| }, |
| { |
| "epoch": 1.4926470588235294, |
| "grad_norm": 13.5, |
| "learning_rate": 1.1229264475743349e-05, |
| "loss": 1.1831, |
| "step": 14210 |
| }, |
| { |
| "epoch": 1.4936974789915967, |
| "grad_norm": 11.5, |
| "learning_rate": 1.1221439749608765e-05, |
| "loss": 1.472, |
| "step": 14220 |
| }, |
| { |
| "epoch": 1.4947478991596639, |
| "grad_norm": 21.5, |
| "learning_rate": 1.121361502347418e-05, |
| "loss": 1.0679, |
| "step": 14230 |
| }, |
| { |
| "epoch": 1.495798319327731, |
| "grad_norm": 20.0, |
| "learning_rate": 1.1205790297339594e-05, |
| "loss": 1.1635, |
| "step": 14240 |
| }, |
| { |
| "epoch": 1.4968487394957983, |
| "grad_norm": 12.25, |
| "learning_rate": 1.1197965571205009e-05, |
| "loss": 1.3566, |
| "step": 14250 |
| }, |
| { |
| "epoch": 1.4978991596638656, |
| "grad_norm": 12.25, |
| "learning_rate": 1.1190140845070424e-05, |
| "loss": 1.3279, |
| "step": 14260 |
| }, |
| { |
| "epoch": 1.4989495798319328, |
| "grad_norm": 18.125, |
| "learning_rate": 1.1182316118935838e-05, |
| "loss": 1.6923, |
| "step": 14270 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 22.375, |
| "learning_rate": 1.1174491392801253e-05, |
| "loss": 1.9083, |
| "step": 14280 |
| }, |
| { |
| "epoch": 1.5010504201680672, |
| "grad_norm": 10.4375, |
| "learning_rate": 1.1166666666666668e-05, |
| "loss": 1.1433, |
| "step": 14290 |
| }, |
| { |
| "epoch": 1.5021008403361344, |
| "grad_norm": 8.9375, |
| "learning_rate": 1.115884194053208e-05, |
| "loss": 1.1706, |
| "step": 14300 |
| }, |
| { |
| "epoch": 1.5031512605042017, |
| "grad_norm": 18.75, |
| "learning_rate": 1.1151017214397497e-05, |
| "loss": 1.2276, |
| "step": 14310 |
| }, |
| { |
| "epoch": 1.504201680672269, |
| "grad_norm": 5.125, |
| "learning_rate": 1.1143192488262913e-05, |
| "loss": 1.0959, |
| "step": 14320 |
| }, |
| { |
| "epoch": 1.5052521008403361, |
| "grad_norm": 12.375, |
| "learning_rate": 1.1135367762128326e-05, |
| "loss": 1.2093, |
| "step": 14330 |
| }, |
| { |
| "epoch": 1.5063025210084033, |
| "grad_norm": 5.6875, |
| "learning_rate": 1.1127543035993743e-05, |
| "loss": 1.4079, |
| "step": 14340 |
| }, |
| { |
| "epoch": 1.5073529411764706, |
| "grad_norm": 16.875, |
| "learning_rate": 1.1119718309859156e-05, |
| "loss": 1.677, |
| "step": 14350 |
| }, |
| { |
| "epoch": 1.5084033613445378, |
| "grad_norm": 13.625, |
| "learning_rate": 1.1111893583724572e-05, |
| "loss": 1.8673, |
| "step": 14360 |
| }, |
| { |
| "epoch": 1.509453781512605, |
| "grad_norm": 26.125, |
| "learning_rate": 1.1104068857589985e-05, |
| "loss": 1.7775, |
| "step": 14370 |
| }, |
| { |
| "epoch": 1.5105042016806722, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.1096244131455401e-05, |
| "loss": 1.238, |
| "step": 14380 |
| }, |
| { |
| "epoch": 1.5115546218487395, |
| "grad_norm": 34.5, |
| "learning_rate": 1.1088419405320814e-05, |
| "loss": 1.2416, |
| "step": 14390 |
| }, |
| { |
| "epoch": 1.5126050420168067, |
| "grad_norm": 6.25, |
| "learning_rate": 1.1080594679186229e-05, |
| "loss": 1.3153, |
| "step": 14400 |
| }, |
| { |
| "epoch": 1.513655462184874, |
| "grad_norm": 19.25, |
| "learning_rate": 1.1072769953051643e-05, |
| "loss": 1.5286, |
| "step": 14410 |
| }, |
| { |
| "epoch": 1.5147058823529411, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.1064945226917058e-05, |
| "loss": 1.6686, |
| "step": 14420 |
| }, |
| { |
| "epoch": 1.5157563025210083, |
| "grad_norm": 13.5, |
| "learning_rate": 1.1057120500782474e-05, |
| "loss": 1.2533, |
| "step": 14430 |
| }, |
| { |
| "epoch": 1.5168067226890756, |
| "grad_norm": 13.25, |
| "learning_rate": 1.1049295774647887e-05, |
| "loss": 1.0326, |
| "step": 14440 |
| }, |
| { |
| "epoch": 1.5178571428571428, |
| "grad_norm": 8.4375, |
| "learning_rate": 1.1041471048513304e-05, |
| "loss": 1.0656, |
| "step": 14450 |
| }, |
| { |
| "epoch": 1.51890756302521, |
| "grad_norm": 13.8125, |
| "learning_rate": 1.1033646322378717e-05, |
| "loss": 1.2902, |
| "step": 14460 |
| }, |
| { |
| "epoch": 1.5199579831932772, |
| "grad_norm": 17.875, |
| "learning_rate": 1.1025821596244133e-05, |
| "loss": 1.4418, |
| "step": 14470 |
| }, |
| { |
| "epoch": 1.5210084033613445, |
| "grad_norm": 14.4375, |
| "learning_rate": 1.1017996870109546e-05, |
| "loss": 0.9938, |
| "step": 14480 |
| }, |
| { |
| "epoch": 1.5220588235294117, |
| "grad_norm": 9.25, |
| "learning_rate": 1.1010172143974962e-05, |
| "loss": 0.9951, |
| "step": 14490 |
| }, |
| { |
| "epoch": 1.523109243697479, |
| "grad_norm": 5.3125, |
| "learning_rate": 1.1002347417840375e-05, |
| "loss": 1.5267, |
| "step": 14500 |
| }, |
| { |
| "epoch": 1.5241596638655461, |
| "grad_norm": 14.375, |
| "learning_rate": 1.0994522691705792e-05, |
| "loss": 1.1844, |
| "step": 14510 |
| }, |
| { |
| "epoch": 1.5252100840336134, |
| "grad_norm": 22.625, |
| "learning_rate": 1.0986697965571206e-05, |
| "loss": 1.5373, |
| "step": 14520 |
| }, |
| { |
| "epoch": 1.5262605042016806, |
| "grad_norm": 15.9375, |
| "learning_rate": 1.0978873239436621e-05, |
| "loss": 1.5569, |
| "step": 14530 |
| }, |
| { |
| "epoch": 1.5273109243697478, |
| "grad_norm": 14.0, |
| "learning_rate": 1.0971048513302036e-05, |
| "loss": 1.4319, |
| "step": 14540 |
| }, |
| { |
| "epoch": 1.528361344537815, |
| "grad_norm": 4.53125, |
| "learning_rate": 1.096322378716745e-05, |
| "loss": 1.3139, |
| "step": 14550 |
| }, |
| { |
| "epoch": 1.5294117647058822, |
| "grad_norm": 20.25, |
| "learning_rate": 1.0955399061032865e-05, |
| "loss": 1.7041, |
| "step": 14560 |
| }, |
| { |
| "epoch": 1.5304621848739495, |
| "grad_norm": 17.875, |
| "learning_rate": 1.094757433489828e-05, |
| "loss": 1.3248, |
| "step": 14570 |
| }, |
| { |
| "epoch": 1.5315126050420167, |
| "grad_norm": 6.0, |
| "learning_rate": 1.0939749608763694e-05, |
| "loss": 0.9728, |
| "step": 14580 |
| }, |
| { |
| "epoch": 1.532563025210084, |
| "grad_norm": 15.1875, |
| "learning_rate": 1.0931924882629107e-05, |
| "loss": 1.0219, |
| "step": 14590 |
| }, |
| { |
| "epoch": 1.5336134453781511, |
| "grad_norm": 21.875, |
| "learning_rate": 1.0924100156494524e-05, |
| "loss": 1.2215, |
| "step": 14600 |
| }, |
| { |
| "epoch": 1.5346638655462184, |
| "grad_norm": 16.5, |
| "learning_rate": 1.0916275430359936e-05, |
| "loss": 1.3004, |
| "step": 14610 |
| }, |
| { |
| "epoch": 1.5357142857142856, |
| "grad_norm": 5.59375, |
| "learning_rate": 1.0908450704225353e-05, |
| "loss": 1.0769, |
| "step": 14620 |
| }, |
| { |
| "epoch": 1.5367647058823528, |
| "grad_norm": 17.875, |
| "learning_rate": 1.090062597809077e-05, |
| "loss": 1.2735, |
| "step": 14630 |
| }, |
| { |
| "epoch": 1.53781512605042, |
| "grad_norm": 20.25, |
| "learning_rate": 1.0892801251956182e-05, |
| "loss": 1.1559, |
| "step": 14640 |
| }, |
| { |
| "epoch": 1.5388655462184873, |
| "grad_norm": 28.75, |
| "learning_rate": 1.0884976525821598e-05, |
| "loss": 1.9156, |
| "step": 14650 |
| }, |
| { |
| "epoch": 1.5399159663865545, |
| "grad_norm": 15.4375, |
| "learning_rate": 1.0877151799687011e-05, |
| "loss": 1.2164, |
| "step": 14660 |
| }, |
| { |
| "epoch": 1.5409663865546217, |
| "grad_norm": 16.5, |
| "learning_rate": 1.0869327073552428e-05, |
| "loss": 1.8587, |
| "step": 14670 |
| }, |
| { |
| "epoch": 1.542016806722689, |
| "grad_norm": 7.15625, |
| "learning_rate": 1.086150234741784e-05, |
| "loss": 1.185, |
| "step": 14680 |
| }, |
| { |
| "epoch": 1.5430672268907561, |
| "grad_norm": 14.5, |
| "learning_rate": 1.0853677621283255e-05, |
| "loss": 1.1462, |
| "step": 14690 |
| }, |
| { |
| "epoch": 1.5441176470588234, |
| "grad_norm": 10.5, |
| "learning_rate": 1.084585289514867e-05, |
| "loss": 1.1659, |
| "step": 14700 |
| }, |
| { |
| "epoch": 1.5451680672268906, |
| "grad_norm": 40.75, |
| "learning_rate": 1.0838028169014085e-05, |
| "loss": 1.4489, |
| "step": 14710 |
| }, |
| { |
| "epoch": 1.5462184873949578, |
| "grad_norm": 15.875, |
| "learning_rate": 1.0830203442879501e-05, |
| "loss": 1.2954, |
| "step": 14720 |
| }, |
| { |
| "epoch": 1.5472689075630253, |
| "grad_norm": 4.3125, |
| "learning_rate": 1.0822378716744914e-05, |
| "loss": 1.5098, |
| "step": 14730 |
| }, |
| { |
| "epoch": 1.5483193277310925, |
| "grad_norm": 51.5, |
| "learning_rate": 1.081455399061033e-05, |
| "loss": 1.268, |
| "step": 14740 |
| }, |
| { |
| "epoch": 1.5493697478991597, |
| "grad_norm": 17.75, |
| "learning_rate": 1.0806729264475743e-05, |
| "loss": 1.5536, |
| "step": 14750 |
| }, |
| { |
| "epoch": 1.550420168067227, |
| "grad_norm": 14.8125, |
| "learning_rate": 1.079890453834116e-05, |
| "loss": 1.1089, |
| "step": 14760 |
| }, |
| { |
| "epoch": 1.5514705882352942, |
| "grad_norm": 9.75, |
| "learning_rate": 1.0791079812206573e-05, |
| "loss": 1.3357, |
| "step": 14770 |
| }, |
| { |
| "epoch": 1.5525210084033614, |
| "grad_norm": 4.65625, |
| "learning_rate": 1.0783255086071989e-05, |
| "loss": 1.5138, |
| "step": 14780 |
| }, |
| { |
| "epoch": 1.5535714285714286, |
| "grad_norm": 42.5, |
| "learning_rate": 1.0775430359937402e-05, |
| "loss": 1.6304, |
| "step": 14790 |
| }, |
| { |
| "epoch": 1.5546218487394958, |
| "grad_norm": 13.25, |
| "learning_rate": 1.0767605633802818e-05, |
| "loss": 1.5183, |
| "step": 14800 |
| }, |
| { |
| "epoch": 1.555672268907563, |
| "grad_norm": 13.0, |
| "learning_rate": 1.0759780907668233e-05, |
| "loss": 1.4326, |
| "step": 14810 |
| }, |
| { |
| "epoch": 1.5567226890756303, |
| "grad_norm": 4.84375, |
| "learning_rate": 1.0751956181533648e-05, |
| "loss": 0.8125, |
| "step": 14820 |
| }, |
| { |
| "epoch": 1.5577731092436975, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.0744131455399062e-05, |
| "loss": 1.1498, |
| "step": 14830 |
| }, |
| { |
| "epoch": 1.5588235294117647, |
| "grad_norm": 5.5625, |
| "learning_rate": 1.0736306729264477e-05, |
| "loss": 1.069, |
| "step": 14840 |
| }, |
| { |
| "epoch": 1.559873949579832, |
| "grad_norm": 15.0, |
| "learning_rate": 1.0728482003129892e-05, |
| "loss": 1.6658, |
| "step": 14850 |
| }, |
| { |
| "epoch": 1.5609243697478992, |
| "grad_norm": 19.375, |
| "learning_rate": 1.0720657276995306e-05, |
| "loss": 1.3619, |
| "step": 14860 |
| }, |
| { |
| "epoch": 1.5619747899159664, |
| "grad_norm": 1312.0, |
| "learning_rate": 1.0712832550860721e-05, |
| "loss": 1.1617, |
| "step": 14870 |
| }, |
| { |
| "epoch": 1.5630252100840336, |
| "grad_norm": 9.875, |
| "learning_rate": 1.0705007824726134e-05, |
| "loss": 1.1236, |
| "step": 14880 |
| }, |
| { |
| "epoch": 1.5640756302521008, |
| "grad_norm": 17.0, |
| "learning_rate": 1.069718309859155e-05, |
| "loss": 1.0404, |
| "step": 14890 |
| }, |
| { |
| "epoch": 1.565126050420168, |
| "grad_norm": 12.9375, |
| "learning_rate": 1.0689358372456963e-05, |
| "loss": 1.2872, |
| "step": 14900 |
| }, |
| { |
| "epoch": 1.5661764705882353, |
| "grad_norm": 12.75, |
| "learning_rate": 1.068153364632238e-05, |
| "loss": 1.3784, |
| "step": 14910 |
| }, |
| { |
| "epoch": 1.5672268907563025, |
| "grad_norm": 6.34375, |
| "learning_rate": 1.0673708920187796e-05, |
| "loss": 1.1165, |
| "step": 14920 |
| }, |
| { |
| "epoch": 1.5682773109243697, |
| "grad_norm": 22.625, |
| "learning_rate": 1.0665884194053209e-05, |
| "loss": 1.2642, |
| "step": 14930 |
| }, |
| { |
| "epoch": 1.569327731092437, |
| "grad_norm": 18.5, |
| "learning_rate": 1.0658059467918625e-05, |
| "loss": 1.3787, |
| "step": 14940 |
| }, |
| { |
| "epoch": 1.5703781512605042, |
| "grad_norm": 11.3125, |
| "learning_rate": 1.0650234741784038e-05, |
| "loss": 1.1048, |
| "step": 14950 |
| }, |
| { |
| "epoch": 1.5714285714285714, |
| "grad_norm": 20.75, |
| "learning_rate": 1.0642410015649454e-05, |
| "loss": 1.2615, |
| "step": 14960 |
| }, |
| { |
| "epoch": 1.5724789915966386, |
| "grad_norm": 19.875, |
| "learning_rate": 1.0634585289514867e-05, |
| "loss": 0.7653, |
| "step": 14970 |
| }, |
| { |
| "epoch": 1.5735294117647058, |
| "grad_norm": 17.875, |
| "learning_rate": 1.0626760563380282e-05, |
| "loss": 1.3239, |
| "step": 14980 |
| }, |
| { |
| "epoch": 1.574579831932773, |
| "grad_norm": 18.375, |
| "learning_rate": 1.0618935837245697e-05, |
| "loss": 1.7513, |
| "step": 14990 |
| }, |
| { |
| "epoch": 1.5756302521008403, |
| "grad_norm": 17.0, |
| "learning_rate": 1.0611111111111111e-05, |
| "loss": 1.4967, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.5766806722689075, |
| "grad_norm": 17.25, |
| "learning_rate": 1.0603286384976528e-05, |
| "loss": 1.2066, |
| "step": 15010 |
| }, |
| { |
| "epoch": 1.5777310924369747, |
| "grad_norm": 5.65625, |
| "learning_rate": 1.059546165884194e-05, |
| "loss": 1.2133, |
| "step": 15020 |
| }, |
| { |
| "epoch": 1.5787815126050422, |
| "grad_norm": 13.5625, |
| "learning_rate": 1.0587636932707357e-05, |
| "loss": 1.6947, |
| "step": 15030 |
| }, |
| { |
| "epoch": 1.5798319327731094, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.057981220657277e-05, |
| "loss": 1.1693, |
| "step": 15040 |
| }, |
| { |
| "epoch": 1.5808823529411766, |
| "grad_norm": 18.75, |
| "learning_rate": 1.0571987480438186e-05, |
| "loss": 1.1638, |
| "step": 15050 |
| }, |
| { |
| "epoch": 1.5819327731092439, |
| "grad_norm": 13.1875, |
| "learning_rate": 1.05641627543036e-05, |
| "loss": 1.203, |
| "step": 15060 |
| }, |
| { |
| "epoch": 1.582983193277311, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.0556338028169016e-05, |
| "loss": 1.6161, |
| "step": 15070 |
| }, |
| { |
| "epoch": 1.5840336134453783, |
| "grad_norm": 11.625, |
| "learning_rate": 1.0548513302034429e-05, |
| "loss": 1.8032, |
| "step": 15080 |
| }, |
| { |
| "epoch": 1.5850840336134455, |
| "grad_norm": 12.125, |
| "learning_rate": 1.0540688575899845e-05, |
| "loss": 1.3926, |
| "step": 15090 |
| }, |
| { |
| "epoch": 1.5861344537815127, |
| "grad_norm": 13.25, |
| "learning_rate": 1.0532863849765258e-05, |
| "loss": 1.7814, |
| "step": 15100 |
| }, |
| { |
| "epoch": 1.58718487394958, |
| "grad_norm": 13.4375, |
| "learning_rate": 1.0525039123630674e-05, |
| "loss": 1.0973, |
| "step": 15110 |
| }, |
| { |
| "epoch": 1.5882352941176472, |
| "grad_norm": 12.875, |
| "learning_rate": 1.0517214397496089e-05, |
| "loss": 1.2523, |
| "step": 15120 |
| }, |
| { |
| "epoch": 1.5892857142857144, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.0509389671361504e-05, |
| "loss": 1.5806, |
| "step": 15130 |
| }, |
| { |
| "epoch": 1.5903361344537816, |
| "grad_norm": 5.53125, |
| "learning_rate": 1.0501564945226918e-05, |
| "loss": 1.2682, |
| "step": 15140 |
| }, |
| { |
| "epoch": 1.5913865546218489, |
| "grad_norm": 21.0, |
| "learning_rate": 1.0493740219092333e-05, |
| "loss": 1.1658, |
| "step": 15150 |
| }, |
| { |
| "epoch": 1.592436974789916, |
| "grad_norm": 9.9375, |
| "learning_rate": 1.0485915492957747e-05, |
| "loss": 1.0203, |
| "step": 15160 |
| }, |
| { |
| "epoch": 1.5934873949579833, |
| "grad_norm": 15.8125, |
| "learning_rate": 1.047809076682316e-05, |
| "loss": 1.3342, |
| "step": 15170 |
| }, |
| { |
| "epoch": 1.5945378151260505, |
| "grad_norm": 19.625, |
| "learning_rate": 1.0470266040688577e-05, |
| "loss": 1.0889, |
| "step": 15180 |
| }, |
| { |
| "epoch": 1.5955882352941178, |
| "grad_norm": 8.5625, |
| "learning_rate": 1.046244131455399e-05, |
| "loss": 1.4284, |
| "step": 15190 |
| }, |
| { |
| "epoch": 1.596638655462185, |
| "grad_norm": 24.375, |
| "learning_rate": 1.0454616588419406e-05, |
| "loss": 1.3882, |
| "step": 15200 |
| }, |
| { |
| "epoch": 1.5976890756302522, |
| "grad_norm": 14.9375, |
| "learning_rate": 1.0446791862284822e-05, |
| "loss": 1.2336, |
| "step": 15210 |
| }, |
| { |
| "epoch": 1.5987394957983194, |
| "grad_norm": 4.0, |
| "learning_rate": 1.0438967136150235e-05, |
| "loss": 0.8792, |
| "step": 15220 |
| }, |
| { |
| "epoch": 1.5997899159663866, |
| "grad_norm": 14.3125, |
| "learning_rate": 1.0431142410015652e-05, |
| "loss": 1.4848, |
| "step": 15230 |
| }, |
| { |
| "epoch": 1.6008403361344539, |
| "grad_norm": 16.625, |
| "learning_rate": 1.0423317683881065e-05, |
| "loss": 1.7036, |
| "step": 15240 |
| }, |
| { |
| "epoch": 1.601890756302521, |
| "grad_norm": 12.625, |
| "learning_rate": 1.0415492957746481e-05, |
| "loss": 1.0141, |
| "step": 15250 |
| }, |
| { |
| "epoch": 1.6029411764705883, |
| "grad_norm": 9.4375, |
| "learning_rate": 1.0407668231611894e-05, |
| "loss": 1.3178, |
| "step": 15260 |
| }, |
| { |
| "epoch": 1.6039915966386555, |
| "grad_norm": 5.28125, |
| "learning_rate": 1.0399843505477309e-05, |
| "loss": 0.9306, |
| "step": 15270 |
| }, |
| { |
| "epoch": 1.6050420168067228, |
| "grad_norm": 16.125, |
| "learning_rate": 1.0392018779342723e-05, |
| "loss": 1.1392, |
| "step": 15280 |
| }, |
| { |
| "epoch": 1.60609243697479, |
| "grad_norm": 8.5625, |
| "learning_rate": 1.0384194053208138e-05, |
| "loss": 0.8321, |
| "step": 15290 |
| }, |
| { |
| "epoch": 1.6071428571428572, |
| "grad_norm": 13.3125, |
| "learning_rate": 1.0376369327073554e-05, |
| "loss": 0.957, |
| "step": 15300 |
| }, |
| { |
| "epoch": 1.6081932773109244, |
| "grad_norm": 17.125, |
| "learning_rate": 1.0368544600938967e-05, |
| "loss": 1.4487, |
| "step": 15310 |
| }, |
| { |
| "epoch": 1.6092436974789917, |
| "grad_norm": 11.6875, |
| "learning_rate": 1.0360719874804384e-05, |
| "loss": 1.2441, |
| "step": 15320 |
| }, |
| { |
| "epoch": 1.6102941176470589, |
| "grad_norm": 16.0, |
| "learning_rate": 1.0352895148669797e-05, |
| "loss": 1.313, |
| "step": 15330 |
| }, |
| { |
| "epoch": 1.611344537815126, |
| "grad_norm": 5.96875, |
| "learning_rate": 1.0345070422535213e-05, |
| "loss": 1.4231, |
| "step": 15340 |
| }, |
| { |
| "epoch": 1.6123949579831933, |
| "grad_norm": 10.125, |
| "learning_rate": 1.0337245696400626e-05, |
| "loss": 1.2076, |
| "step": 15350 |
| }, |
| { |
| "epoch": 1.6134453781512605, |
| "grad_norm": 17.125, |
| "learning_rate": 1.0329420970266042e-05, |
| "loss": 1.6903, |
| "step": 15360 |
| }, |
| { |
| "epoch": 1.6144957983193278, |
| "grad_norm": 5.3125, |
| "learning_rate": 1.0321596244131455e-05, |
| "loss": 0.9224, |
| "step": 15370 |
| }, |
| { |
| "epoch": 1.615546218487395, |
| "grad_norm": 9.9375, |
| "learning_rate": 1.0313771517996872e-05, |
| "loss": 1.5575, |
| "step": 15380 |
| }, |
| { |
| "epoch": 1.6165966386554622, |
| "grad_norm": 10.25, |
| "learning_rate": 1.0305946791862285e-05, |
| "loss": 1.2729, |
| "step": 15390 |
| }, |
| { |
| "epoch": 1.6176470588235294, |
| "grad_norm": 9.0, |
| "learning_rate": 1.0298122065727701e-05, |
| "loss": 1.1269, |
| "step": 15400 |
| }, |
| { |
| "epoch": 1.6186974789915967, |
| "grad_norm": 17.75, |
| "learning_rate": 1.0290297339593116e-05, |
| "loss": 1.5302, |
| "step": 15410 |
| }, |
| { |
| "epoch": 1.6197478991596639, |
| "grad_norm": 13.75, |
| "learning_rate": 1.028247261345853e-05, |
| "loss": 1.2133, |
| "step": 15420 |
| }, |
| { |
| "epoch": 1.620798319327731, |
| "grad_norm": 5.625, |
| "learning_rate": 1.0274647887323945e-05, |
| "loss": 0.8847, |
| "step": 15430 |
| }, |
| { |
| "epoch": 1.6218487394957983, |
| "grad_norm": 20.0, |
| "learning_rate": 1.026682316118936e-05, |
| "loss": 1.1911, |
| "step": 15440 |
| }, |
| { |
| "epoch": 1.6228991596638656, |
| "grad_norm": 8.6875, |
| "learning_rate": 1.0258998435054774e-05, |
| "loss": 1.1295, |
| "step": 15450 |
| }, |
| { |
| "epoch": 1.6239495798319328, |
| "grad_norm": 21.0, |
| "learning_rate": 1.0251173708920189e-05, |
| "loss": 1.133, |
| "step": 15460 |
| }, |
| { |
| "epoch": 1.625, |
| "grad_norm": 11.1875, |
| "learning_rate": 1.0243348982785603e-05, |
| "loss": 2.0244, |
| "step": 15470 |
| }, |
| { |
| "epoch": 1.6260504201680672, |
| "grad_norm": 15.1875, |
| "learning_rate": 1.0235524256651016e-05, |
| "loss": 1.7165, |
| "step": 15480 |
| }, |
| { |
| "epoch": 1.6271008403361344, |
| "grad_norm": 6.0625, |
| "learning_rate": 1.0227699530516433e-05, |
| "loss": 1.1485, |
| "step": 15490 |
| }, |
| { |
| "epoch": 1.6281512605042017, |
| "grad_norm": 17.875, |
| "learning_rate": 1.0219874804381849e-05, |
| "loss": 0.9418, |
| "step": 15500 |
| }, |
| { |
| "epoch": 1.629201680672269, |
| "grad_norm": 17.875, |
| "learning_rate": 1.0212050078247262e-05, |
| "loss": 1.9413, |
| "step": 15510 |
| }, |
| { |
| "epoch": 1.6302521008403361, |
| "grad_norm": 9.5, |
| "learning_rate": 1.0204225352112678e-05, |
| "loss": 1.4225, |
| "step": 15520 |
| }, |
| { |
| "epoch": 1.6313025210084033, |
| "grad_norm": 14.0625, |
| "learning_rate": 1.0196400625978091e-05, |
| "loss": 1.3062, |
| "step": 15530 |
| }, |
| { |
| "epoch": 1.6323529411764706, |
| "grad_norm": 11.75, |
| "learning_rate": 1.0188575899843508e-05, |
| "loss": 1.0521, |
| "step": 15540 |
| }, |
| { |
| "epoch": 1.6334033613445378, |
| "grad_norm": 13.1875, |
| "learning_rate": 1.018075117370892e-05, |
| "loss": 1.7687, |
| "step": 15550 |
| }, |
| { |
| "epoch": 1.634453781512605, |
| "grad_norm": 15.875, |
| "learning_rate": 1.0172926447574335e-05, |
| "loss": 1.1349, |
| "step": 15560 |
| }, |
| { |
| "epoch": 1.6355042016806722, |
| "grad_norm": 23.5, |
| "learning_rate": 1.016510172143975e-05, |
| "loss": 1.6808, |
| "step": 15570 |
| }, |
| { |
| "epoch": 1.6365546218487395, |
| "grad_norm": 22.875, |
| "learning_rate": 1.0157276995305165e-05, |
| "loss": 1.6487, |
| "step": 15580 |
| }, |
| { |
| "epoch": 1.6376050420168067, |
| "grad_norm": 20.875, |
| "learning_rate": 1.014945226917058e-05, |
| "loss": 1.1881, |
| "step": 15590 |
| }, |
| { |
| "epoch": 1.638655462184874, |
| "grad_norm": 12.5625, |
| "learning_rate": 1.0141627543035994e-05, |
| "loss": 1.3333, |
| "step": 15600 |
| }, |
| { |
| "epoch": 1.6397058823529411, |
| "grad_norm": 19.125, |
| "learning_rate": 1.013380281690141e-05, |
| "loss": 1.3079, |
| "step": 15610 |
| }, |
| { |
| "epoch": 1.6407563025210083, |
| "grad_norm": 17.875, |
| "learning_rate": 1.0125978090766823e-05, |
| "loss": 1.3545, |
| "step": 15620 |
| }, |
| { |
| "epoch": 1.6418067226890756, |
| "grad_norm": 16.5, |
| "learning_rate": 1.011815336463224e-05, |
| "loss": 1.3019, |
| "step": 15630 |
| }, |
| { |
| "epoch": 1.6428571428571428, |
| "grad_norm": 15.25, |
| "learning_rate": 1.0110328638497653e-05, |
| "loss": 1.2696, |
| "step": 15640 |
| }, |
| { |
| "epoch": 1.64390756302521, |
| "grad_norm": 17.125, |
| "learning_rate": 1.0102503912363069e-05, |
| "loss": 1.5522, |
| "step": 15650 |
| }, |
| { |
| "epoch": 1.6449579831932772, |
| "grad_norm": 13.8125, |
| "learning_rate": 1.0094679186228482e-05, |
| "loss": 1.0805, |
| "step": 15660 |
| }, |
| { |
| "epoch": 1.6460084033613445, |
| "grad_norm": 11.875, |
| "learning_rate": 1.0086854460093898e-05, |
| "loss": 1.3991, |
| "step": 15670 |
| }, |
| { |
| "epoch": 1.6470588235294117, |
| "grad_norm": 14.125, |
| "learning_rate": 1.0079029733959311e-05, |
| "loss": 1.4535, |
| "step": 15680 |
| }, |
| { |
| "epoch": 1.648109243697479, |
| "grad_norm": 12.375, |
| "learning_rate": 1.0071205007824727e-05, |
| "loss": 1.3899, |
| "step": 15690 |
| }, |
| { |
| "epoch": 1.6491596638655461, |
| "grad_norm": 6.0625, |
| "learning_rate": 1.0063380281690142e-05, |
| "loss": 1.293, |
| "step": 15700 |
| }, |
| { |
| "epoch": 1.6502100840336134, |
| "grad_norm": 11.5, |
| "learning_rate": 1.0055555555555557e-05, |
| "loss": 1.3171, |
| "step": 15710 |
| }, |
| { |
| "epoch": 1.6512605042016806, |
| "grad_norm": 14.875, |
| "learning_rate": 1.0047730829420971e-05, |
| "loss": 1.7538, |
| "step": 15720 |
| }, |
| { |
| "epoch": 1.6523109243697478, |
| "grad_norm": 12.1875, |
| "learning_rate": 1.0039906103286386e-05, |
| "loss": 1.9343, |
| "step": 15730 |
| }, |
| { |
| "epoch": 1.653361344537815, |
| "grad_norm": 38.0, |
| "learning_rate": 1.00320813771518e-05, |
| "loss": 1.2496, |
| "step": 15740 |
| }, |
| { |
| "epoch": 1.6544117647058822, |
| "grad_norm": 5.5625, |
| "learning_rate": 1.0024256651017215e-05, |
| "loss": 1.0771, |
| "step": 15750 |
| }, |
| { |
| "epoch": 1.6554621848739495, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.001643192488263e-05, |
| "loss": 1.5892, |
| "step": 15760 |
| }, |
| { |
| "epoch": 1.6565126050420167, |
| "grad_norm": 18.75, |
| "learning_rate": 1.0008607198748043e-05, |
| "loss": 0.826, |
| "step": 15770 |
| }, |
| { |
| "epoch": 1.657563025210084, |
| "grad_norm": 12.6875, |
| "learning_rate": 1.000078247261346e-05, |
| "loss": 1.4061, |
| "step": 15780 |
| }, |
| { |
| "epoch": 1.6586134453781511, |
| "grad_norm": 12.4375, |
| "learning_rate": 9.992957746478874e-06, |
| "loss": 0.9534, |
| "step": 15790 |
| }, |
| { |
| "epoch": 1.6596638655462184, |
| "grad_norm": 12.625, |
| "learning_rate": 9.985133020344289e-06, |
| "loss": 1.2486, |
| "step": 15800 |
| }, |
| { |
| "epoch": 1.6607142857142856, |
| "grad_norm": 16.75, |
| "learning_rate": 9.977308294209703e-06, |
| "loss": 1.5199, |
| "step": 15810 |
| }, |
| { |
| "epoch": 1.6617647058823528, |
| "grad_norm": 16.625, |
| "learning_rate": 9.969483568075118e-06, |
| "loss": 1.8439, |
| "step": 15820 |
| }, |
| { |
| "epoch": 1.66281512605042, |
| "grad_norm": 17.375, |
| "learning_rate": 9.961658841940533e-06, |
| "loss": 1.3802, |
| "step": 15830 |
| }, |
| { |
| "epoch": 1.6638655462184873, |
| "grad_norm": 5.4375, |
| "learning_rate": 9.953834115805949e-06, |
| "loss": 1.2319, |
| "step": 15840 |
| }, |
| { |
| "epoch": 1.6649159663865545, |
| "grad_norm": 11.0, |
| "learning_rate": 9.946009389671362e-06, |
| "loss": 1.0634, |
| "step": 15850 |
| }, |
| { |
| "epoch": 1.6659663865546217, |
| "grad_norm": 19.125, |
| "learning_rate": 9.938184663536777e-06, |
| "loss": 0.9874, |
| "step": 15860 |
| }, |
| { |
| "epoch": 1.667016806722689, |
| "grad_norm": 15.1875, |
| "learning_rate": 9.930359937402191e-06, |
| "loss": 1.1704, |
| "step": 15870 |
| }, |
| { |
| "epoch": 1.6680672268907561, |
| "grad_norm": 5.03125, |
| "learning_rate": 9.922535211267606e-06, |
| "loss": 1.3389, |
| "step": 15880 |
| }, |
| { |
| "epoch": 1.6691176470588234, |
| "grad_norm": 16.625, |
| "learning_rate": 9.91471048513302e-06, |
| "loss": 1.357, |
| "step": 15890 |
| }, |
| { |
| "epoch": 1.6701680672268906, |
| "grad_norm": 12.75, |
| "learning_rate": 9.906885758998435e-06, |
| "loss": 1.5127, |
| "step": 15900 |
| }, |
| { |
| "epoch": 1.6712184873949578, |
| "grad_norm": 12.3125, |
| "learning_rate": 9.89906103286385e-06, |
| "loss": 1.2742, |
| "step": 15910 |
| }, |
| { |
| "epoch": 1.6722689075630253, |
| "grad_norm": 12.1875, |
| "learning_rate": 9.891236306729264e-06, |
| "loss": 1.429, |
| "step": 15920 |
| }, |
| { |
| "epoch": 1.6733193277310925, |
| "grad_norm": 19.75, |
| "learning_rate": 9.883411580594679e-06, |
| "loss": 0.8065, |
| "step": 15930 |
| }, |
| { |
| "epoch": 1.6743697478991597, |
| "grad_norm": 16.5, |
| "learning_rate": 9.875586854460095e-06, |
| "loss": 1.213, |
| "step": 15940 |
| }, |
| { |
| "epoch": 1.675420168067227, |
| "grad_norm": 28.875, |
| "learning_rate": 9.86776212832551e-06, |
| "loss": 1.1768, |
| "step": 15950 |
| }, |
| { |
| "epoch": 1.6764705882352942, |
| "grad_norm": 13.75, |
| "learning_rate": 9.859937402190925e-06, |
| "loss": 1.6273, |
| "step": 15960 |
| }, |
| { |
| "epoch": 1.6775210084033614, |
| "grad_norm": 13.875, |
| "learning_rate": 9.85211267605634e-06, |
| "loss": 1.6125, |
| "step": 15970 |
| }, |
| { |
| "epoch": 1.6785714285714286, |
| "grad_norm": 12.6875, |
| "learning_rate": 9.844287949921754e-06, |
| "loss": 1.5806, |
| "step": 15980 |
| }, |
| { |
| "epoch": 1.6796218487394958, |
| "grad_norm": 12.375, |
| "learning_rate": 9.836463223787169e-06, |
| "loss": 1.8286, |
| "step": 15990 |
| }, |
| { |
| "epoch": 1.680672268907563, |
| "grad_norm": 13.9375, |
| "learning_rate": 9.828638497652583e-06, |
| "loss": 0.9962, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.6817226890756303, |
| "grad_norm": 13.1875, |
| "learning_rate": 9.820813771517998e-06, |
| "loss": 1.2894, |
| "step": 16010 |
| }, |
| { |
| "epoch": 1.6827731092436975, |
| "grad_norm": 10.125, |
| "learning_rate": 9.812989045383413e-06, |
| "loss": 1.0479, |
| "step": 16020 |
| }, |
| { |
| "epoch": 1.6838235294117647, |
| "grad_norm": 14.0625, |
| "learning_rate": 9.805164319248827e-06, |
| "loss": 1.313, |
| "step": 16030 |
| }, |
| { |
| "epoch": 1.684873949579832, |
| "grad_norm": 18.0, |
| "learning_rate": 9.797339593114242e-06, |
| "loss": 1.4493, |
| "step": 16040 |
| }, |
| { |
| "epoch": 1.6859243697478992, |
| "grad_norm": 20.125, |
| "learning_rate": 9.789514866979657e-06, |
| "loss": 1.5301, |
| "step": 16050 |
| }, |
| { |
| "epoch": 1.6869747899159664, |
| "grad_norm": 15.625, |
| "learning_rate": 9.781690140845071e-06, |
| "loss": 1.2022, |
| "step": 16060 |
| }, |
| { |
| "epoch": 1.6880252100840336, |
| "grad_norm": 13.0625, |
| "learning_rate": 9.773865414710486e-06, |
| "loss": 1.3654, |
| "step": 16070 |
| }, |
| { |
| "epoch": 1.6890756302521008, |
| "grad_norm": 15.6875, |
| "learning_rate": 9.7660406885759e-06, |
| "loss": 1.0885, |
| "step": 16080 |
| }, |
| { |
| "epoch": 1.690126050420168, |
| "grad_norm": 15.25, |
| "learning_rate": 9.758215962441315e-06, |
| "loss": 1.5678, |
| "step": 16090 |
| }, |
| { |
| "epoch": 1.6911764705882353, |
| "grad_norm": 19.875, |
| "learning_rate": 9.75039123630673e-06, |
| "loss": 1.2938, |
| "step": 16100 |
| }, |
| { |
| "epoch": 1.6922268907563025, |
| "grad_norm": 10.8125, |
| "learning_rate": 9.742566510172145e-06, |
| "loss": 1.2581, |
| "step": 16110 |
| }, |
| { |
| "epoch": 1.6932773109243697, |
| "grad_norm": 6.0, |
| "learning_rate": 9.73474178403756e-06, |
| "loss": 1.1328, |
| "step": 16120 |
| }, |
| { |
| "epoch": 1.694327731092437, |
| "grad_norm": 10.1875, |
| "learning_rate": 9.726917057902974e-06, |
| "loss": 1.191, |
| "step": 16130 |
| }, |
| { |
| "epoch": 1.6953781512605042, |
| "grad_norm": 22.0, |
| "learning_rate": 9.71909233176839e-06, |
| "loss": 1.1682, |
| "step": 16140 |
| }, |
| { |
| "epoch": 1.6964285714285714, |
| "grad_norm": 12.0625, |
| "learning_rate": 9.711267605633803e-06, |
| "loss": 1.4293, |
| "step": 16150 |
| }, |
| { |
| "epoch": 1.6974789915966386, |
| "grad_norm": 11.6875, |
| "learning_rate": 9.703442879499218e-06, |
| "loss": 1.5305, |
| "step": 16160 |
| }, |
| { |
| "epoch": 1.6985294117647058, |
| "grad_norm": 13.5625, |
| "learning_rate": 9.695618153364633e-06, |
| "loss": 1.1146, |
| "step": 16170 |
| }, |
| { |
| "epoch": 1.699579831932773, |
| "grad_norm": 6.1875, |
| "learning_rate": 9.687793427230047e-06, |
| "loss": 0.9118, |
| "step": 16180 |
| }, |
| { |
| "epoch": 1.7006302521008403, |
| "grad_norm": 13.6875, |
| "learning_rate": 9.679968701095462e-06, |
| "loss": 1.4518, |
| "step": 16190 |
| }, |
| { |
| "epoch": 1.7016806722689075, |
| "grad_norm": 19.5, |
| "learning_rate": 9.672143974960876e-06, |
| "loss": 1.4933, |
| "step": 16200 |
| }, |
| { |
| "epoch": 1.7027310924369747, |
| "grad_norm": 209.0, |
| "learning_rate": 9.664319248826291e-06, |
| "loss": 1.5073, |
| "step": 16210 |
| }, |
| { |
| "epoch": 1.7037815126050422, |
| "grad_norm": 12.0625, |
| "learning_rate": 9.656494522691706e-06, |
| "loss": 1.0963, |
| "step": 16220 |
| }, |
| { |
| "epoch": 1.7048319327731094, |
| "grad_norm": 11.875, |
| "learning_rate": 9.64866979655712e-06, |
| "loss": 1.6323, |
| "step": 16230 |
| }, |
| { |
| "epoch": 1.7058823529411766, |
| "grad_norm": 12.125, |
| "learning_rate": 9.640845070422537e-06, |
| "loss": 1.0694, |
| "step": 16240 |
| }, |
| { |
| "epoch": 1.7069327731092439, |
| "grad_norm": 24.375, |
| "learning_rate": 9.633020344287951e-06, |
| "loss": 1.269, |
| "step": 16250 |
| }, |
| { |
| "epoch": 1.707983193277311, |
| "grad_norm": 10.8125, |
| "learning_rate": 9.625195618153366e-06, |
| "loss": 1.628, |
| "step": 16260 |
| }, |
| { |
| "epoch": 1.7090336134453783, |
| "grad_norm": 17.625, |
| "learning_rate": 9.61737089201878e-06, |
| "loss": 1.8879, |
| "step": 16270 |
| }, |
| { |
| "epoch": 1.7100840336134455, |
| "grad_norm": 23.75, |
| "learning_rate": 9.609546165884195e-06, |
| "loss": 1.1923, |
| "step": 16280 |
| }, |
| { |
| "epoch": 1.7111344537815127, |
| "grad_norm": 17.125, |
| "learning_rate": 9.60172143974961e-06, |
| "loss": 1.2534, |
| "step": 16290 |
| }, |
| { |
| "epoch": 1.71218487394958, |
| "grad_norm": 8.4375, |
| "learning_rate": 9.593896713615025e-06, |
| "loss": 1.1554, |
| "step": 16300 |
| }, |
| { |
| "epoch": 1.7132352941176472, |
| "grad_norm": 13.0625, |
| "learning_rate": 9.58607198748044e-06, |
| "loss": 1.0627, |
| "step": 16310 |
| }, |
| { |
| "epoch": 1.7142857142857144, |
| "grad_norm": 10.8125, |
| "learning_rate": 9.578247261345854e-06, |
| "loss": 1.0352, |
| "step": 16320 |
| }, |
| { |
| "epoch": 1.7153361344537816, |
| "grad_norm": 10.0, |
| "learning_rate": 9.570422535211269e-06, |
| "loss": 0.916, |
| "step": 16330 |
| }, |
| { |
| "epoch": 1.7163865546218489, |
| "grad_norm": 14.875, |
| "learning_rate": 9.562597809076683e-06, |
| "loss": 1.0697, |
| "step": 16340 |
| }, |
| { |
| "epoch": 1.717436974789916, |
| "grad_norm": 11.5, |
| "learning_rate": 9.554773082942098e-06, |
| "loss": 1.1456, |
| "step": 16350 |
| }, |
| { |
| "epoch": 1.7184873949579833, |
| "grad_norm": 11.9375, |
| "learning_rate": 9.546948356807513e-06, |
| "loss": 1.2569, |
| "step": 16360 |
| }, |
| { |
| "epoch": 1.7195378151260505, |
| "grad_norm": 19.125, |
| "learning_rate": 9.539123630672927e-06, |
| "loss": 1.0225, |
| "step": 16370 |
| }, |
| { |
| "epoch": 1.7205882352941178, |
| "grad_norm": 6.65625, |
| "learning_rate": 9.531298904538342e-06, |
| "loss": 0.8316, |
| "step": 16380 |
| }, |
| { |
| "epoch": 1.721638655462185, |
| "grad_norm": 9.5, |
| "learning_rate": 9.523474178403757e-06, |
| "loss": 1.6279, |
| "step": 16390 |
| }, |
| { |
| "epoch": 1.7226890756302522, |
| "grad_norm": 20.625, |
| "learning_rate": 9.515649452269171e-06, |
| "loss": 1.3566, |
| "step": 16400 |
| }, |
| { |
| "epoch": 1.7237394957983194, |
| "grad_norm": 5.0625, |
| "learning_rate": 9.507824726134586e-06, |
| "loss": 1.3098, |
| "step": 16410 |
| }, |
| { |
| "epoch": 1.7247899159663866, |
| "grad_norm": 18.125, |
| "learning_rate": 9.5e-06, |
| "loss": 1.2304, |
| "step": 16420 |
| }, |
| { |
| "epoch": 1.7258403361344539, |
| "grad_norm": 16.75, |
| "learning_rate": 9.492175273865417e-06, |
| "loss": 1.2579, |
| "step": 16430 |
| }, |
| { |
| "epoch": 1.726890756302521, |
| "grad_norm": 12.75, |
| "learning_rate": 9.48435054773083e-06, |
| "loss": 1.2478, |
| "step": 16440 |
| }, |
| { |
| "epoch": 1.7279411764705883, |
| "grad_norm": 9.8125, |
| "learning_rate": 9.476525821596244e-06, |
| "loss": 1.3145, |
| "step": 16450 |
| }, |
| { |
| "epoch": 1.7289915966386555, |
| "grad_norm": 15.0, |
| "learning_rate": 9.468701095461659e-06, |
| "loss": 1.2866, |
| "step": 16460 |
| }, |
| { |
| "epoch": 1.7300420168067228, |
| "grad_norm": 13.4375, |
| "learning_rate": 9.460876369327074e-06, |
| "loss": 1.5644, |
| "step": 16470 |
| }, |
| { |
| "epoch": 1.73109243697479, |
| "grad_norm": 12.5, |
| "learning_rate": 9.453051643192488e-06, |
| "loss": 1.2161, |
| "step": 16480 |
| }, |
| { |
| "epoch": 1.7321428571428572, |
| "grad_norm": 18.875, |
| "learning_rate": 9.445226917057903e-06, |
| "loss": 1.2717, |
| "step": 16490 |
| }, |
| { |
| "epoch": 1.7331932773109244, |
| "grad_norm": 11.375, |
| "learning_rate": 9.437402190923318e-06, |
| "loss": 1.7148, |
| "step": 16500 |
| }, |
| { |
| "epoch": 1.7342436974789917, |
| "grad_norm": 18.375, |
| "learning_rate": 9.429577464788732e-06, |
| "loss": 1.5477, |
| "step": 16510 |
| }, |
| { |
| "epoch": 1.7352941176470589, |
| "grad_norm": 12.3125, |
| "learning_rate": 9.421752738654147e-06, |
| "loss": 1.2824, |
| "step": 16520 |
| }, |
| { |
| "epoch": 1.736344537815126, |
| "grad_norm": 18.125, |
| "learning_rate": 9.413928012519563e-06, |
| "loss": 1.3925, |
| "step": 16530 |
| }, |
| { |
| "epoch": 1.7373949579831933, |
| "grad_norm": 4.625, |
| "learning_rate": 9.406103286384978e-06, |
| "loss": 1.3081, |
| "step": 16540 |
| }, |
| { |
| "epoch": 1.7384453781512605, |
| "grad_norm": 17.5, |
| "learning_rate": 9.398278560250393e-06, |
| "loss": 1.089, |
| "step": 16550 |
| }, |
| { |
| "epoch": 1.7394957983193278, |
| "grad_norm": 15.75, |
| "learning_rate": 9.390453834115807e-06, |
| "loss": 0.8492, |
| "step": 16560 |
| }, |
| { |
| "epoch": 1.740546218487395, |
| "grad_norm": 203.0, |
| "learning_rate": 9.382629107981222e-06, |
| "loss": 1.4314, |
| "step": 16570 |
| }, |
| { |
| "epoch": 1.7415966386554622, |
| "grad_norm": 18.25, |
| "learning_rate": 9.374804381846637e-06, |
| "loss": 1.9853, |
| "step": 16580 |
| }, |
| { |
| "epoch": 1.7426470588235294, |
| "grad_norm": 21.75, |
| "learning_rate": 9.366979655712051e-06, |
| "loss": 1.6503, |
| "step": 16590 |
| }, |
| { |
| "epoch": 1.7436974789915967, |
| "grad_norm": 9.1875, |
| "learning_rate": 9.359154929577466e-06, |
| "loss": 1.2783, |
| "step": 16600 |
| }, |
| { |
| "epoch": 1.7447478991596639, |
| "grad_norm": 12.6875, |
| "learning_rate": 9.35133020344288e-06, |
| "loss": 1.5263, |
| "step": 16610 |
| }, |
| { |
| "epoch": 1.745798319327731, |
| "grad_norm": 15.375, |
| "learning_rate": 9.343505477308295e-06, |
| "loss": 1.5407, |
| "step": 16620 |
| }, |
| { |
| "epoch": 1.7468487394957983, |
| "grad_norm": 19.75, |
| "learning_rate": 9.33568075117371e-06, |
| "loss": 1.0901, |
| "step": 16630 |
| }, |
| { |
| "epoch": 1.7478991596638656, |
| "grad_norm": 14.3125, |
| "learning_rate": 9.327856025039125e-06, |
| "loss": 1.2885, |
| "step": 16640 |
| }, |
| { |
| "epoch": 1.7489495798319328, |
| "grad_norm": 12.625, |
| "learning_rate": 9.32003129890454e-06, |
| "loss": 1.5592, |
| "step": 16650 |
| }, |
| { |
| "epoch": 1.75, |
| "grad_norm": 8.75, |
| "learning_rate": 9.312206572769954e-06, |
| "loss": 1.2404, |
| "step": 16660 |
| }, |
| { |
| "epoch": 1.7510504201680672, |
| "grad_norm": 13.0, |
| "learning_rate": 9.304381846635369e-06, |
| "loss": 1.0363, |
| "step": 16670 |
| }, |
| { |
| "epoch": 1.7521008403361344, |
| "grad_norm": 12.25, |
| "learning_rate": 9.296557120500783e-06, |
| "loss": 1.1992, |
| "step": 16680 |
| }, |
| { |
| "epoch": 1.7531512605042017, |
| "grad_norm": 14.0, |
| "learning_rate": 9.288732394366198e-06, |
| "loss": 1.1861, |
| "step": 16690 |
| }, |
| { |
| "epoch": 1.754201680672269, |
| "grad_norm": 19.125, |
| "learning_rate": 9.280907668231612e-06, |
| "loss": 1.2102, |
| "step": 16700 |
| }, |
| { |
| "epoch": 1.7552521008403361, |
| "grad_norm": 17.75, |
| "learning_rate": 9.273082942097027e-06, |
| "loss": 1.7202, |
| "step": 16710 |
| }, |
| { |
| "epoch": 1.7563025210084033, |
| "grad_norm": 17.25, |
| "learning_rate": 9.265258215962442e-06, |
| "loss": 1.3753, |
| "step": 16720 |
| }, |
| { |
| "epoch": 1.7573529411764706, |
| "grad_norm": 17.625, |
| "learning_rate": 9.257433489827856e-06, |
| "loss": 1.3309, |
| "step": 16730 |
| }, |
| { |
| "epoch": 1.7584033613445378, |
| "grad_norm": 31.0, |
| "learning_rate": 9.249608763693271e-06, |
| "loss": 1.8675, |
| "step": 16740 |
| }, |
| { |
| "epoch": 1.759453781512605, |
| "grad_norm": 14.375, |
| "learning_rate": 9.241784037558686e-06, |
| "loss": 0.9695, |
| "step": 16750 |
| }, |
| { |
| "epoch": 1.7605042016806722, |
| "grad_norm": 9.3125, |
| "learning_rate": 9.2339593114241e-06, |
| "loss": 1.5124, |
| "step": 16760 |
| }, |
| { |
| "epoch": 1.7615546218487395, |
| "grad_norm": 25.75, |
| "learning_rate": 9.226134585289515e-06, |
| "loss": 1.2201, |
| "step": 16770 |
| }, |
| { |
| "epoch": 1.7626050420168067, |
| "grad_norm": 6.375, |
| "learning_rate": 9.21830985915493e-06, |
| "loss": 1.5277, |
| "step": 16780 |
| }, |
| { |
| "epoch": 1.763655462184874, |
| "grad_norm": 14.5, |
| "learning_rate": 9.210485133020344e-06, |
| "loss": 0.9598, |
| "step": 16790 |
| }, |
| { |
| "epoch": 1.7647058823529411, |
| "grad_norm": 13.375, |
| "learning_rate": 9.202660406885759e-06, |
| "loss": 1.4482, |
| "step": 16800 |
| }, |
| { |
| "epoch": 1.7657563025210083, |
| "grad_norm": 16.875, |
| "learning_rate": 9.194835680751174e-06, |
| "loss": 0.9516, |
| "step": 16810 |
| }, |
| { |
| "epoch": 1.7668067226890756, |
| "grad_norm": 16.375, |
| "learning_rate": 9.187010954616588e-06, |
| "loss": 1.4101, |
| "step": 16820 |
| }, |
| { |
| "epoch": 1.7678571428571428, |
| "grad_norm": 12.125, |
| "learning_rate": 9.179186228482005e-06, |
| "loss": 1.6945, |
| "step": 16830 |
| }, |
| { |
| "epoch": 1.76890756302521, |
| "grad_norm": 5.8125, |
| "learning_rate": 9.17136150234742e-06, |
| "loss": 1.4463, |
| "step": 16840 |
| }, |
| { |
| "epoch": 1.7699579831932772, |
| "grad_norm": 2.46875, |
| "learning_rate": 9.163536776212834e-06, |
| "loss": 0.9211, |
| "step": 16850 |
| }, |
| { |
| "epoch": 1.7710084033613445, |
| "grad_norm": 10.0, |
| "learning_rate": 9.155712050078249e-06, |
| "loss": 1.7824, |
| "step": 16860 |
| }, |
| { |
| "epoch": 1.7720588235294117, |
| "grad_norm": 13.1875, |
| "learning_rate": 9.147887323943663e-06, |
| "loss": 1.5375, |
| "step": 16870 |
| }, |
| { |
| "epoch": 1.773109243697479, |
| "grad_norm": 4.03125, |
| "learning_rate": 9.140062597809078e-06, |
| "loss": 1.0609, |
| "step": 16880 |
| }, |
| { |
| "epoch": 1.7741596638655461, |
| "grad_norm": 17.25, |
| "learning_rate": 9.132237871674493e-06, |
| "loss": 1.4122, |
| "step": 16890 |
| }, |
| { |
| "epoch": 1.7752100840336134, |
| "grad_norm": 13.0625, |
| "learning_rate": 9.124413145539907e-06, |
| "loss": 1.4912, |
| "step": 16900 |
| }, |
| { |
| "epoch": 1.7762605042016806, |
| "grad_norm": 14.625, |
| "learning_rate": 9.116588419405322e-06, |
| "loss": 1.3883, |
| "step": 16910 |
| }, |
| { |
| "epoch": 1.7773109243697478, |
| "grad_norm": 12.75, |
| "learning_rate": 9.108763693270737e-06, |
| "loss": 1.2688, |
| "step": 16920 |
| }, |
| { |
| "epoch": 1.778361344537815, |
| "grad_norm": 11.625, |
| "learning_rate": 9.100938967136151e-06, |
| "loss": 1.1332, |
| "step": 16930 |
| }, |
| { |
| "epoch": 1.7794117647058822, |
| "grad_norm": 5.4375, |
| "learning_rate": 9.093114241001566e-06, |
| "loss": 1.1757, |
| "step": 16940 |
| }, |
| { |
| "epoch": 1.7804621848739495, |
| "grad_norm": 15.25, |
| "learning_rate": 9.08528951486698e-06, |
| "loss": 1.7292, |
| "step": 16950 |
| }, |
| { |
| "epoch": 1.7815126050420167, |
| "grad_norm": 12.3125, |
| "learning_rate": 9.077464788732395e-06, |
| "loss": 1.2453, |
| "step": 16960 |
| }, |
| { |
| "epoch": 1.782563025210084, |
| "grad_norm": 11.5625, |
| "learning_rate": 9.06964006259781e-06, |
| "loss": 1.3372, |
| "step": 16970 |
| }, |
| { |
| "epoch": 1.7836134453781511, |
| "grad_norm": 17.875, |
| "learning_rate": 9.061815336463224e-06, |
| "loss": 1.1219, |
| "step": 16980 |
| }, |
| { |
| "epoch": 1.7846638655462184, |
| "grad_norm": 10.6875, |
| "learning_rate": 9.053990610328639e-06, |
| "loss": 1.6594, |
| "step": 16990 |
| }, |
| { |
| "epoch": 1.7857142857142856, |
| "grad_norm": 3.5625, |
| "learning_rate": 9.046165884194054e-06, |
| "loss": 0.88, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.7867647058823528, |
| "grad_norm": 12.25, |
| "learning_rate": 9.038341158059468e-06, |
| "loss": 0.9889, |
| "step": 17010 |
| }, |
| { |
| "epoch": 1.78781512605042, |
| "grad_norm": 13.0, |
| "learning_rate": 9.030516431924883e-06, |
| "loss": 1.3583, |
| "step": 17020 |
| }, |
| { |
| "epoch": 1.7888655462184873, |
| "grad_norm": 14.625, |
| "learning_rate": 9.022691705790298e-06, |
| "loss": 1.5652, |
| "step": 17030 |
| }, |
| { |
| "epoch": 1.7899159663865545, |
| "grad_norm": 16.25, |
| "learning_rate": 9.014866979655712e-06, |
| "loss": 1.448, |
| "step": 17040 |
| }, |
| { |
| "epoch": 1.7909663865546217, |
| "grad_norm": 21.25, |
| "learning_rate": 9.007042253521127e-06, |
| "loss": 1.4419, |
| "step": 17050 |
| }, |
| { |
| "epoch": 1.792016806722689, |
| "grad_norm": 13.5, |
| "learning_rate": 8.999217527386542e-06, |
| "loss": 1.3056, |
| "step": 17060 |
| }, |
| { |
| "epoch": 1.7930672268907561, |
| "grad_norm": 13.8125, |
| "learning_rate": 8.991392801251956e-06, |
| "loss": 1.2497, |
| "step": 17070 |
| }, |
| { |
| "epoch": 1.7941176470588234, |
| "grad_norm": 19.75, |
| "learning_rate": 8.983568075117371e-06, |
| "loss": 1.1989, |
| "step": 17080 |
| }, |
| { |
| "epoch": 1.7951680672268906, |
| "grad_norm": 25.25, |
| "learning_rate": 8.975743348982786e-06, |
| "loss": 1.2195, |
| "step": 17090 |
| }, |
| { |
| "epoch": 1.7962184873949578, |
| "grad_norm": 14.3125, |
| "learning_rate": 8.9679186228482e-06, |
| "loss": 1.4881, |
| "step": 17100 |
| }, |
| { |
| "epoch": 1.7972689075630253, |
| "grad_norm": 5.9375, |
| "learning_rate": 8.960093896713615e-06, |
| "loss": 0.9233, |
| "step": 17110 |
| }, |
| { |
| "epoch": 1.7983193277310925, |
| "grad_norm": 9.3125, |
| "learning_rate": 8.952269170579031e-06, |
| "loss": 1.3991, |
| "step": 17120 |
| }, |
| { |
| "epoch": 1.7993697478991597, |
| "grad_norm": 16.125, |
| "learning_rate": 8.944444444444446e-06, |
| "loss": 1.5303, |
| "step": 17130 |
| }, |
| { |
| "epoch": 1.800420168067227, |
| "grad_norm": 10.125, |
| "learning_rate": 8.93661971830986e-06, |
| "loss": 1.3232, |
| "step": 17140 |
| }, |
| { |
| "epoch": 1.8014705882352942, |
| "grad_norm": 12.375, |
| "learning_rate": 8.928794992175275e-06, |
| "loss": 1.0356, |
| "step": 17150 |
| }, |
| { |
| "epoch": 1.8025210084033614, |
| "grad_norm": 12.625, |
| "learning_rate": 8.92097026604069e-06, |
| "loss": 1.1945, |
| "step": 17160 |
| }, |
| { |
| "epoch": 1.8035714285714286, |
| "grad_norm": 15.25, |
| "learning_rate": 8.913145539906105e-06, |
| "loss": 1.4963, |
| "step": 17170 |
| }, |
| { |
| "epoch": 1.8046218487394958, |
| "grad_norm": 11.9375, |
| "learning_rate": 8.90532081377152e-06, |
| "loss": 0.9988, |
| "step": 17180 |
| }, |
| { |
| "epoch": 1.805672268907563, |
| "grad_norm": 17.75, |
| "learning_rate": 8.897496087636934e-06, |
| "loss": 1.3638, |
| "step": 17190 |
| }, |
| { |
| "epoch": 1.8067226890756303, |
| "grad_norm": 4.40625, |
| "learning_rate": 8.889671361502349e-06, |
| "loss": 1.2046, |
| "step": 17200 |
| }, |
| { |
| "epoch": 1.8077731092436975, |
| "grad_norm": 20.625, |
| "learning_rate": 8.881846635367763e-06, |
| "loss": 1.053, |
| "step": 17210 |
| }, |
| { |
| "epoch": 1.8088235294117647, |
| "grad_norm": 11.6875, |
| "learning_rate": 8.874021909233178e-06, |
| "loss": 1.6096, |
| "step": 17220 |
| }, |
| { |
| "epoch": 1.809873949579832, |
| "grad_norm": 9.875, |
| "learning_rate": 8.866197183098592e-06, |
| "loss": 1.5234, |
| "step": 17230 |
| }, |
| { |
| "epoch": 1.8109243697478992, |
| "grad_norm": 13.625, |
| "learning_rate": 8.858372456964007e-06, |
| "loss": 1.3558, |
| "step": 17240 |
| }, |
| { |
| "epoch": 1.8119747899159664, |
| "grad_norm": 14.625, |
| "learning_rate": 8.850547730829422e-06, |
| "loss": 1.3757, |
| "step": 17250 |
| }, |
| { |
| "epoch": 1.8130252100840336, |
| "grad_norm": 11.9375, |
| "learning_rate": 8.842723004694836e-06, |
| "loss": 1.2969, |
| "step": 17260 |
| }, |
| { |
| "epoch": 1.8140756302521008, |
| "grad_norm": 33.25, |
| "learning_rate": 8.834898278560251e-06, |
| "loss": 1.258, |
| "step": 17270 |
| }, |
| { |
| "epoch": 1.815126050420168, |
| "grad_norm": 10.9375, |
| "learning_rate": 8.827073552425666e-06, |
| "loss": 1.0944, |
| "step": 17280 |
| }, |
| { |
| "epoch": 1.8161764705882353, |
| "grad_norm": 12.3125, |
| "learning_rate": 8.81924882629108e-06, |
| "loss": 1.3661, |
| "step": 17290 |
| }, |
| { |
| "epoch": 1.8172268907563025, |
| "grad_norm": 37.75, |
| "learning_rate": 8.811424100156495e-06, |
| "loss": 1.1805, |
| "step": 17300 |
| }, |
| { |
| "epoch": 1.8182773109243697, |
| "grad_norm": 16.0, |
| "learning_rate": 8.80359937402191e-06, |
| "loss": 1.4742, |
| "step": 17310 |
| }, |
| { |
| "epoch": 1.819327731092437, |
| "grad_norm": 5.0625, |
| "learning_rate": 8.795774647887324e-06, |
| "loss": 1.4898, |
| "step": 17320 |
| }, |
| { |
| "epoch": 1.8203781512605042, |
| "grad_norm": 5.3125, |
| "learning_rate": 8.787949921752739e-06, |
| "loss": 1.2318, |
| "step": 17330 |
| }, |
| { |
| "epoch": 1.8214285714285714, |
| "grad_norm": 14.875, |
| "learning_rate": 8.780125195618154e-06, |
| "loss": 1.2514, |
| "step": 17340 |
| }, |
| { |
| "epoch": 1.8224789915966386, |
| "grad_norm": 3.25, |
| "learning_rate": 8.772300469483568e-06, |
| "loss": 1.0916, |
| "step": 17350 |
| }, |
| { |
| "epoch": 1.8235294117647058, |
| "grad_norm": 8.8125, |
| "learning_rate": 8.764475743348983e-06, |
| "loss": 1.5622, |
| "step": 17360 |
| }, |
| { |
| "epoch": 1.824579831932773, |
| "grad_norm": 11.9375, |
| "learning_rate": 8.756651017214398e-06, |
| "loss": 1.338, |
| "step": 17370 |
| }, |
| { |
| "epoch": 1.8256302521008403, |
| "grad_norm": 17.625, |
| "learning_rate": 8.748826291079812e-06, |
| "loss": 1.2941, |
| "step": 17380 |
| }, |
| { |
| "epoch": 1.8266806722689075, |
| "grad_norm": 17.875, |
| "learning_rate": 8.741001564945227e-06, |
| "loss": 1.303, |
| "step": 17390 |
| }, |
| { |
| "epoch": 1.8277310924369747, |
| "grad_norm": 14.3125, |
| "learning_rate": 8.733176838810642e-06, |
| "loss": 1.5946, |
| "step": 17400 |
| }, |
| { |
| "epoch": 1.8287815126050422, |
| "grad_norm": 17.375, |
| "learning_rate": 8.725352112676056e-06, |
| "loss": 1.3532, |
| "step": 17410 |
| }, |
| { |
| "epoch": 1.8298319327731094, |
| "grad_norm": 16.25, |
| "learning_rate": 8.717527386541473e-06, |
| "loss": 0.8938, |
| "step": 17420 |
| }, |
| { |
| "epoch": 1.8308823529411766, |
| "grad_norm": 6.21875, |
| "learning_rate": 8.709702660406887e-06, |
| "loss": 0.9492, |
| "step": 17430 |
| }, |
| { |
| "epoch": 1.8319327731092439, |
| "grad_norm": 8.8125, |
| "learning_rate": 8.701877934272302e-06, |
| "loss": 1.4988, |
| "step": 17440 |
| }, |
| { |
| "epoch": 1.832983193277311, |
| "grad_norm": 15.6875, |
| "learning_rate": 8.694053208137717e-06, |
| "loss": 1.0669, |
| "step": 17450 |
| }, |
| { |
| "epoch": 1.8340336134453783, |
| "grad_norm": 18.0, |
| "learning_rate": 8.686228482003131e-06, |
| "loss": 1.4973, |
| "step": 17460 |
| }, |
| { |
| "epoch": 1.8350840336134455, |
| "grad_norm": 9.9375, |
| "learning_rate": 8.678403755868546e-06, |
| "loss": 0.8783, |
| "step": 17470 |
| }, |
| { |
| "epoch": 1.8361344537815127, |
| "grad_norm": 14.375, |
| "learning_rate": 8.67057902973396e-06, |
| "loss": 1.4758, |
| "step": 17480 |
| }, |
| { |
| "epoch": 1.83718487394958, |
| "grad_norm": 18.875, |
| "learning_rate": 8.662754303599375e-06, |
| "loss": 1.4096, |
| "step": 17490 |
| }, |
| { |
| "epoch": 1.8382352941176472, |
| "grad_norm": 18.5, |
| "learning_rate": 8.65492957746479e-06, |
| "loss": 1.1369, |
| "step": 17500 |
| }, |
| { |
| "epoch": 1.8392857142857144, |
| "grad_norm": 4.84375, |
| "learning_rate": 8.647104851330203e-06, |
| "loss": 1.0654, |
| "step": 17510 |
| }, |
| { |
| "epoch": 1.8403361344537816, |
| "grad_norm": 20.375, |
| "learning_rate": 8.639280125195619e-06, |
| "loss": 1.3115, |
| "step": 17520 |
| }, |
| { |
| "epoch": 1.8413865546218489, |
| "grad_norm": 27.5, |
| "learning_rate": 8.631455399061034e-06, |
| "loss": 1.3379, |
| "step": 17530 |
| }, |
| { |
| "epoch": 1.842436974789916, |
| "grad_norm": 11.6875, |
| "learning_rate": 8.623630672926448e-06, |
| "loss": 1.1143, |
| "step": 17540 |
| }, |
| { |
| "epoch": 1.8434873949579833, |
| "grad_norm": 10.9375, |
| "learning_rate": 8.615805946791863e-06, |
| "loss": 1.3278, |
| "step": 17550 |
| }, |
| { |
| "epoch": 1.8445378151260505, |
| "grad_norm": 9.1875, |
| "learning_rate": 8.607981220657278e-06, |
| "loss": 1.3111, |
| "step": 17560 |
| }, |
| { |
| "epoch": 1.8455882352941178, |
| "grad_norm": 23.5, |
| "learning_rate": 8.600156494522692e-06, |
| "loss": 1.5519, |
| "step": 17570 |
| }, |
| { |
| "epoch": 1.846638655462185, |
| "grad_norm": 15.875, |
| "learning_rate": 8.592331768388107e-06, |
| "loss": 1.4849, |
| "step": 17580 |
| }, |
| { |
| "epoch": 1.8476890756302522, |
| "grad_norm": 8.0, |
| "learning_rate": 8.584507042253522e-06, |
| "loss": 0.8678, |
| "step": 17590 |
| }, |
| { |
| "epoch": 1.8487394957983194, |
| "grad_norm": 14.1875, |
| "learning_rate": 8.576682316118936e-06, |
| "loss": 1.3236, |
| "step": 17600 |
| }, |
| { |
| "epoch": 1.8497899159663866, |
| "grad_norm": 12.8125, |
| "learning_rate": 8.568857589984351e-06, |
| "loss": 1.2436, |
| "step": 17610 |
| }, |
| { |
| "epoch": 1.8508403361344539, |
| "grad_norm": 16.625, |
| "learning_rate": 8.561032863849766e-06, |
| "loss": 1.1921, |
| "step": 17620 |
| }, |
| { |
| "epoch": 1.851890756302521, |
| "grad_norm": 5.625, |
| "learning_rate": 8.55320813771518e-06, |
| "loss": 1.3603, |
| "step": 17630 |
| }, |
| { |
| "epoch": 1.8529411764705883, |
| "grad_norm": 10.6875, |
| "learning_rate": 8.545383411580595e-06, |
| "loss": 1.0873, |
| "step": 17640 |
| }, |
| { |
| "epoch": 1.8539915966386555, |
| "grad_norm": 16.875, |
| "learning_rate": 8.53755868544601e-06, |
| "loss": 1.4378, |
| "step": 17650 |
| }, |
| { |
| "epoch": 1.8550420168067228, |
| "grad_norm": 9.9375, |
| "learning_rate": 8.529733959311424e-06, |
| "loss": 1.4792, |
| "step": 17660 |
| }, |
| { |
| "epoch": 1.85609243697479, |
| "grad_norm": 11.75, |
| "learning_rate": 8.521909233176839e-06, |
| "loss": 1.8806, |
| "step": 17670 |
| }, |
| { |
| "epoch": 1.8571428571428572, |
| "grad_norm": 12.5625, |
| "learning_rate": 8.514084507042254e-06, |
| "loss": 1.3804, |
| "step": 17680 |
| }, |
| { |
| "epoch": 1.8581932773109244, |
| "grad_norm": 16.75, |
| "learning_rate": 8.506259780907668e-06, |
| "loss": 1.1978, |
| "step": 17690 |
| }, |
| { |
| "epoch": 1.8592436974789917, |
| "grad_norm": 12.0, |
| "learning_rate": 8.498435054773083e-06, |
| "loss": 0.841, |
| "step": 17700 |
| }, |
| { |
| "epoch": 1.8602941176470589, |
| "grad_norm": 5.375, |
| "learning_rate": 8.4906103286385e-06, |
| "loss": 1.0246, |
| "step": 17710 |
| }, |
| { |
| "epoch": 1.861344537815126, |
| "grad_norm": 16.75, |
| "learning_rate": 8.482785602503914e-06, |
| "loss": 1.5822, |
| "step": 17720 |
| }, |
| { |
| "epoch": 1.8623949579831933, |
| "grad_norm": 16.5, |
| "learning_rate": 8.474960876369329e-06, |
| "loss": 1.3041, |
| "step": 17730 |
| }, |
| { |
| "epoch": 1.8634453781512605, |
| "grad_norm": 13.375, |
| "learning_rate": 8.467136150234743e-06, |
| "loss": 1.4525, |
| "step": 17740 |
| }, |
| { |
| "epoch": 1.8644957983193278, |
| "grad_norm": 12.625, |
| "learning_rate": 8.459311424100158e-06, |
| "loss": 1.4399, |
| "step": 17750 |
| }, |
| { |
| "epoch": 1.865546218487395, |
| "grad_norm": 16.25, |
| "learning_rate": 8.451486697965572e-06, |
| "loss": 1.3268, |
| "step": 17760 |
| }, |
| { |
| "epoch": 1.8665966386554622, |
| "grad_norm": 14.6875, |
| "learning_rate": 8.443661971830987e-06, |
| "loss": 1.6509, |
| "step": 17770 |
| }, |
| { |
| "epoch": 1.8676470588235294, |
| "grad_norm": 17.625, |
| "learning_rate": 8.435837245696402e-06, |
| "loss": 1.6677, |
| "step": 17780 |
| }, |
| { |
| "epoch": 1.8686974789915967, |
| "grad_norm": 14.25, |
| "learning_rate": 8.428012519561816e-06, |
| "loss": 1.2302, |
| "step": 17790 |
| }, |
| { |
| "epoch": 1.8697478991596639, |
| "grad_norm": 4.46875, |
| "learning_rate": 8.420187793427231e-06, |
| "loss": 1.0689, |
| "step": 17800 |
| }, |
| { |
| "epoch": 1.870798319327731, |
| "grad_norm": 5.09375, |
| "learning_rate": 8.412363067292646e-06, |
| "loss": 1.1844, |
| "step": 17810 |
| }, |
| { |
| "epoch": 1.8718487394957983, |
| "grad_norm": 13.25, |
| "learning_rate": 8.40453834115806e-06, |
| "loss": 1.5129, |
| "step": 17820 |
| }, |
| { |
| "epoch": 1.8728991596638656, |
| "grad_norm": 15.125, |
| "learning_rate": 8.396713615023475e-06, |
| "loss": 1.363, |
| "step": 17830 |
| }, |
| { |
| "epoch": 1.8739495798319328, |
| "grad_norm": 16.875, |
| "learning_rate": 8.38888888888889e-06, |
| "loss": 1.2641, |
| "step": 17840 |
| }, |
| { |
| "epoch": 1.875, |
| "grad_norm": 12.5, |
| "learning_rate": 8.381064162754304e-06, |
| "loss": 1.599, |
| "step": 17850 |
| }, |
| { |
| "epoch": 1.8760504201680672, |
| "grad_norm": 4.25, |
| "learning_rate": 8.373239436619719e-06, |
| "loss": 1.3407, |
| "step": 17860 |
| }, |
| { |
| "epoch": 1.8771008403361344, |
| "grad_norm": 11.5, |
| "learning_rate": 8.365414710485134e-06, |
| "loss": 0.8743, |
| "step": 17870 |
| }, |
| { |
| "epoch": 1.8781512605042017, |
| "grad_norm": 12.4375, |
| "learning_rate": 8.357589984350548e-06, |
| "loss": 1.4704, |
| "step": 17880 |
| }, |
| { |
| "epoch": 1.879201680672269, |
| "grad_norm": 8.5, |
| "learning_rate": 8.349765258215963e-06, |
| "loss": 1.2385, |
| "step": 17890 |
| }, |
| { |
| "epoch": 1.8802521008403361, |
| "grad_norm": 13.0625, |
| "learning_rate": 8.341940532081378e-06, |
| "loss": 0.778, |
| "step": 17900 |
| }, |
| { |
| "epoch": 1.8813025210084033, |
| "grad_norm": 11.3125, |
| "learning_rate": 8.334115805946792e-06, |
| "loss": 1.3938, |
| "step": 17910 |
| }, |
| { |
| "epoch": 1.8823529411764706, |
| "grad_norm": 15.4375, |
| "learning_rate": 8.326291079812207e-06, |
| "loss": 1.684, |
| "step": 17920 |
| }, |
| { |
| "epoch": 1.8834033613445378, |
| "grad_norm": 18.25, |
| "learning_rate": 8.318466353677622e-06, |
| "loss": 1.6561, |
| "step": 17930 |
| }, |
| { |
| "epoch": 1.884453781512605, |
| "grad_norm": 17.0, |
| "learning_rate": 8.310641627543036e-06, |
| "loss": 1.5292, |
| "step": 17940 |
| }, |
| { |
| "epoch": 1.8855042016806722, |
| "grad_norm": 10.1875, |
| "learning_rate": 8.302816901408451e-06, |
| "loss": 1.2049, |
| "step": 17950 |
| }, |
| { |
| "epoch": 1.8865546218487395, |
| "grad_norm": 7.375, |
| "learning_rate": 8.294992175273866e-06, |
| "loss": 1.2616, |
| "step": 17960 |
| }, |
| { |
| "epoch": 1.8876050420168067, |
| "grad_norm": 11.0, |
| "learning_rate": 8.28716744913928e-06, |
| "loss": 0.9736, |
| "step": 17970 |
| }, |
| { |
| "epoch": 1.888655462184874, |
| "grad_norm": 18.0, |
| "learning_rate": 8.279342723004695e-06, |
| "loss": 1.194, |
| "step": 17980 |
| }, |
| { |
| "epoch": 1.8897058823529411, |
| "grad_norm": 16.125, |
| "learning_rate": 8.27151799687011e-06, |
| "loss": 1.3858, |
| "step": 17990 |
| }, |
| { |
| "epoch": 1.8907563025210083, |
| "grad_norm": 13.6875, |
| "learning_rate": 8.263693270735524e-06, |
| "loss": 0.965, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.8918067226890756, |
| "grad_norm": 11.75, |
| "learning_rate": 8.25586854460094e-06, |
| "loss": 1.2845, |
| "step": 18010 |
| }, |
| { |
| "epoch": 1.8928571428571428, |
| "grad_norm": 13.5, |
| "learning_rate": 8.248043818466355e-06, |
| "loss": 1.0943, |
| "step": 18020 |
| }, |
| { |
| "epoch": 1.89390756302521, |
| "grad_norm": 16.375, |
| "learning_rate": 8.24021909233177e-06, |
| "loss": 1.4654, |
| "step": 18030 |
| }, |
| { |
| "epoch": 1.8949579831932772, |
| "grad_norm": 15.3125, |
| "learning_rate": 8.232394366197184e-06, |
| "loss": 1.6451, |
| "step": 18040 |
| }, |
| { |
| "epoch": 1.8960084033613445, |
| "grad_norm": 11.8125, |
| "learning_rate": 8.224569640062599e-06, |
| "loss": 1.5634, |
| "step": 18050 |
| }, |
| { |
| "epoch": 1.8970588235294117, |
| "grad_norm": 6.625, |
| "learning_rate": 8.216744913928014e-06, |
| "loss": 1.3746, |
| "step": 18060 |
| }, |
| { |
| "epoch": 1.898109243697479, |
| "grad_norm": 5.40625, |
| "learning_rate": 8.208920187793428e-06, |
| "loss": 0.9587, |
| "step": 18070 |
| }, |
| { |
| "epoch": 1.8991596638655461, |
| "grad_norm": 25.875, |
| "learning_rate": 8.201095461658843e-06, |
| "loss": 1.6194, |
| "step": 18080 |
| }, |
| { |
| "epoch": 1.9002100840336134, |
| "grad_norm": 3.28125, |
| "learning_rate": 8.193270735524258e-06, |
| "loss": 1.1389, |
| "step": 18090 |
| }, |
| { |
| "epoch": 1.9012605042016806, |
| "grad_norm": 4.4375, |
| "learning_rate": 8.18544600938967e-06, |
| "loss": 0.8685, |
| "step": 18100 |
| }, |
| { |
| "epoch": 1.9023109243697478, |
| "grad_norm": 18.875, |
| "learning_rate": 8.177621283255087e-06, |
| "loss": 1.51, |
| "step": 18110 |
| }, |
| { |
| "epoch": 1.903361344537815, |
| "grad_norm": 24.125, |
| "learning_rate": 8.169796557120502e-06, |
| "loss": 1.5197, |
| "step": 18120 |
| }, |
| { |
| "epoch": 1.9044117647058822, |
| "grad_norm": 12.3125, |
| "learning_rate": 8.161971830985916e-06, |
| "loss": 1.4569, |
| "step": 18130 |
| }, |
| { |
| "epoch": 1.9054621848739495, |
| "grad_norm": 4.65625, |
| "learning_rate": 8.154147104851331e-06, |
| "loss": 1.1837, |
| "step": 18140 |
| }, |
| { |
| "epoch": 1.9065126050420167, |
| "grad_norm": 3.765625, |
| "learning_rate": 8.146322378716746e-06, |
| "loss": 1.6231, |
| "step": 18150 |
| }, |
| { |
| "epoch": 1.907563025210084, |
| "grad_norm": 10.3125, |
| "learning_rate": 8.13849765258216e-06, |
| "loss": 1.3244, |
| "step": 18160 |
| }, |
| { |
| "epoch": 1.9086134453781511, |
| "grad_norm": 12.4375, |
| "learning_rate": 8.130672926447575e-06, |
| "loss": 1.1799, |
| "step": 18170 |
| }, |
| { |
| "epoch": 1.9096638655462184, |
| "grad_norm": 13.25, |
| "learning_rate": 8.12284820031299e-06, |
| "loss": 1.0951, |
| "step": 18180 |
| }, |
| { |
| "epoch": 1.9107142857142856, |
| "grad_norm": 4.96875, |
| "learning_rate": 8.115023474178404e-06, |
| "loss": 1.2038, |
| "step": 18190 |
| }, |
| { |
| "epoch": 1.9117647058823528, |
| "grad_norm": 14.5625, |
| "learning_rate": 8.107198748043819e-06, |
| "loss": 1.514, |
| "step": 18200 |
| }, |
| { |
| "epoch": 1.91281512605042, |
| "grad_norm": 15.25, |
| "learning_rate": 8.099374021909234e-06, |
| "loss": 1.1104, |
| "step": 18210 |
| }, |
| { |
| "epoch": 1.9138655462184873, |
| "grad_norm": 15.9375, |
| "learning_rate": 8.091549295774648e-06, |
| "loss": 1.7082, |
| "step": 18220 |
| }, |
| { |
| "epoch": 1.9149159663865545, |
| "grad_norm": 13.375, |
| "learning_rate": 8.083724569640063e-06, |
| "loss": 0.8448, |
| "step": 18230 |
| }, |
| { |
| "epoch": 1.9159663865546217, |
| "grad_norm": 3.328125, |
| "learning_rate": 8.075899843505478e-06, |
| "loss": 1.286, |
| "step": 18240 |
| }, |
| { |
| "epoch": 1.917016806722689, |
| "grad_norm": 11.125, |
| "learning_rate": 8.068075117370892e-06, |
| "loss": 1.6202, |
| "step": 18250 |
| }, |
| { |
| "epoch": 1.9180672268907561, |
| "grad_norm": 5.0, |
| "learning_rate": 8.060250391236307e-06, |
| "loss": 1.2555, |
| "step": 18260 |
| }, |
| { |
| "epoch": 1.9191176470588234, |
| "grad_norm": 11.5, |
| "learning_rate": 8.052425665101721e-06, |
| "loss": 1.323, |
| "step": 18270 |
| }, |
| { |
| "epoch": 1.9201680672268906, |
| "grad_norm": 11.5625, |
| "learning_rate": 8.044600938967136e-06, |
| "loss": 1.8193, |
| "step": 18280 |
| }, |
| { |
| "epoch": 1.9212184873949578, |
| "grad_norm": 12.0625, |
| "learning_rate": 8.03677621283255e-06, |
| "loss": 1.487, |
| "step": 18290 |
| }, |
| { |
| "epoch": 1.9222689075630253, |
| "grad_norm": 12.25, |
| "learning_rate": 8.028951486697967e-06, |
| "loss": 1.0669, |
| "step": 18300 |
| }, |
| { |
| "epoch": 1.9233193277310925, |
| "grad_norm": 16.25, |
| "learning_rate": 8.021126760563382e-06, |
| "loss": 1.2489, |
| "step": 18310 |
| }, |
| { |
| "epoch": 1.9243697478991597, |
| "grad_norm": 15.75, |
| "learning_rate": 8.013302034428796e-06, |
| "loss": 1.2696, |
| "step": 18320 |
| }, |
| { |
| "epoch": 1.925420168067227, |
| "grad_norm": 17.625, |
| "learning_rate": 8.005477308294211e-06, |
| "loss": 1.5837, |
| "step": 18330 |
| }, |
| { |
| "epoch": 1.9264705882352942, |
| "grad_norm": 4.9375, |
| "learning_rate": 7.997652582159626e-06, |
| "loss": 1.2175, |
| "step": 18340 |
| }, |
| { |
| "epoch": 1.9275210084033614, |
| "grad_norm": 19.75, |
| "learning_rate": 7.98982785602504e-06, |
| "loss": 1.5983, |
| "step": 18350 |
| }, |
| { |
| "epoch": 1.9285714285714286, |
| "grad_norm": 13.125, |
| "learning_rate": 7.982003129890455e-06, |
| "loss": 1.6932, |
| "step": 18360 |
| }, |
| { |
| "epoch": 1.9296218487394958, |
| "grad_norm": 12.875, |
| "learning_rate": 7.97417840375587e-06, |
| "loss": 1.0066, |
| "step": 18370 |
| }, |
| { |
| "epoch": 1.930672268907563, |
| "grad_norm": 9.25, |
| "learning_rate": 7.966353677621284e-06, |
| "loss": 1.3433, |
| "step": 18380 |
| }, |
| { |
| "epoch": 1.9317226890756303, |
| "grad_norm": 15.5, |
| "learning_rate": 7.958528951486697e-06, |
| "loss": 1.2296, |
| "step": 18390 |
| }, |
| { |
| "epoch": 1.9327731092436975, |
| "grad_norm": 11.375, |
| "learning_rate": 7.950704225352114e-06, |
| "loss": 0.9645, |
| "step": 18400 |
| }, |
| { |
| "epoch": 1.9338235294117647, |
| "grad_norm": 19.125, |
| "learning_rate": 7.942879499217528e-06, |
| "loss": 1.1675, |
| "step": 18410 |
| }, |
| { |
| "epoch": 1.934873949579832, |
| "grad_norm": 16.25, |
| "learning_rate": 7.935054773082943e-06, |
| "loss": 1.2545, |
| "step": 18420 |
| }, |
| { |
| "epoch": 1.9359243697478992, |
| "grad_norm": 6.40625, |
| "learning_rate": 7.927230046948358e-06, |
| "loss": 1.3921, |
| "step": 18430 |
| }, |
| { |
| "epoch": 1.9369747899159664, |
| "grad_norm": 4.5625, |
| "learning_rate": 7.919405320813772e-06, |
| "loss": 1.333, |
| "step": 18440 |
| }, |
| { |
| "epoch": 1.9380252100840336, |
| "grad_norm": 13.625, |
| "learning_rate": 7.911580594679187e-06, |
| "loss": 1.4191, |
| "step": 18450 |
| }, |
| { |
| "epoch": 1.9390756302521008, |
| "grad_norm": 13.6875, |
| "learning_rate": 7.903755868544602e-06, |
| "loss": 0.9574, |
| "step": 18460 |
| }, |
| { |
| "epoch": 1.940126050420168, |
| "grad_norm": 13.8125, |
| "learning_rate": 7.895931142410016e-06, |
| "loss": 1.4445, |
| "step": 18470 |
| }, |
| { |
| "epoch": 1.9411764705882353, |
| "grad_norm": 8.8125, |
| "learning_rate": 7.888106416275431e-06, |
| "loss": 1.2046, |
| "step": 18480 |
| }, |
| { |
| "epoch": 1.9422268907563025, |
| "grad_norm": 13.375, |
| "learning_rate": 7.880281690140846e-06, |
| "loss": 0.6966, |
| "step": 18490 |
| }, |
| { |
| "epoch": 1.9432773109243697, |
| "grad_norm": 17.875, |
| "learning_rate": 7.87245696400626e-06, |
| "loss": 1.593, |
| "step": 18500 |
| }, |
| { |
| "epoch": 1.944327731092437, |
| "grad_norm": 14.125, |
| "learning_rate": 7.864632237871675e-06, |
| "loss": 1.5295, |
| "step": 18510 |
| }, |
| { |
| "epoch": 1.9453781512605042, |
| "grad_norm": 17.375, |
| "learning_rate": 7.85680751173709e-06, |
| "loss": 1.0856, |
| "step": 18520 |
| }, |
| { |
| "epoch": 1.9464285714285714, |
| "grad_norm": 11.75, |
| "learning_rate": 7.848982785602504e-06, |
| "loss": 1.4568, |
| "step": 18530 |
| }, |
| { |
| "epoch": 1.9474789915966386, |
| "grad_norm": 12.4375, |
| "learning_rate": 7.841158059467919e-06, |
| "loss": 1.2177, |
| "step": 18540 |
| }, |
| { |
| "epoch": 1.9485294117647058, |
| "grad_norm": 7.625, |
| "learning_rate": 7.833333333333333e-06, |
| "loss": 1.1167, |
| "step": 18550 |
| }, |
| { |
| "epoch": 1.949579831932773, |
| "grad_norm": 11.375, |
| "learning_rate": 7.825508607198748e-06, |
| "loss": 0.9677, |
| "step": 18560 |
| }, |
| { |
| "epoch": 1.9506302521008403, |
| "grad_norm": 15.125, |
| "learning_rate": 7.817683881064163e-06, |
| "loss": 1.2355, |
| "step": 18570 |
| }, |
| { |
| "epoch": 1.9516806722689075, |
| "grad_norm": 15.625, |
| "learning_rate": 7.809859154929577e-06, |
| "loss": 1.4345, |
| "step": 18580 |
| }, |
| { |
| "epoch": 1.9527310924369747, |
| "grad_norm": 18.875, |
| "learning_rate": 7.802034428794992e-06, |
| "loss": 1.2863, |
| "step": 18590 |
| }, |
| { |
| "epoch": 1.9537815126050422, |
| "grad_norm": 13.0, |
| "learning_rate": 7.794209702660408e-06, |
| "loss": 1.1896, |
| "step": 18600 |
| }, |
| { |
| "epoch": 1.9548319327731094, |
| "grad_norm": 6.15625, |
| "learning_rate": 7.786384976525823e-06, |
| "loss": 1.1191, |
| "step": 18610 |
| }, |
| { |
| "epoch": 1.9558823529411766, |
| "grad_norm": 7.6875, |
| "learning_rate": 7.778560250391238e-06, |
| "loss": 0.9882, |
| "step": 18620 |
| }, |
| { |
| "epoch": 1.9569327731092439, |
| "grad_norm": 14.9375, |
| "learning_rate": 7.770735524256652e-06, |
| "loss": 1.2986, |
| "step": 18630 |
| }, |
| { |
| "epoch": 1.957983193277311, |
| "grad_norm": 4.78125, |
| "learning_rate": 7.762910798122067e-06, |
| "loss": 1.1171, |
| "step": 18640 |
| }, |
| { |
| "epoch": 1.9590336134453783, |
| "grad_norm": 16.375, |
| "learning_rate": 7.755086071987482e-06, |
| "loss": 1.4327, |
| "step": 18650 |
| }, |
| { |
| "epoch": 1.9600840336134455, |
| "grad_norm": 19.125, |
| "learning_rate": 7.747261345852896e-06, |
| "loss": 1.1288, |
| "step": 18660 |
| }, |
| { |
| "epoch": 1.9611344537815127, |
| "grad_norm": 9.125, |
| "learning_rate": 7.739436619718311e-06, |
| "loss": 0.9525, |
| "step": 18670 |
| }, |
| { |
| "epoch": 1.96218487394958, |
| "grad_norm": 14.625, |
| "learning_rate": 7.731611893583724e-06, |
| "loss": 1.2451, |
| "step": 18680 |
| }, |
| { |
| "epoch": 1.9632352941176472, |
| "grad_norm": 70.0, |
| "learning_rate": 7.723787167449139e-06, |
| "loss": 1.5298, |
| "step": 18690 |
| }, |
| { |
| "epoch": 1.9642857142857144, |
| "grad_norm": 15.5, |
| "learning_rate": 7.715962441314555e-06, |
| "loss": 0.9037, |
| "step": 18700 |
| }, |
| { |
| "epoch": 1.9653361344537816, |
| "grad_norm": 17.75, |
| "learning_rate": 7.70813771517997e-06, |
| "loss": 1.4113, |
| "step": 18710 |
| }, |
| { |
| "epoch": 1.9663865546218489, |
| "grad_norm": 9.3125, |
| "learning_rate": 7.700312989045384e-06, |
| "loss": 1.1319, |
| "step": 18720 |
| }, |
| { |
| "epoch": 1.967436974789916, |
| "grad_norm": 18.625, |
| "learning_rate": 7.692488262910799e-06, |
| "loss": 0.897, |
| "step": 18730 |
| }, |
| { |
| "epoch": 1.9684873949579833, |
| "grad_norm": 18.5, |
| "learning_rate": 7.684663536776214e-06, |
| "loss": 1.8907, |
| "step": 18740 |
| }, |
| { |
| "epoch": 1.9695378151260505, |
| "grad_norm": 15.9375, |
| "learning_rate": 7.676838810641628e-06, |
| "loss": 1.0922, |
| "step": 18750 |
| }, |
| { |
| "epoch": 1.9705882352941178, |
| "grad_norm": 11.625, |
| "learning_rate": 7.669014084507043e-06, |
| "loss": 1.2242, |
| "step": 18760 |
| }, |
| { |
| "epoch": 1.971638655462185, |
| "grad_norm": 8.875, |
| "learning_rate": 7.661189358372457e-06, |
| "loss": 1.3724, |
| "step": 18770 |
| }, |
| { |
| "epoch": 1.9726890756302522, |
| "grad_norm": 16.0, |
| "learning_rate": 7.653364632237872e-06, |
| "loss": 1.4016, |
| "step": 18780 |
| }, |
| { |
| "epoch": 1.9737394957983194, |
| "grad_norm": 13.4375, |
| "learning_rate": 7.645539906103287e-06, |
| "loss": 1.2584, |
| "step": 18790 |
| }, |
| { |
| "epoch": 1.9747899159663866, |
| "grad_norm": 12.75, |
| "learning_rate": 7.637715179968701e-06, |
| "loss": 1.4473, |
| "step": 18800 |
| }, |
| { |
| "epoch": 1.9758403361344539, |
| "grad_norm": 16.625, |
| "learning_rate": 7.629890453834116e-06, |
| "loss": 1.5624, |
| "step": 18810 |
| }, |
| { |
| "epoch": 1.976890756302521, |
| "grad_norm": 15.5625, |
| "learning_rate": 7.622065727699532e-06, |
| "loss": 1.3079, |
| "step": 18820 |
| }, |
| { |
| "epoch": 1.9779411764705883, |
| "grad_norm": 15.125, |
| "learning_rate": 7.614241001564946e-06, |
| "loss": 1.3884, |
| "step": 18830 |
| }, |
| { |
| "epoch": 1.9789915966386555, |
| "grad_norm": 48.0, |
| "learning_rate": 7.606416275430361e-06, |
| "loss": 1.2941, |
| "step": 18840 |
| }, |
| { |
| "epoch": 1.9800420168067228, |
| "grad_norm": 4.46875, |
| "learning_rate": 7.598591549295775e-06, |
| "loss": 1.0213, |
| "step": 18850 |
| }, |
| { |
| "epoch": 1.98109243697479, |
| "grad_norm": 17.5, |
| "learning_rate": 7.590766823161189e-06, |
| "loss": 1.8067, |
| "step": 18860 |
| }, |
| { |
| "epoch": 1.9821428571428572, |
| "grad_norm": 12.3125, |
| "learning_rate": 7.582942097026604e-06, |
| "loss": 1.1539, |
| "step": 18870 |
| }, |
| { |
| "epoch": 1.9831932773109244, |
| "grad_norm": 3.765625, |
| "learning_rate": 7.575117370892019e-06, |
| "loss": 1.2292, |
| "step": 18880 |
| }, |
| { |
| "epoch": 1.9842436974789917, |
| "grad_norm": 16.5, |
| "learning_rate": 7.567292644757435e-06, |
| "loss": 1.4515, |
| "step": 18890 |
| }, |
| { |
| "epoch": 1.9852941176470589, |
| "grad_norm": 19.375, |
| "learning_rate": 7.559467918622849e-06, |
| "loss": 1.7594, |
| "step": 18900 |
| }, |
| { |
| "epoch": 1.986344537815126, |
| "grad_norm": 15.3125, |
| "learning_rate": 7.5516431924882635e-06, |
| "loss": 1.6458, |
| "step": 18910 |
| }, |
| { |
| "epoch": 1.9873949579831933, |
| "grad_norm": 12.375, |
| "learning_rate": 7.543818466353678e-06, |
| "loss": 1.7079, |
| "step": 18920 |
| }, |
| { |
| "epoch": 1.9884453781512605, |
| "grad_norm": 11.9375, |
| "learning_rate": 7.535993740219093e-06, |
| "loss": 1.4056, |
| "step": 18930 |
| }, |
| { |
| "epoch": 1.9894957983193278, |
| "grad_norm": 13.9375, |
| "learning_rate": 7.5281690140845074e-06, |
| "loss": 1.542, |
| "step": 18940 |
| }, |
| { |
| "epoch": 1.990546218487395, |
| "grad_norm": 14.375, |
| "learning_rate": 7.520344287949922e-06, |
| "loss": 1.4334, |
| "step": 18950 |
| }, |
| { |
| "epoch": 1.9915966386554622, |
| "grad_norm": 12.125, |
| "learning_rate": 7.512519561815337e-06, |
| "loss": 1.033, |
| "step": 18960 |
| }, |
| { |
| "epoch": 1.9926470588235294, |
| "grad_norm": 11.8125, |
| "learning_rate": 7.504694835680751e-06, |
| "loss": 1.5305, |
| "step": 18970 |
| }, |
| { |
| "epoch": 1.9936974789915967, |
| "grad_norm": 20.875, |
| "learning_rate": 7.496870109546166e-06, |
| "loss": 1.5138, |
| "step": 18980 |
| }, |
| { |
| "epoch": 1.9947478991596639, |
| "grad_norm": 5.71875, |
| "learning_rate": 7.4890453834115816e-06, |
| "loss": 1.1856, |
| "step": 18990 |
| }, |
| { |
| "epoch": 1.995798319327731, |
| "grad_norm": 12.9375, |
| "learning_rate": 7.481220657276996e-06, |
| "loss": 1.3632, |
| "step": 19000 |
| }, |
| { |
| "epoch": 1.9968487394957983, |
| "grad_norm": 18.0, |
| "learning_rate": 7.473395931142411e-06, |
| "loss": 1.6468, |
| "step": 19010 |
| }, |
| { |
| "epoch": 1.9978991596638656, |
| "grad_norm": 15.6875, |
| "learning_rate": 7.4655712050078255e-06, |
| "loss": 1.5305, |
| "step": 19020 |
| }, |
| { |
| "epoch": 1.9989495798319328, |
| "grad_norm": 38.0, |
| "learning_rate": 7.45774647887324e-06, |
| "loss": 1.3166, |
| "step": 19030 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 15.8125, |
| "learning_rate": 7.449921752738655e-06, |
| "loss": 1.1849, |
| "step": 19040 |
| }, |
| { |
| "epoch": 2.0010504201680672, |
| "grad_norm": 11.5625, |
| "learning_rate": 7.4420970266040695e-06, |
| "loss": 1.2535, |
| "step": 19050 |
| }, |
| { |
| "epoch": 2.0021008403361344, |
| "grad_norm": 18.0, |
| "learning_rate": 7.434272300469484e-06, |
| "loss": 0.9719, |
| "step": 19060 |
| }, |
| { |
| "epoch": 2.0031512605042017, |
| "grad_norm": 11.25, |
| "learning_rate": 7.426447574334899e-06, |
| "loss": 1.3149, |
| "step": 19070 |
| }, |
| { |
| "epoch": 2.004201680672269, |
| "grad_norm": 8.5, |
| "learning_rate": 7.4186228482003134e-06, |
| "loss": 1.257, |
| "step": 19080 |
| }, |
| { |
| "epoch": 2.005252100840336, |
| "grad_norm": 13.5625, |
| "learning_rate": 7.410798122065729e-06, |
| "loss": 1.7853, |
| "step": 19090 |
| }, |
| { |
| "epoch": 2.0063025210084033, |
| "grad_norm": 15.125, |
| "learning_rate": 7.402973395931144e-06, |
| "loss": 1.6606, |
| "step": 19100 |
| }, |
| { |
| "epoch": 2.0073529411764706, |
| "grad_norm": 20.875, |
| "learning_rate": 7.395148669796558e-06, |
| "loss": 1.1459, |
| "step": 19110 |
| }, |
| { |
| "epoch": 2.008403361344538, |
| "grad_norm": 15.1875, |
| "learning_rate": 7.387323943661973e-06, |
| "loss": 1.5116, |
| "step": 19120 |
| }, |
| { |
| "epoch": 2.009453781512605, |
| "grad_norm": 14.375, |
| "learning_rate": 7.3794992175273875e-06, |
| "loss": 1.8419, |
| "step": 19130 |
| }, |
| { |
| "epoch": 2.0105042016806722, |
| "grad_norm": 15.4375, |
| "learning_rate": 7.371674491392802e-06, |
| "loss": 1.3708, |
| "step": 19140 |
| }, |
| { |
| "epoch": 2.0115546218487395, |
| "grad_norm": 13.75, |
| "learning_rate": 7.363849765258216e-06, |
| "loss": 1.3603, |
| "step": 19150 |
| }, |
| { |
| "epoch": 2.0126050420168067, |
| "grad_norm": 13.4375, |
| "learning_rate": 7.356025039123631e-06, |
| "loss": 1.3889, |
| "step": 19160 |
| }, |
| { |
| "epoch": 2.013655462184874, |
| "grad_norm": 12.625, |
| "learning_rate": 7.348200312989045e-06, |
| "loss": 1.4465, |
| "step": 19170 |
| }, |
| { |
| "epoch": 2.014705882352941, |
| "grad_norm": 15.125, |
| "learning_rate": 7.34037558685446e-06, |
| "loss": 1.51, |
| "step": 19180 |
| }, |
| { |
| "epoch": 2.0157563025210083, |
| "grad_norm": 4.15625, |
| "learning_rate": 7.3325508607198755e-06, |
| "loss": 1.547, |
| "step": 19190 |
| }, |
| { |
| "epoch": 2.0168067226890756, |
| "grad_norm": 14.5625, |
| "learning_rate": 7.32472613458529e-06, |
| "loss": 0.7578, |
| "step": 19200 |
| }, |
| { |
| "epoch": 2.017857142857143, |
| "grad_norm": 12.1875, |
| "learning_rate": 7.316901408450705e-06, |
| "loss": 1.379, |
| "step": 19210 |
| }, |
| { |
| "epoch": 2.01890756302521, |
| "grad_norm": 11.75, |
| "learning_rate": 7.309076682316119e-06, |
| "loss": 1.693, |
| "step": 19220 |
| }, |
| { |
| "epoch": 2.0199579831932772, |
| "grad_norm": 6.59375, |
| "learning_rate": 7.301251956181534e-06, |
| "loss": 0.9865, |
| "step": 19230 |
| }, |
| { |
| "epoch": 2.0210084033613445, |
| "grad_norm": 12.125, |
| "learning_rate": 7.293427230046949e-06, |
| "loss": 1.4209, |
| "step": 19240 |
| }, |
| { |
| "epoch": 2.0220588235294117, |
| "grad_norm": 9.8125, |
| "learning_rate": 7.285602503912363e-06, |
| "loss": 1.7937, |
| "step": 19250 |
| }, |
| { |
| "epoch": 2.023109243697479, |
| "grad_norm": 28.125, |
| "learning_rate": 7.277777777777778e-06, |
| "loss": 1.5097, |
| "step": 19260 |
| }, |
| { |
| "epoch": 2.024159663865546, |
| "grad_norm": 14.1875, |
| "learning_rate": 7.269953051643193e-06, |
| "loss": 1.4357, |
| "step": 19270 |
| }, |
| { |
| "epoch": 2.0252100840336134, |
| "grad_norm": 15.625, |
| "learning_rate": 7.262128325508607e-06, |
| "loss": 1.549, |
| "step": 19280 |
| }, |
| { |
| "epoch": 2.0262605042016806, |
| "grad_norm": 8.3125, |
| "learning_rate": 7.254303599374023e-06, |
| "loss": 1.2916, |
| "step": 19290 |
| }, |
| { |
| "epoch": 2.027310924369748, |
| "grad_norm": 15.125, |
| "learning_rate": 7.2464788732394375e-06, |
| "loss": 1.5242, |
| "step": 19300 |
| }, |
| { |
| "epoch": 2.028361344537815, |
| "grad_norm": 12.8125, |
| "learning_rate": 7.238654147104852e-06, |
| "loss": 1.2154, |
| "step": 19310 |
| }, |
| { |
| "epoch": 2.0294117647058822, |
| "grad_norm": 14.75, |
| "learning_rate": 7.230829420970267e-06, |
| "loss": 1.9179, |
| "step": 19320 |
| }, |
| { |
| "epoch": 2.0304621848739495, |
| "grad_norm": 56.75, |
| "learning_rate": 7.2230046948356814e-06, |
| "loss": 1.5349, |
| "step": 19330 |
| }, |
| { |
| "epoch": 2.0315126050420167, |
| "grad_norm": 9.25, |
| "learning_rate": 7.215179968701096e-06, |
| "loss": 1.3553, |
| "step": 19340 |
| }, |
| { |
| "epoch": 2.032563025210084, |
| "grad_norm": 11.9375, |
| "learning_rate": 7.207355242566511e-06, |
| "loss": 1.292, |
| "step": 19350 |
| }, |
| { |
| "epoch": 2.033613445378151, |
| "grad_norm": 12.875, |
| "learning_rate": 7.199530516431925e-06, |
| "loss": 1.8825, |
| "step": 19360 |
| }, |
| { |
| "epoch": 2.0346638655462184, |
| "grad_norm": 11.4375, |
| "learning_rate": 7.19170579029734e-06, |
| "loss": 1.4219, |
| "step": 19370 |
| }, |
| { |
| "epoch": 2.0357142857142856, |
| "grad_norm": 11.625, |
| "learning_rate": 7.1838810641627556e-06, |
| "loss": 1.3718, |
| "step": 19380 |
| }, |
| { |
| "epoch": 2.036764705882353, |
| "grad_norm": 23.25, |
| "learning_rate": 7.17605633802817e-06, |
| "loss": 1.4568, |
| "step": 19390 |
| }, |
| { |
| "epoch": 2.03781512605042, |
| "grad_norm": 13.0, |
| "learning_rate": 7.168231611893585e-06, |
| "loss": 1.4933, |
| "step": 19400 |
| }, |
| { |
| "epoch": 2.0388655462184873, |
| "grad_norm": 4.6875, |
| "learning_rate": 7.1604068857589995e-06, |
| "loss": 1.1494, |
| "step": 19410 |
| }, |
| { |
| "epoch": 2.0399159663865545, |
| "grad_norm": 14.0625, |
| "learning_rate": 7.152582159624414e-06, |
| "loss": 1.1781, |
| "step": 19420 |
| }, |
| { |
| "epoch": 2.0409663865546217, |
| "grad_norm": 13.625, |
| "learning_rate": 7.144757433489829e-06, |
| "loss": 1.5522, |
| "step": 19430 |
| }, |
| { |
| "epoch": 2.042016806722689, |
| "grad_norm": 15.8125, |
| "learning_rate": 7.136932707355243e-06, |
| "loss": 1.2522, |
| "step": 19440 |
| }, |
| { |
| "epoch": 2.043067226890756, |
| "grad_norm": 14.0625, |
| "learning_rate": 7.129107981220657e-06, |
| "loss": 1.0577, |
| "step": 19450 |
| }, |
| { |
| "epoch": 2.0441176470588234, |
| "grad_norm": 15.9375, |
| "learning_rate": 7.121283255086072e-06, |
| "loss": 1.3597, |
| "step": 19460 |
| }, |
| { |
| "epoch": 2.0451680672268906, |
| "grad_norm": 19.125, |
| "learning_rate": 7.113458528951487e-06, |
| "loss": 1.5251, |
| "step": 19470 |
| }, |
| { |
| "epoch": 2.046218487394958, |
| "grad_norm": 12.375, |
| "learning_rate": 7.105633802816903e-06, |
| "loss": 1.4767, |
| "step": 19480 |
| }, |
| { |
| "epoch": 2.047268907563025, |
| "grad_norm": 11.4375, |
| "learning_rate": 7.097809076682317e-06, |
| "loss": 1.2745, |
| "step": 19490 |
| }, |
| { |
| "epoch": 2.0483193277310923, |
| "grad_norm": 7.125, |
| "learning_rate": 7.089984350547731e-06, |
| "loss": 1.2556, |
| "step": 19500 |
| }, |
| { |
| "epoch": 2.0493697478991595, |
| "grad_norm": 17.625, |
| "learning_rate": 7.082159624413146e-06, |
| "loss": 1.4497, |
| "step": 19510 |
| }, |
| { |
| "epoch": 2.0504201680672267, |
| "grad_norm": 8.375, |
| "learning_rate": 7.074334898278561e-06, |
| "loss": 1.151, |
| "step": 19520 |
| }, |
| { |
| "epoch": 2.051470588235294, |
| "grad_norm": 16.5, |
| "learning_rate": 7.066510172143975e-06, |
| "loss": 1.7034, |
| "step": 19530 |
| }, |
| { |
| "epoch": 2.052521008403361, |
| "grad_norm": 11.5625, |
| "learning_rate": 7.05868544600939e-06, |
| "loss": 0.9299, |
| "step": 19540 |
| }, |
| { |
| "epoch": 2.0535714285714284, |
| "grad_norm": 15.625, |
| "learning_rate": 7.050860719874805e-06, |
| "loss": 1.501, |
| "step": 19550 |
| }, |
| { |
| "epoch": 2.0546218487394956, |
| "grad_norm": 18.5, |
| "learning_rate": 7.043035993740219e-06, |
| "loss": 1.1872, |
| "step": 19560 |
| }, |
| { |
| "epoch": 2.055672268907563, |
| "grad_norm": 4.8125, |
| "learning_rate": 7.035211267605634e-06, |
| "loss": 0.9891, |
| "step": 19570 |
| }, |
| { |
| "epoch": 2.05672268907563, |
| "grad_norm": 12.75, |
| "learning_rate": 7.0273865414710495e-06, |
| "loss": 1.0196, |
| "step": 19580 |
| }, |
| { |
| "epoch": 2.0577731092436973, |
| "grad_norm": 19.0, |
| "learning_rate": 7.019561815336464e-06, |
| "loss": 1.3953, |
| "step": 19590 |
| }, |
| { |
| "epoch": 2.0588235294117645, |
| "grad_norm": 11.875, |
| "learning_rate": 7.011737089201879e-06, |
| "loss": 1.0446, |
| "step": 19600 |
| }, |
| { |
| "epoch": 2.0598739495798317, |
| "grad_norm": 14.3125, |
| "learning_rate": 7.003912363067293e-06, |
| "loss": 1.1449, |
| "step": 19610 |
| }, |
| { |
| "epoch": 2.060924369747899, |
| "grad_norm": 11.3125, |
| "learning_rate": 6.996087636932708e-06, |
| "loss": 1.6037, |
| "step": 19620 |
| }, |
| { |
| "epoch": 2.0619747899159666, |
| "grad_norm": 4.0, |
| "learning_rate": 6.988262910798123e-06, |
| "loss": 1.2387, |
| "step": 19630 |
| }, |
| { |
| "epoch": 2.0630252100840334, |
| "grad_norm": 4.5625, |
| "learning_rate": 6.980438184663537e-06, |
| "loss": 0.8792, |
| "step": 19640 |
| }, |
| { |
| "epoch": 2.064075630252101, |
| "grad_norm": 9.0625, |
| "learning_rate": 6.972613458528952e-06, |
| "loss": 1.1865, |
| "step": 19650 |
| }, |
| { |
| "epoch": 2.0651260504201683, |
| "grad_norm": 11.0625, |
| "learning_rate": 6.964788732394367e-06, |
| "loss": 1.3791, |
| "step": 19660 |
| }, |
| { |
| "epoch": 2.0661764705882355, |
| "grad_norm": 15.3125, |
| "learning_rate": 6.956964006259781e-06, |
| "loss": 1.3943, |
| "step": 19670 |
| }, |
| { |
| "epoch": 2.0672268907563027, |
| "grad_norm": 22.875, |
| "learning_rate": 6.949139280125197e-06, |
| "loss": 1.3977, |
| "step": 19680 |
| }, |
| { |
| "epoch": 2.06827731092437, |
| "grad_norm": 14.6875, |
| "learning_rate": 6.9413145539906115e-06, |
| "loss": 1.6686, |
| "step": 19690 |
| }, |
| { |
| "epoch": 2.069327731092437, |
| "grad_norm": 16.25, |
| "learning_rate": 6.933489827856026e-06, |
| "loss": 1.8928, |
| "step": 19700 |
| }, |
| { |
| "epoch": 2.0703781512605044, |
| "grad_norm": 13.1875, |
| "learning_rate": 6.925665101721441e-06, |
| "loss": 1.2138, |
| "step": 19710 |
| }, |
| { |
| "epoch": 2.0714285714285716, |
| "grad_norm": 17.5, |
| "learning_rate": 6.9178403755868554e-06, |
| "loss": 1.5655, |
| "step": 19720 |
| }, |
| { |
| "epoch": 2.072478991596639, |
| "grad_norm": 8.875, |
| "learning_rate": 6.910015649452269e-06, |
| "loss": 1.2248, |
| "step": 19730 |
| }, |
| { |
| "epoch": 2.073529411764706, |
| "grad_norm": 14.0625, |
| "learning_rate": 6.902190923317684e-06, |
| "loss": 1.3255, |
| "step": 19740 |
| }, |
| { |
| "epoch": 2.0745798319327733, |
| "grad_norm": 4.375, |
| "learning_rate": 6.8943661971830986e-06, |
| "loss": 1.1353, |
| "step": 19750 |
| }, |
| { |
| "epoch": 2.0756302521008405, |
| "grad_norm": 21.25, |
| "learning_rate": 6.886541471048513e-06, |
| "loss": 1.2565, |
| "step": 19760 |
| }, |
| { |
| "epoch": 2.0766806722689077, |
| "grad_norm": 13.0625, |
| "learning_rate": 6.878716744913928e-06, |
| "loss": 1.0822, |
| "step": 19770 |
| }, |
| { |
| "epoch": 2.077731092436975, |
| "grad_norm": 18.25, |
| "learning_rate": 6.870892018779343e-06, |
| "loss": 1.4547, |
| "step": 19780 |
| }, |
| { |
| "epoch": 2.078781512605042, |
| "grad_norm": 17.125, |
| "learning_rate": 6.863067292644758e-06, |
| "loss": 1.5542, |
| "step": 19790 |
| }, |
| { |
| "epoch": 2.0798319327731094, |
| "grad_norm": 12.0, |
| "learning_rate": 6.855242566510173e-06, |
| "loss": 1.2621, |
| "step": 19800 |
| }, |
| { |
| "epoch": 2.0808823529411766, |
| "grad_norm": 13.6875, |
| "learning_rate": 6.847417840375587e-06, |
| "loss": 1.1667, |
| "step": 19810 |
| }, |
| { |
| "epoch": 2.081932773109244, |
| "grad_norm": 14.8125, |
| "learning_rate": 6.839593114241002e-06, |
| "loss": 1.8507, |
| "step": 19820 |
| }, |
| { |
| "epoch": 2.082983193277311, |
| "grad_norm": 11.9375, |
| "learning_rate": 6.831768388106417e-06, |
| "loss": 1.0498, |
| "step": 19830 |
| }, |
| { |
| "epoch": 2.0840336134453783, |
| "grad_norm": 13.875, |
| "learning_rate": 6.823943661971831e-06, |
| "loss": 1.2932, |
| "step": 19840 |
| }, |
| { |
| "epoch": 2.0850840336134455, |
| "grad_norm": 12.9375, |
| "learning_rate": 6.816118935837246e-06, |
| "loss": 1.2906, |
| "step": 19850 |
| }, |
| { |
| "epoch": 2.0861344537815127, |
| "grad_norm": 12.1875, |
| "learning_rate": 6.808294209702661e-06, |
| "loss": 1.0509, |
| "step": 19860 |
| }, |
| { |
| "epoch": 2.08718487394958, |
| "grad_norm": 12.8125, |
| "learning_rate": 6.800469483568075e-06, |
| "loss": 1.0825, |
| "step": 19870 |
| }, |
| { |
| "epoch": 2.088235294117647, |
| "grad_norm": 19.25, |
| "learning_rate": 6.792644757433491e-06, |
| "loss": 1.3501, |
| "step": 19880 |
| }, |
| { |
| "epoch": 2.0892857142857144, |
| "grad_norm": 13.0625, |
| "learning_rate": 6.784820031298905e-06, |
| "loss": 1.639, |
| "step": 19890 |
| }, |
| { |
| "epoch": 2.0903361344537816, |
| "grad_norm": 26.0, |
| "learning_rate": 6.77699530516432e-06, |
| "loss": 1.3657, |
| "step": 19900 |
| }, |
| { |
| "epoch": 2.091386554621849, |
| "grad_norm": 19.375, |
| "learning_rate": 6.769170579029735e-06, |
| "loss": 1.62, |
| "step": 19910 |
| }, |
| { |
| "epoch": 2.092436974789916, |
| "grad_norm": 5.46875, |
| "learning_rate": 6.761345852895149e-06, |
| "loss": 1.158, |
| "step": 19920 |
| }, |
| { |
| "epoch": 2.0934873949579833, |
| "grad_norm": 4.78125, |
| "learning_rate": 6.753521126760564e-06, |
| "loss": 0.9588, |
| "step": 19930 |
| }, |
| { |
| "epoch": 2.0945378151260505, |
| "grad_norm": 15.75, |
| "learning_rate": 6.745696400625979e-06, |
| "loss": 1.3956, |
| "step": 19940 |
| }, |
| { |
| "epoch": 2.0955882352941178, |
| "grad_norm": 4.65625, |
| "learning_rate": 6.737871674491393e-06, |
| "loss": 1.192, |
| "step": 19950 |
| }, |
| { |
| "epoch": 2.096638655462185, |
| "grad_norm": 13.0625, |
| "learning_rate": 6.730046948356808e-06, |
| "loss": 1.3621, |
| "step": 19960 |
| }, |
| { |
| "epoch": 2.097689075630252, |
| "grad_norm": 16.0, |
| "learning_rate": 6.7222222222222235e-06, |
| "loss": 1.1402, |
| "step": 19970 |
| }, |
| { |
| "epoch": 2.0987394957983194, |
| "grad_norm": 12.0625, |
| "learning_rate": 6.714397496087638e-06, |
| "loss": 1.1546, |
| "step": 19980 |
| }, |
| { |
| "epoch": 2.0997899159663866, |
| "grad_norm": 4.84375, |
| "learning_rate": 6.706572769953053e-06, |
| "loss": 1.1781, |
| "step": 19990 |
| }, |
| { |
| "epoch": 2.100840336134454, |
| "grad_norm": 11.625, |
| "learning_rate": 6.698748043818467e-06, |
| "loss": 1.2507, |
| "step": 20000 |
| }, |
| { |
| "epoch": 2.101890756302521, |
| "grad_norm": 13.4375, |
| "learning_rate": 6.690923317683882e-06, |
| "loss": 1.3831, |
| "step": 20010 |
| }, |
| { |
| "epoch": 2.1029411764705883, |
| "grad_norm": 8.5, |
| "learning_rate": 6.683098591549296e-06, |
| "loss": 0.8563, |
| "step": 20020 |
| }, |
| { |
| "epoch": 2.1039915966386555, |
| "grad_norm": 15.6875, |
| "learning_rate": 6.6752738654147105e-06, |
| "loss": 1.4442, |
| "step": 20030 |
| }, |
| { |
| "epoch": 2.1050420168067228, |
| "grad_norm": 12.875, |
| "learning_rate": 6.667449139280125e-06, |
| "loss": 1.5468, |
| "step": 20040 |
| }, |
| { |
| "epoch": 2.10609243697479, |
| "grad_norm": 9.4375, |
| "learning_rate": 6.65962441314554e-06, |
| "loss": 1.151, |
| "step": 20050 |
| }, |
| { |
| "epoch": 2.107142857142857, |
| "grad_norm": 9.0625, |
| "learning_rate": 6.6517996870109545e-06, |
| "loss": 1.4075, |
| "step": 20060 |
| }, |
| { |
| "epoch": 2.1081932773109244, |
| "grad_norm": 20.25, |
| "learning_rate": 6.64397496087637e-06, |
| "loss": 1.3844, |
| "step": 20070 |
| }, |
| { |
| "epoch": 2.1092436974789917, |
| "grad_norm": 4.84375, |
| "learning_rate": 6.636150234741785e-06, |
| "loss": 1.6283, |
| "step": 20080 |
| }, |
| { |
| "epoch": 2.110294117647059, |
| "grad_norm": 13.1875, |
| "learning_rate": 6.628325508607199e-06, |
| "loss": 1.0903, |
| "step": 20090 |
| }, |
| { |
| "epoch": 2.111344537815126, |
| "grad_norm": 12.0, |
| "learning_rate": 6.620500782472614e-06, |
| "loss": 1.4893, |
| "step": 20100 |
| }, |
| { |
| "epoch": 2.1123949579831933, |
| "grad_norm": 12.8125, |
| "learning_rate": 6.612676056338029e-06, |
| "loss": 1.5213, |
| "step": 20110 |
| }, |
| { |
| "epoch": 2.1134453781512605, |
| "grad_norm": 13.9375, |
| "learning_rate": 6.604851330203443e-06, |
| "loss": 1.2929, |
| "step": 20120 |
| }, |
| { |
| "epoch": 2.1144957983193278, |
| "grad_norm": 11.4375, |
| "learning_rate": 6.597026604068858e-06, |
| "loss": 1.2527, |
| "step": 20130 |
| }, |
| { |
| "epoch": 2.115546218487395, |
| "grad_norm": 20.625, |
| "learning_rate": 6.5892018779342726e-06, |
| "loss": 1.4411, |
| "step": 20140 |
| }, |
| { |
| "epoch": 2.116596638655462, |
| "grad_norm": 4.09375, |
| "learning_rate": 6.581377151799687e-06, |
| "loss": 1.4455, |
| "step": 20150 |
| }, |
| { |
| "epoch": 2.1176470588235294, |
| "grad_norm": 18.875, |
| "learning_rate": 6.573552425665102e-06, |
| "loss": 1.6437, |
| "step": 20160 |
| }, |
| { |
| "epoch": 2.1186974789915967, |
| "grad_norm": 12.8125, |
| "learning_rate": 6.565727699530517e-06, |
| "loss": 1.1723, |
| "step": 20170 |
| }, |
| { |
| "epoch": 2.119747899159664, |
| "grad_norm": 13.5, |
| "learning_rate": 6.557902973395932e-06, |
| "loss": 1.3895, |
| "step": 20180 |
| }, |
| { |
| "epoch": 2.120798319327731, |
| "grad_norm": 3.9375, |
| "learning_rate": 6.550078247261347e-06, |
| "loss": 1.1102, |
| "step": 20190 |
| }, |
| { |
| "epoch": 2.1218487394957983, |
| "grad_norm": 16.75, |
| "learning_rate": 6.542253521126761e-06, |
| "loss": 1.5825, |
| "step": 20200 |
| }, |
| { |
| "epoch": 2.1228991596638656, |
| "grad_norm": 10.875, |
| "learning_rate": 6.534428794992176e-06, |
| "loss": 1.0678, |
| "step": 20210 |
| }, |
| { |
| "epoch": 2.1239495798319328, |
| "grad_norm": 10.25, |
| "learning_rate": 6.526604068857591e-06, |
| "loss": 1.5383, |
| "step": 20220 |
| }, |
| { |
| "epoch": 2.125, |
| "grad_norm": 18.5, |
| "learning_rate": 6.518779342723005e-06, |
| "loss": 0.979, |
| "step": 20230 |
| }, |
| { |
| "epoch": 2.1260504201680672, |
| "grad_norm": 11.4375, |
| "learning_rate": 6.51095461658842e-06, |
| "loss": 1.2101, |
| "step": 20240 |
| }, |
| { |
| "epoch": 2.1271008403361344, |
| "grad_norm": 5.4375, |
| "learning_rate": 6.503129890453835e-06, |
| "loss": 1.0491, |
| "step": 20250 |
| }, |
| { |
| "epoch": 2.1281512605042017, |
| "grad_norm": 9.375, |
| "learning_rate": 6.495305164319249e-06, |
| "loss": 0.8727, |
| "step": 20260 |
| }, |
| { |
| "epoch": 2.129201680672269, |
| "grad_norm": 24.25, |
| "learning_rate": 6.487480438184665e-06, |
| "loss": 1.3615, |
| "step": 20270 |
| }, |
| { |
| "epoch": 2.130252100840336, |
| "grad_norm": 3.390625, |
| "learning_rate": 6.479655712050079e-06, |
| "loss": 0.8916, |
| "step": 20280 |
| }, |
| { |
| "epoch": 2.1313025210084033, |
| "grad_norm": 10.0625, |
| "learning_rate": 6.471830985915494e-06, |
| "loss": 1.0757, |
| "step": 20290 |
| }, |
| { |
| "epoch": 2.1323529411764706, |
| "grad_norm": 13.5, |
| "learning_rate": 6.464006259780909e-06, |
| "loss": 1.4941, |
| "step": 20300 |
| }, |
| { |
| "epoch": 2.133403361344538, |
| "grad_norm": 4.5, |
| "learning_rate": 6.456181533646323e-06, |
| "loss": 1.1808, |
| "step": 20310 |
| }, |
| { |
| "epoch": 2.134453781512605, |
| "grad_norm": 13.6875, |
| "learning_rate": 6.448356807511737e-06, |
| "loss": 1.61, |
| "step": 20320 |
| }, |
| { |
| "epoch": 2.1355042016806722, |
| "grad_norm": 14.4375, |
| "learning_rate": 6.440532081377152e-06, |
| "loss": 1.0927, |
| "step": 20330 |
| }, |
| { |
| "epoch": 2.1365546218487395, |
| "grad_norm": 12.4375, |
| "learning_rate": 6.4327073552425665e-06, |
| "loss": 0.9832, |
| "step": 20340 |
| }, |
| { |
| "epoch": 2.1376050420168067, |
| "grad_norm": 12.1875, |
| "learning_rate": 6.424882629107981e-06, |
| "loss": 1.8432, |
| "step": 20350 |
| }, |
| { |
| "epoch": 2.138655462184874, |
| "grad_norm": 13.8125, |
| "learning_rate": 6.417057902973396e-06, |
| "loss": 1.3069, |
| "step": 20360 |
| }, |
| { |
| "epoch": 2.139705882352941, |
| "grad_norm": 139.0, |
| "learning_rate": 6.409233176838811e-06, |
| "loss": 1.438, |
| "step": 20370 |
| }, |
| { |
| "epoch": 2.1407563025210083, |
| "grad_norm": 14.0, |
| "learning_rate": 6.401408450704226e-06, |
| "loss": 1.1414, |
| "step": 20380 |
| }, |
| { |
| "epoch": 2.1418067226890756, |
| "grad_norm": 8.875, |
| "learning_rate": 6.3935837245696406e-06, |
| "loss": 1.1154, |
| "step": 20390 |
| }, |
| { |
| "epoch": 2.142857142857143, |
| "grad_norm": 12.3125, |
| "learning_rate": 6.385758998435055e-06, |
| "loss": 1.225, |
| "step": 20400 |
| }, |
| { |
| "epoch": 2.14390756302521, |
| "grad_norm": 12.75, |
| "learning_rate": 6.37793427230047e-06, |
| "loss": 0.8264, |
| "step": 20410 |
| }, |
| { |
| "epoch": 2.1449579831932772, |
| "grad_norm": 41.75, |
| "learning_rate": 6.3701095461658845e-06, |
| "loss": 1.3104, |
| "step": 20420 |
| }, |
| { |
| "epoch": 2.1460084033613445, |
| "grad_norm": 22.75, |
| "learning_rate": 6.362284820031299e-06, |
| "loss": 1.6172, |
| "step": 20430 |
| }, |
| { |
| "epoch": 2.1470588235294117, |
| "grad_norm": 16.875, |
| "learning_rate": 6.354460093896714e-06, |
| "loss": 1.4141, |
| "step": 20440 |
| }, |
| { |
| "epoch": 2.148109243697479, |
| "grad_norm": 17.0, |
| "learning_rate": 6.3466353677621285e-06, |
| "loss": 1.0977, |
| "step": 20450 |
| }, |
| { |
| "epoch": 2.149159663865546, |
| "grad_norm": 11.375, |
| "learning_rate": 6.338810641627543e-06, |
| "loss": 1.2796, |
| "step": 20460 |
| }, |
| { |
| "epoch": 2.1502100840336134, |
| "grad_norm": 15.875, |
| "learning_rate": 6.330985915492959e-06, |
| "loss": 1.1421, |
| "step": 20470 |
| }, |
| { |
| "epoch": 2.1512605042016806, |
| "grad_norm": 11.25, |
| "learning_rate": 6.323161189358373e-06, |
| "loss": 1.1076, |
| "step": 20480 |
| }, |
| { |
| "epoch": 2.152310924369748, |
| "grad_norm": 4.4375, |
| "learning_rate": 6.315336463223788e-06, |
| "loss": 1.0524, |
| "step": 20490 |
| }, |
| { |
| "epoch": 2.153361344537815, |
| "grad_norm": 15.1875, |
| "learning_rate": 6.307511737089203e-06, |
| "loss": 1.2327, |
| "step": 20500 |
| }, |
| { |
| "epoch": 2.1544117647058822, |
| "grad_norm": 16.625, |
| "learning_rate": 6.299687010954617e-06, |
| "loss": 1.4369, |
| "step": 20510 |
| }, |
| { |
| "epoch": 2.1554621848739495, |
| "grad_norm": 12.6875, |
| "learning_rate": 6.291862284820032e-06, |
| "loss": 1.1132, |
| "step": 20520 |
| }, |
| { |
| "epoch": 2.1565126050420167, |
| "grad_norm": 17.0, |
| "learning_rate": 6.2840375586854466e-06, |
| "loss": 1.4176, |
| "step": 20530 |
| }, |
| { |
| "epoch": 2.157563025210084, |
| "grad_norm": 11.25, |
| "learning_rate": 6.276212832550861e-06, |
| "loss": 1.5324, |
| "step": 20540 |
| }, |
| { |
| "epoch": 2.158613445378151, |
| "grad_norm": 9.6875, |
| "learning_rate": 6.268388106416276e-06, |
| "loss": 1.7249, |
| "step": 20550 |
| }, |
| { |
| "epoch": 2.1596638655462184, |
| "grad_norm": 12.125, |
| "learning_rate": 6.260563380281691e-06, |
| "loss": 1.595, |
| "step": 20560 |
| }, |
| { |
| "epoch": 2.1607142857142856, |
| "grad_norm": 8.9375, |
| "learning_rate": 6.252738654147106e-06, |
| "loss": 1.4942, |
| "step": 20570 |
| }, |
| { |
| "epoch": 2.161764705882353, |
| "grad_norm": 14.3125, |
| "learning_rate": 6.244913928012521e-06, |
| "loss": 1.3773, |
| "step": 20580 |
| }, |
| { |
| "epoch": 2.16281512605042, |
| "grad_norm": 28.125, |
| "learning_rate": 6.237089201877935e-06, |
| "loss": 1.6441, |
| "step": 20590 |
| }, |
| { |
| "epoch": 2.1638655462184873, |
| "grad_norm": 16.75, |
| "learning_rate": 6.22926447574335e-06, |
| "loss": 1.139, |
| "step": 20600 |
| }, |
| { |
| "epoch": 2.1649159663865545, |
| "grad_norm": 12.5, |
| "learning_rate": 6.221439749608764e-06, |
| "loss": 1.5922, |
| "step": 20610 |
| }, |
| { |
| "epoch": 2.1659663865546217, |
| "grad_norm": 2.984375, |
| "learning_rate": 6.2136150234741784e-06, |
| "loss": 0.9379, |
| "step": 20620 |
| }, |
| { |
| "epoch": 2.167016806722689, |
| "grad_norm": 10.625, |
| "learning_rate": 6.205790297339593e-06, |
| "loss": 1.0389, |
| "step": 20630 |
| }, |
| { |
| "epoch": 2.168067226890756, |
| "grad_norm": 28.0, |
| "learning_rate": 6.197965571205008e-06, |
| "loss": 1.2278, |
| "step": 20640 |
| }, |
| { |
| "epoch": 2.1691176470588234, |
| "grad_norm": 16.0, |
| "learning_rate": 6.190140845070422e-06, |
| "loss": 1.1912, |
| "step": 20650 |
| }, |
| { |
| "epoch": 2.1701680672268906, |
| "grad_norm": 14.75, |
| "learning_rate": 6.182316118935838e-06, |
| "loss": 1.3608, |
| "step": 20660 |
| }, |
| { |
| "epoch": 2.171218487394958, |
| "grad_norm": 21.125, |
| "learning_rate": 6.1744913928012525e-06, |
| "loss": 1.0686, |
| "step": 20670 |
| }, |
| { |
| "epoch": 2.172268907563025, |
| "grad_norm": 10.125, |
| "learning_rate": 6.166666666666667e-06, |
| "loss": 1.2551, |
| "step": 20680 |
| }, |
| { |
| "epoch": 2.1733193277310923, |
| "grad_norm": 14.4375, |
| "learning_rate": 6.158841940532082e-06, |
| "loss": 1.511, |
| "step": 20690 |
| }, |
| { |
| "epoch": 2.1743697478991595, |
| "grad_norm": 11.5625, |
| "learning_rate": 6.1510172143974965e-06, |
| "loss": 1.4704, |
| "step": 20700 |
| }, |
| { |
| "epoch": 2.1754201680672267, |
| "grad_norm": 12.9375, |
| "learning_rate": 6.143192488262911e-06, |
| "loss": 1.6788, |
| "step": 20710 |
| }, |
| { |
| "epoch": 2.176470588235294, |
| "grad_norm": 22.25, |
| "learning_rate": 6.135367762128326e-06, |
| "loss": 1.4955, |
| "step": 20720 |
| }, |
| { |
| "epoch": 2.177521008403361, |
| "grad_norm": 16.75, |
| "learning_rate": 6.1275430359937405e-06, |
| "loss": 1.3728, |
| "step": 20730 |
| }, |
| { |
| "epoch": 2.1785714285714284, |
| "grad_norm": 13.25, |
| "learning_rate": 6.119718309859155e-06, |
| "loss": 1.213, |
| "step": 20740 |
| }, |
| { |
| "epoch": 2.1796218487394956, |
| "grad_norm": 14.625, |
| "learning_rate": 6.11189358372457e-06, |
| "loss": 1.6045, |
| "step": 20750 |
| }, |
| { |
| "epoch": 2.180672268907563, |
| "grad_norm": 13.75, |
| "learning_rate": 6.104068857589985e-06, |
| "loss": 1.4784, |
| "step": 20760 |
| }, |
| { |
| "epoch": 2.18172268907563, |
| "grad_norm": 11.0, |
| "learning_rate": 6.0962441314554e-06, |
| "loss": 0.9013, |
| "step": 20770 |
| }, |
| { |
| "epoch": 2.1827731092436973, |
| "grad_norm": 20.0, |
| "learning_rate": 6.0884194053208146e-06, |
| "loss": 1.4597, |
| "step": 20780 |
| }, |
| { |
| "epoch": 2.1838235294117645, |
| "grad_norm": 75.0, |
| "learning_rate": 6.080594679186229e-06, |
| "loss": 1.5331, |
| "step": 20790 |
| }, |
| { |
| "epoch": 2.184873949579832, |
| "grad_norm": 11.5625, |
| "learning_rate": 6.072769953051644e-06, |
| "loss": 1.0339, |
| "step": 20800 |
| }, |
| { |
| "epoch": 2.185924369747899, |
| "grad_norm": 12.375, |
| "learning_rate": 6.0649452269170585e-06, |
| "loss": 0.9868, |
| "step": 20810 |
| }, |
| { |
| "epoch": 2.1869747899159666, |
| "grad_norm": 14.1875, |
| "learning_rate": 6.057120500782473e-06, |
| "loss": 0.9225, |
| "step": 20820 |
| }, |
| { |
| "epoch": 2.1880252100840334, |
| "grad_norm": 13.5, |
| "learning_rate": 6.049295774647888e-06, |
| "loss": 1.4781, |
| "step": 20830 |
| }, |
| { |
| "epoch": 2.189075630252101, |
| "grad_norm": 13.625, |
| "learning_rate": 6.0414710485133025e-06, |
| "loss": 1.2404, |
| "step": 20840 |
| }, |
| { |
| "epoch": 2.190126050420168, |
| "grad_norm": 13.6875, |
| "learning_rate": 6.033646322378717e-06, |
| "loss": 1.6465, |
| "step": 20850 |
| }, |
| { |
| "epoch": 2.1911764705882355, |
| "grad_norm": 14.1875, |
| "learning_rate": 6.025821596244133e-06, |
| "loss": 1.2728, |
| "step": 20860 |
| }, |
| { |
| "epoch": 2.1922268907563027, |
| "grad_norm": 3.90625, |
| "learning_rate": 6.017996870109547e-06, |
| "loss": 1.2817, |
| "step": 20870 |
| }, |
| { |
| "epoch": 2.19327731092437, |
| "grad_norm": 6.03125, |
| "learning_rate": 6.010172143974962e-06, |
| "loss": 1.1268, |
| "step": 20880 |
| }, |
| { |
| "epoch": 2.194327731092437, |
| "grad_norm": 6.40625, |
| "learning_rate": 6.002347417840377e-06, |
| "loss": 1.4204, |
| "step": 20890 |
| }, |
| { |
| "epoch": 2.1953781512605044, |
| "grad_norm": 16.0, |
| "learning_rate": 5.99452269170579e-06, |
| "loss": 1.413, |
| "step": 20900 |
| }, |
| { |
| "epoch": 2.1964285714285716, |
| "grad_norm": 19.625, |
| "learning_rate": 5.986697965571205e-06, |
| "loss": 1.2378, |
| "step": 20910 |
| }, |
| { |
| "epoch": 2.197478991596639, |
| "grad_norm": 12.0625, |
| "learning_rate": 5.97887323943662e-06, |
| "loss": 1.3511, |
| "step": 20920 |
| }, |
| { |
| "epoch": 2.198529411764706, |
| "grad_norm": 11.3125, |
| "learning_rate": 5.971048513302034e-06, |
| "loss": 1.9652, |
| "step": 20930 |
| }, |
| { |
| "epoch": 2.1995798319327733, |
| "grad_norm": 17.875, |
| "learning_rate": 5.963223787167449e-06, |
| "loss": 1.3321, |
| "step": 20940 |
| }, |
| { |
| "epoch": 2.2006302521008405, |
| "grad_norm": 11.625, |
| "learning_rate": 5.955399061032864e-06, |
| "loss": 0.9774, |
| "step": 20950 |
| }, |
| { |
| "epoch": 2.2016806722689077, |
| "grad_norm": 79.0, |
| "learning_rate": 5.947574334898279e-06, |
| "loss": 1.2817, |
| "step": 20960 |
| }, |
| { |
| "epoch": 2.202731092436975, |
| "grad_norm": 24.125, |
| "learning_rate": 5.939749608763694e-06, |
| "loss": 1.328, |
| "step": 20970 |
| }, |
| { |
| "epoch": 2.203781512605042, |
| "grad_norm": 4.15625, |
| "learning_rate": 5.9319248826291085e-06, |
| "loss": 1.2104, |
| "step": 20980 |
| }, |
| { |
| "epoch": 2.2048319327731094, |
| "grad_norm": 12.6875, |
| "learning_rate": 5.924100156494523e-06, |
| "loss": 1.7662, |
| "step": 20990 |
| }, |
| { |
| "epoch": 2.2058823529411766, |
| "grad_norm": 25.875, |
| "learning_rate": 5.916275430359938e-06, |
| "loss": 1.0362, |
| "step": 21000 |
| }, |
| { |
| "epoch": 2.206932773109244, |
| "grad_norm": 16.75, |
| "learning_rate": 5.9084507042253524e-06, |
| "loss": 1.5403, |
| "step": 21010 |
| }, |
| { |
| "epoch": 2.207983193277311, |
| "grad_norm": 17.0, |
| "learning_rate": 5.900625978090767e-06, |
| "loss": 1.559, |
| "step": 21020 |
| }, |
| { |
| "epoch": 2.2090336134453783, |
| "grad_norm": 11.625, |
| "learning_rate": 5.892801251956182e-06, |
| "loss": 1.4248, |
| "step": 21030 |
| }, |
| { |
| "epoch": 2.2100840336134455, |
| "grad_norm": 26.875, |
| "learning_rate": 5.884976525821596e-06, |
| "loss": 1.8425, |
| "step": 21040 |
| }, |
| { |
| "epoch": 2.2111344537815127, |
| "grad_norm": 15.0625, |
| "learning_rate": 5.877151799687011e-06, |
| "loss": 1.2505, |
| "step": 21050 |
| }, |
| { |
| "epoch": 2.21218487394958, |
| "grad_norm": 13.0625, |
| "learning_rate": 5.8693270735524265e-06, |
| "loss": 1.2495, |
| "step": 21060 |
| }, |
| { |
| "epoch": 2.213235294117647, |
| "grad_norm": 12.1875, |
| "learning_rate": 5.861502347417841e-06, |
| "loss": 1.0802, |
| "step": 21070 |
| }, |
| { |
| "epoch": 2.2142857142857144, |
| "grad_norm": 66.5, |
| "learning_rate": 5.853677621283256e-06, |
| "loss": 1.3109, |
| "step": 21080 |
| }, |
| { |
| "epoch": 2.2153361344537816, |
| "grad_norm": 21.0, |
| "learning_rate": 5.8458528951486705e-06, |
| "loss": 1.2991, |
| "step": 21090 |
| }, |
| { |
| "epoch": 2.216386554621849, |
| "grad_norm": 12.9375, |
| "learning_rate": 5.838028169014085e-06, |
| "loss": 1.0582, |
| "step": 21100 |
| }, |
| { |
| "epoch": 2.217436974789916, |
| "grad_norm": 8.75, |
| "learning_rate": 5.8302034428795e-06, |
| "loss": 1.3465, |
| "step": 21110 |
| }, |
| { |
| "epoch": 2.2184873949579833, |
| "grad_norm": 12.0, |
| "learning_rate": 5.8223787167449145e-06, |
| "loss": 1.6646, |
| "step": 21120 |
| }, |
| { |
| "epoch": 2.2195378151260505, |
| "grad_norm": 12.4375, |
| "learning_rate": 5.814553990610329e-06, |
| "loss": 0.9414, |
| "step": 21130 |
| }, |
| { |
| "epoch": 2.2205882352941178, |
| "grad_norm": 4.25, |
| "learning_rate": 5.806729264475744e-06, |
| "loss": 1.1275, |
| "step": 21140 |
| }, |
| { |
| "epoch": 2.221638655462185, |
| "grad_norm": 17.5, |
| "learning_rate": 5.798904538341159e-06, |
| "loss": 1.2562, |
| "step": 21150 |
| }, |
| { |
| "epoch": 2.222689075630252, |
| "grad_norm": 28.125, |
| "learning_rate": 5.791079812206574e-06, |
| "loss": 1.1027, |
| "step": 21160 |
| }, |
| { |
| "epoch": 2.2237394957983194, |
| "grad_norm": 13.375, |
| "learning_rate": 5.783255086071989e-06, |
| "loss": 1.8837, |
| "step": 21170 |
| }, |
| { |
| "epoch": 2.2247899159663866, |
| "grad_norm": 13.3125, |
| "learning_rate": 5.775430359937403e-06, |
| "loss": 1.1867, |
| "step": 21180 |
| }, |
| { |
| "epoch": 2.225840336134454, |
| "grad_norm": 13.9375, |
| "learning_rate": 5.767605633802818e-06, |
| "loss": 1.3971, |
| "step": 21190 |
| }, |
| { |
| "epoch": 2.226890756302521, |
| "grad_norm": 22.5, |
| "learning_rate": 5.759780907668232e-06, |
| "loss": 1.2463, |
| "step": 21200 |
| }, |
| { |
| "epoch": 2.2279411764705883, |
| "grad_norm": 15.9375, |
| "learning_rate": 5.751956181533646e-06, |
| "loss": 1.6531, |
| "step": 21210 |
| }, |
| { |
| "epoch": 2.2289915966386555, |
| "grad_norm": 7.40625, |
| "learning_rate": 5.744131455399061e-06, |
| "loss": 1.3287, |
| "step": 21220 |
| }, |
| { |
| "epoch": 2.2300420168067228, |
| "grad_norm": 7.90625, |
| "learning_rate": 5.736306729264476e-06, |
| "loss": 0.915, |
| "step": 21230 |
| }, |
| { |
| "epoch": 2.23109243697479, |
| "grad_norm": 8.6875, |
| "learning_rate": 5.72848200312989e-06, |
| "loss": 1.3027, |
| "step": 21240 |
| }, |
| { |
| "epoch": 2.232142857142857, |
| "grad_norm": 22.0, |
| "learning_rate": 5.720657276995306e-06, |
| "loss": 1.6918, |
| "step": 21250 |
| }, |
| { |
| "epoch": 2.2331932773109244, |
| "grad_norm": 9.75, |
| "learning_rate": 5.7128325508607205e-06, |
| "loss": 0.9006, |
| "step": 21260 |
| }, |
| { |
| "epoch": 2.2342436974789917, |
| "grad_norm": 14.4375, |
| "learning_rate": 5.705007824726135e-06, |
| "loss": 1.3935, |
| "step": 21270 |
| }, |
| { |
| "epoch": 2.235294117647059, |
| "grad_norm": 12.6875, |
| "learning_rate": 5.69718309859155e-06, |
| "loss": 1.321, |
| "step": 21280 |
| }, |
| { |
| "epoch": 2.236344537815126, |
| "grad_norm": 20.625, |
| "learning_rate": 5.689358372456964e-06, |
| "loss": 0.9012, |
| "step": 21290 |
| }, |
| { |
| "epoch": 2.2373949579831933, |
| "grad_norm": 9.9375, |
| "learning_rate": 5.681533646322379e-06, |
| "loss": 0.8183, |
| "step": 21300 |
| }, |
| { |
| "epoch": 2.2384453781512605, |
| "grad_norm": 25.375, |
| "learning_rate": 5.673708920187794e-06, |
| "loss": 1.538, |
| "step": 21310 |
| }, |
| { |
| "epoch": 2.2394957983193278, |
| "grad_norm": 15.625, |
| "learning_rate": 5.665884194053208e-06, |
| "loss": 1.6829, |
| "step": 21320 |
| }, |
| { |
| "epoch": 2.240546218487395, |
| "grad_norm": 11.25, |
| "learning_rate": 5.658059467918623e-06, |
| "loss": 1.3246, |
| "step": 21330 |
| }, |
| { |
| "epoch": 2.241596638655462, |
| "grad_norm": 5.46875, |
| "learning_rate": 5.650234741784038e-06, |
| "loss": 0.9689, |
| "step": 21340 |
| }, |
| { |
| "epoch": 2.2426470588235294, |
| "grad_norm": 12.375, |
| "learning_rate": 5.642410015649453e-06, |
| "loss": 1.1202, |
| "step": 21350 |
| }, |
| { |
| "epoch": 2.2436974789915967, |
| "grad_norm": 11.6875, |
| "learning_rate": 5.634585289514868e-06, |
| "loss": 1.2305, |
| "step": 21360 |
| }, |
| { |
| "epoch": 2.244747899159664, |
| "grad_norm": 5.03125, |
| "learning_rate": 5.6267605633802825e-06, |
| "loss": 0.9779, |
| "step": 21370 |
| }, |
| { |
| "epoch": 2.245798319327731, |
| "grad_norm": 14.25, |
| "learning_rate": 5.618935837245697e-06, |
| "loss": 1.0515, |
| "step": 21380 |
| }, |
| { |
| "epoch": 2.2468487394957983, |
| "grad_norm": 16.5, |
| "learning_rate": 5.611111111111112e-06, |
| "loss": 1.0545, |
| "step": 21390 |
| }, |
| { |
| "epoch": 2.2478991596638656, |
| "grad_norm": 12.9375, |
| "learning_rate": 5.6032863849765264e-06, |
| "loss": 1.1076, |
| "step": 21400 |
| }, |
| { |
| "epoch": 2.2489495798319328, |
| "grad_norm": 15.4375, |
| "learning_rate": 5.595461658841941e-06, |
| "loss": 1.0142, |
| "step": 21410 |
| }, |
| { |
| "epoch": 2.25, |
| "grad_norm": 15.875, |
| "learning_rate": 5.587636932707356e-06, |
| "loss": 1.5, |
| "step": 21420 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 28560, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 7140, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|