| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 32.142857142857146, |
| "eval_steps": 50, |
| "global_step": 450, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.07142857142857142, |
| "grad_norm": 5.383044719696045, |
| "learning_rate": 0.0, |
| "loss": 1.5793, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.14285714285714285, |
| "grad_norm": 3.96494197845459, |
| "learning_rate": 1.2500000000000002e-07, |
| "loss": 1.62, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.21428571428571427, |
| "grad_norm": 4.022143363952637, |
| "learning_rate": 2.5000000000000004e-07, |
| "loss": 1.6035, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.2857142857142857, |
| "grad_norm": 6.808481216430664, |
| "learning_rate": 3.75e-07, |
| "loss": 1.6309, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.35714285714285715, |
| "grad_norm": 4.0089311599731445, |
| "learning_rate": 5.000000000000001e-07, |
| "loss": 1.596, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.42857142857142855, |
| "grad_norm": 6.727263450622559, |
| "learning_rate": 6.25e-07, |
| "loss": 1.6312, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 7.648512840270996, |
| "learning_rate": 7.5e-07, |
| "loss": 1.6206, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.5714285714285714, |
| "grad_norm": 11.115941047668457, |
| "learning_rate": 8.750000000000001e-07, |
| "loss": 1.6133, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.6428571428571429, |
| "grad_norm": 8.82610034942627, |
| "learning_rate": 1.0000000000000002e-06, |
| "loss": 1.6419, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.7142857142857143, |
| "grad_norm": 3.3373453617095947, |
| "learning_rate": 1.125e-06, |
| "loss": 1.5894, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.7857142857142857, |
| "grad_norm": 7.521600723266602, |
| "learning_rate": 1.25e-06, |
| "loss": 1.6188, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.8571428571428571, |
| "grad_norm": 5.497554302215576, |
| "learning_rate": 1.3750000000000002e-06, |
| "loss": 1.6221, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.9285714285714286, |
| "grad_norm": 3.5530192852020264, |
| "learning_rate": 1.5e-06, |
| "loss": 1.6112, |
| "step": 13 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 4.452667713165283, |
| "learning_rate": 1.6250000000000001e-06, |
| "loss": 1.5999, |
| "step": 14 |
| }, |
| { |
| "epoch": 1.0714285714285714, |
| "grad_norm": 2.4583044052124023, |
| "learning_rate": 1.7500000000000002e-06, |
| "loss": 1.5917, |
| "step": 15 |
| }, |
| { |
| "epoch": 1.1428571428571428, |
| "grad_norm": 9.272625923156738, |
| "learning_rate": 1.875e-06, |
| "loss": 1.6092, |
| "step": 16 |
| }, |
| { |
| "epoch": 1.2142857142857142, |
| "grad_norm": 4.451343059539795, |
| "learning_rate": 2.0000000000000003e-06, |
| "loss": 1.5686, |
| "step": 17 |
| }, |
| { |
| "epoch": 1.2857142857142856, |
| "grad_norm": 5.756788730621338, |
| "learning_rate": 2.1250000000000004e-06, |
| "loss": 1.6265, |
| "step": 18 |
| }, |
| { |
| "epoch": 1.3571428571428572, |
| "grad_norm": 4.12752103805542, |
| "learning_rate": 2.25e-06, |
| "loss": 1.6073, |
| "step": 19 |
| }, |
| { |
| "epoch": 1.4285714285714286, |
| "grad_norm": 2.2131590843200684, |
| "learning_rate": 2.375e-06, |
| "loss": 1.5929, |
| "step": 20 |
| }, |
| { |
| "epoch": 1.5, |
| "grad_norm": 7.726499080657959, |
| "learning_rate": 2.5e-06, |
| "loss": 1.6232, |
| "step": 21 |
| }, |
| { |
| "epoch": 1.5714285714285714, |
| "grad_norm": 5.906811714172363, |
| "learning_rate": 2.625e-06, |
| "loss": 1.5777, |
| "step": 22 |
| }, |
| { |
| "epoch": 1.6428571428571428, |
| "grad_norm": 4.695427894592285, |
| "learning_rate": 2.7500000000000004e-06, |
| "loss": 1.591, |
| "step": 23 |
| }, |
| { |
| "epoch": 1.7142857142857144, |
| "grad_norm": 1.5481104850769043, |
| "learning_rate": 2.8750000000000004e-06, |
| "loss": 1.5743, |
| "step": 24 |
| }, |
| { |
| "epoch": 1.7857142857142856, |
| "grad_norm": 4.624354839324951, |
| "learning_rate": 3e-06, |
| "loss": 1.6065, |
| "step": 25 |
| }, |
| { |
| "epoch": 1.8571428571428572, |
| "grad_norm": 5.646456718444824, |
| "learning_rate": 3.125e-06, |
| "loss": 1.5975, |
| "step": 26 |
| }, |
| { |
| "epoch": 1.9285714285714286, |
| "grad_norm": 8.583253860473633, |
| "learning_rate": 3.2500000000000002e-06, |
| "loss": 1.5809, |
| "step": 27 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 1.7897872924804688, |
| "learning_rate": 3.3750000000000003e-06, |
| "loss": 1.5839, |
| "step": 28 |
| }, |
| { |
| "epoch": 2.0714285714285716, |
| "grad_norm": 2.509343385696411, |
| "learning_rate": 3.5000000000000004e-06, |
| "loss": 1.5725, |
| "step": 29 |
| }, |
| { |
| "epoch": 2.142857142857143, |
| "grad_norm": 3.86081600189209, |
| "learning_rate": 3.625e-06, |
| "loss": 1.5766, |
| "step": 30 |
| }, |
| { |
| "epoch": 2.2142857142857144, |
| "grad_norm": 4.29940128326416, |
| "learning_rate": 3.75e-06, |
| "loss": 1.5727, |
| "step": 31 |
| }, |
| { |
| "epoch": 2.2857142857142856, |
| "grad_norm": 3.209973096847534, |
| "learning_rate": 3.875e-06, |
| "loss": 1.5783, |
| "step": 32 |
| }, |
| { |
| "epoch": 2.357142857142857, |
| "grad_norm": 2.813647985458374, |
| "learning_rate": 4.000000000000001e-06, |
| "loss": 1.5687, |
| "step": 33 |
| }, |
| { |
| "epoch": 2.4285714285714284, |
| "grad_norm": 2.750006914138794, |
| "learning_rate": 4.125e-06, |
| "loss": 1.5739, |
| "step": 34 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 4.352213382720947, |
| "learning_rate": 4.250000000000001e-06, |
| "loss": 1.5781, |
| "step": 35 |
| }, |
| { |
| "epoch": 2.571428571428571, |
| "grad_norm": 2.743788957595825, |
| "learning_rate": 4.375e-06, |
| "loss": 1.5731, |
| "step": 36 |
| }, |
| { |
| "epoch": 2.642857142857143, |
| "grad_norm": 5.008453845977783, |
| "learning_rate": 4.5e-06, |
| "loss": 1.5828, |
| "step": 37 |
| }, |
| { |
| "epoch": 2.7142857142857144, |
| "grad_norm": 3.1128811836242676, |
| "learning_rate": 4.625e-06, |
| "loss": 1.5706, |
| "step": 38 |
| }, |
| { |
| "epoch": 2.7857142857142856, |
| "grad_norm": 5.385458946228027, |
| "learning_rate": 4.75e-06, |
| "loss": 1.5834, |
| "step": 39 |
| }, |
| { |
| "epoch": 2.857142857142857, |
| "grad_norm": 1.6079157590866089, |
| "learning_rate": 4.875000000000001e-06, |
| "loss": 1.5721, |
| "step": 40 |
| }, |
| { |
| "epoch": 2.928571428571429, |
| "grad_norm": 3.1016640663146973, |
| "learning_rate": 5e-06, |
| "loss": 1.5763, |
| "step": 41 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 4.650889873504639, |
| "learning_rate": 5.125e-06, |
| "loss": 1.5426, |
| "step": 42 |
| }, |
| { |
| "epoch": 3.0714285714285716, |
| "grad_norm": 1.5727084875106812, |
| "learning_rate": 5.25e-06, |
| "loss": 1.5432, |
| "step": 43 |
| }, |
| { |
| "epoch": 3.142857142857143, |
| "grad_norm": 2.637890577316284, |
| "learning_rate": 5.375e-06, |
| "loss": 1.5472, |
| "step": 44 |
| }, |
| { |
| "epoch": 3.2142857142857144, |
| "grad_norm": 1.8277631998062134, |
| "learning_rate": 5.500000000000001e-06, |
| "loss": 1.5363, |
| "step": 45 |
| }, |
| { |
| "epoch": 3.2857142857142856, |
| "grad_norm": 1.6959493160247803, |
| "learning_rate": 5.625e-06, |
| "loss": 1.5397, |
| "step": 46 |
| }, |
| { |
| "epoch": 3.357142857142857, |
| "grad_norm": 7.91141414642334, |
| "learning_rate": 5.750000000000001e-06, |
| "loss": 1.5392, |
| "step": 47 |
| }, |
| { |
| "epoch": 3.4285714285714284, |
| "grad_norm": 8.705608367919922, |
| "learning_rate": 5.875e-06, |
| "loss": 1.5499, |
| "step": 48 |
| }, |
| { |
| "epoch": 3.5, |
| "grad_norm": 1.4096094369888306, |
| "learning_rate": 6e-06, |
| "loss": 1.5554, |
| "step": 49 |
| }, |
| { |
| "epoch": 3.571428571428571, |
| "grad_norm": 1.2479711771011353, |
| "learning_rate": 6.125e-06, |
| "loss": 1.5537, |
| "step": 50 |
| }, |
| { |
| "epoch": 3.642857142857143, |
| "grad_norm": 1.7531940937042236, |
| "learning_rate": 6.25e-06, |
| "loss": 1.533, |
| "step": 51 |
| }, |
| { |
| "epoch": 3.7142857142857144, |
| "grad_norm": 3.1394143104553223, |
| "learning_rate": 6.375000000000001e-06, |
| "loss": 1.5223, |
| "step": 52 |
| }, |
| { |
| "epoch": 3.7857142857142856, |
| "grad_norm": 5.379184246063232, |
| "learning_rate": 6.5000000000000004e-06, |
| "loss": 1.5204, |
| "step": 53 |
| }, |
| { |
| "epoch": 3.857142857142857, |
| "grad_norm": 1.6642723083496094, |
| "learning_rate": 6.625000000000001e-06, |
| "loss": 1.5195, |
| "step": 54 |
| }, |
| { |
| "epoch": 3.928571428571429, |
| "grad_norm": 1.8299416303634644, |
| "learning_rate": 6.750000000000001e-06, |
| "loss": 1.5111, |
| "step": 55 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 3.942272901535034, |
| "learning_rate": 6.875000000000001e-06, |
| "loss": 1.5134, |
| "step": 56 |
| }, |
| { |
| "epoch": 4.071428571428571, |
| "grad_norm": 1.4287347793579102, |
| "learning_rate": 7.000000000000001e-06, |
| "loss": 1.5119, |
| "step": 57 |
| }, |
| { |
| "epoch": 4.142857142857143, |
| "grad_norm": 2.0702247619628906, |
| "learning_rate": 7.1249999999999995e-06, |
| "loss": 1.4962, |
| "step": 58 |
| }, |
| { |
| "epoch": 4.214285714285714, |
| "grad_norm": 2.863966703414917, |
| "learning_rate": 7.25e-06, |
| "loss": 1.5204, |
| "step": 59 |
| }, |
| { |
| "epoch": 4.285714285714286, |
| "grad_norm": 2.398355722427368, |
| "learning_rate": 7.375e-06, |
| "loss": 1.4955, |
| "step": 60 |
| }, |
| { |
| "epoch": 4.357142857142857, |
| "grad_norm": 1.4015003442764282, |
| "learning_rate": 7.5e-06, |
| "loss": 1.4942, |
| "step": 61 |
| }, |
| { |
| "epoch": 4.428571428571429, |
| "grad_norm": 1.6098459959030151, |
| "learning_rate": 7.625e-06, |
| "loss": 1.4712, |
| "step": 62 |
| }, |
| { |
| "epoch": 4.5, |
| "grad_norm": 4.722957611083984, |
| "learning_rate": 7.75e-06, |
| "loss": 1.5288, |
| "step": 63 |
| }, |
| { |
| "epoch": 4.571428571428571, |
| "grad_norm": 3.6340856552124023, |
| "learning_rate": 7.875e-06, |
| "loss": 1.524, |
| "step": 64 |
| }, |
| { |
| "epoch": 4.642857142857143, |
| "grad_norm": 2.6402580738067627, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 1.5417, |
| "step": 65 |
| }, |
| { |
| "epoch": 4.714285714285714, |
| "grad_norm": 2.7053163051605225, |
| "learning_rate": 8.125000000000001e-06, |
| "loss": 1.5101, |
| "step": 66 |
| }, |
| { |
| "epoch": 4.785714285714286, |
| "grad_norm": 4.306228160858154, |
| "learning_rate": 8.25e-06, |
| "loss": 1.5474, |
| "step": 67 |
| }, |
| { |
| "epoch": 4.857142857142857, |
| "grad_norm": 3.2443113327026367, |
| "learning_rate": 8.375e-06, |
| "loss": 1.5324, |
| "step": 68 |
| }, |
| { |
| "epoch": 4.928571428571429, |
| "grad_norm": 2.3825535774230957, |
| "learning_rate": 8.500000000000002e-06, |
| "loss": 1.522, |
| "step": 69 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 3.2580692768096924, |
| "learning_rate": 8.625e-06, |
| "loss": 1.5179, |
| "step": 70 |
| }, |
| { |
| "epoch": 5.071428571428571, |
| "grad_norm": 1.4018006324768066, |
| "learning_rate": 8.75e-06, |
| "loss": 1.4971, |
| "step": 71 |
| }, |
| { |
| "epoch": 5.142857142857143, |
| "grad_norm": 5.274407386779785, |
| "learning_rate": 8.875e-06, |
| "loss": 1.513, |
| "step": 72 |
| }, |
| { |
| "epoch": 5.214285714285714, |
| "grad_norm": 1.3473955392837524, |
| "learning_rate": 9e-06, |
| "loss": 1.4898, |
| "step": 73 |
| }, |
| { |
| "epoch": 5.285714285714286, |
| "grad_norm": 5.1161346435546875, |
| "learning_rate": 9.125e-06, |
| "loss": 1.5032, |
| "step": 74 |
| }, |
| { |
| "epoch": 5.357142857142857, |
| "grad_norm": 10.268465042114258, |
| "learning_rate": 9.25e-06, |
| "loss": 1.4926, |
| "step": 75 |
| }, |
| { |
| "epoch": 5.428571428571429, |
| "grad_norm": 1.676137089729309, |
| "learning_rate": 9.375000000000001e-06, |
| "loss": 1.4833, |
| "step": 76 |
| }, |
| { |
| "epoch": 5.5, |
| "grad_norm": 1.441231369972229, |
| "learning_rate": 9.5e-06, |
| "loss": 1.4934, |
| "step": 77 |
| }, |
| { |
| "epoch": 5.571428571428571, |
| "grad_norm": 3.917909860610962, |
| "learning_rate": 9.625e-06, |
| "loss": 1.5473, |
| "step": 78 |
| }, |
| { |
| "epoch": 5.642857142857143, |
| "grad_norm": 1.9755581617355347, |
| "learning_rate": 9.750000000000002e-06, |
| "loss": 1.5192, |
| "step": 79 |
| }, |
| { |
| "epoch": 5.714285714285714, |
| "grad_norm": 2.587038516998291, |
| "learning_rate": 9.875000000000001e-06, |
| "loss": 1.5212, |
| "step": 80 |
| }, |
| { |
| "epoch": 5.785714285714286, |
| "grad_norm": 2.155751943588257, |
| "learning_rate": 1e-05, |
| "loss": 1.5337, |
| "step": 81 |
| }, |
| { |
| "epoch": 5.857142857142857, |
| "grad_norm": 2.8446056842803955, |
| "learning_rate": 1.0125e-05, |
| "loss": 1.5174, |
| "step": 82 |
| }, |
| { |
| "epoch": 5.928571428571429, |
| "grad_norm": 2.9765541553497314, |
| "learning_rate": 1.025e-05, |
| "loss": 1.5267, |
| "step": 83 |
| }, |
| { |
| "epoch": 6.0, |
| "grad_norm": 2.4112436771392822, |
| "learning_rate": 1.0375e-05, |
| "loss": 1.4897, |
| "step": 84 |
| }, |
| { |
| "epoch": 6.071428571428571, |
| "grad_norm": 3.6984405517578125, |
| "learning_rate": 1.05e-05, |
| "loss": 1.5304, |
| "step": 85 |
| }, |
| { |
| "epoch": 6.142857142857143, |
| "grad_norm": 3.8474667072296143, |
| "learning_rate": 1.0625e-05, |
| "loss": 1.5232, |
| "step": 86 |
| }, |
| { |
| "epoch": 6.214285714285714, |
| "grad_norm": 2.087263584136963, |
| "learning_rate": 1.075e-05, |
| "loss": 1.5285, |
| "step": 87 |
| }, |
| { |
| "epoch": 6.285714285714286, |
| "grad_norm": 2.1326711177825928, |
| "learning_rate": 1.0875e-05, |
| "loss": 1.5387, |
| "step": 88 |
| }, |
| { |
| "epoch": 6.357142857142857, |
| "grad_norm": 4.559895992279053, |
| "learning_rate": 1.1000000000000001e-05, |
| "loss": 1.524, |
| "step": 89 |
| }, |
| { |
| "epoch": 6.428571428571429, |
| "grad_norm": 1.8320558071136475, |
| "learning_rate": 1.1125000000000001e-05, |
| "loss": 1.5143, |
| "step": 90 |
| }, |
| { |
| "epoch": 6.5, |
| "grad_norm": 1.8108539581298828, |
| "learning_rate": 1.125e-05, |
| "loss": 1.4987, |
| "step": 91 |
| }, |
| { |
| "epoch": 6.571428571428571, |
| "grad_norm": 1.3095402717590332, |
| "learning_rate": 1.1375e-05, |
| "loss": 1.5025, |
| "step": 92 |
| }, |
| { |
| "epoch": 6.642857142857143, |
| "grad_norm": 4.299781799316406, |
| "learning_rate": 1.1500000000000002e-05, |
| "loss": 1.5251, |
| "step": 93 |
| }, |
| { |
| "epoch": 6.714285714285714, |
| "grad_norm": 1.1237058639526367, |
| "learning_rate": 1.1625000000000001e-05, |
| "loss": 1.5185, |
| "step": 94 |
| }, |
| { |
| "epoch": 6.785714285714286, |
| "grad_norm": 6.838643550872803, |
| "learning_rate": 1.175e-05, |
| "loss": 1.5032, |
| "step": 95 |
| }, |
| { |
| "epoch": 6.857142857142857, |
| "grad_norm": 1.886619210243225, |
| "learning_rate": 1.1875e-05, |
| "loss": 1.5394, |
| "step": 96 |
| }, |
| { |
| "epoch": 6.928571428571429, |
| "grad_norm": 2.4293034076690674, |
| "learning_rate": 1.2e-05, |
| "loss": 1.4958, |
| "step": 97 |
| }, |
| { |
| "epoch": 7.0, |
| "grad_norm": 1.3641681671142578, |
| "learning_rate": 1.2125e-05, |
| "loss": 1.5182, |
| "step": 98 |
| }, |
| { |
| "epoch": 7.071428571428571, |
| "grad_norm": 1.6436591148376465, |
| "learning_rate": 1.225e-05, |
| "loss": 1.5007, |
| "step": 99 |
| }, |
| { |
| "epoch": 7.142857142857143, |
| "grad_norm": 1.4990618228912354, |
| "learning_rate": 1.2375000000000001e-05, |
| "loss": 1.4881, |
| "step": 100 |
| }, |
| { |
| "epoch": 7.214285714285714, |
| "grad_norm": 1.2271665334701538, |
| "learning_rate": 1.25e-05, |
| "loss": 1.5017, |
| "step": 101 |
| }, |
| { |
| "epoch": 7.285714285714286, |
| "grad_norm": 1.2291427850723267, |
| "learning_rate": 1.2625e-05, |
| "loss": 1.4979, |
| "step": 102 |
| }, |
| { |
| "epoch": 7.357142857142857, |
| "grad_norm": 5.997069358825684, |
| "learning_rate": 1.2750000000000002e-05, |
| "loss": 1.5238, |
| "step": 103 |
| }, |
| { |
| "epoch": 7.428571428571429, |
| "grad_norm": 1.4330472946166992, |
| "learning_rate": 1.2875000000000001e-05, |
| "loss": 1.5133, |
| "step": 104 |
| }, |
| { |
| "epoch": 7.5, |
| "grad_norm": 1.3329960107803345, |
| "learning_rate": 1.3000000000000001e-05, |
| "loss": 1.5044, |
| "step": 105 |
| }, |
| { |
| "epoch": 7.571428571428571, |
| "grad_norm": 2.7081637382507324, |
| "learning_rate": 1.3125e-05, |
| "loss": 1.4825, |
| "step": 106 |
| }, |
| { |
| "epoch": 7.642857142857143, |
| "grad_norm": 2.8797085285186768, |
| "learning_rate": 1.3250000000000002e-05, |
| "loss": 1.4711, |
| "step": 107 |
| }, |
| { |
| "epoch": 7.714285714285714, |
| "grad_norm": 4.054922580718994, |
| "learning_rate": 1.3375000000000002e-05, |
| "loss": 1.4982, |
| "step": 108 |
| }, |
| { |
| "epoch": 7.785714285714286, |
| "grad_norm": 7.4362311363220215, |
| "learning_rate": 1.3500000000000001e-05, |
| "loss": 1.4786, |
| "step": 109 |
| }, |
| { |
| "epoch": 7.857142857142857, |
| "grad_norm": 2.5698859691619873, |
| "learning_rate": 1.3625e-05, |
| "loss": 1.4572, |
| "step": 110 |
| }, |
| { |
| "epoch": 7.928571428571429, |
| "grad_norm": 2.9194092750549316, |
| "learning_rate": 1.3750000000000002e-05, |
| "loss": 1.4916, |
| "step": 111 |
| }, |
| { |
| "epoch": 8.0, |
| "grad_norm": 2.044605016708374, |
| "learning_rate": 1.3875000000000002e-05, |
| "loss": 1.4879, |
| "step": 112 |
| }, |
| { |
| "epoch": 8.071428571428571, |
| "grad_norm": 1.3414772748947144, |
| "learning_rate": 1.4000000000000001e-05, |
| "loss": 1.4887, |
| "step": 113 |
| }, |
| { |
| "epoch": 8.142857142857142, |
| "grad_norm": 2.0897555351257324, |
| "learning_rate": 1.4125e-05, |
| "loss": 1.4908, |
| "step": 114 |
| }, |
| { |
| "epoch": 8.214285714285714, |
| "grad_norm": 3.721792697906494, |
| "learning_rate": 1.4249999999999999e-05, |
| "loss": 1.4978, |
| "step": 115 |
| }, |
| { |
| "epoch": 8.285714285714286, |
| "grad_norm": 1.3067519664764404, |
| "learning_rate": 1.4374999999999999e-05, |
| "loss": 1.4903, |
| "step": 116 |
| }, |
| { |
| "epoch": 8.357142857142858, |
| "grad_norm": 2.3188929557800293, |
| "learning_rate": 1.45e-05, |
| "loss": 1.475, |
| "step": 117 |
| }, |
| { |
| "epoch": 8.428571428571429, |
| "grad_norm": 2.131817102432251, |
| "learning_rate": 1.4625e-05, |
| "loss": 1.4877, |
| "step": 118 |
| }, |
| { |
| "epoch": 8.5, |
| "grad_norm": 5.011297225952148, |
| "learning_rate": 1.475e-05, |
| "loss": 1.4651, |
| "step": 119 |
| }, |
| { |
| "epoch": 8.571428571428571, |
| "grad_norm": 2.497302770614624, |
| "learning_rate": 1.4875e-05, |
| "loss": 1.4488, |
| "step": 120 |
| }, |
| { |
| "epoch": 8.642857142857142, |
| "grad_norm": 1.3582658767700195, |
| "learning_rate": 1.5e-05, |
| "loss": 1.4446, |
| "step": 121 |
| }, |
| { |
| "epoch": 8.714285714285714, |
| "grad_norm": 5.030256748199463, |
| "learning_rate": 1.5125e-05, |
| "loss": 1.4445, |
| "step": 122 |
| }, |
| { |
| "epoch": 8.785714285714286, |
| "grad_norm": 6.480685234069824, |
| "learning_rate": 1.525e-05, |
| "loss": 1.4853, |
| "step": 123 |
| }, |
| { |
| "epoch": 8.857142857142858, |
| "grad_norm": 9.358418464660645, |
| "learning_rate": 1.5375e-05, |
| "loss": 1.5026, |
| "step": 124 |
| }, |
| { |
| "epoch": 8.928571428571429, |
| "grad_norm": 8.761038780212402, |
| "learning_rate": 1.55e-05, |
| "loss": 1.4884, |
| "step": 125 |
| }, |
| { |
| "epoch": 9.0, |
| "grad_norm": 1.6179229021072388, |
| "learning_rate": 1.5625e-05, |
| "loss": 1.4705, |
| "step": 126 |
| }, |
| { |
| "epoch": 9.071428571428571, |
| "grad_norm": 2.0552783012390137, |
| "learning_rate": 1.575e-05, |
| "loss": 1.4869, |
| "step": 127 |
| }, |
| { |
| "epoch": 9.142857142857142, |
| "grad_norm": 6.032142162322998, |
| "learning_rate": 1.5875e-05, |
| "loss": 1.4461, |
| "step": 128 |
| }, |
| { |
| "epoch": 9.214285714285714, |
| "grad_norm": 2.0145914554595947, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 1.4625, |
| "step": 129 |
| }, |
| { |
| "epoch": 9.285714285714286, |
| "grad_norm": 5.505954742431641, |
| "learning_rate": 1.6125000000000002e-05, |
| "loss": 1.4764, |
| "step": 130 |
| }, |
| { |
| "epoch": 9.357142857142858, |
| "grad_norm": 4.161317348480225, |
| "learning_rate": 1.6250000000000002e-05, |
| "loss": 1.4558, |
| "step": 131 |
| }, |
| { |
| "epoch": 9.428571428571429, |
| "grad_norm": 6.453343391418457, |
| "learning_rate": 1.6375e-05, |
| "loss": 1.4914, |
| "step": 132 |
| }, |
| { |
| "epoch": 9.5, |
| "grad_norm": 2.7060706615448, |
| "learning_rate": 1.65e-05, |
| "loss": 1.4479, |
| "step": 133 |
| }, |
| { |
| "epoch": 9.571428571428571, |
| "grad_norm": 6.37178897857666, |
| "learning_rate": 1.6625e-05, |
| "loss": 1.4951, |
| "step": 134 |
| }, |
| { |
| "epoch": 9.642857142857142, |
| "grad_norm": 1.644932746887207, |
| "learning_rate": 1.675e-05, |
| "loss": 1.4492, |
| "step": 135 |
| }, |
| { |
| "epoch": 9.714285714285714, |
| "grad_norm": 4.173709392547607, |
| "learning_rate": 1.6875000000000004e-05, |
| "loss": 1.4419, |
| "step": 136 |
| }, |
| { |
| "epoch": 9.785714285714286, |
| "grad_norm": 2.661614179611206, |
| "learning_rate": 1.7000000000000003e-05, |
| "loss": 1.4419, |
| "step": 137 |
| }, |
| { |
| "epoch": 9.857142857142858, |
| "grad_norm": 2.468867778778076, |
| "learning_rate": 1.7125000000000003e-05, |
| "loss": 1.416, |
| "step": 138 |
| }, |
| { |
| "epoch": 9.928571428571429, |
| "grad_norm": 1.0798838138580322, |
| "learning_rate": 1.725e-05, |
| "loss": 1.4473, |
| "step": 139 |
| }, |
| { |
| "epoch": 10.0, |
| "grad_norm": 1.5816832780838013, |
| "learning_rate": 1.7375e-05, |
| "loss": 1.4399, |
| "step": 140 |
| }, |
| { |
| "epoch": 10.071428571428571, |
| "grad_norm": 2.063046932220459, |
| "learning_rate": 1.75e-05, |
| "loss": 1.4415, |
| "step": 141 |
| }, |
| { |
| "epoch": 10.142857142857142, |
| "grad_norm": 3.9278879165649414, |
| "learning_rate": 1.7625e-05, |
| "loss": 1.4463, |
| "step": 142 |
| }, |
| { |
| "epoch": 10.214285714285714, |
| "grad_norm": 1.3775101900100708, |
| "learning_rate": 1.775e-05, |
| "loss": 1.4019, |
| "step": 143 |
| }, |
| { |
| "epoch": 10.285714285714286, |
| "grad_norm": 6.007996082305908, |
| "learning_rate": 1.7875e-05, |
| "loss": 1.4282, |
| "step": 144 |
| }, |
| { |
| "epoch": 10.357142857142858, |
| "grad_norm": 4.488779067993164, |
| "learning_rate": 1.8e-05, |
| "loss": 1.451, |
| "step": 145 |
| }, |
| { |
| "epoch": 10.428571428571429, |
| "grad_norm": 5.015976905822754, |
| "learning_rate": 1.8125e-05, |
| "loss": 1.4089, |
| "step": 146 |
| }, |
| { |
| "epoch": 10.5, |
| "grad_norm": 1.4387730360031128, |
| "learning_rate": 1.825e-05, |
| "loss": 1.4249, |
| "step": 147 |
| }, |
| { |
| "epoch": 10.571428571428571, |
| "grad_norm": 2.4665675163269043, |
| "learning_rate": 1.8375e-05, |
| "loss": 1.4327, |
| "step": 148 |
| }, |
| { |
| "epoch": 10.642857142857142, |
| "grad_norm": 1.8020267486572266, |
| "learning_rate": 1.85e-05, |
| "loss": 1.3995, |
| "step": 149 |
| }, |
| { |
| "epoch": 10.714285714285714, |
| "grad_norm": 1.6808390617370605, |
| "learning_rate": 1.8625000000000002e-05, |
| "loss": 1.389, |
| "step": 150 |
| }, |
| { |
| "epoch": 10.785714285714286, |
| "grad_norm": 2.3376588821411133, |
| "learning_rate": 1.8750000000000002e-05, |
| "loss": 1.4748, |
| "step": 151 |
| }, |
| { |
| "epoch": 10.857142857142858, |
| "grad_norm": 2.5257272720336914, |
| "learning_rate": 1.8875e-05, |
| "loss": 1.4221, |
| "step": 152 |
| }, |
| { |
| "epoch": 10.928571428571429, |
| "grad_norm": 1.6355801820755005, |
| "learning_rate": 1.9e-05, |
| "loss": 1.3544, |
| "step": 153 |
| }, |
| { |
| "epoch": 11.0, |
| "grad_norm": 2.1227149963378906, |
| "learning_rate": 1.9125e-05, |
| "loss": 1.3493, |
| "step": 154 |
| }, |
| { |
| "epoch": 11.071428571428571, |
| "grad_norm": 4.7650370597839355, |
| "learning_rate": 1.925e-05, |
| "loss": 1.4708, |
| "step": 155 |
| }, |
| { |
| "epoch": 11.142857142857142, |
| "grad_norm": 2.83673095703125, |
| "learning_rate": 1.9375e-05, |
| "loss": 1.4064, |
| "step": 156 |
| }, |
| { |
| "epoch": 11.214285714285714, |
| "grad_norm": 2.4201812744140625, |
| "learning_rate": 1.9500000000000003e-05, |
| "loss": 1.3221, |
| "step": 157 |
| }, |
| { |
| "epoch": 11.285714285714286, |
| "grad_norm": 2.0276103019714355, |
| "learning_rate": 1.9625000000000003e-05, |
| "loss": 1.2859, |
| "step": 158 |
| }, |
| { |
| "epoch": 11.357142857142858, |
| "grad_norm": 7.189219951629639, |
| "learning_rate": 1.9750000000000002e-05, |
| "loss": 1.3526, |
| "step": 159 |
| }, |
| { |
| "epoch": 11.428571428571429, |
| "grad_norm": 2.9911937713623047, |
| "learning_rate": 1.9875000000000002e-05, |
| "loss": 1.319, |
| "step": 160 |
| }, |
| { |
| "epoch": 11.5, |
| "grad_norm": 7.41688346862793, |
| "learning_rate": 2e-05, |
| "loss": 1.2701, |
| "step": 161 |
| }, |
| { |
| "epoch": 11.571428571428571, |
| "grad_norm": 2.5221216678619385, |
| "learning_rate": 2.0125e-05, |
| "loss": 1.2342, |
| "step": 162 |
| }, |
| { |
| "epoch": 11.642857142857142, |
| "grad_norm": 3.1596360206604004, |
| "learning_rate": 2.025e-05, |
| "loss": 1.2818, |
| "step": 163 |
| }, |
| { |
| "epoch": 11.714285714285714, |
| "grad_norm": 4.858260631561279, |
| "learning_rate": 2.0375e-05, |
| "loss": 1.263, |
| "step": 164 |
| }, |
| { |
| "epoch": 11.785714285714286, |
| "grad_norm": 5.346636772155762, |
| "learning_rate": 2.05e-05, |
| "loss": 1.2936, |
| "step": 165 |
| }, |
| { |
| "epoch": 11.857142857142858, |
| "grad_norm": 4.936079502105713, |
| "learning_rate": 2.0625e-05, |
| "loss": 1.2727, |
| "step": 166 |
| }, |
| { |
| "epoch": 11.928571428571429, |
| "grad_norm": 10.273465156555176, |
| "learning_rate": 2.075e-05, |
| "loss": 1.2605, |
| "step": 167 |
| }, |
| { |
| "epoch": 12.0, |
| "grad_norm": 4.918086051940918, |
| "learning_rate": 2.0875e-05, |
| "loss": 1.2611, |
| "step": 168 |
| }, |
| { |
| "epoch": 12.071428571428571, |
| "grad_norm": 12.318052291870117, |
| "learning_rate": 2.1e-05, |
| "loss": 1.2175, |
| "step": 169 |
| }, |
| { |
| "epoch": 12.142857142857142, |
| "grad_norm": 4.972585201263428, |
| "learning_rate": 2.1125000000000002e-05, |
| "loss": 1.1154, |
| "step": 170 |
| }, |
| { |
| "epoch": 12.214285714285714, |
| "grad_norm": 4.9560441970825195, |
| "learning_rate": 2.125e-05, |
| "loss": 1.1845, |
| "step": 171 |
| }, |
| { |
| "epoch": 12.285714285714286, |
| "grad_norm": 5.026245594024658, |
| "learning_rate": 2.1375e-05, |
| "loss": 1.1234, |
| "step": 172 |
| }, |
| { |
| "epoch": 12.357142857142858, |
| "grad_norm": 7.455307483673096, |
| "learning_rate": 2.15e-05, |
| "loss": 1.1865, |
| "step": 173 |
| }, |
| { |
| "epoch": 12.428571428571429, |
| "grad_norm": 5.8239946365356445, |
| "learning_rate": 2.1625e-05, |
| "loss": 1.2061, |
| "step": 174 |
| }, |
| { |
| "epoch": 12.5, |
| "grad_norm": 7.225159168243408, |
| "learning_rate": 2.175e-05, |
| "loss": 1.2647, |
| "step": 175 |
| }, |
| { |
| "epoch": 12.571428571428571, |
| "grad_norm": 3.666829824447632, |
| "learning_rate": 2.1875e-05, |
| "loss": 1.1374, |
| "step": 176 |
| }, |
| { |
| "epoch": 12.642857142857142, |
| "grad_norm": 6.110956192016602, |
| "learning_rate": 2.2000000000000003e-05, |
| "loss": 1.1125, |
| "step": 177 |
| }, |
| { |
| "epoch": 12.714285714285714, |
| "grad_norm": 10.124720573425293, |
| "learning_rate": 2.2125000000000002e-05, |
| "loss": 1.1762, |
| "step": 178 |
| }, |
| { |
| "epoch": 12.785714285714286, |
| "grad_norm": 6.521277904510498, |
| "learning_rate": 2.2250000000000002e-05, |
| "loss": 1.2196, |
| "step": 179 |
| }, |
| { |
| "epoch": 12.857142857142858, |
| "grad_norm": 3.0751357078552246, |
| "learning_rate": 2.2375000000000002e-05, |
| "loss": 1.1462, |
| "step": 180 |
| }, |
| { |
| "epoch": 12.928571428571429, |
| "grad_norm": 4.803239345550537, |
| "learning_rate": 2.25e-05, |
| "loss": 1.0391, |
| "step": 181 |
| }, |
| { |
| "epoch": 13.0, |
| "grad_norm": 3.005739450454712, |
| "learning_rate": 2.2625e-05, |
| "loss": 1.0753, |
| "step": 182 |
| }, |
| { |
| "epoch": 13.071428571428571, |
| "grad_norm": 4.087983131408691, |
| "learning_rate": 2.275e-05, |
| "loss": 1.2163, |
| "step": 183 |
| }, |
| { |
| "epoch": 13.142857142857142, |
| "grad_norm": 3.475637674331665, |
| "learning_rate": 2.2875e-05, |
| "loss": 1.197, |
| "step": 184 |
| }, |
| { |
| "epoch": 13.214285714285714, |
| "grad_norm": 6.145781517028809, |
| "learning_rate": 2.3000000000000003e-05, |
| "loss": 1.1744, |
| "step": 185 |
| }, |
| { |
| "epoch": 13.285714285714286, |
| "grad_norm": 4.0928544998168945, |
| "learning_rate": 2.3125000000000003e-05, |
| "loss": 1.1189, |
| "step": 186 |
| }, |
| { |
| "epoch": 13.357142857142858, |
| "grad_norm": 9.976082801818848, |
| "learning_rate": 2.3250000000000003e-05, |
| "loss": 1.2488, |
| "step": 187 |
| }, |
| { |
| "epoch": 13.428571428571429, |
| "grad_norm": 6.045898914337158, |
| "learning_rate": 2.3375000000000002e-05, |
| "loss": 1.139, |
| "step": 188 |
| }, |
| { |
| "epoch": 13.5, |
| "grad_norm": 4.3854451179504395, |
| "learning_rate": 2.35e-05, |
| "loss": 1.0929, |
| "step": 189 |
| }, |
| { |
| "epoch": 13.571428571428571, |
| "grad_norm": 3.349039316177368, |
| "learning_rate": 2.3624999999999998e-05, |
| "loss": 1.0427, |
| "step": 190 |
| }, |
| { |
| "epoch": 13.642857142857142, |
| "grad_norm": 3.4710395336151123, |
| "learning_rate": 2.375e-05, |
| "loss": 1.1249, |
| "step": 191 |
| }, |
| { |
| "epoch": 13.714285714285714, |
| "grad_norm": 9.711404800415039, |
| "learning_rate": 2.3875e-05, |
| "loss": 1.2626, |
| "step": 192 |
| }, |
| { |
| "epoch": 13.785714285714286, |
| "grad_norm": 3.2990989685058594, |
| "learning_rate": 2.4e-05, |
| "loss": 1.1687, |
| "step": 193 |
| }, |
| { |
| "epoch": 13.857142857142858, |
| "grad_norm": 2.510340929031372, |
| "learning_rate": 2.4125e-05, |
| "loss": 1.0953, |
| "step": 194 |
| }, |
| { |
| "epoch": 13.928571428571429, |
| "grad_norm": 3.561298131942749, |
| "learning_rate": 2.425e-05, |
| "loss": 1.0411, |
| "step": 195 |
| }, |
| { |
| "epoch": 14.0, |
| "grad_norm": 4.719476699829102, |
| "learning_rate": 2.4375e-05, |
| "loss": 1.0803, |
| "step": 196 |
| }, |
| { |
| "epoch": 14.071428571428571, |
| "grad_norm": 3.3607053756713867, |
| "learning_rate": 2.45e-05, |
| "loss": 1.0709, |
| "step": 197 |
| }, |
| { |
| "epoch": 14.142857142857142, |
| "grad_norm": 4.803561687469482, |
| "learning_rate": 2.4625000000000002e-05, |
| "loss": 1.0335, |
| "step": 198 |
| }, |
| { |
| "epoch": 14.214285714285714, |
| "grad_norm": 3.8269684314727783, |
| "learning_rate": 2.4750000000000002e-05, |
| "loss": 1.0742, |
| "step": 199 |
| }, |
| { |
| "epoch": 14.285714285714286, |
| "grad_norm": 2.7520132064819336, |
| "learning_rate": 2.4875e-05, |
| "loss": 1.17, |
| "step": 200 |
| }, |
| { |
| "epoch": 14.357142857142858, |
| "grad_norm": 5.653459548950195, |
| "learning_rate": 2.5e-05, |
| "loss": 1.0918, |
| "step": 201 |
| }, |
| { |
| "epoch": 14.428571428571429, |
| "grad_norm": 2.774630546569824, |
| "learning_rate": 2.4999980961416097e-05, |
| "loss": 1.058, |
| "step": 202 |
| }, |
| { |
| "epoch": 14.5, |
| "grad_norm": 3.926753520965576, |
| "learning_rate": 2.499992384572238e-05, |
| "loss": 1.0942, |
| "step": 203 |
| }, |
| { |
| "epoch": 14.571428571428571, |
| "grad_norm": 2.9734742641448975, |
| "learning_rate": 2.4999828653092835e-05, |
| "loss": 1.007, |
| "step": 204 |
| }, |
| { |
| "epoch": 14.642857142857142, |
| "grad_norm": 2.9620471000671387, |
| "learning_rate": 2.4999695383817435e-05, |
| "loss": 1.0881, |
| "step": 205 |
| }, |
| { |
| "epoch": 14.714285714285714, |
| "grad_norm": 5.124941349029541, |
| "learning_rate": 2.499952403830214e-05, |
| "loss": 1.0806, |
| "step": 206 |
| }, |
| { |
| "epoch": 14.785714285714286, |
| "grad_norm": 3.0979349613189697, |
| "learning_rate": 2.4999314617068904e-05, |
| "loss": 1.0281, |
| "step": 207 |
| }, |
| { |
| "epoch": 14.857142857142858, |
| "grad_norm": 3.4293224811553955, |
| "learning_rate": 2.4999067120755652e-05, |
| "loss": 1.051, |
| "step": 208 |
| }, |
| { |
| "epoch": 14.928571428571429, |
| "grad_norm": 6.487819194793701, |
| "learning_rate": 2.4998781550116305e-05, |
| "loss": 1.0308, |
| "step": 209 |
| }, |
| { |
| "epoch": 15.0, |
| "grad_norm": 3.3049116134643555, |
| "learning_rate": 2.499845790602076e-05, |
| "loss": 0.9572, |
| "step": 210 |
| }, |
| { |
| "epoch": 15.071428571428571, |
| "grad_norm": 4.562008380889893, |
| "learning_rate": 2.4998096189454893e-05, |
| "loss": 1.0758, |
| "step": 211 |
| }, |
| { |
| "epoch": 15.142857142857142, |
| "grad_norm": 4.130517482757568, |
| "learning_rate": 2.4997696401520555e-05, |
| "loss": 1.0565, |
| "step": 212 |
| }, |
| { |
| "epoch": 15.214285714285714, |
| "grad_norm": 7.463335037231445, |
| "learning_rate": 2.499725854343557e-05, |
| "loss": 0.9738, |
| "step": 213 |
| }, |
| { |
| "epoch": 15.285714285714286, |
| "grad_norm": 5.149586200714111, |
| "learning_rate": 2.4996782616533732e-05, |
| "loss": 1.0234, |
| "step": 214 |
| }, |
| { |
| "epoch": 15.357142857142858, |
| "grad_norm": 5.69240140914917, |
| "learning_rate": 2.499626862226479e-05, |
| "loss": 1.1353, |
| "step": 215 |
| }, |
| { |
| "epoch": 15.428571428571429, |
| "grad_norm": 4.097994327545166, |
| "learning_rate": 2.4995716562194465e-05, |
| "loss": 1.0006, |
| "step": 216 |
| }, |
| { |
| "epoch": 15.5, |
| "grad_norm": 3.709420919418335, |
| "learning_rate": 2.499512643800443e-05, |
| "loss": 0.9626, |
| "step": 217 |
| }, |
| { |
| "epoch": 15.571428571428571, |
| "grad_norm": 6.93643856048584, |
| "learning_rate": 2.4994498251492302e-05, |
| "loss": 0.9727, |
| "step": 218 |
| }, |
| { |
| "epoch": 15.642857142857142, |
| "grad_norm": 2.359558582305908, |
| "learning_rate": 2.4993832004571646e-05, |
| "loss": 1.0169, |
| "step": 219 |
| }, |
| { |
| "epoch": 15.714285714285714, |
| "grad_norm": 3.7421205043792725, |
| "learning_rate": 2.4993127699271966e-05, |
| "loss": 1.0996, |
| "step": 220 |
| }, |
| { |
| "epoch": 15.785714285714286, |
| "grad_norm": 3.670799493789673, |
| "learning_rate": 2.49923853377387e-05, |
| "loss": 1.0078, |
| "step": 221 |
| }, |
| { |
| "epoch": 15.857142857142858, |
| "grad_norm": 3.7862744331359863, |
| "learning_rate": 2.4991604922233204e-05, |
| "loss": 0.9785, |
| "step": 222 |
| }, |
| { |
| "epoch": 15.928571428571429, |
| "grad_norm": 2.4504923820495605, |
| "learning_rate": 2.4990786455132764e-05, |
| "loss": 0.9815, |
| "step": 223 |
| }, |
| { |
| "epoch": 16.0, |
| "grad_norm": 4.063878536224365, |
| "learning_rate": 2.4989929938930576e-05, |
| "loss": 0.9463, |
| "step": 224 |
| }, |
| { |
| "epoch": 16.071428571428573, |
| "grad_norm": 2.9702260494232178, |
| "learning_rate": 2.498903537623573e-05, |
| "loss": 0.9618, |
| "step": 225 |
| }, |
| { |
| "epoch": 16.142857142857142, |
| "grad_norm": 2.672618865966797, |
| "learning_rate": 2.4988102769773227e-05, |
| "loss": 0.9338, |
| "step": 226 |
| }, |
| { |
| "epoch": 16.214285714285715, |
| "grad_norm": 6.381713390350342, |
| "learning_rate": 2.4987132122383936e-05, |
| "loss": 1.0038, |
| "step": 227 |
| }, |
| { |
| "epoch": 16.285714285714285, |
| "grad_norm": 2.369471549987793, |
| "learning_rate": 2.4986123437024627e-05, |
| "loss": 1.0239, |
| "step": 228 |
| }, |
| { |
| "epoch": 16.357142857142858, |
| "grad_norm": 2.5277535915374756, |
| "learning_rate": 2.4985076716767927e-05, |
| "loss": 0.9014, |
| "step": 229 |
| }, |
| { |
| "epoch": 16.428571428571427, |
| "grad_norm": 3.924433946609497, |
| "learning_rate": 2.4983991964802327e-05, |
| "loss": 0.966, |
| "step": 230 |
| }, |
| { |
| "epoch": 16.5, |
| "grad_norm": 3.6045427322387695, |
| "learning_rate": 2.4982869184432174e-05, |
| "loss": 0.9695, |
| "step": 231 |
| }, |
| { |
| "epoch": 16.571428571428573, |
| "grad_norm": 6.360043048858643, |
| "learning_rate": 2.498170837907765e-05, |
| "loss": 1.0349, |
| "step": 232 |
| }, |
| { |
| "epoch": 16.642857142857142, |
| "grad_norm": 4.736248970031738, |
| "learning_rate": 2.4980509552274765e-05, |
| "loss": 0.992, |
| "step": 233 |
| }, |
| { |
| "epoch": 16.714285714285715, |
| "grad_norm": 2.659874677658081, |
| "learning_rate": 2.4979272707675356e-05, |
| "loss": 0.9548, |
| "step": 234 |
| }, |
| { |
| "epoch": 16.785714285714285, |
| "grad_norm": 2.803654909133911, |
| "learning_rate": 2.497799784904707e-05, |
| "loss": 0.8971, |
| "step": 235 |
| }, |
| { |
| "epoch": 16.857142857142858, |
| "grad_norm": 4.661594867706299, |
| "learning_rate": 2.4976684980273338e-05, |
| "loss": 0.9161, |
| "step": 236 |
| }, |
| { |
| "epoch": 16.928571428571427, |
| "grad_norm": 3.7907001972198486, |
| "learning_rate": 2.4975334105353396e-05, |
| "loss": 0.8663, |
| "step": 237 |
| }, |
| { |
| "epoch": 17.0, |
| "grad_norm": 6.289381504058838, |
| "learning_rate": 2.497394522840224e-05, |
| "loss": 0.9157, |
| "step": 238 |
| }, |
| { |
| "epoch": 17.071428571428573, |
| "grad_norm": 3.7386388778686523, |
| "learning_rate": 2.4972518353650626e-05, |
| "loss": 0.9065, |
| "step": 239 |
| }, |
| { |
| "epoch": 17.142857142857142, |
| "grad_norm": 2.243206739425659, |
| "learning_rate": 2.497105348544507e-05, |
| "loss": 0.8066, |
| "step": 240 |
| }, |
| { |
| "epoch": 17.214285714285715, |
| "grad_norm": 3.9942452907562256, |
| "learning_rate": 2.4969550628247805e-05, |
| "loss": 0.9173, |
| "step": 241 |
| }, |
| { |
| "epoch": 17.285714285714285, |
| "grad_norm": 4.815189361572266, |
| "learning_rate": 2.49680097866368e-05, |
| "loss": 0.9629, |
| "step": 242 |
| }, |
| { |
| "epoch": 17.357142857142858, |
| "grad_norm": 5.594153881072998, |
| "learning_rate": 2.4966430965305727e-05, |
| "loss": 0.9673, |
| "step": 243 |
| }, |
| { |
| "epoch": 17.428571428571427, |
| "grad_norm": 4.274055004119873, |
| "learning_rate": 2.4964814169063948e-05, |
| "loss": 0.9222, |
| "step": 244 |
| }, |
| { |
| "epoch": 17.5, |
| "grad_norm": 3.3747873306274414, |
| "learning_rate": 2.4963159402836506e-05, |
| "loss": 0.8992, |
| "step": 245 |
| }, |
| { |
| "epoch": 17.571428571428573, |
| "grad_norm": 4.040309429168701, |
| "learning_rate": 2.49614666716641e-05, |
| "loss": 0.8257, |
| "step": 246 |
| }, |
| { |
| "epoch": 17.642857142857142, |
| "grad_norm": 3.115257740020752, |
| "learning_rate": 2.495973598070309e-05, |
| "loss": 0.9713, |
| "step": 247 |
| }, |
| { |
| "epoch": 17.714285714285715, |
| "grad_norm": 2.5480332374572754, |
| "learning_rate": 2.4957967335225456e-05, |
| "loss": 0.8687, |
| "step": 248 |
| }, |
| { |
| "epoch": 17.785714285714285, |
| "grad_norm": 3.8874218463897705, |
| "learning_rate": 2.4956160740618806e-05, |
| "loss": 0.9432, |
| "step": 249 |
| }, |
| { |
| "epoch": 17.857142857142858, |
| "grad_norm": 3.870405673980713, |
| "learning_rate": 2.495431620238633e-05, |
| "loss": 0.8584, |
| "step": 250 |
| }, |
| { |
| "epoch": 17.928571428571427, |
| "grad_norm": 2.534801483154297, |
| "learning_rate": 2.495243372614682e-05, |
| "loss": 0.9255, |
| "step": 251 |
| }, |
| { |
| "epoch": 18.0, |
| "grad_norm": 2.785574197769165, |
| "learning_rate": 2.495051331763462e-05, |
| "loss": 0.8727, |
| "step": 252 |
| }, |
| { |
| "epoch": 18.071428571428573, |
| "grad_norm": 4.180792331695557, |
| "learning_rate": 2.494855498269963e-05, |
| "loss": 0.9683, |
| "step": 253 |
| }, |
| { |
| "epoch": 18.142857142857142, |
| "grad_norm": 5.634334564208984, |
| "learning_rate": 2.4946558727307277e-05, |
| "loss": 0.9485, |
| "step": 254 |
| }, |
| { |
| "epoch": 18.214285714285715, |
| "grad_norm": 4.938521862030029, |
| "learning_rate": 2.4944524557538503e-05, |
| "loss": 0.878, |
| "step": 255 |
| }, |
| { |
| "epoch": 18.285714285714285, |
| "grad_norm": 3.9465174674987793, |
| "learning_rate": 2.4942452479589735e-05, |
| "loss": 0.9731, |
| "step": 256 |
| }, |
| { |
| "epoch": 18.357142857142858, |
| "grad_norm": 3.601600170135498, |
| "learning_rate": 2.494034249977289e-05, |
| "loss": 0.9057, |
| "step": 257 |
| }, |
| { |
| "epoch": 18.428571428571427, |
| "grad_norm": 3.65264892578125, |
| "learning_rate": 2.4938194624515333e-05, |
| "loss": 0.9126, |
| "step": 258 |
| }, |
| { |
| "epoch": 18.5, |
| "grad_norm": 4.7837677001953125, |
| "learning_rate": 2.4936008860359854e-05, |
| "loss": 0.9847, |
| "step": 259 |
| }, |
| { |
| "epoch": 18.571428571428573, |
| "grad_norm": 4.8172831535339355, |
| "learning_rate": 2.4933785213964677e-05, |
| "loss": 0.9202, |
| "step": 260 |
| }, |
| { |
| "epoch": 18.642857142857142, |
| "grad_norm": 4.068408012390137, |
| "learning_rate": 2.4931523692103418e-05, |
| "loss": 0.9072, |
| "step": 261 |
| }, |
| { |
| "epoch": 18.714285714285715, |
| "grad_norm": 4.106720924377441, |
| "learning_rate": 2.492922430166506e-05, |
| "loss": 0.8837, |
| "step": 262 |
| }, |
| { |
| "epoch": 18.785714285714285, |
| "grad_norm": 3.7029213905334473, |
| "learning_rate": 2.4926887049653943e-05, |
| "loss": 0.9464, |
| "step": 263 |
| }, |
| { |
| "epoch": 18.857142857142858, |
| "grad_norm": 3.3589251041412354, |
| "learning_rate": 2.492451194318975e-05, |
| "loss": 0.8956, |
| "step": 264 |
| }, |
| { |
| "epoch": 18.928571428571427, |
| "grad_norm": 3.4119558334350586, |
| "learning_rate": 2.4922098989507454e-05, |
| "loss": 0.9177, |
| "step": 265 |
| }, |
| { |
| "epoch": 19.0, |
| "grad_norm": 2.3078062534332275, |
| "learning_rate": 2.4919648195957344e-05, |
| "loss": 0.8086, |
| "step": 266 |
| }, |
| { |
| "epoch": 19.071428571428573, |
| "grad_norm": 2.213747978210449, |
| "learning_rate": 2.4917159570004954e-05, |
| "loss": 0.8417, |
| "step": 267 |
| }, |
| { |
| "epoch": 19.142857142857142, |
| "grad_norm": 2.629509687423706, |
| "learning_rate": 2.491463311923108e-05, |
| "loss": 0.9628, |
| "step": 268 |
| }, |
| { |
| "epoch": 19.214285714285715, |
| "grad_norm": 1.877100944519043, |
| "learning_rate": 2.491206885133171e-05, |
| "loss": 0.8174, |
| "step": 269 |
| }, |
| { |
| "epoch": 19.285714285714285, |
| "grad_norm": 2.653271436691284, |
| "learning_rate": 2.490946677411807e-05, |
| "loss": 0.8823, |
| "step": 270 |
| }, |
| { |
| "epoch": 19.357142857142858, |
| "grad_norm": 2.488171339035034, |
| "learning_rate": 2.4906826895516528e-05, |
| "loss": 0.888, |
| "step": 271 |
| }, |
| { |
| "epoch": 19.428571428571427, |
| "grad_norm": 2.5691208839416504, |
| "learning_rate": 2.490414922356861e-05, |
| "loss": 0.87, |
| "step": 272 |
| }, |
| { |
| "epoch": 19.5, |
| "grad_norm": 3.2644972801208496, |
| "learning_rate": 2.4901433766430975e-05, |
| "loss": 0.9013, |
| "step": 273 |
| }, |
| { |
| "epoch": 19.571428571428573, |
| "grad_norm": 5.025469779968262, |
| "learning_rate": 2.4898680532375374e-05, |
| "loss": 0.9013, |
| "step": 274 |
| }, |
| { |
| "epoch": 19.642857142857142, |
| "grad_norm": 3.607896566390991, |
| "learning_rate": 2.489588952978863e-05, |
| "loss": 0.8409, |
| "step": 275 |
| }, |
| { |
| "epoch": 19.714285714285715, |
| "grad_norm": 2.6993868350982666, |
| "learning_rate": 2.4893060767172632e-05, |
| "loss": 0.7845, |
| "step": 276 |
| }, |
| { |
| "epoch": 19.785714285714285, |
| "grad_norm": 2.4249427318573, |
| "learning_rate": 2.489019425314427e-05, |
| "loss": 0.8113, |
| "step": 277 |
| }, |
| { |
| "epoch": 19.857142857142858, |
| "grad_norm": 4.231166362762451, |
| "learning_rate": 2.4887289996435452e-05, |
| "loss": 0.8879, |
| "step": 278 |
| }, |
| { |
| "epoch": 19.928571428571427, |
| "grad_norm": 7.234214782714844, |
| "learning_rate": 2.4884348005893045e-05, |
| "loss": 0.8847, |
| "step": 279 |
| }, |
| { |
| "epoch": 20.0, |
| "grad_norm": 4.228268146514893, |
| "learning_rate": 2.488136829047886e-05, |
| "loss": 0.8504, |
| "step": 280 |
| }, |
| { |
| "epoch": 20.071428571428573, |
| "grad_norm": 2.944539785385132, |
| "learning_rate": 2.487835085926963e-05, |
| "loss": 0.845, |
| "step": 281 |
| }, |
| { |
| "epoch": 20.142857142857142, |
| "grad_norm": 2.7270450592041016, |
| "learning_rate": 2.487529572145697e-05, |
| "loss": 0.8335, |
| "step": 282 |
| }, |
| { |
| "epoch": 20.214285714285715, |
| "grad_norm": 2.7035233974456787, |
| "learning_rate": 2.4872202886347362e-05, |
| "loss": 0.815, |
| "step": 283 |
| }, |
| { |
| "epoch": 20.285714285714285, |
| "grad_norm": 3.4698843955993652, |
| "learning_rate": 2.486907236336212e-05, |
| "loss": 0.875, |
| "step": 284 |
| }, |
| { |
| "epoch": 20.357142857142858, |
| "grad_norm": 4.334729194641113, |
| "learning_rate": 2.4865904162037358e-05, |
| "loss": 0.8447, |
| "step": 285 |
| }, |
| { |
| "epoch": 20.428571428571427, |
| "grad_norm": 3.5744895935058594, |
| "learning_rate": 2.4862698292023963e-05, |
| "loss": 0.792, |
| "step": 286 |
| }, |
| { |
| "epoch": 20.5, |
| "grad_norm": 3.5507006645202637, |
| "learning_rate": 2.4859454763087577e-05, |
| "loss": 0.7948, |
| "step": 287 |
| }, |
| { |
| "epoch": 20.571428571428573, |
| "grad_norm": 2.2967231273651123, |
| "learning_rate": 2.4856173585108544e-05, |
| "loss": 0.7975, |
| "step": 288 |
| }, |
| { |
| "epoch": 20.642857142857142, |
| "grad_norm": 2.5421581268310547, |
| "learning_rate": 2.4852854768081912e-05, |
| "loss": 0.8074, |
| "step": 289 |
| }, |
| { |
| "epoch": 20.714285714285715, |
| "grad_norm": 2.4942188262939453, |
| "learning_rate": 2.4849498322117364e-05, |
| "loss": 0.7962, |
| "step": 290 |
| }, |
| { |
| "epoch": 20.785714285714285, |
| "grad_norm": 4.169875621795654, |
| "learning_rate": 2.4846104257439222e-05, |
| "loss": 0.8181, |
| "step": 291 |
| }, |
| { |
| "epoch": 20.857142857142858, |
| "grad_norm": 6.13093376159668, |
| "learning_rate": 2.4842672584386396e-05, |
| "loss": 0.8434, |
| "step": 292 |
| }, |
| { |
| "epoch": 20.928571428571427, |
| "grad_norm": 3.3157453536987305, |
| "learning_rate": 2.483920331341235e-05, |
| "loss": 0.8108, |
| "step": 293 |
| }, |
| { |
| "epoch": 21.0, |
| "grad_norm": 3.0505266189575195, |
| "learning_rate": 2.4835696455085093e-05, |
| "loss": 0.7412, |
| "step": 294 |
| }, |
| { |
| "epoch": 21.071428571428573, |
| "grad_norm": 5.5504350662231445, |
| "learning_rate": 2.483215202008712e-05, |
| "loss": 0.8437, |
| "step": 295 |
| }, |
| { |
| "epoch": 21.142857142857142, |
| "grad_norm": 6.492306709289551, |
| "learning_rate": 2.4828570019215396e-05, |
| "loss": 0.8533, |
| "step": 296 |
| }, |
| { |
| "epoch": 21.214285714285715, |
| "grad_norm": 3.0854108333587646, |
| "learning_rate": 2.4824950463381314e-05, |
| "loss": 0.7881, |
| "step": 297 |
| }, |
| { |
| "epoch": 21.285714285714285, |
| "grad_norm": 3.006676435470581, |
| "learning_rate": 2.482129336361067e-05, |
| "loss": 0.8156, |
| "step": 298 |
| }, |
| { |
| "epoch": 21.357142857142858, |
| "grad_norm": 3.312669515609741, |
| "learning_rate": 2.481759873104363e-05, |
| "loss": 0.8011, |
| "step": 299 |
| }, |
| { |
| "epoch": 21.428571428571427, |
| "grad_norm": 5.346591949462891, |
| "learning_rate": 2.4813866576934676e-05, |
| "loss": 0.7895, |
| "step": 300 |
| }, |
| { |
| "epoch": 21.5, |
| "grad_norm": 5.8296074867248535, |
| "learning_rate": 2.4810096912652604e-05, |
| "loss": 0.893, |
| "step": 301 |
| }, |
| { |
| "epoch": 21.571428571428573, |
| "grad_norm": 3.022014617919922, |
| "learning_rate": 2.480628974968046e-05, |
| "loss": 0.7875, |
| "step": 302 |
| }, |
| { |
| "epoch": 21.642857142857142, |
| "grad_norm": 3.372377395629883, |
| "learning_rate": 2.4802445099615525e-05, |
| "loss": 0.8107, |
| "step": 303 |
| }, |
| { |
| "epoch": 21.714285714285715, |
| "grad_norm": 4.238394260406494, |
| "learning_rate": 2.479856297416927e-05, |
| "loss": 0.7879, |
| "step": 304 |
| }, |
| { |
| "epoch": 21.785714285714285, |
| "grad_norm": 4.552757263183594, |
| "learning_rate": 2.4794643385167327e-05, |
| "loss": 0.8352, |
| "step": 305 |
| }, |
| { |
| "epoch": 21.857142857142858, |
| "grad_norm": 4.247888088226318, |
| "learning_rate": 2.4790686344549436e-05, |
| "loss": 0.7733, |
| "step": 306 |
| }, |
| { |
| "epoch": 21.928571428571427, |
| "grad_norm": 2.4774329662323, |
| "learning_rate": 2.478669186436943e-05, |
| "loss": 0.7338, |
| "step": 307 |
| }, |
| { |
| "epoch": 22.0, |
| "grad_norm": 2.973707675933838, |
| "learning_rate": 2.478265995679519e-05, |
| "loss": 0.7532, |
| "step": 308 |
| }, |
| { |
| "epoch": 22.071428571428573, |
| "grad_norm": 5.437268257141113, |
| "learning_rate": 2.4778590634108613e-05, |
| "loss": 0.8951, |
| "step": 309 |
| }, |
| { |
| "epoch": 22.142857142857142, |
| "grad_norm": 5.616093635559082, |
| "learning_rate": 2.4774483908705546e-05, |
| "loss": 0.8117, |
| "step": 310 |
| }, |
| { |
| "epoch": 22.214285714285715, |
| "grad_norm": 3.7091429233551025, |
| "learning_rate": 2.4770339793095802e-05, |
| "loss": 0.8166, |
| "step": 311 |
| }, |
| { |
| "epoch": 22.285714285714285, |
| "grad_norm": 3.873770236968994, |
| "learning_rate": 2.4766158299903062e-05, |
| "loss": 0.7786, |
| "step": 312 |
| }, |
| { |
| "epoch": 22.357142857142858, |
| "grad_norm": 3.692065477371216, |
| "learning_rate": 2.4761939441864895e-05, |
| "loss": 0.8148, |
| "step": 313 |
| }, |
| { |
| "epoch": 22.428571428571427, |
| "grad_norm": 4.50544548034668, |
| "learning_rate": 2.4757683231832662e-05, |
| "loss": 0.9163, |
| "step": 314 |
| }, |
| { |
| "epoch": 22.5, |
| "grad_norm": 4.464144229888916, |
| "learning_rate": 2.4753389682771523e-05, |
| "loss": 0.8223, |
| "step": 315 |
| }, |
| { |
| "epoch": 22.571428571428573, |
| "grad_norm": 3.729602575302124, |
| "learning_rate": 2.474905880776037e-05, |
| "loss": 0.8479, |
| "step": 316 |
| }, |
| { |
| "epoch": 22.642857142857142, |
| "grad_norm": 3.674757957458496, |
| "learning_rate": 2.47446906199918e-05, |
| "loss": 0.7319, |
| "step": 317 |
| }, |
| { |
| "epoch": 22.714285714285715, |
| "grad_norm": 4.361830711364746, |
| "learning_rate": 2.4740285132772072e-05, |
| "loss": 0.8054, |
| "step": 318 |
| }, |
| { |
| "epoch": 22.785714285714285, |
| "grad_norm": 4.365119934082031, |
| "learning_rate": 2.4735842359521064e-05, |
| "loss": 0.801, |
| "step": 319 |
| }, |
| { |
| "epoch": 22.857142857142858, |
| "grad_norm": 4.092256546020508, |
| "learning_rate": 2.4731362313772233e-05, |
| "loss": 0.8389, |
| "step": 320 |
| }, |
| { |
| "epoch": 22.928571428571427, |
| "grad_norm": 3.239090919494629, |
| "learning_rate": 2.4726845009172572e-05, |
| "loss": 0.7814, |
| "step": 321 |
| }, |
| { |
| "epoch": 23.0, |
| "grad_norm": 2.6484477519989014, |
| "learning_rate": 2.4722290459482578e-05, |
| "loss": 0.7468, |
| "step": 322 |
| }, |
| { |
| "epoch": 23.071428571428573, |
| "grad_norm": 3.275247573852539, |
| "learning_rate": 2.47176986785762e-05, |
| "loss": 0.827, |
| "step": 323 |
| }, |
| { |
| "epoch": 23.142857142857142, |
| "grad_norm": 3.040330410003662, |
| "learning_rate": 2.47130696804408e-05, |
| "loss": 0.7849, |
| "step": 324 |
| }, |
| { |
| "epoch": 23.214285714285715, |
| "grad_norm": 4.0800395011901855, |
| "learning_rate": 2.47084034791771e-05, |
| "loss": 0.856, |
| "step": 325 |
| }, |
| { |
| "epoch": 23.285714285714285, |
| "grad_norm": 3.5290443897247314, |
| "learning_rate": 2.4703700088999167e-05, |
| "loss": 0.825, |
| "step": 326 |
| }, |
| { |
| "epoch": 23.357142857142858, |
| "grad_norm": 3.670090436935425, |
| "learning_rate": 2.4698959524234346e-05, |
| "loss": 0.8061, |
| "step": 327 |
| }, |
| { |
| "epoch": 23.428571428571427, |
| "grad_norm": 1.9602779150009155, |
| "learning_rate": 2.4694181799323206e-05, |
| "loss": 0.7803, |
| "step": 328 |
| }, |
| { |
| "epoch": 23.5, |
| "grad_norm": 3.8044676780700684, |
| "learning_rate": 2.468936692881954e-05, |
| "loss": 0.767, |
| "step": 329 |
| }, |
| { |
| "epoch": 23.571428571428573, |
| "grad_norm": 3.5850207805633545, |
| "learning_rate": 2.4684514927390274e-05, |
| "loss": 0.7555, |
| "step": 330 |
| }, |
| { |
| "epoch": 23.642857142857142, |
| "grad_norm": 2.4394097328186035, |
| "learning_rate": 2.4679625809815443e-05, |
| "loss": 0.7911, |
| "step": 331 |
| }, |
| { |
| "epoch": 23.714285714285715, |
| "grad_norm": 3.6944406032562256, |
| "learning_rate": 2.467469959098815e-05, |
| "loss": 0.7708, |
| "step": 332 |
| }, |
| { |
| "epoch": 23.785714285714285, |
| "grad_norm": 3.9482243061065674, |
| "learning_rate": 2.4669736285914505e-05, |
| "loss": 0.8035, |
| "step": 333 |
| }, |
| { |
| "epoch": 23.857142857142858, |
| "grad_norm": 4.452454566955566, |
| "learning_rate": 2.4664735909713606e-05, |
| "loss": 0.7837, |
| "step": 334 |
| }, |
| { |
| "epoch": 23.928571428571427, |
| "grad_norm": 3.8159029483795166, |
| "learning_rate": 2.465969847761746e-05, |
| "loss": 0.8188, |
| "step": 335 |
| }, |
| { |
| "epoch": 24.0, |
| "grad_norm": 2.743255138397217, |
| "learning_rate": 2.4654624004970957e-05, |
| "loss": 0.7004, |
| "step": 336 |
| }, |
| { |
| "epoch": 24.071428571428573, |
| "grad_norm": 3.636960506439209, |
| "learning_rate": 2.464951250723183e-05, |
| "loss": 0.7809, |
| "step": 337 |
| }, |
| { |
| "epoch": 24.142857142857142, |
| "grad_norm": 3.9498450756073, |
| "learning_rate": 2.4644363999970576e-05, |
| "loss": 0.7903, |
| "step": 338 |
| }, |
| { |
| "epoch": 24.214285714285715, |
| "grad_norm": 3.0506386756896973, |
| "learning_rate": 2.4639178498870452e-05, |
| "loss": 0.8114, |
| "step": 339 |
| }, |
| { |
| "epoch": 24.285714285714285, |
| "grad_norm": 3.2507994174957275, |
| "learning_rate": 2.4633956019727385e-05, |
| "loss": 0.7556, |
| "step": 340 |
| }, |
| { |
| "epoch": 24.357142857142858, |
| "grad_norm": 5.005919933319092, |
| "learning_rate": 2.4628696578449956e-05, |
| "loss": 0.8456, |
| "step": 341 |
| }, |
| { |
| "epoch": 24.428571428571427, |
| "grad_norm": 3.5123672485351562, |
| "learning_rate": 2.4623400191059335e-05, |
| "loss": 0.7921, |
| "step": 342 |
| }, |
| { |
| "epoch": 24.5, |
| "grad_norm": 3.3725931644439697, |
| "learning_rate": 2.4618066873689238e-05, |
| "loss": 0.7986, |
| "step": 343 |
| }, |
| { |
| "epoch": 24.571428571428573, |
| "grad_norm": 2.6177029609680176, |
| "learning_rate": 2.461269664258587e-05, |
| "loss": 0.7449, |
| "step": 344 |
| }, |
| { |
| "epoch": 24.642857142857142, |
| "grad_norm": 5.391937732696533, |
| "learning_rate": 2.4607289514107888e-05, |
| "loss": 0.7433, |
| "step": 345 |
| }, |
| { |
| "epoch": 24.714285714285715, |
| "grad_norm": 2.888105630874634, |
| "learning_rate": 2.460184550472635e-05, |
| "loss": 0.7079, |
| "step": 346 |
| }, |
| { |
| "epoch": 24.785714285714285, |
| "grad_norm": 2.3010149002075195, |
| "learning_rate": 2.4596364631024643e-05, |
| "loss": 0.7376, |
| "step": 347 |
| }, |
| { |
| "epoch": 24.857142857142858, |
| "grad_norm": 3.590585470199585, |
| "learning_rate": 2.459084690969846e-05, |
| "loss": 0.7532, |
| "step": 348 |
| }, |
| { |
| "epoch": 24.928571428571427, |
| "grad_norm": 6.115037441253662, |
| "learning_rate": 2.4585292357555746e-05, |
| "loss": 0.7568, |
| "step": 349 |
| }, |
| { |
| "epoch": 25.0, |
| "grad_norm": 2.7985963821411133, |
| "learning_rate": 2.457970099151662e-05, |
| "loss": 0.7001, |
| "step": 350 |
| }, |
| { |
| "epoch": 25.071428571428573, |
| "grad_norm": 5.373832702636719, |
| "learning_rate": 2.4574072828613354e-05, |
| "loss": 0.8254, |
| "step": 351 |
| }, |
| { |
| "epoch": 25.142857142857142, |
| "grad_norm": 3.988442897796631, |
| "learning_rate": 2.4568407885990313e-05, |
| "loss": 0.7305, |
| "step": 352 |
| }, |
| { |
| "epoch": 25.214285714285715, |
| "grad_norm": 3.5456085205078125, |
| "learning_rate": 2.4562706180903894e-05, |
| "loss": 0.7546, |
| "step": 353 |
| }, |
| { |
| "epoch": 25.285714285714285, |
| "grad_norm": 3.8878087997436523, |
| "learning_rate": 2.4556967730722478e-05, |
| "loss": 0.794, |
| "step": 354 |
| }, |
| { |
| "epoch": 25.357142857142858, |
| "grad_norm": 3.997696876525879, |
| "learning_rate": 2.455119255292638e-05, |
| "loss": 0.7983, |
| "step": 355 |
| }, |
| { |
| "epoch": 25.428571428571427, |
| "grad_norm": 5.822238445281982, |
| "learning_rate": 2.4545380665107786e-05, |
| "loss": 0.899, |
| "step": 356 |
| }, |
| { |
| "epoch": 25.5, |
| "grad_norm": 3.2456319332122803, |
| "learning_rate": 2.453953208497073e-05, |
| "loss": 0.7284, |
| "step": 357 |
| }, |
| { |
| "epoch": 25.571428571428573, |
| "grad_norm": 3.3598856925964355, |
| "learning_rate": 2.4533646830330986e-05, |
| "loss": 0.8056, |
| "step": 358 |
| }, |
| { |
| "epoch": 25.642857142857142, |
| "grad_norm": 3.6620326042175293, |
| "learning_rate": 2.452772491911607e-05, |
| "loss": 0.7994, |
| "step": 359 |
| }, |
| { |
| "epoch": 25.714285714285715, |
| "grad_norm": 4.005545139312744, |
| "learning_rate": 2.4521766369365142e-05, |
| "loss": 0.7773, |
| "step": 360 |
| }, |
| { |
| "epoch": 25.785714285714285, |
| "grad_norm": 4.023702144622803, |
| "learning_rate": 2.4515771199228987e-05, |
| "loss": 0.7717, |
| "step": 361 |
| }, |
| { |
| "epoch": 25.857142857142858, |
| "grad_norm": 1.9900436401367188, |
| "learning_rate": 2.450973942696993e-05, |
| "loss": 0.7331, |
| "step": 362 |
| }, |
| { |
| "epoch": 25.928571428571427, |
| "grad_norm": 5.718785285949707, |
| "learning_rate": 2.450367107096179e-05, |
| "loss": 0.7818, |
| "step": 363 |
| }, |
| { |
| "epoch": 26.0, |
| "grad_norm": 3.5384066104888916, |
| "learning_rate": 2.449756614968984e-05, |
| "loss": 0.8165, |
| "step": 364 |
| }, |
| { |
| "epoch": 26.071428571428573, |
| "grad_norm": 2.2805933952331543, |
| "learning_rate": 2.449142468175072e-05, |
| "loss": 0.68, |
| "step": 365 |
| }, |
| { |
| "epoch": 26.142857142857142, |
| "grad_norm": 2.818986654281616, |
| "learning_rate": 2.4485246685852413e-05, |
| "loss": 0.6765, |
| "step": 366 |
| }, |
| { |
| "epoch": 26.214285714285715, |
| "grad_norm": 3.089205741882324, |
| "learning_rate": 2.4479032180814166e-05, |
| "loss": 0.6901, |
| "step": 367 |
| }, |
| { |
| "epoch": 26.285714285714285, |
| "grad_norm": 2.990636110305786, |
| "learning_rate": 2.447278118556644e-05, |
| "loss": 0.6839, |
| "step": 368 |
| }, |
| { |
| "epoch": 26.357142857142858, |
| "grad_norm": 2.0169079303741455, |
| "learning_rate": 2.446649371915084e-05, |
| "loss": 0.6499, |
| "step": 369 |
| }, |
| { |
| "epoch": 26.428571428571427, |
| "grad_norm": 2.073349714279175, |
| "learning_rate": 2.4460169800720095e-05, |
| "loss": 0.6312, |
| "step": 370 |
| }, |
| { |
| "epoch": 26.5, |
| "grad_norm": 5.544507026672363, |
| "learning_rate": 2.4453809449537947e-05, |
| "loss": 0.7886, |
| "step": 371 |
| }, |
| { |
| "epoch": 26.571428571428573, |
| "grad_norm": 3.550513505935669, |
| "learning_rate": 2.4447412684979127e-05, |
| "loss": 0.7682, |
| "step": 372 |
| }, |
| { |
| "epoch": 26.642857142857142, |
| "grad_norm": 4.850046634674072, |
| "learning_rate": 2.4440979526529295e-05, |
| "loss": 0.7891, |
| "step": 373 |
| }, |
| { |
| "epoch": 26.714285714285715, |
| "grad_norm": 5.702778339385986, |
| "learning_rate": 2.4434509993784972e-05, |
| "loss": 0.7478, |
| "step": 374 |
| }, |
| { |
| "epoch": 26.785714285714285, |
| "grad_norm": 2.3614633083343506, |
| "learning_rate": 2.4428004106453462e-05, |
| "loss": 0.6622, |
| "step": 375 |
| }, |
| { |
| "epoch": 26.857142857142858, |
| "grad_norm": 4.546297073364258, |
| "learning_rate": 2.4421461884352836e-05, |
| "loss": 0.7229, |
| "step": 376 |
| }, |
| { |
| "epoch": 26.928571428571427, |
| "grad_norm": 4.9091386795043945, |
| "learning_rate": 2.4414883347411836e-05, |
| "loss": 0.7186, |
| "step": 377 |
| }, |
| { |
| "epoch": 27.0, |
| "grad_norm": 3.949092388153076, |
| "learning_rate": 2.440826851566983e-05, |
| "loss": 0.7257, |
| "step": 378 |
| }, |
| { |
| "epoch": 27.071428571428573, |
| "grad_norm": 5.0857439041137695, |
| "learning_rate": 2.4401617409276735e-05, |
| "loss": 0.6953, |
| "step": 379 |
| }, |
| { |
| "epoch": 27.142857142857142, |
| "grad_norm": 4.325730323791504, |
| "learning_rate": 2.439493004849298e-05, |
| "loss": 0.6696, |
| "step": 380 |
| }, |
| { |
| "epoch": 27.214285714285715, |
| "grad_norm": 3.15690541267395, |
| "learning_rate": 2.438820645368942e-05, |
| "loss": 0.7532, |
| "step": 381 |
| }, |
| { |
| "epoch": 27.285714285714285, |
| "grad_norm": 4.08896017074585, |
| "learning_rate": 2.4381446645347297e-05, |
| "loss": 0.7039, |
| "step": 382 |
| }, |
| { |
| "epoch": 27.357142857142858, |
| "grad_norm": 3.996779441833496, |
| "learning_rate": 2.4374650644058156e-05, |
| "loss": 0.7489, |
| "step": 383 |
| }, |
| { |
| "epoch": 27.428571428571427, |
| "grad_norm": 4.629286766052246, |
| "learning_rate": 2.43678184705238e-05, |
| "loss": 0.7634, |
| "step": 384 |
| }, |
| { |
| "epoch": 27.5, |
| "grad_norm": 4.668010234832764, |
| "learning_rate": 2.4360950145556208e-05, |
| "loss": 0.7321, |
| "step": 385 |
| }, |
| { |
| "epoch": 27.571428571428573, |
| "grad_norm": 2.8770227432250977, |
| "learning_rate": 2.4354045690077492e-05, |
| "loss": 0.6417, |
| "step": 386 |
| }, |
| { |
| "epoch": 27.642857142857142, |
| "grad_norm": 4.506302356719971, |
| "learning_rate": 2.4347105125119824e-05, |
| "loss": 0.6698, |
| "step": 387 |
| }, |
| { |
| "epoch": 27.714285714285715, |
| "grad_norm": 2.6643428802490234, |
| "learning_rate": 2.4340128471825362e-05, |
| "loss": 0.6938, |
| "step": 388 |
| }, |
| { |
| "epoch": 27.785714285714285, |
| "grad_norm": 2.8056280612945557, |
| "learning_rate": 2.4333115751446208e-05, |
| "loss": 0.6743, |
| "step": 389 |
| }, |
| { |
| "epoch": 27.857142857142858, |
| "grad_norm": 2.7733471393585205, |
| "learning_rate": 2.4326066985344318e-05, |
| "loss": 0.6573, |
| "step": 390 |
| }, |
| { |
| "epoch": 27.928571428571427, |
| "grad_norm": 5.0971174240112305, |
| "learning_rate": 2.4318982194991463e-05, |
| "loss": 0.6754, |
| "step": 391 |
| }, |
| { |
| "epoch": 28.0, |
| "grad_norm": 6.083596706390381, |
| "learning_rate": 2.4311861401969138e-05, |
| "loss": 0.7646, |
| "step": 392 |
| }, |
| { |
| "epoch": 28.071428571428573, |
| "grad_norm": 4.150708198547363, |
| "learning_rate": 2.4304704627968515e-05, |
| "loss": 0.6958, |
| "step": 393 |
| }, |
| { |
| "epoch": 28.142857142857142, |
| "grad_norm": 3.6580939292907715, |
| "learning_rate": 2.429751189479037e-05, |
| "loss": 0.6848, |
| "step": 394 |
| }, |
| { |
| "epoch": 28.214285714285715, |
| "grad_norm": 2.1017942428588867, |
| "learning_rate": 2.429028322434501e-05, |
| "loss": 0.6576, |
| "step": 395 |
| }, |
| { |
| "epoch": 28.285714285714285, |
| "grad_norm": 1.8062525987625122, |
| "learning_rate": 2.4283018638652234e-05, |
| "loss": 0.6963, |
| "step": 396 |
| }, |
| { |
| "epoch": 28.357142857142858, |
| "grad_norm": 3.1975321769714355, |
| "learning_rate": 2.427571815984121e-05, |
| "loss": 0.6599, |
| "step": 397 |
| }, |
| { |
| "epoch": 28.428571428571427, |
| "grad_norm": 3.0259714126586914, |
| "learning_rate": 2.4268381810150474e-05, |
| "loss": 0.6959, |
| "step": 398 |
| }, |
| { |
| "epoch": 28.5, |
| "grad_norm": 5.031489849090576, |
| "learning_rate": 2.426100961192782e-05, |
| "loss": 0.7208, |
| "step": 399 |
| }, |
| { |
| "epoch": 28.571428571428573, |
| "grad_norm": 3.5637011528015137, |
| "learning_rate": 2.4253601587630236e-05, |
| "loss": 0.6896, |
| "step": 400 |
| }, |
| { |
| "epoch": 28.642857142857142, |
| "grad_norm": 3.333253860473633, |
| "learning_rate": 2.4246157759823855e-05, |
| "loss": 0.6815, |
| "step": 401 |
| }, |
| { |
| "epoch": 28.714285714285715, |
| "grad_norm": 3.422375440597534, |
| "learning_rate": 2.4238678151183863e-05, |
| "loss": 0.6606, |
| "step": 402 |
| }, |
| { |
| "epoch": 28.785714285714285, |
| "grad_norm": 2.458256721496582, |
| "learning_rate": 2.423116278449445e-05, |
| "loss": 0.6484, |
| "step": 403 |
| }, |
| { |
| "epoch": 28.857142857142858, |
| "grad_norm": 4.026810169219971, |
| "learning_rate": 2.4223611682648724e-05, |
| "loss": 0.7167, |
| "step": 404 |
| }, |
| { |
| "epoch": 28.928571428571427, |
| "grad_norm": 3.2683417797088623, |
| "learning_rate": 2.4216024868648644e-05, |
| "loss": 0.6899, |
| "step": 405 |
| }, |
| { |
| "epoch": 29.0, |
| "grad_norm": 3.54608416557312, |
| "learning_rate": 2.4208402365604972e-05, |
| "loss": 0.6843, |
| "step": 406 |
| }, |
| { |
| "epoch": 29.071428571428573, |
| "grad_norm": 4.817044734954834, |
| "learning_rate": 2.420074419673717e-05, |
| "loss": 0.6544, |
| "step": 407 |
| }, |
| { |
| "epoch": 29.142857142857142, |
| "grad_norm": 3.0443451404571533, |
| "learning_rate": 2.4193050385373344e-05, |
| "loss": 0.6134, |
| "step": 408 |
| }, |
| { |
| "epoch": 29.214285714285715, |
| "grad_norm": 6.059633731842041, |
| "learning_rate": 2.418532095495018e-05, |
| "loss": 0.6718, |
| "step": 409 |
| }, |
| { |
| "epoch": 29.285714285714285, |
| "grad_norm": 2.560316324234009, |
| "learning_rate": 2.417755592901287e-05, |
| "loss": 0.6134, |
| "step": 410 |
| }, |
| { |
| "epoch": 29.357142857142858, |
| "grad_norm": 3.06530499458313, |
| "learning_rate": 2.4169755331215023e-05, |
| "loss": 0.6158, |
| "step": 411 |
| }, |
| { |
| "epoch": 29.428571428571427, |
| "grad_norm": 2.439854621887207, |
| "learning_rate": 2.4161919185318617e-05, |
| "loss": 0.6428, |
| "step": 412 |
| }, |
| { |
| "epoch": 29.5, |
| "grad_norm": 4.74956750869751, |
| "learning_rate": 2.4154047515193904e-05, |
| "loss": 0.6847, |
| "step": 413 |
| }, |
| { |
| "epoch": 29.571428571428573, |
| "grad_norm": 3.4892289638519287, |
| "learning_rate": 2.4146140344819363e-05, |
| "loss": 0.6293, |
| "step": 414 |
| }, |
| { |
| "epoch": 29.642857142857142, |
| "grad_norm": 2.6727051734924316, |
| "learning_rate": 2.4138197698281606e-05, |
| "loss": 0.6429, |
| "step": 415 |
| }, |
| { |
| "epoch": 29.714285714285715, |
| "grad_norm": 4.598572254180908, |
| "learning_rate": 2.413021959977531e-05, |
| "loss": 0.6504, |
| "step": 416 |
| }, |
| { |
| "epoch": 29.785714285714285, |
| "grad_norm": 5.181699752807617, |
| "learning_rate": 2.4122206073603142e-05, |
| "loss": 0.6994, |
| "step": 417 |
| }, |
| { |
| "epoch": 29.857142857142858, |
| "grad_norm": 4.484119415283203, |
| "learning_rate": 2.4114157144175703e-05, |
| "loss": 0.7659, |
| "step": 418 |
| }, |
| { |
| "epoch": 29.928571428571427, |
| "grad_norm": 4.818487644195557, |
| "learning_rate": 2.4106072836011422e-05, |
| "loss": 0.7583, |
| "step": 419 |
| }, |
| { |
| "epoch": 30.0, |
| "grad_norm": 3.408137083053589, |
| "learning_rate": 2.40979531737365e-05, |
| "loss": 0.6579, |
| "step": 420 |
| }, |
| { |
| "epoch": 30.071428571428573, |
| "grad_norm": 4.11182975769043, |
| "learning_rate": 2.4089798182084845e-05, |
| "loss": 0.69, |
| "step": 421 |
| }, |
| { |
| "epoch": 30.142857142857142, |
| "grad_norm": 4.751646518707275, |
| "learning_rate": 2.4081607885897966e-05, |
| "loss": 0.7584, |
| "step": 422 |
| }, |
| { |
| "epoch": 30.214285714285715, |
| "grad_norm": 5.190206050872803, |
| "learning_rate": 2.407338231012494e-05, |
| "loss": 0.7468, |
| "step": 423 |
| }, |
| { |
| "epoch": 30.285714285714285, |
| "grad_norm": 3.750694513320923, |
| "learning_rate": 2.406512147982228e-05, |
| "loss": 0.743, |
| "step": 424 |
| }, |
| { |
| "epoch": 30.357142857142858, |
| "grad_norm": 3.4604289531707764, |
| "learning_rate": 2.4056825420153917e-05, |
| "loss": 0.7051, |
| "step": 425 |
| }, |
| { |
| "epoch": 30.428571428571427, |
| "grad_norm": 2.6612353324890137, |
| "learning_rate": 2.4048494156391087e-05, |
| "loss": 0.6852, |
| "step": 426 |
| }, |
| { |
| "epoch": 30.5, |
| "grad_norm": 4.154074192047119, |
| "learning_rate": 2.4040127713912264e-05, |
| "loss": 0.6733, |
| "step": 427 |
| }, |
| { |
| "epoch": 30.571428571428573, |
| "grad_norm": 3.900482654571533, |
| "learning_rate": 2.403172611820308e-05, |
| "loss": 0.6134, |
| "step": 428 |
| }, |
| { |
| "epoch": 30.642857142857142, |
| "grad_norm": 5.209915637969971, |
| "learning_rate": 2.4023289394856257e-05, |
| "loss": 0.7324, |
| "step": 429 |
| }, |
| { |
| "epoch": 30.714285714285715, |
| "grad_norm": 2.668140411376953, |
| "learning_rate": 2.401481756957152e-05, |
| "loss": 0.6655, |
| "step": 430 |
| }, |
| { |
| "epoch": 30.785714285714285, |
| "grad_norm": 3.480163812637329, |
| "learning_rate": 2.4006310668155508e-05, |
| "loss": 0.7295, |
| "step": 431 |
| }, |
| { |
| "epoch": 30.857142857142858, |
| "grad_norm": 2.9695725440979004, |
| "learning_rate": 2.3997768716521723e-05, |
| "loss": 0.6949, |
| "step": 432 |
| }, |
| { |
| "epoch": 30.928571428571427, |
| "grad_norm": 2.298661470413208, |
| "learning_rate": 2.398919174069043e-05, |
| "loss": 0.6687, |
| "step": 433 |
| }, |
| { |
| "epoch": 31.0, |
| "grad_norm": 3.2847728729248047, |
| "learning_rate": 2.398057976678859e-05, |
| "loss": 0.6779, |
| "step": 434 |
| }, |
| { |
| "epoch": 31.071428571428573, |
| "grad_norm": 4.464836120605469, |
| "learning_rate": 2.3971932821049765e-05, |
| "loss": 0.7223, |
| "step": 435 |
| }, |
| { |
| "epoch": 31.142857142857142, |
| "grad_norm": 2.811166763305664, |
| "learning_rate": 2.396325092981405e-05, |
| "loss": 0.6677, |
| "step": 436 |
| }, |
| { |
| "epoch": 31.214285714285715, |
| "grad_norm": 3.1231977939605713, |
| "learning_rate": 2.3954534119527996e-05, |
| "loss": 0.6745, |
| "step": 437 |
| }, |
| { |
| "epoch": 31.285714285714285, |
| "grad_norm": 1.7231147289276123, |
| "learning_rate": 2.3945782416744517e-05, |
| "loss": 0.5596, |
| "step": 438 |
| }, |
| { |
| "epoch": 31.357142857142858, |
| "grad_norm": 2.8085505962371826, |
| "learning_rate": 2.3936995848122812e-05, |
| "loss": 0.5903, |
| "step": 439 |
| }, |
| { |
| "epoch": 31.428571428571427, |
| "grad_norm": 5.913967132568359, |
| "learning_rate": 2.3928174440428297e-05, |
| "loss": 0.6269, |
| "step": 440 |
| }, |
| { |
| "epoch": 31.5, |
| "grad_norm": 3.01706862449646, |
| "learning_rate": 2.391931822053251e-05, |
| "loss": 0.6192, |
| "step": 441 |
| }, |
| { |
| "epoch": 31.571428571428573, |
| "grad_norm": 7.216449737548828, |
| "learning_rate": 2.3910427215413036e-05, |
| "loss": 0.6928, |
| "step": 442 |
| }, |
| { |
| "epoch": 31.642857142857142, |
| "grad_norm": 3.636003255844116, |
| "learning_rate": 2.390150145215341e-05, |
| "loss": 0.6932, |
| "step": 443 |
| }, |
| { |
| "epoch": 31.714285714285715, |
| "grad_norm": 2.978515625, |
| "learning_rate": 2.3892540957943067e-05, |
| "loss": 0.6355, |
| "step": 444 |
| }, |
| { |
| "epoch": 31.785714285714285, |
| "grad_norm": 2.5752205848693848, |
| "learning_rate": 2.3883545760077215e-05, |
| "loss": 0.6208, |
| "step": 445 |
| }, |
| { |
| "epoch": 31.857142857142858, |
| "grad_norm": 2.671752452850342, |
| "learning_rate": 2.3874515885956792e-05, |
| "loss": 0.6078, |
| "step": 446 |
| }, |
| { |
| "epoch": 31.928571428571427, |
| "grad_norm": 3.273820400238037, |
| "learning_rate": 2.386545136308836e-05, |
| "loss": 0.6566, |
| "step": 447 |
| }, |
| { |
| "epoch": 32.0, |
| "grad_norm": 3.485921859741211, |
| "learning_rate": 2.3856352219084024e-05, |
| "loss": 0.606, |
| "step": 448 |
| }, |
| { |
| "epoch": 32.07142857142857, |
| "grad_norm": 2.268240451812744, |
| "learning_rate": 2.384721848166136e-05, |
| "loss": 0.5724, |
| "step": 449 |
| }, |
| { |
| "epoch": 32.142857142857146, |
| "grad_norm": 4.0276689529418945, |
| "learning_rate": 2.3838050178643312e-05, |
| "loss": 0.5929, |
| "step": 450 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 2000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 143, |
| "save_steps": 200, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": false, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 64, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|