diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,19633 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.2734375, + "eval_steps": 500, + "global_step": 2800, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 9.765625e-05, + "grad_norm": 13.125093460083008, + "learning_rate": 1.6666666666666669e-06, + "loss": 6.0473, + "step": 1 + }, + { + "epoch": 0.0001953125, + "grad_norm": 13.109691619873047, + "learning_rate": 3.3333333333333337e-06, + "loss": 6.0151, + "step": 2 + }, + { + "epoch": 0.00029296875, + "grad_norm": 12.975897789001465, + "learning_rate": 5e-06, + "loss": 6.0374, + "step": 3 + }, + { + "epoch": 0.000390625, + "grad_norm": 12.861470222473145, + "learning_rate": 6.6666666666666675e-06, + "loss": 5.9534, + "step": 4 + }, + { + "epoch": 0.00048828125, + "grad_norm": 11.551366806030273, + "learning_rate": 8.333333333333334e-06, + "loss": 5.7805, + "step": 5 + }, + { + "epoch": 0.0005859375, + "grad_norm": 7.968364238739014, + "learning_rate": 1e-05, + "loss": 5.5664, + "step": 6 + }, + { + "epoch": 0.00068359375, + "grad_norm": 6.854646682739258, + "learning_rate": 1.1666666666666668e-05, + "loss": 5.4982, + "step": 7 + }, + { + "epoch": 0.00078125, + "grad_norm": 3.759798049926758, + "learning_rate": 1.3333333333333335e-05, + "loss": 5.1824, + "step": 8 + }, + { + "epoch": 0.00087890625, + "grad_norm": 3.48822283744812, + "learning_rate": 1.5e-05, + "loss": 5.0667, + "step": 9 + }, + { + "epoch": 0.0009765625, + "grad_norm": 3.1011886596679688, + "learning_rate": 1.6666666666666667e-05, + "loss": 5.0028, + "step": 10 + }, + { + "epoch": 0.00107421875, + "grad_norm": 2.5868849754333496, + "learning_rate": 1.8333333333333333e-05, + "loss": 4.7787, + "step": 11 + }, + { + "epoch": 0.001171875, + "grad_norm": 2.378497838973999, + "learning_rate": 2e-05, + "loss": 4.6855, + "step": 12 + }, + { + "epoch": 0.00126953125, + "grad_norm": 2.2101738452911377, + "learning_rate": 2.1666666666666667e-05, + "loss": 4.6042, + "step": 13 + }, + { + "epoch": 0.0013671875, + "grad_norm": 1.9938961267471313, + "learning_rate": 2.3333333333333336e-05, + "loss": 4.3843, + "step": 14 + }, + { + "epoch": 0.00146484375, + "grad_norm": 1.9345463514328003, + "learning_rate": 2.5e-05, + "loss": 4.2548, + "step": 15 + }, + { + "epoch": 0.0015625, + "grad_norm": 1.7730776071548462, + "learning_rate": 2.666666666666667e-05, + "loss": 4.1363, + "step": 16 + }, + { + "epoch": 0.00166015625, + "grad_norm": 1.58334481716156, + "learning_rate": 2.8333333333333332e-05, + "loss": 3.9704, + "step": 17 + }, + { + "epoch": 0.0017578125, + "grad_norm": 1.4827001094818115, + "learning_rate": 3e-05, + "loss": 3.81, + "step": 18 + }, + { + "epoch": 0.00185546875, + "grad_norm": 1.4140430688858032, + "learning_rate": 3.166666666666667e-05, + "loss": 3.7305, + "step": 19 + }, + { + "epoch": 0.001953125, + "grad_norm": 1.3934496641159058, + "learning_rate": 3.3333333333333335e-05, + "loss": 3.5375, + "step": 20 + }, + { + "epoch": 0.00205078125, + "grad_norm": 1.3384771347045898, + "learning_rate": 3.5000000000000004e-05, + "loss": 3.5217, + "step": 21 + }, + { + "epoch": 0.0021484375, + "grad_norm": 1.2228978872299194, + "learning_rate": 3.6666666666666666e-05, + "loss": 3.3742, + "step": 22 + }, + { + "epoch": 0.00224609375, + "grad_norm": 1.1775028705596924, + "learning_rate": 3.8333333333333334e-05, + "loss": 3.2626, + "step": 23 + }, + { + "epoch": 0.00234375, + "grad_norm": 1.0886842012405396, + "learning_rate": 4e-05, + "loss": 3.1864, + "step": 24 + }, + { + "epoch": 0.00244140625, + "grad_norm": 1.0981535911560059, + "learning_rate": 4.1666666666666665e-05, + "loss": 3.0962, + "step": 25 + }, + { + "epoch": 0.0025390625, + "grad_norm": 1.009294867515564, + "learning_rate": 4.3333333333333334e-05, + "loss": 3.0507, + "step": 26 + }, + { + "epoch": 0.00263671875, + "grad_norm": 0.9390103816986084, + "learning_rate": 4.4999999999999996e-05, + "loss": 2.9579, + "step": 27 + }, + { + "epoch": 0.002734375, + "grad_norm": 0.8647847175598145, + "learning_rate": 4.666666666666667e-05, + "loss": 2.8166, + "step": 28 + }, + { + "epoch": 0.00283203125, + "grad_norm": 0.8606237769126892, + "learning_rate": 4.8333333333333334e-05, + "loss": 2.7919, + "step": 29 + }, + { + "epoch": 0.0029296875, + "grad_norm": 0.8069576025009155, + "learning_rate": 5e-05, + "loss": 2.7426, + "step": 30 + }, + { + "epoch": 0.00302734375, + "grad_norm": 0.7007808685302734, + "learning_rate": 5.1666666666666664e-05, + "loss": 2.6926, + "step": 31 + }, + { + "epoch": 0.003125, + "grad_norm": 0.6825646162033081, + "learning_rate": 5.333333333333334e-05, + "loss": 2.6783, + "step": 32 + }, + { + "epoch": 0.00322265625, + "grad_norm": 0.7054751515388489, + "learning_rate": 5.5e-05, + "loss": 2.6239, + "step": 33 + }, + { + "epoch": 0.0033203125, + "grad_norm": 0.7101700305938721, + "learning_rate": 5.6666666666666664e-05, + "loss": 2.5993, + "step": 34 + }, + { + "epoch": 0.00341796875, + "grad_norm": 0.56828773021698, + "learning_rate": 5.833333333333333e-05, + "loss": 2.5258, + "step": 35 + }, + { + "epoch": 0.003515625, + "grad_norm": 0.5513983368873596, + "learning_rate": 6e-05, + "loss": 2.5018, + "step": 36 + }, + { + "epoch": 0.00361328125, + "grad_norm": 0.573391318321228, + "learning_rate": 6.166666666666667e-05, + "loss": 2.4881, + "step": 37 + }, + { + "epoch": 0.0037109375, + "grad_norm": 0.4520920515060425, + "learning_rate": 6.333333333333335e-05, + "loss": 2.4489, + "step": 38 + }, + { + "epoch": 0.00380859375, + "grad_norm": 0.43038809299468994, + "learning_rate": 6.500000000000001e-05, + "loss": 2.4387, + "step": 39 + }, + { + "epoch": 0.00390625, + "grad_norm": 0.4071808457374573, + "learning_rate": 6.666666666666667e-05, + "loss": 2.4261, + "step": 40 + }, + { + "epoch": 0.00400390625, + "grad_norm": 0.3631410300731659, + "learning_rate": 6.833333333333333e-05, + "loss": 2.3656, + "step": 41 + }, + { + "epoch": 0.0041015625, + "grad_norm": 0.33457818627357483, + "learning_rate": 7.000000000000001e-05, + "loss": 2.4011, + "step": 42 + }, + { + "epoch": 0.00419921875, + "grad_norm": 0.3582305610179901, + "learning_rate": 7.166666666666667e-05, + "loss": 2.3662, + "step": 43 + }, + { + "epoch": 0.004296875, + "grad_norm": 0.4090467691421509, + "learning_rate": 7.333333333333333e-05, + "loss": 2.3282, + "step": 44 + }, + { + "epoch": 0.00439453125, + "grad_norm": 0.34438556432724, + "learning_rate": 7.5e-05, + "loss": 2.2991, + "step": 45 + }, + { + "epoch": 0.0044921875, + "grad_norm": 0.29381102323532104, + "learning_rate": 7.666666666666667e-05, + "loss": 2.2506, + "step": 46 + }, + { + "epoch": 0.00458984375, + "grad_norm": 0.4019562304019928, + "learning_rate": 7.833333333333334e-05, + "loss": 2.2494, + "step": 47 + }, + { + "epoch": 0.0046875, + "grad_norm": 0.3183911144733429, + "learning_rate": 8e-05, + "loss": 2.2853, + "step": 48 + }, + { + "epoch": 0.00478515625, + "grad_norm": 0.33041393756866455, + "learning_rate": 8.166666666666667e-05, + "loss": 2.2251, + "step": 49 + }, + { + "epoch": 0.0048828125, + "grad_norm": 0.28936123847961426, + "learning_rate": 8.333333333333333e-05, + "loss": 2.2308, + "step": 50 + }, + { + "epoch": 0.00498046875, + "grad_norm": 0.36125779151916504, + "learning_rate": 8.5e-05, + "loss": 2.2473, + "step": 51 + }, + { + "epoch": 0.005078125, + "grad_norm": 0.2832348942756653, + "learning_rate": 8.666666666666667e-05, + "loss": 2.201, + "step": 52 + }, + { + "epoch": 0.00517578125, + "grad_norm": 0.3250133693218231, + "learning_rate": 8.833333333333333e-05, + "loss": 2.1596, + "step": 53 + }, + { + "epoch": 0.0052734375, + "grad_norm": 0.31502828001976013, + "learning_rate": 8.999999999999999e-05, + "loss": 2.2183, + "step": 54 + }, + { + "epoch": 0.00537109375, + "grad_norm": 0.33068645000457764, + "learning_rate": 9.166666666666667e-05, + "loss": 2.2254, + "step": 55 + }, + { + "epoch": 0.00546875, + "grad_norm": 0.30684661865234375, + "learning_rate": 9.333333333333334e-05, + "loss": 2.2572, + "step": 56 + }, + { + "epoch": 0.00556640625, + "grad_norm": 0.31540846824645996, + "learning_rate": 9.5e-05, + "loss": 2.1594, + "step": 57 + }, + { + "epoch": 0.0056640625, + "grad_norm": 0.23259752988815308, + "learning_rate": 9.666666666666667e-05, + "loss": 2.172, + "step": 58 + }, + { + "epoch": 0.00576171875, + "grad_norm": 0.26666638255119324, + "learning_rate": 9.833333333333333e-05, + "loss": 2.1961, + "step": 59 + }, + { + "epoch": 0.005859375, + "grad_norm": 0.26295527815818787, + "learning_rate": 0.0001, + "loss": 2.1811, + "step": 60 + }, + { + "epoch": 0.00595703125, + "grad_norm": 0.2302207201719284, + "learning_rate": 0.00010166666666666667, + "loss": 2.1431, + "step": 61 + }, + { + "epoch": 0.0060546875, + "grad_norm": 0.468537837266922, + "learning_rate": 0.00010333333333333333, + "loss": 2.133, + "step": 62 + }, + { + "epoch": 0.00615234375, + "grad_norm": 0.27946797013282776, + "learning_rate": 0.000105, + "loss": 2.1917, + "step": 63 + }, + { + "epoch": 0.00625, + "grad_norm": 0.26789286732673645, + "learning_rate": 0.00010666666666666668, + "loss": 2.1417, + "step": 64 + }, + { + "epoch": 0.00634765625, + "grad_norm": 0.22977805137634277, + "learning_rate": 0.00010833333333333334, + "loss": 2.1371, + "step": 65 + }, + { + "epoch": 0.0064453125, + "grad_norm": 0.2590682804584503, + "learning_rate": 0.00011, + "loss": 2.113, + "step": 66 + }, + { + "epoch": 0.00654296875, + "grad_norm": 0.2660805583000183, + "learning_rate": 0.00011166666666666667, + "loss": 2.1346, + "step": 67 + }, + { + "epoch": 0.006640625, + "grad_norm": 0.24538874626159668, + "learning_rate": 0.00011333333333333333, + "loss": 2.1078, + "step": 68 + }, + { + "epoch": 0.00673828125, + "grad_norm": 0.2986001968383789, + "learning_rate": 0.000115, + "loss": 2.1048, + "step": 69 + }, + { + "epoch": 0.0068359375, + "grad_norm": 0.30193910002708435, + "learning_rate": 0.00011666666666666667, + "loss": 2.135, + "step": 70 + }, + { + "epoch": 0.00693359375, + "grad_norm": 0.25153177976608276, + "learning_rate": 0.00011833333333333334, + "loss": 2.1101, + "step": 71 + }, + { + "epoch": 0.00703125, + "grad_norm": 0.7685809135437012, + "learning_rate": 0.00012, + "loss": 2.0833, + "step": 72 + }, + { + "epoch": 0.00712890625, + "grad_norm": 0.24468126893043518, + "learning_rate": 0.00012166666666666668, + "loss": 2.1409, + "step": 73 + }, + { + "epoch": 0.0072265625, + "grad_norm": 0.3070752024650574, + "learning_rate": 0.00012333333333333334, + "loss": 2.1402, + "step": 74 + }, + { + "epoch": 0.00732421875, + "grad_norm": 0.352273553609848, + "learning_rate": 0.000125, + "loss": 2.171, + "step": 75 + }, + { + "epoch": 0.007421875, + "grad_norm": 0.34806111454963684, + "learning_rate": 0.0001266666666666667, + "loss": 2.1155, + "step": 76 + }, + { + "epoch": 0.00751953125, + "grad_norm": 0.317414253950119, + "learning_rate": 0.00012833333333333333, + "loss": 2.0812, + "step": 77 + }, + { + "epoch": 0.0076171875, + "grad_norm": 0.3320539891719818, + "learning_rate": 0.00013000000000000002, + "loss": 2.0987, + "step": 78 + }, + { + "epoch": 0.00771484375, + "grad_norm": 0.3315901756286621, + "learning_rate": 0.00013166666666666665, + "loss": 2.1025, + "step": 79 + }, + { + "epoch": 0.0078125, + "grad_norm": 0.29021286964416504, + "learning_rate": 0.00013333333333333334, + "loss": 2.1474, + "step": 80 + }, + { + "epoch": 0.00791015625, + "grad_norm": 0.2593044936656952, + "learning_rate": 0.000135, + "loss": 2.0667, + "step": 81 + }, + { + "epoch": 0.0080078125, + "grad_norm": 0.3192054033279419, + "learning_rate": 0.00013666666666666666, + "loss": 2.0764, + "step": 82 + }, + { + "epoch": 0.00810546875, + "grad_norm": 0.26053664088249207, + "learning_rate": 0.00013833333333333333, + "loss": 2.0698, + "step": 83 + }, + { + "epoch": 0.008203125, + "grad_norm": 0.24790963530540466, + "learning_rate": 0.00014000000000000001, + "loss": 2.124, + "step": 84 + }, + { + "epoch": 0.00830078125, + "grad_norm": 0.2546316087245941, + "learning_rate": 0.00014166666666666668, + "loss": 2.1025, + "step": 85 + }, + { + "epoch": 0.0083984375, + "grad_norm": 0.22275258600711823, + "learning_rate": 0.00014333333333333334, + "loss": 2.0778, + "step": 86 + }, + { + "epoch": 0.00849609375, + "grad_norm": 0.22855599224567413, + "learning_rate": 0.000145, + "loss": 2.1113, + "step": 87 + }, + { + "epoch": 0.00859375, + "grad_norm": 0.2456902116537094, + "learning_rate": 0.00014666666666666666, + "loss": 2.1399, + "step": 88 + }, + { + "epoch": 0.00869140625, + "grad_norm": 0.22144420444965363, + "learning_rate": 0.00014833333333333335, + "loss": 2.0211, + "step": 89 + }, + { + "epoch": 0.0087890625, + "grad_norm": 0.254894495010376, + "learning_rate": 0.00015, + "loss": 2.1382, + "step": 90 + }, + { + "epoch": 0.00888671875, + "grad_norm": 0.2729082703590393, + "learning_rate": 0.00015166666666666668, + "loss": 2.1271, + "step": 91 + }, + { + "epoch": 0.008984375, + "grad_norm": 0.2564642131328583, + "learning_rate": 0.00015333333333333334, + "loss": 2.0423, + "step": 92 + }, + { + "epoch": 0.00908203125, + "grad_norm": 0.24927419424057007, + "learning_rate": 0.000155, + "loss": 2.0603, + "step": 93 + }, + { + "epoch": 0.0091796875, + "grad_norm": 0.2591923475265503, + "learning_rate": 0.0001566666666666667, + "loss": 2.0522, + "step": 94 + }, + { + "epoch": 0.00927734375, + "grad_norm": 0.3350547254085541, + "learning_rate": 0.00015833333333333332, + "loss": 2.1095, + "step": 95 + }, + { + "epoch": 0.009375, + "grad_norm": 0.30227839946746826, + "learning_rate": 0.00016, + "loss": 2.121, + "step": 96 + }, + { + "epoch": 0.00947265625, + "grad_norm": 0.3027198314666748, + "learning_rate": 0.00016166666666666665, + "loss": 2.0288, + "step": 97 + }, + { + "epoch": 0.0095703125, + "grad_norm": 0.29791951179504395, + "learning_rate": 0.00016333333333333334, + "loss": 2.0471, + "step": 98 + }, + { + "epoch": 0.00966796875, + "grad_norm": 0.30918803811073303, + "learning_rate": 0.000165, + "loss": 2.1036, + "step": 99 + }, + { + "epoch": 0.009765625, + "grad_norm": 0.2510216534137726, + "learning_rate": 0.00016666666666666666, + "loss": 2.0217, + "step": 100 + }, + { + "epoch": 0.00986328125, + "grad_norm": 0.288231760263443, + "learning_rate": 0.00016833333333333335, + "loss": 2.0178, + "step": 101 + }, + { + "epoch": 0.0099609375, + "grad_norm": 0.3326691687107086, + "learning_rate": 0.00017, + "loss": 2.0947, + "step": 102 + }, + { + "epoch": 0.01005859375, + "grad_norm": 0.3552946150302887, + "learning_rate": 0.00017166666666666667, + "loss": 1.9953, + "step": 103 + }, + { + "epoch": 0.01015625, + "grad_norm": 0.34882935881614685, + "learning_rate": 0.00017333333333333334, + "loss": 2.1223, + "step": 104 + }, + { + "epoch": 0.01025390625, + "grad_norm": 0.35487979650497437, + "learning_rate": 0.000175, + "loss": 2.0599, + "step": 105 + }, + { + "epoch": 0.0103515625, + "grad_norm": 0.33561578392982483, + "learning_rate": 0.00017666666666666666, + "loss": 2.0398, + "step": 106 + }, + { + "epoch": 0.01044921875, + "grad_norm": 0.27415409684181213, + "learning_rate": 0.00017833333333333335, + "loss": 2.0982, + "step": 107 + }, + { + "epoch": 0.010546875, + "grad_norm": 0.28592920303344727, + "learning_rate": 0.00017999999999999998, + "loss": 2.0571, + "step": 108 + }, + { + "epoch": 0.01064453125, + "grad_norm": 0.3228552043437958, + "learning_rate": 0.00018166666666666667, + "loss": 1.9923, + "step": 109 + }, + { + "epoch": 0.0107421875, + "grad_norm": 0.29005661606788635, + "learning_rate": 0.00018333333333333334, + "loss": 2.0394, + "step": 110 + }, + { + "epoch": 0.01083984375, + "grad_norm": 0.36677825450897217, + "learning_rate": 0.000185, + "loss": 2.0153, + "step": 111 + }, + { + "epoch": 0.0109375, + "grad_norm": 0.2562806308269501, + "learning_rate": 0.0001866666666666667, + "loss": 2.05, + "step": 112 + }, + { + "epoch": 0.01103515625, + "grad_norm": 0.2748093008995056, + "learning_rate": 0.00018833333333333332, + "loss": 2.0449, + "step": 113 + }, + { + "epoch": 0.0111328125, + "grad_norm": 0.3924459218978882, + "learning_rate": 0.00019, + "loss": 2.0409, + "step": 114 + }, + { + "epoch": 0.01123046875, + "grad_norm": 0.4368191063404083, + "learning_rate": 0.00019166666666666667, + "loss": 2.0178, + "step": 115 + }, + { + "epoch": 0.011328125, + "grad_norm": 0.4236035943031311, + "learning_rate": 0.00019333333333333333, + "loss": 2.0518, + "step": 116 + }, + { + "epoch": 0.01142578125, + "grad_norm": 0.28912851214408875, + "learning_rate": 0.00019500000000000002, + "loss": 2.0058, + "step": 117 + }, + { + "epoch": 0.0115234375, + "grad_norm": 0.38847798109054565, + "learning_rate": 0.00019666666666666666, + "loss": 2.0818, + "step": 118 + }, + { + "epoch": 0.01162109375, + "grad_norm": 0.29429811239242554, + "learning_rate": 0.00019833333333333335, + "loss": 2.004, + "step": 119 + }, + { + "epoch": 0.01171875, + "grad_norm": 0.2969271242618561, + "learning_rate": 0.0002, + "loss": 2.046, + "step": 120 + }, + { + "epoch": 0.01181640625, + "grad_norm": 0.297894150018692, + "learning_rate": 0.00020166666666666667, + "loss": 2.0204, + "step": 121 + }, + { + "epoch": 0.0119140625, + "grad_norm": 0.28794237971305847, + "learning_rate": 0.00020333333333333333, + "loss": 2.0663, + "step": 122 + }, + { + "epoch": 0.01201171875, + "grad_norm": 0.2766033411026001, + "learning_rate": 0.000205, + "loss": 2.0399, + "step": 123 + }, + { + "epoch": 0.012109375, + "grad_norm": 0.3516612648963928, + "learning_rate": 0.00020666666666666666, + "loss": 2.0389, + "step": 124 + }, + { + "epoch": 0.01220703125, + "grad_norm": 0.28204381465911865, + "learning_rate": 0.00020833333333333335, + "loss": 2.0084, + "step": 125 + }, + { + "epoch": 0.0123046875, + "grad_norm": 0.3306240737438202, + "learning_rate": 0.00021, + "loss": 2.013, + "step": 126 + }, + { + "epoch": 0.01240234375, + "grad_norm": 0.3221111595630646, + "learning_rate": 0.00021166666666666667, + "loss": 2.0312, + "step": 127 + }, + { + "epoch": 0.0125, + "grad_norm": 0.2742249071598053, + "learning_rate": 0.00021333333333333336, + "loss": 2.0375, + "step": 128 + }, + { + "epoch": 0.01259765625, + "grad_norm": 0.2785228490829468, + "learning_rate": 0.000215, + "loss": 2.0449, + "step": 129 + }, + { + "epoch": 0.0126953125, + "grad_norm": 0.2666397988796234, + "learning_rate": 0.00021666666666666668, + "loss": 2.0355, + "step": 130 + }, + { + "epoch": 0.01279296875, + "grad_norm": 0.22522135078907013, + "learning_rate": 0.00021833333333333332, + "loss": 1.9693, + "step": 131 + }, + { + "epoch": 0.012890625, + "grad_norm": 0.2724483013153076, + "learning_rate": 0.00022, + "loss": 2.0671, + "step": 132 + }, + { + "epoch": 0.01298828125, + "grad_norm": 0.2655040919780731, + "learning_rate": 0.00022166666666666667, + "loss": 1.9805, + "step": 133 + }, + { + "epoch": 0.0130859375, + "grad_norm": 0.3194504678249359, + "learning_rate": 0.00022333333333333333, + "loss": 2.067, + "step": 134 + }, + { + "epoch": 0.01318359375, + "grad_norm": 0.25511813163757324, + "learning_rate": 0.00022500000000000002, + "loss": 2.0141, + "step": 135 + }, + { + "epoch": 0.01328125, + "grad_norm": 0.3995087444782257, + "learning_rate": 0.00022666666666666666, + "loss": 2.0278, + "step": 136 + }, + { + "epoch": 0.01337890625, + "grad_norm": 0.2623380422592163, + "learning_rate": 0.00022833333333333334, + "loss": 1.9751, + "step": 137 + }, + { + "epoch": 0.0134765625, + "grad_norm": 0.23814889788627625, + "learning_rate": 0.00023, + "loss": 1.969, + "step": 138 + }, + { + "epoch": 0.01357421875, + "grad_norm": 0.23566491901874542, + "learning_rate": 0.00023166666666666667, + "loss": 2.0135, + "step": 139 + }, + { + "epoch": 0.013671875, + "grad_norm": 0.2437373697757721, + "learning_rate": 0.00023333333333333333, + "loss": 2.0044, + "step": 140 + }, + { + "epoch": 0.01376953125, + "grad_norm": 0.2861543595790863, + "learning_rate": 0.000235, + "loss": 2.0378, + "step": 141 + }, + { + "epoch": 0.0138671875, + "grad_norm": 0.318050354719162, + "learning_rate": 0.00023666666666666668, + "loss": 2.0857, + "step": 142 + }, + { + "epoch": 0.01396484375, + "grad_norm": 0.39669227600097656, + "learning_rate": 0.00023833333333333334, + "loss": 2.0535, + "step": 143 + }, + { + "epoch": 0.0140625, + "grad_norm": 0.4359401762485504, + "learning_rate": 0.00024, + "loss": 2.0432, + "step": 144 + }, + { + "epoch": 0.01416015625, + "grad_norm": 0.4532039165496826, + "learning_rate": 0.00024166666666666667, + "loss": 1.9909, + "step": 145 + }, + { + "epoch": 0.0142578125, + "grad_norm": 0.4570695161819458, + "learning_rate": 0.00024333333333333336, + "loss": 2.0123, + "step": 146 + }, + { + "epoch": 0.01435546875, + "grad_norm": 0.36623403429985046, + "learning_rate": 0.000245, + "loss": 2.0179, + "step": 147 + }, + { + "epoch": 0.014453125, + "grad_norm": 0.3069714307785034, + "learning_rate": 0.0002466666666666667, + "loss": 2.0014, + "step": 148 + }, + { + "epoch": 0.01455078125, + "grad_norm": 0.3980304002761841, + "learning_rate": 0.0002483333333333333, + "loss": 2.0489, + "step": 149 + }, + { + "epoch": 0.0146484375, + "grad_norm": 0.31907564401626587, + "learning_rate": 0.00025, + "loss": 2.013, + "step": 150 + }, + { + "epoch": 0.01474609375, + "grad_norm": 0.2952549159526825, + "learning_rate": 0.00025166666666666664, + "loss": 2.0709, + "step": 151 + }, + { + "epoch": 0.01484375, + "grad_norm": 0.29451197385787964, + "learning_rate": 0.0002533333333333334, + "loss": 1.9613, + "step": 152 + }, + { + "epoch": 0.01494140625, + "grad_norm": 0.2893507778644562, + "learning_rate": 0.000255, + "loss": 2.0048, + "step": 153 + }, + { + "epoch": 0.0150390625, + "grad_norm": 0.24850639700889587, + "learning_rate": 0.00025666666666666665, + "loss": 2.0198, + "step": 154 + }, + { + "epoch": 0.01513671875, + "grad_norm": 0.24297639727592468, + "learning_rate": 0.00025833333333333334, + "loss": 2.0561, + "step": 155 + }, + { + "epoch": 0.015234375, + "grad_norm": 0.2777438461780548, + "learning_rate": 0.00026000000000000003, + "loss": 2.0206, + "step": 156 + }, + { + "epoch": 0.01533203125, + "grad_norm": 0.28714093565940857, + "learning_rate": 0.00026166666666666667, + "loss": 1.9891, + "step": 157 + }, + { + "epoch": 0.0154296875, + "grad_norm": 0.25796255469322205, + "learning_rate": 0.0002633333333333333, + "loss": 2.0369, + "step": 158 + }, + { + "epoch": 0.01552734375, + "grad_norm": 0.2391008883714676, + "learning_rate": 0.00026500000000000004, + "loss": 2.0015, + "step": 159 + }, + { + "epoch": 0.015625, + "grad_norm": 0.3203892409801483, + "learning_rate": 0.0002666666666666667, + "loss": 2.0213, + "step": 160 + }, + { + "epoch": 0.01572265625, + "grad_norm": 0.3396870791912079, + "learning_rate": 0.0002683333333333333, + "loss": 2.0333, + "step": 161 + }, + { + "epoch": 0.0158203125, + "grad_norm": 0.315060555934906, + "learning_rate": 0.00027, + "loss": 2.0195, + "step": 162 + }, + { + "epoch": 0.01591796875, + "grad_norm": 0.2672436535358429, + "learning_rate": 0.0002716666666666667, + "loss": 1.9946, + "step": 163 + }, + { + "epoch": 0.016015625, + "grad_norm": 0.2996402382850647, + "learning_rate": 0.00027333333333333333, + "loss": 2.0112, + "step": 164 + }, + { + "epoch": 0.01611328125, + "grad_norm": 0.2894189953804016, + "learning_rate": 0.000275, + "loss": 2.0157, + "step": 165 + }, + { + "epoch": 0.0162109375, + "grad_norm": 0.26241254806518555, + "learning_rate": 0.00027666666666666665, + "loss": 2.0177, + "step": 166 + }, + { + "epoch": 0.01630859375, + "grad_norm": 0.22900305688381195, + "learning_rate": 0.00027833333333333334, + "loss": 1.9834, + "step": 167 + }, + { + "epoch": 0.01640625, + "grad_norm": 0.2373427301645279, + "learning_rate": 0.00028000000000000003, + "loss": 1.9792, + "step": 168 + }, + { + "epoch": 0.01650390625, + "grad_norm": 0.2663004696369171, + "learning_rate": 0.00028166666666666666, + "loss": 2.0203, + "step": 169 + }, + { + "epoch": 0.0166015625, + "grad_norm": 0.31653544306755066, + "learning_rate": 0.00028333333333333335, + "loss": 2.0216, + "step": 170 + }, + { + "epoch": 0.01669921875, + "grad_norm": 0.3077234923839569, + "learning_rate": 0.000285, + "loss": 2.0361, + "step": 171 + }, + { + "epoch": 0.016796875, + "grad_norm": 0.25555703043937683, + "learning_rate": 0.0002866666666666667, + "loss": 2.0102, + "step": 172 + }, + { + "epoch": 0.01689453125, + "grad_norm": 0.29817435145378113, + "learning_rate": 0.0002883333333333333, + "loss": 1.972, + "step": 173 + }, + { + "epoch": 0.0169921875, + "grad_norm": 0.3075692355632782, + "learning_rate": 0.00029, + "loss": 2.0195, + "step": 174 + }, + { + "epoch": 0.01708984375, + "grad_norm": 0.29917964339256287, + "learning_rate": 0.0002916666666666667, + "loss": 1.9972, + "step": 175 + }, + { + "epoch": 0.0171875, + "grad_norm": 0.32018229365348816, + "learning_rate": 0.0002933333333333333, + "loss": 1.9895, + "step": 176 + }, + { + "epoch": 0.01728515625, + "grad_norm": 0.2907097339630127, + "learning_rate": 0.000295, + "loss": 1.9777, + "step": 177 + }, + { + "epoch": 0.0173828125, + "grad_norm": 0.22390642762184143, + "learning_rate": 0.0002966666666666667, + "loss": 1.9961, + "step": 178 + }, + { + "epoch": 0.01748046875, + "grad_norm": 0.25350186228752136, + "learning_rate": 0.00029833333333333334, + "loss": 1.9875, + "step": 179 + }, + { + "epoch": 0.017578125, + "grad_norm": 0.22856706380844116, + "learning_rate": 0.0003, + "loss": 2.0169, + "step": 180 + }, + { + "epoch": 0.01767578125, + "grad_norm": 0.2288493812084198, + "learning_rate": 0.0003016666666666667, + "loss": 2.0238, + "step": 181 + }, + { + "epoch": 0.0177734375, + "grad_norm": 0.27326855063438416, + "learning_rate": 0.00030333333333333335, + "loss": 2.0134, + "step": 182 + }, + { + "epoch": 0.01787109375, + "grad_norm": 0.2447524517774582, + "learning_rate": 0.000305, + "loss": 1.9852, + "step": 183 + }, + { + "epoch": 0.01796875, + "grad_norm": 0.4363366663455963, + "learning_rate": 0.0003066666666666667, + "loss": 1.9921, + "step": 184 + }, + { + "epoch": 0.01806640625, + "grad_norm": 0.4569666385650635, + "learning_rate": 0.00030833333333333337, + "loss": 1.9997, + "step": 185 + }, + { + "epoch": 0.0181640625, + "grad_norm": 0.43348655104637146, + "learning_rate": 0.00031, + "loss": 2.0584, + "step": 186 + }, + { + "epoch": 0.01826171875, + "grad_norm": 0.3844921588897705, + "learning_rate": 0.00031166666666666663, + "loss": 2.0035, + "step": 187 + }, + { + "epoch": 0.018359375, + "grad_norm": 0.3427641987800598, + "learning_rate": 0.0003133333333333334, + "loss": 1.9536, + "step": 188 + }, + { + "epoch": 0.01845703125, + "grad_norm": 0.33557865023612976, + "learning_rate": 0.000315, + "loss": 1.972, + "step": 189 + }, + { + "epoch": 0.0185546875, + "grad_norm": 0.4006612300872803, + "learning_rate": 0.00031666666666666665, + "loss": 2.0652, + "step": 190 + }, + { + "epoch": 0.01865234375, + "grad_norm": 0.3158099055290222, + "learning_rate": 0.00031833333333333334, + "loss": 2.0516, + "step": 191 + }, + { + "epoch": 0.01875, + "grad_norm": 0.3799190819263458, + "learning_rate": 0.00032, + "loss": 2.011, + "step": 192 + }, + { + "epoch": 0.01884765625, + "grad_norm": 0.2948876619338989, + "learning_rate": 0.00032166666666666666, + "loss": 2.0109, + "step": 193 + }, + { + "epoch": 0.0189453125, + "grad_norm": 0.24561335146427155, + "learning_rate": 0.0003233333333333333, + "loss": 2.0264, + "step": 194 + }, + { + "epoch": 0.01904296875, + "grad_norm": 0.24896866083145142, + "learning_rate": 0.00032500000000000004, + "loss": 1.9726, + "step": 195 + }, + { + "epoch": 0.019140625, + "grad_norm": 0.26887547969818115, + "learning_rate": 0.0003266666666666667, + "loss": 2.0036, + "step": 196 + }, + { + "epoch": 0.01923828125, + "grad_norm": 0.3186735212802887, + "learning_rate": 0.0003283333333333333, + "loss": 2.0174, + "step": 197 + }, + { + "epoch": 0.0193359375, + "grad_norm": 0.3317165672779083, + "learning_rate": 0.00033, + "loss": 2.0047, + "step": 198 + }, + { + "epoch": 0.01943359375, + "grad_norm": 0.3068574070930481, + "learning_rate": 0.0003316666666666667, + "loss": 2.0032, + "step": 199 + }, + { + "epoch": 0.01953125, + "grad_norm": 0.29292526841163635, + "learning_rate": 0.0003333333333333333, + "loss": 2.0533, + "step": 200 + }, + { + "epoch": 0.01962890625, + "grad_norm": 0.2519834041595459, + "learning_rate": 0.000335, + "loss": 2.0113, + "step": 201 + }, + { + "epoch": 0.0197265625, + "grad_norm": 0.25766584277153015, + "learning_rate": 0.0003366666666666667, + "loss": 2.0278, + "step": 202 + }, + { + "epoch": 0.01982421875, + "grad_norm": 0.2704983055591583, + "learning_rate": 0.00033833333333333334, + "loss": 1.9725, + "step": 203 + }, + { + "epoch": 0.019921875, + "grad_norm": 0.2882053256034851, + "learning_rate": 0.00034, + "loss": 2.0706, + "step": 204 + }, + { + "epoch": 0.02001953125, + "grad_norm": 0.34524375200271606, + "learning_rate": 0.00034166666666666666, + "loss": 1.9868, + "step": 205 + }, + { + "epoch": 0.0201171875, + "grad_norm": 0.3718552887439728, + "learning_rate": 0.00034333333333333335, + "loss": 1.9596, + "step": 206 + }, + { + "epoch": 0.02021484375, + "grad_norm": 0.2747247815132141, + "learning_rate": 0.000345, + "loss": 2.0125, + "step": 207 + }, + { + "epoch": 0.0203125, + "grad_norm": 0.3062858283519745, + "learning_rate": 0.00034666666666666667, + "loss": 2.0224, + "step": 208 + }, + { + "epoch": 0.02041015625, + "grad_norm": 0.32505863904953003, + "learning_rate": 0.00034833333333333336, + "loss": 2.0376, + "step": 209 + }, + { + "epoch": 0.0205078125, + "grad_norm": 0.35048386454582214, + "learning_rate": 0.00035, + "loss": 2.0268, + "step": 210 + }, + { + "epoch": 0.02060546875, + "grad_norm": 0.31204426288604736, + "learning_rate": 0.0003516666666666667, + "loss": 2.0198, + "step": 211 + }, + { + "epoch": 0.020703125, + "grad_norm": 0.24253524839878082, + "learning_rate": 0.0003533333333333333, + "loss": 2.028, + "step": 212 + }, + { + "epoch": 0.02080078125, + "grad_norm": 0.286915123462677, + "learning_rate": 0.000355, + "loss": 1.9358, + "step": 213 + }, + { + "epoch": 0.0208984375, + "grad_norm": 0.2800680994987488, + "learning_rate": 0.0003566666666666667, + "loss": 1.9905, + "step": 214 + }, + { + "epoch": 0.02099609375, + "grad_norm": 0.2718358635902405, + "learning_rate": 0.00035833333333333333, + "loss": 1.9902, + "step": 215 + }, + { + "epoch": 0.02109375, + "grad_norm": 0.28583604097366333, + "learning_rate": 0.00035999999999999997, + "loss": 1.991, + "step": 216 + }, + { + "epoch": 0.02119140625, + "grad_norm": 0.2911478281021118, + "learning_rate": 0.0003616666666666667, + "loss": 1.973, + "step": 217 + }, + { + "epoch": 0.0212890625, + "grad_norm": 0.3601188361644745, + "learning_rate": 0.00036333333333333335, + "loss": 1.9727, + "step": 218 + }, + { + "epoch": 0.02138671875, + "grad_norm": 0.2888337969779968, + "learning_rate": 0.000365, + "loss": 1.988, + "step": 219 + }, + { + "epoch": 0.021484375, + "grad_norm": 0.25628700852394104, + "learning_rate": 0.00036666666666666667, + "loss": 1.9984, + "step": 220 + }, + { + "epoch": 0.02158203125, + "grad_norm": 0.2637641429901123, + "learning_rate": 0.00036833333333333336, + "loss": 2.0229, + "step": 221 + }, + { + "epoch": 0.0216796875, + "grad_norm": 0.23845899105072021, + "learning_rate": 0.00037, + "loss": 1.9985, + "step": 222 + }, + { + "epoch": 0.02177734375, + "grad_norm": 0.28519535064697266, + "learning_rate": 0.00037166666666666663, + "loss": 2.0061, + "step": 223 + }, + { + "epoch": 0.021875, + "grad_norm": 0.31845173239707947, + "learning_rate": 0.0003733333333333334, + "loss": 2.0081, + "step": 224 + }, + { + "epoch": 0.02197265625, + "grad_norm": 0.3725838363170624, + "learning_rate": 0.000375, + "loss": 2.0032, + "step": 225 + }, + { + "epoch": 0.0220703125, + "grad_norm": 0.49783870577812195, + "learning_rate": 0.00037666666666666664, + "loss": 2.0404, + "step": 226 + }, + { + "epoch": 0.02216796875, + "grad_norm": 0.5059479475021362, + "learning_rate": 0.0003783333333333334, + "loss": 2.0498, + "step": 227 + }, + { + "epoch": 0.022265625, + "grad_norm": 0.461291640996933, + "learning_rate": 0.00038, + "loss": 2.0078, + "step": 228 + }, + { + "epoch": 0.02236328125, + "grad_norm": 0.3970203399658203, + "learning_rate": 0.00038166666666666666, + "loss": 1.9966, + "step": 229 + }, + { + "epoch": 0.0224609375, + "grad_norm": 0.3155679404735565, + "learning_rate": 0.00038333333333333334, + "loss": 2.0263, + "step": 230 + }, + { + "epoch": 0.02255859375, + "grad_norm": 0.32979920506477356, + "learning_rate": 0.00038500000000000003, + "loss": 1.9967, + "step": 231 + }, + { + "epoch": 0.02265625, + "grad_norm": 0.27470117807388306, + "learning_rate": 0.00038666666666666667, + "loss": 2.0461, + "step": 232 + }, + { + "epoch": 0.02275390625, + "grad_norm": 0.2981088161468506, + "learning_rate": 0.0003883333333333333, + "loss": 1.9944, + "step": 233 + }, + { + "epoch": 0.0228515625, + "grad_norm": 0.3496599495410919, + "learning_rate": 0.00039000000000000005, + "loss": 1.9907, + "step": 234 + }, + { + "epoch": 0.02294921875, + "grad_norm": 0.3318106532096863, + "learning_rate": 0.0003916666666666667, + "loss": 2.0576, + "step": 235 + }, + { + "epoch": 0.023046875, + "grad_norm": 0.29498377442359924, + "learning_rate": 0.0003933333333333333, + "loss": 2.0242, + "step": 236 + }, + { + "epoch": 0.02314453125, + "grad_norm": 0.2970214784145355, + "learning_rate": 0.000395, + "loss": 2.0087, + "step": 237 + }, + { + "epoch": 0.0232421875, + "grad_norm": 0.37431418895721436, + "learning_rate": 0.0003966666666666667, + "loss": 2.0657, + "step": 238 + }, + { + "epoch": 0.02333984375, + "grad_norm": 0.30095174908638, + "learning_rate": 0.00039833333333333333, + "loss": 2.0217, + "step": 239 + }, + { + "epoch": 0.0234375, + "grad_norm": 0.24695053696632385, + "learning_rate": 0.0004, + "loss": 1.9833, + "step": 240 + }, + { + "epoch": 0.02353515625, + "grad_norm": 0.2923540771007538, + "learning_rate": 0.00040166666666666665, + "loss": 2.0272, + "step": 241 + }, + { + "epoch": 0.0236328125, + "grad_norm": 0.2788209915161133, + "learning_rate": 0.00040333333333333334, + "loss": 2.0104, + "step": 242 + }, + { + "epoch": 0.02373046875, + "grad_norm": 0.2529614567756653, + "learning_rate": 0.00040500000000000003, + "loss": 2.003, + "step": 243 + }, + { + "epoch": 0.023828125, + "grad_norm": 0.2551966905593872, + "learning_rate": 0.00040666666666666667, + "loss": 2.001, + "step": 244 + }, + { + "epoch": 0.02392578125, + "grad_norm": 0.2613292634487152, + "learning_rate": 0.00040833333333333336, + "loss": 1.9822, + "step": 245 + }, + { + "epoch": 0.0240234375, + "grad_norm": 0.3060430884361267, + "learning_rate": 0.00041, + "loss": 2.0024, + "step": 246 + }, + { + "epoch": 0.02412109375, + "grad_norm": 0.33755916357040405, + "learning_rate": 0.0004116666666666667, + "loss": 2.0023, + "step": 247 + }, + { + "epoch": 0.02421875, + "grad_norm": 0.33021774888038635, + "learning_rate": 0.0004133333333333333, + "loss": 1.9086, + "step": 248 + }, + { + "epoch": 0.02431640625, + "grad_norm": 0.26662060618400574, + "learning_rate": 0.000415, + "loss": 2.009, + "step": 249 + }, + { + "epoch": 0.0244140625, + "grad_norm": 0.27698251605033875, + "learning_rate": 0.0004166666666666667, + "loss": 2.0183, + "step": 250 + }, + { + "epoch": 0.02451171875, + "grad_norm": 0.2582184970378876, + "learning_rate": 0.00041833333333333333, + "loss": 1.9932, + "step": 251 + }, + { + "epoch": 0.024609375, + "grad_norm": 0.28684699535369873, + "learning_rate": 0.00042, + "loss": 2.0021, + "step": 252 + }, + { + "epoch": 0.02470703125, + "grad_norm": 0.33535540103912354, + "learning_rate": 0.0004216666666666667, + "loss": 2.0387, + "step": 253 + }, + { + "epoch": 0.0248046875, + "grad_norm": 0.3330588638782501, + "learning_rate": 0.00042333333333333334, + "loss": 1.9776, + "step": 254 + }, + { + "epoch": 0.02490234375, + "grad_norm": 0.27919256687164307, + "learning_rate": 0.000425, + "loss": 2.0111, + "step": 255 + }, + { + "epoch": 0.025, + "grad_norm": 0.25296416878700256, + "learning_rate": 0.0004266666666666667, + "loss": 1.9755, + "step": 256 + }, + { + "epoch": 0.02509765625, + "grad_norm": 0.31288138031959534, + "learning_rate": 0.00042833333333333335, + "loss": 2.032, + "step": 257 + }, + { + "epoch": 0.0251953125, + "grad_norm": 0.3360923230648041, + "learning_rate": 0.00043, + "loss": 2.0837, + "step": 258 + }, + { + "epoch": 0.02529296875, + "grad_norm": 0.36317816376686096, + "learning_rate": 0.0004316666666666667, + "loss": 1.9696, + "step": 259 + }, + { + "epoch": 0.025390625, + "grad_norm": 0.366953581571579, + "learning_rate": 0.00043333333333333337, + "loss": 2.031, + "step": 260 + }, + { + "epoch": 0.02548828125, + "grad_norm": 0.34289368987083435, + "learning_rate": 0.000435, + "loss": 1.9968, + "step": 261 + }, + { + "epoch": 0.0255859375, + "grad_norm": 0.35170793533325195, + "learning_rate": 0.00043666666666666664, + "loss": 1.9963, + "step": 262 + }, + { + "epoch": 0.02568359375, + "grad_norm": 0.28625521063804626, + "learning_rate": 0.0004383333333333334, + "loss": 1.9932, + "step": 263 + }, + { + "epoch": 0.02578125, + "grad_norm": 0.2861610949039459, + "learning_rate": 0.00044, + "loss": 2.0297, + "step": 264 + }, + { + "epoch": 0.02587890625, + "grad_norm": 0.30467647314071655, + "learning_rate": 0.00044166666666666665, + "loss": 1.992, + "step": 265 + }, + { + "epoch": 0.0259765625, + "grad_norm": 0.31711357831954956, + "learning_rate": 0.00044333333333333334, + "loss": 2.0135, + "step": 266 + }, + { + "epoch": 0.02607421875, + "grad_norm": 0.341530978679657, + "learning_rate": 0.00044500000000000003, + "loss": 1.983, + "step": 267 + }, + { + "epoch": 0.026171875, + "grad_norm": 0.373901903629303, + "learning_rate": 0.00044666666666666666, + "loss": 2.0048, + "step": 268 + }, + { + "epoch": 0.02626953125, + "grad_norm": 0.3105134665966034, + "learning_rate": 0.0004483333333333333, + "loss": 2.0368, + "step": 269 + }, + { + "epoch": 0.0263671875, + "grad_norm": 0.29363134503364563, + "learning_rate": 0.00045000000000000004, + "loss": 1.978, + "step": 270 + }, + { + "epoch": 0.02646484375, + "grad_norm": 0.3060167133808136, + "learning_rate": 0.0004516666666666667, + "loss": 1.9479, + "step": 271 + }, + { + "epoch": 0.0265625, + "grad_norm": 0.30803290009498596, + "learning_rate": 0.0004533333333333333, + "loss": 1.9662, + "step": 272 + }, + { + "epoch": 0.02666015625, + "grad_norm": 0.3324045240879059, + "learning_rate": 0.000455, + "loss": 2.0299, + "step": 273 + }, + { + "epoch": 0.0267578125, + "grad_norm": 0.39051148295402527, + "learning_rate": 0.0004566666666666667, + "loss": 1.9856, + "step": 274 + }, + { + "epoch": 0.02685546875, + "grad_norm": 0.4288715124130249, + "learning_rate": 0.0004583333333333333, + "loss": 2.0264, + "step": 275 + }, + { + "epoch": 0.026953125, + "grad_norm": 0.34478962421417236, + "learning_rate": 0.00046, + "loss": 1.9824, + "step": 276 + }, + { + "epoch": 0.02705078125, + "grad_norm": 0.2766290009021759, + "learning_rate": 0.0004616666666666667, + "loss": 2.0066, + "step": 277 + }, + { + "epoch": 0.0271484375, + "grad_norm": 0.2508682608604431, + "learning_rate": 0.00046333333333333334, + "loss": 1.9663, + "step": 278 + }, + { + "epoch": 0.02724609375, + "grad_norm": 0.26924827694892883, + "learning_rate": 0.000465, + "loss": 1.9903, + "step": 279 + }, + { + "epoch": 0.02734375, + "grad_norm": 0.27668496966362, + "learning_rate": 0.00046666666666666666, + "loss": 2.0097, + "step": 280 + }, + { + "epoch": 0.02744140625, + "grad_norm": 0.25026220083236694, + "learning_rate": 0.00046833333333333335, + "loss": 2.0583, + "step": 281 + }, + { + "epoch": 0.0275390625, + "grad_norm": 0.2158055454492569, + "learning_rate": 0.00047, + "loss": 2.0137, + "step": 282 + }, + { + "epoch": 0.02763671875, + "grad_norm": 0.22540244460105896, + "learning_rate": 0.0004716666666666667, + "loss": 1.994, + "step": 283 + }, + { + "epoch": 0.027734375, + "grad_norm": 0.26405519247055054, + "learning_rate": 0.00047333333333333336, + "loss": 2.0221, + "step": 284 + }, + { + "epoch": 0.02783203125, + "grad_norm": 0.2979099452495575, + "learning_rate": 0.000475, + "loss": 2.0047, + "step": 285 + }, + { + "epoch": 0.0279296875, + "grad_norm": 0.34131935238838196, + "learning_rate": 0.0004766666666666667, + "loss": 1.9907, + "step": 286 + }, + { + "epoch": 0.02802734375, + "grad_norm": 0.37178686261177063, + "learning_rate": 0.0004783333333333333, + "loss": 1.9806, + "step": 287 + }, + { + "epoch": 0.028125, + "grad_norm": 0.36835598945617676, + "learning_rate": 0.00048, + "loss": 2.0134, + "step": 288 + }, + { + "epoch": 0.02822265625, + "grad_norm": 0.29690125584602356, + "learning_rate": 0.0004816666666666667, + "loss": 2.0261, + "step": 289 + }, + { + "epoch": 0.0283203125, + "grad_norm": 0.2690771818161011, + "learning_rate": 0.00048333333333333334, + "loss": 1.9718, + "step": 290 + }, + { + "epoch": 0.02841796875, + "grad_norm": 0.3377201557159424, + "learning_rate": 0.00048499999999999997, + "loss": 1.99, + "step": 291 + }, + { + "epoch": 0.028515625, + "grad_norm": 0.34973010420799255, + "learning_rate": 0.0004866666666666667, + "loss": 1.9721, + "step": 292 + }, + { + "epoch": 0.02861328125, + "grad_norm": 0.3172457218170166, + "learning_rate": 0.0004883333333333333, + "loss": 1.9928, + "step": 293 + }, + { + "epoch": 0.0287109375, + "grad_norm": 0.34357598423957825, + "learning_rate": 0.00049, + "loss": 1.9995, + "step": 294 + }, + { + "epoch": 0.02880859375, + "grad_norm": 0.3824540376663208, + "learning_rate": 0.0004916666666666666, + "loss": 1.9772, + "step": 295 + }, + { + "epoch": 0.02890625, + "grad_norm": 0.3704535663127899, + "learning_rate": 0.0004933333333333334, + "loss": 1.9829, + "step": 296 + }, + { + "epoch": 0.02900390625, + "grad_norm": 0.2571757733821869, + "learning_rate": 0.000495, + "loss": 1.9966, + "step": 297 + }, + { + "epoch": 0.0291015625, + "grad_norm": 0.3970927894115448, + "learning_rate": 0.0004966666666666666, + "loss": 1.9841, + "step": 298 + }, + { + "epoch": 0.02919921875, + "grad_norm": 0.3420144319534302, + "learning_rate": 0.0004983333333333334, + "loss": 1.9747, + "step": 299 + }, + { + "epoch": 0.029296875, + "grad_norm": 0.30147823691368103, + "learning_rate": 0.0005, + "loss": 1.9999, + "step": 300 + }, + { + "epoch": 0.02939453125, + "grad_norm": 0.33727970719337463, + "learning_rate": 0.0004999999887622467, + "loss": 2.0084, + "step": 301 + }, + { + "epoch": 0.0294921875, + "grad_norm": 0.29407384991645813, + "learning_rate": 0.0004999999550489878, + "loss": 2.0144, + "step": 302 + }, + { + "epoch": 0.02958984375, + "grad_norm": 0.3489755094051361, + "learning_rate": 0.0004999998988602267, + "loss": 2.0058, + "step": 303 + }, + { + "epoch": 0.0296875, + "grad_norm": 0.3327770233154297, + "learning_rate": 0.0004999998201959691, + "loss": 2.0166, + "step": 304 + }, + { + "epoch": 0.02978515625, + "grad_norm": 0.2923370599746704, + "learning_rate": 0.0004999997190562227, + "loss": 2.0148, + "step": 305 + }, + { + "epoch": 0.0298828125, + "grad_norm": 0.31616437435150146, + "learning_rate": 0.0004999995954409976, + "loss": 1.9772, + "step": 306 + }, + { + "epoch": 0.02998046875, + "grad_norm": 0.22982288897037506, + "learning_rate": 0.0004999994493503064, + "loss": 2.0584, + "step": 307 + }, + { + "epoch": 0.030078125, + "grad_norm": 0.2886744737625122, + "learning_rate": 0.0004999992807841634, + "loss": 2.0114, + "step": 308 + }, + { + "epoch": 0.03017578125, + "grad_norm": 0.3027271032333374, + "learning_rate": 0.0004999990897425856, + "loss": 2.011, + "step": 309 + }, + { + "epoch": 0.0302734375, + "grad_norm": 0.3191162645816803, + "learning_rate": 0.0004999988762255922, + "loss": 1.9962, + "step": 310 + }, + { + "epoch": 0.03037109375, + "grad_norm": 0.34986981749534607, + "learning_rate": 0.0004999986402332042, + "loss": 1.9612, + "step": 311 + }, + { + "epoch": 0.03046875, + "grad_norm": 0.36431390047073364, + "learning_rate": 0.0004999983817654454, + "loss": 2.045, + "step": 312 + }, + { + "epoch": 0.03056640625, + "grad_norm": 0.4198042154312134, + "learning_rate": 0.0004999981008223416, + "loss": 2.0132, + "step": 313 + }, + { + "epoch": 0.0306640625, + "grad_norm": 0.43374890089035034, + "learning_rate": 0.0004999977974039207, + "loss": 2.0578, + "step": 314 + }, + { + "epoch": 0.03076171875, + "grad_norm": 0.3654812276363373, + "learning_rate": 0.0004999974715102132, + "loss": 1.9721, + "step": 315 + }, + { + "epoch": 0.030859375, + "grad_norm": 0.29420921206474304, + "learning_rate": 0.0004999971231412517, + "loss": 2.0296, + "step": 316 + }, + { + "epoch": 0.03095703125, + "grad_norm": 0.31979072093963623, + "learning_rate": 0.0004999967522970708, + "loss": 1.9623, + "step": 317 + }, + { + "epoch": 0.0310546875, + "grad_norm": 0.3570129871368408, + "learning_rate": 0.0004999963589777076, + "loss": 1.9966, + "step": 318 + }, + { + "epoch": 0.03115234375, + "grad_norm": 0.29743143916130066, + "learning_rate": 0.0004999959431832016, + "loss": 1.9535, + "step": 319 + }, + { + "epoch": 0.03125, + "grad_norm": 0.23641493916511536, + "learning_rate": 0.000499995504913594, + "loss": 2.0304, + "step": 320 + }, + { + "epoch": 0.03134765625, + "grad_norm": 0.3556622564792633, + "learning_rate": 0.0004999950441689288, + "loss": 1.971, + "step": 321 + }, + { + "epoch": 0.0314453125, + "grad_norm": 0.323939710855484, + "learning_rate": 0.0004999945609492519, + "loss": 2.0153, + "step": 322 + }, + { + "epoch": 0.03154296875, + "grad_norm": 0.22797244787216187, + "learning_rate": 0.0004999940552546118, + "loss": 1.9807, + "step": 323 + }, + { + "epoch": 0.031640625, + "grad_norm": 0.2641647458076477, + "learning_rate": 0.0004999935270850587, + "loss": 1.9988, + "step": 324 + }, + { + "epoch": 0.03173828125, + "grad_norm": 0.25289344787597656, + "learning_rate": 0.0004999929764406455, + "loss": 1.9748, + "step": 325 + }, + { + "epoch": 0.0318359375, + "grad_norm": 0.2377796769142151, + "learning_rate": 0.0004999924033214274, + "loss": 1.9983, + "step": 326 + }, + { + "epoch": 0.03193359375, + "grad_norm": 0.2711915671825409, + "learning_rate": 0.0004999918077274612, + "loss": 1.9643, + "step": 327 + }, + { + "epoch": 0.03203125, + "grad_norm": 0.2866462767124176, + "learning_rate": 0.0004999911896588068, + "loss": 2.0196, + "step": 328 + }, + { + "epoch": 0.03212890625, + "grad_norm": 0.25075578689575195, + "learning_rate": 0.0004999905491155257, + "loss": 2.0426, + "step": 329 + }, + { + "epoch": 0.0322265625, + "grad_norm": 0.266648530960083, + "learning_rate": 0.000499989886097682, + "loss": 2.0318, + "step": 330 + }, + { + "epoch": 0.03232421875, + "grad_norm": 0.22494247555732727, + "learning_rate": 0.0004999892006053421, + "loss": 1.9839, + "step": 331 + }, + { + "epoch": 0.032421875, + "grad_norm": 0.3179854452610016, + "learning_rate": 0.0004999884926385741, + "loss": 1.9981, + "step": 332 + }, + { + "epoch": 0.03251953125, + "grad_norm": 0.2754990756511688, + "learning_rate": 0.000499987762197449, + "loss": 1.9879, + "step": 333 + }, + { + "epoch": 0.0326171875, + "grad_norm": 0.2669137418270111, + "learning_rate": 0.0004999870092820395, + "loss": 2.0083, + "step": 334 + }, + { + "epoch": 0.03271484375, + "grad_norm": 0.31376007199287415, + "learning_rate": 0.0004999862338924212, + "loss": 2.0166, + "step": 335 + }, + { + "epoch": 0.0328125, + "grad_norm": 0.32645899057388306, + "learning_rate": 0.0004999854360286712, + "loss": 2.0019, + "step": 336 + }, + { + "epoch": 0.03291015625, + "grad_norm": 0.3550071120262146, + "learning_rate": 0.0004999846156908692, + "loss": 1.9744, + "step": 337 + }, + { + "epoch": 0.0330078125, + "grad_norm": 0.30171769857406616, + "learning_rate": 0.0004999837728790975, + "loss": 2.0231, + "step": 338 + }, + { + "epoch": 0.03310546875, + "grad_norm": 0.23128142952919006, + "learning_rate": 0.00049998290759344, + "loss": 2.001, + "step": 339 + }, + { + "epoch": 0.033203125, + "grad_norm": 0.2433364987373352, + "learning_rate": 0.0004999820198339832, + "loss": 1.9945, + "step": 340 + }, + { + "epoch": 0.03330078125, + "grad_norm": 0.30562201142311096, + "learning_rate": 0.0004999811096008159, + "loss": 2.0228, + "step": 341 + }, + { + "epoch": 0.0333984375, + "grad_norm": 0.33020487427711487, + "learning_rate": 0.0004999801768940287, + "loss": 1.9691, + "step": 342 + }, + { + "epoch": 0.03349609375, + "grad_norm": 0.3125375807285309, + "learning_rate": 0.0004999792217137151, + "loss": 1.9728, + "step": 343 + }, + { + "epoch": 0.03359375, + "grad_norm": 0.26769348978996277, + "learning_rate": 0.0004999782440599702, + "loss": 2.0066, + "step": 344 + }, + { + "epoch": 0.03369140625, + "grad_norm": 0.3596431314945221, + "learning_rate": 0.0004999772439328921, + "loss": 2.0426, + "step": 345 + }, + { + "epoch": 0.0337890625, + "grad_norm": 0.34117281436920166, + "learning_rate": 0.0004999762213325803, + "loss": 1.9985, + "step": 346 + }, + { + "epoch": 0.03388671875, + "grad_norm": 0.2922564148902893, + "learning_rate": 0.0004999751762591371, + "loss": 2.0346, + "step": 347 + }, + { + "epoch": 0.033984375, + "grad_norm": 0.2813419997692108, + "learning_rate": 0.0004999741087126669, + "loss": 2.0573, + "step": 348 + }, + { + "epoch": 0.03408203125, + "grad_norm": 0.27346089482307434, + "learning_rate": 0.0004999730186932764, + "loss": 1.9655, + "step": 349 + }, + { + "epoch": 0.0341796875, + "grad_norm": 0.24704065918922424, + "learning_rate": 0.0004999719062010745, + "loss": 1.9542, + "step": 350 + }, + { + "epoch": 0.03427734375, + "grad_norm": 0.24761976301670074, + "learning_rate": 0.000499970771236172, + "loss": 1.9526, + "step": 351 + }, + { + "epoch": 0.034375, + "grad_norm": 0.2508860230445862, + "learning_rate": 0.0004999696137986826, + "loss": 2.0119, + "step": 352 + }, + { + "epoch": 0.03447265625, + "grad_norm": 0.24160990118980408, + "learning_rate": 0.0004999684338887219, + "loss": 2.0148, + "step": 353 + }, + { + "epoch": 0.0345703125, + "grad_norm": 0.2879098355770111, + "learning_rate": 0.0004999672315064076, + "loss": 2.0206, + "step": 354 + }, + { + "epoch": 0.03466796875, + "grad_norm": 0.28366580605506897, + "learning_rate": 0.0004999660066518601, + "loss": 2.0737, + "step": 355 + }, + { + "epoch": 0.034765625, + "grad_norm": 0.3401612937450409, + "learning_rate": 0.0004999647593252013, + "loss": 1.9868, + "step": 356 + }, + { + "epoch": 0.03486328125, + "grad_norm": 0.4696269631385803, + "learning_rate": 0.0004999634895265562, + "loss": 2.0079, + "step": 357 + }, + { + "epoch": 0.0349609375, + "grad_norm": 0.596168041229248, + "learning_rate": 0.0004999621972560515, + "loss": 2.0084, + "step": 358 + }, + { + "epoch": 0.03505859375, + "grad_norm": 0.5319205522537231, + "learning_rate": 0.0004999608825138162, + "loss": 2.0084, + "step": 359 + }, + { + "epoch": 0.03515625, + "grad_norm": 0.24298223853111267, + "learning_rate": 0.0004999595452999818, + "loss": 1.9183, + "step": 360 + }, + { + "epoch": 0.03525390625, + "grad_norm": 0.39650923013687134, + "learning_rate": 0.0004999581856146817, + "loss": 1.9886, + "step": 361 + }, + { + "epoch": 0.0353515625, + "grad_norm": 0.3819667100906372, + "learning_rate": 0.0004999568034580518, + "loss": 2.0047, + "step": 362 + }, + { + "epoch": 0.03544921875, + "grad_norm": 0.3298279345035553, + "learning_rate": 0.0004999553988302303, + "loss": 1.9466, + "step": 363 + }, + { + "epoch": 0.035546875, + "grad_norm": 0.2343115359544754, + "learning_rate": 0.0004999539717313573, + "loss": 1.9828, + "step": 364 + }, + { + "epoch": 0.03564453125, + "grad_norm": 0.31238943338394165, + "learning_rate": 0.0004999525221615755, + "loss": 2.0721, + "step": 365 + }, + { + "epoch": 0.0357421875, + "grad_norm": 0.27481910586357117, + "learning_rate": 0.0004999510501210295, + "loss": 2.0106, + "step": 366 + }, + { + "epoch": 0.03583984375, + "grad_norm": 0.24870915710926056, + "learning_rate": 0.0004999495556098666, + "loss": 1.9846, + "step": 367 + }, + { + "epoch": 0.0359375, + "grad_norm": 0.2666539251804352, + "learning_rate": 0.0004999480386282359, + "loss": 1.9988, + "step": 368 + }, + { + "epoch": 0.03603515625, + "grad_norm": 0.23231451213359833, + "learning_rate": 0.000499946499176289, + "loss": 2.0235, + "step": 369 + }, + { + "epoch": 0.0361328125, + "grad_norm": 0.19276577234268188, + "learning_rate": 0.0004999449372541798, + "loss": 1.9832, + "step": 370 + }, + { + "epoch": 0.03623046875, + "grad_norm": 0.23545822501182556, + "learning_rate": 0.000499943352862064, + "loss": 1.9879, + "step": 371 + }, + { + "epoch": 0.036328125, + "grad_norm": 0.2608807384967804, + "learning_rate": 0.0004999417460001002, + "loss": 2.047, + "step": 372 + }, + { + "epoch": 0.03642578125, + "grad_norm": 0.27395492792129517, + "learning_rate": 0.0004999401166684487, + "loss": 1.9496, + "step": 373 + }, + { + "epoch": 0.0365234375, + "grad_norm": 0.24661333858966827, + "learning_rate": 0.0004999384648672724, + "loss": 1.9924, + "step": 374 + }, + { + "epoch": 0.03662109375, + "grad_norm": 0.23559828102588654, + "learning_rate": 0.0004999367905967362, + "loss": 1.979, + "step": 375 + }, + { + "epoch": 0.03671875, + "grad_norm": 0.2556357681751251, + "learning_rate": 0.0004999350938570074, + "loss": 1.9793, + "step": 376 + }, + { + "epoch": 0.03681640625, + "grad_norm": 0.29081088304519653, + "learning_rate": 0.0004999333746482555, + "loss": 1.962, + "step": 377 + }, + { + "epoch": 0.0369140625, + "grad_norm": 0.2583800256252289, + "learning_rate": 0.0004999316329706521, + "loss": 1.9765, + "step": 378 + }, + { + "epoch": 0.03701171875, + "grad_norm": 0.2506580054759979, + "learning_rate": 0.0004999298688243714, + "loss": 2.0302, + "step": 379 + }, + { + "epoch": 0.037109375, + "grad_norm": 0.289530485868454, + "learning_rate": 0.0004999280822095895, + "loss": 1.9514, + "step": 380 + }, + { + "epoch": 0.03720703125, + "grad_norm": 0.24479152262210846, + "learning_rate": 0.0004999262731264848, + "loss": 1.9742, + "step": 381 + }, + { + "epoch": 0.0373046875, + "grad_norm": 0.23676550388336182, + "learning_rate": 0.0004999244415752381, + "loss": 2.0161, + "step": 382 + }, + { + "epoch": 0.03740234375, + "grad_norm": 0.2447502315044403, + "learning_rate": 0.0004999225875560323, + "loss": 2.0114, + "step": 383 + }, + { + "epoch": 0.0375, + "grad_norm": 0.20993874967098236, + "learning_rate": 0.0004999207110690528, + "loss": 1.9786, + "step": 384 + }, + { + "epoch": 0.03759765625, + "grad_norm": 0.23405557870864868, + "learning_rate": 0.0004999188121144867, + "loss": 2.0004, + "step": 385 + }, + { + "epoch": 0.0376953125, + "grad_norm": 0.2310025990009308, + "learning_rate": 0.0004999168906925238, + "loss": 1.924, + "step": 386 + }, + { + "epoch": 0.03779296875, + "grad_norm": 0.2507460117340088, + "learning_rate": 0.0004999149468033564, + "loss": 2.0301, + "step": 387 + }, + { + "epoch": 0.037890625, + "grad_norm": 0.30916762351989746, + "learning_rate": 0.0004999129804471782, + "loss": 1.9948, + "step": 388 + }, + { + "epoch": 0.03798828125, + "grad_norm": 0.42094507813453674, + "learning_rate": 0.0004999109916241858, + "loss": 2.0128, + "step": 389 + }, + { + "epoch": 0.0380859375, + "grad_norm": 0.5398088693618774, + "learning_rate": 0.0004999089803345779, + "loss": 1.9866, + "step": 390 + }, + { + "epoch": 0.03818359375, + "grad_norm": 0.4601620137691498, + "learning_rate": 0.0004999069465785554, + "loss": 1.9807, + "step": 391 + }, + { + "epoch": 0.03828125, + "grad_norm": 0.24873095750808716, + "learning_rate": 0.0004999048903563213, + "loss": 1.9225, + "step": 392 + }, + { + "epoch": 0.03837890625, + "grad_norm": 0.37363573908805847, + "learning_rate": 0.0004999028116680814, + "loss": 2.0212, + "step": 393 + }, + { + "epoch": 0.0384765625, + "grad_norm": 0.27399107813835144, + "learning_rate": 0.0004999007105140428, + "loss": 1.9797, + "step": 394 + }, + { + "epoch": 0.03857421875, + "grad_norm": 0.2717953622341156, + "learning_rate": 0.0004998985868944158, + "loss": 2.0119, + "step": 395 + }, + { + "epoch": 0.038671875, + "grad_norm": 0.2766648530960083, + "learning_rate": 0.0004998964408094124, + "loss": 1.9936, + "step": 396 + }, + { + "epoch": 0.03876953125, + "grad_norm": 0.2989843487739563, + "learning_rate": 0.0004998942722592469, + "loss": 2.0261, + "step": 397 + }, + { + "epoch": 0.0388671875, + "grad_norm": 0.31954097747802734, + "learning_rate": 0.000499892081244136, + "loss": 1.9713, + "step": 398 + }, + { + "epoch": 0.03896484375, + "grad_norm": 0.31199777126312256, + "learning_rate": 0.0004998898677642987, + "loss": 1.9548, + "step": 399 + }, + { + "epoch": 0.0390625, + "grad_norm": 0.2475418746471405, + "learning_rate": 0.0004998876318199557, + "loss": 1.9878, + "step": 400 + }, + { + "epoch": 0.03916015625, + "grad_norm": 0.25001809000968933, + "learning_rate": 0.0004998853734113308, + "loss": 1.991, + "step": 401 + }, + { + "epoch": 0.0392578125, + "grad_norm": 0.2468329817056656, + "learning_rate": 0.0004998830925386492, + "loss": 1.9988, + "step": 402 + }, + { + "epoch": 0.03935546875, + "grad_norm": 0.23273305594921112, + "learning_rate": 0.000499880789202139, + "loss": 2.0138, + "step": 403 + }, + { + "epoch": 0.039453125, + "grad_norm": 0.21512626111507416, + "learning_rate": 0.0004998784634020303, + "loss": 1.9326, + "step": 404 + }, + { + "epoch": 0.03955078125, + "grad_norm": 0.2123369723558426, + "learning_rate": 0.0004998761151385554, + "loss": 1.9954, + "step": 405 + }, + { + "epoch": 0.0396484375, + "grad_norm": 0.20985403656959534, + "learning_rate": 0.0004998737444119488, + "loss": 1.996, + "step": 406 + }, + { + "epoch": 0.03974609375, + "grad_norm": 0.19709332287311554, + "learning_rate": 0.0004998713512224473, + "loss": 1.9642, + "step": 407 + }, + { + "epoch": 0.03984375, + "grad_norm": 0.21997478604316711, + "learning_rate": 0.00049986893557029, + "loss": 1.9925, + "step": 408 + }, + { + "epoch": 0.03994140625, + "grad_norm": 0.2419801503419876, + "learning_rate": 0.0004998664974557182, + "loss": 2.0262, + "step": 409 + }, + { + "epoch": 0.0400390625, + "grad_norm": 0.2527197003364563, + "learning_rate": 0.0004998640368789754, + "loss": 2.0011, + "step": 410 + }, + { + "epoch": 0.04013671875, + "grad_norm": 0.2539777159690857, + "learning_rate": 0.0004998615538403074, + "loss": 1.9799, + "step": 411 + }, + { + "epoch": 0.040234375, + "grad_norm": 0.2419251948595047, + "learning_rate": 0.0004998590483399623, + "loss": 2.0035, + "step": 412 + }, + { + "epoch": 0.04033203125, + "grad_norm": 0.26659199595451355, + "learning_rate": 0.0004998565203781904, + "loss": 1.9743, + "step": 413 + }, + { + "epoch": 0.0404296875, + "grad_norm": 0.3216726779937744, + "learning_rate": 0.0004998539699552441, + "loss": 1.979, + "step": 414 + }, + { + "epoch": 0.04052734375, + "grad_norm": 0.2780512571334839, + "learning_rate": 0.0004998513970713783, + "loss": 1.9892, + "step": 415 + }, + { + "epoch": 0.040625, + "grad_norm": 0.2947809100151062, + "learning_rate": 0.00049984880172685, + "loss": 1.9836, + "step": 416 + }, + { + "epoch": 0.04072265625, + "grad_norm": 0.33937209844589233, + "learning_rate": 0.0004998461839219182, + "loss": 2.0007, + "step": 417 + }, + { + "epoch": 0.0408203125, + "grad_norm": 0.3530365228652954, + "learning_rate": 0.0004998435436568446, + "loss": 1.9462, + "step": 418 + }, + { + "epoch": 0.04091796875, + "grad_norm": 0.34433260560035706, + "learning_rate": 0.000499840880931893, + "loss": 1.9733, + "step": 419 + }, + { + "epoch": 0.041015625, + "grad_norm": 0.323081910610199, + "learning_rate": 0.0004998381957473293, + "loss": 1.9522, + "step": 420 + }, + { + "epoch": 0.04111328125, + "grad_norm": 0.22895868122577667, + "learning_rate": 0.0004998354881034217, + "loss": 1.9672, + "step": 421 + }, + { + "epoch": 0.0412109375, + "grad_norm": 0.2543105185031891, + "learning_rate": 0.0004998327580004408, + "loss": 1.9429, + "step": 422 + }, + { + "epoch": 0.04130859375, + "grad_norm": 0.2795581519603729, + "learning_rate": 0.0004998300054386591, + "loss": 1.9902, + "step": 423 + }, + { + "epoch": 0.04140625, + "grad_norm": 0.2201048880815506, + "learning_rate": 0.0004998272304183517, + "loss": 1.9365, + "step": 424 + }, + { + "epoch": 0.04150390625, + "grad_norm": 0.26653197407722473, + "learning_rate": 0.0004998244329397958, + "loss": 2.0178, + "step": 425 + }, + { + "epoch": 0.0416015625, + "grad_norm": 0.26959821581840515, + "learning_rate": 0.0004998216130032708, + "loss": 1.9625, + "step": 426 + }, + { + "epoch": 0.04169921875, + "grad_norm": 0.2673629820346832, + "learning_rate": 0.0004998187706090584, + "loss": 1.995, + "step": 427 + }, + { + "epoch": 0.041796875, + "grad_norm": 0.3190925121307373, + "learning_rate": 0.0004998159057574426, + "loss": 2.0094, + "step": 428 + }, + { + "epoch": 0.04189453125, + "grad_norm": 0.3376927673816681, + "learning_rate": 0.0004998130184487094, + "loss": 2.0535, + "step": 429 + }, + { + "epoch": 0.0419921875, + "grad_norm": 0.35333067178726196, + "learning_rate": 0.0004998101086831474, + "loss": 2.0256, + "step": 430 + }, + { + "epoch": 0.04208984375, + "grad_norm": 0.3051585853099823, + "learning_rate": 0.0004998071764610471, + "loss": 1.9798, + "step": 431 + }, + { + "epoch": 0.0421875, + "grad_norm": 0.21881523728370667, + "learning_rate": 0.0004998042217827015, + "loss": 1.9726, + "step": 432 + }, + { + "epoch": 0.04228515625, + "grad_norm": 0.2296576052904129, + "learning_rate": 0.0004998012446484057, + "loss": 1.9675, + "step": 433 + }, + { + "epoch": 0.0423828125, + "grad_norm": 0.24055050313472748, + "learning_rate": 0.0004997982450584572, + "loss": 1.9488, + "step": 434 + }, + { + "epoch": 0.04248046875, + "grad_norm": 0.22267191112041473, + "learning_rate": 0.0004997952230131555, + "loss": 1.9617, + "step": 435 + }, + { + "epoch": 0.042578125, + "grad_norm": 0.2590855658054352, + "learning_rate": 0.0004997921785128026, + "loss": 1.9745, + "step": 436 + }, + { + "epoch": 0.04267578125, + "grad_norm": 0.2995806336402893, + "learning_rate": 0.0004997891115577025, + "loss": 1.9667, + "step": 437 + }, + { + "epoch": 0.0427734375, + "grad_norm": 0.3083277642726898, + "learning_rate": 0.0004997860221481616, + "loss": 1.9661, + "step": 438 + }, + { + "epoch": 0.04287109375, + "grad_norm": 0.31743094325065613, + "learning_rate": 0.0004997829102844885, + "loss": 2.0172, + "step": 439 + }, + { + "epoch": 0.04296875, + "grad_norm": 0.37507855892181396, + "learning_rate": 0.0004997797759669941, + "loss": 2.0211, + "step": 440 + }, + { + "epoch": 0.04306640625, + "grad_norm": 0.35785287618637085, + "learning_rate": 0.0004997766191959914, + "loss": 2.038, + "step": 441 + }, + { + "epoch": 0.0431640625, + "grad_norm": 0.3099903166294098, + "learning_rate": 0.0004997734399717958, + "loss": 1.9879, + "step": 442 + }, + { + "epoch": 0.04326171875, + "grad_norm": 0.32211577892303467, + "learning_rate": 0.0004997702382947248, + "loss": 1.987, + "step": 443 + }, + { + "epoch": 0.043359375, + "grad_norm": 0.24999700486660004, + "learning_rate": 0.0004997670141650984, + "loss": 1.9765, + "step": 444 + }, + { + "epoch": 0.04345703125, + "grad_norm": 0.23301391303539276, + "learning_rate": 0.0004997637675832386, + "loss": 1.9592, + "step": 445 + }, + { + "epoch": 0.0435546875, + "grad_norm": 0.22780932486057281, + "learning_rate": 0.0004997604985494694, + "loss": 1.9762, + "step": 446 + }, + { + "epoch": 0.04365234375, + "grad_norm": 0.2347191423177719, + "learning_rate": 0.0004997572070641178, + "loss": 1.9791, + "step": 447 + }, + { + "epoch": 0.04375, + "grad_norm": 0.3288608193397522, + "learning_rate": 0.0004997538931275123, + "loss": 1.9664, + "step": 448 + }, + { + "epoch": 0.04384765625, + "grad_norm": 0.29897212982177734, + "learning_rate": 0.000499750556739984, + "loss": 1.9876, + "step": 449 + }, + { + "epoch": 0.0439453125, + "grad_norm": 0.24389693140983582, + "learning_rate": 0.0004997471979018663, + "loss": 1.9742, + "step": 450 + }, + { + "epoch": 0.04404296875, + "grad_norm": 0.3172384798526764, + "learning_rate": 0.0004997438166134945, + "loss": 1.969, + "step": 451 + }, + { + "epoch": 0.044140625, + "grad_norm": 0.22951611876487732, + "learning_rate": 0.0004997404128752065, + "loss": 1.9832, + "step": 452 + }, + { + "epoch": 0.04423828125, + "grad_norm": 0.2356499582529068, + "learning_rate": 0.0004997369866873423, + "loss": 2.0236, + "step": 453 + }, + { + "epoch": 0.0443359375, + "grad_norm": 0.2687929570674896, + "learning_rate": 0.000499733538050244, + "loss": 1.9521, + "step": 454 + }, + { + "epoch": 0.04443359375, + "grad_norm": 0.2512679398059845, + "learning_rate": 0.0004997300669642564, + "loss": 1.9754, + "step": 455 + }, + { + "epoch": 0.04453125, + "grad_norm": 0.22110989689826965, + "learning_rate": 0.0004997265734297259, + "loss": 1.9931, + "step": 456 + }, + { + "epoch": 0.04462890625, + "grad_norm": 0.2548464834690094, + "learning_rate": 0.0004997230574470017, + "loss": 2.001, + "step": 457 + }, + { + "epoch": 0.0447265625, + "grad_norm": 0.3107970356941223, + "learning_rate": 0.0004997195190164349, + "loss": 1.9855, + "step": 458 + }, + { + "epoch": 0.04482421875, + "grad_norm": 0.2989678680896759, + "learning_rate": 0.0004997159581383789, + "loss": 1.9646, + "step": 459 + }, + { + "epoch": 0.044921875, + "grad_norm": 0.3651764690876007, + "learning_rate": 0.0004997123748131896, + "loss": 1.9865, + "step": 460 + }, + { + "epoch": 0.04501953125, + "grad_norm": 0.41693389415740967, + "learning_rate": 0.0004997087690412248, + "loss": 1.9756, + "step": 461 + }, + { + "epoch": 0.0451171875, + "grad_norm": 0.425959974527359, + "learning_rate": 0.0004997051408228447, + "loss": 1.9887, + "step": 462 + }, + { + "epoch": 0.04521484375, + "grad_norm": 0.38966843485832214, + "learning_rate": 0.0004997014901584118, + "loss": 1.9532, + "step": 463 + }, + { + "epoch": 0.0453125, + "grad_norm": 0.29931432008743286, + "learning_rate": 0.0004996978170482906, + "loss": 1.9806, + "step": 464 + }, + { + "epoch": 0.04541015625, + "grad_norm": 0.23077397048473358, + "learning_rate": 0.0004996941214928482, + "loss": 2.0258, + "step": 465 + }, + { + "epoch": 0.0455078125, + "grad_norm": 0.32101985812187195, + "learning_rate": 0.0004996904034924536, + "loss": 2.0023, + "step": 466 + }, + { + "epoch": 0.04560546875, + "grad_norm": 0.3225870430469513, + "learning_rate": 0.0004996866630474783, + "loss": 1.9385, + "step": 467 + }, + { + "epoch": 0.045703125, + "grad_norm": 0.2592551112174988, + "learning_rate": 0.0004996829001582959, + "loss": 2.0176, + "step": 468 + }, + { + "epoch": 0.04580078125, + "grad_norm": 0.24367393553256989, + "learning_rate": 0.0004996791148252824, + "loss": 1.9844, + "step": 469 + }, + { + "epoch": 0.0458984375, + "grad_norm": 0.26423218846321106, + "learning_rate": 0.0004996753070488156, + "loss": 1.9675, + "step": 470 + }, + { + "epoch": 0.04599609375, + "grad_norm": 0.27801692485809326, + "learning_rate": 0.0004996714768292762, + "loss": 1.9265, + "step": 471 + }, + { + "epoch": 0.04609375, + "grad_norm": 0.3210003077983856, + "learning_rate": 0.0004996676241670467, + "loss": 1.9937, + "step": 472 + }, + { + "epoch": 0.04619140625, + "grad_norm": 0.3082759380340576, + "learning_rate": 0.0004996637490625118, + "loss": 1.9997, + "step": 473 + }, + { + "epoch": 0.0462890625, + "grad_norm": 0.2486785650253296, + "learning_rate": 0.0004996598515160589, + "loss": 2.0687, + "step": 474 + }, + { + "epoch": 0.04638671875, + "grad_norm": 0.31927499175071716, + "learning_rate": 0.000499655931528077, + "loss": 1.9853, + "step": 475 + }, + { + "epoch": 0.046484375, + "grad_norm": 0.2736678123474121, + "learning_rate": 0.0004996519890989578, + "loss": 1.9916, + "step": 476 + }, + { + "epoch": 0.04658203125, + "grad_norm": 0.29522353410720825, + "learning_rate": 0.0004996480242290952, + "loss": 1.9554, + "step": 477 + }, + { + "epoch": 0.0466796875, + "grad_norm": 0.26841938495635986, + "learning_rate": 0.0004996440369188851, + "loss": 1.9275, + "step": 478 + }, + { + "epoch": 0.04677734375, + "grad_norm": 0.2695488929748535, + "learning_rate": 0.0004996400271687259, + "loss": 1.9898, + "step": 479 + }, + { + "epoch": 0.046875, + "grad_norm": 0.36634454131126404, + "learning_rate": 0.0004996359949790181, + "loss": 1.9799, + "step": 480 + }, + { + "epoch": 0.04697265625, + "grad_norm": 0.37612974643707275, + "learning_rate": 0.0004996319403501647, + "loss": 1.9645, + "step": 481 + }, + { + "epoch": 0.0470703125, + "grad_norm": 0.29007261991500854, + "learning_rate": 0.0004996278632825703, + "loss": 1.9486, + "step": 482 + }, + { + "epoch": 0.04716796875, + "grad_norm": 0.30373314023017883, + "learning_rate": 0.0004996237637766424, + "loss": 1.9913, + "step": 483 + }, + { + "epoch": 0.047265625, + "grad_norm": 0.2949075400829315, + "learning_rate": 0.0004996196418327906, + "loss": 1.9894, + "step": 484 + }, + { + "epoch": 0.04736328125, + "grad_norm": 0.30601680278778076, + "learning_rate": 0.0004996154974514264, + "loss": 1.9835, + "step": 485 + }, + { + "epoch": 0.0474609375, + "grad_norm": 0.2896345853805542, + "learning_rate": 0.000499611330632964, + "loss": 2.0043, + "step": 486 + }, + { + "epoch": 0.04755859375, + "grad_norm": 0.3836851418018341, + "learning_rate": 0.0004996071413778195, + "loss": 1.9782, + "step": 487 + }, + { + "epoch": 0.04765625, + "grad_norm": 0.4167245626449585, + "learning_rate": 0.0004996029296864114, + "loss": 1.9792, + "step": 488 + }, + { + "epoch": 0.04775390625, + "grad_norm": 0.3247157335281372, + "learning_rate": 0.0004995986955591606, + "loss": 1.9548, + "step": 489 + }, + { + "epoch": 0.0478515625, + "grad_norm": 0.36726462841033936, + "learning_rate": 0.0004995944389964897, + "loss": 1.9844, + "step": 490 + }, + { + "epoch": 0.04794921875, + "grad_norm": 0.27440258860588074, + "learning_rate": 0.0004995901599988241, + "loss": 1.9608, + "step": 491 + }, + { + "epoch": 0.048046875, + "grad_norm": 0.271771103143692, + "learning_rate": 0.0004995858585665912, + "loss": 2.022, + "step": 492 + }, + { + "epoch": 0.04814453125, + "grad_norm": 0.3079121708869934, + "learning_rate": 0.0004995815347002208, + "loss": 1.9804, + "step": 493 + }, + { + "epoch": 0.0482421875, + "grad_norm": 0.2976233661174774, + "learning_rate": 0.0004995771884001445, + "loss": 1.9679, + "step": 494 + }, + { + "epoch": 0.04833984375, + "grad_norm": 0.31318438053131104, + "learning_rate": 0.0004995728196667969, + "loss": 2.0342, + "step": 495 + }, + { + "epoch": 0.0484375, + "grad_norm": 0.2397848516702652, + "learning_rate": 0.0004995684285006139, + "loss": 1.9801, + "step": 496 + }, + { + "epoch": 0.04853515625, + "grad_norm": 0.2205589860677719, + "learning_rate": 0.0004995640149020346, + "loss": 2.0088, + "step": 497 + }, + { + "epoch": 0.0486328125, + "grad_norm": 0.23892101645469666, + "learning_rate": 0.0004995595788714995, + "loss": 1.9888, + "step": 498 + }, + { + "epoch": 0.04873046875, + "grad_norm": 0.2999497354030609, + "learning_rate": 0.000499555120409452, + "loss": 1.9896, + "step": 499 + }, + { + "epoch": 0.048828125, + "grad_norm": 0.3309103548526764, + "learning_rate": 0.0004995506395163372, + "loss": 1.9394, + "step": 500 + }, + { + "epoch": 0.04892578125, + "grad_norm": 0.36243122816085815, + "learning_rate": 0.000499546136192603, + "loss": 1.9604, + "step": 501 + }, + { + "epoch": 0.0490234375, + "grad_norm": 0.3052852153778076, + "learning_rate": 0.0004995416104386991, + "loss": 1.9887, + "step": 502 + }, + { + "epoch": 0.04912109375, + "grad_norm": 0.2957259714603424, + "learning_rate": 0.0004995370622550775, + "loss": 1.9876, + "step": 503 + }, + { + "epoch": 0.04921875, + "grad_norm": 0.3991003632545471, + "learning_rate": 0.0004995324916421926, + "loss": 1.9932, + "step": 504 + }, + { + "epoch": 0.04931640625, + "grad_norm": 0.27617356181144714, + "learning_rate": 0.000499527898600501, + "loss": 1.9387, + "step": 505 + }, + { + "epoch": 0.0494140625, + "grad_norm": 0.2820669412612915, + "learning_rate": 0.0004995232831304614, + "loss": 2.012, + "step": 506 + }, + { + "epoch": 0.04951171875, + "grad_norm": 0.3026665449142456, + "learning_rate": 0.0004995186452325351, + "loss": 1.9709, + "step": 507 + }, + { + "epoch": 0.049609375, + "grad_norm": 0.2331731766462326, + "learning_rate": 0.000499513984907185, + "loss": 1.9617, + "step": 508 + }, + { + "epoch": 0.04970703125, + "grad_norm": 0.2791639268398285, + "learning_rate": 0.0004995093021548768, + "loss": 1.9635, + "step": 509 + }, + { + "epoch": 0.0498046875, + "grad_norm": 0.257519006729126, + "learning_rate": 0.0004995045969760785, + "loss": 1.8908, + "step": 510 + }, + { + "epoch": 0.04990234375, + "grad_norm": 0.26278433203697205, + "learning_rate": 0.0004994998693712598, + "loss": 1.9418, + "step": 511 + }, + { + "epoch": 0.05, + "grad_norm": 0.2534024715423584, + "learning_rate": 0.0004994951193408929, + "loss": 1.9821, + "step": 512 + }, + { + "epoch": 0.05009765625, + "grad_norm": 0.272348552942276, + "learning_rate": 0.0004994903468854527, + "loss": 1.982, + "step": 513 + }, + { + "epoch": 0.0501953125, + "grad_norm": 0.23883765935897827, + "learning_rate": 0.0004994855520054154, + "loss": 2.006, + "step": 514 + }, + { + "epoch": 0.05029296875, + "grad_norm": 0.25627830624580383, + "learning_rate": 0.0004994807347012603, + "loss": 2.0008, + "step": 515 + }, + { + "epoch": 0.050390625, + "grad_norm": 0.33997049927711487, + "learning_rate": 0.0004994758949734686, + "loss": 2.0339, + "step": 516 + }, + { + "epoch": 0.05048828125, + "grad_norm": 0.4115971028804779, + "learning_rate": 0.0004994710328225236, + "loss": 1.9982, + "step": 517 + }, + { + "epoch": 0.0505859375, + "grad_norm": 0.4417625069618225, + "learning_rate": 0.000499466148248911, + "loss": 1.9742, + "step": 518 + }, + { + "epoch": 0.05068359375, + "grad_norm": 0.4123833477497101, + "learning_rate": 0.0004994612412531189, + "loss": 2.0148, + "step": 519 + }, + { + "epoch": 0.05078125, + "grad_norm": 0.2234133630990982, + "learning_rate": 0.0004994563118356373, + "loss": 1.9679, + "step": 520 + }, + { + "epoch": 0.05087890625, + "grad_norm": 0.35439473390579224, + "learning_rate": 0.0004994513599969586, + "loss": 1.9602, + "step": 521 + }, + { + "epoch": 0.0509765625, + "grad_norm": 0.32346996665000916, + "learning_rate": 0.0004994463857375776, + "loss": 2.0106, + "step": 522 + }, + { + "epoch": 0.05107421875, + "grad_norm": 0.24506594240665436, + "learning_rate": 0.000499441389057991, + "loss": 1.993, + "step": 523 + }, + { + "epoch": 0.051171875, + "grad_norm": 0.26808494329452515, + "learning_rate": 0.000499436369958698, + "loss": 1.9764, + "step": 524 + }, + { + "epoch": 0.05126953125, + "grad_norm": 0.21936193108558655, + "learning_rate": 0.0004994313284401999, + "loss": 1.9619, + "step": 525 + }, + { + "epoch": 0.0513671875, + "grad_norm": 0.2545083165168762, + "learning_rate": 0.0004994262645030005, + "loss": 1.9825, + "step": 526 + }, + { + "epoch": 0.05146484375, + "grad_norm": 0.26330727338790894, + "learning_rate": 0.0004994211781476055, + "loss": 1.9512, + "step": 527 + }, + { + "epoch": 0.0515625, + "grad_norm": 0.2708059549331665, + "learning_rate": 0.0004994160693745229, + "loss": 1.9761, + "step": 528 + }, + { + "epoch": 0.05166015625, + "grad_norm": 0.2882954478263855, + "learning_rate": 0.0004994109381842632, + "loss": 2.0223, + "step": 529 + }, + { + "epoch": 0.0517578125, + "grad_norm": 0.25243133306503296, + "learning_rate": 0.0004994057845773389, + "loss": 1.976, + "step": 530 + }, + { + "epoch": 0.05185546875, + "grad_norm": 0.23666390776634216, + "learning_rate": 0.0004994006085542648, + "loss": 1.9729, + "step": 531 + }, + { + "epoch": 0.051953125, + "grad_norm": 0.24242763221263885, + "learning_rate": 0.0004993954101155578, + "loss": 1.9452, + "step": 532 + }, + { + "epoch": 0.05205078125, + "grad_norm": 0.259750634431839, + "learning_rate": 0.0004993901892617373, + "loss": 1.9582, + "step": 533 + }, + { + "epoch": 0.0521484375, + "grad_norm": 0.28516751527786255, + "learning_rate": 0.0004993849459933249, + "loss": 1.9367, + "step": 534 + }, + { + "epoch": 0.05224609375, + "grad_norm": 0.26534441113471985, + "learning_rate": 0.0004993796803108442, + "loss": 2.0137, + "step": 535 + }, + { + "epoch": 0.05234375, + "grad_norm": 0.28771716356277466, + "learning_rate": 0.0004993743922148213, + "loss": 1.927, + "step": 536 + }, + { + "epoch": 0.05244140625, + "grad_norm": 0.2747785449028015, + "learning_rate": 0.0004993690817057844, + "loss": 1.9729, + "step": 537 + }, + { + "epoch": 0.0525390625, + "grad_norm": 0.29758942127227783, + "learning_rate": 0.0004993637487842639, + "loss": 1.9713, + "step": 538 + }, + { + "epoch": 0.05263671875, + "grad_norm": 0.3059535026550293, + "learning_rate": 0.0004993583934507927, + "loss": 2.0313, + "step": 539 + }, + { + "epoch": 0.052734375, + "grad_norm": 0.36820656061172485, + "learning_rate": 0.0004993530157059056, + "loss": 1.9476, + "step": 540 + }, + { + "epoch": 0.05283203125, + "grad_norm": 0.39971432089805603, + "learning_rate": 0.0004993476155501396, + "loss": 2.0238, + "step": 541 + }, + { + "epoch": 0.0529296875, + "grad_norm": 0.3369678854942322, + "learning_rate": 0.0004993421929840346, + "loss": 1.9502, + "step": 542 + }, + { + "epoch": 0.05302734375, + "grad_norm": 0.3733840882778168, + "learning_rate": 0.000499336748008132, + "loss": 2.0114, + "step": 543 + }, + { + "epoch": 0.053125, + "grad_norm": 0.3251579999923706, + "learning_rate": 0.0004993312806229757, + "loss": 1.964, + "step": 544 + }, + { + "epoch": 0.05322265625, + "grad_norm": 0.4093327522277832, + "learning_rate": 0.0004993257908291117, + "loss": 1.9853, + "step": 545 + }, + { + "epoch": 0.0533203125, + "grad_norm": 0.26201343536376953, + "learning_rate": 0.0004993202786270888, + "loss": 2.0297, + "step": 546 + }, + { + "epoch": 0.05341796875, + "grad_norm": 0.27842357754707336, + "learning_rate": 0.0004993147440174572, + "loss": 2.003, + "step": 547 + }, + { + "epoch": 0.053515625, + "grad_norm": 0.2930687367916107, + "learning_rate": 0.00049930918700077, + "loss": 1.9634, + "step": 548 + }, + { + "epoch": 0.05361328125, + "grad_norm": 0.2799973785877228, + "learning_rate": 0.0004993036075775821, + "loss": 2.0023, + "step": 549 + }, + { + "epoch": 0.0537109375, + "grad_norm": 0.35084354877471924, + "learning_rate": 0.000499298005748451, + "loss": 2.0508, + "step": 550 + }, + { + "epoch": 0.05380859375, + "grad_norm": 0.2895161211490631, + "learning_rate": 0.0004992923815139362, + "loss": 2.0074, + "step": 551 + }, + { + "epoch": 0.05390625, + "grad_norm": 0.2637045979499817, + "learning_rate": 0.0004992867348745997, + "loss": 1.9559, + "step": 552 + }, + { + "epoch": 0.05400390625, + "grad_norm": 0.2837629020214081, + "learning_rate": 0.0004992810658310052, + "loss": 1.9603, + "step": 553 + }, + { + "epoch": 0.0541015625, + "grad_norm": 0.2644243836402893, + "learning_rate": 0.0004992753743837193, + "loss": 1.9767, + "step": 554 + }, + { + "epoch": 0.05419921875, + "grad_norm": 0.21815134584903717, + "learning_rate": 0.0004992696605333103, + "loss": 1.9684, + "step": 555 + }, + { + "epoch": 0.054296875, + "grad_norm": 0.28167852759361267, + "learning_rate": 0.0004992639242803492, + "loss": 1.97, + "step": 556 + }, + { + "epoch": 0.05439453125, + "grad_norm": 0.26942357420921326, + "learning_rate": 0.0004992581656254087, + "loss": 1.9825, + "step": 557 + }, + { + "epoch": 0.0544921875, + "grad_norm": 0.25316697359085083, + "learning_rate": 0.0004992523845690644, + "loss": 1.9698, + "step": 558 + }, + { + "epoch": 0.05458984375, + "grad_norm": 0.29587239027023315, + "learning_rate": 0.0004992465811118934, + "loss": 2.0063, + "step": 559 + }, + { + "epoch": 0.0546875, + "grad_norm": 0.3209396004676819, + "learning_rate": 0.0004992407552544757, + "loss": 2.0002, + "step": 560 + }, + { + "epoch": 0.05478515625, + "grad_norm": 0.2541654109954834, + "learning_rate": 0.0004992349069973931, + "loss": 2.0039, + "step": 561 + }, + { + "epoch": 0.0548828125, + "grad_norm": 0.34102505445480347, + "learning_rate": 0.0004992290363412298, + "loss": 1.9181, + "step": 562 + }, + { + "epoch": 0.05498046875, + "grad_norm": 0.36123788356781006, + "learning_rate": 0.0004992231432865723, + "loss": 2.042, + "step": 563 + }, + { + "epoch": 0.055078125, + "grad_norm": 0.27344051003456116, + "learning_rate": 0.0004992172278340093, + "loss": 2.0236, + "step": 564 + }, + { + "epoch": 0.05517578125, + "grad_norm": 0.34781453013420105, + "learning_rate": 0.0004992112899841315, + "loss": 2.0017, + "step": 565 + }, + { + "epoch": 0.0552734375, + "grad_norm": 0.2822319567203522, + "learning_rate": 0.0004992053297375322, + "loss": 1.9607, + "step": 566 + }, + { + "epoch": 0.05537109375, + "grad_norm": 0.27945882081985474, + "learning_rate": 0.0004991993470948066, + "loss": 1.9707, + "step": 567 + }, + { + "epoch": 0.05546875, + "grad_norm": 0.31535595655441284, + "learning_rate": 0.0004991933420565527, + "loss": 1.9708, + "step": 568 + }, + { + "epoch": 0.05556640625, + "grad_norm": 0.2658466398715973, + "learning_rate": 0.00049918731462337, + "loss": 1.9763, + "step": 569 + }, + { + "epoch": 0.0556640625, + "grad_norm": 0.3618657886981964, + "learning_rate": 0.0004991812647958607, + "loss": 2.0154, + "step": 570 + }, + { + "epoch": 0.05576171875, + "grad_norm": 0.34552058577537537, + "learning_rate": 0.000499175192574629, + "loss": 1.978, + "step": 571 + }, + { + "epoch": 0.055859375, + "grad_norm": 0.32903313636779785, + "learning_rate": 0.0004991690979602817, + "loss": 1.9762, + "step": 572 + }, + { + "epoch": 0.05595703125, + "grad_norm": 0.30814382433891296, + "learning_rate": 0.0004991629809534275, + "loss": 1.9944, + "step": 573 + }, + { + "epoch": 0.0560546875, + "grad_norm": 0.21156492829322815, + "learning_rate": 0.0004991568415546775, + "loss": 1.9919, + "step": 574 + }, + { + "epoch": 0.05615234375, + "grad_norm": 0.32283446192741394, + "learning_rate": 0.0004991506797646446, + "loss": 1.9697, + "step": 575 + }, + { + "epoch": 0.05625, + "grad_norm": 0.2915102541446686, + "learning_rate": 0.0004991444955839447, + "loss": 1.9736, + "step": 576 + }, + { + "epoch": 0.05634765625, + "grad_norm": 0.28996148705482483, + "learning_rate": 0.0004991382890131955, + "loss": 2.002, + "step": 577 + }, + { + "epoch": 0.0564453125, + "grad_norm": 0.3859401345252991, + "learning_rate": 0.0004991320600530168, + "loss": 1.9753, + "step": 578 + }, + { + "epoch": 0.05654296875, + "grad_norm": 0.28169432282447815, + "learning_rate": 0.0004991258087040312, + "loss": 1.9798, + "step": 579 + }, + { + "epoch": 0.056640625, + "grad_norm": 0.30669984221458435, + "learning_rate": 0.0004991195349668626, + "loss": 2.0007, + "step": 580 + }, + { + "epoch": 0.05673828125, + "grad_norm": 0.26832297444343567, + "learning_rate": 0.000499113238842138, + "loss": 2.0146, + "step": 581 + }, + { + "epoch": 0.0568359375, + "grad_norm": 0.25695785880088806, + "learning_rate": 0.0004991069203304865, + "loss": 1.9814, + "step": 582 + }, + { + "epoch": 0.05693359375, + "grad_norm": 0.25700533390045166, + "learning_rate": 0.0004991005794325389, + "loss": 1.9623, + "step": 583 + }, + { + "epoch": 0.05703125, + "grad_norm": 0.20786869525909424, + "learning_rate": 0.0004990942161489288, + "loss": 1.9728, + "step": 584 + }, + { + "epoch": 0.05712890625, + "grad_norm": 0.23988646268844604, + "learning_rate": 0.0004990878304802918, + "loss": 1.9749, + "step": 585 + }, + { + "epoch": 0.0572265625, + "grad_norm": 0.2418268769979477, + "learning_rate": 0.0004990814224272658, + "loss": 1.928, + "step": 586 + }, + { + "epoch": 0.05732421875, + "grad_norm": 0.27739882469177246, + "learning_rate": 0.0004990749919904909, + "loss": 1.964, + "step": 587 + }, + { + "epoch": 0.057421875, + "grad_norm": 0.2773842513561249, + "learning_rate": 0.0004990685391706094, + "loss": 1.9589, + "step": 588 + }, + { + "epoch": 0.05751953125, + "grad_norm": 0.24483682215213776, + "learning_rate": 0.0004990620639682659, + "loss": 1.9965, + "step": 589 + }, + { + "epoch": 0.0576171875, + "grad_norm": 0.27345889806747437, + "learning_rate": 0.0004990555663841071, + "loss": 1.9444, + "step": 590 + }, + { + "epoch": 0.05771484375, + "grad_norm": 0.26006460189819336, + "learning_rate": 0.0004990490464187824, + "loss": 1.9391, + "step": 591 + }, + { + "epoch": 0.0578125, + "grad_norm": 0.1935306191444397, + "learning_rate": 0.0004990425040729427, + "loss": 1.9679, + "step": 592 + }, + { + "epoch": 0.05791015625, + "grad_norm": 0.26823100447654724, + "learning_rate": 0.0004990359393472418, + "loss": 2.0055, + "step": 593 + }, + { + "epoch": 0.0580078125, + "grad_norm": 0.2706959545612335, + "learning_rate": 0.0004990293522423352, + "loss": 1.9853, + "step": 594 + }, + { + "epoch": 0.05810546875, + "grad_norm": 0.2099694162607193, + "learning_rate": 0.0004990227427588811, + "loss": 1.959, + "step": 595 + }, + { + "epoch": 0.058203125, + "grad_norm": 0.22345881164073944, + "learning_rate": 0.0004990161108975398, + "loss": 1.9867, + "step": 596 + }, + { + "epoch": 0.05830078125, + "grad_norm": 0.34050315618515015, + "learning_rate": 0.0004990094566589734, + "loss": 1.9555, + "step": 597 + }, + { + "epoch": 0.0583984375, + "grad_norm": 0.5298358798027039, + "learning_rate": 0.0004990027800438468, + "loss": 1.9824, + "step": 598 + }, + { + "epoch": 0.05849609375, + "grad_norm": 0.6408056616783142, + "learning_rate": 0.0004989960810528271, + "loss": 1.9647, + "step": 599 + }, + { + "epoch": 0.05859375, + "grad_norm": 0.4058885872364044, + "learning_rate": 0.0004989893596865833, + "loss": 1.9564, + "step": 600 + }, + { + "epoch": 0.05869140625, + "grad_norm": 0.2911423146724701, + "learning_rate": 0.0004989826159457869, + "loss": 1.9485, + "step": 601 + }, + { + "epoch": 0.0587890625, + "grad_norm": 0.40600210428237915, + "learning_rate": 0.0004989758498311114, + "loss": 1.993, + "step": 602 + }, + { + "epoch": 0.05888671875, + "grad_norm": 0.3920167088508606, + "learning_rate": 0.0004989690613432327, + "loss": 2.0188, + "step": 603 + }, + { + "epoch": 0.058984375, + "grad_norm": 0.2924116551876068, + "learning_rate": 0.0004989622504828291, + "loss": 1.9749, + "step": 604 + }, + { + "epoch": 0.05908203125, + "grad_norm": 0.33497628569602966, + "learning_rate": 0.0004989554172505807, + "loss": 1.9358, + "step": 605 + }, + { + "epoch": 0.0591796875, + "grad_norm": 0.31751033663749695, + "learning_rate": 0.0004989485616471702, + "loss": 1.9886, + "step": 606 + }, + { + "epoch": 0.05927734375, + "grad_norm": 0.2204602211713791, + "learning_rate": 0.0004989416836732825, + "loss": 1.9715, + "step": 607 + }, + { + "epoch": 0.059375, + "grad_norm": 0.2980511784553528, + "learning_rate": 0.0004989347833296044, + "loss": 1.9479, + "step": 608 + }, + { + "epoch": 0.05947265625, + "grad_norm": 0.2874164879322052, + "learning_rate": 0.0004989278606168253, + "loss": 1.9717, + "step": 609 + }, + { + "epoch": 0.0595703125, + "grad_norm": 0.2526591718196869, + "learning_rate": 0.0004989209155356368, + "loss": 2.0009, + "step": 610 + }, + { + "epoch": 0.05966796875, + "grad_norm": 0.2890442907810211, + "learning_rate": 0.0004989139480867327, + "loss": 1.9862, + "step": 611 + }, + { + "epoch": 0.059765625, + "grad_norm": 0.26463401317596436, + "learning_rate": 0.0004989069582708088, + "loss": 2.0143, + "step": 612 + }, + { + "epoch": 0.05986328125, + "grad_norm": 0.19906644523143768, + "learning_rate": 0.0004988999460885634, + "loss": 1.9845, + "step": 613 + }, + { + "epoch": 0.0599609375, + "grad_norm": 0.28752684593200684, + "learning_rate": 0.000498892911540697, + "loss": 1.9421, + "step": 614 + }, + { + "epoch": 0.06005859375, + "grad_norm": 0.2739415466785431, + "learning_rate": 0.0004988858546279123, + "loss": 1.967, + "step": 615 + }, + { + "epoch": 0.06015625, + "grad_norm": 0.27040478587150574, + "learning_rate": 0.000498878775350914, + "loss": 2.0001, + "step": 616 + }, + { + "epoch": 0.06025390625, + "grad_norm": 0.28345200419425964, + "learning_rate": 0.0004988716737104096, + "loss": 1.9637, + "step": 617 + }, + { + "epoch": 0.0603515625, + "grad_norm": 0.2515316903591156, + "learning_rate": 0.0004988645497071082, + "loss": 1.9511, + "step": 618 + }, + { + "epoch": 0.06044921875, + "grad_norm": 0.22784312069416046, + "learning_rate": 0.0004988574033417216, + "loss": 1.9752, + "step": 619 + }, + { + "epoch": 0.060546875, + "grad_norm": 0.23128612339496613, + "learning_rate": 0.0004988502346149636, + "loss": 1.9803, + "step": 620 + }, + { + "epoch": 0.06064453125, + "grad_norm": 0.22022667527198792, + "learning_rate": 0.0004988430435275503, + "loss": 1.9592, + "step": 621 + }, + { + "epoch": 0.0607421875, + "grad_norm": 0.2470719963312149, + "learning_rate": 0.0004988358300802, + "loss": 2.0063, + "step": 622 + }, + { + "epoch": 0.06083984375, + "grad_norm": 0.22507692873477936, + "learning_rate": 0.0004988285942736332, + "loss": 1.9873, + "step": 623 + }, + { + "epoch": 0.0609375, + "grad_norm": 0.2285120040178299, + "learning_rate": 0.0004988213361085729, + "loss": 1.973, + "step": 624 + }, + { + "epoch": 0.06103515625, + "grad_norm": 0.2436297982931137, + "learning_rate": 0.000498814055585744, + "loss": 1.9786, + "step": 625 + }, + { + "epoch": 0.0611328125, + "grad_norm": 0.2736089825630188, + "learning_rate": 0.0004988067527058737, + "loss": 1.9552, + "step": 626 + }, + { + "epoch": 0.06123046875, + "grad_norm": 0.2813490033149719, + "learning_rate": 0.0004987994274696917, + "loss": 2.0173, + "step": 627 + }, + { + "epoch": 0.061328125, + "grad_norm": 0.3024250864982605, + "learning_rate": 0.0004987920798779294, + "loss": 1.9232, + "step": 628 + }, + { + "epoch": 0.06142578125, + "grad_norm": 0.22795617580413818, + "learning_rate": 0.000498784709931321, + "loss": 1.9981, + "step": 629 + }, + { + "epoch": 0.0615234375, + "grad_norm": 0.2585217356681824, + "learning_rate": 0.0004987773176306026, + "loss": 2.0197, + "step": 630 + }, + { + "epoch": 0.06162109375, + "grad_norm": 0.27120688557624817, + "learning_rate": 0.0004987699029765127, + "loss": 1.961, + "step": 631 + }, + { + "epoch": 0.06171875, + "grad_norm": 0.23527094721794128, + "learning_rate": 0.000498762465969792, + "loss": 1.9996, + "step": 632 + }, + { + "epoch": 0.06181640625, + "grad_norm": 0.2519530653953552, + "learning_rate": 0.0004987550066111832, + "loss": 1.9926, + "step": 633 + }, + { + "epoch": 0.0619140625, + "grad_norm": 0.3199564218521118, + "learning_rate": 0.0004987475249014315, + "loss": 2.0066, + "step": 634 + }, + { + "epoch": 0.06201171875, + "grad_norm": 0.3830641806125641, + "learning_rate": 0.0004987400208412843, + "loss": 1.9368, + "step": 635 + }, + { + "epoch": 0.062109375, + "grad_norm": 0.3777483403682709, + "learning_rate": 0.0004987324944314913, + "loss": 1.9792, + "step": 636 + }, + { + "epoch": 0.06220703125, + "grad_norm": 0.420107901096344, + "learning_rate": 0.0004987249456728041, + "loss": 1.9765, + "step": 637 + }, + { + "epoch": 0.0623046875, + "grad_norm": 0.3748738467693329, + "learning_rate": 0.0004987173745659768, + "loss": 1.9826, + "step": 638 + }, + { + "epoch": 0.06240234375, + "grad_norm": 0.3142974078655243, + "learning_rate": 0.0004987097811117658, + "loss": 1.9393, + "step": 639 + }, + { + "epoch": 0.0625, + "grad_norm": 0.24331647157669067, + "learning_rate": 0.0004987021653109296, + "loss": 1.9305, + "step": 640 + }, + { + "epoch": 0.06259765625, + "grad_norm": 0.2792535424232483, + "learning_rate": 0.0004986945271642289, + "loss": 1.9437, + "step": 641 + }, + { + "epoch": 0.0626953125, + "grad_norm": 0.24984711408615112, + "learning_rate": 0.0004986868666724267, + "loss": 1.9838, + "step": 642 + }, + { + "epoch": 0.06279296875, + "grad_norm": 0.2477482557296753, + "learning_rate": 0.0004986791838362881, + "loss": 1.9981, + "step": 643 + }, + { + "epoch": 0.062890625, + "grad_norm": 0.3012976348400116, + "learning_rate": 0.0004986714786565807, + "loss": 1.9963, + "step": 644 + }, + { + "epoch": 0.06298828125, + "grad_norm": 0.2480735331773758, + "learning_rate": 0.0004986637511340742, + "loss": 1.9567, + "step": 645 + }, + { + "epoch": 0.0630859375, + "grad_norm": 0.20119979977607727, + "learning_rate": 0.0004986560012695403, + "loss": 2.0012, + "step": 646 + }, + { + "epoch": 0.06318359375, + "grad_norm": 0.27725833654403687, + "learning_rate": 0.0004986482290637535, + "loss": 1.9751, + "step": 647 + }, + { + "epoch": 0.06328125, + "grad_norm": 0.27348190546035767, + "learning_rate": 0.0004986404345174898, + "loss": 1.976, + "step": 648 + }, + { + "epoch": 0.06337890625, + "grad_norm": 0.29639676213264465, + "learning_rate": 0.000498632617631528, + "loss": 1.9835, + "step": 649 + }, + { + "epoch": 0.0634765625, + "grad_norm": 0.2911996841430664, + "learning_rate": 0.000498624778406649, + "loss": 2.0107, + "step": 650 + }, + { + "epoch": 0.06357421875, + "grad_norm": 0.30200451612472534, + "learning_rate": 0.0004986169168436355, + "loss": 1.9433, + "step": 651 + }, + { + "epoch": 0.063671875, + "grad_norm": 0.288396954536438, + "learning_rate": 0.0004986090329432734, + "loss": 1.9759, + "step": 652 + }, + { + "epoch": 0.06376953125, + "grad_norm": 0.26006969809532166, + "learning_rate": 0.0004986011267063497, + "loss": 1.977, + "step": 653 + }, + { + "epoch": 0.0638671875, + "grad_norm": 0.2467355877161026, + "learning_rate": 0.0004985931981336544, + "loss": 1.9892, + "step": 654 + }, + { + "epoch": 0.06396484375, + "grad_norm": 0.2180376946926117, + "learning_rate": 0.0004985852472259797, + "loss": 2.0012, + "step": 655 + }, + { + "epoch": 0.0640625, + "grad_norm": 0.23811106383800507, + "learning_rate": 0.0004985772739841193, + "loss": 1.969, + "step": 656 + }, + { + "epoch": 0.06416015625, + "grad_norm": 0.2517051696777344, + "learning_rate": 0.00049856927840887, + "loss": 1.9607, + "step": 657 + }, + { + "epoch": 0.0642578125, + "grad_norm": 0.2900746464729309, + "learning_rate": 0.0004985612605010305, + "loss": 1.9865, + "step": 658 + }, + { + "epoch": 0.06435546875, + "grad_norm": 0.2779560983181, + "learning_rate": 0.0004985532202614017, + "loss": 1.9787, + "step": 659 + }, + { + "epoch": 0.064453125, + "grad_norm": 0.2686167061328888, + "learning_rate": 0.0004985451576907865, + "loss": 1.9279, + "step": 660 + }, + { + "epoch": 0.06455078125, + "grad_norm": 0.25136813521385193, + "learning_rate": 0.0004985370727899907, + "loss": 1.9502, + "step": 661 + }, + { + "epoch": 0.0646484375, + "grad_norm": 0.26508840918540955, + "learning_rate": 0.0004985289655598216, + "loss": 1.9588, + "step": 662 + }, + { + "epoch": 0.06474609375, + "grad_norm": 0.2178860753774643, + "learning_rate": 0.000498520836001089, + "loss": 1.9445, + "step": 663 + }, + { + "epoch": 0.06484375, + "grad_norm": 0.21833141148090363, + "learning_rate": 0.0004985126841146052, + "loss": 1.9963, + "step": 664 + }, + { + "epoch": 0.06494140625, + "grad_norm": 0.21572785079479218, + "learning_rate": 0.0004985045099011844, + "loss": 1.9789, + "step": 665 + }, + { + "epoch": 0.0650390625, + "grad_norm": 0.23386414349079132, + "learning_rate": 0.0004984963133616432, + "loss": 1.9829, + "step": 666 + }, + { + "epoch": 0.06513671875, + "grad_norm": 0.25145047903060913, + "learning_rate": 0.0004984880944968003, + "loss": 2.0106, + "step": 667 + }, + { + "epoch": 0.065234375, + "grad_norm": 0.23240795731544495, + "learning_rate": 0.0004984798533074767, + "loss": 1.999, + "step": 668 + }, + { + "epoch": 0.06533203125, + "grad_norm": 0.22223520278930664, + "learning_rate": 0.0004984715897944954, + "loss": 1.9872, + "step": 669 + }, + { + "epoch": 0.0654296875, + "grad_norm": 0.2768714725971222, + "learning_rate": 0.0004984633039586823, + "loss": 1.9656, + "step": 670 + }, + { + "epoch": 0.06552734375, + "grad_norm": 0.35761523246765137, + "learning_rate": 0.0004984549958008646, + "loss": 1.9984, + "step": 671 + }, + { + "epoch": 0.065625, + "grad_norm": 0.4054514169692993, + "learning_rate": 0.0004984466653218726, + "loss": 1.9406, + "step": 672 + }, + { + "epoch": 0.06572265625, + "grad_norm": 0.40501922369003296, + "learning_rate": 0.0004984383125225383, + "loss": 1.9682, + "step": 673 + }, + { + "epoch": 0.0658203125, + "grad_norm": 0.31225842237472534, + "learning_rate": 0.0004984299374036961, + "loss": 1.9833, + "step": 674 + }, + { + "epoch": 0.06591796875, + "grad_norm": 0.24159114062786102, + "learning_rate": 0.0004984215399661825, + "loss": 1.9634, + "step": 675 + }, + { + "epoch": 0.066015625, + "grad_norm": 0.2537291944026947, + "learning_rate": 0.0004984131202108364, + "loss": 1.9563, + "step": 676 + }, + { + "epoch": 0.06611328125, + "grad_norm": 0.29919305443763733, + "learning_rate": 0.0004984046781384988, + "loss": 1.9594, + "step": 677 + }, + { + "epoch": 0.0662109375, + "grad_norm": 0.28458645939826965, + "learning_rate": 0.0004983962137500133, + "loss": 1.9313, + "step": 678 + }, + { + "epoch": 0.06630859375, + "grad_norm": 0.2523960471153259, + "learning_rate": 0.0004983877270462249, + "loss": 1.9616, + "step": 679 + }, + { + "epoch": 0.06640625, + "grad_norm": 0.2605716288089752, + "learning_rate": 0.0004983792180279816, + "loss": 1.9631, + "step": 680 + }, + { + "epoch": 0.06650390625, + "grad_norm": 0.280000239610672, + "learning_rate": 0.0004983706866961335, + "loss": 1.9945, + "step": 681 + }, + { + "epoch": 0.0666015625, + "grad_norm": 0.2881770730018616, + "learning_rate": 0.0004983621330515326, + "loss": 1.9639, + "step": 682 + }, + { + "epoch": 0.06669921875, + "grad_norm": 0.27240094542503357, + "learning_rate": 0.0004983535570950334, + "loss": 1.9226, + "step": 683 + }, + { + "epoch": 0.066796875, + "grad_norm": 0.2580428123474121, + "learning_rate": 0.0004983449588274926, + "loss": 1.9625, + "step": 684 + }, + { + "epoch": 0.06689453125, + "grad_norm": 0.2719865143299103, + "learning_rate": 0.0004983363382497692, + "loss": 1.9941, + "step": 685 + }, + { + "epoch": 0.0669921875, + "grad_norm": 0.2593652904033661, + "learning_rate": 0.000498327695362724, + "loss": 1.9558, + "step": 686 + }, + { + "epoch": 0.06708984375, + "grad_norm": 0.2575826346874237, + "learning_rate": 0.0004983190301672207, + "loss": 2.0188, + "step": 687 + }, + { + "epoch": 0.0671875, + "grad_norm": 0.23726949095726013, + "learning_rate": 0.0004983103426641246, + "loss": 1.9568, + "step": 688 + }, + { + "epoch": 0.06728515625, + "grad_norm": 0.3027707040309906, + "learning_rate": 0.0004983016328543037, + "loss": 1.9708, + "step": 689 + }, + { + "epoch": 0.0673828125, + "grad_norm": 0.29431256651878357, + "learning_rate": 0.0004982929007386279, + "loss": 1.9668, + "step": 690 + }, + { + "epoch": 0.06748046875, + "grad_norm": 0.24073566496372223, + "learning_rate": 0.0004982841463179694, + "loss": 1.9746, + "step": 691 + }, + { + "epoch": 0.067578125, + "grad_norm": 0.27540677785873413, + "learning_rate": 0.000498275369593203, + "loss": 1.868, + "step": 692 + }, + { + "epoch": 0.06767578125, + "grad_norm": 0.2665644884109497, + "learning_rate": 0.0004982665705652052, + "loss": 1.9528, + "step": 693 + }, + { + "epoch": 0.0677734375, + "grad_norm": 0.2117265909910202, + "learning_rate": 0.0004982577492348549, + "loss": 1.9977, + "step": 694 + }, + { + "epoch": 0.06787109375, + "grad_norm": 0.30092543363571167, + "learning_rate": 0.0004982489056030334, + "loss": 1.9636, + "step": 695 + }, + { + "epoch": 0.06796875, + "grad_norm": 0.2956363558769226, + "learning_rate": 0.0004982400396706239, + "loss": 1.9519, + "step": 696 + }, + { + "epoch": 0.06806640625, + "grad_norm": 0.3209246098995209, + "learning_rate": 0.0004982311514385124, + "loss": 1.9761, + "step": 697 + }, + { + "epoch": 0.0681640625, + "grad_norm": 0.4527253806591034, + "learning_rate": 0.0004982222409075863, + "loss": 1.9405, + "step": 698 + }, + { + "epoch": 0.06826171875, + "grad_norm": 0.4838014245033264, + "learning_rate": 0.000498213308078736, + "loss": 1.9434, + "step": 699 + }, + { + "epoch": 0.068359375, + "grad_norm": 0.4865509569644928, + "learning_rate": 0.0004982043529528536, + "loss": 1.9706, + "step": 700 + }, + { + "epoch": 0.06845703125, + "grad_norm": 0.38092029094696045, + "learning_rate": 0.0004981953755308339, + "loss": 1.9168, + "step": 701 + }, + { + "epoch": 0.0685546875, + "grad_norm": 0.24092066287994385, + "learning_rate": 0.0004981863758135735, + "loss": 1.9693, + "step": 702 + }, + { + "epoch": 0.06865234375, + "grad_norm": 0.3893338143825531, + "learning_rate": 0.0004981773538019714, + "loss": 1.9812, + "step": 703 + }, + { + "epoch": 0.06875, + "grad_norm": 0.3019930422306061, + "learning_rate": 0.0004981683094969287, + "loss": 1.9207, + "step": 704 + }, + { + "epoch": 0.06884765625, + "grad_norm": 0.2746503949165344, + "learning_rate": 0.0004981592428993492, + "loss": 1.9416, + "step": 705 + }, + { + "epoch": 0.0689453125, + "grad_norm": 0.2998049557209015, + "learning_rate": 0.0004981501540101382, + "loss": 1.9685, + "step": 706 + }, + { + "epoch": 0.06904296875, + "grad_norm": 0.2986246645450592, + "learning_rate": 0.0004981410428302037, + "loss": 1.9827, + "step": 707 + }, + { + "epoch": 0.069140625, + "grad_norm": 0.28714433312416077, + "learning_rate": 0.000498131909360456, + "loss": 1.904, + "step": 708 + }, + { + "epoch": 0.06923828125, + "grad_norm": 0.21930216252803802, + "learning_rate": 0.0004981227536018073, + "loss": 1.966, + "step": 709 + }, + { + "epoch": 0.0693359375, + "grad_norm": 0.2309812754392624, + "learning_rate": 0.0004981135755551722, + "loss": 1.9983, + "step": 710 + }, + { + "epoch": 0.06943359375, + "grad_norm": 0.2016872614622116, + "learning_rate": 0.0004981043752214674, + "loss": 1.9382, + "step": 711 + }, + { + "epoch": 0.06953125, + "grad_norm": 0.235287606716156, + "learning_rate": 0.0004980951526016122, + "loss": 1.9914, + "step": 712 + }, + { + "epoch": 0.06962890625, + "grad_norm": 0.23780131340026855, + "learning_rate": 0.0004980859076965276, + "loss": 1.9404, + "step": 713 + }, + { + "epoch": 0.0697265625, + "grad_norm": 0.2195412665605545, + "learning_rate": 0.0004980766405071372, + "loss": 1.9526, + "step": 714 + }, + { + "epoch": 0.06982421875, + "grad_norm": 0.2391882836818695, + "learning_rate": 0.0004980673510343668, + "loss": 1.9179, + "step": 715 + }, + { + "epoch": 0.069921875, + "grad_norm": 0.27623385190963745, + "learning_rate": 0.000498058039279144, + "loss": 1.9519, + "step": 716 + }, + { + "epoch": 0.07001953125, + "grad_norm": 0.27679604291915894, + "learning_rate": 0.0004980487052423994, + "loss": 2.0292, + "step": 717 + }, + { + "epoch": 0.0701171875, + "grad_norm": 0.2968669533729553, + "learning_rate": 0.0004980393489250651, + "loss": 1.9162, + "step": 718 + }, + { + "epoch": 0.07021484375, + "grad_norm": 0.2479199320077896, + "learning_rate": 0.0004980299703280758, + "loss": 1.9544, + "step": 719 + }, + { + "epoch": 0.0703125, + "grad_norm": 0.2597293555736542, + "learning_rate": 0.0004980205694523682, + "loss": 1.9273, + "step": 720 + }, + { + "epoch": 0.07041015625, + "grad_norm": 0.23868051171302795, + "learning_rate": 0.0004980111462988817, + "loss": 1.9527, + "step": 721 + }, + { + "epoch": 0.0705078125, + "grad_norm": 0.28150075674057007, + "learning_rate": 0.0004980017008685573, + "loss": 1.9523, + "step": 722 + }, + { + "epoch": 0.07060546875, + "grad_norm": 0.3689873516559601, + "learning_rate": 0.0004979922331623387, + "loss": 1.9167, + "step": 723 + }, + { + "epoch": 0.070703125, + "grad_norm": 0.380206823348999, + "learning_rate": 0.0004979827431811713, + "loss": 1.9025, + "step": 724 + }, + { + "epoch": 0.07080078125, + "grad_norm": 0.2919938266277313, + "learning_rate": 0.0004979732309260036, + "loss": 1.9996, + "step": 725 + }, + { + "epoch": 0.0708984375, + "grad_norm": 0.302680104970932, + "learning_rate": 0.0004979636963977853, + "loss": 1.95, + "step": 726 + }, + { + "epoch": 0.07099609375, + "grad_norm": 0.25307831168174744, + "learning_rate": 0.0004979541395974691, + "loss": 1.9723, + "step": 727 + }, + { + "epoch": 0.07109375, + "grad_norm": 0.304147869348526, + "learning_rate": 0.0004979445605260096, + "loss": 2.0359, + "step": 728 + }, + { + "epoch": 0.07119140625, + "grad_norm": 0.35432276129722595, + "learning_rate": 0.0004979349591843635, + "loss": 1.9384, + "step": 729 + }, + { + "epoch": 0.0712890625, + "grad_norm": 0.23866906762123108, + "learning_rate": 0.00049792533557349, + "loss": 1.9341, + "step": 730 + }, + { + "epoch": 0.07138671875, + "grad_norm": 0.30183711647987366, + "learning_rate": 0.0004979156896943504, + "loss": 1.9603, + "step": 731 + }, + { + "epoch": 0.071484375, + "grad_norm": 0.30881768465042114, + "learning_rate": 0.0004979060215479083, + "loss": 1.982, + "step": 732 + }, + { + "epoch": 0.07158203125, + "grad_norm": 0.2996484041213989, + "learning_rate": 0.0004978963311351295, + "loss": 1.9536, + "step": 733 + }, + { + "epoch": 0.0716796875, + "grad_norm": 0.33594587445259094, + "learning_rate": 0.0004978866184569818, + "loss": 1.9813, + "step": 734 + }, + { + "epoch": 0.07177734375, + "grad_norm": 0.30377864837646484, + "learning_rate": 0.0004978768835144355, + "loss": 1.9417, + "step": 735 + }, + { + "epoch": 0.071875, + "grad_norm": 0.34858009219169617, + "learning_rate": 0.0004978671263084631, + "loss": 1.9368, + "step": 736 + }, + { + "epoch": 0.07197265625, + "grad_norm": 0.28119683265686035, + "learning_rate": 0.0004978573468400392, + "loss": 1.9223, + "step": 737 + }, + { + "epoch": 0.0720703125, + "grad_norm": 0.2478085160255432, + "learning_rate": 0.0004978475451101408, + "loss": 1.9744, + "step": 738 + }, + { + "epoch": 0.07216796875, + "grad_norm": 0.2461906224489212, + "learning_rate": 0.0004978377211197466, + "loss": 1.961, + "step": 739 + }, + { + "epoch": 0.072265625, + "grad_norm": 0.2495601922273636, + "learning_rate": 0.0004978278748698385, + "loss": 1.9602, + "step": 740 + }, + { + "epoch": 0.07236328125, + "grad_norm": 0.3028644323348999, + "learning_rate": 0.0004978180063613997, + "loss": 1.95, + "step": 741 + }, + { + "epoch": 0.0724609375, + "grad_norm": 0.3690338730812073, + "learning_rate": 0.000497808115595416, + "loss": 1.9625, + "step": 742 + }, + { + "epoch": 0.07255859375, + "grad_norm": 0.3791324198246002, + "learning_rate": 0.0004977982025728755, + "loss": 1.9747, + "step": 743 + }, + { + "epoch": 0.07265625, + "grad_norm": 0.29515188932418823, + "learning_rate": 0.0004977882672947683, + "loss": 1.9774, + "step": 744 + }, + { + "epoch": 0.07275390625, + "grad_norm": 0.29207196831703186, + "learning_rate": 0.000497778309762087, + "loss": 1.9704, + "step": 745 + }, + { + "epoch": 0.0728515625, + "grad_norm": 0.23698067665100098, + "learning_rate": 0.0004977683299758262, + "loss": 1.978, + "step": 746 + }, + { + "epoch": 0.07294921875, + "grad_norm": 0.2712165415287018, + "learning_rate": 0.0004977583279369827, + "loss": 1.9524, + "step": 747 + }, + { + "epoch": 0.073046875, + "grad_norm": 0.32161638140678406, + "learning_rate": 0.0004977483036465558, + "loss": 1.9532, + "step": 748 + }, + { + "epoch": 0.07314453125, + "grad_norm": 0.25917574763298035, + "learning_rate": 0.0004977382571055467, + "loss": 1.9307, + "step": 749 + }, + { + "epoch": 0.0732421875, + "grad_norm": 0.27562615275382996, + "learning_rate": 0.000497728188314959, + "loss": 1.9842, + "step": 750 + }, + { + "epoch": 0.07333984375, + "grad_norm": 0.23242716491222382, + "learning_rate": 0.0004977180972757986, + "loss": 1.9097, + "step": 751 + }, + { + "epoch": 0.0734375, + "grad_norm": 0.3111313283443451, + "learning_rate": 0.0004977079839890732, + "loss": 1.9812, + "step": 752 + }, + { + "epoch": 0.07353515625, + "grad_norm": 0.3295767605304718, + "learning_rate": 0.0004976978484557933, + "loss": 1.9206, + "step": 753 + }, + { + "epoch": 0.0736328125, + "grad_norm": 0.3592395782470703, + "learning_rate": 0.0004976876906769713, + "loss": 1.9817, + "step": 754 + }, + { + "epoch": 0.07373046875, + "grad_norm": 0.3530953824520111, + "learning_rate": 0.0004976775106536218, + "loss": 1.9831, + "step": 755 + }, + { + "epoch": 0.073828125, + "grad_norm": 0.3064524829387665, + "learning_rate": 0.0004976673083867617, + "loss": 1.9207, + "step": 756 + }, + { + "epoch": 0.07392578125, + "grad_norm": 0.281597763299942, + "learning_rate": 0.0004976570838774102, + "loss": 1.9781, + "step": 757 + }, + { + "epoch": 0.0740234375, + "grad_norm": 0.296041876077652, + "learning_rate": 0.0004976468371265886, + "loss": 1.9708, + "step": 758 + }, + { + "epoch": 0.07412109375, + "grad_norm": 0.3518696427345276, + "learning_rate": 0.0004976365681353203, + "loss": 1.9333, + "step": 759 + }, + { + "epoch": 0.07421875, + "grad_norm": 0.21943148970603943, + "learning_rate": 0.0004976262769046313, + "loss": 1.9416, + "step": 760 + }, + { + "epoch": 0.07431640625, + "grad_norm": 0.2540123760700226, + "learning_rate": 0.0004976159634355495, + "loss": 1.989, + "step": 761 + }, + { + "epoch": 0.0744140625, + "grad_norm": 0.3120715320110321, + "learning_rate": 0.0004976056277291053, + "loss": 1.95, + "step": 762 + }, + { + "epoch": 0.07451171875, + "grad_norm": 0.23797550797462463, + "learning_rate": 0.0004975952697863308, + "loss": 1.9534, + "step": 763 + }, + { + "epoch": 0.074609375, + "grad_norm": 0.30815449357032776, + "learning_rate": 0.000497584889608261, + "loss": 2.0283, + "step": 764 + }, + { + "epoch": 0.07470703125, + "grad_norm": 0.37625932693481445, + "learning_rate": 0.0004975744871959327, + "loss": 1.9658, + "step": 765 + }, + { + "epoch": 0.0748046875, + "grad_norm": 0.27407678961753845, + "learning_rate": 0.0004975640625503848, + "loss": 1.9458, + "step": 766 + }, + { + "epoch": 0.07490234375, + "grad_norm": 0.26577049493789673, + "learning_rate": 0.000497553615672659, + "loss": 1.9667, + "step": 767 + }, + { + "epoch": 0.075, + "grad_norm": 0.29821062088012695, + "learning_rate": 0.0004975431465637983, + "loss": 1.9805, + "step": 768 + }, + { + "epoch": 0.07509765625, + "grad_norm": 0.2582527995109558, + "learning_rate": 0.0004975326552248491, + "loss": 1.9904, + "step": 769 + }, + { + "epoch": 0.0751953125, + "grad_norm": 0.2533649802207947, + "learning_rate": 0.0004975221416568591, + "loss": 1.9383, + "step": 770 + }, + { + "epoch": 0.07529296875, + "grad_norm": 0.253592312335968, + "learning_rate": 0.0004975116058608784, + "loss": 1.9855, + "step": 771 + }, + { + "epoch": 0.075390625, + "grad_norm": 0.2658921182155609, + "learning_rate": 0.0004975010478379596, + "loss": 1.9845, + "step": 772 + }, + { + "epoch": 0.07548828125, + "grad_norm": 0.25861868262290955, + "learning_rate": 0.0004974904675891573, + "loss": 1.9776, + "step": 773 + }, + { + "epoch": 0.0755859375, + "grad_norm": 0.272682249546051, + "learning_rate": 0.0004974798651155284, + "loss": 1.9357, + "step": 774 + }, + { + "epoch": 0.07568359375, + "grad_norm": 0.2574002742767334, + "learning_rate": 0.0004974692404181319, + "loss": 1.9677, + "step": 775 + }, + { + "epoch": 0.07578125, + "grad_norm": 0.3008319139480591, + "learning_rate": 0.0004974585934980293, + "loss": 1.983, + "step": 776 + }, + { + "epoch": 0.07587890625, + "grad_norm": 0.3242790699005127, + "learning_rate": 0.0004974479243562839, + "loss": 1.9846, + "step": 777 + }, + { + "epoch": 0.0759765625, + "grad_norm": 0.29044070839881897, + "learning_rate": 0.0004974372329939617, + "loss": 2.0012, + "step": 778 + }, + { + "epoch": 0.07607421875, + "grad_norm": 0.32275035977363586, + "learning_rate": 0.0004974265194121303, + "loss": 1.9842, + "step": 779 + }, + { + "epoch": 0.076171875, + "grad_norm": 0.31752893328666687, + "learning_rate": 0.0004974157836118604, + "loss": 1.9368, + "step": 780 + }, + { + "epoch": 0.07626953125, + "grad_norm": 0.27947962284088135, + "learning_rate": 0.0004974050255942239, + "loss": 1.9794, + "step": 781 + }, + { + "epoch": 0.0763671875, + "grad_norm": 0.29937905073165894, + "learning_rate": 0.000497394245360296, + "loss": 1.9597, + "step": 782 + }, + { + "epoch": 0.07646484375, + "grad_norm": 0.2969702482223511, + "learning_rate": 0.0004973834429111529, + "loss": 1.945, + "step": 783 + }, + { + "epoch": 0.0765625, + "grad_norm": 0.3729354441165924, + "learning_rate": 0.0004973726182478741, + "loss": 1.9032, + "step": 784 + }, + { + "epoch": 0.07666015625, + "grad_norm": 0.32856106758117676, + "learning_rate": 0.0004973617713715408, + "loss": 1.9551, + "step": 785 + }, + { + "epoch": 0.0767578125, + "grad_norm": 0.23239392042160034, + "learning_rate": 0.0004973509022832364, + "loss": 1.9632, + "step": 786 + }, + { + "epoch": 0.07685546875, + "grad_norm": 0.23353087902069092, + "learning_rate": 0.000497340010984047, + "loss": 1.9514, + "step": 787 + }, + { + "epoch": 0.076953125, + "grad_norm": 0.23280592262744904, + "learning_rate": 0.00049732909747506, + "loss": 1.9696, + "step": 788 + }, + { + "epoch": 0.07705078125, + "grad_norm": 0.2193138748407364, + "learning_rate": 0.0004973181617573658, + "loss": 1.9734, + "step": 789 + }, + { + "epoch": 0.0771484375, + "grad_norm": 0.3234734833240509, + "learning_rate": 0.0004973072038320569, + "loss": 1.9874, + "step": 790 + }, + { + "epoch": 0.07724609375, + "grad_norm": 0.35144004225730896, + "learning_rate": 0.0004972962237002278, + "loss": 1.9331, + "step": 791 + }, + { + "epoch": 0.07734375, + "grad_norm": 0.37716230750083923, + "learning_rate": 0.0004972852213629753, + "loss": 1.9731, + "step": 792 + }, + { + "epoch": 0.07744140625, + "grad_norm": 0.3221200108528137, + "learning_rate": 0.0004972741968213985, + "loss": 2.0043, + "step": 793 + }, + { + "epoch": 0.0775390625, + "grad_norm": 0.23167745769023895, + "learning_rate": 0.0004972631500765985, + "loss": 1.9634, + "step": 794 + }, + { + "epoch": 0.07763671875, + "grad_norm": 0.4177049398422241, + "learning_rate": 0.0004972520811296789, + "loss": 1.9533, + "step": 795 + }, + { + "epoch": 0.077734375, + "grad_norm": 0.39038336277008057, + "learning_rate": 0.0004972409899817455, + "loss": 1.9508, + "step": 796 + }, + { + "epoch": 0.07783203125, + "grad_norm": 0.4033074676990509, + "learning_rate": 0.0004972298766339061, + "loss": 1.965, + "step": 797 + }, + { + "epoch": 0.0779296875, + "grad_norm": 0.3271097242832184, + "learning_rate": 0.0004972187410872707, + "loss": 1.9893, + "step": 798 + }, + { + "epoch": 0.07802734375, + "grad_norm": 0.30553460121154785, + "learning_rate": 0.0004972075833429516, + "loss": 1.9735, + "step": 799 + }, + { + "epoch": 0.078125, + "grad_norm": 0.28286099433898926, + "learning_rate": 0.0004971964034020637, + "loss": 1.9704, + "step": 800 + }, + { + "epoch": 0.07822265625, + "grad_norm": 0.28064650297164917, + "learning_rate": 0.0004971852012657235, + "loss": 1.9452, + "step": 801 + }, + { + "epoch": 0.0783203125, + "grad_norm": 0.2370154857635498, + "learning_rate": 0.0004971739769350502, + "loss": 2.0068, + "step": 802 + }, + { + "epoch": 0.07841796875, + "grad_norm": 0.26157453656196594, + "learning_rate": 0.0004971627304111648, + "loss": 1.9209, + "step": 803 + }, + { + "epoch": 0.078515625, + "grad_norm": 0.2244575023651123, + "learning_rate": 0.0004971514616951908, + "loss": 1.9952, + "step": 804 + }, + { + "epoch": 0.07861328125, + "grad_norm": 0.241064190864563, + "learning_rate": 0.0004971401707882538, + "loss": 1.9616, + "step": 805 + }, + { + "epoch": 0.0787109375, + "grad_norm": 0.23037207126617432, + "learning_rate": 0.0004971288576914819, + "loss": 1.9792, + "step": 806 + }, + { + "epoch": 0.07880859375, + "grad_norm": 0.2162441909313202, + "learning_rate": 0.0004971175224060049, + "loss": 1.9556, + "step": 807 + }, + { + "epoch": 0.07890625, + "grad_norm": 0.2614160180091858, + "learning_rate": 0.0004971061649329551, + "loss": 1.9247, + "step": 808 + }, + { + "epoch": 0.07900390625, + "grad_norm": 0.22268271446228027, + "learning_rate": 0.0004970947852734672, + "loss": 1.9596, + "step": 809 + }, + { + "epoch": 0.0791015625, + "grad_norm": 0.26265689730644226, + "learning_rate": 0.0004970833834286779, + "loss": 1.9722, + "step": 810 + }, + { + "epoch": 0.07919921875, + "grad_norm": 0.29413190484046936, + "learning_rate": 0.000497071959399726, + "loss": 1.9808, + "step": 811 + }, + { + "epoch": 0.079296875, + "grad_norm": 0.3389902114868164, + "learning_rate": 0.0004970605131877527, + "loss": 1.9986, + "step": 812 + }, + { + "epoch": 0.07939453125, + "grad_norm": 0.3456060290336609, + "learning_rate": 0.0004970490447939016, + "loss": 1.9556, + "step": 813 + }, + { + "epoch": 0.0794921875, + "grad_norm": 0.3039393126964569, + "learning_rate": 0.0004970375542193179, + "loss": 1.9359, + "step": 814 + }, + { + "epoch": 0.07958984375, + "grad_norm": 0.46128201484680176, + "learning_rate": 0.0004970260414651497, + "loss": 1.935, + "step": 815 + }, + { + "epoch": 0.0796875, + "grad_norm": 0.5321182608604431, + "learning_rate": 0.000497014506532547, + "loss": 1.9767, + "step": 816 + }, + { + "epoch": 0.07978515625, + "grad_norm": 0.39562714099884033, + "learning_rate": 0.0004970029494226618, + "loss": 1.9803, + "step": 817 + }, + { + "epoch": 0.0798828125, + "grad_norm": 0.2717466950416565, + "learning_rate": 0.0004969913701366488, + "loss": 1.9675, + "step": 818 + }, + { + "epoch": 0.07998046875, + "grad_norm": 0.32809481024742126, + "learning_rate": 0.0004969797686756646, + "loss": 1.9869, + "step": 819 + }, + { + "epoch": 0.080078125, + "grad_norm": 0.3525324761867523, + "learning_rate": 0.0004969681450408681, + "loss": 1.9858, + "step": 820 + }, + { + "epoch": 0.08017578125, + "grad_norm": 0.30448198318481445, + "learning_rate": 0.0004969564992334202, + "loss": 2.0067, + "step": 821 + }, + { + "epoch": 0.0802734375, + "grad_norm": 0.27088144421577454, + "learning_rate": 0.0004969448312544845, + "loss": 2.0068, + "step": 822 + }, + { + "epoch": 0.08037109375, + "grad_norm": 0.26881691813468933, + "learning_rate": 0.0004969331411052265, + "loss": 1.9418, + "step": 823 + }, + { + "epoch": 0.08046875, + "grad_norm": 0.31323063373565674, + "learning_rate": 0.0004969214287868138, + "loss": 1.9662, + "step": 824 + }, + { + "epoch": 0.08056640625, + "grad_norm": 0.25669077038764954, + "learning_rate": 0.0004969096943004163, + "loss": 1.9733, + "step": 825 + }, + { + "epoch": 0.0806640625, + "grad_norm": 0.27999046444892883, + "learning_rate": 0.0004968979376472065, + "loss": 1.9611, + "step": 826 + }, + { + "epoch": 0.08076171875, + "grad_norm": 0.35361799597740173, + "learning_rate": 0.0004968861588283584, + "loss": 1.9035, + "step": 827 + }, + { + "epoch": 0.080859375, + "grad_norm": 0.290341317653656, + "learning_rate": 0.0004968743578450489, + "loss": 2.0007, + "step": 828 + }, + { + "epoch": 0.08095703125, + "grad_norm": 0.29945844411849976, + "learning_rate": 0.0004968625346984565, + "loss": 1.9724, + "step": 829 + }, + { + "epoch": 0.0810546875, + "grad_norm": 0.2630455493927002, + "learning_rate": 0.0004968506893897626, + "loss": 1.9765, + "step": 830 + }, + { + "epoch": 0.08115234375, + "grad_norm": 0.2183736264705658, + "learning_rate": 0.0004968388219201501, + "loss": 1.9414, + "step": 831 + }, + { + "epoch": 0.08125, + "grad_norm": 0.24703919887542725, + "learning_rate": 0.0004968269322908046, + "loss": 1.9725, + "step": 832 + }, + { + "epoch": 0.08134765625, + "grad_norm": 0.22808632254600525, + "learning_rate": 0.0004968150205029139, + "loss": 1.9121, + "step": 833 + }, + { + "epoch": 0.0814453125, + "grad_norm": 0.18958304822444916, + "learning_rate": 0.0004968030865576677, + "loss": 1.971, + "step": 834 + }, + { + "epoch": 0.08154296875, + "grad_norm": 0.22013933956623077, + "learning_rate": 0.0004967911304562581, + "loss": 1.9703, + "step": 835 + }, + { + "epoch": 0.081640625, + "grad_norm": 0.21636781096458435, + "learning_rate": 0.0004967791521998795, + "loss": 1.992, + "step": 836 + }, + { + "epoch": 0.08173828125, + "grad_norm": 0.20808343589305878, + "learning_rate": 0.0004967671517897284, + "loss": 1.9881, + "step": 837 + }, + { + "epoch": 0.0818359375, + "grad_norm": 0.23455387353897095, + "learning_rate": 0.0004967551292270034, + "loss": 1.9635, + "step": 838 + }, + { + "epoch": 0.08193359375, + "grad_norm": 0.2605065703392029, + "learning_rate": 0.0004967430845129057, + "loss": 1.8877, + "step": 839 + }, + { + "epoch": 0.08203125, + "grad_norm": 0.30832260847091675, + "learning_rate": 0.0004967310176486382, + "loss": 1.9557, + "step": 840 + }, + { + "epoch": 0.08212890625, + "grad_norm": 0.32088643312454224, + "learning_rate": 0.0004967189286354066, + "loss": 1.9352, + "step": 841 + }, + { + "epoch": 0.0822265625, + "grad_norm": 0.3073631823062897, + "learning_rate": 0.0004967068174744181, + "loss": 1.9487, + "step": 842 + }, + { + "epoch": 0.08232421875, + "grad_norm": 0.23467223346233368, + "learning_rate": 0.0004966946841668828, + "loss": 1.9514, + "step": 843 + }, + { + "epoch": 0.082421875, + "grad_norm": 0.2447347342967987, + "learning_rate": 0.0004966825287140125, + "loss": 1.9435, + "step": 844 + }, + { + "epoch": 0.08251953125, + "grad_norm": 0.2715449631214142, + "learning_rate": 0.0004966703511170215, + "loss": 1.9591, + "step": 845 + }, + { + "epoch": 0.0826171875, + "grad_norm": 0.28944098949432373, + "learning_rate": 0.0004966581513771263, + "loss": 1.952, + "step": 846 + }, + { + "epoch": 0.08271484375, + "grad_norm": 0.25582170486450195, + "learning_rate": 0.0004966459294955455, + "loss": 1.9486, + "step": 847 + }, + { + "epoch": 0.0828125, + "grad_norm": 0.305513858795166, + "learning_rate": 0.0004966336854734999, + "loss": 1.9416, + "step": 848 + }, + { + "epoch": 0.08291015625, + "grad_norm": 0.28799137473106384, + "learning_rate": 0.0004966214193122127, + "loss": 1.9417, + "step": 849 + }, + { + "epoch": 0.0830078125, + "grad_norm": 0.22209681570529938, + "learning_rate": 0.000496609131012909, + "loss": 1.9641, + "step": 850 + }, + { + "epoch": 0.08310546875, + "grad_norm": 0.35411134362220764, + "learning_rate": 0.0004965968205768165, + "loss": 1.9497, + "step": 851 + }, + { + "epoch": 0.083203125, + "grad_norm": 0.2821413576602936, + "learning_rate": 0.0004965844880051647, + "loss": 1.9398, + "step": 852 + }, + { + "epoch": 0.08330078125, + "grad_norm": 0.31592175364494324, + "learning_rate": 0.0004965721332991856, + "loss": 1.9407, + "step": 853 + }, + { + "epoch": 0.0833984375, + "grad_norm": 0.3579123318195343, + "learning_rate": 0.0004965597564601134, + "loss": 1.9767, + "step": 854 + }, + { + "epoch": 0.08349609375, + "grad_norm": 0.3371833860874176, + "learning_rate": 0.0004965473574891845, + "loss": 2.0025, + "step": 855 + }, + { + "epoch": 0.08359375, + "grad_norm": 0.39282771944999695, + "learning_rate": 0.0004965349363876371, + "loss": 1.9303, + "step": 856 + }, + { + "epoch": 0.08369140625, + "grad_norm": 0.34869179129600525, + "learning_rate": 0.0004965224931567122, + "loss": 1.9613, + "step": 857 + }, + { + "epoch": 0.0837890625, + "grad_norm": 0.2918208837509155, + "learning_rate": 0.0004965100277976528, + "loss": 1.9242, + "step": 858 + }, + { + "epoch": 0.08388671875, + "grad_norm": 0.25440847873687744, + "learning_rate": 0.0004964975403117041, + "loss": 1.9223, + "step": 859 + }, + { + "epoch": 0.083984375, + "grad_norm": 0.27069318294525146, + "learning_rate": 0.0004964850307001132, + "loss": 1.9711, + "step": 860 + }, + { + "epoch": 0.08408203125, + "grad_norm": 0.27405187487602234, + "learning_rate": 0.0004964724989641301, + "loss": 1.9646, + "step": 861 + }, + { + "epoch": 0.0841796875, + "grad_norm": 0.3116970956325531, + "learning_rate": 0.0004964599451050064, + "loss": 1.9231, + "step": 862 + }, + { + "epoch": 0.08427734375, + "grad_norm": 0.3815697729587555, + "learning_rate": 0.000496447369123996, + "loss": 1.9773, + "step": 863 + }, + { + "epoch": 0.084375, + "grad_norm": 0.32103797793388367, + "learning_rate": 0.0004964347710223555, + "loss": 1.936, + "step": 864 + }, + { + "epoch": 0.08447265625, + "grad_norm": 0.42826011776924133, + "learning_rate": 0.0004964221508013429, + "loss": 1.9315, + "step": 865 + }, + { + "epoch": 0.0845703125, + "grad_norm": 0.3091580271720886, + "learning_rate": 0.0004964095084622192, + "loss": 1.9362, + "step": 866 + }, + { + "epoch": 0.08466796875, + "grad_norm": 0.2816343605518341, + "learning_rate": 0.0004963968440062471, + "loss": 2.0076, + "step": 867 + }, + { + "epoch": 0.084765625, + "grad_norm": 0.35303476452827454, + "learning_rate": 0.0004963841574346917, + "loss": 1.9399, + "step": 868 + }, + { + "epoch": 0.08486328125, + "grad_norm": 0.27021846175193787, + "learning_rate": 0.0004963714487488202, + "loss": 1.9556, + "step": 869 + }, + { + "epoch": 0.0849609375, + "grad_norm": 0.28174805641174316, + "learning_rate": 0.0004963587179499021, + "loss": 1.9749, + "step": 870 + }, + { + "epoch": 0.08505859375, + "grad_norm": 0.2774275243282318, + "learning_rate": 0.0004963459650392093, + "loss": 1.9457, + "step": 871 + }, + { + "epoch": 0.08515625, + "grad_norm": 0.2178190052509308, + "learning_rate": 0.0004963331900180154, + "loss": 2.0192, + "step": 872 + }, + { + "epoch": 0.08525390625, + "grad_norm": 0.2983119487762451, + "learning_rate": 0.0004963203928875967, + "loss": 1.9659, + "step": 873 + }, + { + "epoch": 0.0853515625, + "grad_norm": 0.2878938913345337, + "learning_rate": 0.0004963075736492315, + "loss": 1.9581, + "step": 874 + }, + { + "epoch": 0.08544921875, + "grad_norm": 0.19974899291992188, + "learning_rate": 0.0004962947323042002, + "loss": 1.9898, + "step": 875 + }, + { + "epoch": 0.085546875, + "grad_norm": 0.31860581040382385, + "learning_rate": 0.0004962818688537857, + "loss": 1.9737, + "step": 876 + }, + { + "epoch": 0.08564453125, + "grad_norm": 0.2651256322860718, + "learning_rate": 0.0004962689832992729, + "loss": 1.9977, + "step": 877 + }, + { + "epoch": 0.0857421875, + "grad_norm": 0.286821573972702, + "learning_rate": 0.0004962560756419489, + "loss": 1.9317, + "step": 878 + }, + { + "epoch": 0.08583984375, + "grad_norm": 0.3837340772151947, + "learning_rate": 0.0004962431458831032, + "loss": 1.9419, + "step": 879 + }, + { + "epoch": 0.0859375, + "grad_norm": 0.32222527265548706, + "learning_rate": 0.0004962301940240272, + "loss": 1.9603, + "step": 880 + }, + { + "epoch": 0.08603515625, + "grad_norm": 0.33181512355804443, + "learning_rate": 0.0004962172200660146, + "loss": 1.9646, + "step": 881 + }, + { + "epoch": 0.0861328125, + "grad_norm": 0.2813611626625061, + "learning_rate": 0.0004962042240103617, + "loss": 1.9706, + "step": 882 + }, + { + "epoch": 0.08623046875, + "grad_norm": 0.2933270335197449, + "learning_rate": 0.0004961912058583664, + "loss": 1.9529, + "step": 883 + }, + { + "epoch": 0.086328125, + "grad_norm": 0.31641921401023865, + "learning_rate": 0.0004961781656113292, + "loss": 1.9437, + "step": 884 + }, + { + "epoch": 0.08642578125, + "grad_norm": 0.39271247386932373, + "learning_rate": 0.0004961651032705528, + "loss": 1.957, + "step": 885 + }, + { + "epoch": 0.0865234375, + "grad_norm": 0.3720405697822571, + "learning_rate": 0.000496152018837342, + "loss": 1.9393, + "step": 886 + }, + { + "epoch": 0.08662109375, + "grad_norm": 0.27104172110557556, + "learning_rate": 0.0004961389123130037, + "loss": 1.9547, + "step": 887 + }, + { + "epoch": 0.08671875, + "grad_norm": 0.24182896316051483, + "learning_rate": 0.0004961257836988472, + "loss": 1.9456, + "step": 888 + }, + { + "epoch": 0.08681640625, + "grad_norm": 0.22501450777053833, + "learning_rate": 0.0004961126329961838, + "loss": 1.9336, + "step": 889 + }, + { + "epoch": 0.0869140625, + "grad_norm": 0.2969515323638916, + "learning_rate": 0.0004960994602063273, + "loss": 1.9415, + "step": 890 + }, + { + "epoch": 0.08701171875, + "grad_norm": 0.3117569386959076, + "learning_rate": 0.0004960862653305934, + "loss": 1.8993, + "step": 891 + }, + { + "epoch": 0.087109375, + "grad_norm": 0.32299551367759705, + "learning_rate": 0.0004960730483703004, + "loss": 1.9879, + "step": 892 + }, + { + "epoch": 0.08720703125, + "grad_norm": 0.335592657327652, + "learning_rate": 0.0004960598093267682, + "loss": 1.9665, + "step": 893 + }, + { + "epoch": 0.0873046875, + "grad_norm": 0.21160855889320374, + "learning_rate": 0.0004960465482013197, + "loss": 1.8927, + "step": 894 + }, + { + "epoch": 0.08740234375, + "grad_norm": 0.30621418356895447, + "learning_rate": 0.0004960332649952792, + "loss": 1.9828, + "step": 895 + }, + { + "epoch": 0.0875, + "grad_norm": 0.3541114330291748, + "learning_rate": 0.0004960199597099737, + "loss": 1.9478, + "step": 896 + }, + { + "epoch": 0.08759765625, + "grad_norm": 0.31596067547798157, + "learning_rate": 0.0004960066323467323, + "loss": 1.9355, + "step": 897 + }, + { + "epoch": 0.0876953125, + "grad_norm": 0.3170725107192993, + "learning_rate": 0.0004959932829068863, + "loss": 1.9501, + "step": 898 + }, + { + "epoch": 0.08779296875, + "grad_norm": 0.26953962445259094, + "learning_rate": 0.0004959799113917691, + "loss": 1.9488, + "step": 899 + }, + { + "epoch": 0.087890625, + "grad_norm": 0.23792020976543427, + "learning_rate": 0.0004959665178027165, + "loss": 1.941, + "step": 900 + }, + { + "epoch": 0.08798828125, + "grad_norm": 0.2375611513853073, + "learning_rate": 0.0004959531021410664, + "loss": 1.9133, + "step": 901 + }, + { + "epoch": 0.0880859375, + "grad_norm": 0.20919479429721832, + "learning_rate": 0.0004959396644081588, + "loss": 1.9624, + "step": 902 + }, + { + "epoch": 0.08818359375, + "grad_norm": 0.22488316893577576, + "learning_rate": 0.0004959262046053361, + "loss": 1.9408, + "step": 903 + }, + { + "epoch": 0.08828125, + "grad_norm": 0.2249651849269867, + "learning_rate": 0.0004959127227339427, + "loss": 2.0022, + "step": 904 + }, + { + "epoch": 0.08837890625, + "grad_norm": 0.19234181940555573, + "learning_rate": 0.0004958992187953256, + "loss": 1.9221, + "step": 905 + }, + { + "epoch": 0.0884765625, + "grad_norm": 0.21266022324562073, + "learning_rate": 0.0004958856927908334, + "loss": 1.9796, + "step": 906 + }, + { + "epoch": 0.08857421875, + "grad_norm": 0.22636987268924713, + "learning_rate": 0.0004958721447218174, + "loss": 1.9799, + "step": 907 + }, + { + "epoch": 0.088671875, + "grad_norm": 0.21743637323379517, + "learning_rate": 0.0004958585745896308, + "loss": 1.9156, + "step": 908 + }, + { + "epoch": 0.08876953125, + "grad_norm": 0.316840797662735, + "learning_rate": 0.0004958449823956294, + "loss": 1.9856, + "step": 909 + }, + { + "epoch": 0.0888671875, + "grad_norm": 0.2583603262901306, + "learning_rate": 0.0004958313681411706, + "loss": 1.9601, + "step": 910 + }, + { + "epoch": 0.08896484375, + "grad_norm": 0.2872125208377838, + "learning_rate": 0.0004958177318276146, + "loss": 1.9526, + "step": 911 + }, + { + "epoch": 0.0890625, + "grad_norm": 0.37896475195884705, + "learning_rate": 0.0004958040734563235, + "loss": 1.9465, + "step": 912 + }, + { + "epoch": 0.08916015625, + "grad_norm": 0.3056401014328003, + "learning_rate": 0.0004957903930286616, + "loss": 1.9344, + "step": 913 + }, + { + "epoch": 0.0892578125, + "grad_norm": 0.35092082619667053, + "learning_rate": 0.0004957766905459953, + "loss": 1.9294, + "step": 914 + }, + { + "epoch": 0.08935546875, + "grad_norm": 0.2681847810745239, + "learning_rate": 0.0004957629660096937, + "loss": 1.9436, + "step": 915 + }, + { + "epoch": 0.089453125, + "grad_norm": 0.25762611627578735, + "learning_rate": 0.0004957492194211275, + "loss": 1.931, + "step": 916 + }, + { + "epoch": 0.08955078125, + "grad_norm": 0.24634388089179993, + "learning_rate": 0.00049573545078167, + "loss": 1.9373, + "step": 917 + }, + { + "epoch": 0.0896484375, + "grad_norm": 0.27374333143234253, + "learning_rate": 0.0004957216600926965, + "loss": 2.046, + "step": 918 + }, + { + "epoch": 0.08974609375, + "grad_norm": 0.338375061750412, + "learning_rate": 0.0004957078473555845, + "loss": 2.0259, + "step": 919 + }, + { + "epoch": 0.08984375, + "grad_norm": 0.27678367495536804, + "learning_rate": 0.0004956940125717139, + "loss": 1.9947, + "step": 920 + }, + { + "epoch": 0.08994140625, + "grad_norm": 0.266079306602478, + "learning_rate": 0.0004956801557424666, + "loss": 2.0097, + "step": 921 + }, + { + "epoch": 0.0900390625, + "grad_norm": 0.28650638461112976, + "learning_rate": 0.0004956662768692268, + "loss": 1.9893, + "step": 922 + }, + { + "epoch": 0.09013671875, + "grad_norm": 0.2717742323875427, + "learning_rate": 0.0004956523759533809, + "loss": 1.976, + "step": 923 + }, + { + "epoch": 0.090234375, + "grad_norm": 0.3267741799354553, + "learning_rate": 0.0004956384529963174, + "loss": 1.9803, + "step": 924 + }, + { + "epoch": 0.09033203125, + "grad_norm": 0.32973724603652954, + "learning_rate": 0.0004956245079994272, + "loss": 1.942, + "step": 925 + }, + { + "epoch": 0.0904296875, + "grad_norm": 0.3257177770137787, + "learning_rate": 0.0004956105409641031, + "loss": 1.9584, + "step": 926 + }, + { + "epoch": 0.09052734375, + "grad_norm": 0.31881803274154663, + "learning_rate": 0.0004955965518917405, + "loss": 1.9386, + "step": 927 + }, + { + "epoch": 0.090625, + "grad_norm": 0.2814694046974182, + "learning_rate": 0.0004955825407837365, + "loss": 1.9114, + "step": 928 + }, + { + "epoch": 0.09072265625, + "grad_norm": 0.24232172966003418, + "learning_rate": 0.000495568507641491, + "loss": 1.954, + "step": 929 + }, + { + "epoch": 0.0908203125, + "grad_norm": 0.2062203288078308, + "learning_rate": 0.0004955544524664056, + "loss": 1.9682, + "step": 930 + }, + { + "epoch": 0.09091796875, + "grad_norm": 0.2777712047100067, + "learning_rate": 0.0004955403752598844, + "loss": 1.983, + "step": 931 + }, + { + "epoch": 0.091015625, + "grad_norm": 0.31253954768180847, + "learning_rate": 0.0004955262760233335, + "loss": 1.9211, + "step": 932 + }, + { + "epoch": 0.09111328125, + "grad_norm": 0.292059987783432, + "learning_rate": 0.0004955121547581613, + "loss": 1.9745, + "step": 933 + }, + { + "epoch": 0.0912109375, + "grad_norm": 0.309123694896698, + "learning_rate": 0.0004954980114657783, + "loss": 1.966, + "step": 934 + }, + { + "epoch": 0.09130859375, + "grad_norm": 0.2765352129936218, + "learning_rate": 0.0004954838461475975, + "loss": 1.9535, + "step": 935 + }, + { + "epoch": 0.09140625, + "grad_norm": 0.2576250731945038, + "learning_rate": 0.0004954696588050338, + "loss": 1.9002, + "step": 936 + }, + { + "epoch": 0.09150390625, + "grad_norm": 0.3167097866535187, + "learning_rate": 0.0004954554494395044, + "loss": 1.9249, + "step": 937 + }, + { + "epoch": 0.0916015625, + "grad_norm": 0.32505497336387634, + "learning_rate": 0.0004954412180524285, + "loss": 1.9669, + "step": 938 + }, + { + "epoch": 0.09169921875, + "grad_norm": 0.3114219307899475, + "learning_rate": 0.000495426964645228, + "loss": 1.9276, + "step": 939 + }, + { + "epoch": 0.091796875, + "grad_norm": 0.2786170542240143, + "learning_rate": 0.0004954126892193264, + "loss": 1.965, + "step": 940 + }, + { + "epoch": 0.09189453125, + "grad_norm": 0.2753596901893616, + "learning_rate": 0.00049539839177615, + "loss": 1.9925, + "step": 941 + }, + { + "epoch": 0.0919921875, + "grad_norm": 0.29401320219039917, + "learning_rate": 0.0004953840723171268, + "loss": 1.9772, + "step": 942 + }, + { + "epoch": 0.09208984375, + "grad_norm": 0.3157658278942108, + "learning_rate": 0.0004953697308436871, + "loss": 2.0019, + "step": 943 + }, + { + "epoch": 0.0921875, + "grad_norm": 0.29067566990852356, + "learning_rate": 0.0004953553673572636, + "loss": 1.9753, + "step": 944 + }, + { + "epoch": 0.09228515625, + "grad_norm": 0.28693830966949463, + "learning_rate": 0.0004953409818592911, + "loss": 1.9436, + "step": 945 + }, + { + "epoch": 0.0923828125, + "grad_norm": 0.2536534070968628, + "learning_rate": 0.0004953265743512065, + "loss": 1.9801, + "step": 946 + }, + { + "epoch": 0.09248046875, + "grad_norm": 0.3041675090789795, + "learning_rate": 0.000495312144834449, + "loss": 1.9378, + "step": 947 + }, + { + "epoch": 0.092578125, + "grad_norm": 0.3237321376800537, + "learning_rate": 0.0004952976933104602, + "loss": 1.9781, + "step": 948 + }, + { + "epoch": 0.09267578125, + "grad_norm": 0.25767290592193604, + "learning_rate": 0.0004952832197806835, + "loss": 1.9031, + "step": 949 + }, + { + "epoch": 0.0927734375, + "grad_norm": 0.25587791204452515, + "learning_rate": 0.0004952687242465645, + "loss": 1.9405, + "step": 950 + }, + { + "epoch": 0.09287109375, + "grad_norm": 0.2901047468185425, + "learning_rate": 0.0004952542067095515, + "loss": 1.9567, + "step": 951 + }, + { + "epoch": 0.09296875, + "grad_norm": 0.31517183780670166, + "learning_rate": 0.0004952396671710945, + "loss": 1.9464, + "step": 952 + }, + { + "epoch": 0.09306640625, + "grad_norm": 0.30896711349487305, + "learning_rate": 0.0004952251056326458, + "loss": 1.9096, + "step": 953 + }, + { + "epoch": 0.0931640625, + "grad_norm": 0.26281988620758057, + "learning_rate": 0.0004952105220956602, + "loss": 1.9616, + "step": 954 + }, + { + "epoch": 0.09326171875, + "grad_norm": 0.23971204459667206, + "learning_rate": 0.0004951959165615942, + "loss": 1.9928, + "step": 955 + }, + { + "epoch": 0.093359375, + "grad_norm": 0.22277623414993286, + "learning_rate": 0.000495181289031907, + "loss": 1.9652, + "step": 956 + }, + { + "epoch": 0.09345703125, + "grad_norm": 0.26211845874786377, + "learning_rate": 0.0004951666395080596, + "loss": 1.9617, + "step": 957 + }, + { + "epoch": 0.0935546875, + "grad_norm": 0.24874535202980042, + "learning_rate": 0.0004951519679915155, + "loss": 1.9509, + "step": 958 + }, + { + "epoch": 0.09365234375, + "grad_norm": 0.21750757098197937, + "learning_rate": 0.0004951372744837401, + "loss": 1.966, + "step": 959 + }, + { + "epoch": 0.09375, + "grad_norm": 0.2798830270767212, + "learning_rate": 0.0004951225589862012, + "loss": 1.9574, + "step": 960 + }, + { + "epoch": 0.09384765625, + "grad_norm": 0.33166754245758057, + "learning_rate": 0.0004951078215003688, + "loss": 1.9672, + "step": 961 + }, + { + "epoch": 0.0939453125, + "grad_norm": 0.3122573792934418, + "learning_rate": 0.0004950930620277149, + "loss": 1.9205, + "step": 962 + }, + { + "epoch": 0.09404296875, + "grad_norm": 0.233395054936409, + "learning_rate": 0.0004950782805697141, + "loss": 1.9389, + "step": 963 + }, + { + "epoch": 0.094140625, + "grad_norm": 0.2523380219936371, + "learning_rate": 0.0004950634771278427, + "loss": 1.9643, + "step": 964 + }, + { + "epoch": 0.09423828125, + "grad_norm": 0.2774633765220642, + "learning_rate": 0.0004950486517035796, + "loss": 1.9467, + "step": 965 + }, + { + "epoch": 0.0943359375, + "grad_norm": 0.2782750129699707, + "learning_rate": 0.0004950338042984056, + "loss": 1.9296, + "step": 966 + }, + { + "epoch": 0.09443359375, + "grad_norm": 0.3408234417438507, + "learning_rate": 0.0004950189349138039, + "loss": 1.9705, + "step": 967 + }, + { + "epoch": 0.09453125, + "grad_norm": 0.27266553044319153, + "learning_rate": 0.0004950040435512598, + "loss": 1.9292, + "step": 968 + }, + { + "epoch": 0.09462890625, + "grad_norm": 0.2276991307735443, + "learning_rate": 0.0004949891302122607, + "loss": 1.9793, + "step": 969 + }, + { + "epoch": 0.0947265625, + "grad_norm": 0.23914872109889984, + "learning_rate": 0.0004949741948982965, + "loss": 1.9632, + "step": 970 + }, + { + "epoch": 0.09482421875, + "grad_norm": 0.24836085736751556, + "learning_rate": 0.000494959237610859, + "loss": 1.9304, + "step": 971 + }, + { + "epoch": 0.094921875, + "grad_norm": 0.1932704746723175, + "learning_rate": 0.0004949442583514425, + "loss": 1.9483, + "step": 972 + }, + { + "epoch": 0.09501953125, + "grad_norm": 0.27378395199775696, + "learning_rate": 0.0004949292571215429, + "loss": 1.9687, + "step": 973 + }, + { + "epoch": 0.0951171875, + "grad_norm": 0.3175489008426666, + "learning_rate": 0.0004949142339226591, + "loss": 2.0036, + "step": 974 + }, + { + "epoch": 0.09521484375, + "grad_norm": 0.34563618898391724, + "learning_rate": 0.0004948991887562914, + "loss": 1.9429, + "step": 975 + }, + { + "epoch": 0.0953125, + "grad_norm": 0.2931719720363617, + "learning_rate": 0.000494884121623943, + "loss": 1.9424, + "step": 976 + }, + { + "epoch": 0.09541015625, + "grad_norm": 0.29774370789527893, + "learning_rate": 0.0004948690325271188, + "loss": 1.9297, + "step": 977 + }, + { + "epoch": 0.0955078125, + "grad_norm": 0.31513288617134094, + "learning_rate": 0.0004948539214673261, + "loss": 1.9917, + "step": 978 + }, + { + "epoch": 0.09560546875, + "grad_norm": 0.2150924950838089, + "learning_rate": 0.0004948387884460744, + "loss": 1.9529, + "step": 979 + }, + { + "epoch": 0.095703125, + "grad_norm": 0.27547362446784973, + "learning_rate": 0.0004948236334648753, + "loss": 1.9885, + "step": 980 + }, + { + "epoch": 0.09580078125, + "grad_norm": 0.3271898925304413, + "learning_rate": 0.0004948084565252427, + "loss": 1.9232, + "step": 981 + }, + { + "epoch": 0.0958984375, + "grad_norm": 0.2923772931098938, + "learning_rate": 0.0004947932576286928, + "loss": 1.9429, + "step": 982 + }, + { + "epoch": 0.09599609375, + "grad_norm": 0.2936458885669708, + "learning_rate": 0.0004947780367767434, + "loss": 1.9773, + "step": 983 + }, + { + "epoch": 0.09609375, + "grad_norm": 0.29706791043281555, + "learning_rate": 0.0004947627939709153, + "loss": 1.9284, + "step": 984 + }, + { + "epoch": 0.09619140625, + "grad_norm": 0.22721995413303375, + "learning_rate": 0.000494747529212731, + "loss": 1.9281, + "step": 985 + }, + { + "epoch": 0.0962890625, + "grad_norm": 0.27041006088256836, + "learning_rate": 0.0004947322425037152, + "loss": 1.9735, + "step": 986 + }, + { + "epoch": 0.09638671875, + "grad_norm": 0.26037418842315674, + "learning_rate": 0.0004947169338453952, + "loss": 1.9426, + "step": 987 + }, + { + "epoch": 0.096484375, + "grad_norm": 0.23383264243602753, + "learning_rate": 0.0004947016032392999, + "loss": 1.923, + "step": 988 + }, + { + "epoch": 0.09658203125, + "grad_norm": 0.29238423705101013, + "learning_rate": 0.0004946862506869609, + "loss": 1.9803, + "step": 989 + }, + { + "epoch": 0.0966796875, + "grad_norm": 0.223761647939682, + "learning_rate": 0.0004946708761899116, + "loss": 1.9377, + "step": 990 + }, + { + "epoch": 0.09677734375, + "grad_norm": 0.2669995427131653, + "learning_rate": 0.0004946554797496881, + "loss": 1.9694, + "step": 991 + }, + { + "epoch": 0.096875, + "grad_norm": 0.2847992181777954, + "learning_rate": 0.0004946400613678279, + "loss": 1.9501, + "step": 992 + }, + { + "epoch": 0.09697265625, + "grad_norm": 0.25099360942840576, + "learning_rate": 0.0004946246210458715, + "loss": 1.9707, + "step": 993 + }, + { + "epoch": 0.0970703125, + "grad_norm": 0.23306165635585785, + "learning_rate": 0.0004946091587853611, + "loss": 1.9174, + "step": 994 + }, + { + "epoch": 0.09716796875, + "grad_norm": 0.2907635569572449, + "learning_rate": 0.0004945936745878414, + "loss": 1.9521, + "step": 995 + }, + { + "epoch": 0.097265625, + "grad_norm": 0.3234463334083557, + "learning_rate": 0.0004945781684548591, + "loss": 1.9667, + "step": 996 + }, + { + "epoch": 0.09736328125, + "grad_norm": 0.36099013686180115, + "learning_rate": 0.0004945626403879629, + "loss": 1.9257, + "step": 997 + }, + { + "epoch": 0.0974609375, + "grad_norm": 0.3735397160053253, + "learning_rate": 0.0004945470903887041, + "loss": 1.9512, + "step": 998 + }, + { + "epoch": 0.09755859375, + "grad_norm": 0.3653644025325775, + "learning_rate": 0.0004945315184586359, + "loss": 1.9664, + "step": 999 + }, + { + "epoch": 0.09765625, + "grad_norm": 0.2931279242038727, + "learning_rate": 0.000494515924599314, + "loss": 1.9209, + "step": 1000 + }, + { + "epoch": 0.09775390625, + "grad_norm": 0.31677475571632385, + "learning_rate": 0.000494500308812296, + "loss": 1.9989, + "step": 1001 + }, + { + "epoch": 0.0978515625, + "grad_norm": 0.2640211582183838, + "learning_rate": 0.0004944846710991416, + "loss": 1.8792, + "step": 1002 + }, + { + "epoch": 0.09794921875, + "grad_norm": 0.3207566738128662, + "learning_rate": 0.0004944690114614132, + "loss": 1.955, + "step": 1003 + }, + { + "epoch": 0.098046875, + "grad_norm": 0.29306915402412415, + "learning_rate": 0.0004944533299006749, + "loss": 1.9264, + "step": 1004 + }, + { + "epoch": 0.09814453125, + "grad_norm": 0.2783614993095398, + "learning_rate": 0.000494437626418493, + "loss": 1.9341, + "step": 1005 + }, + { + "epoch": 0.0982421875, + "grad_norm": 0.346975713968277, + "learning_rate": 0.0004944219010164361, + "loss": 1.9599, + "step": 1006 + }, + { + "epoch": 0.09833984375, + "grad_norm": 0.2799683213233948, + "learning_rate": 0.0004944061536960755, + "loss": 1.9201, + "step": 1007 + }, + { + "epoch": 0.0984375, + "grad_norm": 0.27122223377227783, + "learning_rate": 0.0004943903844589837, + "loss": 1.9685, + "step": 1008 + }, + { + "epoch": 0.09853515625, + "grad_norm": 0.34742453694343567, + "learning_rate": 0.0004943745933067362, + "loss": 1.9533, + "step": 1009 + }, + { + "epoch": 0.0986328125, + "grad_norm": 0.2686299681663513, + "learning_rate": 0.0004943587802409103, + "loss": 1.9149, + "step": 1010 + }, + { + "epoch": 0.09873046875, + "grad_norm": 0.3223266303539276, + "learning_rate": 0.0004943429452630856, + "loss": 1.8887, + "step": 1011 + }, + { + "epoch": 0.098828125, + "grad_norm": 0.2767395079135895, + "learning_rate": 0.0004943270883748439, + "loss": 1.9756, + "step": 1012 + }, + { + "epoch": 0.09892578125, + "grad_norm": 0.2761443257331848, + "learning_rate": 0.0004943112095777691, + "loss": 1.981, + "step": 1013 + }, + { + "epoch": 0.0990234375, + "grad_norm": 0.22587864100933075, + "learning_rate": 0.0004942953088734474, + "loss": 1.939, + "step": 1014 + }, + { + "epoch": 0.09912109375, + "grad_norm": 0.2596238851547241, + "learning_rate": 0.0004942793862634671, + "loss": 1.9546, + "step": 1015 + }, + { + "epoch": 0.09921875, + "grad_norm": 0.3035464584827423, + "learning_rate": 0.0004942634417494188, + "loss": 2.0001, + "step": 1016 + }, + { + "epoch": 0.09931640625, + "grad_norm": 0.32868078351020813, + "learning_rate": 0.0004942474753328952, + "loss": 1.9248, + "step": 1017 + }, + { + "epoch": 0.0994140625, + "grad_norm": 0.34833693504333496, + "learning_rate": 0.000494231487015491, + "loss": 1.9148, + "step": 1018 + }, + { + "epoch": 0.09951171875, + "grad_norm": 0.2557753920555115, + "learning_rate": 0.0004942154767988037, + "loss": 1.9585, + "step": 1019 + }, + { + "epoch": 0.099609375, + "grad_norm": 0.25750526785850525, + "learning_rate": 0.0004941994446844321, + "loss": 2.0257, + "step": 1020 + }, + { + "epoch": 0.09970703125, + "grad_norm": 0.22088824212551117, + "learning_rate": 0.000494183390673978, + "loss": 1.9373, + "step": 1021 + }, + { + "epoch": 0.0998046875, + "grad_norm": 0.23292425274848938, + "learning_rate": 0.000494167314769045, + "loss": 1.9462, + "step": 1022 + }, + { + "epoch": 0.09990234375, + "grad_norm": 0.24202801287174225, + "learning_rate": 0.0004941512169712389, + "loss": 2.0237, + "step": 1023 + }, + { + "epoch": 0.1, + "grad_norm": 0.28507405519485474, + "learning_rate": 0.0004941350972821678, + "loss": 1.9373, + "step": 1024 + }, + { + "epoch": 0.10009765625, + "grad_norm": 0.22370004653930664, + "learning_rate": 0.0004941189557034417, + "loss": 1.9332, + "step": 1025 + }, + { + "epoch": 0.1001953125, + "grad_norm": 0.2707929313182831, + "learning_rate": 0.0004941027922366732, + "loss": 1.9908, + "step": 1026 + }, + { + "epoch": 0.10029296875, + "grad_norm": 0.29618075489997864, + "learning_rate": 0.0004940866068834769, + "loss": 1.9317, + "step": 1027 + }, + { + "epoch": 0.100390625, + "grad_norm": 0.27780136466026306, + "learning_rate": 0.0004940703996454695, + "loss": 1.9225, + "step": 1028 + }, + { + "epoch": 0.10048828125, + "grad_norm": 0.33955708146095276, + "learning_rate": 0.0004940541705242698, + "loss": 1.9232, + "step": 1029 + }, + { + "epoch": 0.1005859375, + "grad_norm": 0.3132137954235077, + "learning_rate": 0.0004940379195214992, + "loss": 1.976, + "step": 1030 + }, + { + "epoch": 0.10068359375, + "grad_norm": 0.2742595076560974, + "learning_rate": 0.000494021646638781, + "loss": 1.9138, + "step": 1031 + }, + { + "epoch": 0.10078125, + "grad_norm": 0.2921305298805237, + "learning_rate": 0.0004940053518777406, + "loss": 1.9178, + "step": 1032 + }, + { + "epoch": 0.10087890625, + "grad_norm": 0.2586468458175659, + "learning_rate": 0.0004939890352400058, + "loss": 1.9385, + "step": 1033 + }, + { + "epoch": 0.1009765625, + "grad_norm": 0.30906325578689575, + "learning_rate": 0.0004939726967272064, + "loss": 1.9368, + "step": 1034 + }, + { + "epoch": 0.10107421875, + "grad_norm": 0.3483060300350189, + "learning_rate": 0.0004939563363409745, + "loss": 1.9922, + "step": 1035 + }, + { + "epoch": 0.101171875, + "grad_norm": 0.3559792637825012, + "learning_rate": 0.0004939399540829445, + "loss": 1.9564, + "step": 1036 + }, + { + "epoch": 0.10126953125, + "grad_norm": 0.33506283164024353, + "learning_rate": 0.0004939235499547526, + "loss": 1.9784, + "step": 1037 + }, + { + "epoch": 0.1013671875, + "grad_norm": 0.23956939578056335, + "learning_rate": 0.0004939071239580377, + "loss": 1.9755, + "step": 1038 + }, + { + "epoch": 0.10146484375, + "grad_norm": 0.33348920941352844, + "learning_rate": 0.0004938906760944403, + "loss": 1.9128, + "step": 1039 + }, + { + "epoch": 0.1015625, + "grad_norm": 0.2843862771987915, + "learning_rate": 0.0004938742063656036, + "loss": 1.9297, + "step": 1040 + }, + { + "epoch": 0.10166015625, + "grad_norm": 0.2671546936035156, + "learning_rate": 0.0004938577147731727, + "loss": 1.9391, + "step": 1041 + }, + { + "epoch": 0.1017578125, + "grad_norm": 0.2899993360042572, + "learning_rate": 0.000493841201318795, + "loss": 1.9198, + "step": 1042 + }, + { + "epoch": 0.10185546875, + "grad_norm": 0.20722518861293793, + "learning_rate": 0.0004938246660041201, + "loss": 1.9295, + "step": 1043 + }, + { + "epoch": 0.101953125, + "grad_norm": 0.22667360305786133, + "learning_rate": 0.0004938081088307997, + "loss": 1.9371, + "step": 1044 + }, + { + "epoch": 0.10205078125, + "grad_norm": 0.275579035282135, + "learning_rate": 0.0004937915298004877, + "loss": 1.9383, + "step": 1045 + }, + { + "epoch": 0.1021484375, + "grad_norm": 0.2900732159614563, + "learning_rate": 0.0004937749289148401, + "loss": 1.9805, + "step": 1046 + }, + { + "epoch": 0.10224609375, + "grad_norm": 0.26479572057724, + "learning_rate": 0.0004937583061755153, + "loss": 1.911, + "step": 1047 + }, + { + "epoch": 0.10234375, + "grad_norm": 0.2967575490474701, + "learning_rate": 0.0004937416615841739, + "loss": 1.9654, + "step": 1048 + }, + { + "epoch": 0.10244140625, + "grad_norm": 0.2840502858161926, + "learning_rate": 0.0004937249951424782, + "loss": 1.9332, + "step": 1049 + }, + { + "epoch": 0.1025390625, + "grad_norm": 0.2664666175842285, + "learning_rate": 0.0004937083068520931, + "loss": 1.973, + "step": 1050 + }, + { + "epoch": 0.10263671875, + "grad_norm": 0.2535867989063263, + "learning_rate": 0.000493691596714686, + "loss": 1.9421, + "step": 1051 + }, + { + "epoch": 0.102734375, + "grad_norm": 0.3005189895629883, + "learning_rate": 0.0004936748647319256, + "loss": 1.9434, + "step": 1052 + }, + { + "epoch": 0.10283203125, + "grad_norm": 0.2663765549659729, + "learning_rate": 0.0004936581109054837, + "loss": 1.9947, + "step": 1053 + }, + { + "epoch": 0.1029296875, + "grad_norm": 0.29048940539360046, + "learning_rate": 0.0004936413352370335, + "loss": 1.9737, + "step": 1054 + }, + { + "epoch": 0.10302734375, + "grad_norm": 0.2594558596611023, + "learning_rate": 0.000493624537728251, + "loss": 1.9601, + "step": 1055 + }, + { + "epoch": 0.103125, + "grad_norm": 0.27011555433273315, + "learning_rate": 0.000493607718380814, + "loss": 1.9459, + "step": 1056 + }, + { + "epoch": 0.10322265625, + "grad_norm": 0.3043757677078247, + "learning_rate": 0.0004935908771964026, + "loss": 1.9494, + "step": 1057 + }, + { + "epoch": 0.1033203125, + "grad_norm": 0.26345208287239075, + "learning_rate": 0.0004935740141766991, + "loss": 1.896, + "step": 1058 + }, + { + "epoch": 0.10341796875, + "grad_norm": 0.20400577783584595, + "learning_rate": 0.000493557129323388, + "loss": 1.9276, + "step": 1059 + }, + { + "epoch": 0.103515625, + "grad_norm": 0.2836177349090576, + "learning_rate": 0.0004935402226381558, + "loss": 1.921, + "step": 1060 + }, + { + "epoch": 0.10361328125, + "grad_norm": 0.2904714047908783, + "learning_rate": 0.0004935232941226916, + "loss": 1.9631, + "step": 1061 + }, + { + "epoch": 0.1037109375, + "grad_norm": 0.25980454683303833, + "learning_rate": 0.0004935063437786863, + "loss": 1.9757, + "step": 1062 + }, + { + "epoch": 0.10380859375, + "grad_norm": 0.27394455671310425, + "learning_rate": 0.0004934893716078329, + "loss": 1.9647, + "step": 1063 + }, + { + "epoch": 0.10390625, + "grad_norm": 0.23419839143753052, + "learning_rate": 0.000493472377611827, + "loss": 1.962, + "step": 1064 + }, + { + "epoch": 0.10400390625, + "grad_norm": 0.2323913723230362, + "learning_rate": 0.0004934553617923661, + "loss": 1.9759, + "step": 1065 + }, + { + "epoch": 0.1041015625, + "grad_norm": 0.2700064778327942, + "learning_rate": 0.0004934383241511498, + "loss": 1.9206, + "step": 1066 + }, + { + "epoch": 0.10419921875, + "grad_norm": 0.3530958294868469, + "learning_rate": 0.0004934212646898802, + "loss": 1.979, + "step": 1067 + }, + { + "epoch": 0.104296875, + "grad_norm": 0.30405712127685547, + "learning_rate": 0.0004934041834102612, + "loss": 1.951, + "step": 1068 + }, + { + "epoch": 0.10439453125, + "grad_norm": 0.2923583984375, + "learning_rate": 0.0004933870803139992, + "loss": 2.0158, + "step": 1069 + }, + { + "epoch": 0.1044921875, + "grad_norm": 0.23782993853092194, + "learning_rate": 0.0004933699554028027, + "loss": 1.896, + "step": 1070 + }, + { + "epoch": 0.10458984375, + "grad_norm": 0.3045034110546112, + "learning_rate": 0.0004933528086783821, + "loss": 1.9555, + "step": 1071 + }, + { + "epoch": 0.1046875, + "grad_norm": 0.40851131081581116, + "learning_rate": 0.0004933356401424505, + "loss": 1.9108, + "step": 1072 + }, + { + "epoch": 0.10478515625, + "grad_norm": 0.39787429571151733, + "learning_rate": 0.0004933184497967226, + "loss": 1.9472, + "step": 1073 + }, + { + "epoch": 0.1048828125, + "grad_norm": 0.24665354192256927, + "learning_rate": 0.0004933012376429158, + "loss": 1.9617, + "step": 1074 + }, + { + "epoch": 0.10498046875, + "grad_norm": 0.28025394678115845, + "learning_rate": 0.0004932840036827492, + "loss": 1.898, + "step": 1075 + }, + { + "epoch": 0.105078125, + "grad_norm": 0.3374732434749603, + "learning_rate": 0.0004932667479179445, + "loss": 1.9323, + "step": 1076 + }, + { + "epoch": 0.10517578125, + "grad_norm": 0.2941461503505707, + "learning_rate": 0.0004932494703502254, + "loss": 1.964, + "step": 1077 + }, + { + "epoch": 0.1052734375, + "grad_norm": 0.2827446758747101, + "learning_rate": 0.0004932321709813177, + "loss": 1.9554, + "step": 1078 + }, + { + "epoch": 0.10537109375, + "grad_norm": 0.32132452726364136, + "learning_rate": 0.0004932148498129495, + "loss": 1.9333, + "step": 1079 + }, + { + "epoch": 0.10546875, + "grad_norm": 0.2732277810573578, + "learning_rate": 0.000493197506846851, + "loss": 1.932, + "step": 1080 + }, + { + "epoch": 0.10556640625, + "grad_norm": 0.2594139575958252, + "learning_rate": 0.0004931801420847545, + "loss": 1.9447, + "step": 1081 + }, + { + "epoch": 0.1056640625, + "grad_norm": 0.24399201571941376, + "learning_rate": 0.0004931627555283948, + "loss": 1.9163, + "step": 1082 + }, + { + "epoch": 0.10576171875, + "grad_norm": 0.22696708142757416, + "learning_rate": 0.0004931453471795087, + "loss": 1.9734, + "step": 1083 + }, + { + "epoch": 0.105859375, + "grad_norm": 0.25717827677726746, + "learning_rate": 0.0004931279170398349, + "loss": 1.9767, + "step": 1084 + }, + { + "epoch": 0.10595703125, + "grad_norm": 0.23899690806865692, + "learning_rate": 0.0004931104651111146, + "loss": 1.9519, + "step": 1085 + }, + { + "epoch": 0.1060546875, + "grad_norm": 0.27689501643180847, + "learning_rate": 0.0004930929913950911, + "loss": 1.9582, + "step": 1086 + }, + { + "epoch": 0.10615234375, + "grad_norm": 0.23959669470787048, + "learning_rate": 0.00049307549589351, + "loss": 1.9369, + "step": 1087 + }, + { + "epoch": 0.10625, + "grad_norm": 0.25673651695251465, + "learning_rate": 0.0004930579786081188, + "loss": 1.9711, + "step": 1088 + }, + { + "epoch": 0.10634765625, + "grad_norm": 0.30868396162986755, + "learning_rate": 0.0004930404395406674, + "loss": 1.9539, + "step": 1089 + }, + { + "epoch": 0.1064453125, + "grad_norm": 0.2462042272090912, + "learning_rate": 0.0004930228786929076, + "loss": 1.9061, + "step": 1090 + }, + { + "epoch": 0.10654296875, + "grad_norm": 0.24275164306163788, + "learning_rate": 0.0004930052960665938, + "loss": 1.8992, + "step": 1091 + }, + { + "epoch": 0.106640625, + "grad_norm": 0.27242550253868103, + "learning_rate": 0.0004929876916634823, + "loss": 1.8997, + "step": 1092 + }, + { + "epoch": 0.10673828125, + "grad_norm": 0.2655372619628906, + "learning_rate": 0.0004929700654853316, + "loss": 1.8905, + "step": 1093 + }, + { + "epoch": 0.1068359375, + "grad_norm": 0.27779421210289, + "learning_rate": 0.0004929524175339024, + "loss": 1.9485, + "step": 1094 + }, + { + "epoch": 0.10693359375, + "grad_norm": 0.3308311402797699, + "learning_rate": 0.0004929347478109576, + "loss": 1.896, + "step": 1095 + }, + { + "epoch": 0.10703125, + "grad_norm": 0.2988120913505554, + "learning_rate": 0.0004929170563182621, + "loss": 1.9508, + "step": 1096 + }, + { + "epoch": 0.10712890625, + "grad_norm": 0.24604608118534088, + "learning_rate": 0.0004928993430575834, + "loss": 1.9912, + "step": 1097 + }, + { + "epoch": 0.1072265625, + "grad_norm": 0.31434887647628784, + "learning_rate": 0.0004928816080306906, + "loss": 1.9592, + "step": 1098 + }, + { + "epoch": 0.10732421875, + "grad_norm": 0.21227046847343445, + "learning_rate": 0.0004928638512393555, + "loss": 1.9523, + "step": 1099 + }, + { + "epoch": 0.107421875, + "grad_norm": 0.26643091440200806, + "learning_rate": 0.0004928460726853518, + "loss": 1.9506, + "step": 1100 + }, + { + "epoch": 0.10751953125, + "grad_norm": 0.2661186754703522, + "learning_rate": 0.0004928282723704554, + "loss": 1.9425, + "step": 1101 + }, + { + "epoch": 0.1076171875, + "grad_norm": 0.25391289591789246, + "learning_rate": 0.0004928104502964442, + "loss": 1.9551, + "step": 1102 + }, + { + "epoch": 0.10771484375, + "grad_norm": 0.276909202337265, + "learning_rate": 0.0004927926064650988, + "loss": 1.9416, + "step": 1103 + }, + { + "epoch": 0.1078125, + "grad_norm": 0.25135815143585205, + "learning_rate": 0.0004927747408782014, + "loss": 1.9197, + "step": 1104 + }, + { + "epoch": 0.10791015625, + "grad_norm": 0.2529551684856415, + "learning_rate": 0.0004927568535375368, + "loss": 1.9561, + "step": 1105 + }, + { + "epoch": 0.1080078125, + "grad_norm": 0.2697699964046478, + "learning_rate": 0.0004927389444448915, + "loss": 1.9369, + "step": 1106 + }, + { + "epoch": 0.10810546875, + "grad_norm": 0.2805427014827728, + "learning_rate": 0.0004927210136020548, + "loss": 1.9535, + "step": 1107 + }, + { + "epoch": 0.108203125, + "grad_norm": 0.33747565746307373, + "learning_rate": 0.0004927030610108176, + "loss": 1.9582, + "step": 1108 + }, + { + "epoch": 0.10830078125, + "grad_norm": 0.30215132236480713, + "learning_rate": 0.0004926850866729734, + "loss": 1.9245, + "step": 1109 + }, + { + "epoch": 0.1083984375, + "grad_norm": 0.3483729064464569, + "learning_rate": 0.0004926670905903174, + "loss": 1.9959, + "step": 1110 + }, + { + "epoch": 0.10849609375, + "grad_norm": 0.23263247311115265, + "learning_rate": 0.0004926490727646476, + "loss": 1.9224, + "step": 1111 + }, + { + "epoch": 0.10859375, + "grad_norm": 0.3023853898048401, + "learning_rate": 0.0004926310331977633, + "loss": 1.9493, + "step": 1112 + }, + { + "epoch": 0.10869140625, + "grad_norm": 0.3592170774936676, + "learning_rate": 0.0004926129718914672, + "loss": 1.9082, + "step": 1113 + }, + { + "epoch": 0.1087890625, + "grad_norm": 0.29466864466667175, + "learning_rate": 0.0004925948888475628, + "loss": 2.016, + "step": 1114 + }, + { + "epoch": 0.10888671875, + "grad_norm": 0.3136419355869293, + "learning_rate": 0.0004925767840678569, + "loss": 1.9592, + "step": 1115 + }, + { + "epoch": 0.108984375, + "grad_norm": 0.21686695516109467, + "learning_rate": 0.0004925586575541578, + "loss": 1.9081, + "step": 1116 + }, + { + "epoch": 0.10908203125, + "grad_norm": 0.2520851790904999, + "learning_rate": 0.0004925405093082761, + "loss": 1.9418, + "step": 1117 + }, + { + "epoch": 0.1091796875, + "grad_norm": 0.28278669714927673, + "learning_rate": 0.0004925223393320247, + "loss": 1.9815, + "step": 1118 + }, + { + "epoch": 0.10927734375, + "grad_norm": 0.2773968577384949, + "learning_rate": 0.0004925041476272189, + "loss": 1.8888, + "step": 1119 + }, + { + "epoch": 0.109375, + "grad_norm": 0.24399970471858978, + "learning_rate": 0.0004924859341956755, + "loss": 1.9548, + "step": 1120 + }, + { + "epoch": 0.10947265625, + "grad_norm": 0.26073598861694336, + "learning_rate": 0.0004924676990392141, + "loss": 1.9112, + "step": 1121 + }, + { + "epoch": 0.1095703125, + "grad_norm": 0.24541547894477844, + "learning_rate": 0.0004924494421596562, + "loss": 1.914, + "step": 1122 + }, + { + "epoch": 0.10966796875, + "grad_norm": 0.21719145774841309, + "learning_rate": 0.0004924311635588254, + "loss": 1.9204, + "step": 1123 + }, + { + "epoch": 0.109765625, + "grad_norm": 0.295460045337677, + "learning_rate": 0.0004924128632385476, + "loss": 1.9545, + "step": 1124 + }, + { + "epoch": 0.10986328125, + "grad_norm": 0.3058072626590729, + "learning_rate": 0.0004923945412006509, + "loss": 1.9048, + "step": 1125 + }, + { + "epoch": 0.1099609375, + "grad_norm": 0.19512003660202026, + "learning_rate": 0.0004923761974469655, + "loss": 1.9581, + "step": 1126 + }, + { + "epoch": 0.11005859375, + "grad_norm": 0.21684713661670685, + "learning_rate": 0.0004923578319793237, + "loss": 1.8765, + "step": 1127 + }, + { + "epoch": 0.11015625, + "grad_norm": 0.2229132503271103, + "learning_rate": 0.0004923394447995601, + "loss": 1.8599, + "step": 1128 + }, + { + "epoch": 0.11025390625, + "grad_norm": 0.20479273796081543, + "learning_rate": 0.0004923210359095115, + "loss": 1.938, + "step": 1129 + }, + { + "epoch": 0.1103515625, + "grad_norm": 0.2477833479642868, + "learning_rate": 0.0004923026053110167, + "loss": 1.9655, + "step": 1130 + }, + { + "epoch": 0.11044921875, + "grad_norm": 0.23927846550941467, + "learning_rate": 0.0004922841530059167, + "loss": 1.9461, + "step": 1131 + }, + { + "epoch": 0.110546875, + "grad_norm": 0.24189342558383942, + "learning_rate": 0.0004922656789960547, + "loss": 1.9338, + "step": 1132 + }, + { + "epoch": 0.11064453125, + "grad_norm": 0.20208942890167236, + "learning_rate": 0.0004922471832832763, + "loss": 1.973, + "step": 1133 + }, + { + "epoch": 0.1107421875, + "grad_norm": 0.20967181026935577, + "learning_rate": 0.0004922286658694289, + "loss": 1.9391, + "step": 1134 + }, + { + "epoch": 0.11083984375, + "grad_norm": 0.2538667917251587, + "learning_rate": 0.0004922101267563622, + "loss": 1.9159, + "step": 1135 + }, + { + "epoch": 0.1109375, + "grad_norm": 0.2945930063724518, + "learning_rate": 0.0004921915659459281, + "loss": 1.915, + "step": 1136 + }, + { + "epoch": 0.11103515625, + "grad_norm": 0.3631344437599182, + "learning_rate": 0.0004921729834399808, + "loss": 1.9926, + "step": 1137 + }, + { + "epoch": 0.1111328125, + "grad_norm": 0.32187792658805847, + "learning_rate": 0.0004921543792403765, + "loss": 1.9137, + "step": 1138 + }, + { + "epoch": 0.11123046875, + "grad_norm": 0.276929646730423, + "learning_rate": 0.0004921357533489734, + "loss": 1.9427, + "step": 1139 + }, + { + "epoch": 0.111328125, + "grad_norm": 0.20455695688724518, + "learning_rate": 0.0004921171057676323, + "loss": 1.9118, + "step": 1140 + }, + { + "epoch": 0.11142578125, + "grad_norm": 0.2824450731277466, + "learning_rate": 0.0004920984364982157, + "loss": 1.9471, + "step": 1141 + }, + { + "epoch": 0.1115234375, + "grad_norm": 0.2931687831878662, + "learning_rate": 0.0004920797455425887, + "loss": 1.9546, + "step": 1142 + }, + { + "epoch": 0.11162109375, + "grad_norm": 0.31234028935432434, + "learning_rate": 0.0004920610329026181, + "loss": 1.9466, + "step": 1143 + }, + { + "epoch": 0.11171875, + "grad_norm": 0.39190223813056946, + "learning_rate": 0.0004920422985801735, + "loss": 1.9394, + "step": 1144 + }, + { + "epoch": 0.11181640625, + "grad_norm": 0.26808232069015503, + "learning_rate": 0.0004920235425771261, + "loss": 1.9278, + "step": 1145 + }, + { + "epoch": 0.1119140625, + "grad_norm": 0.22590310871601105, + "learning_rate": 0.0004920047648953493, + "loss": 1.9467, + "step": 1146 + }, + { + "epoch": 0.11201171875, + "grad_norm": 0.2531168758869171, + "learning_rate": 0.0004919859655367192, + "loss": 1.958, + "step": 1147 + }, + { + "epoch": 0.112109375, + "grad_norm": 0.207502081990242, + "learning_rate": 0.0004919671445031132, + "loss": 2.0148, + "step": 1148 + }, + { + "epoch": 0.11220703125, + "grad_norm": 0.23596623539924622, + "learning_rate": 0.0004919483017964118, + "loss": 1.9617, + "step": 1149 + }, + { + "epoch": 0.1123046875, + "grad_norm": 0.25810694694519043, + "learning_rate": 0.000491929437418497, + "loss": 2.0262, + "step": 1150 + }, + { + "epoch": 0.11240234375, + "grad_norm": 0.27257680892944336, + "learning_rate": 0.0004919105513712533, + "loss": 1.9285, + "step": 1151 + }, + { + "epoch": 0.1125, + "grad_norm": 0.2861190736293793, + "learning_rate": 0.000491891643656567, + "loss": 1.965, + "step": 1152 + }, + { + "epoch": 0.11259765625, + "grad_norm": 0.2790696620941162, + "learning_rate": 0.0004918727142763271, + "loss": 1.9572, + "step": 1153 + }, + { + "epoch": 0.1126953125, + "grad_norm": 0.35226336121559143, + "learning_rate": 0.0004918537632324243, + "loss": 1.9338, + "step": 1154 + }, + { + "epoch": 0.11279296875, + "grad_norm": 0.2735171616077423, + "learning_rate": 0.0004918347905267517, + "loss": 1.9304, + "step": 1155 + }, + { + "epoch": 0.112890625, + "grad_norm": 0.2629276514053345, + "learning_rate": 0.0004918157961612046, + "loss": 1.9383, + "step": 1156 + }, + { + "epoch": 0.11298828125, + "grad_norm": 0.3583827614784241, + "learning_rate": 0.0004917967801376802, + "loss": 1.9448, + "step": 1157 + }, + { + "epoch": 0.1130859375, + "grad_norm": 0.2546387016773224, + "learning_rate": 0.0004917777424580781, + "loss": 1.9506, + "step": 1158 + }, + { + "epoch": 0.11318359375, + "grad_norm": 0.35346466302871704, + "learning_rate": 0.0004917586831243001, + "loss": 1.9207, + "step": 1159 + }, + { + "epoch": 0.11328125, + "grad_norm": 0.2943922281265259, + "learning_rate": 0.0004917396021382499, + "loss": 1.961, + "step": 1160 + }, + { + "epoch": 0.11337890625, + "grad_norm": 0.24972833693027496, + "learning_rate": 0.0004917204995018336, + "loss": 1.9271, + "step": 1161 + }, + { + "epoch": 0.1134765625, + "grad_norm": 0.31200850009918213, + "learning_rate": 0.0004917013752169594, + "loss": 1.9444, + "step": 1162 + }, + { + "epoch": 0.11357421875, + "grad_norm": 0.3580279052257538, + "learning_rate": 0.0004916822292855375, + "loss": 1.9132, + "step": 1163 + }, + { + "epoch": 0.113671875, + "grad_norm": 0.31269049644470215, + "learning_rate": 0.0004916630617094807, + "loss": 1.9283, + "step": 1164 + }, + { + "epoch": 0.11376953125, + "grad_norm": 0.25159773230552673, + "learning_rate": 0.0004916438724907033, + "loss": 1.9018, + "step": 1165 + }, + { + "epoch": 0.1138671875, + "grad_norm": 0.22316916286945343, + "learning_rate": 0.0004916246616311225, + "loss": 1.8821, + "step": 1166 + }, + { + "epoch": 0.11396484375, + "grad_norm": 0.31150946021080017, + "learning_rate": 0.0004916054291326571, + "loss": 1.9565, + "step": 1167 + }, + { + "epoch": 0.1140625, + "grad_norm": 0.29112792015075684, + "learning_rate": 0.0004915861749972281, + "loss": 1.9582, + "step": 1168 + }, + { + "epoch": 0.11416015625, + "grad_norm": 0.26860764622688293, + "learning_rate": 0.0004915668992267593, + "loss": 1.9299, + "step": 1169 + }, + { + "epoch": 0.1142578125, + "grad_norm": 0.24470628798007965, + "learning_rate": 0.0004915476018231757, + "loss": 1.9131, + "step": 1170 + }, + { + "epoch": 0.11435546875, + "grad_norm": 0.26279789209365845, + "learning_rate": 0.0004915282827884051, + "loss": 1.9055, + "step": 1171 + }, + { + "epoch": 0.114453125, + "grad_norm": 0.21272587776184082, + "learning_rate": 0.0004915089421243773, + "loss": 1.9694, + "step": 1172 + }, + { + "epoch": 0.11455078125, + "grad_norm": 0.21711157262325287, + "learning_rate": 0.0004914895798330244, + "loss": 1.9644, + "step": 1173 + }, + { + "epoch": 0.1146484375, + "grad_norm": 0.21599583327770233, + "learning_rate": 0.0004914701959162803, + "loss": 1.9528, + "step": 1174 + }, + { + "epoch": 0.11474609375, + "grad_norm": 0.22919407486915588, + "learning_rate": 0.0004914507903760814, + "loss": 1.9353, + "step": 1175 + }, + { + "epoch": 0.11484375, + "grad_norm": 0.22974969446659088, + "learning_rate": 0.0004914313632143661, + "loss": 1.9148, + "step": 1176 + }, + { + "epoch": 0.11494140625, + "grad_norm": 0.2370641529560089, + "learning_rate": 0.000491411914433075, + "loss": 1.9056, + "step": 1177 + }, + { + "epoch": 0.1150390625, + "grad_norm": 0.27402329444885254, + "learning_rate": 0.000491392444034151, + "loss": 1.9483, + "step": 1178 + }, + { + "epoch": 0.11513671875, + "grad_norm": 0.2953847050666809, + "learning_rate": 0.0004913729520195388, + "loss": 1.9553, + "step": 1179 + }, + { + "epoch": 0.115234375, + "grad_norm": 0.26085543632507324, + "learning_rate": 0.0004913534383911857, + "loss": 1.9307, + "step": 1180 + }, + { + "epoch": 0.11533203125, + "grad_norm": 0.2782340943813324, + "learning_rate": 0.0004913339031510408, + "loss": 1.9733, + "step": 1181 + }, + { + "epoch": 0.1154296875, + "grad_norm": 0.2568228244781494, + "learning_rate": 0.0004913143463010554, + "loss": 1.9402, + "step": 1182 + }, + { + "epoch": 0.11552734375, + "grad_norm": 0.2584892809391022, + "learning_rate": 0.0004912947678431833, + "loss": 1.9254, + "step": 1183 + }, + { + "epoch": 0.115625, + "grad_norm": 0.2621140778064728, + "learning_rate": 0.00049127516777938, + "loss": 1.9323, + "step": 1184 + }, + { + "epoch": 0.11572265625, + "grad_norm": 0.28239843249320984, + "learning_rate": 0.0004912555461116035, + "loss": 1.9447, + "step": 1185 + }, + { + "epoch": 0.1158203125, + "grad_norm": 0.3147210478782654, + "learning_rate": 0.0004912359028418138, + "loss": 1.8867, + "step": 1186 + }, + { + "epoch": 0.11591796875, + "grad_norm": 0.25541990995407104, + "learning_rate": 0.0004912162379719731, + "loss": 1.871, + "step": 1187 + }, + { + "epoch": 0.116015625, + "grad_norm": 0.24328139424324036, + "learning_rate": 0.0004911965515040457, + "loss": 1.9316, + "step": 1188 + }, + { + "epoch": 0.11611328125, + "grad_norm": 0.20230042934417725, + "learning_rate": 0.0004911768434399981, + "loss": 1.9298, + "step": 1189 + }, + { + "epoch": 0.1162109375, + "grad_norm": 0.32389941811561584, + "learning_rate": 0.0004911571137817991, + "loss": 1.8941, + "step": 1190 + }, + { + "epoch": 0.11630859375, + "grad_norm": 0.3173327147960663, + "learning_rate": 0.0004911373625314194, + "loss": 1.9222, + "step": 1191 + }, + { + "epoch": 0.11640625, + "grad_norm": 0.32637453079223633, + "learning_rate": 0.0004911175896908319, + "loss": 1.9273, + "step": 1192 + }, + { + "epoch": 0.11650390625, + "grad_norm": 0.307746022939682, + "learning_rate": 0.0004910977952620119, + "loss": 1.9352, + "step": 1193 + }, + { + "epoch": 0.1166015625, + "grad_norm": 0.21165402233600616, + "learning_rate": 0.0004910779792469365, + "loss": 1.8867, + "step": 1194 + }, + { + "epoch": 0.11669921875, + "grad_norm": 0.3059009909629822, + "learning_rate": 0.0004910581416475854, + "loss": 1.9187, + "step": 1195 + }, + { + "epoch": 0.116796875, + "grad_norm": 0.2690452039241791, + "learning_rate": 0.00049103828246594, + "loss": 1.9723, + "step": 1196 + }, + { + "epoch": 0.11689453125, + "grad_norm": 0.25487321615219116, + "learning_rate": 0.0004910184017039841, + "loss": 1.9525, + "step": 1197 + }, + { + "epoch": 0.1169921875, + "grad_norm": 0.33227404952049255, + "learning_rate": 0.0004909984993637036, + "loss": 1.9062, + "step": 1198 + }, + { + "epoch": 0.11708984375, + "grad_norm": 0.32993388175964355, + "learning_rate": 0.0004909785754470866, + "loss": 1.9327, + "step": 1199 + }, + { + "epoch": 0.1171875, + "grad_norm": 0.28574901819229126, + "learning_rate": 0.0004909586299561233, + "loss": 1.9481, + "step": 1200 + }, + { + "epoch": 0.11728515625, + "grad_norm": 0.27981141209602356, + "learning_rate": 0.0004909386628928061, + "loss": 1.9284, + "step": 1201 + }, + { + "epoch": 0.1173828125, + "grad_norm": 0.26131895184516907, + "learning_rate": 0.0004909186742591296, + "loss": 1.9276, + "step": 1202 + }, + { + "epoch": 0.11748046875, + "grad_norm": 0.3690519630908966, + "learning_rate": 0.0004908986640570903, + "loss": 1.8733, + "step": 1203 + }, + { + "epoch": 0.117578125, + "grad_norm": 0.38982152938842773, + "learning_rate": 0.000490878632288687, + "loss": 1.9204, + "step": 1204 + }, + { + "epoch": 0.11767578125, + "grad_norm": 0.2588847875595093, + "learning_rate": 0.000490858578955921, + "loss": 1.8941, + "step": 1205 + }, + { + "epoch": 0.1177734375, + "grad_norm": 0.3161073625087738, + "learning_rate": 0.0004908385040607954, + "loss": 1.9077, + "step": 1206 + }, + { + "epoch": 0.11787109375, + "grad_norm": 0.26444011926651, + "learning_rate": 0.0004908184076053153, + "loss": 1.9471, + "step": 1207 + }, + { + "epoch": 0.11796875, + "grad_norm": 0.28895479440689087, + "learning_rate": 0.0004907982895914883, + "loss": 1.9159, + "step": 1208 + }, + { + "epoch": 0.11806640625, + "grad_norm": 0.264992356300354, + "learning_rate": 0.000490778150021324, + "loss": 1.9383, + "step": 1209 + }, + { + "epoch": 0.1181640625, + "grad_norm": 0.25046393275260925, + "learning_rate": 0.0004907579888968341, + "loss": 1.9756, + "step": 1210 + }, + { + "epoch": 0.11826171875, + "grad_norm": 0.2875134348869324, + "learning_rate": 0.0004907378062200325, + "loss": 1.9368, + "step": 1211 + }, + { + "epoch": 0.118359375, + "grad_norm": 0.24127423763275146, + "learning_rate": 0.0004907176019929353, + "loss": 1.9399, + "step": 1212 + }, + { + "epoch": 0.11845703125, + "grad_norm": 0.25520575046539307, + "learning_rate": 0.0004906973762175608, + "loss": 1.9662, + "step": 1213 + }, + { + "epoch": 0.1185546875, + "grad_norm": 0.23595774173736572, + "learning_rate": 0.0004906771288959293, + "loss": 1.9302, + "step": 1214 + }, + { + "epoch": 0.11865234375, + "grad_norm": 0.2919490337371826, + "learning_rate": 0.0004906568600300635, + "loss": 1.9421, + "step": 1215 + }, + { + "epoch": 0.11875, + "grad_norm": 0.23202334344387054, + "learning_rate": 0.0004906365696219877, + "loss": 1.8934, + "step": 1216 + }, + { + "epoch": 0.11884765625, + "grad_norm": 0.2639678716659546, + "learning_rate": 0.0004906162576737292, + "loss": 1.9272, + "step": 1217 + }, + { + "epoch": 0.1189453125, + "grad_norm": 0.24962055683135986, + "learning_rate": 0.0004905959241873165, + "loss": 1.9605, + "step": 1218 + }, + { + "epoch": 0.11904296875, + "grad_norm": 0.2519899308681488, + "learning_rate": 0.0004905755691647812, + "loss": 1.9757, + "step": 1219 + }, + { + "epoch": 0.119140625, + "grad_norm": 0.2044687420129776, + "learning_rate": 0.0004905551926081563, + "loss": 1.9817, + "step": 1220 + }, + { + "epoch": 0.11923828125, + "grad_norm": 0.21942391991615295, + "learning_rate": 0.0004905347945194772, + "loss": 1.9705, + "step": 1221 + }, + { + "epoch": 0.1193359375, + "grad_norm": 0.26614969968795776, + "learning_rate": 0.0004905143749007817, + "loss": 1.9375, + "step": 1222 + }, + { + "epoch": 0.11943359375, + "grad_norm": 0.23212748765945435, + "learning_rate": 0.0004904939337541093, + "loss": 1.9728, + "step": 1223 + }, + { + "epoch": 0.11953125, + "grad_norm": 0.28004127740859985, + "learning_rate": 0.0004904734710815022, + "loss": 1.9082, + "step": 1224 + }, + { + "epoch": 0.11962890625, + "grad_norm": 0.3012614846229553, + "learning_rate": 0.0004904529868850042, + "loss": 1.9347, + "step": 1225 + }, + { + "epoch": 0.1197265625, + "grad_norm": 0.2914253771305084, + "learning_rate": 0.0004904324811666615, + "loss": 1.9435, + "step": 1226 + }, + { + "epoch": 0.11982421875, + "grad_norm": 0.2315436154603958, + "learning_rate": 0.0004904119539285225, + "loss": 1.9503, + "step": 1227 + }, + { + "epoch": 0.119921875, + "grad_norm": 0.2043173611164093, + "learning_rate": 0.0004903914051726377, + "loss": 1.937, + "step": 1228 + }, + { + "epoch": 0.12001953125, + "grad_norm": 0.26184800267219543, + "learning_rate": 0.0004903708349010598, + "loss": 1.9459, + "step": 1229 + }, + { + "epoch": 0.1201171875, + "grad_norm": 0.24898020923137665, + "learning_rate": 0.0004903502431158434, + "loss": 1.935, + "step": 1230 + }, + { + "epoch": 0.12021484375, + "grad_norm": 0.29706010222435, + "learning_rate": 0.0004903296298190456, + "loss": 1.9741, + "step": 1231 + }, + { + "epoch": 0.1203125, + "grad_norm": 0.3235405385494232, + "learning_rate": 0.0004903089950127254, + "loss": 1.9319, + "step": 1232 + }, + { + "epoch": 0.12041015625, + "grad_norm": 0.25629666447639465, + "learning_rate": 0.0004902883386989441, + "loss": 1.9533, + "step": 1233 + }, + { + "epoch": 0.1205078125, + "grad_norm": 0.23315580189228058, + "learning_rate": 0.0004902676608797649, + "loss": 1.9033, + "step": 1234 + }, + { + "epoch": 0.12060546875, + "grad_norm": 0.2336481362581253, + "learning_rate": 0.0004902469615572537, + "loss": 1.9123, + "step": 1235 + }, + { + "epoch": 0.120703125, + "grad_norm": 0.27162161469459534, + "learning_rate": 0.0004902262407334779, + "loss": 1.9253, + "step": 1236 + }, + { + "epoch": 0.12080078125, + "grad_norm": 0.2913827896118164, + "learning_rate": 0.0004902054984105073, + "loss": 1.9526, + "step": 1237 + }, + { + "epoch": 0.1208984375, + "grad_norm": 0.30284538865089417, + "learning_rate": 0.0004901847345904141, + "loss": 1.9588, + "step": 1238 + }, + { + "epoch": 0.12099609375, + "grad_norm": 0.2579716742038727, + "learning_rate": 0.000490163949275272, + "loss": 1.892, + "step": 1239 + }, + { + "epoch": 0.12109375, + "grad_norm": 0.21740810573101044, + "learning_rate": 0.0004901431424671577, + "loss": 1.9584, + "step": 1240 + }, + { + "epoch": 0.12119140625, + "grad_norm": 0.3371858298778534, + "learning_rate": 0.0004901223141681496, + "loss": 1.9415, + "step": 1241 + }, + { + "epoch": 0.1212890625, + "grad_norm": 0.34596753120422363, + "learning_rate": 0.000490101464380328, + "loss": 1.9384, + "step": 1242 + }, + { + "epoch": 0.12138671875, + "grad_norm": 0.2591091990470886, + "learning_rate": 0.0004900805931057757, + "loss": 1.9433, + "step": 1243 + }, + { + "epoch": 0.121484375, + "grad_norm": 0.27759140729904175, + "learning_rate": 0.0004900597003465777, + "loss": 1.9124, + "step": 1244 + }, + { + "epoch": 0.12158203125, + "grad_norm": 0.2503783106803894, + "learning_rate": 0.0004900387861048208, + "loss": 1.9426, + "step": 1245 + }, + { + "epoch": 0.1216796875, + "grad_norm": 0.36085277795791626, + "learning_rate": 0.0004900178503825942, + "loss": 1.9141, + "step": 1246 + }, + { + "epoch": 0.12177734375, + "grad_norm": 0.2831665873527527, + "learning_rate": 0.0004899968931819893, + "loss": 1.9515, + "step": 1247 + }, + { + "epoch": 0.121875, + "grad_norm": 0.26685988903045654, + "learning_rate": 0.0004899759145050996, + "loss": 1.9752, + "step": 1248 + }, + { + "epoch": 0.12197265625, + "grad_norm": 0.3139730393886566, + "learning_rate": 0.0004899549143540205, + "loss": 1.9298, + "step": 1249 + }, + { + "epoch": 0.1220703125, + "grad_norm": 0.20647141337394714, + "learning_rate": 0.0004899338927308497, + "loss": 1.9233, + "step": 1250 + }, + { + "epoch": 0.12216796875, + "grad_norm": 0.293544203042984, + "learning_rate": 0.0004899128496376871, + "loss": 1.9534, + "step": 1251 + }, + { + "epoch": 0.122265625, + "grad_norm": 0.28824836015701294, + "learning_rate": 0.0004898917850766349, + "loss": 1.96, + "step": 1252 + }, + { + "epoch": 0.12236328125, + "grad_norm": 0.23198992013931274, + "learning_rate": 0.0004898706990497971, + "loss": 1.9275, + "step": 1253 + }, + { + "epoch": 0.1224609375, + "grad_norm": 0.27970659732818604, + "learning_rate": 0.00048984959155928, + "loss": 1.8998, + "step": 1254 + }, + { + "epoch": 0.12255859375, + "grad_norm": 0.2355661541223526, + "learning_rate": 0.0004898284626071921, + "loss": 1.9313, + "step": 1255 + }, + { + "epoch": 0.12265625, + "grad_norm": 0.3456537127494812, + "learning_rate": 0.0004898073121956441, + "loss": 1.9164, + "step": 1256 + }, + { + "epoch": 0.12275390625, + "grad_norm": 0.32490289211273193, + "learning_rate": 0.0004897861403267486, + "loss": 1.9568, + "step": 1257 + }, + { + "epoch": 0.1228515625, + "grad_norm": 0.37740960717201233, + "learning_rate": 0.0004897649470026205, + "loss": 1.9618, + "step": 1258 + }, + { + "epoch": 0.12294921875, + "grad_norm": 0.34424281120300293, + "learning_rate": 0.0004897437322253767, + "loss": 1.9192, + "step": 1259 + }, + { + "epoch": 0.123046875, + "grad_norm": 0.1956419050693512, + "learning_rate": 0.0004897224959971368, + "loss": 1.9128, + "step": 1260 + }, + { + "epoch": 0.12314453125, + "grad_norm": 0.30533185601234436, + "learning_rate": 0.0004897012383200216, + "loss": 1.9378, + "step": 1261 + }, + { + "epoch": 0.1232421875, + "grad_norm": 0.28745076060295105, + "learning_rate": 0.0004896799591961548, + "loss": 1.9734, + "step": 1262 + }, + { + "epoch": 0.12333984375, + "grad_norm": 0.3063562512397766, + "learning_rate": 0.000489658658627662, + "loss": 1.9713, + "step": 1263 + }, + { + "epoch": 0.1234375, + "grad_norm": 0.26775872707366943, + "learning_rate": 0.000489637336616671, + "loss": 1.9305, + "step": 1264 + }, + { + "epoch": 0.12353515625, + "grad_norm": 0.2045804262161255, + "learning_rate": 0.0004896159931653114, + "loss": 1.9018, + "step": 1265 + }, + { + "epoch": 0.1236328125, + "grad_norm": 0.26752471923828125, + "learning_rate": 0.0004895946282757155, + "loss": 1.9334, + "step": 1266 + }, + { + "epoch": 0.12373046875, + "grad_norm": 0.3504626452922821, + "learning_rate": 0.0004895732419500174, + "loss": 1.9678, + "step": 1267 + }, + { + "epoch": 0.123828125, + "grad_norm": 0.2588043808937073, + "learning_rate": 0.0004895518341903534, + "loss": 1.9336, + "step": 1268 + }, + { + "epoch": 0.12392578125, + "grad_norm": 0.23393654823303223, + "learning_rate": 0.0004895304049988619, + "loss": 1.9386, + "step": 1269 + }, + { + "epoch": 0.1240234375, + "grad_norm": 0.3162434697151184, + "learning_rate": 0.0004895089543776834, + "loss": 1.9752, + "step": 1270 + }, + { + "epoch": 0.12412109375, + "grad_norm": 0.28987085819244385, + "learning_rate": 0.0004894874823289609, + "loss": 1.9346, + "step": 1271 + }, + { + "epoch": 0.12421875, + "grad_norm": 0.2837126553058624, + "learning_rate": 0.000489465988854839, + "loss": 1.9375, + "step": 1272 + }, + { + "epoch": 0.12431640625, + "grad_norm": 0.26402387022972107, + "learning_rate": 0.0004894444739574649, + "loss": 1.9235, + "step": 1273 + }, + { + "epoch": 0.1244140625, + "grad_norm": 0.21077674627304077, + "learning_rate": 0.0004894229376389875, + "loss": 1.9524, + "step": 1274 + }, + { + "epoch": 0.12451171875, + "grad_norm": 0.18328504264354706, + "learning_rate": 0.0004894013799015583, + "loss": 1.9496, + "step": 1275 + }, + { + "epoch": 0.124609375, + "grad_norm": 0.22258712351322174, + "learning_rate": 0.0004893798007473307, + "loss": 1.9661, + "step": 1276 + }, + { + "epoch": 0.12470703125, + "grad_norm": 0.20786426961421967, + "learning_rate": 0.0004893582001784604, + "loss": 1.9263, + "step": 1277 + }, + { + "epoch": 0.1248046875, + "grad_norm": 0.1959957331418991, + "learning_rate": 0.0004893365781971047, + "loss": 1.9525, + "step": 1278 + }, + { + "epoch": 0.12490234375, + "grad_norm": 0.2622356712818146, + "learning_rate": 0.0004893149348054238, + "loss": 1.9433, + "step": 1279 + }, + { + "epoch": 0.125, + "grad_norm": 0.2329346388578415, + "learning_rate": 0.0004892932700055797, + "loss": 1.9293, + "step": 1280 + }, + { + "epoch": 0.12509765625, + "grad_norm": 0.26422178745269775, + "learning_rate": 0.0004892715837997361, + "loss": 1.9252, + "step": 1281 + }, + { + "epoch": 0.1251953125, + "grad_norm": 0.30737435817718506, + "learning_rate": 0.0004892498761900597, + "loss": 1.9171, + "step": 1282 + }, + { + "epoch": 0.12529296875, + "grad_norm": 0.30651819705963135, + "learning_rate": 0.0004892281471787187, + "loss": 1.937, + "step": 1283 + }, + { + "epoch": 0.125390625, + "grad_norm": 0.23027795553207397, + "learning_rate": 0.0004892063967678838, + "loss": 1.9491, + "step": 1284 + }, + { + "epoch": 0.12548828125, + "grad_norm": 0.30681830644607544, + "learning_rate": 0.0004891846249597276, + "loss": 1.905, + "step": 1285 + }, + { + "epoch": 0.1255859375, + "grad_norm": 0.34682074189186096, + "learning_rate": 0.0004891628317564247, + "loss": 1.9157, + "step": 1286 + }, + { + "epoch": 0.12568359375, + "grad_norm": 0.26123619079589844, + "learning_rate": 0.0004891410171601523, + "loss": 1.9382, + "step": 1287 + }, + { + "epoch": 0.12578125, + "grad_norm": 0.32031431794166565, + "learning_rate": 0.0004891191811730894, + "loss": 1.9234, + "step": 1288 + }, + { + "epoch": 0.12587890625, + "grad_norm": 0.24917498230934143, + "learning_rate": 0.0004890973237974172, + "loss": 1.919, + "step": 1289 + }, + { + "epoch": 0.1259765625, + "grad_norm": 0.29161348938941956, + "learning_rate": 0.0004890754450353192, + "loss": 1.9405, + "step": 1290 + }, + { + "epoch": 0.12607421875, + "grad_norm": 0.31046807765960693, + "learning_rate": 0.0004890535448889807, + "loss": 1.9298, + "step": 1291 + }, + { + "epoch": 0.126171875, + "grad_norm": 0.30187463760375977, + "learning_rate": 0.0004890316233605893, + "loss": 1.9329, + "step": 1292 + }, + { + "epoch": 0.12626953125, + "grad_norm": 0.3013293445110321, + "learning_rate": 0.000489009680452335, + "loss": 1.9228, + "step": 1293 + }, + { + "epoch": 0.1263671875, + "grad_norm": 0.2492775321006775, + "learning_rate": 0.0004889877161664096, + "loss": 1.9565, + "step": 1294 + }, + { + "epoch": 0.12646484375, + "grad_norm": 0.29035332798957825, + "learning_rate": 0.000488965730505007, + "loss": 1.9184, + "step": 1295 + }, + { + "epoch": 0.1265625, + "grad_norm": 0.1983688622713089, + "learning_rate": 0.0004889437234703236, + "loss": 1.97, + "step": 1296 + }, + { + "epoch": 0.12666015625, + "grad_norm": 0.24074336886405945, + "learning_rate": 0.0004889216950645576, + "loss": 1.9289, + "step": 1297 + }, + { + "epoch": 0.1267578125, + "grad_norm": 0.2601425051689148, + "learning_rate": 0.0004888996452899093, + "loss": 1.9374, + "step": 1298 + }, + { + "epoch": 0.12685546875, + "grad_norm": 0.21484848856925964, + "learning_rate": 0.0004888775741485816, + "loss": 1.9075, + "step": 1299 + }, + { + "epoch": 0.126953125, + "grad_norm": 0.2370765060186386, + "learning_rate": 0.0004888554816427789, + "loss": 1.9131, + "step": 1300 + }, + { + "epoch": 0.12705078125, + "grad_norm": 0.22826816141605377, + "learning_rate": 0.0004888333677747082, + "loss": 1.9514, + "step": 1301 + }, + { + "epoch": 0.1271484375, + "grad_norm": 0.2141917198896408, + "learning_rate": 0.0004888112325465785, + "loss": 1.8406, + "step": 1302 + }, + { + "epoch": 0.12724609375, + "grad_norm": 0.22987738251686096, + "learning_rate": 0.0004887890759606008, + "loss": 1.9498, + "step": 1303 + }, + { + "epoch": 0.12734375, + "grad_norm": 0.2592550814151764, + "learning_rate": 0.0004887668980189885, + "loss": 1.9427, + "step": 1304 + }, + { + "epoch": 0.12744140625, + "grad_norm": 0.24788032472133636, + "learning_rate": 0.0004887446987239567, + "loss": 1.9718, + "step": 1305 + }, + { + "epoch": 0.1275390625, + "grad_norm": 0.3112640380859375, + "learning_rate": 0.0004887224780777233, + "loss": 1.972, + "step": 1306 + }, + { + "epoch": 0.12763671875, + "grad_norm": 0.2711445987224579, + "learning_rate": 0.0004887002360825077, + "loss": 1.9636, + "step": 1307 + }, + { + "epoch": 0.127734375, + "grad_norm": 0.19959181547164917, + "learning_rate": 0.0004886779727405318, + "loss": 1.9142, + "step": 1308 + }, + { + "epoch": 0.12783203125, + "grad_norm": 0.24054476618766785, + "learning_rate": 0.0004886556880540192, + "loss": 1.8724, + "step": 1309 + }, + { + "epoch": 0.1279296875, + "grad_norm": 0.21965253353118896, + "learning_rate": 0.0004886333820251963, + "loss": 1.9295, + "step": 1310 + }, + { + "epoch": 0.12802734375, + "grad_norm": 0.2383204847574234, + "learning_rate": 0.0004886110546562912, + "loss": 1.8945, + "step": 1311 + }, + { + "epoch": 0.128125, + "grad_norm": 0.20343053340911865, + "learning_rate": 0.0004885887059495342, + "loss": 1.9364, + "step": 1312 + }, + { + "epoch": 0.12822265625, + "grad_norm": 0.24133317172527313, + "learning_rate": 0.0004885663359071576, + "loss": 1.9023, + "step": 1313 + }, + { + "epoch": 0.1283203125, + "grad_norm": 0.27090492844581604, + "learning_rate": 0.0004885439445313962, + "loss": 1.9007, + "step": 1314 + }, + { + "epoch": 0.12841796875, + "grad_norm": 0.2153930515050888, + "learning_rate": 0.0004885215318244865, + "loss": 1.9475, + "step": 1315 + }, + { + "epoch": 0.128515625, + "grad_norm": 0.3661288022994995, + "learning_rate": 0.0004884990977886673, + "loss": 1.9858, + "step": 1316 + }, + { + "epoch": 0.12861328125, + "grad_norm": 0.4454726576805115, + "learning_rate": 0.0004884766424261798, + "loss": 1.9207, + "step": 1317 + }, + { + "epoch": 0.1287109375, + "grad_norm": 0.4249671995639801, + "learning_rate": 0.000488454165739267, + "loss": 1.9889, + "step": 1318 + }, + { + "epoch": 0.12880859375, + "grad_norm": 0.3520338237285614, + "learning_rate": 0.0004884316677301739, + "loss": 1.9564, + "step": 1319 + }, + { + "epoch": 0.12890625, + "grad_norm": 0.28052136301994324, + "learning_rate": 0.0004884091484011482, + "loss": 1.9677, + "step": 1320 + }, + { + "epoch": 0.12900390625, + "grad_norm": 0.31618887186050415, + "learning_rate": 0.0004883866077544391, + "loss": 1.9306, + "step": 1321 + }, + { + "epoch": 0.1291015625, + "grad_norm": 0.29120492935180664, + "learning_rate": 0.0004883640457922984, + "loss": 1.9205, + "step": 1322 + }, + { + "epoch": 0.12919921875, + "grad_norm": 0.23774267733097076, + "learning_rate": 0.0004883414625169797, + "loss": 1.9078, + "step": 1323 + }, + { + "epoch": 0.129296875, + "grad_norm": 0.3183814287185669, + "learning_rate": 0.0004883188579307389, + "loss": 1.7927, + "step": 1324 + }, + { + "epoch": 0.12939453125, + "grad_norm": 0.2845805585384369, + "learning_rate": 0.0004882962320358341, + "loss": 1.888, + "step": 1325 + }, + { + "epoch": 0.1294921875, + "grad_norm": 0.2877943515777588, + "learning_rate": 0.00048827358483452527, + "loss": 1.9307, + "step": 1326 + }, + { + "epoch": 0.12958984375, + "grad_norm": 0.2753991186618805, + "learning_rate": 0.0004882509163290748, + "loss": 1.9306, + "step": 1327 + }, + { + "epoch": 0.1296875, + "grad_norm": 0.2753439247608185, + "learning_rate": 0.00048822822652174705, + "loss": 1.9573, + "step": 1328 + }, + { + "epoch": 0.12978515625, + "grad_norm": 0.2690608501434326, + "learning_rate": 0.00048820551541480845, + "loss": 1.9229, + "step": 1329 + }, + { + "epoch": 0.1298828125, + "grad_norm": 0.2276460975408554, + "learning_rate": 0.00048818278301052776, + "loss": 1.9871, + "step": 1330 + }, + { + "epoch": 0.12998046875, + "grad_norm": 0.23562243580818176, + "learning_rate": 0.00048816002931117557, + "loss": 1.9489, + "step": 1331 + }, + { + "epoch": 0.130078125, + "grad_norm": 0.27421703934669495, + "learning_rate": 0.00048813725431902496, + "loss": 1.922, + "step": 1332 + }, + { + "epoch": 0.13017578125, + "grad_norm": 0.1969674974679947, + "learning_rate": 0.00048811445803635087, + "loss": 1.929, + "step": 1333 + }, + { + "epoch": 0.1302734375, + "grad_norm": 0.22562848031520844, + "learning_rate": 0.00048809164046543043, + "loss": 1.9183, + "step": 1334 + }, + { + "epoch": 0.13037109375, + "grad_norm": 0.2235979586839676, + "learning_rate": 0.0004880688016085429, + "loss": 1.9155, + "step": 1335 + }, + { + "epoch": 0.13046875, + "grad_norm": 0.23072493076324463, + "learning_rate": 0.00048804594146796977, + "loss": 1.9647, + "step": 1336 + }, + { + "epoch": 0.13056640625, + "grad_norm": 0.2528524398803711, + "learning_rate": 0.00048802306004599446, + "loss": 1.979, + "step": 1337 + }, + { + "epoch": 0.1306640625, + "grad_norm": 0.2586860954761505, + "learning_rate": 0.00048800015734490267, + "loss": 1.9371, + "step": 1338 + }, + { + "epoch": 0.13076171875, + "grad_norm": 0.21609428524971008, + "learning_rate": 0.0004879772333669822, + "loss": 1.9319, + "step": 1339 + }, + { + "epoch": 0.130859375, + "grad_norm": 0.28810006380081177, + "learning_rate": 0.0004879542881145229, + "loss": 1.9218, + "step": 1340 + }, + { + "epoch": 0.13095703125, + "grad_norm": 0.288849413394928, + "learning_rate": 0.0004879313215898168, + "loss": 1.929, + "step": 1341 + }, + { + "epoch": 0.1310546875, + "grad_norm": 0.3168569803237915, + "learning_rate": 0.00048790833379515813, + "loss": 1.9276, + "step": 1342 + }, + { + "epoch": 0.13115234375, + "grad_norm": 0.3000844120979309, + "learning_rate": 0.00048788532473284307, + "loss": 1.9337, + "step": 1343 + }, + { + "epoch": 0.13125, + "grad_norm": 0.2524905204772949, + "learning_rate": 0.00048786229440517004, + "loss": 1.9105, + "step": 1344 + }, + { + "epoch": 0.13134765625, + "grad_norm": 0.25065815448760986, + "learning_rate": 0.0004878392428144396, + "loss": 1.9161, + "step": 1345 + }, + { + "epoch": 0.1314453125, + "grad_norm": 0.3235273063182831, + "learning_rate": 0.00048781616996295443, + "loss": 1.9159, + "step": 1346 + }, + { + "epoch": 0.13154296875, + "grad_norm": 0.23726877570152283, + "learning_rate": 0.0004877930758530192, + "loss": 1.9063, + "step": 1347 + }, + { + "epoch": 0.131640625, + "grad_norm": 0.2991824150085449, + "learning_rate": 0.000487769960486941, + "loss": 1.9885, + "step": 1348 + }, + { + "epoch": 0.13173828125, + "grad_norm": 0.24819223582744598, + "learning_rate": 0.0004877468238670287, + "loss": 1.9311, + "step": 1349 + }, + { + "epoch": 0.1318359375, + "grad_norm": 0.23198595643043518, + "learning_rate": 0.00048772366599559333, + "loss": 1.9278, + "step": 1350 + }, + { + "epoch": 0.13193359375, + "grad_norm": 0.3012297451496124, + "learning_rate": 0.0004877004868749483, + "loss": 1.9258, + "step": 1351 + }, + { + "epoch": 0.13203125, + "grad_norm": 0.21619653701782227, + "learning_rate": 0.0004876772865074091, + "loss": 1.9121, + "step": 1352 + }, + { + "epoch": 0.13212890625, + "grad_norm": 0.24853789806365967, + "learning_rate": 0.00048765406489529315, + "loss": 1.9533, + "step": 1353 + }, + { + "epoch": 0.1322265625, + "grad_norm": 0.33258283138275146, + "learning_rate": 0.00048763082204091995, + "loss": 1.9424, + "step": 1354 + }, + { + "epoch": 0.13232421875, + "grad_norm": 0.3056190609931946, + "learning_rate": 0.00048760755794661137, + "loss": 1.9605, + "step": 1355 + }, + { + "epoch": 0.132421875, + "grad_norm": 0.20714829862117767, + "learning_rate": 0.00048758427261469135, + "loss": 1.9416, + "step": 1356 + }, + { + "epoch": 0.13251953125, + "grad_norm": 0.2573312520980835, + "learning_rate": 0.0004875609660474857, + "loss": 1.9198, + "step": 1357 + }, + { + "epoch": 0.1326171875, + "grad_norm": 0.24284306168556213, + "learning_rate": 0.0004875376382473228, + "loss": 1.9308, + "step": 1358 + }, + { + "epoch": 0.13271484375, + "grad_norm": 0.20341083407402039, + "learning_rate": 0.0004875142892165327, + "loss": 1.8933, + "step": 1359 + }, + { + "epoch": 0.1328125, + "grad_norm": 0.3155312240123749, + "learning_rate": 0.0004874909189574478, + "loss": 1.8825, + "step": 1360 + }, + { + "epoch": 0.13291015625, + "grad_norm": 0.2829397916793823, + "learning_rate": 0.0004874675274724026, + "loss": 1.9229, + "step": 1361 + }, + { + "epoch": 0.1330078125, + "grad_norm": 0.34132909774780273, + "learning_rate": 0.00048744411476373374, + "loss": 1.8935, + "step": 1362 + }, + { + "epoch": 0.13310546875, + "grad_norm": 0.41804420948028564, + "learning_rate": 0.0004874206808337799, + "loss": 1.9241, + "step": 1363 + }, + { + "epoch": 0.133203125, + "grad_norm": 0.3885767161846161, + "learning_rate": 0.0004873972256848819, + "loss": 1.9391, + "step": 1364 + }, + { + "epoch": 0.13330078125, + "grad_norm": 0.3422619104385376, + "learning_rate": 0.00048737374931938273, + "loss": 1.9113, + "step": 1365 + }, + { + "epoch": 0.1333984375, + "grad_norm": 0.33838555216789246, + "learning_rate": 0.00048735025173962746, + "loss": 1.8967, + "step": 1366 + }, + { + "epoch": 0.13349609375, + "grad_norm": 0.3656925857067108, + "learning_rate": 0.0004873267329479633, + "loss": 1.9405, + "step": 1367 + }, + { + "epoch": 0.13359375, + "grad_norm": 0.3086741268634796, + "learning_rate": 0.00048730319294673955, + "loss": 1.9518, + "step": 1368 + }, + { + "epoch": 0.13369140625, + "grad_norm": 0.24022363126277924, + "learning_rate": 0.0004872796317383077, + "loss": 1.9368, + "step": 1369 + }, + { + "epoch": 0.1337890625, + "grad_norm": 0.2861846089363098, + "learning_rate": 0.00048725604932502116, + "loss": 1.9186, + "step": 1370 + }, + { + "epoch": 0.13388671875, + "grad_norm": 0.29110145568847656, + "learning_rate": 0.0004872324457092358, + "loss": 1.8963, + "step": 1371 + }, + { + "epoch": 0.133984375, + "grad_norm": 0.2232580929994583, + "learning_rate": 0.0004872088208933093, + "loss": 1.929, + "step": 1372 + }, + { + "epoch": 0.13408203125, + "grad_norm": 0.3331020772457123, + "learning_rate": 0.0004871851748796016, + "loss": 1.936, + "step": 1373 + }, + { + "epoch": 0.1341796875, + "grad_norm": 0.2353833019733429, + "learning_rate": 0.00048716150767047473, + "loss": 1.9047, + "step": 1374 + }, + { + "epoch": 0.13427734375, + "grad_norm": 0.25953710079193115, + "learning_rate": 0.0004871378192682928, + "loss": 1.9812, + "step": 1375 + }, + { + "epoch": 0.134375, + "grad_norm": 0.24382802844047546, + "learning_rate": 0.0004871141096754222, + "loss": 1.9599, + "step": 1376 + }, + { + "epoch": 0.13447265625, + "grad_norm": 0.2496114820241928, + "learning_rate": 0.00048709037889423107, + "loss": 1.9094, + "step": 1377 + }, + { + "epoch": 0.1345703125, + "grad_norm": 0.26899224519729614, + "learning_rate": 0.0004870666269270901, + "loss": 1.9467, + "step": 1378 + }, + { + "epoch": 0.13466796875, + "grad_norm": 0.23490363359451294, + "learning_rate": 0.00048704285377637184, + "loss": 1.9682, + "step": 1379 + }, + { + "epoch": 0.134765625, + "grad_norm": 0.23190240561962128, + "learning_rate": 0.000487019059444451, + "loss": 1.9324, + "step": 1380 + }, + { + "epoch": 0.13486328125, + "grad_norm": 0.2239055186510086, + "learning_rate": 0.0004869952439337045, + "loss": 1.995, + "step": 1381 + }, + { + "epoch": 0.1349609375, + "grad_norm": 0.23679253458976746, + "learning_rate": 0.00048697140724651114, + "loss": 1.8685, + "step": 1382 + }, + { + "epoch": 0.13505859375, + "grad_norm": 0.2584695816040039, + "learning_rate": 0.0004869475493852521, + "loss": 1.9421, + "step": 1383 + }, + { + "epoch": 0.13515625, + "grad_norm": 0.219711571931839, + "learning_rate": 0.0004869236703523106, + "loss": 1.9326, + "step": 1384 + }, + { + "epoch": 0.13525390625, + "grad_norm": 0.2176481932401657, + "learning_rate": 0.00048689977015007197, + "loss": 1.8959, + "step": 1385 + }, + { + "epoch": 0.1353515625, + "grad_norm": 0.22419661283493042, + "learning_rate": 0.0004868758487809235, + "loss": 1.9095, + "step": 1386 + }, + { + "epoch": 0.13544921875, + "grad_norm": 0.255984365940094, + "learning_rate": 0.0004868519062472548, + "loss": 1.9756, + "step": 1387 + }, + { + "epoch": 0.135546875, + "grad_norm": 0.217428520321846, + "learning_rate": 0.0004868279425514575, + "loss": 1.9481, + "step": 1388 + }, + { + "epoch": 0.13564453125, + "grad_norm": 0.19990497827529907, + "learning_rate": 0.00048680395769592534, + "loss": 1.9279, + "step": 1389 + }, + { + "epoch": 0.1357421875, + "grad_norm": 0.2209380716085434, + "learning_rate": 0.00048677995168305426, + "loss": 1.9148, + "step": 1390 + }, + { + "epoch": 0.13583984375, + "grad_norm": 0.25395241379737854, + "learning_rate": 0.00048675592451524226, + "loss": 1.9688, + "step": 1391 + }, + { + "epoch": 0.1359375, + "grad_norm": 0.1944066882133484, + "learning_rate": 0.0004867318761948892, + "loss": 1.9005, + "step": 1392 + }, + { + "epoch": 0.13603515625, + "grad_norm": 0.22825607657432556, + "learning_rate": 0.0004867078067243977, + "loss": 1.9538, + "step": 1393 + }, + { + "epoch": 0.1361328125, + "grad_norm": 0.23758959770202637, + "learning_rate": 0.0004866837161061717, + "loss": 1.9887, + "step": 1394 + }, + { + "epoch": 0.13623046875, + "grad_norm": 0.26341712474823, + "learning_rate": 0.00048665960434261796, + "loss": 1.9217, + "step": 1395 + }, + { + "epoch": 0.136328125, + "grad_norm": 0.283316045999527, + "learning_rate": 0.0004866354714361448, + "loss": 1.9576, + "step": 1396 + }, + { + "epoch": 0.13642578125, + "grad_norm": 0.32178181409835815, + "learning_rate": 0.000486611317389163, + "loss": 1.9651, + "step": 1397 + }, + { + "epoch": 0.1365234375, + "grad_norm": 0.4065782129764557, + "learning_rate": 0.0004865871422040853, + "loss": 1.897, + "step": 1398 + }, + { + "epoch": 0.13662109375, + "grad_norm": 0.3661766052246094, + "learning_rate": 0.00048656294588332653, + "loss": 1.9472, + "step": 1399 + }, + { + "epoch": 0.13671875, + "grad_norm": 0.24434155225753784, + "learning_rate": 0.0004865387284293038, + "loss": 1.9179, + "step": 1400 + }, + { + "epoch": 0.13681640625, + "grad_norm": 0.4203675091266632, + "learning_rate": 0.0004865144898444361, + "loss": 1.9583, + "step": 1401 + }, + { + "epoch": 0.1369140625, + "grad_norm": 0.4013305902481079, + "learning_rate": 0.0004864902301311448, + "loss": 1.9322, + "step": 1402 + }, + { + "epoch": 0.13701171875, + "grad_norm": 0.30299681425094604, + "learning_rate": 0.00048646594929185313, + "loss": 1.9152, + "step": 1403 + }, + { + "epoch": 0.137109375, + "grad_norm": 0.30768755078315735, + "learning_rate": 0.0004864416473289865, + "loss": 1.966, + "step": 1404 + }, + { + "epoch": 0.13720703125, + "grad_norm": 0.2759581506252289, + "learning_rate": 0.00048641732424497256, + "loss": 1.9211, + "step": 1405 + }, + { + "epoch": 0.1373046875, + "grad_norm": 0.2976212799549103, + "learning_rate": 0.00048639298004224087, + "loss": 1.9376, + "step": 1406 + }, + { + "epoch": 0.13740234375, + "grad_norm": 0.21789424121379852, + "learning_rate": 0.0004863686147232232, + "loss": 1.994, + "step": 1407 + }, + { + "epoch": 0.1375, + "grad_norm": 0.2543160617351532, + "learning_rate": 0.0004863442282903535, + "loss": 1.9153, + "step": 1408 + }, + { + "epoch": 0.13759765625, + "grad_norm": 0.2789456248283386, + "learning_rate": 0.0004863198207460678, + "loss": 1.9238, + "step": 1409 + }, + { + "epoch": 0.1376953125, + "grad_norm": 0.21548853814601898, + "learning_rate": 0.0004862953920928041, + "loss": 1.9003, + "step": 1410 + }, + { + "epoch": 0.13779296875, + "grad_norm": 0.2607603967189789, + "learning_rate": 0.00048627094233300266, + "loss": 1.951, + "step": 1411 + }, + { + "epoch": 0.137890625, + "grad_norm": 0.21658223867416382, + "learning_rate": 0.00048624647146910575, + "loss": 1.9818, + "step": 1412 + }, + { + "epoch": 0.13798828125, + "grad_norm": 0.24333196878433228, + "learning_rate": 0.00048622197950355786, + "loss": 1.8933, + "step": 1413 + }, + { + "epoch": 0.1380859375, + "grad_norm": 0.22348099946975708, + "learning_rate": 0.0004861974664388054, + "loss": 1.9302, + "step": 1414 + }, + { + "epoch": 0.13818359375, + "grad_norm": 0.1976805180311203, + "learning_rate": 0.0004861729322772971, + "loss": 1.9044, + "step": 1415 + }, + { + "epoch": 0.13828125, + "grad_norm": 0.22617708146572113, + "learning_rate": 0.00048614837702148374, + "loss": 1.9445, + "step": 1416 + }, + { + "epoch": 0.13837890625, + "grad_norm": 0.2243851125240326, + "learning_rate": 0.0004861238006738181, + "loss": 1.9215, + "step": 1417 + }, + { + "epoch": 0.1384765625, + "grad_norm": 0.21161694824695587, + "learning_rate": 0.0004860992032367551, + "loss": 1.9443, + "step": 1418 + }, + { + "epoch": 0.13857421875, + "grad_norm": 0.22872799634933472, + "learning_rate": 0.0004860745847127519, + "loss": 1.8174, + "step": 1419 + }, + { + "epoch": 0.138671875, + "grad_norm": 0.27401795983314514, + "learning_rate": 0.0004860499451042677, + "loss": 1.919, + "step": 1420 + }, + { + "epoch": 0.13876953125, + "grad_norm": 0.2309410721063614, + "learning_rate": 0.00048602528441376365, + "loss": 1.9025, + "step": 1421 + }, + { + "epoch": 0.1388671875, + "grad_norm": 0.21750116348266602, + "learning_rate": 0.0004860006026437033, + "loss": 1.9242, + "step": 1422 + }, + { + "epoch": 0.13896484375, + "grad_norm": 0.24858328700065613, + "learning_rate": 0.0004859758997965519, + "loss": 1.9562, + "step": 1423 + }, + { + "epoch": 0.1390625, + "grad_norm": 0.23306035995483398, + "learning_rate": 0.00048595117587477724, + "loss": 1.9504, + "step": 1424 + }, + { + "epoch": 0.13916015625, + "grad_norm": 0.24605897068977356, + "learning_rate": 0.00048592643088084904, + "loss": 1.9502, + "step": 1425 + }, + { + "epoch": 0.1392578125, + "grad_norm": 0.23687416315078735, + "learning_rate": 0.0004859016648172389, + "loss": 1.9035, + "step": 1426 + }, + { + "epoch": 0.13935546875, + "grad_norm": 0.33239808678627014, + "learning_rate": 0.00048587687768642104, + "loss": 1.9469, + "step": 1427 + }, + { + "epoch": 0.139453125, + "grad_norm": 0.3443644344806671, + "learning_rate": 0.00048585206949087117, + "loss": 1.9126, + "step": 1428 + }, + { + "epoch": 0.13955078125, + "grad_norm": 0.4044901430606842, + "learning_rate": 0.0004858272402330676, + "loss": 1.928, + "step": 1429 + }, + { + "epoch": 0.1396484375, + "grad_norm": 0.3553960919380188, + "learning_rate": 0.0004858023899154905, + "loss": 1.9782, + "step": 1430 + }, + { + "epoch": 0.13974609375, + "grad_norm": 0.24459035694599152, + "learning_rate": 0.00048577751854062207, + "loss": 1.9228, + "step": 1431 + }, + { + "epoch": 0.13984375, + "grad_norm": 0.26586851477622986, + "learning_rate": 0.000485752626110947, + "loss": 1.9607, + "step": 1432 + }, + { + "epoch": 0.13994140625, + "grad_norm": 0.2949119210243225, + "learning_rate": 0.0004857277126289516, + "loss": 1.9418, + "step": 1433 + }, + { + "epoch": 0.1400390625, + "grad_norm": 0.23309734463691711, + "learning_rate": 0.0004857027780971246, + "loss": 1.9036, + "step": 1434 + }, + { + "epoch": 0.14013671875, + "grad_norm": 0.27574148774147034, + "learning_rate": 0.00048567782251795674, + "loss": 1.883, + "step": 1435 + }, + { + "epoch": 0.140234375, + "grad_norm": 0.2791823446750641, + "learning_rate": 0.0004856528458939409, + "loss": 1.9021, + "step": 1436 + }, + { + "epoch": 0.14033203125, + "grad_norm": 0.3588283061981201, + "learning_rate": 0.00048562784822757187, + "loss": 1.9417, + "step": 1437 + }, + { + "epoch": 0.1404296875, + "grad_norm": 0.3069411814212799, + "learning_rate": 0.00048560282952134687, + "loss": 1.9055, + "step": 1438 + }, + { + "epoch": 0.14052734375, + "grad_norm": 0.3290352523326874, + "learning_rate": 0.00048557778977776495, + "loss": 1.9674, + "step": 1439 + }, + { + "epoch": 0.140625, + "grad_norm": 0.21290095150470734, + "learning_rate": 0.0004855527289993274, + "loss": 1.9099, + "step": 1440 + }, + { + "epoch": 0.14072265625, + "grad_norm": 0.28244757652282715, + "learning_rate": 0.00048552764718853755, + "loss": 1.9821, + "step": 1441 + }, + { + "epoch": 0.1408203125, + "grad_norm": 0.2707923948764801, + "learning_rate": 0.0004855025443479008, + "loss": 1.92, + "step": 1442 + }, + { + "epoch": 0.14091796875, + "grad_norm": 0.2617279589176178, + "learning_rate": 0.0004854774204799248, + "loss": 1.9425, + "step": 1443 + }, + { + "epoch": 0.141015625, + "grad_norm": 0.2797471880912781, + "learning_rate": 0.00048545227558711914, + "loss": 1.9239, + "step": 1444 + }, + { + "epoch": 0.14111328125, + "grad_norm": 0.21120323240756989, + "learning_rate": 0.00048542710967199563, + "loss": 1.9323, + "step": 1445 + }, + { + "epoch": 0.1412109375, + "grad_norm": 0.23532937467098236, + "learning_rate": 0.00048540192273706807, + "loss": 1.9033, + "step": 1446 + }, + { + "epoch": 0.14130859375, + "grad_norm": 0.26372474431991577, + "learning_rate": 0.00048537671478485233, + "loss": 1.9109, + "step": 1447 + }, + { + "epoch": 0.14140625, + "grad_norm": 0.2063819169998169, + "learning_rate": 0.0004853514858178667, + "loss": 1.8868, + "step": 1448 + }, + { + "epoch": 0.14150390625, + "grad_norm": 0.2495461255311966, + "learning_rate": 0.0004853262358386311, + "loss": 1.9261, + "step": 1449 + }, + { + "epoch": 0.1416015625, + "grad_norm": 0.24477644264698029, + "learning_rate": 0.00048530096484966786, + "loss": 1.8921, + "step": 1450 + }, + { + "epoch": 0.14169921875, + "grad_norm": 0.20758818089962006, + "learning_rate": 0.0004852756728535013, + "loss": 1.9274, + "step": 1451 + }, + { + "epoch": 0.141796875, + "grad_norm": 0.24920731782913208, + "learning_rate": 0.000485250359852658, + "loss": 1.9203, + "step": 1452 + }, + { + "epoch": 0.14189453125, + "grad_norm": 0.27062875032424927, + "learning_rate": 0.00048522502584966626, + "loss": 1.9225, + "step": 1453 + }, + { + "epoch": 0.1419921875, + "grad_norm": 0.26489493250846863, + "learning_rate": 0.00048519967084705696, + "loss": 1.9797, + "step": 1454 + }, + { + "epoch": 0.14208984375, + "grad_norm": 0.2739894688129425, + "learning_rate": 0.0004851742948473627, + "loss": 1.8824, + "step": 1455 + }, + { + "epoch": 0.1421875, + "grad_norm": 0.24199414253234863, + "learning_rate": 0.0004851488978531184, + "loss": 1.9439, + "step": 1456 + }, + { + "epoch": 0.14228515625, + "grad_norm": 0.21669942140579224, + "learning_rate": 0.00048512347986686094, + "loss": 1.8901, + "step": 1457 + }, + { + "epoch": 0.1423828125, + "grad_norm": 0.27475011348724365, + "learning_rate": 0.0004850980408911294, + "loss": 1.877, + "step": 1458 + }, + { + "epoch": 0.14248046875, + "grad_norm": 0.24100442230701447, + "learning_rate": 0.0004850725809284648, + "loss": 1.9286, + "step": 1459 + }, + { + "epoch": 0.142578125, + "grad_norm": 0.24343852698802948, + "learning_rate": 0.0004850470999814105, + "loss": 1.9503, + "step": 1460 + }, + { + "epoch": 0.14267578125, + "grad_norm": 0.22666993737220764, + "learning_rate": 0.0004850215980525118, + "loss": 1.8722, + "step": 1461 + }, + { + "epoch": 0.1427734375, + "grad_norm": 0.22877848148345947, + "learning_rate": 0.00048499607514431597, + "loss": 1.8726, + "step": 1462 + }, + { + "epoch": 0.14287109375, + "grad_norm": 0.21896860003471375, + "learning_rate": 0.0004849705312593727, + "loss": 1.9109, + "step": 1463 + }, + { + "epoch": 0.14296875, + "grad_norm": 0.2896305024623871, + "learning_rate": 0.00048494496640023356, + "loss": 1.9293, + "step": 1464 + }, + { + "epoch": 0.14306640625, + "grad_norm": 0.2690875232219696, + "learning_rate": 0.00048491938056945214, + "loss": 1.9191, + "step": 1465 + }, + { + "epoch": 0.1431640625, + "grad_norm": 0.27440106868743896, + "learning_rate": 0.0004848937737695844, + "loss": 1.9729, + "step": 1466 + }, + { + "epoch": 0.14326171875, + "grad_norm": 0.26410380005836487, + "learning_rate": 0.000484868146003188, + "loss": 1.8782, + "step": 1467 + }, + { + "epoch": 0.143359375, + "grad_norm": 0.2948254644870758, + "learning_rate": 0.00048484249727282326, + "loss": 1.9726, + "step": 1468 + }, + { + "epoch": 0.14345703125, + "grad_norm": 0.34213653206825256, + "learning_rate": 0.000484816827581052, + "loss": 1.8893, + "step": 1469 + }, + { + "epoch": 0.1435546875, + "grad_norm": 0.2864801287651062, + "learning_rate": 0.0004847911369304385, + "loss": 1.9717, + "step": 1470 + }, + { + "epoch": 0.14365234375, + "grad_norm": 0.23252104222774506, + "learning_rate": 0.00048476542532354896, + "loss": 1.9256, + "step": 1471 + }, + { + "epoch": 0.14375, + "grad_norm": 0.24983392655849457, + "learning_rate": 0.00048473969276295175, + "loss": 1.9275, + "step": 1472 + }, + { + "epoch": 0.14384765625, + "grad_norm": 0.24153786897659302, + "learning_rate": 0.0004847139392512175, + "loss": 1.8821, + "step": 1473 + }, + { + "epoch": 0.1439453125, + "grad_norm": 0.2148951143026352, + "learning_rate": 0.00048468816479091853, + "loss": 1.9125, + "step": 1474 + }, + { + "epoch": 0.14404296875, + "grad_norm": 0.20793981850147247, + "learning_rate": 0.00048466236938462956, + "loss": 1.9336, + "step": 1475 + }, + { + "epoch": 0.144140625, + "grad_norm": 0.22007182240486145, + "learning_rate": 0.0004846365530349273, + "loss": 1.9383, + "step": 1476 + }, + { + "epoch": 0.14423828125, + "grad_norm": 0.21164564788341522, + "learning_rate": 0.00048461071574439063, + "loss": 1.9188, + "step": 1477 + }, + { + "epoch": 0.1443359375, + "grad_norm": 0.26241186261177063, + "learning_rate": 0.00048458485751560045, + "loss": 1.9348, + "step": 1478 + }, + { + "epoch": 0.14443359375, + "grad_norm": 0.19936083257198334, + "learning_rate": 0.00048455897835113974, + "loss": 1.8605, + "step": 1479 + }, + { + "epoch": 0.14453125, + "grad_norm": 0.2293226420879364, + "learning_rate": 0.0004845330782535937, + "loss": 1.8666, + "step": 1480 + }, + { + "epoch": 0.14462890625, + "grad_norm": 0.27344003319740295, + "learning_rate": 0.0004845071572255494, + "loss": 1.9281, + "step": 1481 + }, + { + "epoch": 0.1447265625, + "grad_norm": 0.2886092960834503, + "learning_rate": 0.0004844812152695961, + "loss": 1.9106, + "step": 1482 + }, + { + "epoch": 0.14482421875, + "grad_norm": 0.345197856426239, + "learning_rate": 0.00048445525238832526, + "loss": 1.8825, + "step": 1483 + }, + { + "epoch": 0.144921875, + "grad_norm": 0.32034623622894287, + "learning_rate": 0.0004844292685843303, + "loss": 1.9103, + "step": 1484 + }, + { + "epoch": 0.14501953125, + "grad_norm": 0.30518102645874023, + "learning_rate": 0.00048440326386020685, + "loss": 1.9614, + "step": 1485 + }, + { + "epoch": 0.1451171875, + "grad_norm": 0.28948330879211426, + "learning_rate": 0.00048437723821855244, + "loss": 1.8966, + "step": 1486 + }, + { + "epoch": 0.14521484375, + "grad_norm": 0.28401121497154236, + "learning_rate": 0.00048435119166196684, + "loss": 1.9455, + "step": 1487 + }, + { + "epoch": 0.1453125, + "grad_norm": 0.2352132648229599, + "learning_rate": 0.0004843251241930519, + "loss": 1.9285, + "step": 1488 + }, + { + "epoch": 0.14541015625, + "grad_norm": 0.231564462184906, + "learning_rate": 0.0004842990358144115, + "loss": 1.9204, + "step": 1489 + }, + { + "epoch": 0.1455078125, + "grad_norm": 0.2393004596233368, + "learning_rate": 0.00048427292652865165, + "loss": 1.9449, + "step": 1490 + }, + { + "epoch": 0.14560546875, + "grad_norm": 0.2513711154460907, + "learning_rate": 0.00048424679633838046, + "loss": 1.9092, + "step": 1491 + }, + { + "epoch": 0.145703125, + "grad_norm": 0.22026249766349792, + "learning_rate": 0.000484220645246208, + "loss": 1.9439, + "step": 1492 + }, + { + "epoch": 0.14580078125, + "grad_norm": 0.29993200302124023, + "learning_rate": 0.0004841944732547467, + "loss": 1.8934, + "step": 1493 + }, + { + "epoch": 0.1458984375, + "grad_norm": 0.27871406078338623, + "learning_rate": 0.0004841682803666108, + "loss": 1.8947, + "step": 1494 + }, + { + "epoch": 0.14599609375, + "grad_norm": 0.21927416324615479, + "learning_rate": 0.0004841420665844166, + "loss": 1.9472, + "step": 1495 + }, + { + "epoch": 0.14609375, + "grad_norm": 0.24884743988513947, + "learning_rate": 0.0004841158319107829, + "loss": 1.9099, + "step": 1496 + }, + { + "epoch": 0.14619140625, + "grad_norm": 0.256428986787796, + "learning_rate": 0.0004840895763483302, + "loss": 1.9266, + "step": 1497 + }, + { + "epoch": 0.1462890625, + "grad_norm": 0.2777554392814636, + "learning_rate": 0.0004840632998996812, + "loss": 1.9764, + "step": 1498 + }, + { + "epoch": 0.14638671875, + "grad_norm": 0.3199704587459564, + "learning_rate": 0.00048403700256746066, + "loss": 1.9391, + "step": 1499 + }, + { + "epoch": 0.146484375, + "grad_norm": 0.33837810158729553, + "learning_rate": 0.00048401068435429545, + "loss": 1.9131, + "step": 1500 + }, + { + "epoch": 0.14658203125, + "grad_norm": 0.25376418232917786, + "learning_rate": 0.00048398434526281456, + "loss": 1.9391, + "step": 1501 + }, + { + "epoch": 0.1466796875, + "grad_norm": 0.2878507673740387, + "learning_rate": 0.00048395798529564897, + "loss": 1.9226, + "step": 1502 + }, + { + "epoch": 0.14677734375, + "grad_norm": 0.3166121244430542, + "learning_rate": 0.0004839316044554319, + "loss": 1.9024, + "step": 1503 + }, + { + "epoch": 0.146875, + "grad_norm": 0.30799803137779236, + "learning_rate": 0.00048390520274479857, + "loss": 1.9076, + "step": 1504 + }, + { + "epoch": 0.14697265625, + "grad_norm": 0.3012866675853729, + "learning_rate": 0.0004838787801663861, + "loss": 1.9282, + "step": 1505 + }, + { + "epoch": 0.1470703125, + "grad_norm": 0.29956120252609253, + "learning_rate": 0.000483852336722834, + "loss": 1.8934, + "step": 1506 + }, + { + "epoch": 0.14716796875, + "grad_norm": 0.2174024134874344, + "learning_rate": 0.0004838258724167838, + "loss": 1.9434, + "step": 1507 + }, + { + "epoch": 0.147265625, + "grad_norm": 0.31407755613327026, + "learning_rate": 0.0004837993872508789, + "loss": 1.8654, + "step": 1508 + }, + { + "epoch": 0.14736328125, + "grad_norm": 0.22638960182666779, + "learning_rate": 0.00048377288122776506, + "loss": 1.937, + "step": 1509 + }, + { + "epoch": 0.1474609375, + "grad_norm": 0.26045146584510803, + "learning_rate": 0.0004837463543500899, + "loss": 1.8986, + "step": 1510 + }, + { + "epoch": 0.14755859375, + "grad_norm": 0.2935525178909302, + "learning_rate": 0.0004837198066205034, + "loss": 1.9554, + "step": 1511 + }, + { + "epoch": 0.14765625, + "grad_norm": 0.20349502563476562, + "learning_rate": 0.0004836932380416571, + "loss": 1.893, + "step": 1512 + }, + { + "epoch": 0.14775390625, + "grad_norm": 0.28056201338768005, + "learning_rate": 0.0004836666486162053, + "loss": 1.983, + "step": 1513 + }, + { + "epoch": 0.1478515625, + "grad_norm": 0.27571672201156616, + "learning_rate": 0.00048364003834680395, + "loss": 1.9399, + "step": 1514 + }, + { + "epoch": 0.14794921875, + "grad_norm": 0.28008097410202026, + "learning_rate": 0.0004836134072361111, + "loss": 1.9549, + "step": 1515 + }, + { + "epoch": 0.148046875, + "grad_norm": 0.2211456298828125, + "learning_rate": 0.00048358675528678704, + "loss": 1.9245, + "step": 1516 + }, + { + "epoch": 0.14814453125, + "grad_norm": 0.259030818939209, + "learning_rate": 0.00048356008250149404, + "loss": 1.8919, + "step": 1517 + }, + { + "epoch": 0.1482421875, + "grad_norm": 0.2772839665412903, + "learning_rate": 0.00048353338888289645, + "loss": 1.9386, + "step": 1518 + }, + { + "epoch": 0.14833984375, + "grad_norm": 0.3579663336277008, + "learning_rate": 0.00048350667443366075, + "loss": 1.9375, + "step": 1519 + }, + { + "epoch": 0.1484375, + "grad_norm": 0.37625831365585327, + "learning_rate": 0.0004834799391564555, + "loss": 1.9508, + "step": 1520 + }, + { + "epoch": 0.14853515625, + "grad_norm": 0.23583106696605682, + "learning_rate": 0.0004834531830539512, + "loss": 1.8934, + "step": 1521 + }, + { + "epoch": 0.1486328125, + "grad_norm": 0.2464006394147873, + "learning_rate": 0.00048342640612882075, + "loss": 1.8932, + "step": 1522 + }, + { + "epoch": 0.14873046875, + "grad_norm": 0.30233830213546753, + "learning_rate": 0.0004833996083837388, + "loss": 1.9117, + "step": 1523 + }, + { + "epoch": 0.148828125, + "grad_norm": 0.23098327219486237, + "learning_rate": 0.0004833727898213822, + "loss": 1.9229, + "step": 1524 + }, + { + "epoch": 0.14892578125, + "grad_norm": 0.2418309450149536, + "learning_rate": 0.00048334595044442996, + "loss": 1.9369, + "step": 1525 + }, + { + "epoch": 0.1490234375, + "grad_norm": 0.2360529899597168, + "learning_rate": 0.000483319090255563, + "loss": 1.8828, + "step": 1526 + }, + { + "epoch": 0.14912109375, + "grad_norm": 0.24173954129219055, + "learning_rate": 0.0004832922092574645, + "loss": 1.9349, + "step": 1527 + }, + { + "epoch": 0.14921875, + "grad_norm": 0.21543891727924347, + "learning_rate": 0.0004832653074528196, + "loss": 1.9317, + "step": 1528 + }, + { + "epoch": 0.14931640625, + "grad_norm": 0.21203261613845825, + "learning_rate": 0.0004832383848443155, + "loss": 1.9261, + "step": 1529 + }, + { + "epoch": 0.1494140625, + "grad_norm": 0.19579768180847168, + "learning_rate": 0.0004832114414346417, + "loss": 1.878, + "step": 1530 + }, + { + "epoch": 0.14951171875, + "grad_norm": 0.23105336725711823, + "learning_rate": 0.0004831844772264894, + "loss": 1.8957, + "step": 1531 + }, + { + "epoch": 0.149609375, + "grad_norm": 0.1797371655702591, + "learning_rate": 0.0004831574922225521, + "loss": 1.9036, + "step": 1532 + }, + { + "epoch": 0.14970703125, + "grad_norm": 0.22427886724472046, + "learning_rate": 0.00048313048642552553, + "loss": 1.9143, + "step": 1533 + }, + { + "epoch": 0.1498046875, + "grad_norm": 0.2837050259113312, + "learning_rate": 0.00048310345983810723, + "loss": 1.9411, + "step": 1534 + }, + { + "epoch": 0.14990234375, + "grad_norm": 0.25001662969589233, + "learning_rate": 0.0004830764124629969, + "loss": 1.9157, + "step": 1535 + }, + { + "epoch": 0.15, + "grad_norm": 0.21143823862075806, + "learning_rate": 0.00048304934430289635, + "loss": 1.9654, + "step": 1536 + }, + { + "epoch": 0.15009765625, + "grad_norm": 0.27027738094329834, + "learning_rate": 0.00048302225536050954, + "loss": 1.8961, + "step": 1537 + }, + { + "epoch": 0.1501953125, + "grad_norm": 0.2757401168346405, + "learning_rate": 0.00048299514563854225, + "loss": 1.9537, + "step": 1538 + }, + { + "epoch": 0.15029296875, + "grad_norm": 0.30425819754600525, + "learning_rate": 0.0004829680151397025, + "loss": 1.8713, + "step": 1539 + }, + { + "epoch": 0.150390625, + "grad_norm": 0.3185950815677643, + "learning_rate": 0.0004829408638667006, + "loss": 1.987, + "step": 1540 + }, + { + "epoch": 0.15048828125, + "grad_norm": 0.29616236686706543, + "learning_rate": 0.0004829136918222486, + "loss": 1.9329, + "step": 1541 + }, + { + "epoch": 0.1505859375, + "grad_norm": 0.2886502146720886, + "learning_rate": 0.00048288649900906066, + "loss": 1.9178, + "step": 1542 + }, + { + "epoch": 0.15068359375, + "grad_norm": 0.27268344163894653, + "learning_rate": 0.0004828592854298532, + "loss": 1.941, + "step": 1543 + }, + { + "epoch": 0.15078125, + "grad_norm": 0.27281153202056885, + "learning_rate": 0.00048283205108734463, + "loss": 1.9668, + "step": 1544 + }, + { + "epoch": 0.15087890625, + "grad_norm": 0.34445685148239136, + "learning_rate": 0.0004828047959842554, + "loss": 1.9153, + "step": 1545 + }, + { + "epoch": 0.1509765625, + "grad_norm": 0.3492426872253418, + "learning_rate": 0.0004827775201233079, + "loss": 1.9123, + "step": 1546 + }, + { + "epoch": 0.15107421875, + "grad_norm": 0.23972700536251068, + "learning_rate": 0.000482750223507227, + "loss": 1.9189, + "step": 1547 + }, + { + "epoch": 0.151171875, + "grad_norm": 0.28324854373931885, + "learning_rate": 0.0004827229061387393, + "loss": 1.9113, + "step": 1548 + }, + { + "epoch": 0.15126953125, + "grad_norm": 0.26321306824684143, + "learning_rate": 0.0004826955680205734, + "loss": 1.9568, + "step": 1549 + }, + { + "epoch": 0.1513671875, + "grad_norm": 0.26853737235069275, + "learning_rate": 0.0004826682091554604, + "loss": 1.8798, + "step": 1550 + }, + { + "epoch": 0.15146484375, + "grad_norm": 0.32173535227775574, + "learning_rate": 0.00048264082954613296, + "loss": 1.9564, + "step": 1551 + }, + { + "epoch": 0.1515625, + "grad_norm": 0.2380555272102356, + "learning_rate": 0.0004826134291953263, + "loss": 1.883, + "step": 1552 + }, + { + "epoch": 0.15166015625, + "grad_norm": 0.251919150352478, + "learning_rate": 0.0004825860081057773, + "loss": 1.8876, + "step": 1553 + }, + { + "epoch": 0.1517578125, + "grad_norm": 0.26018527150154114, + "learning_rate": 0.00048255856628022514, + "loss": 1.9115, + "step": 1554 + }, + { + "epoch": 0.15185546875, + "grad_norm": 0.21712715923786163, + "learning_rate": 0.00048253110372141104, + "loss": 1.9097, + "step": 1555 + }, + { + "epoch": 0.151953125, + "grad_norm": 0.2382095605134964, + "learning_rate": 0.0004825036204320782, + "loss": 1.937, + "step": 1556 + }, + { + "epoch": 0.15205078125, + "grad_norm": 0.231217160820961, + "learning_rate": 0.000482476116414972, + "loss": 1.9344, + "step": 1557 + }, + { + "epoch": 0.1521484375, + "grad_norm": 0.20458881556987762, + "learning_rate": 0.0004824485916728398, + "loss": 1.9036, + "step": 1558 + }, + { + "epoch": 0.15224609375, + "grad_norm": 0.205721914768219, + "learning_rate": 0.00048242104620843125, + "loss": 1.8998, + "step": 1559 + }, + { + "epoch": 0.15234375, + "grad_norm": 0.23993369936943054, + "learning_rate": 0.0004823934800244976, + "loss": 1.8706, + "step": 1560 + }, + { + "epoch": 0.15244140625, + "grad_norm": 0.23396044969558716, + "learning_rate": 0.00048236589312379276, + "loss": 1.8834, + "step": 1561 + }, + { + "epoch": 0.1525390625, + "grad_norm": 0.24282585084438324, + "learning_rate": 0.0004823382855090723, + "loss": 1.8873, + "step": 1562 + }, + { + "epoch": 0.15263671875, + "grad_norm": 0.25903937220573425, + "learning_rate": 0.00048231065718309387, + "loss": 1.9357, + "step": 1563 + }, + { + "epoch": 0.152734375, + "grad_norm": 0.24471600353717804, + "learning_rate": 0.0004822830081486175, + "loss": 1.8975, + "step": 1564 + }, + { + "epoch": 0.15283203125, + "grad_norm": 0.21125410497188568, + "learning_rate": 0.00048225533840840493, + "loss": 1.9513, + "step": 1565 + }, + { + "epoch": 0.1529296875, + "grad_norm": 0.22737543284893036, + "learning_rate": 0.00048222764796522017, + "loss": 1.8962, + "step": 1566 + }, + { + "epoch": 0.15302734375, + "grad_norm": 0.1857408881187439, + "learning_rate": 0.0004821999368218292, + "loss": 1.9054, + "step": 1567 + }, + { + "epoch": 0.153125, + "grad_norm": 0.23223114013671875, + "learning_rate": 0.0004821722049810002, + "loss": 1.9118, + "step": 1568 + }, + { + "epoch": 0.15322265625, + "grad_norm": 0.25117412209510803, + "learning_rate": 0.0004821444524455034, + "loss": 1.8889, + "step": 1569 + }, + { + "epoch": 0.1533203125, + "grad_norm": 0.23292681574821472, + "learning_rate": 0.00048211667921811084, + "loss": 1.9495, + "step": 1570 + }, + { + "epoch": 0.15341796875, + "grad_norm": 0.24508163332939148, + "learning_rate": 0.0004820888853015969, + "loss": 1.922, + "step": 1571 + }, + { + "epoch": 0.153515625, + "grad_norm": 0.26448142528533936, + "learning_rate": 0.000482061070698738, + "loss": 1.9811, + "step": 1572 + }, + { + "epoch": 0.15361328125, + "grad_norm": 0.24274107813835144, + "learning_rate": 0.0004820332354123125, + "loss": 1.9134, + "step": 1573 + }, + { + "epoch": 0.1537109375, + "grad_norm": 0.2016661912202835, + "learning_rate": 0.00048200537944510097, + "loss": 1.9151, + "step": 1574 + }, + { + "epoch": 0.15380859375, + "grad_norm": 0.21885213255882263, + "learning_rate": 0.00048197750279988595, + "loss": 1.9613, + "step": 1575 + }, + { + "epoch": 0.15390625, + "grad_norm": 0.2705588936805725, + "learning_rate": 0.000481949605479452, + "loss": 1.9535, + "step": 1576 + }, + { + "epoch": 0.15400390625, + "grad_norm": 0.31828904151916504, + "learning_rate": 0.00048192168748658595, + "loss": 1.9592, + "step": 1577 + }, + { + "epoch": 0.1541015625, + "grad_norm": 0.38848698139190674, + "learning_rate": 0.0004818937488240764, + "loss": 1.9027, + "step": 1578 + }, + { + "epoch": 0.15419921875, + "grad_norm": 0.2857527732849121, + "learning_rate": 0.00048186578949471434, + "loss": 1.9888, + "step": 1579 + }, + { + "epoch": 0.154296875, + "grad_norm": 0.30096855759620667, + "learning_rate": 0.0004818378095012925, + "loss": 1.9199, + "step": 1580 + }, + { + "epoch": 0.15439453125, + "grad_norm": 0.3158320486545563, + "learning_rate": 0.00048180980884660594, + "loss": 1.9081, + "step": 1581 + }, + { + "epoch": 0.1544921875, + "grad_norm": 0.22765566408634186, + "learning_rate": 0.00048178178753345157, + "loss": 1.9181, + "step": 1582 + }, + { + "epoch": 0.15458984375, + "grad_norm": 0.28936147689819336, + "learning_rate": 0.0004817537455646286, + "loss": 1.9424, + "step": 1583 + }, + { + "epoch": 0.1546875, + "grad_norm": 0.3088884651660919, + "learning_rate": 0.00048172568294293816, + "loss": 1.9088, + "step": 1584 + }, + { + "epoch": 0.15478515625, + "grad_norm": 0.2758021950721741, + "learning_rate": 0.00048169759967118343, + "loss": 1.9298, + "step": 1585 + }, + { + "epoch": 0.1548828125, + "grad_norm": 0.2999780774116516, + "learning_rate": 0.0004816694957521696, + "loss": 1.9335, + "step": 1586 + }, + { + "epoch": 0.15498046875, + "grad_norm": 0.22048184275627136, + "learning_rate": 0.0004816413711887041, + "loss": 1.8843, + "step": 1587 + }, + { + "epoch": 0.155078125, + "grad_norm": 0.26630088686943054, + "learning_rate": 0.0004816132259835963, + "loss": 1.9073, + "step": 1588 + }, + { + "epoch": 0.15517578125, + "grad_norm": 0.2615686058998108, + "learning_rate": 0.0004815850601396577, + "loss": 1.9313, + "step": 1589 + }, + { + "epoch": 0.1552734375, + "grad_norm": 0.2473047822713852, + "learning_rate": 0.00048155687365970167, + "loss": 1.8878, + "step": 1590 + }, + { + "epoch": 0.15537109375, + "grad_norm": 0.3328106999397278, + "learning_rate": 0.000481528666546544, + "loss": 1.9424, + "step": 1591 + }, + { + "epoch": 0.15546875, + "grad_norm": 0.26094701886177063, + "learning_rate": 0.0004815004388030022, + "loss": 1.9623, + "step": 1592 + }, + { + "epoch": 0.15556640625, + "grad_norm": 0.3564736247062683, + "learning_rate": 0.0004814721904318959, + "loss": 1.8966, + "step": 1593 + }, + { + "epoch": 0.1556640625, + "grad_norm": 0.3765439987182617, + "learning_rate": 0.000481443921436047, + "loss": 1.9189, + "step": 1594 + }, + { + "epoch": 0.15576171875, + "grad_norm": 0.36775973439216614, + "learning_rate": 0.0004814156318182793, + "loss": 1.9427, + "step": 1595 + }, + { + "epoch": 0.155859375, + "grad_norm": 0.3822444975376129, + "learning_rate": 0.00048138732158141867, + "loss": 1.9329, + "step": 1596 + }, + { + "epoch": 0.15595703125, + "grad_norm": 0.26008525490760803, + "learning_rate": 0.00048135899072829295, + "loss": 1.928, + "step": 1597 + }, + { + "epoch": 0.1560546875, + "grad_norm": 0.32837969064712524, + "learning_rate": 0.0004813306392617324, + "loss": 1.9409, + "step": 1598 + }, + { + "epoch": 0.15615234375, + "grad_norm": 0.30874279141426086, + "learning_rate": 0.0004813022671845687, + "loss": 1.9378, + "step": 1599 + }, + { + "epoch": 0.15625, + "grad_norm": 0.2988525629043579, + "learning_rate": 0.00048127387449963633, + "loss": 1.9091, + "step": 1600 + }, + { + "epoch": 0.15634765625, + "grad_norm": 0.4251779615879059, + "learning_rate": 0.0004812454612097713, + "loss": 1.9693, + "step": 1601 + }, + { + "epoch": 0.1564453125, + "grad_norm": 0.2827276885509491, + "learning_rate": 0.00048121702731781184, + "loss": 1.9177, + "step": 1602 + }, + { + "epoch": 0.15654296875, + "grad_norm": 0.24983742833137512, + "learning_rate": 0.00048118857282659834, + "loss": 1.9235, + "step": 1603 + }, + { + "epoch": 0.156640625, + "grad_norm": 0.25737494230270386, + "learning_rate": 0.000481160097738973, + "loss": 1.9678, + "step": 1604 + }, + { + "epoch": 0.15673828125, + "grad_norm": 0.38471952080726624, + "learning_rate": 0.00048113160205778035, + "loss": 1.9403, + "step": 1605 + }, + { + "epoch": 0.1568359375, + "grad_norm": 0.30553486943244934, + "learning_rate": 0.0004811030857858667, + "loss": 1.939, + "step": 1606 + }, + { + "epoch": 0.15693359375, + "grad_norm": 0.3170030117034912, + "learning_rate": 0.0004810745489260808, + "loss": 1.9331, + "step": 1607 + }, + { + "epoch": 0.15703125, + "grad_norm": 0.3244478404521942, + "learning_rate": 0.0004810459914812731, + "loss": 1.8986, + "step": 1608 + }, + { + "epoch": 0.15712890625, + "grad_norm": 0.24035608768463135, + "learning_rate": 0.0004810174134542963, + "loss": 1.9542, + "step": 1609 + }, + { + "epoch": 0.1572265625, + "grad_norm": 0.29910752177238464, + "learning_rate": 0.00048098881484800505, + "loss": 1.8521, + "step": 1610 + }, + { + "epoch": 0.15732421875, + "grad_norm": 0.2832294702529907, + "learning_rate": 0.000480960195665256, + "loss": 1.9567, + "step": 1611 + }, + { + "epoch": 0.157421875, + "grad_norm": 0.3093940317630768, + "learning_rate": 0.0004809315559089081, + "loss": 1.9332, + "step": 1612 + }, + { + "epoch": 0.15751953125, + "grad_norm": 0.2190847098827362, + "learning_rate": 0.00048090289558182215, + "loss": 1.8979, + "step": 1613 + }, + { + "epoch": 0.1576171875, + "grad_norm": 0.2461078017950058, + "learning_rate": 0.00048087421468686107, + "loss": 1.9032, + "step": 1614 + }, + { + "epoch": 0.15771484375, + "grad_norm": 0.2852519750595093, + "learning_rate": 0.0004808455132268898, + "loss": 1.9457, + "step": 1615 + }, + { + "epoch": 0.1578125, + "grad_norm": 0.23334845900535583, + "learning_rate": 0.00048081679120477543, + "loss": 1.8718, + "step": 1616 + }, + { + "epoch": 0.15791015625, + "grad_norm": 0.2714453637599945, + "learning_rate": 0.000480788048623387, + "loss": 1.9212, + "step": 1617 + }, + { + "epoch": 0.1580078125, + "grad_norm": 0.2445923238992691, + "learning_rate": 0.0004807592854855955, + "loss": 1.9322, + "step": 1618 + }, + { + "epoch": 0.15810546875, + "grad_norm": 0.2522878050804138, + "learning_rate": 0.0004807305017942744, + "loss": 1.939, + "step": 1619 + }, + { + "epoch": 0.158203125, + "grad_norm": 0.28525444865226746, + "learning_rate": 0.0004807016975522987, + "loss": 1.929, + "step": 1620 + }, + { + "epoch": 0.15830078125, + "grad_norm": 0.22004222869873047, + "learning_rate": 0.0004806728727625458, + "loss": 1.9431, + "step": 1621 + }, + { + "epoch": 0.1583984375, + "grad_norm": 0.20370328426361084, + "learning_rate": 0.000480644027427895, + "loss": 1.9063, + "step": 1622 + }, + { + "epoch": 0.15849609375, + "grad_norm": 0.21263106167316437, + "learning_rate": 0.00048061516155122763, + "loss": 1.9098, + "step": 1623 + }, + { + "epoch": 0.15859375, + "grad_norm": 0.25402888655662537, + "learning_rate": 0.0004805862751354273, + "loss": 1.8932, + "step": 1624 + }, + { + "epoch": 0.15869140625, + "grad_norm": 0.21452878415584564, + "learning_rate": 0.0004805573681833794, + "loss": 1.9674, + "step": 1625 + }, + { + "epoch": 0.1587890625, + "grad_norm": 0.2378520965576172, + "learning_rate": 0.0004805284406979715, + "loss": 1.8841, + "step": 1626 + }, + { + "epoch": 0.15888671875, + "grad_norm": 0.20481383800506592, + "learning_rate": 0.0004804994926820932, + "loss": 1.8836, + "step": 1627 + }, + { + "epoch": 0.158984375, + "grad_norm": 0.19195064902305603, + "learning_rate": 0.00048047052413863607, + "loss": 1.9694, + "step": 1628 + }, + { + "epoch": 0.15908203125, + "grad_norm": 0.228655144572258, + "learning_rate": 0.0004804415350704939, + "loss": 1.9323, + "step": 1629 + }, + { + "epoch": 0.1591796875, + "grad_norm": 0.2156544327735901, + "learning_rate": 0.0004804125254805625, + "loss": 1.9236, + "step": 1630 + }, + { + "epoch": 0.15927734375, + "grad_norm": 0.24159860610961914, + "learning_rate": 0.00048038349537173953, + "loss": 1.9611, + "step": 1631 + }, + { + "epoch": 0.159375, + "grad_norm": 0.21641957759857178, + "learning_rate": 0.0004803544447469249, + "loss": 1.9237, + "step": 1632 + }, + { + "epoch": 0.15947265625, + "grad_norm": 0.21140074729919434, + "learning_rate": 0.0004803253736090205, + "loss": 1.9297, + "step": 1633 + }, + { + "epoch": 0.1595703125, + "grad_norm": 0.23588809370994568, + "learning_rate": 0.00048029628196093047, + "loss": 1.8995, + "step": 1634 + }, + { + "epoch": 0.15966796875, + "grad_norm": 0.2083786576986313, + "learning_rate": 0.00048026716980556044, + "loss": 1.9154, + "step": 1635 + }, + { + "epoch": 0.159765625, + "grad_norm": 0.21067486703395844, + "learning_rate": 0.00048023803714581867, + "loss": 1.9032, + "step": 1636 + }, + { + "epoch": 0.15986328125, + "grad_norm": 0.19560249149799347, + "learning_rate": 0.0004802088839846153, + "loss": 1.9352, + "step": 1637 + }, + { + "epoch": 0.1599609375, + "grad_norm": 0.21477989852428436, + "learning_rate": 0.0004801797103248624, + "loss": 1.9192, + "step": 1638 + }, + { + "epoch": 0.16005859375, + "grad_norm": 0.27302122116088867, + "learning_rate": 0.0004801505161694741, + "loss": 1.9177, + "step": 1639 + }, + { + "epoch": 0.16015625, + "grad_norm": 0.251449316740036, + "learning_rate": 0.00048012130152136676, + "loss": 1.9383, + "step": 1640 + }, + { + "epoch": 0.16025390625, + "grad_norm": 0.29725855588912964, + "learning_rate": 0.00048009206638345857, + "loss": 1.9027, + "step": 1641 + }, + { + "epoch": 0.1603515625, + "grad_norm": 0.3364620804786682, + "learning_rate": 0.00048006281075866996, + "loss": 1.8999, + "step": 1642 + }, + { + "epoch": 0.16044921875, + "grad_norm": 0.33240416646003723, + "learning_rate": 0.00048003353464992323, + "loss": 1.9023, + "step": 1643 + }, + { + "epoch": 0.160546875, + "grad_norm": 0.29456380009651184, + "learning_rate": 0.0004800042380601428, + "loss": 1.8874, + "step": 1644 + }, + { + "epoch": 0.16064453125, + "grad_norm": 0.20352846384048462, + "learning_rate": 0.00047997492099225515, + "loss": 1.8813, + "step": 1645 + }, + { + "epoch": 0.1607421875, + "grad_norm": 0.2666397988796234, + "learning_rate": 0.0004799455834491889, + "loss": 1.9582, + "step": 1646 + }, + { + "epoch": 0.16083984375, + "grad_norm": 0.34426215291023254, + "learning_rate": 0.0004799162254338744, + "loss": 1.9372, + "step": 1647 + }, + { + "epoch": 0.1609375, + "grad_norm": 0.3887472450733185, + "learning_rate": 0.00047988684694924447, + "loss": 1.9381, + "step": 1648 + }, + { + "epoch": 0.16103515625, + "grad_norm": 0.3413529694080353, + "learning_rate": 0.0004798574479982336, + "loss": 1.8966, + "step": 1649 + }, + { + "epoch": 0.1611328125, + "grad_norm": 0.2743453085422516, + "learning_rate": 0.00047982802858377853, + "loss": 1.9228, + "step": 1650 + }, + { + "epoch": 0.16123046875, + "grad_norm": 0.29950737953186035, + "learning_rate": 0.00047979858870881805, + "loss": 1.9204, + "step": 1651 + }, + { + "epoch": 0.161328125, + "grad_norm": 0.273075133562088, + "learning_rate": 0.00047976912837629295, + "loss": 1.8683, + "step": 1652 + }, + { + "epoch": 0.16142578125, + "grad_norm": 0.2547522485256195, + "learning_rate": 0.00047973964758914596, + "loss": 1.9275, + "step": 1653 + }, + { + "epoch": 0.1615234375, + "grad_norm": 0.2716279923915863, + "learning_rate": 0.00047971014635032205, + "loss": 1.8917, + "step": 1654 + }, + { + "epoch": 0.16162109375, + "grad_norm": 0.28852027654647827, + "learning_rate": 0.00047968062466276803, + "loss": 1.9418, + "step": 1655 + }, + { + "epoch": 0.16171875, + "grad_norm": 0.19788789749145508, + "learning_rate": 0.00047965108252943293, + "loss": 1.922, + "step": 1656 + }, + { + "epoch": 0.16181640625, + "grad_norm": 0.33596405386924744, + "learning_rate": 0.0004796215199532678, + "loss": 1.9291, + "step": 1657 + }, + { + "epoch": 0.1619140625, + "grad_norm": 0.32864540815353394, + "learning_rate": 0.00047959193693722553, + "loss": 1.9793, + "step": 1658 + }, + { + "epoch": 0.16201171875, + "grad_norm": 0.25629329681396484, + "learning_rate": 0.0004795623334842613, + "loss": 1.9453, + "step": 1659 + }, + { + "epoch": 0.162109375, + "grad_norm": 0.2877620458602905, + "learning_rate": 0.0004795327095973322, + "loss": 1.9225, + "step": 1660 + }, + { + "epoch": 0.16220703125, + "grad_norm": 0.24507132172584534, + "learning_rate": 0.00047950306527939744, + "loss": 1.8917, + "step": 1661 + }, + { + "epoch": 0.1623046875, + "grad_norm": 0.2875724732875824, + "learning_rate": 0.0004794734005334182, + "loss": 1.8819, + "step": 1662 + }, + { + "epoch": 0.16240234375, + "grad_norm": 0.2515362501144409, + "learning_rate": 0.0004794437153623577, + "loss": 1.9232, + "step": 1663 + }, + { + "epoch": 0.1625, + "grad_norm": 0.27037444710731506, + "learning_rate": 0.00047941400976918123, + "loss": 1.9567, + "step": 1664 + }, + { + "epoch": 0.16259765625, + "grad_norm": 0.29190242290496826, + "learning_rate": 0.0004793842837568562, + "loss": 1.9266, + "step": 1665 + }, + { + "epoch": 0.1626953125, + "grad_norm": 0.26770535111427307, + "learning_rate": 0.0004793545373283518, + "loss": 1.9163, + "step": 1666 + }, + { + "epoch": 0.16279296875, + "grad_norm": 0.25330448150634766, + "learning_rate": 0.0004793247704866396, + "loss": 1.9182, + "step": 1667 + }, + { + "epoch": 0.162890625, + "grad_norm": 0.27300313115119934, + "learning_rate": 0.00047929498323469295, + "loss": 1.9592, + "step": 1668 + }, + { + "epoch": 0.16298828125, + "grad_norm": 0.21413999795913696, + "learning_rate": 0.00047926517557548746, + "loss": 1.919, + "step": 1669 + }, + { + "epoch": 0.1630859375, + "grad_norm": 0.27660825848579407, + "learning_rate": 0.0004792353475120004, + "loss": 1.9176, + "step": 1670 + }, + { + "epoch": 0.16318359375, + "grad_norm": 0.2682628631591797, + "learning_rate": 0.00047920549904721156, + "loss": 1.9381, + "step": 1671 + }, + { + "epoch": 0.16328125, + "grad_norm": 0.2719988226890564, + "learning_rate": 0.00047917563018410247, + "loss": 1.9137, + "step": 1672 + }, + { + "epoch": 0.16337890625, + "grad_norm": 0.23097728192806244, + "learning_rate": 0.0004791457409256568, + "loss": 1.9128, + "step": 1673 + }, + { + "epoch": 0.1634765625, + "grad_norm": 0.27505284547805786, + "learning_rate": 0.00047911583127486015, + "loss": 1.9026, + "step": 1674 + }, + { + "epoch": 0.16357421875, + "grad_norm": 0.31115567684173584, + "learning_rate": 0.0004790859012347002, + "loss": 1.9438, + "step": 1675 + }, + { + "epoch": 0.163671875, + "grad_norm": 0.2766979932785034, + "learning_rate": 0.0004790559508081668, + "loss": 1.9389, + "step": 1676 + }, + { + "epoch": 0.16376953125, + "grad_norm": 0.25364476442337036, + "learning_rate": 0.0004790259799982517, + "loss": 1.9621, + "step": 1677 + }, + { + "epoch": 0.1638671875, + "grad_norm": 0.2380513697862625, + "learning_rate": 0.0004789959888079487, + "loss": 1.8891, + "step": 1678 + }, + { + "epoch": 0.16396484375, + "grad_norm": 0.21375243365764618, + "learning_rate": 0.0004789659772402536, + "loss": 1.8864, + "step": 1679 + }, + { + "epoch": 0.1640625, + "grad_norm": 0.2307070791721344, + "learning_rate": 0.00047893594529816445, + "loss": 1.8975, + "step": 1680 + }, + { + "epoch": 0.16416015625, + "grad_norm": 0.20362551510334015, + "learning_rate": 0.000478905892984681, + "loss": 1.9489, + "step": 1681 + }, + { + "epoch": 0.1642578125, + "grad_norm": 0.23438306152820587, + "learning_rate": 0.00047887582030280536, + "loss": 1.9874, + "step": 1682 + }, + { + "epoch": 0.16435546875, + "grad_norm": 0.19733543694019318, + "learning_rate": 0.00047884572725554137, + "loss": 1.9147, + "step": 1683 + }, + { + "epoch": 0.164453125, + "grad_norm": 0.2748172879219055, + "learning_rate": 0.0004788156138458952, + "loss": 1.863, + "step": 1684 + }, + { + "epoch": 0.16455078125, + "grad_norm": 0.26911303400993347, + "learning_rate": 0.00047878548007687476, + "loss": 1.8852, + "step": 1685 + }, + { + "epoch": 0.1646484375, + "grad_norm": 0.20826198160648346, + "learning_rate": 0.0004787553259514903, + "loss": 1.933, + "step": 1686 + }, + { + "epoch": 0.16474609375, + "grad_norm": 0.3136061429977417, + "learning_rate": 0.00047872515147275393, + "loss": 1.9264, + "step": 1687 + }, + { + "epoch": 0.16484375, + "grad_norm": 0.22931191325187683, + "learning_rate": 0.0004786949566436797, + "loss": 1.9175, + "step": 1688 + }, + { + "epoch": 0.16494140625, + "grad_norm": 0.24129685759544373, + "learning_rate": 0.0004786647414672839, + "loss": 1.9114, + "step": 1689 + }, + { + "epoch": 0.1650390625, + "grad_norm": 0.2148384302854538, + "learning_rate": 0.00047863450594658473, + "loss": 1.9189, + "step": 1690 + }, + { + "epoch": 0.16513671875, + "grad_norm": 0.24560751020908356, + "learning_rate": 0.0004786042500846025, + "loss": 1.9459, + "step": 1691 + }, + { + "epoch": 0.165234375, + "grad_norm": 0.2328975796699524, + "learning_rate": 0.0004785739738843594, + "loss": 1.9037, + "step": 1692 + }, + { + "epoch": 0.16533203125, + "grad_norm": 0.19054313004016876, + "learning_rate": 0.00047854367734887984, + "loss": 1.9035, + "step": 1693 + }, + { + "epoch": 0.1654296875, + "grad_norm": 0.19610174000263214, + "learning_rate": 0.00047851336048119016, + "loss": 1.8709, + "step": 1694 + }, + { + "epoch": 0.16552734375, + "grad_norm": 0.22446346282958984, + "learning_rate": 0.0004784830232843186, + "loss": 1.8866, + "step": 1695 + }, + { + "epoch": 0.165625, + "grad_norm": 0.22055386006832123, + "learning_rate": 0.0004784526657612959, + "loss": 1.9104, + "step": 1696 + }, + { + "epoch": 0.16572265625, + "grad_norm": 0.19776830077171326, + "learning_rate": 0.0004784222879151542, + "loss": 1.9302, + "step": 1697 + }, + { + "epoch": 0.1658203125, + "grad_norm": 0.19348515570163727, + "learning_rate": 0.0004783918897489282, + "loss": 1.944, + "step": 1698 + }, + { + "epoch": 0.16591796875, + "grad_norm": 0.21035566926002502, + "learning_rate": 0.0004783614712656542, + "loss": 1.9697, + "step": 1699 + }, + { + "epoch": 0.166015625, + "grad_norm": 0.20167845487594604, + "learning_rate": 0.0004783310324683709, + "loss": 1.9087, + "step": 1700 + }, + { + "epoch": 0.16611328125, + "grad_norm": 0.24482861161231995, + "learning_rate": 0.0004783005733601188, + "loss": 1.9316, + "step": 1701 + }, + { + "epoch": 0.1662109375, + "grad_norm": 0.2270205020904541, + "learning_rate": 0.00047827009394394045, + "loss": 1.885, + "step": 1702 + }, + { + "epoch": 0.16630859375, + "grad_norm": 0.2608849108219147, + "learning_rate": 0.0004782395942228806, + "loss": 1.9639, + "step": 1703 + }, + { + "epoch": 0.16640625, + "grad_norm": 0.29610663652420044, + "learning_rate": 0.0004782090741999858, + "loss": 1.9248, + "step": 1704 + }, + { + "epoch": 0.16650390625, + "grad_norm": 0.44481170177459717, + "learning_rate": 0.0004781785338783047, + "loss": 1.9352, + "step": 1705 + }, + { + "epoch": 0.1666015625, + "grad_norm": 0.4372797906398773, + "learning_rate": 0.0004781479732608881, + "loss": 1.9557, + "step": 1706 + }, + { + "epoch": 0.16669921875, + "grad_norm": 0.2859649956226349, + "learning_rate": 0.00047811739235078876, + "loss": 1.9753, + "step": 1707 + }, + { + "epoch": 0.166796875, + "grad_norm": 0.26134198904037476, + "learning_rate": 0.00047808679115106135, + "loss": 1.9513, + "step": 1708 + }, + { + "epoch": 0.16689453125, + "grad_norm": 0.30859240889549255, + "learning_rate": 0.0004780561696647627, + "loss": 1.9084, + "step": 1709 + }, + { + "epoch": 0.1669921875, + "grad_norm": 0.33629563450813293, + "learning_rate": 0.0004780255278949516, + "loss": 1.928, + "step": 1710 + }, + { + "epoch": 0.16708984375, + "grad_norm": 0.2672228217124939, + "learning_rate": 0.0004779948658446889, + "loss": 1.9033, + "step": 1711 + }, + { + "epoch": 0.1671875, + "grad_norm": 0.2845448851585388, + "learning_rate": 0.00047796418351703744, + "loss": 1.8892, + "step": 1712 + }, + { + "epoch": 0.16728515625, + "grad_norm": 0.236465185880661, + "learning_rate": 0.0004779334809150622, + "loss": 1.9363, + "step": 1713 + }, + { + "epoch": 0.1673828125, + "grad_norm": 0.2997010052204132, + "learning_rate": 0.00047790275804183004, + "loss": 1.8917, + "step": 1714 + }, + { + "epoch": 0.16748046875, + "grad_norm": 0.3848778307437897, + "learning_rate": 0.00047787201490040985, + "loss": 1.959, + "step": 1715 + }, + { + "epoch": 0.167578125, + "grad_norm": 0.2715388834476471, + "learning_rate": 0.00047784125149387277, + "loss": 1.9463, + "step": 1716 + }, + { + "epoch": 0.16767578125, + "grad_norm": 0.3423638343811035, + "learning_rate": 0.0004778104678252916, + "loss": 1.9393, + "step": 1717 + }, + { + "epoch": 0.1677734375, + "grad_norm": 0.3216153383255005, + "learning_rate": 0.0004777796638977414, + "loss": 1.9688, + "step": 1718 + }, + { + "epoch": 0.16787109375, + "grad_norm": 0.24108828604221344, + "learning_rate": 0.00047774883971429935, + "loss": 1.8822, + "step": 1719 + }, + { + "epoch": 0.16796875, + "grad_norm": 0.3516473174095154, + "learning_rate": 0.0004777179952780443, + "loss": 1.912, + "step": 1720 + }, + { + "epoch": 0.16806640625, + "grad_norm": 0.2251645028591156, + "learning_rate": 0.0004776871305920575, + "loss": 1.8989, + "step": 1721 + }, + { + "epoch": 0.1681640625, + "grad_norm": 0.282797634601593, + "learning_rate": 0.000477656245659422, + "loss": 1.8483, + "step": 1722 + }, + { + "epoch": 0.16826171875, + "grad_norm": 0.3045579195022583, + "learning_rate": 0.0004776253404832229, + "loss": 1.8934, + "step": 1723 + }, + { + "epoch": 0.168359375, + "grad_norm": 0.27416738867759705, + "learning_rate": 0.00047759441506654745, + "loss": 1.924, + "step": 1724 + }, + { + "epoch": 0.16845703125, + "grad_norm": 0.2600431442260742, + "learning_rate": 0.0004775634694124847, + "loss": 1.9507, + "step": 1725 + }, + { + "epoch": 0.1685546875, + "grad_norm": 0.22953875362873077, + "learning_rate": 0.0004775325035241259, + "loss": 1.9062, + "step": 1726 + }, + { + "epoch": 0.16865234375, + "grad_norm": 0.20656803250312805, + "learning_rate": 0.00047750151740456426, + "loss": 1.9312, + "step": 1727 + }, + { + "epoch": 0.16875, + "grad_norm": 0.21857194602489471, + "learning_rate": 0.0004774705110568951, + "loss": 1.931, + "step": 1728 + }, + { + "epoch": 0.16884765625, + "grad_norm": 0.2585659325122833, + "learning_rate": 0.0004774394844842155, + "loss": 1.9149, + "step": 1729 + }, + { + "epoch": 0.1689453125, + "grad_norm": 0.23686473071575165, + "learning_rate": 0.000477408437689625, + "loss": 1.9316, + "step": 1730 + }, + { + "epoch": 0.16904296875, + "grad_norm": 0.21281464397907257, + "learning_rate": 0.0004773773706762247, + "loss": 1.9278, + "step": 1731 + }, + { + "epoch": 0.169140625, + "grad_norm": 0.2731045186519623, + "learning_rate": 0.00047734628344711796, + "loss": 1.9088, + "step": 1732 + }, + { + "epoch": 0.16923828125, + "grad_norm": 0.21235254406929016, + "learning_rate": 0.00047731517600541014, + "loss": 1.9418, + "step": 1733 + }, + { + "epoch": 0.1693359375, + "grad_norm": 0.2536609172821045, + "learning_rate": 0.0004772840483542085, + "loss": 1.9651, + "step": 1734 + }, + { + "epoch": 0.16943359375, + "grad_norm": 0.22243930399417877, + "learning_rate": 0.00047725290049662264, + "loss": 1.9158, + "step": 1735 + }, + { + "epoch": 0.16953125, + "grad_norm": 0.24414226412773132, + "learning_rate": 0.00047722173243576374, + "loss": 1.9167, + "step": 1736 + }, + { + "epoch": 0.16962890625, + "grad_norm": 0.25493788719177246, + "learning_rate": 0.0004771905441747453, + "loss": 1.9244, + "step": 1737 + }, + { + "epoch": 0.1697265625, + "grad_norm": 0.21703946590423584, + "learning_rate": 0.0004771593357166827, + "loss": 1.8882, + "step": 1738 + }, + { + "epoch": 0.16982421875, + "grad_norm": 0.27353623509407043, + "learning_rate": 0.00047712810706469354, + "loss": 1.9134, + "step": 1739 + }, + { + "epoch": 0.169921875, + "grad_norm": 0.2163049578666687, + "learning_rate": 0.0004770968582218971, + "loss": 1.9226, + "step": 1740 + }, + { + "epoch": 0.17001953125, + "grad_norm": 0.1948620080947876, + "learning_rate": 0.000477065589191415, + "loss": 1.8663, + "step": 1741 + }, + { + "epoch": 0.1701171875, + "grad_norm": 0.22654829919338226, + "learning_rate": 0.0004770342999763706, + "loss": 1.9327, + "step": 1742 + }, + { + "epoch": 0.17021484375, + "grad_norm": 0.24024905264377594, + "learning_rate": 0.00047700299057988957, + "loss": 1.9581, + "step": 1743 + }, + { + "epoch": 0.1703125, + "grad_norm": 0.2727830111980438, + "learning_rate": 0.0004769716610050993, + "loss": 1.9158, + "step": 1744 + }, + { + "epoch": 0.17041015625, + "grad_norm": 0.2848643362522125, + "learning_rate": 0.0004769403112551295, + "loss": 1.8943, + "step": 1745 + }, + { + "epoch": 0.1705078125, + "grad_norm": 0.29621654748916626, + "learning_rate": 0.00047690894133311157, + "loss": 1.9182, + "step": 1746 + }, + { + "epoch": 0.17060546875, + "grad_norm": 0.317327618598938, + "learning_rate": 0.00047687755124217914, + "loss": 1.9624, + "step": 1747 + }, + { + "epoch": 0.170703125, + "grad_norm": 0.3084838390350342, + "learning_rate": 0.00047684614098546783, + "loss": 1.919, + "step": 1748 + }, + { + "epoch": 0.17080078125, + "grad_norm": 0.26487037539482117, + "learning_rate": 0.00047681471056611527, + "loss": 1.9005, + "step": 1749 + }, + { + "epoch": 0.1708984375, + "grad_norm": 0.2641773223876953, + "learning_rate": 0.000476783259987261, + "loss": 1.9085, + "step": 1750 + }, + { + "epoch": 0.17099609375, + "grad_norm": 0.30503541231155396, + "learning_rate": 0.00047675178925204667, + "loss": 1.9148, + "step": 1751 + }, + { + "epoch": 0.17109375, + "grad_norm": 0.2899084985256195, + "learning_rate": 0.00047672029836361596, + "loss": 1.9369, + "step": 1752 + }, + { + "epoch": 0.17119140625, + "grad_norm": 0.3330906629562378, + "learning_rate": 0.00047668878732511457, + "loss": 1.8529, + "step": 1753 + }, + { + "epoch": 0.1712890625, + "grad_norm": 0.32898756861686707, + "learning_rate": 0.00047665725613969015, + "loss": 1.8918, + "step": 1754 + }, + { + "epoch": 0.17138671875, + "grad_norm": 0.21951165795326233, + "learning_rate": 0.0004766257048104923, + "loss": 1.8868, + "step": 1755 + }, + { + "epoch": 0.171484375, + "grad_norm": 0.20533274114131927, + "learning_rate": 0.00047659413334067276, + "loss": 1.8993, + "step": 1756 + }, + { + "epoch": 0.17158203125, + "grad_norm": 0.19753561913967133, + "learning_rate": 0.0004765625417333853, + "loss": 1.913, + "step": 1757 + }, + { + "epoch": 0.1716796875, + "grad_norm": 0.20068642497062683, + "learning_rate": 0.0004765309299917856, + "loss": 1.8639, + "step": 1758 + }, + { + "epoch": 0.17177734375, + "grad_norm": 0.22603540122509003, + "learning_rate": 0.00047649929811903143, + "loss": 1.9014, + "step": 1759 + }, + { + "epoch": 0.171875, + "grad_norm": 0.18157346546649933, + "learning_rate": 0.00047646764611828247, + "loss": 1.8956, + "step": 1760 + }, + { + "epoch": 0.17197265625, + "grad_norm": 0.21239453554153442, + "learning_rate": 0.0004764359739927005, + "loss": 1.9636, + "step": 1761 + }, + { + "epoch": 0.1720703125, + "grad_norm": 0.25655651092529297, + "learning_rate": 0.0004764042817454494, + "loss": 1.8901, + "step": 1762 + }, + { + "epoch": 0.17216796875, + "grad_norm": 0.20679731667041779, + "learning_rate": 0.0004763725693796947, + "loss": 1.8908, + "step": 1763 + }, + { + "epoch": 0.172265625, + "grad_norm": 0.20754674077033997, + "learning_rate": 0.0004763408368986043, + "loss": 1.9445, + "step": 1764 + }, + { + "epoch": 0.17236328125, + "grad_norm": 0.18912455439567566, + "learning_rate": 0.0004763090843053481, + "loss": 1.9176, + "step": 1765 + }, + { + "epoch": 0.1724609375, + "grad_norm": 0.20957696437835693, + "learning_rate": 0.0004762773116030978, + "loss": 1.9206, + "step": 1766 + }, + { + "epoch": 0.17255859375, + "grad_norm": 0.1931106299161911, + "learning_rate": 0.0004762455187950272, + "loss": 1.9051, + "step": 1767 + }, + { + "epoch": 0.17265625, + "grad_norm": 0.218103289604187, + "learning_rate": 0.00047621370588431217, + "loss": 1.9326, + "step": 1768 + }, + { + "epoch": 0.17275390625, + "grad_norm": 0.20649409294128418, + "learning_rate": 0.0004761818728741305, + "loss": 1.9156, + "step": 1769 + }, + { + "epoch": 0.1728515625, + "grad_norm": 0.2224937528371811, + "learning_rate": 0.000476150019767662, + "loss": 1.9181, + "step": 1770 + }, + { + "epoch": 0.17294921875, + "grad_norm": 0.30575814843177795, + "learning_rate": 0.00047611814656808865, + "loss": 1.9276, + "step": 1771 + }, + { + "epoch": 0.173046875, + "grad_norm": 0.34954264760017395, + "learning_rate": 0.0004760862532785942, + "loss": 1.8818, + "step": 1772 + }, + { + "epoch": 0.17314453125, + "grad_norm": 0.2726197838783264, + "learning_rate": 0.0004760543399023644, + "loss": 1.9279, + "step": 1773 + }, + { + "epoch": 0.1732421875, + "grad_norm": 0.2664092779159546, + "learning_rate": 0.00047602240644258726, + "loss": 1.9055, + "step": 1774 + }, + { + "epoch": 0.17333984375, + "grad_norm": 0.22661474347114563, + "learning_rate": 0.00047599045290245277, + "loss": 1.9406, + "step": 1775 + }, + { + "epoch": 0.1734375, + "grad_norm": 0.20825809240341187, + "learning_rate": 0.00047595847928515243, + "loss": 1.9189, + "step": 1776 + }, + { + "epoch": 0.17353515625, + "grad_norm": 0.200863778591156, + "learning_rate": 0.0004759264855938805, + "loss": 1.8985, + "step": 1777 + }, + { + "epoch": 0.1736328125, + "grad_norm": 0.18549345433712006, + "learning_rate": 0.00047589447183183254, + "loss": 1.8855, + "step": 1778 + }, + { + "epoch": 0.17373046875, + "grad_norm": 0.19504952430725098, + "learning_rate": 0.00047586243800220673, + "loss": 1.9494, + "step": 1779 + }, + { + "epoch": 0.173828125, + "grad_norm": 0.2109365612268448, + "learning_rate": 0.0004758303841082029, + "loss": 1.9223, + "step": 1780 + }, + { + "epoch": 0.17392578125, + "grad_norm": 0.20636534690856934, + "learning_rate": 0.00047579831015302277, + "loss": 1.8939, + "step": 1781 + }, + { + "epoch": 0.1740234375, + "grad_norm": 0.2035193145275116, + "learning_rate": 0.0004757662161398704, + "loss": 1.933, + "step": 1782 + }, + { + "epoch": 0.17412109375, + "grad_norm": 0.20754744112491608, + "learning_rate": 0.0004757341020719517, + "loss": 1.8626, + "step": 1783 + }, + { + "epoch": 0.17421875, + "grad_norm": 0.29153457283973694, + "learning_rate": 0.00047570196795247455, + "loss": 1.8666, + "step": 1784 + }, + { + "epoch": 0.17431640625, + "grad_norm": 0.36487746238708496, + "learning_rate": 0.00047566981378464884, + "loss": 1.8917, + "step": 1785 + }, + { + "epoch": 0.1744140625, + "grad_norm": 0.32802101969718933, + "learning_rate": 0.0004756376395716865, + "loss": 1.8797, + "step": 1786 + }, + { + "epoch": 0.17451171875, + "grad_norm": 0.2676317095756531, + "learning_rate": 0.0004756054453168015, + "loss": 1.9127, + "step": 1787 + }, + { + "epoch": 0.174609375, + "grad_norm": 0.21236379444599152, + "learning_rate": 0.0004755732310232097, + "loss": 1.8628, + "step": 1788 + }, + { + "epoch": 0.17470703125, + "grad_norm": 0.2659597396850586, + "learning_rate": 0.000475540996694129, + "loss": 1.8725, + "step": 1789 + }, + { + "epoch": 0.1748046875, + "grad_norm": 0.2773102819919586, + "learning_rate": 0.00047550874233277937, + "loss": 1.9394, + "step": 1790 + }, + { + "epoch": 0.17490234375, + "grad_norm": 0.2603963613510132, + "learning_rate": 0.00047547646794238277, + "loss": 1.9164, + "step": 1791 + }, + { + "epoch": 0.175, + "grad_norm": 0.21353478729724884, + "learning_rate": 0.00047544417352616305, + "loss": 1.9091, + "step": 1792 + }, + { + "epoch": 0.17509765625, + "grad_norm": 0.29758280515670776, + "learning_rate": 0.0004754118590873462, + "loss": 1.9137, + "step": 1793 + }, + { + "epoch": 0.1751953125, + "grad_norm": 0.27226370573043823, + "learning_rate": 0.0004753795246291601, + "loss": 1.944, + "step": 1794 + }, + { + "epoch": 0.17529296875, + "grad_norm": 0.3030240833759308, + "learning_rate": 0.0004753471701548347, + "loss": 1.8922, + "step": 1795 + }, + { + "epoch": 0.175390625, + "grad_norm": 0.2778298258781433, + "learning_rate": 0.0004753147956676019, + "loss": 1.9321, + "step": 1796 + }, + { + "epoch": 0.17548828125, + "grad_norm": 0.2760380506515503, + "learning_rate": 0.0004752824011706956, + "loss": 1.9063, + "step": 1797 + }, + { + "epoch": 0.1755859375, + "grad_norm": 0.2793836295604706, + "learning_rate": 0.00047524998666735184, + "loss": 1.9186, + "step": 1798 + }, + { + "epoch": 0.17568359375, + "grad_norm": 0.25465574860572815, + "learning_rate": 0.00047521755216080845, + "loss": 1.9247, + "step": 1799 + }, + { + "epoch": 0.17578125, + "grad_norm": 0.23424822092056274, + "learning_rate": 0.00047518509765430536, + "loss": 1.924, + "step": 1800 + }, + { + "epoch": 0.17587890625, + "grad_norm": 0.25181999802589417, + "learning_rate": 0.00047515262315108456, + "loss": 1.9536, + "step": 1801 + }, + { + "epoch": 0.1759765625, + "grad_norm": 0.2854491174221039, + "learning_rate": 0.00047512012865438976, + "loss": 1.9098, + "step": 1802 + }, + { + "epoch": 0.17607421875, + "grad_norm": 0.3238297998905182, + "learning_rate": 0.0004750876141674671, + "loss": 1.8833, + "step": 1803 + }, + { + "epoch": 0.176171875, + "grad_norm": 0.23803670704364777, + "learning_rate": 0.0004750550796935644, + "loss": 1.9685, + "step": 1804 + }, + { + "epoch": 0.17626953125, + "grad_norm": 0.2600926458835602, + "learning_rate": 0.00047502252523593154, + "loss": 1.913, + "step": 1805 + }, + { + "epoch": 0.1763671875, + "grad_norm": 0.30193689465522766, + "learning_rate": 0.00047498995079782046, + "loss": 1.9453, + "step": 1806 + }, + { + "epoch": 0.17646484375, + "grad_norm": 0.2594965994358063, + "learning_rate": 0.0004749573563824851, + "loss": 1.8932, + "step": 1807 + }, + { + "epoch": 0.1765625, + "grad_norm": 0.25451624393463135, + "learning_rate": 0.0004749247419931812, + "loss": 1.9312, + "step": 1808 + }, + { + "epoch": 0.17666015625, + "grad_norm": 0.2194494754076004, + "learning_rate": 0.00047489210763316674, + "loss": 1.8783, + "step": 1809 + }, + { + "epoch": 0.1767578125, + "grad_norm": 0.2316436916589737, + "learning_rate": 0.00047485945330570173, + "loss": 1.8924, + "step": 1810 + }, + { + "epoch": 0.17685546875, + "grad_norm": 0.2362855076789856, + "learning_rate": 0.00047482677901404785, + "loss": 1.8883, + "step": 1811 + }, + { + "epoch": 0.176953125, + "grad_norm": 0.19301187992095947, + "learning_rate": 0.00047479408476146906, + "loss": 1.8778, + "step": 1812 + }, + { + "epoch": 0.17705078125, + "grad_norm": 0.2641616463661194, + "learning_rate": 0.0004747613705512313, + "loss": 1.8932, + "step": 1813 + }, + { + "epoch": 0.1771484375, + "grad_norm": 0.24684876203536987, + "learning_rate": 0.00047472863638660227, + "loss": 1.8961, + "step": 1814 + }, + { + "epoch": 0.17724609375, + "grad_norm": 0.26180848479270935, + "learning_rate": 0.0004746958822708519, + "loss": 1.8992, + "step": 1815 + }, + { + "epoch": 0.17734375, + "grad_norm": 0.21258847415447235, + "learning_rate": 0.0004746631082072521, + "loss": 1.9513, + "step": 1816 + }, + { + "epoch": 0.17744140625, + "grad_norm": 0.23707158863544464, + "learning_rate": 0.00047463031419907656, + "loss": 1.9015, + "step": 1817 + }, + { + "epoch": 0.1775390625, + "grad_norm": 0.2716965675354004, + "learning_rate": 0.0004745975002496012, + "loss": 1.9023, + "step": 1818 + }, + { + "epoch": 0.17763671875, + "grad_norm": 0.19453154504299164, + "learning_rate": 0.0004745646663621038, + "loss": 1.8883, + "step": 1819 + }, + { + "epoch": 0.177734375, + "grad_norm": 0.28068456053733826, + "learning_rate": 0.00047453181253986437, + "loss": 1.893, + "step": 1820 + }, + { + "epoch": 0.17783203125, + "grad_norm": 0.24416638910770416, + "learning_rate": 0.0004744989387861645, + "loss": 1.9337, + "step": 1821 + }, + { + "epoch": 0.1779296875, + "grad_norm": 0.24311622977256775, + "learning_rate": 0.000474466045104288, + "loss": 1.9139, + "step": 1822 + }, + { + "epoch": 0.17802734375, + "grad_norm": 0.3018793761730194, + "learning_rate": 0.00047443313149752073, + "loss": 1.8941, + "step": 1823 + }, + { + "epoch": 0.178125, + "grad_norm": 0.2428286373615265, + "learning_rate": 0.00047440019796915044, + "loss": 1.9238, + "step": 1824 + }, + { + "epoch": 0.17822265625, + "grad_norm": 0.2676742970943451, + "learning_rate": 0.0004743672445224669, + "loss": 1.8603, + "step": 1825 + }, + { + "epoch": 0.1783203125, + "grad_norm": 0.27228862047195435, + "learning_rate": 0.00047433427116076184, + "loss": 1.9091, + "step": 1826 + }, + { + "epoch": 0.17841796875, + "grad_norm": 0.23455603420734406, + "learning_rate": 0.0004743012778873291, + "loss": 1.8429, + "step": 1827 + }, + { + "epoch": 0.178515625, + "grad_norm": 0.26803067326545715, + "learning_rate": 0.0004742682647054643, + "loss": 1.8386, + "step": 1828 + }, + { + "epoch": 0.17861328125, + "grad_norm": 0.30642521381378174, + "learning_rate": 0.0004742352316184652, + "loss": 1.8997, + "step": 1829 + }, + { + "epoch": 0.1787109375, + "grad_norm": 0.23812299966812134, + "learning_rate": 0.00047420217862963155, + "loss": 1.9538, + "step": 1830 + }, + { + "epoch": 0.17880859375, + "grad_norm": 0.21221864223480225, + "learning_rate": 0.000474169105742265, + "loss": 1.9395, + "step": 1831 + }, + { + "epoch": 0.17890625, + "grad_norm": 0.2294473797082901, + "learning_rate": 0.0004741360129596693, + "loss": 1.9051, + "step": 1832 + }, + { + "epoch": 0.17900390625, + "grad_norm": 0.27911970019340515, + "learning_rate": 0.00047410290028515004, + "loss": 1.9032, + "step": 1833 + }, + { + "epoch": 0.1791015625, + "grad_norm": 0.3017365336418152, + "learning_rate": 0.000474069767722015, + "loss": 1.9179, + "step": 1834 + }, + { + "epoch": 0.17919921875, + "grad_norm": 0.21640396118164062, + "learning_rate": 0.0004740366152735738, + "loss": 1.9204, + "step": 1835 + }, + { + "epoch": 0.179296875, + "grad_norm": 0.2559264302253723, + "learning_rate": 0.0004740034429431379, + "loss": 1.9145, + "step": 1836 + }, + { + "epoch": 0.17939453125, + "grad_norm": 0.24724526703357697, + "learning_rate": 0.00047397025073402116, + "loss": 1.9107, + "step": 1837 + }, + { + "epoch": 0.1794921875, + "grad_norm": 0.20995844900608063, + "learning_rate": 0.0004739370386495391, + "loss": 1.9408, + "step": 1838 + }, + { + "epoch": 0.17958984375, + "grad_norm": 0.24171245098114014, + "learning_rate": 0.00047390380669300923, + "loss": 1.9023, + "step": 1839 + }, + { + "epoch": 0.1796875, + "grad_norm": 0.2987540066242218, + "learning_rate": 0.00047387055486775123, + "loss": 1.932, + "step": 1840 + }, + { + "epoch": 0.17978515625, + "grad_norm": 0.2575046420097351, + "learning_rate": 0.0004738372831770866, + "loss": 1.8957, + "step": 1841 + }, + { + "epoch": 0.1798828125, + "grad_norm": 0.26643261313438416, + "learning_rate": 0.00047380399162433903, + "loss": 1.927, + "step": 1842 + }, + { + "epoch": 0.17998046875, + "grad_norm": 0.25104060769081116, + "learning_rate": 0.00047377068021283385, + "loss": 1.9379, + "step": 1843 + }, + { + "epoch": 0.180078125, + "grad_norm": 0.2125697284936905, + "learning_rate": 0.00047373734894589877, + "loss": 1.9262, + "step": 1844 + }, + { + "epoch": 0.18017578125, + "grad_norm": 0.3097136616706848, + "learning_rate": 0.0004737039978268631, + "loss": 1.9351, + "step": 1845 + }, + { + "epoch": 0.1802734375, + "grad_norm": 0.28434544801712036, + "learning_rate": 0.00047367062685905835, + "loss": 1.9399, + "step": 1846 + }, + { + "epoch": 0.18037109375, + "grad_norm": 0.2924177944660187, + "learning_rate": 0.00047363723604581815, + "loss": 1.8924, + "step": 1847 + }, + { + "epoch": 0.18046875, + "grad_norm": 0.2927818298339844, + "learning_rate": 0.0004736038253904778, + "loss": 1.8868, + "step": 1848 + }, + { + "epoch": 0.18056640625, + "grad_norm": 0.24946413934230804, + "learning_rate": 0.0004735703948963747, + "loss": 1.9145, + "step": 1849 + }, + { + "epoch": 0.1806640625, + "grad_norm": 0.2521260380744934, + "learning_rate": 0.00047353694456684844, + "loss": 1.9579, + "step": 1850 + }, + { + "epoch": 0.18076171875, + "grad_norm": 0.30722346901893616, + "learning_rate": 0.0004735034744052402, + "loss": 1.9279, + "step": 1851 + }, + { + "epoch": 0.180859375, + "grad_norm": 0.24488766491413116, + "learning_rate": 0.00047346998441489353, + "loss": 1.9055, + "step": 1852 + }, + { + "epoch": 0.18095703125, + "grad_norm": 0.29834532737731934, + "learning_rate": 0.00047343647459915364, + "loss": 1.9511, + "step": 1853 + }, + { + "epoch": 0.1810546875, + "grad_norm": 0.252108097076416, + "learning_rate": 0.000473402944961368, + "loss": 1.8903, + "step": 1854 + }, + { + "epoch": 0.18115234375, + "grad_norm": 0.21066221594810486, + "learning_rate": 0.00047336939550488575, + "loss": 1.9162, + "step": 1855 + }, + { + "epoch": 0.18125, + "grad_norm": 0.23784129321575165, + "learning_rate": 0.0004733358262330584, + "loss": 1.8824, + "step": 1856 + }, + { + "epoch": 0.18134765625, + "grad_norm": 0.20668673515319824, + "learning_rate": 0.000473302237149239, + "loss": 1.9197, + "step": 1857 + }, + { + "epoch": 0.1814453125, + "grad_norm": 0.21304477751255035, + "learning_rate": 0.00047326862825678296, + "loss": 1.883, + "step": 1858 + }, + { + "epoch": 0.18154296875, + "grad_norm": 0.240382581949234, + "learning_rate": 0.0004732349995590474, + "loss": 1.9065, + "step": 1859 + }, + { + "epoch": 0.181640625, + "grad_norm": 0.27892357110977173, + "learning_rate": 0.00047320135105939165, + "loss": 1.9338, + "step": 1860 + }, + { + "epoch": 0.18173828125, + "grad_norm": 0.22719697654247284, + "learning_rate": 0.0004731676827611768, + "loss": 1.8927, + "step": 1861 + }, + { + "epoch": 0.1818359375, + "grad_norm": 0.19273899495601654, + "learning_rate": 0.00047313399466776605, + "loss": 1.9132, + "step": 1862 + }, + { + "epoch": 0.18193359375, + "grad_norm": 0.2563495934009552, + "learning_rate": 0.0004731002867825246, + "loss": 1.9206, + "step": 1863 + }, + { + "epoch": 0.18203125, + "grad_norm": 0.29519474506378174, + "learning_rate": 0.0004730665591088195, + "loss": 1.9254, + "step": 1864 + }, + { + "epoch": 0.18212890625, + "grad_norm": 0.2755618095397949, + "learning_rate": 0.0004730328116500198, + "loss": 1.9348, + "step": 1865 + }, + { + "epoch": 0.1822265625, + "grad_norm": 0.25873810052871704, + "learning_rate": 0.00047299904440949663, + "loss": 1.9285, + "step": 1866 + }, + { + "epoch": 0.18232421875, + "grad_norm": 0.19655470550060272, + "learning_rate": 0.0004729652573906231, + "loss": 1.9268, + "step": 1867 + }, + { + "epoch": 0.182421875, + "grad_norm": 0.288585901260376, + "learning_rate": 0.00047293145059677403, + "loss": 1.9379, + "step": 1868 + }, + { + "epoch": 0.18251953125, + "grad_norm": 0.2560109794139862, + "learning_rate": 0.00047289762403132673, + "loss": 1.9258, + "step": 1869 + }, + { + "epoch": 0.1826171875, + "grad_norm": 0.2627304196357727, + "learning_rate": 0.00047286377769766, + "loss": 1.8885, + "step": 1870 + }, + { + "epoch": 0.18271484375, + "grad_norm": 0.3108537793159485, + "learning_rate": 0.0004728299115991547, + "loss": 1.9106, + "step": 1871 + }, + { + "epoch": 0.1828125, + "grad_norm": 0.24468936026096344, + "learning_rate": 0.00047279602573919386, + "loss": 1.9436, + "step": 1872 + }, + { + "epoch": 0.18291015625, + "grad_norm": 0.3040415048599243, + "learning_rate": 0.0004727621201211623, + "loss": 1.977, + "step": 1873 + }, + { + "epoch": 0.1830078125, + "grad_norm": 0.3214859366416931, + "learning_rate": 0.0004727281947484471, + "loss": 1.9259, + "step": 1874 + }, + { + "epoch": 0.18310546875, + "grad_norm": 0.3081786632537842, + "learning_rate": 0.0004726942496244369, + "loss": 1.8833, + "step": 1875 + }, + { + "epoch": 0.183203125, + "grad_norm": 0.2123297154903412, + "learning_rate": 0.0004726602847525225, + "loss": 1.9327, + "step": 1876 + }, + { + "epoch": 0.18330078125, + "grad_norm": 0.20726516842842102, + "learning_rate": 0.00047262630013609694, + "loss": 1.932, + "step": 1877 + }, + { + "epoch": 0.1833984375, + "grad_norm": 0.2831561863422394, + "learning_rate": 0.0004725922957785547, + "loss": 1.8966, + "step": 1878 + }, + { + "epoch": 0.18349609375, + "grad_norm": 0.25883564352989197, + "learning_rate": 0.00047255827168329255, + "loss": 1.8872, + "step": 1879 + }, + { + "epoch": 0.18359375, + "grad_norm": 0.24307870864868164, + "learning_rate": 0.00047252422785370936, + "loss": 1.935, + "step": 1880 + }, + { + "epoch": 0.18369140625, + "grad_norm": 0.2524338662624359, + "learning_rate": 0.00047249016429320567, + "loss": 1.887, + "step": 1881 + }, + { + "epoch": 0.1837890625, + "grad_norm": 0.22779789566993713, + "learning_rate": 0.0004724560810051843, + "loss": 1.9536, + "step": 1882 + }, + { + "epoch": 0.18388671875, + "grad_norm": 0.23726516962051392, + "learning_rate": 0.0004724219779930496, + "loss": 1.9267, + "step": 1883 + }, + { + "epoch": 0.183984375, + "grad_norm": 0.2780471742153168, + "learning_rate": 0.0004723878552602083, + "loss": 1.9097, + "step": 1884 + }, + { + "epoch": 0.18408203125, + "grad_norm": 0.2215142548084259, + "learning_rate": 0.000472353712810069, + "loss": 1.8851, + "step": 1885 + }, + { + "epoch": 0.1841796875, + "grad_norm": 0.22764411568641663, + "learning_rate": 0.0004723195506460422, + "loss": 1.9108, + "step": 1886 + }, + { + "epoch": 0.18427734375, + "grad_norm": 0.2349424958229065, + "learning_rate": 0.0004722853687715404, + "loss": 1.9079, + "step": 1887 + }, + { + "epoch": 0.184375, + "grad_norm": 0.23532803356647491, + "learning_rate": 0.00047225116718997804, + "loss": 1.9198, + "step": 1888 + }, + { + "epoch": 0.18447265625, + "grad_norm": 0.2555566728115082, + "learning_rate": 0.00047221694590477147, + "loss": 1.8961, + "step": 1889 + }, + { + "epoch": 0.1845703125, + "grad_norm": 0.30815714597702026, + "learning_rate": 0.00047218270491933934, + "loss": 1.8792, + "step": 1890 + }, + { + "epoch": 0.18466796875, + "grad_norm": 0.21387292444705963, + "learning_rate": 0.00047214844423710173, + "loss": 1.918, + "step": 1891 + }, + { + "epoch": 0.184765625, + "grad_norm": 0.24553145468235016, + "learning_rate": 0.00047211416386148123, + "loss": 1.8908, + "step": 1892 + }, + { + "epoch": 0.18486328125, + "grad_norm": 0.3307762145996094, + "learning_rate": 0.00047207986379590197, + "loss": 1.9298, + "step": 1893 + }, + { + "epoch": 0.1849609375, + "grad_norm": 0.22250519692897797, + "learning_rate": 0.00047204554404379036, + "loss": 1.886, + "step": 1894 + }, + { + "epoch": 0.18505859375, + "grad_norm": 0.2863367199897766, + "learning_rate": 0.0004720112046085745, + "loss": 1.8674, + "step": 1895 + }, + { + "epoch": 0.18515625, + "grad_norm": 0.34826767444610596, + "learning_rate": 0.00047197684549368465, + "loss": 1.9282, + "step": 1896 + }, + { + "epoch": 0.18525390625, + "grad_norm": 0.289986252784729, + "learning_rate": 0.00047194246670255295, + "loss": 1.9078, + "step": 1897 + }, + { + "epoch": 0.1853515625, + "grad_norm": 0.29067522287368774, + "learning_rate": 0.0004719080682386137, + "loss": 1.8943, + "step": 1898 + }, + { + "epoch": 0.18544921875, + "grad_norm": 0.2805711627006531, + "learning_rate": 0.0004718736501053028, + "loss": 1.8729, + "step": 1899 + }, + { + "epoch": 0.185546875, + "grad_norm": 0.1988450288772583, + "learning_rate": 0.0004718392123060584, + "loss": 1.897, + "step": 1900 + }, + { + "epoch": 0.18564453125, + "grad_norm": 0.28334251046180725, + "learning_rate": 0.0004718047548443206, + "loss": 1.8796, + "step": 1901 + }, + { + "epoch": 0.1857421875, + "grad_norm": 0.24370847642421722, + "learning_rate": 0.00047177027772353134, + "loss": 1.9136, + "step": 1902 + }, + { + "epoch": 0.18583984375, + "grad_norm": 0.25666430592536926, + "learning_rate": 0.0004717357809471344, + "loss": 1.9396, + "step": 1903 + }, + { + "epoch": 0.1859375, + "grad_norm": 0.3168308138847351, + "learning_rate": 0.00047170126451857603, + "loss": 1.8627, + "step": 1904 + }, + { + "epoch": 0.18603515625, + "grad_norm": 0.3242661952972412, + "learning_rate": 0.0004716667284413039, + "loss": 1.9534, + "step": 1905 + }, + { + "epoch": 0.1861328125, + "grad_norm": 0.23754087090492249, + "learning_rate": 0.0004716321727187678, + "loss": 1.935, + "step": 1906 + }, + { + "epoch": 0.18623046875, + "grad_norm": 0.2671366035938263, + "learning_rate": 0.0004715975973544198, + "loss": 1.9543, + "step": 1907 + }, + { + "epoch": 0.186328125, + "grad_norm": 0.2637098729610443, + "learning_rate": 0.00047156300235171353, + "loss": 1.8575, + "step": 1908 + }, + { + "epoch": 0.18642578125, + "grad_norm": 0.22229737043380737, + "learning_rate": 0.0004715283877141046, + "loss": 1.8757, + "step": 1909 + }, + { + "epoch": 0.1865234375, + "grad_norm": 0.2838973104953766, + "learning_rate": 0.00047149375344505084, + "loss": 1.9114, + "step": 1910 + }, + { + "epoch": 0.18662109375, + "grad_norm": 0.30752044916152954, + "learning_rate": 0.00047145909954801195, + "loss": 1.8879, + "step": 1911 + }, + { + "epoch": 0.18671875, + "grad_norm": 0.2020583301782608, + "learning_rate": 0.0004714244260264495, + "loss": 1.8957, + "step": 1912 + }, + { + "epoch": 0.18681640625, + "grad_norm": 0.25839710235595703, + "learning_rate": 0.000471389732883827, + "loss": 1.9356, + "step": 1913 + }, + { + "epoch": 0.1869140625, + "grad_norm": 0.2770155668258667, + "learning_rate": 0.0004713550201236101, + "loss": 1.9127, + "step": 1914 + }, + { + "epoch": 0.18701171875, + "grad_norm": 0.20789627730846405, + "learning_rate": 0.0004713202877492661, + "loss": 1.9254, + "step": 1915 + }, + { + "epoch": 0.187109375, + "grad_norm": 0.2471790611743927, + "learning_rate": 0.00047128553576426477, + "loss": 1.958, + "step": 1916 + }, + { + "epoch": 0.18720703125, + "grad_norm": 0.2624439597129822, + "learning_rate": 0.0004712507641720772, + "loss": 1.9403, + "step": 1917 + }, + { + "epoch": 0.1873046875, + "grad_norm": 0.2492697685956955, + "learning_rate": 0.00047121597297617704, + "loss": 1.8928, + "step": 1918 + }, + { + "epoch": 0.18740234375, + "grad_norm": 0.20261016488075256, + "learning_rate": 0.0004711811621800394, + "loss": 1.925, + "step": 1919 + }, + { + "epoch": 0.1875, + "grad_norm": 0.20547007024288177, + "learning_rate": 0.0004711463317871417, + "loss": 1.9058, + "step": 1920 + }, + { + "epoch": 0.18759765625, + "grad_norm": 0.24086418747901917, + "learning_rate": 0.0004711114818009632, + "loss": 1.9377, + "step": 1921 + }, + { + "epoch": 0.1876953125, + "grad_norm": 0.28701573610305786, + "learning_rate": 0.00047107661222498497, + "loss": 1.8958, + "step": 1922 + }, + { + "epoch": 0.18779296875, + "grad_norm": 0.2246810346841812, + "learning_rate": 0.0004710417230626904, + "loss": 1.8759, + "step": 1923 + }, + { + "epoch": 0.187890625, + "grad_norm": 0.1889760047197342, + "learning_rate": 0.00047100681431756433, + "loss": 1.947, + "step": 1924 + }, + { + "epoch": 0.18798828125, + "grad_norm": 0.22808609902858734, + "learning_rate": 0.0004709718859930941, + "loss": 1.9428, + "step": 1925 + }, + { + "epoch": 0.1880859375, + "grad_norm": 0.2610793709754944, + "learning_rate": 0.00047093693809276843, + "loss": 1.964, + "step": 1926 + }, + { + "epoch": 0.18818359375, + "grad_norm": 0.30671024322509766, + "learning_rate": 0.00047090197062007864, + "loss": 1.9101, + "step": 1927 + }, + { + "epoch": 0.18828125, + "grad_norm": 0.27946990728378296, + "learning_rate": 0.0004708669835785175, + "loss": 1.9487, + "step": 1928 + }, + { + "epoch": 0.18837890625, + "grad_norm": 0.2439918965101242, + "learning_rate": 0.0004708319769715799, + "loss": 1.9037, + "step": 1929 + }, + { + "epoch": 0.1884765625, + "grad_norm": 0.2517712414264679, + "learning_rate": 0.0004707969508027627, + "loss": 1.9059, + "step": 1930 + }, + { + "epoch": 0.18857421875, + "grad_norm": 0.2609822750091553, + "learning_rate": 0.0004707619050755648, + "loss": 1.9119, + "step": 1931 + }, + { + "epoch": 0.188671875, + "grad_norm": 0.2705680727958679, + "learning_rate": 0.0004707268397934868, + "loss": 1.9264, + "step": 1932 + }, + { + "epoch": 0.18876953125, + "grad_norm": 0.27703019976615906, + "learning_rate": 0.00047069175496003147, + "loss": 1.8859, + "step": 1933 + }, + { + "epoch": 0.1888671875, + "grad_norm": 0.4434245228767395, + "learning_rate": 0.00047065665057870355, + "loss": 2.0193, + "step": 1934 + }, + { + "epoch": 0.18896484375, + "grad_norm": 0.25824934244155884, + "learning_rate": 0.0004706215266530096, + "loss": 1.9491, + "step": 1935 + }, + { + "epoch": 0.1890625, + "grad_norm": 0.3010888397693634, + "learning_rate": 0.00047058638318645815, + "loss": 1.9461, + "step": 1936 + }, + { + "epoch": 0.18916015625, + "grad_norm": 0.2592420279979706, + "learning_rate": 0.0004705512201825597, + "loss": 1.9329, + "step": 1937 + }, + { + "epoch": 0.1892578125, + "grad_norm": 0.2857649326324463, + "learning_rate": 0.0004705160376448269, + "loss": 1.913, + "step": 1938 + }, + { + "epoch": 0.18935546875, + "grad_norm": 0.3039281964302063, + "learning_rate": 0.000470480835576774, + "loss": 1.9253, + "step": 1939 + }, + { + "epoch": 0.189453125, + "grad_norm": 0.23744364082813263, + "learning_rate": 0.00047044561398191744, + "loss": 1.9015, + "step": 1940 + }, + { + "epoch": 0.18955078125, + "grad_norm": 0.24613125622272491, + "learning_rate": 0.0004704103728637756, + "loss": 1.9214, + "step": 1941 + }, + { + "epoch": 0.1896484375, + "grad_norm": 0.2505404055118561, + "learning_rate": 0.0004703751122258686, + "loss": 1.9532, + "step": 1942 + }, + { + "epoch": 0.18974609375, + "grad_norm": 0.2505582571029663, + "learning_rate": 0.0004703398320717188, + "loss": 1.932, + "step": 1943 + }, + { + "epoch": 0.18984375, + "grad_norm": 0.2656853199005127, + "learning_rate": 0.00047030453240485035, + "loss": 1.8801, + "step": 1944 + }, + { + "epoch": 0.18994140625, + "grad_norm": 0.2200378179550171, + "learning_rate": 0.0004702692132287892, + "loss": 1.9601, + "step": 1945 + }, + { + "epoch": 0.1900390625, + "grad_norm": 0.260431170463562, + "learning_rate": 0.00047023387454706375, + "loss": 1.9112, + "step": 1946 + }, + { + "epoch": 0.19013671875, + "grad_norm": 0.2725105881690979, + "learning_rate": 0.00047019851636320384, + "loss": 1.9616, + "step": 1947 + }, + { + "epoch": 0.190234375, + "grad_norm": 0.21269826591014862, + "learning_rate": 0.0004701631386807414, + "loss": 1.9063, + "step": 1948 + }, + { + "epoch": 0.19033203125, + "grad_norm": 0.2665625512599945, + "learning_rate": 0.0004701277415032104, + "loss": 1.9361, + "step": 1949 + }, + { + "epoch": 0.1904296875, + "grad_norm": 0.2765948176383972, + "learning_rate": 0.0004700923248341467, + "loss": 1.9528, + "step": 1950 + }, + { + "epoch": 0.19052734375, + "grad_norm": 0.22930294275283813, + "learning_rate": 0.00047005688867708814, + "loss": 1.8957, + "step": 1951 + }, + { + "epoch": 0.190625, + "grad_norm": 0.22256456315517426, + "learning_rate": 0.00047002143303557435, + "loss": 1.9489, + "step": 1952 + }, + { + "epoch": 0.19072265625, + "grad_norm": 0.24130624532699585, + "learning_rate": 0.00046998595791314726, + "loss": 1.9574, + "step": 1953 + }, + { + "epoch": 0.1908203125, + "grad_norm": 0.292375773191452, + "learning_rate": 0.00046995046331335026, + "loss": 1.9176, + "step": 1954 + }, + { + "epoch": 0.19091796875, + "grad_norm": 0.3528386950492859, + "learning_rate": 0.0004699149492397292, + "loss": 1.9511, + "step": 1955 + }, + { + "epoch": 0.191015625, + "grad_norm": 0.26164865493774414, + "learning_rate": 0.00046987941569583147, + "loss": 1.9094, + "step": 1956 + }, + { + "epoch": 0.19111328125, + "grad_norm": 0.27652114629745483, + "learning_rate": 0.0004698438626852066, + "loss": 1.9011, + "step": 1957 + }, + { + "epoch": 0.1912109375, + "grad_norm": 0.31081855297088623, + "learning_rate": 0.00046980829021140596, + "loss": 1.9041, + "step": 1958 + }, + { + "epoch": 0.19130859375, + "grad_norm": 0.29934486746788025, + "learning_rate": 0.00046977269827798304, + "loss": 1.9301, + "step": 1959 + }, + { + "epoch": 0.19140625, + "grad_norm": 0.31369319558143616, + "learning_rate": 0.000469737086888493, + "loss": 1.9194, + "step": 1960 + }, + { + "epoch": 0.19150390625, + "grad_norm": 0.21880319714546204, + "learning_rate": 0.00046970145604649326, + "loss": 1.9029, + "step": 1961 + }, + { + "epoch": 0.1916015625, + "grad_norm": 0.21874625980854034, + "learning_rate": 0.00046966580575554304, + "loss": 1.9073, + "step": 1962 + }, + { + "epoch": 0.19169921875, + "grad_norm": 0.27970215678215027, + "learning_rate": 0.0004696301360192033, + "loss": 1.9589, + "step": 1963 + }, + { + "epoch": 0.191796875, + "grad_norm": 0.22823897004127502, + "learning_rate": 0.00046959444684103725, + "loss": 1.8801, + "step": 1964 + }, + { + "epoch": 0.19189453125, + "grad_norm": 0.22956214845180511, + "learning_rate": 0.0004695587382246099, + "loss": 1.908, + "step": 1965 + }, + { + "epoch": 0.1919921875, + "grad_norm": 0.2367463856935501, + "learning_rate": 0.00046952301017348826, + "loss": 1.9049, + "step": 1966 + }, + { + "epoch": 0.19208984375, + "grad_norm": 0.2357739508152008, + "learning_rate": 0.0004694872626912412, + "loss": 1.9173, + "step": 1967 + }, + { + "epoch": 0.1921875, + "grad_norm": 0.25406715273857117, + "learning_rate": 0.0004694514957814396, + "loss": 1.9309, + "step": 1968 + }, + { + "epoch": 0.19228515625, + "grad_norm": 0.2415982335805893, + "learning_rate": 0.0004694157094476563, + "loss": 1.8976, + "step": 1969 + }, + { + "epoch": 0.1923828125, + "grad_norm": 0.2648661434650421, + "learning_rate": 0.000469379903693466, + "loss": 1.9001, + "step": 1970 + }, + { + "epoch": 0.19248046875, + "grad_norm": 0.186269611120224, + "learning_rate": 0.00046934407852244534, + "loss": 1.8917, + "step": 1971 + }, + { + "epoch": 0.192578125, + "grad_norm": 0.30947110056877136, + "learning_rate": 0.00046930823393817296, + "loss": 1.8892, + "step": 1972 + }, + { + "epoch": 0.19267578125, + "grad_norm": 0.2893858850002289, + "learning_rate": 0.00046927236994422945, + "loss": 1.9225, + "step": 1973 + }, + { + "epoch": 0.1927734375, + "grad_norm": 0.24984197318553925, + "learning_rate": 0.0004692364865441973, + "loss": 1.9367, + "step": 1974 + }, + { + "epoch": 0.19287109375, + "grad_norm": 0.27524879574775696, + "learning_rate": 0.0004692005837416609, + "loss": 1.9141, + "step": 1975 + }, + { + "epoch": 0.19296875, + "grad_norm": 0.22581790387630463, + "learning_rate": 0.00046916466154020656, + "loss": 1.9093, + "step": 1976 + }, + { + "epoch": 0.19306640625, + "grad_norm": 0.2497783899307251, + "learning_rate": 0.0004691287199434228, + "loss": 1.9136, + "step": 1977 + }, + { + "epoch": 0.1931640625, + "grad_norm": 0.22558996081352234, + "learning_rate": 0.0004690927589548997, + "loss": 1.9372, + "step": 1978 + }, + { + "epoch": 0.19326171875, + "grad_norm": 0.2091677039861679, + "learning_rate": 0.0004690567785782295, + "loss": 1.9425, + "step": 1979 + }, + { + "epoch": 0.193359375, + "grad_norm": 0.26007091999053955, + "learning_rate": 0.0004690207788170063, + "loss": 1.8893, + "step": 1980 + }, + { + "epoch": 0.19345703125, + "grad_norm": 0.2551250755786896, + "learning_rate": 0.0004689847596748261, + "loss": 1.924, + "step": 1981 + }, + { + "epoch": 0.1935546875, + "grad_norm": 0.30021947622299194, + "learning_rate": 0.000468948721155287, + "loss": 1.9103, + "step": 1982 + }, + { + "epoch": 0.19365234375, + "grad_norm": 0.2922191917896271, + "learning_rate": 0.0004689126632619889, + "loss": 1.9284, + "step": 1983 + }, + { + "epoch": 0.19375, + "grad_norm": 0.24520905315876007, + "learning_rate": 0.00046887658599853373, + "loss": 1.8806, + "step": 1984 + }, + { + "epoch": 0.19384765625, + "grad_norm": 0.27852240204811096, + "learning_rate": 0.00046884048936852513, + "loss": 1.9375, + "step": 1985 + }, + { + "epoch": 0.1939453125, + "grad_norm": 0.26879197359085083, + "learning_rate": 0.0004688043733755689, + "loss": 1.9054, + "step": 1986 + }, + { + "epoch": 0.19404296875, + "grad_norm": 0.2576071321964264, + "learning_rate": 0.00046876823802327274, + "loss": 1.9758, + "step": 1987 + }, + { + "epoch": 0.194140625, + "grad_norm": 0.24778784811496735, + "learning_rate": 0.0004687320833152463, + "loss": 1.8767, + "step": 1988 + }, + { + "epoch": 0.19423828125, + "grad_norm": 0.2549983859062195, + "learning_rate": 0.000468695909255101, + "loss": 1.9532, + "step": 1989 + }, + { + "epoch": 0.1943359375, + "grad_norm": 0.24673303961753845, + "learning_rate": 0.00046865971584645027, + "loss": 1.9304, + "step": 1990 + }, + { + "epoch": 0.19443359375, + "grad_norm": 0.23759478330612183, + "learning_rate": 0.00046862350309290977, + "loss": 1.8887, + "step": 1991 + }, + { + "epoch": 0.19453125, + "grad_norm": 0.28576332330703735, + "learning_rate": 0.00046858727099809643, + "loss": 1.8766, + "step": 1992 + }, + { + "epoch": 0.19462890625, + "grad_norm": 0.23105448484420776, + "learning_rate": 0.00046855101956563, + "loss": 1.8956, + "step": 1993 + }, + { + "epoch": 0.1947265625, + "grad_norm": 0.27505502104759216, + "learning_rate": 0.0004685147487991312, + "loss": 1.8915, + "step": 1994 + }, + { + "epoch": 0.19482421875, + "grad_norm": 0.2024049013853073, + "learning_rate": 0.00046847845870222354, + "loss": 1.8735, + "step": 1995 + }, + { + "epoch": 0.194921875, + "grad_norm": 0.26097309589385986, + "learning_rate": 0.0004684421492785318, + "loss": 1.9298, + "step": 1996 + }, + { + "epoch": 0.19501953125, + "grad_norm": 0.2358173280954361, + "learning_rate": 0.0004684058205316832, + "loss": 1.9121, + "step": 1997 + }, + { + "epoch": 0.1951171875, + "grad_norm": 0.2725095748901367, + "learning_rate": 0.00046836947246530646, + "loss": 1.8684, + "step": 1998 + }, + { + "epoch": 0.19521484375, + "grad_norm": 0.25193265080451965, + "learning_rate": 0.0004683331050830326, + "loss": 1.9188, + "step": 1999 + }, + { + "epoch": 0.1953125, + "grad_norm": 0.23056617379188538, + "learning_rate": 0.00046829671838849424, + "loss": 1.8628, + "step": 2000 + }, + { + "epoch": 0.19541015625, + "grad_norm": 0.21980516612529755, + "learning_rate": 0.00046826031238532623, + "loss": 1.9392, + "step": 2001 + }, + { + "epoch": 0.1955078125, + "grad_norm": 0.244726300239563, + "learning_rate": 0.0004682238870771651, + "loss": 1.9079, + "step": 2002 + }, + { + "epoch": 0.19560546875, + "grad_norm": 0.21440590918064117, + "learning_rate": 0.00046818744246764955, + "loss": 1.9059, + "step": 2003 + }, + { + "epoch": 0.195703125, + "grad_norm": 0.25402185320854187, + "learning_rate": 0.0004681509785604199, + "loss": 1.9226, + "step": 2004 + }, + { + "epoch": 0.19580078125, + "grad_norm": 0.24775418639183044, + "learning_rate": 0.00046811449535911863, + "loss": 1.8361, + "step": 2005 + }, + { + "epoch": 0.1958984375, + "grad_norm": 0.2603761553764343, + "learning_rate": 0.0004680779928673902, + "loss": 1.9252, + "step": 2006 + }, + { + "epoch": 0.19599609375, + "grad_norm": 0.25276273488998413, + "learning_rate": 0.0004680414710888808, + "loss": 1.9072, + "step": 2007 + }, + { + "epoch": 0.19609375, + "grad_norm": 0.23971259593963623, + "learning_rate": 0.00046800493002723854, + "loss": 1.9073, + "step": 2008 + }, + { + "epoch": 0.19619140625, + "grad_norm": 0.21876166760921478, + "learning_rate": 0.00046796836968611373, + "loss": 1.8839, + "step": 2009 + }, + { + "epoch": 0.1962890625, + "grad_norm": 0.22767984867095947, + "learning_rate": 0.00046793179006915825, + "loss": 1.8852, + "step": 2010 + }, + { + "epoch": 0.19638671875, + "grad_norm": 0.1997627317905426, + "learning_rate": 0.0004678951911800262, + "loss": 1.8831, + "step": 2011 + }, + { + "epoch": 0.196484375, + "grad_norm": 0.23563456535339355, + "learning_rate": 0.00046785857302237345, + "loss": 1.8791, + "step": 2012 + }, + { + "epoch": 0.19658203125, + "grad_norm": 0.1879824995994568, + "learning_rate": 0.0004678219355998578, + "loss": 1.8986, + "step": 2013 + }, + { + "epoch": 0.1966796875, + "grad_norm": 0.26914268732070923, + "learning_rate": 0.00046778527891613907, + "loss": 1.9269, + "step": 2014 + }, + { + "epoch": 0.19677734375, + "grad_norm": 0.20528461039066315, + "learning_rate": 0.000467748602974879, + "loss": 1.9475, + "step": 2015 + }, + { + "epoch": 0.196875, + "grad_norm": 0.260937362909317, + "learning_rate": 0.0004677119077797409, + "loss": 1.9036, + "step": 2016 + }, + { + "epoch": 0.19697265625, + "grad_norm": 0.233351469039917, + "learning_rate": 0.0004676751933343906, + "loss": 1.9155, + "step": 2017 + }, + { + "epoch": 0.1970703125, + "grad_norm": 0.24918298423290253, + "learning_rate": 0.0004676384596424954, + "loss": 1.9169, + "step": 2018 + }, + { + "epoch": 0.19716796875, + "grad_norm": 0.331938236951828, + "learning_rate": 0.00046760170670772473, + "loss": 1.8655, + "step": 2019 + }, + { + "epoch": 0.197265625, + "grad_norm": 0.31642693281173706, + "learning_rate": 0.0004675649345337498, + "loss": 1.926, + "step": 2020 + }, + { + "epoch": 0.19736328125, + "grad_norm": 0.28673624992370605, + "learning_rate": 0.00046752814312424394, + "loss": 1.8992, + "step": 2021 + }, + { + "epoch": 0.1974609375, + "grad_norm": 0.27498120069503784, + "learning_rate": 0.0004674913324828822, + "loss": 1.9125, + "step": 2022 + }, + { + "epoch": 0.19755859375, + "grad_norm": 0.3634410500526428, + "learning_rate": 0.0004674545026133417, + "loss": 1.9216, + "step": 2023 + }, + { + "epoch": 0.19765625, + "grad_norm": 0.2195715606212616, + "learning_rate": 0.00046741765351930137, + "loss": 1.9071, + "step": 2024 + }, + { + "epoch": 0.19775390625, + "grad_norm": 0.30638429522514343, + "learning_rate": 0.0004673807852044421, + "loss": 1.9037, + "step": 2025 + }, + { + "epoch": 0.1978515625, + "grad_norm": 0.30112913250923157, + "learning_rate": 0.0004673438976724468, + "loss": 1.9117, + "step": 2026 + }, + { + "epoch": 0.19794921875, + "grad_norm": 0.2405427247285843, + "learning_rate": 0.00046730699092700003, + "loss": 1.8978, + "step": 2027 + }, + { + "epoch": 0.198046875, + "grad_norm": 0.38917067646980286, + "learning_rate": 0.0004672700649717886, + "loss": 1.9055, + "step": 2028 + }, + { + "epoch": 0.19814453125, + "grad_norm": 0.31037402153015137, + "learning_rate": 0.000467233119810501, + "loss": 1.8982, + "step": 2029 + }, + { + "epoch": 0.1982421875, + "grad_norm": 0.31107863783836365, + "learning_rate": 0.00046719615544682784, + "loss": 1.8812, + "step": 2030 + }, + { + "epoch": 0.19833984375, + "grad_norm": 0.3467962145805359, + "learning_rate": 0.00046715917188446134, + "loss": 1.8957, + "step": 2031 + }, + { + "epoch": 0.1984375, + "grad_norm": 0.25533223152160645, + "learning_rate": 0.0004671221691270961, + "loss": 1.9409, + "step": 2032 + }, + { + "epoch": 0.19853515625, + "grad_norm": 0.3249166011810303, + "learning_rate": 0.0004670851471784281, + "loss": 1.9037, + "step": 2033 + }, + { + "epoch": 0.1986328125, + "grad_norm": 0.2961365282535553, + "learning_rate": 0.00046704810604215565, + "loss": 1.8862, + "step": 2034 + }, + { + "epoch": 0.19873046875, + "grad_norm": 0.30742213129997253, + "learning_rate": 0.0004670110457219788, + "loss": 1.8787, + "step": 2035 + }, + { + "epoch": 0.198828125, + "grad_norm": 0.20632188022136688, + "learning_rate": 0.00046697396622159956, + "loss": 1.8716, + "step": 2036 + }, + { + "epoch": 0.19892578125, + "grad_norm": 0.3098668158054352, + "learning_rate": 0.0004669368675447218, + "loss": 1.9131, + "step": 2037 + }, + { + "epoch": 0.1990234375, + "grad_norm": 0.210395947098732, + "learning_rate": 0.00046689974969505143, + "loss": 1.8658, + "step": 2038 + }, + { + "epoch": 0.19912109375, + "grad_norm": 0.2865216135978699, + "learning_rate": 0.00046686261267629605, + "loss": 1.9041, + "step": 2039 + }, + { + "epoch": 0.19921875, + "grad_norm": 0.2984568774700165, + "learning_rate": 0.00046682545649216544, + "loss": 1.9122, + "step": 2040 + }, + { + "epoch": 0.19931640625, + "grad_norm": 0.285224586725235, + "learning_rate": 0.00046678828114637126, + "loss": 1.9139, + "step": 2041 + }, + { + "epoch": 0.1994140625, + "grad_norm": 0.23279589414596558, + "learning_rate": 0.0004667510866426268, + "loss": 1.8941, + "step": 2042 + }, + { + "epoch": 0.19951171875, + "grad_norm": 0.20549869537353516, + "learning_rate": 0.0004667138729846475, + "loss": 1.8947, + "step": 2043 + }, + { + "epoch": 0.199609375, + "grad_norm": 0.26036345958709717, + "learning_rate": 0.0004666766401761508, + "loss": 1.9111, + "step": 2044 + }, + { + "epoch": 0.19970703125, + "grad_norm": 0.2697281837463379, + "learning_rate": 0.00046663938822085583, + "loss": 1.909, + "step": 2045 + }, + { + "epoch": 0.1998046875, + "grad_norm": 0.23096928000450134, + "learning_rate": 0.00046660211712248374, + "loss": 1.9102, + "step": 2046 + }, + { + "epoch": 0.19990234375, + "grad_norm": 0.21505551040172577, + "learning_rate": 0.00046656482688475754, + "loss": 1.9257, + "step": 2047 + }, + { + "epoch": 0.2, + "grad_norm": 0.2264963835477829, + "learning_rate": 0.0004665275175114024, + "loss": 1.9256, + "step": 2048 + }, + { + "epoch": 0.20009765625, + "grad_norm": 0.29406315088272095, + "learning_rate": 0.0004664901890061449, + "loss": 1.9512, + "step": 2049 + }, + { + "epoch": 0.2001953125, + "grad_norm": 0.2077530175447464, + "learning_rate": 0.0004664528413727139, + "loss": 1.9162, + "step": 2050 + }, + { + "epoch": 0.20029296875, + "grad_norm": 0.23434431850910187, + "learning_rate": 0.00046641547461484035, + "loss": 1.8639, + "step": 2051 + }, + { + "epoch": 0.200390625, + "grad_norm": 0.3102989196777344, + "learning_rate": 0.0004663780887362566, + "loss": 1.8729, + "step": 2052 + }, + { + "epoch": 0.20048828125, + "grad_norm": 0.2750473618507385, + "learning_rate": 0.0004663406837406973, + "loss": 1.8936, + "step": 2053 + }, + { + "epoch": 0.2005859375, + "grad_norm": 0.24825125932693481, + "learning_rate": 0.00046630325963189866, + "loss": 1.8843, + "step": 2054 + }, + { + "epoch": 0.20068359375, + "grad_norm": 0.2034183144569397, + "learning_rate": 0.0004662658164135993, + "loss": 1.9049, + "step": 2055 + }, + { + "epoch": 0.20078125, + "grad_norm": 0.26106199622154236, + "learning_rate": 0.0004662283540895394, + "loss": 1.9034, + "step": 2056 + }, + { + "epoch": 0.20087890625, + "grad_norm": 0.22128944098949432, + "learning_rate": 0.0004661908726634609, + "loss": 1.8846, + "step": 2057 + }, + { + "epoch": 0.2009765625, + "grad_norm": 0.19975769519805908, + "learning_rate": 0.0004661533721391081, + "loss": 1.9336, + "step": 2058 + }, + { + "epoch": 0.20107421875, + "grad_norm": 0.21513822674751282, + "learning_rate": 0.0004661158525202269, + "loss": 1.8887, + "step": 2059 + }, + { + "epoch": 0.201171875, + "grad_norm": 0.2806141972541809, + "learning_rate": 0.00046607831381056513, + "loss": 1.8993, + "step": 2060 + }, + { + "epoch": 0.20126953125, + "grad_norm": 0.24645020067691803, + "learning_rate": 0.00046604075601387263, + "loss": 1.9206, + "step": 2061 + }, + { + "epoch": 0.2013671875, + "grad_norm": 0.19376933574676514, + "learning_rate": 0.0004660031791339011, + "loss": 1.9024, + "step": 2062 + }, + { + "epoch": 0.20146484375, + "grad_norm": 0.3149716556072235, + "learning_rate": 0.00046596558317440415, + "loss": 1.9294, + "step": 2063 + }, + { + "epoch": 0.2015625, + "grad_norm": 0.2759716808795929, + "learning_rate": 0.00046592796813913715, + "loss": 1.8748, + "step": 2064 + }, + { + "epoch": 0.20166015625, + "grad_norm": 0.2037348598241806, + "learning_rate": 0.0004658903340318577, + "loss": 1.8803, + "step": 2065 + }, + { + "epoch": 0.2017578125, + "grad_norm": 0.21185782551765442, + "learning_rate": 0.0004658526808563248, + "loss": 1.9478, + "step": 2066 + }, + { + "epoch": 0.20185546875, + "grad_norm": 0.2070467174053192, + "learning_rate": 0.0004658150086163001, + "loss": 1.9131, + "step": 2067 + }, + { + "epoch": 0.201953125, + "grad_norm": 0.22816209495067596, + "learning_rate": 0.00046577731731554646, + "loss": 1.9053, + "step": 2068 + }, + { + "epoch": 0.20205078125, + "grad_norm": 0.2772216796875, + "learning_rate": 0.0004657396069578289, + "loss": 1.9048, + "step": 2069 + }, + { + "epoch": 0.2021484375, + "grad_norm": 0.27881917357444763, + "learning_rate": 0.0004657018775469145, + "loss": 1.9161, + "step": 2070 + }, + { + "epoch": 0.20224609375, + "grad_norm": 0.2583853602409363, + "learning_rate": 0.0004656641290865719, + "loss": 1.8961, + "step": 2071 + }, + { + "epoch": 0.20234375, + "grad_norm": 0.2798794209957123, + "learning_rate": 0.00046562636158057203, + "loss": 1.8912, + "step": 2072 + }, + { + "epoch": 0.20244140625, + "grad_norm": 0.25222232937812805, + "learning_rate": 0.0004655885750326874, + "loss": 1.9522, + "step": 2073 + }, + { + "epoch": 0.2025390625, + "grad_norm": 0.2819882929325104, + "learning_rate": 0.0004655507694466925, + "loss": 1.8891, + "step": 2074 + }, + { + "epoch": 0.20263671875, + "grad_norm": 0.21219240128993988, + "learning_rate": 0.0004655129448263639, + "loss": 1.9383, + "step": 2075 + }, + { + "epoch": 0.202734375, + "grad_norm": 0.2394660860300064, + "learning_rate": 0.00046547510117547997, + "loss": 1.9034, + "step": 2076 + }, + { + "epoch": 0.20283203125, + "grad_norm": 0.22925956547260284, + "learning_rate": 0.00046543723849782085, + "loss": 1.9189, + "step": 2077 + }, + { + "epoch": 0.2029296875, + "grad_norm": 0.27168065309524536, + "learning_rate": 0.00046539935679716876, + "loss": 1.8803, + "step": 2078 + }, + { + "epoch": 0.20302734375, + "grad_norm": 0.3125360906124115, + "learning_rate": 0.00046536145607730767, + "loss": 1.9144, + "step": 2079 + }, + { + "epoch": 0.203125, + "grad_norm": 0.2742547392845154, + "learning_rate": 0.00046532353634202355, + "loss": 1.9011, + "step": 2080 + }, + { + "epoch": 0.20322265625, + "grad_norm": 0.30199679732322693, + "learning_rate": 0.0004652855975951044, + "loss": 1.8952, + "step": 2081 + }, + { + "epoch": 0.2033203125, + "grad_norm": 0.2881486117839813, + "learning_rate": 0.0004652476398403397, + "loss": 1.8888, + "step": 2082 + }, + { + "epoch": 0.20341796875, + "grad_norm": 0.30965685844421387, + "learning_rate": 0.0004652096630815212, + "loss": 1.9286, + "step": 2083 + }, + { + "epoch": 0.203515625, + "grad_norm": 0.38387784361839294, + "learning_rate": 0.00046517166732244254, + "loss": 1.922, + "step": 2084 + }, + { + "epoch": 0.20361328125, + "grad_norm": 0.35733339190483093, + "learning_rate": 0.00046513365256689907, + "loss": 1.9086, + "step": 2085 + }, + { + "epoch": 0.2037109375, + "grad_norm": 0.28068143129348755, + "learning_rate": 0.00046509561881868814, + "loss": 1.8819, + "step": 2086 + }, + { + "epoch": 0.20380859375, + "grad_norm": 0.2930566668510437, + "learning_rate": 0.00046505756608160895, + "loss": 1.8737, + "step": 2087 + }, + { + "epoch": 0.20390625, + "grad_norm": 0.2810053527355194, + "learning_rate": 0.0004650194943594627, + "loss": 1.9112, + "step": 2088 + }, + { + "epoch": 0.20400390625, + "grad_norm": 0.2667030394077301, + "learning_rate": 0.00046498140365605243, + "loss": 1.9051, + "step": 2089 + }, + { + "epoch": 0.2041015625, + "grad_norm": 0.20515574514865875, + "learning_rate": 0.00046494329397518284, + "loss": 1.9115, + "step": 2090 + }, + { + "epoch": 0.20419921875, + "grad_norm": 0.24670284986495972, + "learning_rate": 0.00046490516532066107, + "loss": 1.885, + "step": 2091 + }, + { + "epoch": 0.204296875, + "grad_norm": 0.2236049324274063, + "learning_rate": 0.0004648670176962957, + "loss": 1.9239, + "step": 2092 + }, + { + "epoch": 0.20439453125, + "grad_norm": 0.20630759000778198, + "learning_rate": 0.00046482885110589727, + "loss": 1.9762, + "step": 2093 + }, + { + "epoch": 0.2044921875, + "grad_norm": 0.25111427903175354, + "learning_rate": 0.0004647906655532784, + "loss": 1.9006, + "step": 2094 + }, + { + "epoch": 0.20458984375, + "grad_norm": 0.2501910626888275, + "learning_rate": 0.0004647524610422535, + "loss": 1.9586, + "step": 2095 + }, + { + "epoch": 0.2046875, + "grad_norm": 0.27904069423675537, + "learning_rate": 0.00046471423757663867, + "loss": 1.8846, + "step": 2096 + }, + { + "epoch": 0.20478515625, + "grad_norm": 0.24683643877506256, + "learning_rate": 0.0004646759951602523, + "loss": 1.8941, + "step": 2097 + }, + { + "epoch": 0.2048828125, + "grad_norm": 0.21275731921195984, + "learning_rate": 0.00046463773379691424, + "loss": 1.9168, + "step": 2098 + }, + { + "epoch": 0.20498046875, + "grad_norm": 0.2217152714729309, + "learning_rate": 0.0004645994534904468, + "loss": 1.8554, + "step": 2099 + }, + { + "epoch": 0.205078125, + "grad_norm": 0.20951475203037262, + "learning_rate": 0.0004645611542446736, + "loss": 1.9226, + "step": 2100 + }, + { + "epoch": 0.20517578125, + "grad_norm": 0.1962258368730545, + "learning_rate": 0.00046452283606342046, + "loss": 1.9244, + "step": 2101 + }, + { + "epoch": 0.2052734375, + "grad_norm": 0.25039345026016235, + "learning_rate": 0.000464484498950515, + "loss": 1.9237, + "step": 2102 + }, + { + "epoch": 0.20537109375, + "grad_norm": 0.2845058739185333, + "learning_rate": 0.00046444614290978674, + "loss": 1.8962, + "step": 2103 + }, + { + "epoch": 0.20546875, + "grad_norm": 0.22429201006889343, + "learning_rate": 0.0004644077679450672, + "loss": 1.9632, + "step": 2104 + }, + { + "epoch": 0.20556640625, + "grad_norm": 0.20088529586791992, + "learning_rate": 0.00046436937406018963, + "loss": 1.9211, + "step": 2105 + }, + { + "epoch": 0.2056640625, + "grad_norm": 0.2091323882341385, + "learning_rate": 0.0004643309612589893, + "loss": 1.913, + "step": 2106 + }, + { + "epoch": 0.20576171875, + "grad_norm": 0.24969804286956787, + "learning_rate": 0.00046429252954530316, + "loss": 1.8726, + "step": 2107 + }, + { + "epoch": 0.205859375, + "grad_norm": 0.30802738666534424, + "learning_rate": 0.00046425407892297025, + "loss": 1.9256, + "step": 2108 + }, + { + "epoch": 0.20595703125, + "grad_norm": 0.27234166860580444, + "learning_rate": 0.0004642156093958317, + "loss": 1.9548, + "step": 2109 + }, + { + "epoch": 0.2060546875, + "grad_norm": 0.19717903435230255, + "learning_rate": 0.00046417712096772994, + "loss": 1.9694, + "step": 2110 + }, + { + "epoch": 0.20615234375, + "grad_norm": 0.3146752715110779, + "learning_rate": 0.0004641386136425098, + "loss": 1.922, + "step": 2111 + }, + { + "epoch": 0.20625, + "grad_norm": 0.29423826932907104, + "learning_rate": 0.0004641000874240178, + "loss": 1.8961, + "step": 2112 + }, + { + "epoch": 0.20634765625, + "grad_norm": 0.23182527720928192, + "learning_rate": 0.0004640615423161022, + "loss": 1.9062, + "step": 2113 + }, + { + "epoch": 0.2064453125, + "grad_norm": 0.3178667426109314, + "learning_rate": 0.00046402297832261354, + "loss": 1.9098, + "step": 2114 + }, + { + "epoch": 0.20654296875, + "grad_norm": 0.3038199841976166, + "learning_rate": 0.000463984395447404, + "loss": 1.8738, + "step": 2115 + }, + { + "epoch": 0.206640625, + "grad_norm": 0.29474422335624695, + "learning_rate": 0.00046394579369432755, + "loss": 1.8506, + "step": 2116 + }, + { + "epoch": 0.20673828125, + "grad_norm": 0.30578935146331787, + "learning_rate": 0.00046390717306724023, + "loss": 1.8775, + "step": 2117 + }, + { + "epoch": 0.2068359375, + "grad_norm": 0.27422910928726196, + "learning_rate": 0.0004638685335699999, + "loss": 1.9101, + "step": 2118 + }, + { + "epoch": 0.20693359375, + "grad_norm": 0.2473389059305191, + "learning_rate": 0.00046382987520646624, + "loss": 1.9342, + "step": 2119 + }, + { + "epoch": 0.20703125, + "grad_norm": 0.2475956827402115, + "learning_rate": 0.00046379119798050094, + "loss": 1.9242, + "step": 2120 + }, + { + "epoch": 0.20712890625, + "grad_norm": 0.2836126983165741, + "learning_rate": 0.0004637525018959675, + "loss": 1.9263, + "step": 2121 + }, + { + "epoch": 0.2072265625, + "grad_norm": 0.20722109079360962, + "learning_rate": 0.00046371378695673137, + "loss": 1.9077, + "step": 2122 + }, + { + "epoch": 0.20732421875, + "grad_norm": 0.2630356550216675, + "learning_rate": 0.00046367505316665975, + "loss": 1.9109, + "step": 2123 + }, + { + "epoch": 0.207421875, + "grad_norm": 0.26636844873428345, + "learning_rate": 0.0004636363005296218, + "loss": 1.9011, + "step": 2124 + }, + { + "epoch": 0.20751953125, + "grad_norm": 0.3061290681362152, + "learning_rate": 0.0004635975290494887, + "loss": 1.8875, + "step": 2125 + }, + { + "epoch": 0.2076171875, + "grad_norm": 0.2455659955739975, + "learning_rate": 0.00046355873873013315, + "loss": 1.8745, + "step": 2126 + }, + { + "epoch": 0.20771484375, + "grad_norm": 0.226226806640625, + "learning_rate": 0.0004635199295754301, + "loss": 1.9186, + "step": 2127 + }, + { + "epoch": 0.2078125, + "grad_norm": 0.27755168080329895, + "learning_rate": 0.0004634811015892563, + "loss": 1.878, + "step": 2128 + }, + { + "epoch": 0.20791015625, + "grad_norm": 0.24024835228919983, + "learning_rate": 0.00046344225477549015, + "loss": 1.8886, + "step": 2129 + }, + { + "epoch": 0.2080078125, + "grad_norm": 0.2625919282436371, + "learning_rate": 0.0004634033891380122, + "loss": 1.8997, + "step": 2130 + }, + { + "epoch": 0.20810546875, + "grad_norm": 0.2913138270378113, + "learning_rate": 0.00046336450468070485, + "loss": 1.8711, + "step": 2131 + }, + { + "epoch": 0.208203125, + "grad_norm": 0.24820668995380402, + "learning_rate": 0.00046332560140745216, + "loss": 1.9243, + "step": 2132 + }, + { + "epoch": 0.20830078125, + "grad_norm": 0.2551608085632324, + "learning_rate": 0.00046328667932214035, + "loss": 1.8974, + "step": 2133 + }, + { + "epoch": 0.2083984375, + "grad_norm": 0.28112420439720154, + "learning_rate": 0.0004632477384286574, + "loss": 1.9122, + "step": 2134 + }, + { + "epoch": 0.20849609375, + "grad_norm": 0.23654119670391083, + "learning_rate": 0.0004632087787308931, + "loss": 1.93, + "step": 2135 + }, + { + "epoch": 0.20859375, + "grad_norm": 0.26119500398635864, + "learning_rate": 0.00046316980023273905, + "loss": 1.8936, + "step": 2136 + }, + { + "epoch": 0.20869140625, + "grad_norm": 0.26187509298324585, + "learning_rate": 0.0004631308029380891, + "loss": 1.9257, + "step": 2137 + }, + { + "epoch": 0.2087890625, + "grad_norm": 0.26531320810317993, + "learning_rate": 0.00046309178685083856, + "loss": 1.9203, + "step": 2138 + }, + { + "epoch": 0.20888671875, + "grad_norm": 0.28738391399383545, + "learning_rate": 0.00046305275197488496, + "loss": 1.9092, + "step": 2139 + }, + { + "epoch": 0.208984375, + "grad_norm": 0.2203925997018814, + "learning_rate": 0.0004630136983141273, + "loss": 1.9086, + "step": 2140 + }, + { + "epoch": 0.20908203125, + "grad_norm": 0.337566077709198, + "learning_rate": 0.0004629746258724669, + "loss": 1.896, + "step": 2141 + }, + { + "epoch": 0.2091796875, + "grad_norm": 0.22496585547924042, + "learning_rate": 0.0004629355346538067, + "loss": 1.9216, + "step": 2142 + }, + { + "epoch": 0.20927734375, + "grad_norm": 0.22534677386283875, + "learning_rate": 0.0004628964246620516, + "loss": 1.9074, + "step": 2143 + }, + { + "epoch": 0.209375, + "grad_norm": 0.24914197623729706, + "learning_rate": 0.0004628572959011082, + "loss": 1.9033, + "step": 2144 + }, + { + "epoch": 0.20947265625, + "grad_norm": 0.20345696806907654, + "learning_rate": 0.0004628181483748852, + "loss": 1.9218, + "step": 2145 + }, + { + "epoch": 0.2095703125, + "grad_norm": 0.22977322340011597, + "learning_rate": 0.00046277898208729306, + "loss": 1.8171, + "step": 2146 + }, + { + "epoch": 0.20966796875, + "grad_norm": 0.24085818231105804, + "learning_rate": 0.0004627397970422443, + "loss": 1.867, + "step": 2147 + }, + { + "epoch": 0.209765625, + "grad_norm": 0.26558682322502136, + "learning_rate": 0.00046270059324365296, + "loss": 1.9171, + "step": 2148 + }, + { + "epoch": 0.20986328125, + "grad_norm": 0.27644988894462585, + "learning_rate": 0.00046266137069543524, + "loss": 1.908, + "step": 2149 + }, + { + "epoch": 0.2099609375, + "grad_norm": 0.2609714865684509, + "learning_rate": 0.00046262212940150914, + "loss": 1.9649, + "step": 2150 + }, + { + "epoch": 0.21005859375, + "grad_norm": 0.27116766571998596, + "learning_rate": 0.00046258286936579453, + "loss": 1.9123, + "step": 2151 + }, + { + "epoch": 0.21015625, + "grad_norm": 0.21386384963989258, + "learning_rate": 0.0004625435905922131, + "loss": 1.9453, + "step": 2152 + }, + { + "epoch": 0.21025390625, + "grad_norm": 0.21030083298683167, + "learning_rate": 0.00046250429308468833, + "loss": 1.9349, + "step": 2153 + }, + { + "epoch": 0.2103515625, + "grad_norm": 0.267782062292099, + "learning_rate": 0.0004624649768471459, + "loss": 1.9266, + "step": 2154 + }, + { + "epoch": 0.21044921875, + "grad_norm": 0.2850502133369446, + "learning_rate": 0.0004624256418835131, + "loss": 1.884, + "step": 2155 + }, + { + "epoch": 0.210546875, + "grad_norm": 0.3452877700328827, + "learning_rate": 0.0004623862881977191, + "loss": 1.9651, + "step": 2156 + }, + { + "epoch": 0.21064453125, + "grad_norm": 0.2617965638637543, + "learning_rate": 0.00046234691579369504, + "loss": 1.8952, + "step": 2157 + }, + { + "epoch": 0.2107421875, + "grad_norm": 0.2741365134716034, + "learning_rate": 0.00046230752467537375, + "loss": 1.8951, + "step": 2158 + }, + { + "epoch": 0.21083984375, + "grad_norm": 0.37253716588020325, + "learning_rate": 0.00046226811484669015, + "loss": 1.9242, + "step": 2159 + }, + { + "epoch": 0.2109375, + "grad_norm": 0.28392404317855835, + "learning_rate": 0.00046222868631158105, + "loss": 1.9025, + "step": 2160 + }, + { + "epoch": 0.21103515625, + "grad_norm": 0.2524137794971466, + "learning_rate": 0.00046218923907398473, + "loss": 1.9166, + "step": 2161 + }, + { + "epoch": 0.2111328125, + "grad_norm": 0.2966980040073395, + "learning_rate": 0.0004621497731378419, + "loss": 1.9083, + "step": 2162 + }, + { + "epoch": 0.21123046875, + "grad_norm": 0.28292515873908997, + "learning_rate": 0.0004621102885070946, + "loss": 1.9049, + "step": 2163 + }, + { + "epoch": 0.211328125, + "grad_norm": 0.2480284720659256, + "learning_rate": 0.00046207078518568717, + "loss": 1.9063, + "step": 2164 + }, + { + "epoch": 0.21142578125, + "grad_norm": 0.2493746429681778, + "learning_rate": 0.0004620312631775657, + "loss": 1.8636, + "step": 2165 + }, + { + "epoch": 0.2115234375, + "grad_norm": 0.19225762784481049, + "learning_rate": 0.0004619917224866779, + "loss": 1.8916, + "step": 2166 + }, + { + "epoch": 0.21162109375, + "grad_norm": 0.24277755618095398, + "learning_rate": 0.00046195216311697355, + "loss": 1.9265, + "step": 2167 + }, + { + "epoch": 0.21171875, + "grad_norm": 0.2281288206577301, + "learning_rate": 0.0004619125850724045, + "loss": 1.9044, + "step": 2168 + }, + { + "epoch": 0.21181640625, + "grad_norm": 0.24480199813842773, + "learning_rate": 0.000461872988356924, + "loss": 1.8734, + "step": 2169 + }, + { + "epoch": 0.2119140625, + "grad_norm": 0.2577948570251465, + "learning_rate": 0.0004618333729744876, + "loss": 1.9017, + "step": 2170 + }, + { + "epoch": 0.21201171875, + "grad_norm": 0.22131475806236267, + "learning_rate": 0.00046179373892905233, + "loss": 1.9071, + "step": 2171 + }, + { + "epoch": 0.212109375, + "grad_norm": 0.24380597472190857, + "learning_rate": 0.00046175408622457743, + "loss": 1.9291, + "step": 2172 + }, + { + "epoch": 0.21220703125, + "grad_norm": 0.25301221013069153, + "learning_rate": 0.0004617144148650238, + "loss": 1.8888, + "step": 2173 + }, + { + "epoch": 0.2123046875, + "grad_norm": 0.2644136846065521, + "learning_rate": 0.00046167472485435424, + "loss": 1.8995, + "step": 2174 + }, + { + "epoch": 0.21240234375, + "grad_norm": 0.2486433982849121, + "learning_rate": 0.0004616350161965335, + "loss": 1.9229, + "step": 2175 + }, + { + "epoch": 0.2125, + "grad_norm": 0.3275326192378998, + "learning_rate": 0.0004615952888955281, + "loss": 1.9174, + "step": 2176 + }, + { + "epoch": 0.21259765625, + "grad_norm": 0.22458401322364807, + "learning_rate": 0.00046155554295530636, + "loss": 1.8847, + "step": 2177 + }, + { + "epoch": 0.2126953125, + "grad_norm": 0.2412051111459732, + "learning_rate": 0.0004615157783798387, + "loss": 1.9229, + "step": 2178 + }, + { + "epoch": 0.21279296875, + "grad_norm": 0.21972574293613434, + "learning_rate": 0.00046147599517309707, + "loss": 1.9243, + "step": 2179 + }, + { + "epoch": 0.212890625, + "grad_norm": 0.2897114157676697, + "learning_rate": 0.0004614361933390555, + "loss": 1.9192, + "step": 2180 + }, + { + "epoch": 0.21298828125, + "grad_norm": 0.2781403660774231, + "learning_rate": 0.00046139637288169003, + "loss": 1.8995, + "step": 2181 + }, + { + "epoch": 0.2130859375, + "grad_norm": 0.2788412272930145, + "learning_rate": 0.00046135653380497814, + "loss": 1.9041, + "step": 2182 + }, + { + "epoch": 0.21318359375, + "grad_norm": 0.28828147053718567, + "learning_rate": 0.0004613166761128996, + "loss": 1.9016, + "step": 2183 + }, + { + "epoch": 0.21328125, + "grad_norm": 0.26774659752845764, + "learning_rate": 0.00046127679980943563, + "loss": 1.9703, + "step": 2184 + }, + { + "epoch": 0.21337890625, + "grad_norm": 0.2517338991165161, + "learning_rate": 0.00046123690489856966, + "loss": 1.8761, + "step": 2185 + }, + { + "epoch": 0.2134765625, + "grad_norm": 0.25602057576179504, + "learning_rate": 0.00046119699138428676, + "loss": 1.9538, + "step": 2186 + }, + { + "epoch": 0.21357421875, + "grad_norm": 0.20147521793842316, + "learning_rate": 0.0004611570592705741, + "loss": 1.8697, + "step": 2187 + }, + { + "epoch": 0.213671875, + "grad_norm": 0.22156640887260437, + "learning_rate": 0.0004611171085614203, + "loss": 1.8982, + "step": 2188 + }, + { + "epoch": 0.21376953125, + "grad_norm": 0.23269224166870117, + "learning_rate": 0.00046107713926081626, + "loss": 1.8925, + "step": 2189 + }, + { + "epoch": 0.2138671875, + "grad_norm": 0.21519561111927032, + "learning_rate": 0.00046103715137275447, + "loss": 1.8476, + "step": 2190 + }, + { + "epoch": 0.21396484375, + "grad_norm": 0.2535529136657715, + "learning_rate": 0.0004609971449012294, + "loss": 1.901, + "step": 2191 + }, + { + "epoch": 0.2140625, + "grad_norm": 0.24704895913600922, + "learning_rate": 0.0004609571198502374, + "loss": 1.8946, + "step": 2192 + }, + { + "epoch": 0.21416015625, + "grad_norm": 0.24865074455738068, + "learning_rate": 0.00046091707622377646, + "loss": 1.8576, + "step": 2193 + }, + { + "epoch": 0.2142578125, + "grad_norm": 0.2663979232311249, + "learning_rate": 0.00046087701402584683, + "loss": 1.89, + "step": 2194 + }, + { + "epoch": 0.21435546875, + "grad_norm": 0.2134150117635727, + "learning_rate": 0.00046083693326045004, + "loss": 1.9338, + "step": 2195 + }, + { + "epoch": 0.214453125, + "grad_norm": 0.20297788083553314, + "learning_rate": 0.00046079683393159004, + "loss": 1.9193, + "step": 2196 + }, + { + "epoch": 0.21455078125, + "grad_norm": 0.22827771306037903, + "learning_rate": 0.0004607567160432724, + "loss": 1.9089, + "step": 2197 + }, + { + "epoch": 0.2146484375, + "grad_norm": 0.2234564572572708, + "learning_rate": 0.0004607165795995044, + "loss": 1.9069, + "step": 2198 + }, + { + "epoch": 0.21474609375, + "grad_norm": 0.23381231725215912, + "learning_rate": 0.0004606764246042955, + "loss": 1.9059, + "step": 2199 + }, + { + "epoch": 0.21484375, + "grad_norm": 0.24767504632472992, + "learning_rate": 0.0004606362510616567, + "loss": 1.9448, + "step": 2200 + }, + { + "epoch": 0.21494140625, + "grad_norm": 0.23719260096549988, + "learning_rate": 0.00046059605897560095, + "loss": 1.8839, + "step": 2201 + }, + { + "epoch": 0.2150390625, + "grad_norm": 0.35658907890319824, + "learning_rate": 0.00046055584835014325, + "loss": 1.912, + "step": 2202 + }, + { + "epoch": 0.21513671875, + "grad_norm": 0.35971537232398987, + "learning_rate": 0.0004605156191893001, + "loss": 1.9476, + "step": 2203 + }, + { + "epoch": 0.215234375, + "grad_norm": 0.242756187915802, + "learning_rate": 0.0004604753714970901, + "loss": 1.8945, + "step": 2204 + }, + { + "epoch": 0.21533203125, + "grad_norm": 0.299610435962677, + "learning_rate": 0.0004604351052775337, + "loss": 1.9209, + "step": 2205 + }, + { + "epoch": 0.2154296875, + "grad_norm": 0.21338649094104767, + "learning_rate": 0.000460394820534653, + "loss": 1.8975, + "step": 2206 + }, + { + "epoch": 0.21552734375, + "grad_norm": 0.2629339098930359, + "learning_rate": 0.00046035451727247226, + "loss": 1.8938, + "step": 2207 + }, + { + "epoch": 0.215625, + "grad_norm": 0.2574734091758728, + "learning_rate": 0.00046031419549501733, + "loss": 1.9167, + "step": 2208 + }, + { + "epoch": 0.21572265625, + "grad_norm": 0.20906251668930054, + "learning_rate": 0.000460273855206316, + "loss": 1.8635, + "step": 2209 + }, + { + "epoch": 0.2158203125, + "grad_norm": 0.265655517578125, + "learning_rate": 0.00046023349641039786, + "loss": 1.9223, + "step": 2210 + }, + { + "epoch": 0.21591796875, + "grad_norm": 0.308352530002594, + "learning_rate": 0.0004601931191112945, + "loss": 1.9127, + "step": 2211 + }, + { + "epoch": 0.216015625, + "grad_norm": 0.26350700855255127, + "learning_rate": 0.0004601527233130392, + "loss": 1.8908, + "step": 2212 + }, + { + "epoch": 0.21611328125, + "grad_norm": 0.28734543919563293, + "learning_rate": 0.00046011230901966716, + "loss": 1.9426, + "step": 2213 + }, + { + "epoch": 0.2162109375, + "grad_norm": 0.24283497035503387, + "learning_rate": 0.0004600718762352154, + "loss": 1.915, + "step": 2214 + }, + { + "epoch": 0.21630859375, + "grad_norm": 0.24215170741081238, + "learning_rate": 0.00046003142496372275, + "loss": 1.9154, + "step": 2215 + }, + { + "epoch": 0.21640625, + "grad_norm": 0.25598227977752686, + "learning_rate": 0.0004599909552092301, + "loss": 1.9225, + "step": 2216 + }, + { + "epoch": 0.21650390625, + "grad_norm": 0.2031700313091278, + "learning_rate": 0.00045995046697577975, + "loss": 1.8965, + "step": 2217 + }, + { + "epoch": 0.2166015625, + "grad_norm": 0.28480711579322815, + "learning_rate": 0.0004599099602674163, + "loss": 1.9212, + "step": 2218 + }, + { + "epoch": 0.21669921875, + "grad_norm": 0.22612528502941132, + "learning_rate": 0.000459869435088186, + "loss": 1.9439, + "step": 2219 + }, + { + "epoch": 0.216796875, + "grad_norm": 0.22312787175178528, + "learning_rate": 0.0004598288914421369, + "loss": 1.9238, + "step": 2220 + }, + { + "epoch": 0.21689453125, + "grad_norm": 0.24936683475971222, + "learning_rate": 0.000459788329333319, + "loss": 1.8923, + "step": 2221 + }, + { + "epoch": 0.2169921875, + "grad_norm": 0.2038477212190628, + "learning_rate": 0.00045974774876578406, + "loss": 1.8993, + "step": 2222 + }, + { + "epoch": 0.21708984375, + "grad_norm": 0.20470038056373596, + "learning_rate": 0.00045970714974358576, + "loss": 1.901, + "step": 2223 + }, + { + "epoch": 0.2171875, + "grad_norm": 0.20012640953063965, + "learning_rate": 0.00045966653227077955, + "loss": 1.8809, + "step": 2224 + }, + { + "epoch": 0.21728515625, + "grad_norm": 0.1954420655965805, + "learning_rate": 0.0004596258963514228, + "loss": 1.9468, + "step": 2225 + }, + { + "epoch": 0.2173828125, + "grad_norm": 0.19190649688243866, + "learning_rate": 0.00045958524198957463, + "loss": 1.8548, + "step": 2226 + }, + { + "epoch": 0.21748046875, + "grad_norm": 0.2360774725675583, + "learning_rate": 0.000459544569189296, + "loss": 1.8779, + "step": 2227 + }, + { + "epoch": 0.217578125, + "grad_norm": 0.24308523535728455, + "learning_rate": 0.0004595038779546499, + "loss": 1.9122, + "step": 2228 + }, + { + "epoch": 0.21767578125, + "grad_norm": 0.31237441301345825, + "learning_rate": 0.00045946316828970093, + "loss": 1.9101, + "step": 2229 + }, + { + "epoch": 0.2177734375, + "grad_norm": 0.26211169362068176, + "learning_rate": 0.00045942244019851557, + "loss": 1.8804, + "step": 2230 + }, + { + "epoch": 0.21787109375, + "grad_norm": 0.2568318843841553, + "learning_rate": 0.0004593816936851623, + "loss": 1.8625, + "step": 2231 + }, + { + "epoch": 0.21796875, + "grad_norm": 0.2667257487773895, + "learning_rate": 0.00045934092875371134, + "loss": 1.8856, + "step": 2232 + }, + { + "epoch": 0.21806640625, + "grad_norm": 0.24612100422382355, + "learning_rate": 0.0004593001454082347, + "loss": 1.8921, + "step": 2233 + }, + { + "epoch": 0.2181640625, + "grad_norm": 0.2511996626853943, + "learning_rate": 0.0004592593436528063, + "loss": 1.8955, + "step": 2234 + }, + { + "epoch": 0.21826171875, + "grad_norm": 0.2945897877216339, + "learning_rate": 0.0004592185234915019, + "loss": 1.8911, + "step": 2235 + }, + { + "epoch": 0.218359375, + "grad_norm": 0.34499260783195496, + "learning_rate": 0.00045917768492839895, + "loss": 1.9234, + "step": 2236 + }, + { + "epoch": 0.21845703125, + "grad_norm": 0.20355650782585144, + "learning_rate": 0.000459136827967577, + "loss": 1.8892, + "step": 2237 + }, + { + "epoch": 0.2185546875, + "grad_norm": 0.31010961532592773, + "learning_rate": 0.0004590959526131172, + "loss": 1.8915, + "step": 2238 + }, + { + "epoch": 0.21865234375, + "grad_norm": 0.3139292001724243, + "learning_rate": 0.0004590550588691028, + "loss": 1.9084, + "step": 2239 + }, + { + "epoch": 0.21875, + "grad_norm": 0.2442852258682251, + "learning_rate": 0.00045901414673961844, + "loss": 1.8947, + "step": 2240 + }, + { + "epoch": 0.21884765625, + "grad_norm": 0.27955976128578186, + "learning_rate": 0.00045897321622875105, + "loss": 1.8675, + "step": 2241 + }, + { + "epoch": 0.2189453125, + "grad_norm": 0.1819334477186203, + "learning_rate": 0.00045893226734058923, + "loss": 1.8999, + "step": 2242 + }, + { + "epoch": 0.21904296875, + "grad_norm": 0.24583211541175842, + "learning_rate": 0.00045889130007922347, + "loss": 1.8846, + "step": 2243 + }, + { + "epoch": 0.219140625, + "grad_norm": 0.2537577450275421, + "learning_rate": 0.00045885031444874597, + "loss": 1.9059, + "step": 2244 + }, + { + "epoch": 0.21923828125, + "grad_norm": 0.22242943942546844, + "learning_rate": 0.00045880931045325074, + "loss": 1.8862, + "step": 2245 + }, + { + "epoch": 0.2193359375, + "grad_norm": 0.2518673241138458, + "learning_rate": 0.0004587682880968338, + "loss": 1.9075, + "step": 2246 + }, + { + "epoch": 0.21943359375, + "grad_norm": 0.2151048630475998, + "learning_rate": 0.00045872724738359297, + "loss": 1.9299, + "step": 2247 + }, + { + "epoch": 0.21953125, + "grad_norm": 0.24207067489624023, + "learning_rate": 0.00045868618831762775, + "loss": 1.8986, + "step": 2248 + }, + { + "epoch": 0.21962890625, + "grad_norm": 0.2483375072479248, + "learning_rate": 0.0004586451109030397, + "loss": 1.9244, + "step": 2249 + }, + { + "epoch": 0.2197265625, + "grad_norm": 0.19953475892543793, + "learning_rate": 0.000458604015143932, + "loss": 1.9284, + "step": 2250 + }, + { + "epoch": 0.21982421875, + "grad_norm": 0.20768243074417114, + "learning_rate": 0.00045856290104440974, + "loss": 1.8846, + "step": 2251 + }, + { + "epoch": 0.219921875, + "grad_norm": 0.21332871913909912, + "learning_rate": 0.0004585217686085799, + "loss": 1.8777, + "step": 2252 + }, + { + "epoch": 0.22001953125, + "grad_norm": 0.22202058136463165, + "learning_rate": 0.0004584806178405512, + "loss": 1.8778, + "step": 2253 + }, + { + "epoch": 0.2201171875, + "grad_norm": 0.22092659771442413, + "learning_rate": 0.00045843944874443435, + "loss": 1.9006, + "step": 2254 + }, + { + "epoch": 0.22021484375, + "grad_norm": 0.2076462358236313, + "learning_rate": 0.0004583982613243416, + "loss": 1.8781, + "step": 2255 + }, + { + "epoch": 0.2203125, + "grad_norm": 0.26965686678886414, + "learning_rate": 0.0004583570555843874, + "loss": 1.8818, + "step": 2256 + }, + { + "epoch": 0.22041015625, + "grad_norm": 0.23709794878959656, + "learning_rate": 0.0004583158315286877, + "loss": 1.799, + "step": 2257 + }, + { + "epoch": 0.2205078125, + "grad_norm": 0.21832555532455444, + "learning_rate": 0.00045827458916136043, + "loss": 1.9321, + "step": 2258 + }, + { + "epoch": 0.22060546875, + "grad_norm": 0.268564909696579, + "learning_rate": 0.0004582333284865254, + "loss": 1.9043, + "step": 2259 + }, + { + "epoch": 0.220703125, + "grad_norm": 0.2340027242898941, + "learning_rate": 0.00045819204950830426, + "loss": 1.8435, + "step": 2260 + }, + { + "epoch": 0.22080078125, + "grad_norm": 0.2653496563434601, + "learning_rate": 0.00045815075223082016, + "loss": 1.8857, + "step": 2261 + }, + { + "epoch": 0.2208984375, + "grad_norm": 0.2956698536872864, + "learning_rate": 0.00045810943665819853, + "loss": 1.9683, + "step": 2262 + }, + { + "epoch": 0.22099609375, + "grad_norm": 0.2561390697956085, + "learning_rate": 0.0004580681027945663, + "loss": 1.9258, + "step": 2263 + }, + { + "epoch": 0.22109375, + "grad_norm": 0.2162081003189087, + "learning_rate": 0.00045802675064405266, + "loss": 1.9093, + "step": 2264 + }, + { + "epoch": 0.22119140625, + "grad_norm": 0.2608626186847687, + "learning_rate": 0.0004579853802107879, + "loss": 1.8742, + "step": 2265 + }, + { + "epoch": 0.2212890625, + "grad_norm": 0.2352752387523651, + "learning_rate": 0.0004579439914989049, + "loss": 1.9224, + "step": 2266 + }, + { + "epoch": 0.22138671875, + "grad_norm": 0.20996929705142975, + "learning_rate": 0.00045790258451253776, + "loss": 1.9081, + "step": 2267 + }, + { + "epoch": 0.221484375, + "grad_norm": 0.2689194083213806, + "learning_rate": 0.00045786115925582283, + "loss": 1.928, + "step": 2268 + }, + { + "epoch": 0.22158203125, + "grad_norm": 0.27555009722709656, + "learning_rate": 0.0004578197157328981, + "loss": 1.8619, + "step": 2269 + }, + { + "epoch": 0.2216796875, + "grad_norm": 0.23493286967277527, + "learning_rate": 0.00045777825394790344, + "loss": 1.9058, + "step": 2270 + }, + { + "epoch": 0.22177734375, + "grad_norm": 0.18501000106334686, + "learning_rate": 0.0004577367739049804, + "loss": 1.935, + "step": 2271 + }, + { + "epoch": 0.221875, + "grad_norm": 0.2200579047203064, + "learning_rate": 0.00045769527560827263, + "loss": 1.9219, + "step": 2272 + }, + { + "epoch": 0.22197265625, + "grad_norm": 0.24917781352996826, + "learning_rate": 0.0004576537590619254, + "loss": 1.8729, + "step": 2273 + }, + { + "epoch": 0.2220703125, + "grad_norm": 0.2548181116580963, + "learning_rate": 0.0004576122242700857, + "loss": 1.8637, + "step": 2274 + }, + { + "epoch": 0.22216796875, + "grad_norm": 0.27391573786735535, + "learning_rate": 0.0004575706712369027, + "loss": 1.9145, + "step": 2275 + }, + { + "epoch": 0.222265625, + "grad_norm": 0.21581903100013733, + "learning_rate": 0.00045752909996652695, + "loss": 1.9498, + "step": 2276 + }, + { + "epoch": 0.22236328125, + "grad_norm": 0.2738785147666931, + "learning_rate": 0.00045748751046311125, + "loss": 1.9297, + "step": 2277 + }, + { + "epoch": 0.2224609375, + "grad_norm": 0.2592880427837372, + "learning_rate": 0.00045744590273080987, + "loss": 1.9323, + "step": 2278 + }, + { + "epoch": 0.22255859375, + "grad_norm": 0.21176159381866455, + "learning_rate": 0.00045740427677377926, + "loss": 1.9299, + "step": 2279 + }, + { + "epoch": 0.22265625, + "grad_norm": 0.2628064751625061, + "learning_rate": 0.00045736263259617727, + "loss": 1.9047, + "step": 2280 + }, + { + "epoch": 0.22275390625, + "grad_norm": 0.199008047580719, + "learning_rate": 0.00045732097020216393, + "loss": 1.8896, + "step": 2281 + }, + { + "epoch": 0.2228515625, + "grad_norm": 0.2588101923465729, + "learning_rate": 0.00045727928959590086, + "loss": 1.9043, + "step": 2282 + }, + { + "epoch": 0.22294921875, + "grad_norm": 0.2765114903450012, + "learning_rate": 0.00045723759078155165, + "loss": 1.8701, + "step": 2283 + }, + { + "epoch": 0.223046875, + "grad_norm": 0.22020654380321503, + "learning_rate": 0.0004571958737632815, + "loss": 1.842, + "step": 2284 + }, + { + "epoch": 0.22314453125, + "grad_norm": 0.25397929549217224, + "learning_rate": 0.0004571541385452577, + "loss": 1.9356, + "step": 2285 + }, + { + "epoch": 0.2232421875, + "grad_norm": 0.3181978464126587, + "learning_rate": 0.0004571123851316492, + "loss": 1.8812, + "step": 2286 + }, + { + "epoch": 0.22333984375, + "grad_norm": 0.24803847074508667, + "learning_rate": 0.0004570706135266268, + "loss": 1.9017, + "step": 2287 + }, + { + "epoch": 0.2234375, + "grad_norm": 0.2591163218021393, + "learning_rate": 0.00045702882373436317, + "loss": 1.8597, + "step": 2288 + }, + { + "epoch": 0.22353515625, + "grad_norm": 0.2701427936553955, + "learning_rate": 0.00045698701575903265, + "loss": 1.9202, + "step": 2289 + }, + { + "epoch": 0.2236328125, + "grad_norm": 0.259224534034729, + "learning_rate": 0.00045694518960481145, + "loss": 1.8726, + "step": 2290 + }, + { + "epoch": 0.22373046875, + "grad_norm": 0.2605198919773102, + "learning_rate": 0.00045690334527587786, + "loss": 1.8942, + "step": 2291 + }, + { + "epoch": 0.223828125, + "grad_norm": 0.2564171254634857, + "learning_rate": 0.00045686148277641143, + "loss": 1.8893, + "step": 2292 + }, + { + "epoch": 0.22392578125, + "grad_norm": 0.2904933989048004, + "learning_rate": 0.0004568196021105941, + "loss": 1.8712, + "step": 2293 + }, + { + "epoch": 0.2240234375, + "grad_norm": 0.19899466633796692, + "learning_rate": 0.0004567777032826092, + "loss": 1.882, + "step": 2294 + }, + { + "epoch": 0.22412109375, + "grad_norm": 0.3275931477546692, + "learning_rate": 0.00045673578629664227, + "loss": 1.9256, + "step": 2295 + }, + { + "epoch": 0.22421875, + "grad_norm": 0.25155848264694214, + "learning_rate": 0.0004566938511568802, + "loss": 1.8914, + "step": 2296 + }, + { + "epoch": 0.22431640625, + "grad_norm": 0.2259829044342041, + "learning_rate": 0.00045665189786751214, + "loss": 1.9638, + "step": 2297 + }, + { + "epoch": 0.2244140625, + "grad_norm": 0.21916180849075317, + "learning_rate": 0.0004566099264327288, + "loss": 1.8701, + "step": 2298 + }, + { + "epoch": 0.22451171875, + "grad_norm": 0.2307112216949463, + "learning_rate": 0.0004565679368567227, + "loss": 1.9039, + "step": 2299 + }, + { + "epoch": 0.224609375, + "grad_norm": 0.25406181812286377, + "learning_rate": 0.00045652592914368826, + "loss": 1.8874, + "step": 2300 + }, + { + "epoch": 0.22470703125, + "grad_norm": 0.23334388434886932, + "learning_rate": 0.0004564839032978216, + "loss": 1.9329, + "step": 2301 + }, + { + "epoch": 0.2248046875, + "grad_norm": 0.312667578458786, + "learning_rate": 0.0004564418593233209, + "loss": 1.8981, + "step": 2302 + }, + { + "epoch": 0.22490234375, + "grad_norm": 0.21629926562309265, + "learning_rate": 0.00045639979722438586, + "loss": 1.9121, + "step": 2303 + }, + { + "epoch": 0.225, + "grad_norm": 0.2639700174331665, + "learning_rate": 0.0004563577170052182, + "loss": 1.8917, + "step": 2304 + }, + { + "epoch": 0.22509765625, + "grad_norm": 0.28482767939567566, + "learning_rate": 0.00045631561867002114, + "loss": 1.8615, + "step": 2305 + }, + { + "epoch": 0.2251953125, + "grad_norm": 0.21128278970718384, + "learning_rate": 0.0004562735022230002, + "loss": 1.8742, + "step": 2306 + }, + { + "epoch": 0.22529296875, + "grad_norm": 0.22739477455615997, + "learning_rate": 0.00045623136766836226, + "loss": 1.8953, + "step": 2307 + }, + { + "epoch": 0.225390625, + "grad_norm": 0.2244565635919571, + "learning_rate": 0.0004561892150103163, + "loss": 1.8773, + "step": 2308 + }, + { + "epoch": 0.22548828125, + "grad_norm": 0.24845881760120392, + "learning_rate": 0.00045614704425307295, + "loss": 1.9182, + "step": 2309 + }, + { + "epoch": 0.2255859375, + "grad_norm": 0.24752497673034668, + "learning_rate": 0.0004561048554008447, + "loss": 1.8817, + "step": 2310 + }, + { + "epoch": 0.22568359375, + "grad_norm": 0.20452168583869934, + "learning_rate": 0.00045606264845784577, + "loss": 1.8662, + "step": 2311 + }, + { + "epoch": 0.22578125, + "grad_norm": 0.22137118875980377, + "learning_rate": 0.00045602042342829244, + "loss": 1.8966, + "step": 2312 + }, + { + "epoch": 0.22587890625, + "grad_norm": 0.2197723239660263, + "learning_rate": 0.00045597818031640244, + "loss": 1.9152, + "step": 2313 + }, + { + "epoch": 0.2259765625, + "grad_norm": 0.19033396244049072, + "learning_rate": 0.0004559359191263955, + "loss": 1.92, + "step": 2314 + }, + { + "epoch": 0.22607421875, + "grad_norm": 0.21937741339206696, + "learning_rate": 0.0004558936398624933, + "loss": 1.9554, + "step": 2315 + }, + { + "epoch": 0.226171875, + "grad_norm": 0.20096588134765625, + "learning_rate": 0.000455851342528919, + "loss": 1.9401, + "step": 2316 + }, + { + "epoch": 0.22626953125, + "grad_norm": 0.22804000973701477, + "learning_rate": 0.00045580902712989783, + "loss": 1.8753, + "step": 2317 + }, + { + "epoch": 0.2263671875, + "grad_norm": 0.2619505226612091, + "learning_rate": 0.0004557666936696566, + "loss": 1.8913, + "step": 2318 + }, + { + "epoch": 0.22646484375, + "grad_norm": 0.2752479016780853, + "learning_rate": 0.0004557243421524241, + "loss": 1.8744, + "step": 2319 + }, + { + "epoch": 0.2265625, + "grad_norm": 0.2984503209590912, + "learning_rate": 0.0004556819725824309, + "loss": 1.8951, + "step": 2320 + }, + { + "epoch": 0.22666015625, + "grad_norm": 0.22869320213794708, + "learning_rate": 0.0004556395849639094, + "loss": 1.9453, + "step": 2321 + }, + { + "epoch": 0.2267578125, + "grad_norm": 0.2314172238111496, + "learning_rate": 0.0004555971793010937, + "loss": 1.8549, + "step": 2322 + }, + { + "epoch": 0.22685546875, + "grad_norm": 0.2702145278453827, + "learning_rate": 0.0004555547555982197, + "loss": 1.8916, + "step": 2323 + }, + { + "epoch": 0.226953125, + "grad_norm": 0.29122287034988403, + "learning_rate": 0.00045551231385952516, + "loss": 1.9072, + "step": 2324 + }, + { + "epoch": 0.22705078125, + "grad_norm": 0.26239168643951416, + "learning_rate": 0.0004554698540892497, + "loss": 1.9243, + "step": 2325 + }, + { + "epoch": 0.2271484375, + "grad_norm": 0.25048989057540894, + "learning_rate": 0.0004554273762916346, + "loss": 1.8799, + "step": 2326 + }, + { + "epoch": 0.22724609375, + "grad_norm": 0.21093380451202393, + "learning_rate": 0.0004553848804709231, + "loss": 1.928, + "step": 2327 + }, + { + "epoch": 0.22734375, + "grad_norm": 0.24841107428073883, + "learning_rate": 0.00045534236663136, + "loss": 1.867, + "step": 2328 + }, + { + "epoch": 0.22744140625, + "grad_norm": 0.3298115134239197, + "learning_rate": 0.0004552998347771923, + "loss": 1.9387, + "step": 2329 + }, + { + "epoch": 0.2275390625, + "grad_norm": 0.2536830008029938, + "learning_rate": 0.0004552572849126683, + "loss": 1.8998, + "step": 2330 + }, + { + "epoch": 0.22763671875, + "grad_norm": 0.25619152188301086, + "learning_rate": 0.0004552147170420386, + "loss": 1.9069, + "step": 2331 + }, + { + "epoch": 0.227734375, + "grad_norm": 0.21483200788497925, + "learning_rate": 0.00045517213116955514, + "loss": 1.8838, + "step": 2332 + }, + { + "epoch": 0.22783203125, + "grad_norm": 0.24529071152210236, + "learning_rate": 0.00045512952729947196, + "loss": 1.8954, + "step": 2333 + }, + { + "epoch": 0.2279296875, + "grad_norm": 0.2747213840484619, + "learning_rate": 0.0004550869054360448, + "loss": 1.8707, + "step": 2334 + }, + { + "epoch": 0.22802734375, + "grad_norm": 0.24370868504047394, + "learning_rate": 0.00045504426558353123, + "loss": 1.8723, + "step": 2335 + }, + { + "epoch": 0.228125, + "grad_norm": 0.2064986526966095, + "learning_rate": 0.0004550016077461906, + "loss": 1.893, + "step": 2336 + }, + { + "epoch": 0.22822265625, + "grad_norm": 0.28048041462898254, + "learning_rate": 0.000454958931928284, + "loss": 1.8731, + "step": 2337 + }, + { + "epoch": 0.2283203125, + "grad_norm": 0.2515961527824402, + "learning_rate": 0.0004549162381340744, + "loss": 1.9322, + "step": 2338 + }, + { + "epoch": 0.22841796875, + "grad_norm": 0.21575289964675903, + "learning_rate": 0.0004548735263678266, + "loss": 1.8622, + "step": 2339 + }, + { + "epoch": 0.228515625, + "grad_norm": 0.22191932797431946, + "learning_rate": 0.00045483079663380686, + "loss": 1.9339, + "step": 2340 + }, + { + "epoch": 0.22861328125, + "grad_norm": 0.27168577909469604, + "learning_rate": 0.0004547880489362838, + "loss": 1.8946, + "step": 2341 + }, + { + "epoch": 0.2287109375, + "grad_norm": 0.2825971841812134, + "learning_rate": 0.0004547452832795275, + "loss": 1.9222, + "step": 2342 + }, + { + "epoch": 0.22880859375, + "grad_norm": 0.2461196780204773, + "learning_rate": 0.00045470249966780976, + "loss": 1.8803, + "step": 2343 + }, + { + "epoch": 0.22890625, + "grad_norm": 0.2580738067626953, + "learning_rate": 0.00045465969810540427, + "loss": 1.9044, + "step": 2344 + }, + { + "epoch": 0.22900390625, + "grad_norm": 0.22742857038974762, + "learning_rate": 0.0004546168785965866, + "loss": 1.8423, + "step": 2345 + }, + { + "epoch": 0.2291015625, + "grad_norm": 0.2607462406158447, + "learning_rate": 0.00045457404114563406, + "loss": 1.9238, + "step": 2346 + }, + { + "epoch": 0.22919921875, + "grad_norm": 0.21950286626815796, + "learning_rate": 0.00045453118575682565, + "loss": 1.8932, + "step": 2347 + }, + { + "epoch": 0.229296875, + "grad_norm": 0.2564532160758972, + "learning_rate": 0.0004544883124344423, + "loss": 1.8941, + "step": 2348 + }, + { + "epoch": 0.22939453125, + "grad_norm": 0.2837105393409729, + "learning_rate": 0.0004544454211827667, + "loss": 1.9385, + "step": 2349 + }, + { + "epoch": 0.2294921875, + "grad_norm": 0.268484503030777, + "learning_rate": 0.00045440251200608316, + "loss": 1.8984, + "step": 2350 + }, + { + "epoch": 0.22958984375, + "grad_norm": 0.3416820466518402, + "learning_rate": 0.0004543595849086782, + "loss": 1.7954, + "step": 2351 + }, + { + "epoch": 0.2296875, + "grad_norm": 0.3154846727848053, + "learning_rate": 0.00045431663989483955, + "loss": 1.8759, + "step": 2352 + }, + { + "epoch": 0.22978515625, + "grad_norm": 0.29342061281204224, + "learning_rate": 0.00045427367696885727, + "loss": 1.8975, + "step": 2353 + }, + { + "epoch": 0.2298828125, + "grad_norm": 0.27414578199386597, + "learning_rate": 0.00045423069613502275, + "loss": 1.8983, + "step": 2354 + }, + { + "epoch": 0.22998046875, + "grad_norm": 0.2255915254354477, + "learning_rate": 0.00045418769739762966, + "loss": 1.9371, + "step": 2355 + }, + { + "epoch": 0.230078125, + "grad_norm": 0.21754321455955505, + "learning_rate": 0.000454144680760973, + "loss": 1.8771, + "step": 2356 + }, + { + "epoch": 0.23017578125, + "grad_norm": 0.26277226209640503, + "learning_rate": 0.0004541016462293498, + "loss": 1.8672, + "step": 2357 + }, + { + "epoch": 0.2302734375, + "grad_norm": 0.22035324573516846, + "learning_rate": 0.0004540585938070589, + "loss": 1.8724, + "step": 2358 + }, + { + "epoch": 0.23037109375, + "grad_norm": 0.22731097042560577, + "learning_rate": 0.00045401552349840077, + "loss": 1.9341, + "step": 2359 + }, + { + "epoch": 0.23046875, + "grad_norm": 0.23395659029483795, + "learning_rate": 0.00045397243530767773, + "loss": 1.9038, + "step": 2360 + }, + { + "epoch": 0.23056640625, + "grad_norm": 0.24531114101409912, + "learning_rate": 0.000453929329239194, + "loss": 1.8877, + "step": 2361 + }, + { + "epoch": 0.2306640625, + "grad_norm": 0.25629279017448425, + "learning_rate": 0.00045388620529725546, + "loss": 1.8831, + "step": 2362 + }, + { + "epoch": 0.23076171875, + "grad_norm": 0.24789990484714508, + "learning_rate": 0.00045384306348616977, + "loss": 1.8992, + "step": 2363 + }, + { + "epoch": 0.230859375, + "grad_norm": 0.23063692450523376, + "learning_rate": 0.00045379990381024644, + "loss": 1.8645, + "step": 2364 + }, + { + "epoch": 0.23095703125, + "grad_norm": 0.274730920791626, + "learning_rate": 0.0004537567262737968, + "loss": 1.8846, + "step": 2365 + }, + { + "epoch": 0.2310546875, + "grad_norm": 0.21056944131851196, + "learning_rate": 0.00045371353088113377, + "loss": 1.8942, + "step": 2366 + }, + { + "epoch": 0.23115234375, + "grad_norm": 0.2247105836868286, + "learning_rate": 0.0004536703176365723, + "loss": 1.9012, + "step": 2367 + }, + { + "epoch": 0.23125, + "grad_norm": 0.23797035217285156, + "learning_rate": 0.00045362708654442897, + "loss": 1.8886, + "step": 2368 + }, + { + "epoch": 0.23134765625, + "grad_norm": 0.2110954225063324, + "learning_rate": 0.0004535838376090222, + "loss": 1.8937, + "step": 2369 + }, + { + "epoch": 0.2314453125, + "grad_norm": 0.2745433747768402, + "learning_rate": 0.00045354057083467217, + "loss": 1.8799, + "step": 2370 + }, + { + "epoch": 0.23154296875, + "grad_norm": 0.20481610298156738, + "learning_rate": 0.0004534972862257008, + "loss": 1.8282, + "step": 2371 + }, + { + "epoch": 0.231640625, + "grad_norm": 0.4868389368057251, + "learning_rate": 0.00045345398378643186, + "loss": 1.8975, + "step": 2372 + }, + { + "epoch": 0.23173828125, + "grad_norm": 0.2749464213848114, + "learning_rate": 0.00045341066352119096, + "loss": 1.8683, + "step": 2373 + }, + { + "epoch": 0.2318359375, + "grad_norm": 0.2716321051120758, + "learning_rate": 0.00045336732543430536, + "loss": 1.8833, + "step": 2374 + }, + { + "epoch": 0.23193359375, + "grad_norm": 0.2876909077167511, + "learning_rate": 0.00045332396953010415, + "loss": 1.881, + "step": 2375 + }, + { + "epoch": 0.23203125, + "grad_norm": 0.2515881061553955, + "learning_rate": 0.0004532805958129181, + "loss": 1.9023, + "step": 2376 + }, + { + "epoch": 0.23212890625, + "grad_norm": 0.2852384150028229, + "learning_rate": 0.00045323720428708003, + "loss": 1.924, + "step": 2377 + }, + { + "epoch": 0.2322265625, + "grad_norm": 0.22907400131225586, + "learning_rate": 0.0004531937949569243, + "loss": 1.8765, + "step": 2378 + }, + { + "epoch": 0.23232421875, + "grad_norm": 0.2648322582244873, + "learning_rate": 0.0004531503678267871, + "loss": 1.9247, + "step": 2379 + }, + { + "epoch": 0.232421875, + "grad_norm": 0.28034675121307373, + "learning_rate": 0.00045310692290100637, + "loss": 1.8948, + "step": 2380 + }, + { + "epoch": 0.23251953125, + "grad_norm": 0.29645946621894836, + "learning_rate": 0.00045306346018392197, + "loss": 1.9011, + "step": 2381 + }, + { + "epoch": 0.2326171875, + "grad_norm": 0.24943086504936218, + "learning_rate": 0.0004530199796798754, + "loss": 1.9089, + "step": 2382 + }, + { + "epoch": 0.23271484375, + "grad_norm": 0.279883474111557, + "learning_rate": 0.00045297648139321, + "loss": 1.8779, + "step": 2383 + }, + { + "epoch": 0.2328125, + "grad_norm": 0.22576992213726044, + "learning_rate": 0.00045293296532827074, + "loss": 1.9037, + "step": 2384 + }, + { + "epoch": 0.23291015625, + "grad_norm": 0.2487352192401886, + "learning_rate": 0.0004528894314894047, + "loss": 1.9089, + "step": 2385 + }, + { + "epoch": 0.2330078125, + "grad_norm": 0.25155317783355713, + "learning_rate": 0.0004528458798809603, + "loss": 1.8979, + "step": 2386 + }, + { + "epoch": 0.23310546875, + "grad_norm": 0.25363826751708984, + "learning_rate": 0.00045280231050728815, + "loss": 1.9251, + "step": 2387 + }, + { + "epoch": 0.233203125, + "grad_norm": 0.2635079026222229, + "learning_rate": 0.0004527587233727404, + "loss": 1.9152, + "step": 2388 + }, + { + "epoch": 0.23330078125, + "grad_norm": 0.2683078944683075, + "learning_rate": 0.0004527151184816709, + "loss": 1.8777, + "step": 2389 + }, + { + "epoch": 0.2333984375, + "grad_norm": 0.23932483792304993, + "learning_rate": 0.00045267149583843555, + "loss": 1.8953, + "step": 2390 + }, + { + "epoch": 0.23349609375, + "grad_norm": 0.20382820069789886, + "learning_rate": 0.00045262785544739173, + "loss": 1.9141, + "step": 2391 + }, + { + "epoch": 0.23359375, + "grad_norm": 0.21496914327144623, + "learning_rate": 0.00045258419731289884, + "loss": 1.9053, + "step": 2392 + }, + { + "epoch": 0.23369140625, + "grad_norm": 0.3160604238510132, + "learning_rate": 0.00045254052143931786, + "loss": 1.9264, + "step": 2393 + }, + { + "epoch": 0.2337890625, + "grad_norm": 0.3175326883792877, + "learning_rate": 0.0004524968278310117, + "loss": 1.8834, + "step": 2394 + }, + { + "epoch": 0.23388671875, + "grad_norm": 0.2783551514148712, + "learning_rate": 0.0004524531164923448, + "loss": 1.8688, + "step": 2395 + }, + { + "epoch": 0.233984375, + "grad_norm": 0.24283266067504883, + "learning_rate": 0.0004524093874276838, + "loss": 1.9118, + "step": 2396 + }, + { + "epoch": 0.23408203125, + "grad_norm": 0.26176488399505615, + "learning_rate": 0.0004523656406413967, + "loss": 1.9251, + "step": 2397 + }, + { + "epoch": 0.2341796875, + "grad_norm": 0.24710685014724731, + "learning_rate": 0.0004523218761378533, + "loss": 1.8714, + "step": 2398 + }, + { + "epoch": 0.23427734375, + "grad_norm": 0.21796317398548126, + "learning_rate": 0.00045227809392142546, + "loss": 1.8935, + "step": 2399 + }, + { + "epoch": 0.234375, + "grad_norm": 0.25704821944236755, + "learning_rate": 0.00045223429399648664, + "loss": 1.876, + "step": 2400 + }, + { + "epoch": 0.23447265625, + "grad_norm": 0.2729525864124298, + "learning_rate": 0.0004521904763674119, + "loss": 1.9012, + "step": 2401 + }, + { + "epoch": 0.2345703125, + "grad_norm": 0.18755966424942017, + "learning_rate": 0.00045214664103857845, + "loss": 1.8964, + "step": 2402 + }, + { + "epoch": 0.23466796875, + "grad_norm": 0.2789023816585541, + "learning_rate": 0.0004521027880143648, + "loss": 1.8949, + "step": 2403 + }, + { + "epoch": 0.234765625, + "grad_norm": 0.21166810393333435, + "learning_rate": 0.00045205891729915176, + "loss": 1.9133, + "step": 2404 + }, + { + "epoch": 0.23486328125, + "grad_norm": 0.22952231764793396, + "learning_rate": 0.00045201502889732144, + "loss": 1.9143, + "step": 2405 + }, + { + "epoch": 0.2349609375, + "grad_norm": 0.23634375631809235, + "learning_rate": 0.00045197112281325784, + "loss": 1.8645, + "step": 2406 + }, + { + "epoch": 0.23505859375, + "grad_norm": 0.25142142176628113, + "learning_rate": 0.00045192719905134705, + "loss": 1.9343, + "step": 2407 + }, + { + "epoch": 0.23515625, + "grad_norm": 0.2976498007774353, + "learning_rate": 0.0004518832576159764, + "loss": 1.918, + "step": 2408 + }, + { + "epoch": 0.23525390625, + "grad_norm": 0.2667110562324524, + "learning_rate": 0.0004518392985115354, + "loss": 1.9093, + "step": 2409 + }, + { + "epoch": 0.2353515625, + "grad_norm": 0.2608035206794739, + "learning_rate": 0.00045179532174241516, + "loss": 1.9234, + "step": 2410 + }, + { + "epoch": 0.23544921875, + "grad_norm": 0.26505768299102783, + "learning_rate": 0.0004517513273130085, + "loss": 1.8844, + "step": 2411 + }, + { + "epoch": 0.235546875, + "grad_norm": 0.31293320655822754, + "learning_rate": 0.0004517073152277101, + "loss": 1.907, + "step": 2412 + }, + { + "epoch": 0.23564453125, + "grad_norm": 0.27297648787498474, + "learning_rate": 0.00045166328549091643, + "loss": 1.9158, + "step": 2413 + }, + { + "epoch": 0.2357421875, + "grad_norm": 0.2741674780845642, + "learning_rate": 0.00045161923810702563, + "loss": 1.9011, + "step": 2414 + }, + { + "epoch": 0.23583984375, + "grad_norm": 0.29327699542045593, + "learning_rate": 0.00045157517308043774, + "loss": 1.894, + "step": 2415 + }, + { + "epoch": 0.2359375, + "grad_norm": 0.28643742203712463, + "learning_rate": 0.00045153109041555425, + "loss": 1.8993, + "step": 2416 + }, + { + "epoch": 0.23603515625, + "grad_norm": 0.46471482515335083, + "learning_rate": 0.00045148699011677873, + "loss": 1.9068, + "step": 2417 + }, + { + "epoch": 0.2361328125, + "grad_norm": 0.331559956073761, + "learning_rate": 0.0004514428721885165, + "loss": 1.9106, + "step": 2418 + }, + { + "epoch": 0.23623046875, + "grad_norm": 0.2610962688922882, + "learning_rate": 0.00045139873663517445, + "loss": 1.8757, + "step": 2419 + }, + { + "epoch": 0.236328125, + "grad_norm": 0.32770147919654846, + "learning_rate": 0.0004513545834611614, + "loss": 1.9095, + "step": 2420 + }, + { + "epoch": 0.23642578125, + "grad_norm": 0.2246219664812088, + "learning_rate": 0.00045131041267088787, + "loss": 1.8643, + "step": 2421 + }, + { + "epoch": 0.2365234375, + "grad_norm": 0.31902652978897095, + "learning_rate": 0.00045126622426876596, + "loss": 1.9061, + "step": 2422 + }, + { + "epoch": 0.23662109375, + "grad_norm": 0.20017661154270172, + "learning_rate": 0.00045122201825920983, + "loss": 1.9162, + "step": 2423 + }, + { + "epoch": 0.23671875, + "grad_norm": 0.30280932784080505, + "learning_rate": 0.0004511777946466353, + "loss": 1.8944, + "step": 2424 + }, + { + "epoch": 0.23681640625, + "grad_norm": 0.20749028027057648, + "learning_rate": 0.00045113355343545994, + "loss": 1.871, + "step": 2425 + }, + { + "epoch": 0.2369140625, + "grad_norm": 0.2936987280845642, + "learning_rate": 0.000451089294630103, + "loss": 1.8808, + "step": 2426 + }, + { + "epoch": 0.23701171875, + "grad_norm": 0.18236121535301208, + "learning_rate": 0.0004510450182349854, + "loss": 1.8534, + "step": 2427 + }, + { + "epoch": 0.237109375, + "grad_norm": 0.25331661105155945, + "learning_rate": 0.0004510007242545302, + "loss": 1.8707, + "step": 2428 + }, + { + "epoch": 0.23720703125, + "grad_norm": 0.2117713838815689, + "learning_rate": 0.00045095641269316184, + "loss": 1.9254, + "step": 2429 + }, + { + "epoch": 0.2373046875, + "grad_norm": 0.25148898363113403, + "learning_rate": 0.0004509120835553067, + "loss": 1.8973, + "step": 2430 + }, + { + "epoch": 0.23740234375, + "grad_norm": 0.24941754341125488, + "learning_rate": 0.0004508677368453929, + "loss": 1.8605, + "step": 2431 + }, + { + "epoch": 0.2375, + "grad_norm": 0.20781292021274567, + "learning_rate": 0.00045082337256785015, + "loss": 1.8802, + "step": 2432 + }, + { + "epoch": 0.23759765625, + "grad_norm": 0.2777664363384247, + "learning_rate": 0.0004507789907271102, + "loss": 1.8958, + "step": 2433 + }, + { + "epoch": 0.2376953125, + "grad_norm": 0.2164555788040161, + "learning_rate": 0.0004507345913276063, + "loss": 1.8586, + "step": 2434 + }, + { + "epoch": 0.23779296875, + "grad_norm": 0.2553316354751587, + "learning_rate": 0.0004506901743737736, + "loss": 1.8892, + "step": 2435 + }, + { + "epoch": 0.237890625, + "grad_norm": 0.23472066223621368, + "learning_rate": 0.0004506457398700489, + "loss": 1.8896, + "step": 2436 + }, + { + "epoch": 0.23798828125, + "grad_norm": 0.2276161015033722, + "learning_rate": 0.00045060128782087094, + "loss": 1.8851, + "step": 2437 + }, + { + "epoch": 0.2380859375, + "grad_norm": 0.21676012873649597, + "learning_rate": 0.00045055681823068006, + "loss": 1.8854, + "step": 2438 + }, + { + "epoch": 0.23818359375, + "grad_norm": 0.22366516292095184, + "learning_rate": 0.00045051233110391823, + "loss": 1.9309, + "step": 2439 + }, + { + "epoch": 0.23828125, + "grad_norm": 0.28527531027793884, + "learning_rate": 0.00045046782644502946, + "loss": 1.8901, + "step": 2440 + }, + { + "epoch": 0.23837890625, + "grad_norm": 0.3597792088985443, + "learning_rate": 0.0004504233042584594, + "loss": 1.8695, + "step": 2441 + }, + { + "epoch": 0.2384765625, + "grad_norm": 0.2790103554725647, + "learning_rate": 0.00045037876454865526, + "loss": 1.8756, + "step": 2442 + }, + { + "epoch": 0.23857421875, + "grad_norm": 0.3042321503162384, + "learning_rate": 0.0004503342073200663, + "loss": 1.8643, + "step": 2443 + }, + { + "epoch": 0.238671875, + "grad_norm": 0.24617038667201996, + "learning_rate": 0.0004502896325771433, + "loss": 1.9004, + "step": 2444 + }, + { + "epoch": 0.23876953125, + "grad_norm": 0.29559648036956787, + "learning_rate": 0.000450245040324339, + "loss": 1.9126, + "step": 2445 + }, + { + "epoch": 0.2388671875, + "grad_norm": 0.28401464223861694, + "learning_rate": 0.0004502004305661077, + "loss": 1.8639, + "step": 2446 + }, + { + "epoch": 0.23896484375, + "grad_norm": 0.2113400399684906, + "learning_rate": 0.00045015580330690547, + "loss": 1.8863, + "step": 2447 + }, + { + "epoch": 0.2390625, + "grad_norm": 0.28481796383857727, + "learning_rate": 0.0004501111585511903, + "loss": 1.8826, + "step": 2448 + }, + { + "epoch": 0.23916015625, + "grad_norm": 0.25684455037117004, + "learning_rate": 0.00045006649630342164, + "loss": 1.8952, + "step": 2449 + }, + { + "epoch": 0.2392578125, + "grad_norm": 0.23974579572677612, + "learning_rate": 0.00045002181656806105, + "loss": 1.8615, + "step": 2450 + }, + { + "epoch": 0.23935546875, + "grad_norm": 0.24866057932376862, + "learning_rate": 0.0004499771193495715, + "loss": 1.9343, + "step": 2451 + }, + { + "epoch": 0.239453125, + "grad_norm": 0.22083763778209686, + "learning_rate": 0.0004499324046524179, + "loss": 1.8349, + "step": 2452 + }, + { + "epoch": 0.23955078125, + "grad_norm": 0.33529961109161377, + "learning_rate": 0.00044988767248106676, + "loss": 1.9153, + "step": 2453 + }, + { + "epoch": 0.2396484375, + "grad_norm": 0.26694998145103455, + "learning_rate": 0.0004498429228399866, + "loss": 1.8993, + "step": 2454 + }, + { + "epoch": 0.23974609375, + "grad_norm": 0.29110854864120483, + "learning_rate": 0.0004497981557336474, + "loss": 1.9531, + "step": 2455 + }, + { + "epoch": 0.23984375, + "grad_norm": 0.3439907729625702, + "learning_rate": 0.00044975337116652097, + "loss": 1.8862, + "step": 2456 + }, + { + "epoch": 0.23994140625, + "grad_norm": 0.2403963953256607, + "learning_rate": 0.00044970856914308097, + "loss": 1.8965, + "step": 2457 + }, + { + "epoch": 0.2400390625, + "grad_norm": 0.28895774483680725, + "learning_rate": 0.00044966374966780266, + "loss": 1.8895, + "step": 2458 + }, + { + "epoch": 0.24013671875, + "grad_norm": 0.2802990972995758, + "learning_rate": 0.00044961891274516313, + "loss": 1.8932, + "step": 2459 + }, + { + "epoch": 0.240234375, + "grad_norm": 0.251863032579422, + "learning_rate": 0.0004495740583796413, + "loss": 1.9222, + "step": 2460 + }, + { + "epoch": 0.24033203125, + "grad_norm": 0.22680626809597015, + "learning_rate": 0.0004495291865757175, + "loss": 1.8676, + "step": 2461 + }, + { + "epoch": 0.2404296875, + "grad_norm": 0.24981464445590973, + "learning_rate": 0.0004494842973378742, + "loss": 1.8868, + "step": 2462 + }, + { + "epoch": 0.24052734375, + "grad_norm": 0.19781389832496643, + "learning_rate": 0.0004494393906705954, + "loss": 1.9079, + "step": 2463 + }, + { + "epoch": 0.240625, + "grad_norm": 0.22644692659378052, + "learning_rate": 0.00044939446657836686, + "loss": 1.9074, + "step": 2464 + }, + { + "epoch": 0.24072265625, + "grad_norm": 0.2107515037059784, + "learning_rate": 0.0004493495250656761, + "loss": 1.8994, + "step": 2465 + }, + { + "epoch": 0.2408203125, + "grad_norm": 0.18987129628658295, + "learning_rate": 0.00044930456613701236, + "loss": 1.9397, + "step": 2466 + }, + { + "epoch": 0.24091796875, + "grad_norm": 0.21435169875621796, + "learning_rate": 0.00044925958979686676, + "loss": 1.9007, + "step": 2467 + }, + { + "epoch": 0.241015625, + "grad_norm": 0.1948396861553192, + "learning_rate": 0.00044921459604973184, + "loss": 1.8762, + "step": 2468 + }, + { + "epoch": 0.24111328125, + "grad_norm": 0.2153381109237671, + "learning_rate": 0.00044916958490010217, + "loss": 1.9325, + "step": 2469 + }, + { + "epoch": 0.2412109375, + "grad_norm": 0.21237823367118835, + "learning_rate": 0.00044912455635247404, + "loss": 1.9006, + "step": 2470 + }, + { + "epoch": 0.24130859375, + "grad_norm": 0.19011497497558594, + "learning_rate": 0.0004490795104113453, + "loss": 1.9169, + "step": 2471 + }, + { + "epoch": 0.24140625, + "grad_norm": 0.2259521484375, + "learning_rate": 0.0004490344470812157, + "loss": 1.8984, + "step": 2472 + }, + { + "epoch": 0.24150390625, + "grad_norm": 0.1522991955280304, + "learning_rate": 0.0004489893663665866, + "loss": 1.838, + "step": 2473 + }, + { + "epoch": 0.2416015625, + "grad_norm": 0.21235786378383636, + "learning_rate": 0.00044894426827196124, + "loss": 1.9033, + "step": 2474 + }, + { + "epoch": 0.24169921875, + "grad_norm": 0.20965880155563354, + "learning_rate": 0.0004488991528018445, + "loss": 1.9381, + "step": 2475 + }, + { + "epoch": 0.241796875, + "grad_norm": 0.25868213176727295, + "learning_rate": 0.0004488540199607429, + "loss": 1.8907, + "step": 2476 + }, + { + "epoch": 0.24189453125, + "grad_norm": 0.2545277178287506, + "learning_rate": 0.00044880886975316503, + "loss": 1.8905, + "step": 2477 + }, + { + "epoch": 0.2419921875, + "grad_norm": 0.17712853848934174, + "learning_rate": 0.00044876370218362085, + "loss": 1.9022, + "step": 2478 + }, + { + "epoch": 0.24208984375, + "grad_norm": 0.24900510907173157, + "learning_rate": 0.0004487185172566222, + "loss": 1.9247, + "step": 2479 + }, + { + "epoch": 0.2421875, + "grad_norm": 0.24943552911281586, + "learning_rate": 0.0004486733149766827, + "loss": 1.9378, + "step": 2480 + }, + { + "epoch": 0.24228515625, + "grad_norm": 0.27004435658454895, + "learning_rate": 0.0004486280953483177, + "loss": 1.8644, + "step": 2481 + }, + { + "epoch": 0.2423828125, + "grad_norm": 0.21977658569812775, + "learning_rate": 0.0004485828583760441, + "loss": 1.9166, + "step": 2482 + }, + { + "epoch": 0.24248046875, + "grad_norm": 0.17602773010730743, + "learning_rate": 0.0004485376040643808, + "loss": 1.8733, + "step": 2483 + }, + { + "epoch": 0.242578125, + "grad_norm": 0.2778310179710388, + "learning_rate": 0.0004484923324178483, + "loss": 1.8788, + "step": 2484 + }, + { + "epoch": 0.24267578125, + "grad_norm": 0.33434948325157166, + "learning_rate": 0.00044844704344096863, + "loss": 1.8932, + "step": 2485 + }, + { + "epoch": 0.2427734375, + "grad_norm": 0.2541770040988922, + "learning_rate": 0.00044840173713826603, + "loss": 1.8727, + "step": 2486 + }, + { + "epoch": 0.24287109375, + "grad_norm": 0.21049551665782928, + "learning_rate": 0.00044835641351426614, + "loss": 1.8567, + "step": 2487 + }, + { + "epoch": 0.24296875, + "grad_norm": 0.2456684708595276, + "learning_rate": 0.0004483110725734962, + "loss": 1.907, + "step": 2488 + }, + { + "epoch": 0.24306640625, + "grad_norm": 0.2708995044231415, + "learning_rate": 0.0004482657143204857, + "loss": 1.9146, + "step": 2489 + }, + { + "epoch": 0.2431640625, + "grad_norm": 0.2767466604709625, + "learning_rate": 0.00044822033875976513, + "loss": 1.9058, + "step": 2490 + }, + { + "epoch": 0.24326171875, + "grad_norm": 0.21536017954349518, + "learning_rate": 0.00044817494589586746, + "loss": 1.8867, + "step": 2491 + }, + { + "epoch": 0.243359375, + "grad_norm": 0.32470428943634033, + "learning_rate": 0.0004481295357333268, + "loss": 1.8651, + "step": 2492 + }, + { + "epoch": 0.24345703125, + "grad_norm": 0.35625115036964417, + "learning_rate": 0.00044808410827667936, + "loss": 1.8959, + "step": 2493 + }, + { + "epoch": 0.2435546875, + "grad_norm": 0.2434857189655304, + "learning_rate": 0.0004480386635304629, + "loss": 1.8805, + "step": 2494 + }, + { + "epoch": 0.24365234375, + "grad_norm": 0.2814987003803253, + "learning_rate": 0.000447993201499217, + "loss": 1.9388, + "step": 2495 + }, + { + "epoch": 0.24375, + "grad_norm": 0.3357986807823181, + "learning_rate": 0.00044794772218748275, + "loss": 1.9182, + "step": 2496 + }, + { + "epoch": 0.24384765625, + "grad_norm": 0.26190170645713806, + "learning_rate": 0.00044790222559980313, + "loss": 1.9124, + "step": 2497 + }, + { + "epoch": 0.2439453125, + "grad_norm": 0.2307833731174469, + "learning_rate": 0.0004478567117407231, + "loss": 1.9049, + "step": 2498 + }, + { + "epoch": 0.24404296875, + "grad_norm": 0.24210835993289948, + "learning_rate": 0.00044781118061478897, + "loss": 1.8924, + "step": 2499 + }, + { + "epoch": 0.244140625, + "grad_norm": 0.2340785562992096, + "learning_rate": 0.00044776563222654876, + "loss": 1.8846, + "step": 2500 + }, + { + "epoch": 0.24423828125, + "grad_norm": 0.20149527490139008, + "learning_rate": 0.0004477200665805525, + "loss": 1.8752, + "step": 2501 + }, + { + "epoch": 0.2443359375, + "grad_norm": 0.21381939947605133, + "learning_rate": 0.00044767448368135176, + "loss": 1.8537, + "step": 2502 + }, + { + "epoch": 0.24443359375, + "grad_norm": 0.20310865342617035, + "learning_rate": 0.0004476288835334999, + "loss": 1.926, + "step": 2503 + }, + { + "epoch": 0.24453125, + "grad_norm": 0.21173352003097534, + "learning_rate": 0.0004475832661415518, + "loss": 1.9143, + "step": 2504 + }, + { + "epoch": 0.24462890625, + "grad_norm": 0.22699382901191711, + "learning_rate": 0.0004475376315100645, + "loss": 1.9025, + "step": 2505 + }, + { + "epoch": 0.2447265625, + "grad_norm": 0.2502591907978058, + "learning_rate": 0.0004474919796435963, + "loss": 1.9133, + "step": 2506 + }, + { + "epoch": 0.24482421875, + "grad_norm": 0.18316835165023804, + "learning_rate": 0.00044744631054670743, + "loss": 1.8408, + "step": 2507 + }, + { + "epoch": 0.244921875, + "grad_norm": 0.2320968210697174, + "learning_rate": 0.00044740062422395994, + "loss": 1.924, + "step": 2508 + }, + { + "epoch": 0.24501953125, + "grad_norm": 0.22780783474445343, + "learning_rate": 0.0004473549206799174, + "loss": 1.8782, + "step": 2509 + }, + { + "epoch": 0.2451171875, + "grad_norm": 0.22676895558834076, + "learning_rate": 0.0004473091999191452, + "loss": 1.8904, + "step": 2510 + }, + { + "epoch": 0.24521484375, + "grad_norm": 0.22365321218967438, + "learning_rate": 0.0004472634619462105, + "loss": 1.9029, + "step": 2511 + }, + { + "epoch": 0.2453125, + "grad_norm": 0.24881736934185028, + "learning_rate": 0.00044721770676568203, + "loss": 1.8722, + "step": 2512 + }, + { + "epoch": 0.24541015625, + "grad_norm": 0.1900773048400879, + "learning_rate": 0.0004471719343821303, + "loss": 1.9069, + "step": 2513 + }, + { + "epoch": 0.2455078125, + "grad_norm": 0.253597229719162, + "learning_rate": 0.00044712614480012773, + "loss": 1.9053, + "step": 2514 + }, + { + "epoch": 0.24560546875, + "grad_norm": 0.2389373481273651, + "learning_rate": 0.0004470803380242481, + "loss": 1.892, + "step": 2515 + }, + { + "epoch": 0.245703125, + "grad_norm": 0.21958309412002563, + "learning_rate": 0.0004470345140590673, + "loss": 1.9074, + "step": 2516 + }, + { + "epoch": 0.24580078125, + "grad_norm": 0.24225091934204102, + "learning_rate": 0.00044698867290916263, + "loss": 1.8854, + "step": 2517 + }, + { + "epoch": 0.2458984375, + "grad_norm": 0.19851315021514893, + "learning_rate": 0.0004469428145791132, + "loss": 1.9327, + "step": 2518 + }, + { + "epoch": 0.24599609375, + "grad_norm": 0.27460166811943054, + "learning_rate": 0.0004468969390734998, + "loss": 1.8819, + "step": 2519 + }, + { + "epoch": 0.24609375, + "grad_norm": 0.250698447227478, + "learning_rate": 0.00044685104639690515, + "loss": 1.8329, + "step": 2520 + }, + { + "epoch": 0.24619140625, + "grad_norm": 0.22741656005382538, + "learning_rate": 0.0004468051365539133, + "loss": 1.9376, + "step": 2521 + }, + { + "epoch": 0.2462890625, + "grad_norm": 0.2052220106124878, + "learning_rate": 0.00044675920954911045, + "loss": 1.9158, + "step": 2522 + }, + { + "epoch": 0.24638671875, + "grad_norm": 0.20518803596496582, + "learning_rate": 0.0004467132653870842, + "loss": 1.8723, + "step": 2523 + }, + { + "epoch": 0.246484375, + "grad_norm": 0.2157951146364212, + "learning_rate": 0.00044666730407242407, + "loss": 1.8463, + "step": 2524 + }, + { + "epoch": 0.24658203125, + "grad_norm": 0.2367670238018036, + "learning_rate": 0.000446621325609721, + "loss": 1.9047, + "step": 2525 + }, + { + "epoch": 0.2466796875, + "grad_norm": 0.2216240018606186, + "learning_rate": 0.00044657533000356793, + "loss": 1.8725, + "step": 2526 + }, + { + "epoch": 0.24677734375, + "grad_norm": 0.20380742847919464, + "learning_rate": 0.00044652931725855946, + "loss": 1.9393, + "step": 2527 + }, + { + "epoch": 0.246875, + "grad_norm": 0.2439001351594925, + "learning_rate": 0.00044648328737929176, + "loss": 1.8832, + "step": 2528 + }, + { + "epoch": 0.24697265625, + "grad_norm": 0.19296857714653015, + "learning_rate": 0.00044643724037036293, + "loss": 1.8894, + "step": 2529 + }, + { + "epoch": 0.2470703125, + "grad_norm": 0.23245202004909515, + "learning_rate": 0.00044639117623637253, + "loss": 1.9035, + "step": 2530 + }, + { + "epoch": 0.24716796875, + "grad_norm": 0.23975121974945068, + "learning_rate": 0.00044634509498192197, + "loss": 1.9336, + "step": 2531 + }, + { + "epoch": 0.247265625, + "grad_norm": 0.20720966160297394, + "learning_rate": 0.0004462989966116145, + "loss": 1.9538, + "step": 2532 + }, + { + "epoch": 0.24736328125, + "grad_norm": 0.23715616762638092, + "learning_rate": 0.0004462528811300548, + "loss": 1.8944, + "step": 2533 + }, + { + "epoch": 0.2474609375, + "grad_norm": 0.22838465869426727, + "learning_rate": 0.00044620674854184937, + "loss": 1.8472, + "step": 2534 + }, + { + "epoch": 0.24755859375, + "grad_norm": 0.19445407390594482, + "learning_rate": 0.00044616059885160657, + "loss": 1.9021, + "step": 2535 + }, + { + "epoch": 0.24765625, + "grad_norm": 0.22580035030841827, + "learning_rate": 0.00044611443206393634, + "loss": 1.9035, + "step": 2536 + }, + { + "epoch": 0.24775390625, + "grad_norm": 0.18850651383399963, + "learning_rate": 0.0004460682481834503, + "loss": 1.8769, + "step": 2537 + }, + { + "epoch": 0.2478515625, + "grad_norm": 0.20952773094177246, + "learning_rate": 0.0004460220472147617, + "loss": 1.9035, + "step": 2538 + }, + { + "epoch": 0.24794921875, + "grad_norm": 0.2210458666086197, + "learning_rate": 0.00044597582916248567, + "loss": 1.882, + "step": 2539 + }, + { + "epoch": 0.248046875, + "grad_norm": 0.22584138810634613, + "learning_rate": 0.0004459295940312391, + "loss": 1.9031, + "step": 2540 + }, + { + "epoch": 0.24814453125, + "grad_norm": 0.22548091411590576, + "learning_rate": 0.0004458833418256404, + "loss": 1.9215, + "step": 2541 + }, + { + "epoch": 0.2482421875, + "grad_norm": 0.22244450449943542, + "learning_rate": 0.0004458370725503097, + "loss": 1.8845, + "step": 2542 + }, + { + "epoch": 0.24833984375, + "grad_norm": 0.25378334522247314, + "learning_rate": 0.00044579078620986896, + "loss": 1.908, + "step": 2543 + }, + { + "epoch": 0.2484375, + "grad_norm": 0.3173237144947052, + "learning_rate": 0.0004457444828089417, + "loss": 1.9347, + "step": 2544 + }, + { + "epoch": 0.24853515625, + "grad_norm": 0.3610992431640625, + "learning_rate": 0.00044569816235215333, + "loss": 1.8956, + "step": 2545 + }, + { + "epoch": 0.2486328125, + "grad_norm": 0.24675659835338593, + "learning_rate": 0.00044565182484413075, + "loss": 1.905, + "step": 2546 + }, + { + "epoch": 0.24873046875, + "grad_norm": 0.22318989038467407, + "learning_rate": 0.0004456054702895027, + "loss": 1.8726, + "step": 2547 + }, + { + "epoch": 0.248828125, + "grad_norm": 0.3162762522697449, + "learning_rate": 0.0004455590986928996, + "loss": 1.8969, + "step": 2548 + }, + { + "epoch": 0.24892578125, + "grad_norm": 0.2868858575820923, + "learning_rate": 0.0004455127100589536, + "loss": 1.8387, + "step": 2549 + }, + { + "epoch": 0.2490234375, + "grad_norm": 0.20986978709697723, + "learning_rate": 0.0004454663043922984, + "loss": 1.86, + "step": 2550 + }, + { + "epoch": 0.24912109375, + "grad_norm": 0.20684708654880524, + "learning_rate": 0.0004454198816975697, + "loss": 1.8546, + "step": 2551 + }, + { + "epoch": 0.24921875, + "grad_norm": 0.2441062480211258, + "learning_rate": 0.00044537344197940457, + "loss": 1.8749, + "step": 2552 + }, + { + "epoch": 0.24931640625, + "grad_norm": 0.2610075771808624, + "learning_rate": 0.00044532698524244184, + "loss": 1.8743, + "step": 2553 + }, + { + "epoch": 0.2494140625, + "grad_norm": 0.21725575625896454, + "learning_rate": 0.0004452805114913223, + "loss": 1.942, + "step": 2554 + }, + { + "epoch": 0.24951171875, + "grad_norm": 0.3064901828765869, + "learning_rate": 0.0004452340207306883, + "loss": 1.8555, + "step": 2555 + }, + { + "epoch": 0.249609375, + "grad_norm": 0.3561396300792694, + "learning_rate": 0.00044518751296518367, + "loss": 1.9103, + "step": 2556 + }, + { + "epoch": 0.24970703125, + "grad_norm": 0.22716249525547028, + "learning_rate": 0.0004451409881994543, + "loss": 1.8788, + "step": 2557 + }, + { + "epoch": 0.2498046875, + "grad_norm": 0.20097126066684723, + "learning_rate": 0.00044509444643814745, + "loss": 1.8081, + "step": 2558 + }, + { + "epoch": 0.24990234375, + "grad_norm": 0.2304898500442505, + "learning_rate": 0.0004450478876859123, + "loss": 1.8918, + "step": 2559 + }, + { + "epoch": 0.25, + "grad_norm": 0.22153149545192719, + "learning_rate": 0.00044500131194739976, + "loss": 1.8949, + "step": 2560 + }, + { + "epoch": 0.25009765625, + "grad_norm": 0.21728168427944183, + "learning_rate": 0.0004449547192272622, + "loss": 1.8991, + "step": 2561 + }, + { + "epoch": 0.2501953125, + "grad_norm": 0.191215381026268, + "learning_rate": 0.00044490810953015376, + "loss": 1.8753, + "step": 2562 + }, + { + "epoch": 0.25029296875, + "grad_norm": 0.20973442494869232, + "learning_rate": 0.00044486148286073047, + "loss": 1.8833, + "step": 2563 + }, + { + "epoch": 0.250390625, + "grad_norm": 0.16450902819633484, + "learning_rate": 0.00044481483922364986, + "loss": 1.8508, + "step": 2564 + }, + { + "epoch": 0.25048828125, + "grad_norm": 0.19612371921539307, + "learning_rate": 0.00044476817862357135, + "loss": 1.8916, + "step": 2565 + }, + { + "epoch": 0.2505859375, + "grad_norm": 0.17144112288951874, + "learning_rate": 0.00044472150106515565, + "loss": 1.9255, + "step": 2566 + }, + { + "epoch": 0.25068359375, + "grad_norm": 0.2133835256099701, + "learning_rate": 0.00044467480655306565, + "loss": 1.8678, + "step": 2567 + }, + { + "epoch": 0.25078125, + "grad_norm": 0.21594852209091187, + "learning_rate": 0.0004446280950919657, + "loss": 1.8686, + "step": 2568 + }, + { + "epoch": 0.25087890625, + "grad_norm": 0.24870210886001587, + "learning_rate": 0.00044458136668652185, + "loss": 1.9342, + "step": 2569 + }, + { + "epoch": 0.2509765625, + "grad_norm": 0.2506871223449707, + "learning_rate": 0.0004445346213414017, + "loss": 1.9142, + "step": 2570 + }, + { + "epoch": 0.25107421875, + "grad_norm": 0.21057821810245514, + "learning_rate": 0.0004444878590612749, + "loss": 1.8809, + "step": 2571 + }, + { + "epoch": 0.251171875, + "grad_norm": 0.2578639090061188, + "learning_rate": 0.0004444410798508125, + "loss": 1.8706, + "step": 2572 + }, + { + "epoch": 0.25126953125, + "grad_norm": 0.23633147776126862, + "learning_rate": 0.00044439428371468724, + "loss": 1.8912, + "step": 2573 + }, + { + "epoch": 0.2513671875, + "grad_norm": 0.222493976354599, + "learning_rate": 0.00044434747065757383, + "loss": 1.8508, + "step": 2574 + }, + { + "epoch": 0.25146484375, + "grad_norm": 0.2404196411371231, + "learning_rate": 0.00044430064068414843, + "loss": 1.8879, + "step": 2575 + }, + { + "epoch": 0.2515625, + "grad_norm": 0.2464357167482376, + "learning_rate": 0.0004442537937990889, + "loss": 1.9289, + "step": 2576 + }, + { + "epoch": 0.25166015625, + "grad_norm": 0.2768436372280121, + "learning_rate": 0.0004442069300070747, + "loss": 1.8553, + "step": 2577 + }, + { + "epoch": 0.2517578125, + "grad_norm": 0.22688689827919006, + "learning_rate": 0.0004441600493127873, + "loss": 1.8665, + "step": 2578 + }, + { + "epoch": 0.25185546875, + "grad_norm": 0.28142839670181274, + "learning_rate": 0.0004441131517209096, + "loss": 1.8982, + "step": 2579 + }, + { + "epoch": 0.251953125, + "grad_norm": 0.22705985605716705, + "learning_rate": 0.0004440662372361262, + "loss": 1.8778, + "step": 2580 + }, + { + "epoch": 0.25205078125, + "grad_norm": 0.24169547855854034, + "learning_rate": 0.0004440193058631236, + "loss": 1.9192, + "step": 2581 + }, + { + "epoch": 0.2521484375, + "grad_norm": 0.2988471984863281, + "learning_rate": 0.00044397235760658963, + "loss": 1.8754, + "step": 2582 + }, + { + "epoch": 0.25224609375, + "grad_norm": 0.343605101108551, + "learning_rate": 0.0004439253924712141, + "loss": 1.8933, + "step": 2583 + }, + { + "epoch": 0.25234375, + "grad_norm": 0.3238380551338196, + "learning_rate": 0.00044387841046168844, + "loss": 1.8798, + "step": 2584 + }, + { + "epoch": 0.25244140625, + "grad_norm": 0.21151965856552124, + "learning_rate": 0.0004438314115827057, + "loss": 1.88, + "step": 2585 + }, + { + "epoch": 0.2525390625, + "grad_norm": 0.3310568630695343, + "learning_rate": 0.00044378439583896064, + "loss": 1.9042, + "step": 2586 + }, + { + "epoch": 0.25263671875, + "grad_norm": 0.30944469571113586, + "learning_rate": 0.00044373736323514973, + "loss": 1.8837, + "step": 2587 + }, + { + "epoch": 0.252734375, + "grad_norm": 0.23433548212051392, + "learning_rate": 0.00044369031377597114, + "loss": 1.9025, + "step": 2588 + }, + { + "epoch": 0.25283203125, + "grad_norm": 0.2772672176361084, + "learning_rate": 0.0004436432474661246, + "loss": 1.8573, + "step": 2589 + }, + { + "epoch": 0.2529296875, + "grad_norm": 0.251200407743454, + "learning_rate": 0.00044359616431031176, + "loss": 1.869, + "step": 2590 + }, + { + "epoch": 0.25302734375, + "grad_norm": 0.28617560863494873, + "learning_rate": 0.0004435490643132357, + "loss": 1.91, + "step": 2591 + }, + { + "epoch": 0.253125, + "grad_norm": 0.23629052937030792, + "learning_rate": 0.0004435019474796013, + "loss": 1.9149, + "step": 2592 + }, + { + "epoch": 0.25322265625, + "grad_norm": 0.2794038653373718, + "learning_rate": 0.0004434548138141152, + "loss": 1.9415, + "step": 2593 + }, + { + "epoch": 0.2533203125, + "grad_norm": 0.22255775332450867, + "learning_rate": 0.0004434076633214855, + "loss": 1.8657, + "step": 2594 + }, + { + "epoch": 0.25341796875, + "grad_norm": 0.28047165274620056, + "learning_rate": 0.00044336049600642225, + "loss": 1.9193, + "step": 2595 + }, + { + "epoch": 0.253515625, + "grad_norm": 0.16849219799041748, + "learning_rate": 0.0004433133118736369, + "loss": 1.8809, + "step": 2596 + }, + { + "epoch": 0.25361328125, + "grad_norm": 0.26295021176338196, + "learning_rate": 0.00044326611092784297, + "loss": 1.8732, + "step": 2597 + }, + { + "epoch": 0.2537109375, + "grad_norm": 0.2265648990869522, + "learning_rate": 0.00044321889317375517, + "loss": 1.8861, + "step": 2598 + }, + { + "epoch": 0.25380859375, + "grad_norm": 0.28177958726882935, + "learning_rate": 0.0004431716586160902, + "loss": 1.9275, + "step": 2599 + }, + { + "epoch": 0.25390625, + "grad_norm": 0.34475746750831604, + "learning_rate": 0.0004431244072595665, + "loss": 1.8884, + "step": 2600 + }, + { + "epoch": 0.25400390625, + "grad_norm": 0.32186293601989746, + "learning_rate": 0.0004430771391089038, + "loss": 1.9329, + "step": 2601 + }, + { + "epoch": 0.2541015625, + "grad_norm": 0.2824218273162842, + "learning_rate": 0.00044302985416882405, + "loss": 1.9011, + "step": 2602 + }, + { + "epoch": 0.25419921875, + "grad_norm": 0.33732739090919495, + "learning_rate": 0.0004429825524440505, + "loss": 1.9148, + "step": 2603 + }, + { + "epoch": 0.254296875, + "grad_norm": 0.23914110660552979, + "learning_rate": 0.00044293523393930807, + "loss": 1.866, + "step": 2604 + }, + { + "epoch": 0.25439453125, + "grad_norm": 0.2791132926940918, + "learning_rate": 0.0004428878986593236, + "loss": 1.8849, + "step": 2605 + }, + { + "epoch": 0.2544921875, + "grad_norm": 0.20789825916290283, + "learning_rate": 0.0004428405466088253, + "loss": 1.9007, + "step": 2606 + }, + { + "epoch": 0.25458984375, + "grad_norm": 0.27687475085258484, + "learning_rate": 0.0004427931777925435, + "loss": 1.8936, + "step": 2607 + }, + { + "epoch": 0.2546875, + "grad_norm": 0.29951584339141846, + "learning_rate": 0.00044274579221520964, + "loss": 1.8735, + "step": 2608 + }, + { + "epoch": 0.25478515625, + "grad_norm": 0.25301870703697205, + "learning_rate": 0.00044269838988155726, + "loss": 1.8829, + "step": 2609 + }, + { + "epoch": 0.2548828125, + "grad_norm": 0.254138320684433, + "learning_rate": 0.0004426509707963214, + "loss": 1.8881, + "step": 2610 + }, + { + "epoch": 0.25498046875, + "grad_norm": 0.2853766679763794, + "learning_rate": 0.00044260353496423883, + "loss": 1.9376, + "step": 2611 + }, + { + "epoch": 0.255078125, + "grad_norm": 0.22820377349853516, + "learning_rate": 0.00044255608239004795, + "loss": 1.8961, + "step": 2612 + }, + { + "epoch": 0.25517578125, + "grad_norm": 0.269482284784317, + "learning_rate": 0.00044250861307848884, + "loss": 1.8892, + "step": 2613 + }, + { + "epoch": 0.2552734375, + "grad_norm": 0.21227295696735382, + "learning_rate": 0.0004424611270343033, + "loss": 1.9393, + "step": 2614 + }, + { + "epoch": 0.25537109375, + "grad_norm": 0.21992765367031097, + "learning_rate": 0.00044241362426223463, + "loss": 1.8699, + "step": 2615 + }, + { + "epoch": 0.25546875, + "grad_norm": 0.26295095682144165, + "learning_rate": 0.0004423661047670282, + "loss": 1.9119, + "step": 2616 + }, + { + "epoch": 0.25556640625, + "grad_norm": 0.1936780959367752, + "learning_rate": 0.00044231856855343055, + "loss": 1.8939, + "step": 2617 + }, + { + "epoch": 0.2556640625, + "grad_norm": 0.27874645590782166, + "learning_rate": 0.0004422710156261903, + "loss": 1.9553, + "step": 2618 + }, + { + "epoch": 0.25576171875, + "grad_norm": 0.2704470753669739, + "learning_rate": 0.0004422234459900574, + "loss": 1.8742, + "step": 2619 + }, + { + "epoch": 0.255859375, + "grad_norm": 0.29514744877815247, + "learning_rate": 0.00044217585964978365, + "loss": 1.8803, + "step": 2620 + }, + { + "epoch": 0.25595703125, + "grad_norm": 0.22899918258190155, + "learning_rate": 0.0004421282566101226, + "loss": 1.8898, + "step": 2621 + }, + { + "epoch": 0.2560546875, + "grad_norm": 0.3023613691329956, + "learning_rate": 0.00044208063687582944, + "loss": 1.917, + "step": 2622 + }, + { + "epoch": 0.25615234375, + "grad_norm": 0.2797189950942993, + "learning_rate": 0.0004420330004516608, + "loss": 1.89, + "step": 2623 + }, + { + "epoch": 0.25625, + "grad_norm": 0.23178371787071228, + "learning_rate": 0.0004419853473423751, + "loss": 1.8528, + "step": 2624 + }, + { + "epoch": 0.25634765625, + "grad_norm": 0.2943453788757324, + "learning_rate": 0.0004419376775527326, + "loss": 1.8675, + "step": 2625 + }, + { + "epoch": 0.2564453125, + "grad_norm": 0.2193489670753479, + "learning_rate": 0.00044188999108749516, + "loss": 1.8843, + "step": 2626 + }, + { + "epoch": 0.25654296875, + "grad_norm": 0.2941931486129761, + "learning_rate": 0.00044184228795142607, + "loss": 1.8976, + "step": 2627 + }, + { + "epoch": 0.256640625, + "grad_norm": 0.22620061039924622, + "learning_rate": 0.00044179456814929044, + "loss": 1.8674, + "step": 2628 + }, + { + "epoch": 0.25673828125, + "grad_norm": 0.2290557473897934, + "learning_rate": 0.0004417468316858552, + "loss": 1.9054, + "step": 2629 + }, + { + "epoch": 0.2568359375, + "grad_norm": 0.24180883169174194, + "learning_rate": 0.00044169907856588865, + "loss": 1.8484, + "step": 2630 + }, + { + "epoch": 0.25693359375, + "grad_norm": 0.19765223562717438, + "learning_rate": 0.00044165130879416097, + "loss": 1.8613, + "step": 2631 + }, + { + "epoch": 0.25703125, + "grad_norm": 0.20855452120304108, + "learning_rate": 0.00044160352237544406, + "loss": 1.8621, + "step": 2632 + }, + { + "epoch": 0.25712890625, + "grad_norm": 0.23405295610427856, + "learning_rate": 0.00044155571931451105, + "loss": 1.8958, + "step": 2633 + }, + { + "epoch": 0.2572265625, + "grad_norm": 0.27464863657951355, + "learning_rate": 0.0004415078996161374, + "loss": 1.8959, + "step": 2634 + }, + { + "epoch": 0.25732421875, + "grad_norm": 0.2642795145511627, + "learning_rate": 0.00044146006328509956, + "loss": 1.8968, + "step": 2635 + }, + { + "epoch": 0.257421875, + "grad_norm": 0.29816219210624695, + "learning_rate": 0.00044141221032617626, + "loss": 1.8845, + "step": 2636 + }, + { + "epoch": 0.25751953125, + "grad_norm": 0.251708984375, + "learning_rate": 0.0004413643407441473, + "loss": 1.9025, + "step": 2637 + }, + { + "epoch": 0.2576171875, + "grad_norm": 0.28949257731437683, + "learning_rate": 0.0004413164545437946, + "loss": 1.8946, + "step": 2638 + }, + { + "epoch": 0.25771484375, + "grad_norm": 0.24704940617084503, + "learning_rate": 0.0004412685517299015, + "loss": 1.8745, + "step": 2639 + }, + { + "epoch": 0.2578125, + "grad_norm": 0.25325044989585876, + "learning_rate": 0.0004412206323072532, + "loss": 1.8796, + "step": 2640 + }, + { + "epoch": 0.25791015625, + "grad_norm": 0.22802412509918213, + "learning_rate": 0.00044117269628063616, + "loss": 1.909, + "step": 2641 + }, + { + "epoch": 0.2580078125, + "grad_norm": 0.21644683182239532, + "learning_rate": 0.00044112474365483903, + "loss": 1.846, + "step": 2642 + }, + { + "epoch": 0.25810546875, + "grad_norm": 0.20541459321975708, + "learning_rate": 0.00044107677443465165, + "loss": 1.8854, + "step": 2643 + }, + { + "epoch": 0.258203125, + "grad_norm": 0.2531701326370239, + "learning_rate": 0.00044102878862486587, + "loss": 1.9159, + "step": 2644 + }, + { + "epoch": 0.25830078125, + "grad_norm": 0.20743289589881897, + "learning_rate": 0.00044098078623027495, + "loss": 1.8419, + "step": 2645 + }, + { + "epoch": 0.2583984375, + "grad_norm": 0.23902501165866852, + "learning_rate": 0.00044093276725567403, + "loss": 1.883, + "step": 2646 + }, + { + "epoch": 0.25849609375, + "grad_norm": 0.260051965713501, + "learning_rate": 0.00044088473170585964, + "loss": 1.9066, + "step": 2647 + }, + { + "epoch": 0.25859375, + "grad_norm": 0.22198320925235748, + "learning_rate": 0.00044083667958563007, + "loss": 1.8814, + "step": 2648 + }, + { + "epoch": 0.25869140625, + "grad_norm": 0.20246773958206177, + "learning_rate": 0.0004407886108997855, + "loss": 1.923, + "step": 2649 + }, + { + "epoch": 0.2587890625, + "grad_norm": 0.26774612069129944, + "learning_rate": 0.0004407405256531274, + "loss": 1.8811, + "step": 2650 + }, + { + "epoch": 0.25888671875, + "grad_norm": 0.2821747064590454, + "learning_rate": 0.0004406924238504592, + "loss": 1.9521, + "step": 2651 + }, + { + "epoch": 0.258984375, + "grad_norm": 0.2747217118740082, + "learning_rate": 0.0004406443054965857, + "loss": 1.863, + "step": 2652 + }, + { + "epoch": 0.25908203125, + "grad_norm": 0.24007029831409454, + "learning_rate": 0.0004405961705963135, + "loss": 1.8777, + "step": 2653 + }, + { + "epoch": 0.2591796875, + "grad_norm": 0.2279479056596756, + "learning_rate": 0.000440548019154451, + "loss": 1.8902, + "step": 2654 + }, + { + "epoch": 0.25927734375, + "grad_norm": 0.24569466710090637, + "learning_rate": 0.00044049985117580796, + "loss": 1.8834, + "step": 2655 + }, + { + "epoch": 0.259375, + "grad_norm": 0.22550983726978302, + "learning_rate": 0.00044045166666519604, + "loss": 1.8953, + "step": 2656 + }, + { + "epoch": 0.25947265625, + "grad_norm": 0.2546038031578064, + "learning_rate": 0.00044040346562742834, + "loss": 1.8834, + "step": 2657 + }, + { + "epoch": 0.2595703125, + "grad_norm": 0.21834610402584076, + "learning_rate": 0.00044035524806731975, + "loss": 1.8543, + "step": 2658 + }, + { + "epoch": 0.25966796875, + "grad_norm": 0.22710272669792175, + "learning_rate": 0.0004403070139896869, + "loss": 1.8995, + "step": 2659 + }, + { + "epoch": 0.259765625, + "grad_norm": 0.19748175144195557, + "learning_rate": 0.0004402587633993477, + "loss": 1.864, + "step": 2660 + }, + { + "epoch": 0.25986328125, + "grad_norm": 0.23321975767612457, + "learning_rate": 0.0004402104963011222, + "loss": 1.9256, + "step": 2661 + }, + { + "epoch": 0.2599609375, + "grad_norm": 0.19426730275154114, + "learning_rate": 0.0004401622126998317, + "loss": 1.8956, + "step": 2662 + }, + { + "epoch": 0.26005859375, + "grad_norm": 0.2513621747493744, + "learning_rate": 0.0004401139126002993, + "loss": 1.9298, + "step": 2663 + }, + { + "epoch": 0.26015625, + "grad_norm": 0.2774468660354614, + "learning_rate": 0.00044006559600734993, + "loss": 1.8946, + "step": 2664 + }, + { + "epoch": 0.26025390625, + "grad_norm": 0.2116554081439972, + "learning_rate": 0.00044001726292580977, + "loss": 1.8713, + "step": 2665 + }, + { + "epoch": 0.2603515625, + "grad_norm": 0.24314185976982117, + "learning_rate": 0.000439968913360507, + "loss": 1.8807, + "step": 2666 + }, + { + "epoch": 0.26044921875, + "grad_norm": 0.22041741013526917, + "learning_rate": 0.00043992054731627135, + "loss": 1.9022, + "step": 2667 + }, + { + "epoch": 0.260546875, + "grad_norm": 0.214686781167984, + "learning_rate": 0.00043987216479793404, + "loss": 1.8687, + "step": 2668 + }, + { + "epoch": 0.26064453125, + "grad_norm": 0.254207581281662, + "learning_rate": 0.0004398237658103281, + "loss": 1.8489, + "step": 2669 + }, + { + "epoch": 0.2607421875, + "grad_norm": 0.21913522481918335, + "learning_rate": 0.0004397753503582881, + "loss": 1.9079, + "step": 2670 + }, + { + "epoch": 0.26083984375, + "grad_norm": 0.2150515615940094, + "learning_rate": 0.0004397269184466505, + "loss": 1.8985, + "step": 2671 + }, + { + "epoch": 0.2609375, + "grad_norm": 0.26243162155151367, + "learning_rate": 0.000439678470080253, + "loss": 1.8512, + "step": 2672 + }, + { + "epoch": 0.26103515625, + "grad_norm": 0.20125195384025574, + "learning_rate": 0.0004396300052639353, + "loss": 1.8621, + "step": 2673 + }, + { + "epoch": 0.2611328125, + "grad_norm": 0.2900947630405426, + "learning_rate": 0.00043958152400253855, + "loss": 1.9354, + "step": 2674 + }, + { + "epoch": 0.26123046875, + "grad_norm": 0.2357524037361145, + "learning_rate": 0.00043953302630090566, + "loss": 1.9136, + "step": 2675 + }, + { + "epoch": 0.261328125, + "grad_norm": 0.23963971436023712, + "learning_rate": 0.000439484512163881, + "loss": 1.867, + "step": 2676 + }, + { + "epoch": 0.26142578125, + "grad_norm": 0.27860602736473083, + "learning_rate": 0.00043943598159631075, + "loss": 1.9149, + "step": 2677 + }, + { + "epoch": 0.2615234375, + "grad_norm": 0.22854892909526825, + "learning_rate": 0.00043938743460304275, + "loss": 1.8415, + "step": 2678 + }, + { + "epoch": 0.26162109375, + "grad_norm": 0.27040791511535645, + "learning_rate": 0.0004393388711889264, + "loss": 1.9017, + "step": 2679 + }, + { + "epoch": 0.26171875, + "grad_norm": 0.2952374219894409, + "learning_rate": 0.0004392902913588128, + "loss": 1.9406, + "step": 2680 + }, + { + "epoch": 0.26181640625, + "grad_norm": 0.27148759365081787, + "learning_rate": 0.0004392416951175544, + "loss": 1.8999, + "step": 2681 + }, + { + "epoch": 0.2619140625, + "grad_norm": 0.24533145129680634, + "learning_rate": 0.00043919308247000575, + "loss": 1.8835, + "step": 2682 + }, + { + "epoch": 0.26201171875, + "grad_norm": 0.2174772322177887, + "learning_rate": 0.0004391444534210229, + "loss": 1.8424, + "step": 2683 + }, + { + "epoch": 0.262109375, + "grad_norm": 0.19030643999576569, + "learning_rate": 0.0004390958079754633, + "loss": 1.8736, + "step": 2684 + }, + { + "epoch": 0.26220703125, + "grad_norm": 0.26530101895332336, + "learning_rate": 0.0004390471461381862, + "loss": 1.8988, + "step": 2685 + }, + { + "epoch": 0.2623046875, + "grad_norm": 0.24065977334976196, + "learning_rate": 0.0004389984679140525, + "loss": 1.9052, + "step": 2686 + }, + { + "epoch": 0.26240234375, + "grad_norm": 0.2812477946281433, + "learning_rate": 0.00043894977330792486, + "loss": 1.8987, + "step": 2687 + }, + { + "epoch": 0.2625, + "grad_norm": 0.2505251169204712, + "learning_rate": 0.00043890106232466724, + "loss": 1.8922, + "step": 2688 + }, + { + "epoch": 0.26259765625, + "grad_norm": 0.2238491326570511, + "learning_rate": 0.0004388523349691455, + "loss": 1.9183, + "step": 2689 + }, + { + "epoch": 0.2626953125, + "grad_norm": 0.26334211230278015, + "learning_rate": 0.00043880359124622714, + "loss": 1.8899, + "step": 2690 + }, + { + "epoch": 0.26279296875, + "grad_norm": 0.24818876385688782, + "learning_rate": 0.0004387548311607812, + "loss": 1.9063, + "step": 2691 + }, + { + "epoch": 0.262890625, + "grad_norm": 0.22725768387317657, + "learning_rate": 0.0004387060547176784, + "loss": 1.8828, + "step": 2692 + }, + { + "epoch": 0.26298828125, + "grad_norm": 0.2095717042684555, + "learning_rate": 0.00043865726192179096, + "loss": 1.8658, + "step": 2693 + }, + { + "epoch": 0.2630859375, + "grad_norm": 0.2227015644311905, + "learning_rate": 0.000438608452777993, + "loss": 1.9546, + "step": 2694 + }, + { + "epoch": 0.26318359375, + "grad_norm": 0.20986926555633545, + "learning_rate": 0.00043855962729116, + "loss": 1.878, + "step": 2695 + }, + { + "epoch": 0.26328125, + "grad_norm": 0.24293240904808044, + "learning_rate": 0.00043851078546616924, + "loss": 1.889, + "step": 2696 + }, + { + "epoch": 0.26337890625, + "grad_norm": 0.23340186476707458, + "learning_rate": 0.0004384619273078996, + "loss": 1.9285, + "step": 2697 + }, + { + "epoch": 0.2634765625, + "grad_norm": 0.22993077337741852, + "learning_rate": 0.00043841305282123157, + "loss": 1.9004, + "step": 2698 + }, + { + "epoch": 0.26357421875, + "grad_norm": 0.25231555104255676, + "learning_rate": 0.00043836416201104727, + "loss": 1.886, + "step": 2699 + }, + { + "epoch": 0.263671875, + "grad_norm": 0.18800829350948334, + "learning_rate": 0.0004383152548822304, + "loss": 1.9054, + "step": 2700 + }, + { + "epoch": 0.26376953125, + "grad_norm": 0.31697878241539, + "learning_rate": 0.0004382663314396665, + "loss": 1.8387, + "step": 2701 + }, + { + "epoch": 0.2638671875, + "grad_norm": 0.31109264492988586, + "learning_rate": 0.00043821739168824246, + "loss": 1.922, + "step": 2702 + }, + { + "epoch": 0.26396484375, + "grad_norm": 0.23077085614204407, + "learning_rate": 0.00043816843563284697, + "loss": 1.8771, + "step": 2703 + }, + { + "epoch": 0.2640625, + "grad_norm": 0.31502625346183777, + "learning_rate": 0.0004381194632783703, + "loss": 1.8655, + "step": 2704 + }, + { + "epoch": 0.26416015625, + "grad_norm": 0.24782267212867737, + "learning_rate": 0.00043807047462970434, + "loss": 1.8778, + "step": 2705 + }, + { + "epoch": 0.2642578125, + "grad_norm": 0.25926321744918823, + "learning_rate": 0.00043802146969174267, + "loss": 1.8437, + "step": 2706 + }, + { + "epoch": 0.26435546875, + "grad_norm": 0.27319055795669556, + "learning_rate": 0.0004379724484693804, + "loss": 1.8433, + "step": 2707 + }, + { + "epoch": 0.264453125, + "grad_norm": 0.23524881899356842, + "learning_rate": 0.0004379234109675143, + "loss": 1.9148, + "step": 2708 + }, + { + "epoch": 0.26455078125, + "grad_norm": 0.28222718834877014, + "learning_rate": 0.0004378743571910429, + "loss": 1.9035, + "step": 2709 + }, + { + "epoch": 0.2646484375, + "grad_norm": 0.271083801984787, + "learning_rate": 0.00043782528714486613, + "loss": 1.8691, + "step": 2710 + }, + { + "epoch": 0.26474609375, + "grad_norm": 0.23110386729240417, + "learning_rate": 0.0004377762008338856, + "loss": 1.8794, + "step": 2711 + }, + { + "epoch": 0.26484375, + "grad_norm": 0.24147644639015198, + "learning_rate": 0.0004377270982630048, + "loss": 1.8577, + "step": 2712 + }, + { + "epoch": 0.26494140625, + "grad_norm": 0.23246973752975464, + "learning_rate": 0.0004376779794371284, + "loss": 1.8876, + "step": 2713 + }, + { + "epoch": 0.2650390625, + "grad_norm": 0.23443640768527985, + "learning_rate": 0.00043762884436116315, + "loss": 1.8693, + "step": 2714 + }, + { + "epoch": 0.26513671875, + "grad_norm": 0.2223389744758606, + "learning_rate": 0.00043757969304001704, + "loss": 1.8308, + "step": 2715 + }, + { + "epoch": 0.265234375, + "grad_norm": 0.25246211886405945, + "learning_rate": 0.0004375305254785999, + "loss": 1.8723, + "step": 2716 + }, + { + "epoch": 0.26533203125, + "grad_norm": 0.2547222077846527, + "learning_rate": 0.0004374813416818232, + "loss": 1.8826, + "step": 2717 + }, + { + "epoch": 0.2654296875, + "grad_norm": 0.2161521464586258, + "learning_rate": 0.0004374321416545999, + "loss": 1.9017, + "step": 2718 + }, + { + "epoch": 0.26552734375, + "grad_norm": 0.2493196278810501, + "learning_rate": 0.0004373829254018447, + "loss": 1.86, + "step": 2719 + }, + { + "epoch": 0.265625, + "grad_norm": 0.2375672310590744, + "learning_rate": 0.00043733369292847386, + "loss": 1.8784, + "step": 2720 + }, + { + "epoch": 0.26572265625, + "grad_norm": 0.34022462368011475, + "learning_rate": 0.00043728444423940516, + "loss": 1.8915, + "step": 2721 + }, + { + "epoch": 0.2658203125, + "grad_norm": 0.25120794773101807, + "learning_rate": 0.0004372351793395582, + "loss": 1.8894, + "step": 2722 + }, + { + "epoch": 0.26591796875, + "grad_norm": 0.2032475620508194, + "learning_rate": 0.0004371858982338542, + "loss": 1.8942, + "step": 2723 + }, + { + "epoch": 0.266015625, + "grad_norm": 0.2685857117176056, + "learning_rate": 0.00043713660092721573, + "loss": 1.9001, + "step": 2724 + }, + { + "epoch": 0.26611328125, + "grad_norm": 0.296835333108902, + "learning_rate": 0.00043708728742456723, + "loss": 1.8994, + "step": 2725 + }, + { + "epoch": 0.2662109375, + "grad_norm": 0.27104151248931885, + "learning_rate": 0.00043703795773083467, + "loss": 1.9119, + "step": 2726 + }, + { + "epoch": 0.26630859375, + "grad_norm": 0.25494450330734253, + "learning_rate": 0.0004369886118509457, + "loss": 1.9188, + "step": 2727 + }, + { + "epoch": 0.26640625, + "grad_norm": 0.3688213527202606, + "learning_rate": 0.0004369392497898294, + "loss": 1.9021, + "step": 2728 + }, + { + "epoch": 0.26650390625, + "grad_norm": 0.24721217155456543, + "learning_rate": 0.00043688987155241676, + "loss": 1.8895, + "step": 2729 + }, + { + "epoch": 0.2666015625, + "grad_norm": 0.290102481842041, + "learning_rate": 0.0004368404771436402, + "loss": 1.8565, + "step": 2730 + }, + { + "epoch": 0.26669921875, + "grad_norm": 0.3468306064605713, + "learning_rate": 0.0004367910665684338, + "loss": 1.8795, + "step": 2731 + }, + { + "epoch": 0.266796875, + "grad_norm": 0.23682735860347748, + "learning_rate": 0.00043674163983173304, + "loss": 1.8749, + "step": 2732 + }, + { + "epoch": 0.26689453125, + "grad_norm": 0.24279217422008514, + "learning_rate": 0.00043669219693847543, + "loss": 1.887, + "step": 2733 + }, + { + "epoch": 0.2669921875, + "grad_norm": 0.25867339968681335, + "learning_rate": 0.0004366427378935998, + "loss": 1.8708, + "step": 2734 + }, + { + "epoch": 0.26708984375, + "grad_norm": 0.28321221470832825, + "learning_rate": 0.0004365932627020467, + "loss": 1.9077, + "step": 2735 + }, + { + "epoch": 0.2671875, + "grad_norm": 0.2243577092885971, + "learning_rate": 0.00043654377136875827, + "loss": 1.8902, + "step": 2736 + }, + { + "epoch": 0.26728515625, + "grad_norm": 0.27794283628463745, + "learning_rate": 0.0004364942638986782, + "loss": 1.8659, + "step": 2737 + }, + { + "epoch": 0.2673828125, + "grad_norm": 0.2588692903518677, + "learning_rate": 0.00043644474029675193, + "loss": 1.8494, + "step": 2738 + }, + { + "epoch": 0.26748046875, + "grad_norm": 0.22354848682880402, + "learning_rate": 0.00043639520056792627, + "loss": 1.895, + "step": 2739 + }, + { + "epoch": 0.267578125, + "grad_norm": 0.3032642900943756, + "learning_rate": 0.00043634564471715, + "loss": 1.8643, + "step": 2740 + }, + { + "epoch": 0.26767578125, + "grad_norm": 0.25967633724212646, + "learning_rate": 0.0004362960727493732, + "loss": 1.8712, + "step": 2741 + }, + { + "epoch": 0.2677734375, + "grad_norm": 0.2637724280357361, + "learning_rate": 0.0004362464846695476, + "loss": 1.8512, + "step": 2742 + }, + { + "epoch": 0.26787109375, + "grad_norm": 0.2793140709400177, + "learning_rate": 0.00043619688048262677, + "loss": 1.9086, + "step": 2743 + }, + { + "epoch": 0.26796875, + "grad_norm": 0.2188546061515808, + "learning_rate": 0.00043614726019356565, + "loss": 1.8649, + "step": 2744 + }, + { + "epoch": 0.26806640625, + "grad_norm": 0.21244989335536957, + "learning_rate": 0.00043609762380732083, + "loss": 1.8744, + "step": 2745 + }, + { + "epoch": 0.2681640625, + "grad_norm": 0.2601993978023529, + "learning_rate": 0.00043604797132885064, + "loss": 1.8649, + "step": 2746 + }, + { + "epoch": 0.26826171875, + "grad_norm": 0.2278032749891281, + "learning_rate": 0.00043599830276311487, + "loss": 1.8662, + "step": 2747 + }, + { + "epoch": 0.268359375, + "grad_norm": 0.24005573987960815, + "learning_rate": 0.000435948618115075, + "loss": 1.8916, + "step": 2748 + }, + { + "epoch": 0.26845703125, + "grad_norm": 0.28929322957992554, + "learning_rate": 0.00043589891738969393, + "loss": 1.9566, + "step": 2749 + }, + { + "epoch": 0.2685546875, + "grad_norm": 0.21914543211460114, + "learning_rate": 0.00043584920059193654, + "loss": 1.9043, + "step": 2750 + }, + { + "epoch": 0.26865234375, + "grad_norm": 0.24009595811367035, + "learning_rate": 0.000435799467726769, + "loss": 1.8493, + "step": 2751 + }, + { + "epoch": 0.26875, + "grad_norm": 0.20415683090686798, + "learning_rate": 0.0004357497187991593, + "loss": 1.8954, + "step": 2752 + }, + { + "epoch": 0.26884765625, + "grad_norm": 0.1941215842962265, + "learning_rate": 0.0004356999538140766, + "loss": 1.8634, + "step": 2753 + }, + { + "epoch": 0.2689453125, + "grad_norm": 0.2202758938074112, + "learning_rate": 0.0004356501727764923, + "loss": 1.8706, + "step": 2754 + }, + { + "epoch": 0.26904296875, + "grad_norm": 0.21078895032405853, + "learning_rate": 0.000435600375691379, + "loss": 1.8942, + "step": 2755 + }, + { + "epoch": 0.269140625, + "grad_norm": 0.21176692843437195, + "learning_rate": 0.00043555056256371083, + "loss": 1.8814, + "step": 2756 + }, + { + "epoch": 0.26923828125, + "grad_norm": 0.20595595240592957, + "learning_rate": 0.00043550073339846394, + "loss": 1.8919, + "step": 2757 + }, + { + "epoch": 0.2693359375, + "grad_norm": 0.2053648680448532, + "learning_rate": 0.00043545088820061564, + "loss": 1.9387, + "step": 2758 + }, + { + "epoch": 0.26943359375, + "grad_norm": 0.2367970496416092, + "learning_rate": 0.00043540102697514513, + "loss": 1.8863, + "step": 2759 + }, + { + "epoch": 0.26953125, + "grad_norm": 0.1842896044254303, + "learning_rate": 0.0004353511497270329, + "loss": 1.8805, + "step": 2760 + }, + { + "epoch": 0.26962890625, + "grad_norm": 0.25769177079200745, + "learning_rate": 0.00043530125646126154, + "loss": 1.8833, + "step": 2761 + }, + { + "epoch": 0.2697265625, + "grad_norm": 0.2760721743106842, + "learning_rate": 0.0004352513471828148, + "loss": 1.8924, + "step": 2762 + }, + { + "epoch": 0.26982421875, + "grad_norm": 0.17177268862724304, + "learning_rate": 0.0004352014218966781, + "loss": 1.8441, + "step": 2763 + }, + { + "epoch": 0.269921875, + "grad_norm": 0.22797243297100067, + "learning_rate": 0.0004351514806078387, + "loss": 1.8753, + "step": 2764 + }, + { + "epoch": 0.27001953125, + "grad_norm": 0.22956639528274536, + "learning_rate": 0.0004351015233212851, + "loss": 1.9303, + "step": 2765 + }, + { + "epoch": 0.2701171875, + "grad_norm": 0.20838238298892975, + "learning_rate": 0.0004350515500420077, + "loss": 1.8689, + "step": 2766 + }, + { + "epoch": 0.27021484375, + "grad_norm": 0.2128564864397049, + "learning_rate": 0.00043500156077499834, + "loss": 1.9195, + "step": 2767 + }, + { + "epoch": 0.2703125, + "grad_norm": 0.21279101073741913, + "learning_rate": 0.00043495155552525074, + "loss": 1.8944, + "step": 2768 + }, + { + "epoch": 0.27041015625, + "grad_norm": 0.22925125062465668, + "learning_rate": 0.00043490153429775964, + "loss": 1.8632, + "step": 2769 + }, + { + "epoch": 0.2705078125, + "grad_norm": 0.23544587194919586, + "learning_rate": 0.00043485149709752193, + "loss": 1.9066, + "step": 2770 + }, + { + "epoch": 0.27060546875, + "grad_norm": 0.22419407963752747, + "learning_rate": 0.00043480144392953567, + "loss": 1.8416, + "step": 2771 + }, + { + "epoch": 0.270703125, + "grad_norm": 0.21770869195461273, + "learning_rate": 0.00043475137479880103, + "loss": 1.9117, + "step": 2772 + }, + { + "epoch": 0.27080078125, + "grad_norm": 0.2358229011297226, + "learning_rate": 0.0004347012897103192, + "loss": 1.9187, + "step": 2773 + }, + { + "epoch": 0.2708984375, + "grad_norm": 0.30673930048942566, + "learning_rate": 0.00043465118866909346, + "loss": 1.8928, + "step": 2774 + }, + { + "epoch": 0.27099609375, + "grad_norm": 0.21168282628059387, + "learning_rate": 0.00043460107168012837, + "loss": 1.9002, + "step": 2775 + }, + { + "epoch": 0.27109375, + "grad_norm": 0.22865557670593262, + "learning_rate": 0.0004345509387484301, + "loss": 1.8956, + "step": 2776 + }, + { + "epoch": 0.27119140625, + "grad_norm": 0.3253214955329895, + "learning_rate": 0.00043450078987900654, + "loss": 1.9049, + "step": 2777 + }, + { + "epoch": 0.2712890625, + "grad_norm": 0.2774486541748047, + "learning_rate": 0.00043445062507686713, + "loss": 1.8817, + "step": 2778 + }, + { + "epoch": 0.27138671875, + "grad_norm": 0.22928157448768616, + "learning_rate": 0.00043440044434702287, + "loss": 1.9167, + "step": 2779 + }, + { + "epoch": 0.271484375, + "grad_norm": 0.2905370891094208, + "learning_rate": 0.0004343502476944864, + "loss": 1.8434, + "step": 2780 + }, + { + "epoch": 0.27158203125, + "grad_norm": 0.19285933673381805, + "learning_rate": 0.0004343000351242719, + "loss": 1.8287, + "step": 2781 + }, + { + "epoch": 0.2716796875, + "grad_norm": 0.25820720195770264, + "learning_rate": 0.0004342498066413951, + "loss": 1.8488, + "step": 2782 + }, + { + "epoch": 0.27177734375, + "grad_norm": 0.2616766095161438, + "learning_rate": 0.00043419956225087354, + "loss": 1.8627, + "step": 2783 + }, + { + "epoch": 0.271875, + "grad_norm": 0.18560141324996948, + "learning_rate": 0.000434149301957726, + "loss": 1.8272, + "step": 2784 + }, + { + "epoch": 0.27197265625, + "grad_norm": 0.3137165307998657, + "learning_rate": 0.0004340990257669732, + "loss": 1.8952, + "step": 2785 + }, + { + "epoch": 0.2720703125, + "grad_norm": 0.231571763753891, + "learning_rate": 0.00043404873368363726, + "loss": 1.9044, + "step": 2786 + }, + { + "epoch": 0.27216796875, + "grad_norm": 0.2827242314815521, + "learning_rate": 0.00043399842571274175, + "loss": 1.8984, + "step": 2787 + }, + { + "epoch": 0.272265625, + "grad_norm": 0.2350980043411255, + "learning_rate": 0.0004339481018593121, + "loss": 1.8902, + "step": 2788 + }, + { + "epoch": 0.27236328125, + "grad_norm": 0.2104049175977707, + "learning_rate": 0.00043389776212837537, + "loss": 1.8853, + "step": 2789 + }, + { + "epoch": 0.2724609375, + "grad_norm": 0.3253656327724457, + "learning_rate": 0.0004338474065249598, + "loss": 1.8982, + "step": 2790 + }, + { + "epoch": 0.27255859375, + "grad_norm": 0.2520630359649658, + "learning_rate": 0.0004337970350540956, + "loss": 1.8806, + "step": 2791 + }, + { + "epoch": 0.27265625, + "grad_norm": 0.2999001145362854, + "learning_rate": 0.0004337466477208145, + "loss": 1.9165, + "step": 2792 + }, + { + "epoch": 0.27275390625, + "grad_norm": 0.20815780758857727, + "learning_rate": 0.0004336962445301495, + "loss": 1.9357, + "step": 2793 + }, + { + "epoch": 0.2728515625, + "grad_norm": 0.24039196968078613, + "learning_rate": 0.0004336458254871357, + "loss": 1.84, + "step": 2794 + }, + { + "epoch": 0.27294921875, + "grad_norm": 0.24862824380397797, + "learning_rate": 0.0004335953905968094, + "loss": 1.8991, + "step": 2795 + }, + { + "epoch": 0.273046875, + "grad_norm": 0.2109254151582718, + "learning_rate": 0.0004335449398642086, + "loss": 1.9077, + "step": 2796 + }, + { + "epoch": 0.27314453125, + "grad_norm": 0.27603134512901306, + "learning_rate": 0.00043349447329437286, + "loss": 1.8703, + "step": 2797 + }, + { + "epoch": 0.2732421875, + "grad_norm": 0.20653647184371948, + "learning_rate": 0.00043344399089234343, + "loss": 1.9106, + "step": 2798 + }, + { + "epoch": 0.27333984375, + "grad_norm": 0.26815375685691833, + "learning_rate": 0.000433393492663163, + "loss": 1.8934, + "step": 2799 + }, + { + "epoch": 0.2734375, + "grad_norm": 0.2597907781600952, + "learning_rate": 0.0004333429786118758, + "loss": 1.8716, + "step": 2800 + } + ], + "logging_steps": 1.0, + "max_steps": 10240, + "num_input_tokens_seen": 0, + "num_train_epochs": 9223372036854775807, + "save_steps": 200, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.7143016646613402e+19, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +}