diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,25245 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.9835860342189457, + "eval_steps": 899, + "global_step": 3596, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005564056196967589, + "grad_norm": 1.3203125, + "learning_rate": 2e-05, + "loss": 1.7806, + "step": 1 + }, + { + "epoch": 0.0005564056196967589, + "eval_loss": 1.7791756391525269, + "eval_runtime": 1750.0366, + "eval_samples_per_second": 2.681, + "eval_steps_per_second": 1.341, + "step": 1 + }, + { + "epoch": 0.0011128112393935178, + "grad_norm": 1.1875, + "learning_rate": 4e-05, + "loss": 1.7467, + "step": 2 + }, + { + "epoch": 0.0016692168590902768, + "grad_norm": 1.1484375, + "learning_rate": 6e-05, + "loss": 1.6933, + "step": 3 + }, + { + "epoch": 0.0022256224787870356, + "grad_norm": 1.0390625, + "learning_rate": 8e-05, + "loss": 1.6916, + "step": 4 + }, + { + "epoch": 0.0027820280984837947, + "grad_norm": 1.078125, + "learning_rate": 0.0001, + "loss": 1.6973, + "step": 5 + }, + { + "epoch": 0.0033384337181805537, + "grad_norm": 0.9140625, + "learning_rate": 0.00012, + "loss": 1.6518, + "step": 6 + }, + { + "epoch": 0.0038948393378773127, + "grad_norm": 0.796875, + "learning_rate": 0.00014, + "loss": 1.6383, + "step": 7 + }, + { + "epoch": 0.004451244957574071, + "grad_norm": 0.75, + "learning_rate": 0.00016, + "loss": 1.5714, + "step": 8 + }, + { + "epoch": 0.005007650577270831, + "grad_norm": 0.70703125, + "learning_rate": 0.00018, + "loss": 1.553, + "step": 9 + }, + { + "epoch": 0.005564056196967589, + "grad_norm": 0.734375, + "learning_rate": 0.0002, + "loss": 1.5321, + "step": 10 + }, + { + "epoch": 0.006120461816664348, + "grad_norm": 0.84765625, + "learning_rate": 0.0001999999829570899, + "loss": 1.5589, + "step": 11 + }, + { + "epoch": 0.006676867436361107, + "grad_norm": 0.72265625, + "learning_rate": 0.0001999999318283654, + "loss": 1.4519, + "step": 12 + }, + { + "epoch": 0.007233273056057866, + "grad_norm": 0.6640625, + "learning_rate": 0.0001999998466138439, + "loss": 1.4035, + "step": 13 + }, + { + "epoch": 0.007789678675754625, + "grad_norm": 0.68359375, + "learning_rate": 0.0001999997273135545, + "loss": 1.4071, + "step": 14 + }, + { + "epoch": 0.008346084295451384, + "grad_norm": 0.6796875, + "learning_rate": 0.0001999995739275378, + "loss": 1.2862, + "step": 15 + }, + { + "epoch": 0.008902489915148143, + "grad_norm": 0.609375, + "learning_rate": 0.00019999938645584616, + "loss": 1.2801, + "step": 16 + }, + { + "epoch": 0.009458895534844901, + "grad_norm": 0.5390625, + "learning_rate": 0.0001999991648985434, + "loss": 1.2358, + "step": 17 + }, + { + "epoch": 0.010015301154541661, + "grad_norm": 0.56640625, + "learning_rate": 0.0001999989092557051, + "loss": 1.2122, + "step": 18 + }, + { + "epoch": 0.01057170677423842, + "grad_norm": 0.56640625, + "learning_rate": 0.00019999861952741839, + "loss": 1.2294, + "step": 19 + }, + { + "epoch": 0.011128112393935179, + "grad_norm": 0.53125, + "learning_rate": 0.00019999829571378197, + "loss": 1.2641, + "step": 20 + }, + { + "epoch": 0.011684518013631937, + "grad_norm": 0.482421875, + "learning_rate": 0.0001999979378149063, + "loss": 1.2373, + "step": 21 + }, + { + "epoch": 0.012240923633328696, + "grad_norm": 0.474609375, + "learning_rate": 0.0001999975458309133, + "loss": 1.2389, + "step": 22 + }, + { + "epoch": 0.012797329253025456, + "grad_norm": 0.47265625, + "learning_rate": 0.00019999711976193663, + "loss": 1.2716, + "step": 23 + }, + { + "epoch": 0.013353734872722215, + "grad_norm": 0.455078125, + "learning_rate": 0.0001999966596081215, + "loss": 1.2038, + "step": 24 + }, + { + "epoch": 0.013910140492418973, + "grad_norm": 0.462890625, + "learning_rate": 0.00019999616536962473, + "loss": 1.2588, + "step": 25 + }, + { + "epoch": 0.014466546112115732, + "grad_norm": 0.451171875, + "learning_rate": 0.00019999563704661483, + "loss": 1.198, + "step": 26 + }, + { + "epoch": 0.01502295173181249, + "grad_norm": 0.4296875, + "learning_rate": 0.00019999507463927188, + "loss": 1.2047, + "step": 27 + }, + { + "epoch": 0.01557935735150925, + "grad_norm": 0.431640625, + "learning_rate": 0.00019999447814778755, + "loss": 1.1884, + "step": 28 + }, + { + "epoch": 0.01613576297120601, + "grad_norm": 0.435546875, + "learning_rate": 0.0001999938475723652, + "loss": 1.2156, + "step": 29 + }, + { + "epoch": 0.016692168590902768, + "grad_norm": 0.4375, + "learning_rate": 0.00019999318291321968, + "loss": 1.1636, + "step": 30 + }, + { + "epoch": 0.017248574210599527, + "grad_norm": 0.4140625, + "learning_rate": 0.00019999248417057767, + "loss": 1.1857, + "step": 31 + }, + { + "epoch": 0.017804979830296285, + "grad_norm": 0.439453125, + "learning_rate": 0.00019999175134467726, + "loss": 1.1862, + "step": 32 + }, + { + "epoch": 0.018361385449993044, + "grad_norm": 0.41015625, + "learning_rate": 0.00019999098443576827, + "loss": 1.2155, + "step": 33 + }, + { + "epoch": 0.018917791069689802, + "grad_norm": 0.427734375, + "learning_rate": 0.00019999018344411207, + "loss": 1.1009, + "step": 34 + }, + { + "epoch": 0.019474196689386564, + "grad_norm": 0.421875, + "learning_rate": 0.00019998934836998173, + "loss": 1.1714, + "step": 35 + }, + { + "epoch": 0.020030602309083323, + "grad_norm": 0.42578125, + "learning_rate": 0.00019998847921366187, + "loss": 1.0824, + "step": 36 + }, + { + "epoch": 0.02058700792878008, + "grad_norm": 0.390625, + "learning_rate": 0.00019998757597544875, + "loss": 1.0703, + "step": 37 + }, + { + "epoch": 0.02114341354847684, + "grad_norm": 0.392578125, + "learning_rate": 0.00019998663865565023, + "loss": 1.1183, + "step": 38 + }, + { + "epoch": 0.0216998191681736, + "grad_norm": 0.400390625, + "learning_rate": 0.00019998566725458586, + "loss": 1.1328, + "step": 39 + }, + { + "epoch": 0.022256224787870357, + "grad_norm": 0.396484375, + "learning_rate": 0.0001999846617725867, + "loss": 1.1076, + "step": 40 + }, + { + "epoch": 0.022812630407567116, + "grad_norm": 0.39453125, + "learning_rate": 0.00019998362220999548, + "loss": 1.1143, + "step": 41 + }, + { + "epoch": 0.023369036027263874, + "grad_norm": 0.419921875, + "learning_rate": 0.00019998254856716657, + "loss": 1.1768, + "step": 42 + }, + { + "epoch": 0.023925441646960633, + "grad_norm": 0.400390625, + "learning_rate": 0.0001999814408444659, + "loss": 1.0872, + "step": 43 + }, + { + "epoch": 0.02448184726665739, + "grad_norm": 0.396484375, + "learning_rate": 0.00019998029904227103, + "loss": 1.122, + "step": 44 + }, + { + "epoch": 0.025038252886354154, + "grad_norm": 0.39453125, + "learning_rate": 0.0001999791231609712, + "loss": 1.0025, + "step": 45 + }, + { + "epoch": 0.025594658506050912, + "grad_norm": 0.4140625, + "learning_rate": 0.00019997791320096718, + "loss": 1.1672, + "step": 46 + }, + { + "epoch": 0.02615106412574767, + "grad_norm": 0.40234375, + "learning_rate": 0.00019997666916267144, + "loss": 1.1096, + "step": 47 + }, + { + "epoch": 0.02670746974544443, + "grad_norm": 0.39453125, + "learning_rate": 0.000199975391046508, + "loss": 1.1446, + "step": 48 + }, + { + "epoch": 0.027263875365141188, + "grad_norm": 0.400390625, + "learning_rate": 0.0001999740788529125, + "loss": 1.1254, + "step": 49 + }, + { + "epoch": 0.027820280984837947, + "grad_norm": 0.396484375, + "learning_rate": 0.00019997273258233223, + "loss": 1.1143, + "step": 50 + }, + { + "epoch": 0.028376686604534705, + "grad_norm": 0.404296875, + "learning_rate": 0.00019997135223522605, + "loss": 1.1123, + "step": 51 + }, + { + "epoch": 0.028933092224231464, + "grad_norm": 0.384765625, + "learning_rate": 0.00019996993781206448, + "loss": 1.0469, + "step": 52 + }, + { + "epoch": 0.029489497843928222, + "grad_norm": 0.388671875, + "learning_rate": 0.00019996848931332965, + "loss": 1.1145, + "step": 53 + }, + { + "epoch": 0.03004590346362498, + "grad_norm": 0.38671875, + "learning_rate": 0.0001999670067395153, + "loss": 1.0489, + "step": 54 + }, + { + "epoch": 0.030602309083321743, + "grad_norm": 0.390625, + "learning_rate": 0.00019996549009112675, + "loss": 1.0671, + "step": 55 + }, + { + "epoch": 0.0311587147030185, + "grad_norm": 0.3828125, + "learning_rate": 0.00019996393936868098, + "loss": 1.0341, + "step": 56 + }, + { + "epoch": 0.03171512032271526, + "grad_norm": 0.3828125, + "learning_rate": 0.00019996235457270654, + "loss": 1.0725, + "step": 57 + }, + { + "epoch": 0.03227152594241202, + "grad_norm": 0.388671875, + "learning_rate": 0.00019996073570374367, + "loss": 1.1031, + "step": 58 + }, + { + "epoch": 0.032827931562108774, + "grad_norm": 0.380859375, + "learning_rate": 0.0001999590827623441, + "loss": 0.9974, + "step": 59 + }, + { + "epoch": 0.033384337181805536, + "grad_norm": 0.392578125, + "learning_rate": 0.00019995739574907133, + "loss": 1.1398, + "step": 60 + }, + { + "epoch": 0.0339407428015023, + "grad_norm": 0.40234375, + "learning_rate": 0.00019995567466450035, + "loss": 1.0728, + "step": 61 + }, + { + "epoch": 0.03449714842119905, + "grad_norm": 0.380859375, + "learning_rate": 0.00019995391950921782, + "loss": 0.9649, + "step": 62 + }, + { + "epoch": 0.035053554040895815, + "grad_norm": 0.384765625, + "learning_rate": 0.000199952130283822, + "loss": 1.1536, + "step": 63 + }, + { + "epoch": 0.03560995966059257, + "grad_norm": 0.376953125, + "learning_rate": 0.0001999503069889227, + "loss": 1.0135, + "step": 64 + }, + { + "epoch": 0.03616636528028933, + "grad_norm": 0.39453125, + "learning_rate": 0.0001999484496251415, + "loss": 1.0599, + "step": 65 + }, + { + "epoch": 0.03672277089998609, + "grad_norm": 0.384765625, + "learning_rate": 0.00019994655819311144, + "loss": 1.0521, + "step": 66 + }, + { + "epoch": 0.03727917651968285, + "grad_norm": 0.388671875, + "learning_rate": 0.00019994463269347725, + "loss": 1.0737, + "step": 67 + }, + { + "epoch": 0.037835582139379605, + "grad_norm": 0.388671875, + "learning_rate": 0.00019994267312689525, + "loss": 1.0856, + "step": 68 + }, + { + "epoch": 0.03839198775907637, + "grad_norm": 0.365234375, + "learning_rate": 0.00019994067949403337, + "loss": 0.9674, + "step": 69 + }, + { + "epoch": 0.03894839337877313, + "grad_norm": 0.390625, + "learning_rate": 0.00019993865179557117, + "loss": 1.1069, + "step": 70 + }, + { + "epoch": 0.039504798998469884, + "grad_norm": 0.3828125, + "learning_rate": 0.00019993659003219978, + "loss": 1.0198, + "step": 71 + }, + { + "epoch": 0.040061204618166646, + "grad_norm": 0.384765625, + "learning_rate": 0.00019993449420462197, + "loss": 1.0864, + "step": 72 + }, + { + "epoch": 0.0406176102378634, + "grad_norm": 0.38671875, + "learning_rate": 0.00019993236431355217, + "loss": 1.037, + "step": 73 + }, + { + "epoch": 0.04117401585756016, + "grad_norm": 0.39453125, + "learning_rate": 0.00019993020035971633, + "loss": 1.0724, + "step": 74 + }, + { + "epoch": 0.04173042147725692, + "grad_norm": 0.369140625, + "learning_rate": 0.00019992800234385201, + "loss": 0.9854, + "step": 75 + }, + { + "epoch": 0.04228682709695368, + "grad_norm": 0.37890625, + "learning_rate": 0.00019992577026670854, + "loss": 1.0655, + "step": 76 + }, + { + "epoch": 0.042843232716650435, + "grad_norm": 0.392578125, + "learning_rate": 0.00019992350412904663, + "loss": 1.0487, + "step": 77 + }, + { + "epoch": 0.0433996383363472, + "grad_norm": 0.369140625, + "learning_rate": 0.00019992120393163876, + "loss": 0.9593, + "step": 78 + }, + { + "epoch": 0.04395604395604396, + "grad_norm": 0.40234375, + "learning_rate": 0.00019991886967526897, + "loss": 1.1132, + "step": 79 + }, + { + "epoch": 0.044512449575740715, + "grad_norm": 0.36328125, + "learning_rate": 0.00019991650136073287, + "loss": 0.9343, + "step": 80 + }, + { + "epoch": 0.04506885519543748, + "grad_norm": 0.369140625, + "learning_rate": 0.0001999140989888378, + "loss": 1.0228, + "step": 81 + }, + { + "epoch": 0.04562526081513423, + "grad_norm": 0.384765625, + "learning_rate": 0.00019991166256040256, + "loss": 1.0239, + "step": 82 + }, + { + "epoch": 0.046181666434830994, + "grad_norm": 0.38671875, + "learning_rate": 0.00019990919207625764, + "loss": 1.0706, + "step": 83 + }, + { + "epoch": 0.04673807205452775, + "grad_norm": 0.38671875, + "learning_rate": 0.00019990668753724515, + "loss": 1.0069, + "step": 84 + }, + { + "epoch": 0.04729447767422451, + "grad_norm": 0.38671875, + "learning_rate": 0.00019990414894421876, + "loss": 1.0511, + "step": 85 + }, + { + "epoch": 0.047850883293921266, + "grad_norm": 0.37890625, + "learning_rate": 0.0001999015762980438, + "loss": 0.9398, + "step": 86 + }, + { + "epoch": 0.04840728891361803, + "grad_norm": 0.396484375, + "learning_rate": 0.00019989896959959712, + "loss": 1.1496, + "step": 87 + }, + { + "epoch": 0.04896369453331478, + "grad_norm": 0.369140625, + "learning_rate": 0.00019989632884976727, + "loss": 1.0488, + "step": 88 + }, + { + "epoch": 0.049520100153011545, + "grad_norm": 0.39453125, + "learning_rate": 0.00019989365404945436, + "loss": 1.0843, + "step": 89 + }, + { + "epoch": 0.05007650577270831, + "grad_norm": 0.375, + "learning_rate": 0.00019989094519957015, + "loss": 1.0822, + "step": 90 + }, + { + "epoch": 0.05063291139240506, + "grad_norm": 0.373046875, + "learning_rate": 0.00019988820230103796, + "loss": 1.0301, + "step": 91 + }, + { + "epoch": 0.051189317012101825, + "grad_norm": 0.380859375, + "learning_rate": 0.0001998854253547927, + "loss": 1.0098, + "step": 92 + }, + { + "epoch": 0.05174572263179858, + "grad_norm": 0.361328125, + "learning_rate": 0.00019988261436178093, + "loss": 0.9847, + "step": 93 + }, + { + "epoch": 0.05230212825149534, + "grad_norm": 0.38671875, + "learning_rate": 0.0001998797693229608, + "loss": 1.0233, + "step": 94 + }, + { + "epoch": 0.0528585338711921, + "grad_norm": 0.388671875, + "learning_rate": 0.0001998768902393021, + "loss": 1.0747, + "step": 95 + }, + { + "epoch": 0.05341493949088886, + "grad_norm": 0.3828125, + "learning_rate": 0.00019987397711178614, + "loss": 0.9823, + "step": 96 + }, + { + "epoch": 0.053971345110585614, + "grad_norm": 0.390625, + "learning_rate": 0.0001998710299414059, + "loss": 1.0238, + "step": 97 + }, + { + "epoch": 0.054527750730282376, + "grad_norm": 0.38671875, + "learning_rate": 0.00019986804872916593, + "loss": 0.9832, + "step": 98 + }, + { + "epoch": 0.05508415634997914, + "grad_norm": 0.373046875, + "learning_rate": 0.00019986503347608245, + "loss": 0.9833, + "step": 99 + }, + { + "epoch": 0.05564056196967589, + "grad_norm": 0.392578125, + "learning_rate": 0.0001998619841831832, + "loss": 1.0971, + "step": 100 + }, + { + "epoch": 0.056196967589372655, + "grad_norm": 0.376953125, + "learning_rate": 0.0001998589008515076, + "loss": 1.0077, + "step": 101 + }, + { + "epoch": 0.05675337320906941, + "grad_norm": 0.390625, + "learning_rate": 0.0001998557834821065, + "loss": 1.0066, + "step": 102 + }, + { + "epoch": 0.05730977882876617, + "grad_norm": 0.380859375, + "learning_rate": 0.00019985263207604264, + "loss": 1.0267, + "step": 103 + }, + { + "epoch": 0.05786618444846293, + "grad_norm": 0.373046875, + "learning_rate": 0.0001998494466343901, + "loss": 0.975, + "step": 104 + }, + { + "epoch": 0.05842259006815969, + "grad_norm": 0.3828125, + "learning_rate": 0.00019984622715823474, + "loss": 1.0134, + "step": 105 + }, + { + "epoch": 0.058978995687856445, + "grad_norm": 0.392578125, + "learning_rate": 0.00019984297364867385, + "loss": 1.1247, + "step": 106 + }, + { + "epoch": 0.05953540130755321, + "grad_norm": 0.376953125, + "learning_rate": 0.00019983968610681648, + "loss": 0.974, + "step": 107 + }, + { + "epoch": 0.06009180692724996, + "grad_norm": 0.380859375, + "learning_rate": 0.0001998363645337832, + "loss": 1.0099, + "step": 108 + }, + { + "epoch": 0.060648212546946724, + "grad_norm": 0.3515625, + "learning_rate": 0.00019983300893070623, + "loss": 0.8479, + "step": 109 + }, + { + "epoch": 0.061204618166643486, + "grad_norm": 0.375, + "learning_rate": 0.00019982961929872927, + "loss": 1.0034, + "step": 110 + }, + { + "epoch": 0.06176102378634024, + "grad_norm": 0.380859375, + "learning_rate": 0.00019982619563900782, + "loss": 1.0165, + "step": 111 + }, + { + "epoch": 0.062317429406037, + "grad_norm": 0.390625, + "learning_rate": 0.00019982273795270876, + "loss": 1.0338, + "step": 112 + }, + { + "epoch": 0.06287383502573377, + "grad_norm": 0.392578125, + "learning_rate": 0.00019981924624101071, + "loss": 1.0465, + "step": 113 + }, + { + "epoch": 0.06343024064543051, + "grad_norm": 0.369140625, + "learning_rate": 0.00019981572050510387, + "loss": 0.9284, + "step": 114 + }, + { + "epoch": 0.06398664626512728, + "grad_norm": 0.40234375, + "learning_rate": 0.00019981216074619, + "loss": 1.074, + "step": 115 + }, + { + "epoch": 0.06454305188482404, + "grad_norm": 0.375, + "learning_rate": 0.00019980856696548242, + "loss": 0.8981, + "step": 116 + }, + { + "epoch": 0.0650994575045208, + "grad_norm": 0.375, + "learning_rate": 0.00019980493916420616, + "loss": 1.0605, + "step": 117 + }, + { + "epoch": 0.06565586312421755, + "grad_norm": 0.37890625, + "learning_rate": 0.00019980127734359782, + "loss": 1.0433, + "step": 118 + }, + { + "epoch": 0.06621226874391431, + "grad_norm": 0.38671875, + "learning_rate": 0.00019979758150490546, + "loss": 1.0483, + "step": 119 + }, + { + "epoch": 0.06676867436361107, + "grad_norm": 0.392578125, + "learning_rate": 0.0001997938516493889, + "loss": 1.0543, + "step": 120 + }, + { + "epoch": 0.06732507998330783, + "grad_norm": 0.376953125, + "learning_rate": 0.0001997900877783195, + "loss": 1.0162, + "step": 121 + }, + { + "epoch": 0.0678814856030046, + "grad_norm": 0.369140625, + "learning_rate": 0.0001997862898929802, + "loss": 0.8468, + "step": 122 + }, + { + "epoch": 0.06843789122270134, + "grad_norm": 0.37890625, + "learning_rate": 0.00019978245799466554, + "loss": 1.0417, + "step": 123 + }, + { + "epoch": 0.0689942968423981, + "grad_norm": 0.3828125, + "learning_rate": 0.00019977859208468161, + "loss": 0.9901, + "step": 124 + }, + { + "epoch": 0.06955070246209487, + "grad_norm": 0.404296875, + "learning_rate": 0.00019977469216434619, + "loss": 1.0117, + "step": 125 + }, + { + "epoch": 0.07010710808179163, + "grad_norm": 0.375, + "learning_rate": 0.0001997707582349886, + "loss": 0.9719, + "step": 126 + }, + { + "epoch": 0.07066351370148838, + "grad_norm": 0.369140625, + "learning_rate": 0.00019976679029794968, + "loss": 0.9176, + "step": 127 + }, + { + "epoch": 0.07121991932118514, + "grad_norm": 0.400390625, + "learning_rate": 0.00019976278835458203, + "loss": 1.0633, + "step": 128 + }, + { + "epoch": 0.0717763249408819, + "grad_norm": 0.380859375, + "learning_rate": 0.00019975875240624968, + "loss": 0.9845, + "step": 129 + }, + { + "epoch": 0.07233273056057866, + "grad_norm": 0.376953125, + "learning_rate": 0.00019975468245432833, + "loss": 0.9352, + "step": 130 + }, + { + "epoch": 0.07288913618027543, + "grad_norm": 0.3828125, + "learning_rate": 0.0001997505785002053, + "loss": 1.0966, + "step": 131 + }, + { + "epoch": 0.07344554179997217, + "grad_norm": 0.380859375, + "learning_rate": 0.0001997464405452794, + "loss": 0.9278, + "step": 132 + }, + { + "epoch": 0.07400194741966894, + "grad_norm": 0.359375, + "learning_rate": 0.0001997422685909611, + "loss": 0.9244, + "step": 133 + }, + { + "epoch": 0.0745583530393657, + "grad_norm": 0.3671875, + "learning_rate": 0.00019973806263867247, + "loss": 0.8941, + "step": 134 + }, + { + "epoch": 0.07511475865906246, + "grad_norm": 0.380859375, + "learning_rate": 0.00019973382268984713, + "loss": 0.9702, + "step": 135 + }, + { + "epoch": 0.07567116427875921, + "grad_norm": 0.376953125, + "learning_rate": 0.00019972954874593026, + "loss": 0.9809, + "step": 136 + }, + { + "epoch": 0.07622756989845597, + "grad_norm": 0.37890625, + "learning_rate": 0.00019972524080837873, + "loss": 1.0604, + "step": 137 + }, + { + "epoch": 0.07678397551815273, + "grad_norm": 0.361328125, + "learning_rate": 0.00019972089887866092, + "loss": 0.9702, + "step": 138 + }, + { + "epoch": 0.0773403811378495, + "grad_norm": 0.37890625, + "learning_rate": 0.00019971652295825681, + "loss": 0.9903, + "step": 139 + }, + { + "epoch": 0.07789678675754626, + "grad_norm": 0.36328125, + "learning_rate": 0.00019971211304865795, + "loss": 0.979, + "step": 140 + }, + { + "epoch": 0.078453192377243, + "grad_norm": 0.380859375, + "learning_rate": 0.00019970766915136746, + "loss": 1.1067, + "step": 141 + }, + { + "epoch": 0.07900959799693977, + "grad_norm": 0.357421875, + "learning_rate": 0.0001997031912679002, + "loss": 0.966, + "step": 142 + }, + { + "epoch": 0.07956600361663653, + "grad_norm": 0.375, + "learning_rate": 0.00019969867939978236, + "loss": 1.0207, + "step": 143 + }, + { + "epoch": 0.08012240923633329, + "grad_norm": 0.37890625, + "learning_rate": 0.00019969413354855195, + "loss": 0.9127, + "step": 144 + }, + { + "epoch": 0.08067881485603004, + "grad_norm": 0.400390625, + "learning_rate": 0.0001996895537157584, + "loss": 1.0944, + "step": 145 + }, + { + "epoch": 0.0812352204757268, + "grad_norm": 0.35546875, + "learning_rate": 0.00019968493990296277, + "loss": 0.948, + "step": 146 + }, + { + "epoch": 0.08179162609542356, + "grad_norm": 0.3828125, + "learning_rate": 0.00019968029211173778, + "loss": 0.9824, + "step": 147 + }, + { + "epoch": 0.08234803171512033, + "grad_norm": 0.37890625, + "learning_rate": 0.00019967561034366763, + "loss": 0.9328, + "step": 148 + }, + { + "epoch": 0.08290443733481709, + "grad_norm": 0.37109375, + "learning_rate": 0.00019967089460034814, + "loss": 1.0162, + "step": 149 + }, + { + "epoch": 0.08346084295451384, + "grad_norm": 0.33984375, + "learning_rate": 0.0001996661448833867, + "loss": 0.7696, + "step": 150 + }, + { + "epoch": 0.0840172485742106, + "grad_norm": 0.361328125, + "learning_rate": 0.00019966136119440233, + "loss": 0.9133, + "step": 151 + }, + { + "epoch": 0.08457365419390736, + "grad_norm": 0.384765625, + "learning_rate": 0.00019965654353502554, + "loss": 1.0071, + "step": 152 + }, + { + "epoch": 0.08513005981360412, + "grad_norm": 0.392578125, + "learning_rate": 0.0001996516919068985, + "loss": 0.9797, + "step": 153 + }, + { + "epoch": 0.08568646543330087, + "grad_norm": 0.388671875, + "learning_rate": 0.00019964680631167492, + "loss": 0.9521, + "step": 154 + }, + { + "epoch": 0.08624287105299763, + "grad_norm": 0.392578125, + "learning_rate": 0.0001996418867510201, + "loss": 1.034, + "step": 155 + }, + { + "epoch": 0.0867992766726944, + "grad_norm": 0.373046875, + "learning_rate": 0.00019963693322661087, + "loss": 0.9971, + "step": 156 + }, + { + "epoch": 0.08735568229239116, + "grad_norm": 0.37890625, + "learning_rate": 0.00019963194574013573, + "loss": 1.0262, + "step": 157 + }, + { + "epoch": 0.08791208791208792, + "grad_norm": 0.380859375, + "learning_rate": 0.0001996269242932947, + "loss": 0.9913, + "step": 158 + }, + { + "epoch": 0.08846849353178467, + "grad_norm": 0.408203125, + "learning_rate": 0.00019962186888779936, + "loss": 1.0822, + "step": 159 + }, + { + "epoch": 0.08902489915148143, + "grad_norm": 0.36328125, + "learning_rate": 0.00019961677952537292, + "loss": 0.8287, + "step": 160 + }, + { + "epoch": 0.08958130477117819, + "grad_norm": 0.37890625, + "learning_rate": 0.0001996116562077501, + "loss": 0.9663, + "step": 161 + }, + { + "epoch": 0.09013771039087495, + "grad_norm": 0.333984375, + "learning_rate": 0.00019960649893667722, + "loss": 0.7657, + "step": 162 + }, + { + "epoch": 0.0906941160105717, + "grad_norm": 0.35546875, + "learning_rate": 0.0001996013077139122, + "loss": 0.8837, + "step": 163 + }, + { + "epoch": 0.09125052163026846, + "grad_norm": 0.376953125, + "learning_rate": 0.0001995960825412245, + "loss": 0.9727, + "step": 164 + }, + { + "epoch": 0.09180692724996523, + "grad_norm": 0.3671875, + "learning_rate": 0.0001995908234203952, + "loss": 0.9438, + "step": 165 + }, + { + "epoch": 0.09236333286966199, + "grad_norm": 0.369140625, + "learning_rate": 0.00019958553035321686, + "loss": 0.9364, + "step": 166 + }, + { + "epoch": 0.09291973848935874, + "grad_norm": 0.353515625, + "learning_rate": 0.00019958020334149366, + "loss": 0.9435, + "step": 167 + }, + { + "epoch": 0.0934761441090555, + "grad_norm": 0.384765625, + "learning_rate": 0.00019957484238704139, + "loss": 1.034, + "step": 168 + }, + { + "epoch": 0.09403254972875226, + "grad_norm": 0.37109375, + "learning_rate": 0.00019956944749168737, + "loss": 0.9662, + "step": 169 + }, + { + "epoch": 0.09458895534844902, + "grad_norm": 0.38671875, + "learning_rate": 0.00019956401865727053, + "loss": 1.0057, + "step": 170 + }, + { + "epoch": 0.09514536096814578, + "grad_norm": 0.361328125, + "learning_rate": 0.00019955855588564125, + "loss": 0.9569, + "step": 171 + }, + { + "epoch": 0.09570176658784253, + "grad_norm": 0.37109375, + "learning_rate": 0.00019955305917866164, + "loss": 0.9651, + "step": 172 + }, + { + "epoch": 0.0962581722075393, + "grad_norm": 0.3515625, + "learning_rate": 0.00019954752853820523, + "loss": 0.8911, + "step": 173 + }, + { + "epoch": 0.09681457782723606, + "grad_norm": 0.3515625, + "learning_rate": 0.00019954196396615721, + "loss": 0.9656, + "step": 174 + }, + { + "epoch": 0.09737098344693282, + "grad_norm": 0.373046875, + "learning_rate": 0.00019953636546441431, + "loss": 0.8851, + "step": 175 + }, + { + "epoch": 0.09792738906662957, + "grad_norm": 0.357421875, + "learning_rate": 0.00019953073303488486, + "loss": 0.91, + "step": 176 + }, + { + "epoch": 0.09848379468632633, + "grad_norm": 0.3828125, + "learning_rate": 0.00019952506667948868, + "loss": 1.0532, + "step": 177 + }, + { + "epoch": 0.09904020030602309, + "grad_norm": 0.384765625, + "learning_rate": 0.00019951936640015723, + "loss": 1.052, + "step": 178 + }, + { + "epoch": 0.09959660592571985, + "grad_norm": 0.36328125, + "learning_rate": 0.00019951363219883344, + "loss": 0.9436, + "step": 179 + }, + { + "epoch": 0.10015301154541661, + "grad_norm": 0.3828125, + "learning_rate": 0.0001995078640774719, + "loss": 1.0067, + "step": 180 + }, + { + "epoch": 0.10070941716511336, + "grad_norm": 0.353515625, + "learning_rate": 0.00019950206203803874, + "loss": 0.8923, + "step": 181 + }, + { + "epoch": 0.10126582278481013, + "grad_norm": 0.3671875, + "learning_rate": 0.00019949622608251156, + "loss": 0.9681, + "step": 182 + }, + { + "epoch": 0.10182222840450689, + "grad_norm": 0.3828125, + "learning_rate": 0.00019949035621287967, + "loss": 0.9994, + "step": 183 + }, + { + "epoch": 0.10237863402420365, + "grad_norm": 0.3671875, + "learning_rate": 0.00019948445243114384, + "loss": 1.0139, + "step": 184 + }, + { + "epoch": 0.1029350396439004, + "grad_norm": 0.365234375, + "learning_rate": 0.0001994785147393164, + "loss": 0.9698, + "step": 185 + }, + { + "epoch": 0.10349144526359716, + "grad_norm": 0.373046875, + "learning_rate": 0.0001994725431394213, + "loss": 0.9727, + "step": 186 + }, + { + "epoch": 0.10404785088329392, + "grad_norm": 0.34765625, + "learning_rate": 0.00019946653763349394, + "loss": 0.9109, + "step": 187 + }, + { + "epoch": 0.10460425650299068, + "grad_norm": 0.380859375, + "learning_rate": 0.0001994604982235814, + "loss": 0.9881, + "step": 188 + }, + { + "epoch": 0.10516066212268745, + "grad_norm": 0.3828125, + "learning_rate": 0.00019945442491174227, + "loss": 1.0062, + "step": 189 + }, + { + "epoch": 0.1057170677423842, + "grad_norm": 0.375, + "learning_rate": 0.00019944831770004665, + "loss": 0.9374, + "step": 190 + }, + { + "epoch": 0.10627347336208096, + "grad_norm": 0.3671875, + "learning_rate": 0.00019944217659057626, + "loss": 0.907, + "step": 191 + }, + { + "epoch": 0.10682987898177772, + "grad_norm": 0.365234375, + "learning_rate": 0.00019943600158542436, + "loss": 0.9287, + "step": 192 + }, + { + "epoch": 0.10738628460147448, + "grad_norm": 0.376953125, + "learning_rate": 0.00019942979268669573, + "loss": 0.9948, + "step": 193 + }, + { + "epoch": 0.10794269022117123, + "grad_norm": 0.373046875, + "learning_rate": 0.0001994235498965067, + "loss": 0.9651, + "step": 194 + }, + { + "epoch": 0.10849909584086799, + "grad_norm": 0.3671875, + "learning_rate": 0.00019941727321698521, + "loss": 0.9267, + "step": 195 + }, + { + "epoch": 0.10905550146056475, + "grad_norm": 0.349609375, + "learning_rate": 0.00019941096265027074, + "loss": 0.8551, + "step": 196 + }, + { + "epoch": 0.10961190708026151, + "grad_norm": 0.369140625, + "learning_rate": 0.00019940461819851425, + "loss": 0.9683, + "step": 197 + }, + { + "epoch": 0.11016831269995828, + "grad_norm": 0.357421875, + "learning_rate": 0.00019939823986387834, + "loss": 0.9127, + "step": 198 + }, + { + "epoch": 0.11072471831965502, + "grad_norm": 0.373046875, + "learning_rate": 0.00019939182764853707, + "loss": 1.0557, + "step": 199 + }, + { + "epoch": 0.11128112393935179, + "grad_norm": 0.37890625, + "learning_rate": 0.00019938538155467615, + "loss": 0.9967, + "step": 200 + }, + { + "epoch": 0.11183752955904855, + "grad_norm": 0.376953125, + "learning_rate": 0.00019937890158449272, + "loss": 0.9391, + "step": 201 + }, + { + "epoch": 0.11239393517874531, + "grad_norm": 2.109375, + "learning_rate": 0.00019937238774019557, + "loss": 1.9551, + "step": 202 + }, + { + "epoch": 0.11295034079844206, + "grad_norm": 0.37890625, + "learning_rate": 0.000199365840024005, + "loss": 0.9682, + "step": 203 + }, + { + "epoch": 0.11350674641813882, + "grad_norm": 0.373046875, + "learning_rate": 0.00019935925843815285, + "loss": 0.961, + "step": 204 + }, + { + "epoch": 0.11406315203783558, + "grad_norm": 0.42578125, + "learning_rate": 0.0001993526429848825, + "loss": 1.0718, + "step": 205 + }, + { + "epoch": 0.11461955765753234, + "grad_norm": 0.37890625, + "learning_rate": 0.0001993459936664489, + "loss": 0.9232, + "step": 206 + }, + { + "epoch": 0.1151759632772291, + "grad_norm": 0.396484375, + "learning_rate": 0.0001993393104851185, + "loss": 1.0022, + "step": 207 + }, + { + "epoch": 0.11573236889692586, + "grad_norm": 0.37890625, + "learning_rate": 0.00019933259344316934, + "loss": 0.9549, + "step": 208 + }, + { + "epoch": 0.11628877451662262, + "grad_norm": 0.357421875, + "learning_rate": 0.00019932584254289096, + "loss": 0.9472, + "step": 209 + }, + { + "epoch": 0.11684518013631938, + "grad_norm": 2.859375, + "learning_rate": 0.00019931905778658443, + "loss": 1.9077, + "step": 210 + }, + { + "epoch": 0.11740158575601614, + "grad_norm": 0.388671875, + "learning_rate": 0.00019931223917656247, + "loss": 1.0227, + "step": 211 + }, + { + "epoch": 0.11795799137571289, + "grad_norm": 0.404296875, + "learning_rate": 0.00019930538671514923, + "loss": 1.0234, + "step": 212 + }, + { + "epoch": 0.11851439699540965, + "grad_norm": 0.373046875, + "learning_rate": 0.00019929850040468035, + "loss": 0.9573, + "step": 213 + }, + { + "epoch": 0.11907080261510641, + "grad_norm": 0.365234375, + "learning_rate": 0.0001992915802475032, + "loss": 0.9576, + "step": 214 + }, + { + "epoch": 0.11962720823480318, + "grad_norm": 0.365234375, + "learning_rate": 0.0001992846262459765, + "loss": 0.9715, + "step": 215 + }, + { + "epoch": 0.12018361385449992, + "grad_norm": 0.376953125, + "learning_rate": 0.0001992776384024706, + "loss": 0.8743, + "step": 216 + }, + { + "epoch": 0.12074001947419669, + "grad_norm": 0.365234375, + "learning_rate": 0.00019927061671936737, + "loss": 0.8957, + "step": 217 + }, + { + "epoch": 0.12129642509389345, + "grad_norm": 0.37890625, + "learning_rate": 0.0001992635611990602, + "loss": 0.9718, + "step": 218 + }, + { + "epoch": 0.12185283071359021, + "grad_norm": 0.369140625, + "learning_rate": 0.00019925647184395402, + "loss": 0.8916, + "step": 219 + }, + { + "epoch": 0.12240923633328697, + "grad_norm": 0.361328125, + "learning_rate": 0.0001992493486564653, + "loss": 0.9546, + "step": 220 + }, + { + "epoch": 0.12296564195298372, + "grad_norm": 0.369140625, + "learning_rate": 0.00019924219163902204, + "loss": 0.9285, + "step": 221 + }, + { + "epoch": 0.12352204757268048, + "grad_norm": 0.36328125, + "learning_rate": 0.00019923500079406373, + "loss": 0.9095, + "step": 222 + }, + { + "epoch": 0.12407845319237724, + "grad_norm": 0.390625, + "learning_rate": 0.00019922777612404148, + "loss": 0.9921, + "step": 223 + }, + { + "epoch": 0.124634858812074, + "grad_norm": 0.376953125, + "learning_rate": 0.00019922051763141788, + "loss": 0.9756, + "step": 224 + }, + { + "epoch": 0.12519126443177075, + "grad_norm": 0.345703125, + "learning_rate": 0.00019921322531866702, + "loss": 0.832, + "step": 225 + }, + { + "epoch": 0.12574767005146753, + "grad_norm": 0.322265625, + "learning_rate": 0.00019920589918827455, + "loss": 0.741, + "step": 226 + }, + { + "epoch": 0.12630407567116428, + "grad_norm": 0.3671875, + "learning_rate": 0.0001991985392427376, + "loss": 0.8839, + "step": 227 + }, + { + "epoch": 0.12686048129086103, + "grad_norm": 0.37890625, + "learning_rate": 0.00019919114548456497, + "loss": 1.0329, + "step": 228 + }, + { + "epoch": 0.1274168869105578, + "grad_norm": 0.369140625, + "learning_rate": 0.0001991837179162768, + "loss": 0.9777, + "step": 229 + }, + { + "epoch": 0.12797329253025455, + "grad_norm": 0.357421875, + "learning_rate": 0.00019917625654040485, + "loss": 0.9311, + "step": 230 + }, + { + "epoch": 0.12852969814995133, + "grad_norm": 0.38671875, + "learning_rate": 0.0001991687613594924, + "loss": 0.9973, + "step": 231 + }, + { + "epoch": 0.12908610376964808, + "grad_norm": 0.375, + "learning_rate": 0.00019916123237609428, + "loss": 1.054, + "step": 232 + }, + { + "epoch": 0.12964250938934482, + "grad_norm": 0.36328125, + "learning_rate": 0.00019915366959277674, + "loss": 0.886, + "step": 233 + }, + { + "epoch": 0.1301989150090416, + "grad_norm": 0.37109375, + "learning_rate": 0.00019914607301211765, + "loss": 0.9879, + "step": 234 + }, + { + "epoch": 0.13075532062873835, + "grad_norm": 0.333984375, + "learning_rate": 0.00019913844263670638, + "loss": 0.8272, + "step": 235 + }, + { + "epoch": 0.1313117262484351, + "grad_norm": 0.37890625, + "learning_rate": 0.00019913077846914377, + "loss": 0.8881, + "step": 236 + }, + { + "epoch": 0.13186813186813187, + "grad_norm": 0.37109375, + "learning_rate": 0.00019912308051204226, + "loss": 0.9737, + "step": 237 + }, + { + "epoch": 0.13242453748782862, + "grad_norm": 0.40625, + "learning_rate": 0.00019911534876802572, + "loss": 1.1547, + "step": 238 + }, + { + "epoch": 0.1329809431075254, + "grad_norm": 0.37109375, + "learning_rate": 0.00019910758323972961, + "loss": 0.9975, + "step": 239 + }, + { + "epoch": 0.13353734872722214, + "grad_norm": 0.36328125, + "learning_rate": 0.00019909978392980087, + "loss": 0.9454, + "step": 240 + }, + { + "epoch": 0.1340937543469189, + "grad_norm": 0.37890625, + "learning_rate": 0.00019909195084089792, + "loss": 0.969, + "step": 241 + }, + { + "epoch": 0.13465015996661567, + "grad_norm": 0.357421875, + "learning_rate": 0.00019908408397569075, + "loss": 1.0012, + "step": 242 + }, + { + "epoch": 0.13520656558631242, + "grad_norm": 0.365234375, + "learning_rate": 0.0001990761833368609, + "loss": 0.9292, + "step": 243 + }, + { + "epoch": 0.1357629712060092, + "grad_norm": 0.35546875, + "learning_rate": 0.0001990682489271013, + "loss": 0.8401, + "step": 244 + }, + { + "epoch": 0.13631937682570594, + "grad_norm": 0.36328125, + "learning_rate": 0.00019906028074911645, + "loss": 0.9614, + "step": 245 + }, + { + "epoch": 0.1368757824454027, + "grad_norm": 0.388671875, + "learning_rate": 0.00019905227880562246, + "loss": 0.9454, + "step": 246 + }, + { + "epoch": 0.13743218806509946, + "grad_norm": 0.349609375, + "learning_rate": 0.0001990442430993468, + "loss": 0.8745, + "step": 247 + }, + { + "epoch": 0.1379885936847962, + "grad_norm": 0.361328125, + "learning_rate": 0.00019903617363302846, + "loss": 0.9185, + "step": 248 + }, + { + "epoch": 0.138544999304493, + "grad_norm": 0.37109375, + "learning_rate": 0.00019902807040941806, + "loss": 0.9842, + "step": 249 + }, + { + "epoch": 0.13910140492418974, + "grad_norm": 0.37890625, + "learning_rate": 0.00019901993343127763, + "loss": 1.0057, + "step": 250 + }, + { + "epoch": 0.13965781054388648, + "grad_norm": 0.357421875, + "learning_rate": 0.0001990117627013807, + "loss": 0.8891, + "step": 251 + }, + { + "epoch": 0.14021421616358326, + "grad_norm": 0.361328125, + "learning_rate": 0.00019900355822251233, + "loss": 0.9715, + "step": 252 + }, + { + "epoch": 0.14077062178328, + "grad_norm": 0.34375, + "learning_rate": 0.0001989953199974691, + "loss": 0.8475, + "step": 253 + }, + { + "epoch": 0.14132702740297676, + "grad_norm": 0.3515625, + "learning_rate": 0.0001989870480290591, + "loss": 0.9162, + "step": 254 + }, + { + "epoch": 0.14188343302267353, + "grad_norm": 0.349609375, + "learning_rate": 0.00019897874232010186, + "loss": 0.8788, + "step": 255 + }, + { + "epoch": 0.14243983864237028, + "grad_norm": 0.7734375, + "learning_rate": 0.0001989704028734285, + "loss": 1.8645, + "step": 256 + }, + { + "epoch": 0.14299624426206706, + "grad_norm": 0.53125, + "learning_rate": 0.00019896202969188148, + "loss": 1.0641, + "step": 257 + }, + { + "epoch": 0.1435526498817638, + "grad_norm": 0.37109375, + "learning_rate": 0.000198953622778315, + "loss": 1.0014, + "step": 258 + }, + { + "epoch": 0.14410905550146055, + "grad_norm": 0.39453125, + "learning_rate": 0.00019894518213559457, + "loss": 1.0273, + "step": 259 + }, + { + "epoch": 0.14466546112115733, + "grad_norm": 0.361328125, + "learning_rate": 0.0001989367077665972, + "loss": 0.9224, + "step": 260 + }, + { + "epoch": 0.14522186674085408, + "grad_norm": 0.353515625, + "learning_rate": 0.00019892819967421154, + "loss": 0.9585, + "step": 261 + }, + { + "epoch": 0.14577827236055085, + "grad_norm": 0.3828125, + "learning_rate": 0.0001989196578613376, + "loss": 0.9612, + "step": 262 + }, + { + "epoch": 0.1463346779802476, + "grad_norm": 0.3515625, + "learning_rate": 0.0001989110823308869, + "loss": 0.8635, + "step": 263 + }, + { + "epoch": 0.14689108359994435, + "grad_norm": 0.37109375, + "learning_rate": 0.0001989024730857825, + "loss": 0.9478, + "step": 264 + }, + { + "epoch": 0.14744748921964113, + "grad_norm": 0.376953125, + "learning_rate": 0.00019889383012895896, + "loss": 1.0103, + "step": 265 + }, + { + "epoch": 0.14800389483933787, + "grad_norm": 0.353515625, + "learning_rate": 0.00019888515346336226, + "loss": 0.8712, + "step": 266 + }, + { + "epoch": 0.14856030045903465, + "grad_norm": 0.3671875, + "learning_rate": 0.00019887644309195, + "loss": 0.9338, + "step": 267 + }, + { + "epoch": 0.1491167060787314, + "grad_norm": 0.345703125, + "learning_rate": 0.00019886769901769104, + "loss": 0.9435, + "step": 268 + }, + { + "epoch": 0.14967311169842815, + "grad_norm": 0.369140625, + "learning_rate": 0.00019885892124356597, + "loss": 1.0054, + "step": 269 + }, + { + "epoch": 0.15022951731812492, + "grad_norm": 0.375, + "learning_rate": 0.00019885010977256675, + "loss": 0.9563, + "step": 270 + }, + { + "epoch": 0.15078592293782167, + "grad_norm": 0.36328125, + "learning_rate": 0.0001988412646076968, + "loss": 0.889, + "step": 271 + }, + { + "epoch": 0.15134232855751842, + "grad_norm": 0.34375, + "learning_rate": 0.00019883238575197113, + "loss": 0.8797, + "step": 272 + }, + { + "epoch": 0.1518987341772152, + "grad_norm": 0.357421875, + "learning_rate": 0.00019882347320841615, + "loss": 0.9045, + "step": 273 + }, + { + "epoch": 0.15245513979691194, + "grad_norm": 0.34765625, + "learning_rate": 0.00019881452698006973, + "loss": 0.8376, + "step": 274 + }, + { + "epoch": 0.15301154541660872, + "grad_norm": 0.361328125, + "learning_rate": 0.0001988055470699813, + "loss": 0.9139, + "step": 275 + }, + { + "epoch": 0.15356795103630547, + "grad_norm": 0.361328125, + "learning_rate": 0.00019879653348121176, + "loss": 0.9675, + "step": 276 + }, + { + "epoch": 0.15412435665600221, + "grad_norm": 0.345703125, + "learning_rate": 0.00019878748621683344, + "loss": 0.8359, + "step": 277 + }, + { + "epoch": 0.154680762275699, + "grad_norm": 0.353515625, + "learning_rate": 0.00019877840527993018, + "loss": 0.9296, + "step": 278 + }, + { + "epoch": 0.15523716789539574, + "grad_norm": 0.357421875, + "learning_rate": 0.00019876929067359726, + "loss": 0.9716, + "step": 279 + }, + { + "epoch": 0.15579357351509251, + "grad_norm": 0.37109375, + "learning_rate": 0.00019876014240094154, + "loss": 0.9313, + "step": 280 + }, + { + "epoch": 0.15634997913478926, + "grad_norm": 0.357421875, + "learning_rate": 0.0001987509604650812, + "loss": 0.9064, + "step": 281 + }, + { + "epoch": 0.156906384754486, + "grad_norm": 0.369140625, + "learning_rate": 0.00019874174486914602, + "loss": 1.0199, + "step": 282 + }, + { + "epoch": 0.1574627903741828, + "grad_norm": 0.365234375, + "learning_rate": 0.00019873249561627723, + "loss": 0.9184, + "step": 283 + }, + { + "epoch": 0.15801919599387954, + "grad_norm": 0.337890625, + "learning_rate": 0.0001987232127096275, + "loss": 0.8171, + "step": 284 + }, + { + "epoch": 0.15857560161357628, + "grad_norm": 0.376953125, + "learning_rate": 0.00019871389615236094, + "loss": 0.9977, + "step": 285 + }, + { + "epoch": 0.15913200723327306, + "grad_norm": 0.369140625, + "learning_rate": 0.00019870454594765323, + "loss": 0.9063, + "step": 286 + }, + { + "epoch": 0.1596884128529698, + "grad_norm": 0.3671875, + "learning_rate": 0.00019869516209869144, + "loss": 1.0768, + "step": 287 + }, + { + "epoch": 0.16024481847266658, + "grad_norm": 0.3671875, + "learning_rate": 0.00019868574460867416, + "loss": 0.9082, + "step": 288 + }, + { + "epoch": 0.16080122409236333, + "grad_norm": 0.36328125, + "learning_rate": 0.00019867629348081138, + "loss": 0.9355, + "step": 289 + }, + { + "epoch": 0.16135762971206008, + "grad_norm": 0.3515625, + "learning_rate": 0.0001986668087183246, + "loss": 0.8966, + "step": 290 + }, + { + "epoch": 0.16191403533175686, + "grad_norm": 0.3515625, + "learning_rate": 0.0001986572903244468, + "loss": 0.8964, + "step": 291 + }, + { + "epoch": 0.1624704409514536, + "grad_norm": 0.3671875, + "learning_rate": 0.00019864773830242244, + "loss": 0.9721, + "step": 292 + }, + { + "epoch": 0.16302684657115038, + "grad_norm": 0.36328125, + "learning_rate": 0.0001986381526555073, + "loss": 0.9809, + "step": 293 + }, + { + "epoch": 0.16358325219084713, + "grad_norm": 0.376953125, + "learning_rate": 0.0001986285333869688, + "loss": 0.9851, + "step": 294 + }, + { + "epoch": 0.16413965781054388, + "grad_norm": 0.349609375, + "learning_rate": 0.00019861888050008575, + "loss": 0.8559, + "step": 295 + }, + { + "epoch": 0.16469606343024065, + "grad_norm": 0.35546875, + "learning_rate": 0.00019860919399814837, + "loss": 0.9342, + "step": 296 + }, + { + "epoch": 0.1652524690499374, + "grad_norm": 0.375, + "learning_rate": 0.0001985994738844584, + "loss": 0.9079, + "step": 297 + }, + { + "epoch": 0.16580887466963418, + "grad_norm": 0.33984375, + "learning_rate": 0.00019858972016232907, + "loss": 0.8145, + "step": 298 + }, + { + "epoch": 0.16636528028933092, + "grad_norm": 0.361328125, + "learning_rate": 0.00019857993283508497, + "loss": 0.9047, + "step": 299 + }, + { + "epoch": 0.16692168590902767, + "grad_norm": 0.373046875, + "learning_rate": 0.00019857011190606215, + "loss": 0.9684, + "step": 300 + }, + { + "epoch": 0.16747809152872445, + "grad_norm": 0.380859375, + "learning_rate": 0.00019856025737860828, + "loss": 1.0272, + "step": 301 + }, + { + "epoch": 0.1680344971484212, + "grad_norm": 0.359375, + "learning_rate": 0.0001985503692560822, + "loss": 0.9, + "step": 302 + }, + { + "epoch": 0.16859090276811795, + "grad_norm": 0.376953125, + "learning_rate": 0.0001985404475418545, + "loss": 1.0117, + "step": 303 + }, + { + "epoch": 0.16914730838781472, + "grad_norm": 0.3515625, + "learning_rate": 0.00019853049223930697, + "loss": 0.8626, + "step": 304 + }, + { + "epoch": 0.16970371400751147, + "grad_norm": 0.359375, + "learning_rate": 0.00019852050335183305, + "loss": 0.9383, + "step": 305 + }, + { + "epoch": 0.17026011962720825, + "grad_norm": 0.357421875, + "learning_rate": 0.00019851048088283742, + "loss": 0.9119, + "step": 306 + }, + { + "epoch": 0.170816525246905, + "grad_norm": 0.373046875, + "learning_rate": 0.00019850042483573644, + "loss": 0.9732, + "step": 307 + }, + { + "epoch": 0.17137293086660174, + "grad_norm": 0.373046875, + "learning_rate": 0.00019849033521395773, + "loss": 0.9586, + "step": 308 + }, + { + "epoch": 0.17192933648629852, + "grad_norm": 0.37109375, + "learning_rate": 0.00019848021202094042, + "loss": 0.943, + "step": 309 + }, + { + "epoch": 0.17248574210599527, + "grad_norm": 0.361328125, + "learning_rate": 0.0001984700552601351, + "loss": 0.9244, + "step": 310 + }, + { + "epoch": 0.17304214772569204, + "grad_norm": 0.369140625, + "learning_rate": 0.00019845986493500378, + "loss": 1.0124, + "step": 311 + }, + { + "epoch": 0.1735985533453888, + "grad_norm": 0.3671875, + "learning_rate": 0.00019844964104901992, + "loss": 1.0024, + "step": 312 + }, + { + "epoch": 0.17415495896508554, + "grad_norm": 0.3671875, + "learning_rate": 0.00019843938360566843, + "loss": 0.856, + "step": 313 + }, + { + "epoch": 0.1747113645847823, + "grad_norm": 0.369140625, + "learning_rate": 0.0001984290926084456, + "loss": 0.9451, + "step": 314 + }, + { + "epoch": 0.17526777020447906, + "grad_norm": 0.3671875, + "learning_rate": 0.00019841876806085923, + "loss": 0.9995, + "step": 315 + }, + { + "epoch": 0.17582417582417584, + "grad_norm": 0.376953125, + "learning_rate": 0.00019840840996642852, + "loss": 0.9539, + "step": 316 + }, + { + "epoch": 0.17638058144387259, + "grad_norm": 0.357421875, + "learning_rate": 0.00019839801832868413, + "loss": 0.9055, + "step": 317 + }, + { + "epoch": 0.17693698706356933, + "grad_norm": 0.376953125, + "learning_rate": 0.0001983875931511681, + "loss": 0.8727, + "step": 318 + }, + { + "epoch": 0.1774933926832661, + "grad_norm": 0.36328125, + "learning_rate": 0.00019837713443743397, + "loss": 0.899, + "step": 319 + }, + { + "epoch": 0.17804979830296286, + "grad_norm": 0.3515625, + "learning_rate": 0.00019836664219104666, + "loss": 0.8607, + "step": 320 + }, + { + "epoch": 0.1786062039226596, + "grad_norm": 0.359375, + "learning_rate": 0.00019835611641558253, + "loss": 1.0124, + "step": 321 + }, + { + "epoch": 0.17916260954235638, + "grad_norm": 0.369140625, + "learning_rate": 0.0001983455571146294, + "loss": 1.0013, + "step": 322 + }, + { + "epoch": 0.17971901516205313, + "grad_norm": 0.34765625, + "learning_rate": 0.00019833496429178652, + "loss": 0.8477, + "step": 323 + }, + { + "epoch": 0.1802754207817499, + "grad_norm": 0.357421875, + "learning_rate": 0.00019832433795066446, + "loss": 0.889, + "step": 324 + }, + { + "epoch": 0.18083182640144665, + "grad_norm": 0.3359375, + "learning_rate": 0.00019831367809488537, + "loss": 0.789, + "step": 325 + }, + { + "epoch": 0.1813882320211434, + "grad_norm": 0.34765625, + "learning_rate": 0.0001983029847280827, + "loss": 0.8853, + "step": 326 + }, + { + "epoch": 0.18194463764084018, + "grad_norm": 0.3671875, + "learning_rate": 0.0001982922578539014, + "loss": 0.9427, + "step": 327 + }, + { + "epoch": 0.18250104326053693, + "grad_norm": 0.361328125, + "learning_rate": 0.0001982814974759978, + "loss": 0.8221, + "step": 328 + }, + { + "epoch": 0.1830574488802337, + "grad_norm": 0.384765625, + "learning_rate": 0.00019827070359803968, + "loss": 0.9266, + "step": 329 + }, + { + "epoch": 0.18361385449993045, + "grad_norm": 0.38671875, + "learning_rate": 0.0001982598762237062, + "loss": 1.0117, + "step": 330 + }, + { + "epoch": 0.1841702601196272, + "grad_norm": 0.416015625, + "learning_rate": 0.00019824901535668796, + "loss": 0.9188, + "step": 331 + }, + { + "epoch": 0.18472666573932398, + "grad_norm": 0.3828125, + "learning_rate": 0.000198238121000687, + "loss": 0.952, + "step": 332 + }, + { + "epoch": 0.18528307135902072, + "grad_norm": 0.353515625, + "learning_rate": 0.00019822719315941672, + "loss": 0.9938, + "step": 333 + }, + { + "epoch": 0.18583947697871747, + "grad_norm": 0.3515625, + "learning_rate": 0.000198216231836602, + "loss": 0.8886, + "step": 334 + }, + { + "epoch": 0.18639588259841425, + "grad_norm": 0.361328125, + "learning_rate": 0.00019820523703597908, + "loss": 0.8996, + "step": 335 + }, + { + "epoch": 0.186952288218111, + "grad_norm": 0.359375, + "learning_rate": 0.0001981942087612956, + "loss": 0.8635, + "step": 336 + }, + { + "epoch": 0.18750869383780777, + "grad_norm": 0.37890625, + "learning_rate": 0.0001981831470163107, + "loss": 1.029, + "step": 337 + }, + { + "epoch": 0.18806509945750452, + "grad_norm": 0.359375, + "learning_rate": 0.0001981720518047948, + "loss": 0.9771, + "step": 338 + }, + { + "epoch": 0.18862150507720127, + "grad_norm": 0.36328125, + "learning_rate": 0.0001981609231305298, + "loss": 0.919, + "step": 339 + }, + { + "epoch": 0.18917791069689804, + "grad_norm": 0.359375, + "learning_rate": 0.00019814976099730904, + "loss": 0.9555, + "step": 340 + }, + { + "epoch": 0.1897343163165948, + "grad_norm": 0.3828125, + "learning_rate": 0.00019813856540893722, + "loss": 1.0713, + "step": 341 + }, + { + "epoch": 0.19029072193629157, + "grad_norm": 0.36328125, + "learning_rate": 0.00019812733636923044, + "loss": 0.9679, + "step": 342 + }, + { + "epoch": 0.19084712755598832, + "grad_norm": 0.3515625, + "learning_rate": 0.0001981160738820162, + "loss": 0.9542, + "step": 343 + }, + { + "epoch": 0.19140353317568506, + "grad_norm": 0.37109375, + "learning_rate": 0.0001981047779511334, + "loss": 0.991, + "step": 344 + }, + { + "epoch": 0.19195993879538184, + "grad_norm": 0.498046875, + "learning_rate": 0.0001980934485804324, + "loss": 0.9589, + "step": 345 + }, + { + "epoch": 0.1925163444150786, + "grad_norm": 0.3515625, + "learning_rate": 0.00019808208577377486, + "loss": 0.9242, + "step": 346 + }, + { + "epoch": 0.19307275003477536, + "grad_norm": 0.375, + "learning_rate": 0.0001980706895350339, + "loss": 0.9109, + "step": 347 + }, + { + "epoch": 0.1936291556544721, + "grad_norm": 0.37890625, + "learning_rate": 0.00019805925986809403, + "loss": 1.0326, + "step": 348 + }, + { + "epoch": 0.19418556127416886, + "grad_norm": 0.369140625, + "learning_rate": 0.00019804779677685112, + "loss": 0.9841, + "step": 349 + }, + { + "epoch": 0.19474196689386564, + "grad_norm": 0.359375, + "learning_rate": 0.0001980363002652125, + "loss": 0.9203, + "step": 350 + }, + { + "epoch": 0.19529837251356238, + "grad_norm": 0.33984375, + "learning_rate": 0.0001980247703370968, + "loss": 0.8824, + "step": 351 + }, + { + "epoch": 0.19585477813325913, + "grad_norm": 0.326171875, + "learning_rate": 0.00019801320699643416, + "loss": 0.7956, + "step": 352 + }, + { + "epoch": 0.1964111837529559, + "grad_norm": 0.369140625, + "learning_rate": 0.00019800161024716598, + "loss": 0.922, + "step": 353 + }, + { + "epoch": 0.19696758937265266, + "grad_norm": 0.369140625, + "learning_rate": 0.00019798998009324512, + "loss": 0.9232, + "step": 354 + }, + { + "epoch": 0.19752399499234943, + "grad_norm": 0.34765625, + "learning_rate": 0.00019797831653863582, + "loss": 0.8868, + "step": 355 + }, + { + "epoch": 0.19808040061204618, + "grad_norm": 0.3515625, + "learning_rate": 0.0001979666195873137, + "loss": 0.9103, + "step": 356 + }, + { + "epoch": 0.19863680623174293, + "grad_norm": 0.353515625, + "learning_rate": 0.00019795488924326574, + "loss": 0.8998, + "step": 357 + }, + { + "epoch": 0.1991932118514397, + "grad_norm": 0.361328125, + "learning_rate": 0.00019794312551049039, + "loss": 0.8973, + "step": 358 + }, + { + "epoch": 0.19974961747113645, + "grad_norm": 0.369140625, + "learning_rate": 0.00019793132839299732, + "loss": 0.9458, + "step": 359 + }, + { + "epoch": 0.20030602309083323, + "grad_norm": 0.3671875, + "learning_rate": 0.00019791949789480775, + "loss": 1.0083, + "step": 360 + }, + { + "epoch": 0.20086242871052998, + "grad_norm": 0.35546875, + "learning_rate": 0.00019790763401995415, + "loss": 0.9661, + "step": 361 + }, + { + "epoch": 0.20141883433022673, + "grad_norm": 0.365234375, + "learning_rate": 0.00019789573677248047, + "loss": 0.8475, + "step": 362 + }, + { + "epoch": 0.2019752399499235, + "grad_norm": 0.345703125, + "learning_rate": 0.00019788380615644196, + "loss": 0.8632, + "step": 363 + }, + { + "epoch": 0.20253164556962025, + "grad_norm": 0.359375, + "learning_rate": 0.00019787184217590523, + "loss": 0.9522, + "step": 364 + }, + { + "epoch": 0.20308805118931703, + "grad_norm": 0.3671875, + "learning_rate": 0.00019785984483494838, + "loss": 0.8749, + "step": 365 + }, + { + "epoch": 0.20364445680901377, + "grad_norm": 0.337890625, + "learning_rate": 0.00019784781413766073, + "loss": 0.8714, + "step": 366 + }, + { + "epoch": 0.20420086242871052, + "grad_norm": 0.373046875, + "learning_rate": 0.0001978357500881431, + "loss": 0.9581, + "step": 367 + }, + { + "epoch": 0.2047572680484073, + "grad_norm": 0.37109375, + "learning_rate": 0.00019782365269050756, + "loss": 1.0496, + "step": 368 + }, + { + "epoch": 0.20531367366810405, + "grad_norm": 0.34765625, + "learning_rate": 0.00019781152194887764, + "loss": 0.8804, + "step": 369 + }, + { + "epoch": 0.2058700792878008, + "grad_norm": 0.3359375, + "learning_rate": 0.00019779935786738822, + "loss": 0.8474, + "step": 370 + }, + { + "epoch": 0.20642648490749757, + "grad_norm": 0.349609375, + "learning_rate": 0.0001977871604501855, + "loss": 0.8413, + "step": 371 + }, + { + "epoch": 0.20698289052719432, + "grad_norm": 0.3515625, + "learning_rate": 0.00019777492970142707, + "loss": 0.8636, + "step": 372 + }, + { + "epoch": 0.2075392961468911, + "grad_norm": 0.345703125, + "learning_rate": 0.00019776266562528192, + "loss": 0.8351, + "step": 373 + }, + { + "epoch": 0.20809570176658784, + "grad_norm": 0.330078125, + "learning_rate": 0.0001977503682259303, + "loss": 0.8751, + "step": 374 + }, + { + "epoch": 0.2086521073862846, + "grad_norm": 0.3515625, + "learning_rate": 0.00019773803750756393, + "loss": 0.9113, + "step": 375 + }, + { + "epoch": 0.20920851300598137, + "grad_norm": 0.3671875, + "learning_rate": 0.0001977256734743858, + "loss": 0.9433, + "step": 376 + }, + { + "epoch": 0.20976491862567812, + "grad_norm": 0.34375, + "learning_rate": 0.0001977132761306103, + "loss": 0.8863, + "step": 377 + }, + { + "epoch": 0.2103213242453749, + "grad_norm": 0.34765625, + "learning_rate": 0.00019770084548046318, + "loss": 0.8073, + "step": 378 + }, + { + "epoch": 0.21087772986507164, + "grad_norm": 0.345703125, + "learning_rate": 0.00019768838152818153, + "loss": 0.9198, + "step": 379 + }, + { + "epoch": 0.2114341354847684, + "grad_norm": 0.357421875, + "learning_rate": 0.0001976758842780138, + "loss": 0.9204, + "step": 380 + }, + { + "epoch": 0.21199054110446516, + "grad_norm": 0.359375, + "learning_rate": 0.00019766335373421973, + "loss": 0.9714, + "step": 381 + }, + { + "epoch": 0.2125469467241619, + "grad_norm": 0.3671875, + "learning_rate": 0.00019765078990107048, + "loss": 1.0368, + "step": 382 + }, + { + "epoch": 0.21310335234385866, + "grad_norm": 0.353515625, + "learning_rate": 0.0001976381927828486, + "loss": 0.8881, + "step": 383 + }, + { + "epoch": 0.21365975796355544, + "grad_norm": 0.359375, + "learning_rate": 0.00019762556238384784, + "loss": 1.0236, + "step": 384 + }, + { + "epoch": 0.21421616358325218, + "grad_norm": 0.37109375, + "learning_rate": 0.0001976128987083734, + "loss": 0.9628, + "step": 385 + }, + { + "epoch": 0.21477256920294896, + "grad_norm": 0.349609375, + "learning_rate": 0.0001976002017607418, + "loss": 0.8861, + "step": 386 + }, + { + "epoch": 0.2153289748226457, + "grad_norm": 0.380859375, + "learning_rate": 0.00019758747154528092, + "loss": 1.0946, + "step": 387 + }, + { + "epoch": 0.21588538044234246, + "grad_norm": 0.36328125, + "learning_rate": 0.00019757470806632994, + "loss": 1.0047, + "step": 388 + }, + { + "epoch": 0.21644178606203923, + "grad_norm": 0.35546875, + "learning_rate": 0.00019756191132823937, + "loss": 0.8831, + "step": 389 + }, + { + "epoch": 0.21699819168173598, + "grad_norm": 0.373046875, + "learning_rate": 0.00019754908133537113, + "loss": 0.8999, + "step": 390 + }, + { + "epoch": 0.21755459730143276, + "grad_norm": 0.361328125, + "learning_rate": 0.00019753621809209842, + "loss": 0.8368, + "step": 391 + }, + { + "epoch": 0.2181110029211295, + "grad_norm": 0.345703125, + "learning_rate": 0.00019752332160280576, + "loss": 0.9262, + "step": 392 + }, + { + "epoch": 0.21866740854082625, + "grad_norm": 0.365234375, + "learning_rate": 0.000197510391871889, + "loss": 0.9496, + "step": 393 + }, + { + "epoch": 0.21922381416052303, + "grad_norm": 0.353515625, + "learning_rate": 0.0001974974289037554, + "loss": 0.9303, + "step": 394 + }, + { + "epoch": 0.21978021978021978, + "grad_norm": 0.375, + "learning_rate": 0.0001974844327028235, + "loss": 0.924, + "step": 395 + }, + { + "epoch": 0.22033662539991655, + "grad_norm": 0.3671875, + "learning_rate": 0.00019747140327352306, + "loss": 0.9763, + "step": 396 + }, + { + "epoch": 0.2208930310196133, + "grad_norm": 0.3359375, + "learning_rate": 0.0001974583406202954, + "loss": 0.7589, + "step": 397 + }, + { + "epoch": 0.22144943663931005, + "grad_norm": 0.35546875, + "learning_rate": 0.00019744524474759296, + "loss": 0.9391, + "step": 398 + }, + { + "epoch": 0.22200584225900682, + "grad_norm": 0.376953125, + "learning_rate": 0.0001974321156598796, + "loss": 0.9961, + "step": 399 + }, + { + "epoch": 0.22256224787870357, + "grad_norm": 0.35546875, + "learning_rate": 0.00019741895336163046, + "loss": 0.9214, + "step": 400 + }, + { + "epoch": 0.22311865349840032, + "grad_norm": 0.359375, + "learning_rate": 0.000197405757857332, + "loss": 0.8915, + "step": 401 + }, + { + "epoch": 0.2236750591180971, + "grad_norm": 0.337890625, + "learning_rate": 0.00019739252915148208, + "loss": 0.8334, + "step": 402 + }, + { + "epoch": 0.22423146473779385, + "grad_norm": 0.359375, + "learning_rate": 0.00019737926724858976, + "loss": 0.9007, + "step": 403 + }, + { + "epoch": 0.22478787035749062, + "grad_norm": 0.353515625, + "learning_rate": 0.0001973659721531755, + "loss": 0.8694, + "step": 404 + }, + { + "epoch": 0.22534427597718737, + "grad_norm": 0.34375, + "learning_rate": 0.00019735264386977098, + "loss": 0.9033, + "step": 405 + }, + { + "epoch": 0.22590068159688412, + "grad_norm": 0.353515625, + "learning_rate": 0.00019733928240291932, + "loss": 0.8603, + "step": 406 + }, + { + "epoch": 0.2264570872165809, + "grad_norm": 0.365234375, + "learning_rate": 0.00019732588775717486, + "loss": 1.024, + "step": 407 + }, + { + "epoch": 0.22701349283627764, + "grad_norm": 0.345703125, + "learning_rate": 0.00019731245993710327, + "loss": 0.8826, + "step": 408 + }, + { + "epoch": 0.22756989845597442, + "grad_norm": 0.345703125, + "learning_rate": 0.00019729899894728158, + "loss": 0.8364, + "step": 409 + }, + { + "epoch": 0.22812630407567117, + "grad_norm": 0.373046875, + "learning_rate": 0.000197285504792298, + "loss": 1.0239, + "step": 410 + }, + { + "epoch": 0.22868270969536791, + "grad_norm": 0.3515625, + "learning_rate": 0.00019727197747675218, + "loss": 0.9286, + "step": 411 + }, + { + "epoch": 0.2292391153150647, + "grad_norm": 0.3515625, + "learning_rate": 0.000197258417005255, + "loss": 0.8623, + "step": 412 + }, + { + "epoch": 0.22979552093476144, + "grad_norm": 0.349609375, + "learning_rate": 0.00019724482338242866, + "loss": 0.8571, + "step": 413 + }, + { + "epoch": 0.2303519265544582, + "grad_norm": 0.36328125, + "learning_rate": 0.00019723119661290664, + "loss": 1.0023, + "step": 414 + }, + { + "epoch": 0.23090833217415496, + "grad_norm": 0.345703125, + "learning_rate": 0.00019721753670133376, + "loss": 0.8866, + "step": 415 + }, + { + "epoch": 0.2314647377938517, + "grad_norm": 0.33203125, + "learning_rate": 0.0001972038436523661, + "loss": 0.7956, + "step": 416 + }, + { + "epoch": 0.2320211434135485, + "grad_norm": 0.369140625, + "learning_rate": 0.00019719011747067108, + "loss": 0.9645, + "step": 417 + }, + { + "epoch": 0.23257754903324523, + "grad_norm": 0.37890625, + "learning_rate": 0.00019717635816092732, + "loss": 0.9324, + "step": 418 + }, + { + "epoch": 0.23313395465294198, + "grad_norm": 0.361328125, + "learning_rate": 0.00019716256572782482, + "loss": 0.9626, + "step": 419 + }, + { + "epoch": 0.23369036027263876, + "grad_norm": 0.34375, + "learning_rate": 0.0001971487401760649, + "loss": 0.8419, + "step": 420 + }, + { + "epoch": 0.2342467658923355, + "grad_norm": 0.33984375, + "learning_rate": 0.00019713488151036003, + "loss": 0.8471, + "step": 421 + }, + { + "epoch": 0.23480317151203228, + "grad_norm": 0.361328125, + "learning_rate": 0.00019712098973543408, + "loss": 1.0177, + "step": 422 + }, + { + "epoch": 0.23535957713172903, + "grad_norm": 0.35546875, + "learning_rate": 0.0001971070648560222, + "loss": 0.944, + "step": 423 + }, + { + "epoch": 0.23591598275142578, + "grad_norm": 0.3359375, + "learning_rate": 0.0001970931068768708, + "loss": 0.8007, + "step": 424 + }, + { + "epoch": 0.23647238837112255, + "grad_norm": 0.3671875, + "learning_rate": 0.0001970791158027375, + "loss": 0.9658, + "step": 425 + }, + { + "epoch": 0.2370287939908193, + "grad_norm": 0.35546875, + "learning_rate": 0.00019706509163839136, + "loss": 0.8966, + "step": 426 + }, + { + "epoch": 0.23758519961051608, + "grad_norm": 0.365234375, + "learning_rate": 0.0001970510343886126, + "loss": 1.0448, + "step": 427 + }, + { + "epoch": 0.23814160523021283, + "grad_norm": 0.37109375, + "learning_rate": 0.0001970369440581927, + "loss": 0.937, + "step": 428 + }, + { + "epoch": 0.23869801084990958, + "grad_norm": 0.3515625, + "learning_rate": 0.00019702282065193455, + "loss": 0.9181, + "step": 429 + }, + { + "epoch": 0.23925441646960635, + "grad_norm": 0.34375, + "learning_rate": 0.00019700866417465216, + "loss": 0.9101, + "step": 430 + }, + { + "epoch": 0.2398108220893031, + "grad_norm": 0.359375, + "learning_rate": 0.00019699447463117093, + "loss": 0.9438, + "step": 431 + }, + { + "epoch": 0.24036722770899985, + "grad_norm": 0.34765625, + "learning_rate": 0.00019698025202632743, + "loss": 0.8315, + "step": 432 + }, + { + "epoch": 0.24092363332869662, + "grad_norm": 0.353515625, + "learning_rate": 0.00019696599636496962, + "loss": 0.8385, + "step": 433 + }, + { + "epoch": 0.24148003894839337, + "grad_norm": 0.33984375, + "learning_rate": 0.0001969517076519566, + "loss": 0.8667, + "step": 434 + }, + { + "epoch": 0.24203644456809015, + "grad_norm": 0.349609375, + "learning_rate": 0.00019693738589215883, + "loss": 0.881, + "step": 435 + }, + { + "epoch": 0.2425928501877869, + "grad_norm": 0.36328125, + "learning_rate": 0.00019692303109045795, + "loss": 0.9509, + "step": 436 + }, + { + "epoch": 0.24314925580748364, + "grad_norm": 0.357421875, + "learning_rate": 0.00019690864325174699, + "loss": 0.9065, + "step": 437 + }, + { + "epoch": 0.24370566142718042, + "grad_norm": 0.380859375, + "learning_rate": 0.0001968942223809301, + "loss": 1.072, + "step": 438 + }, + { + "epoch": 0.24426206704687717, + "grad_norm": 0.333984375, + "learning_rate": 0.0001968797684829228, + "loss": 0.9068, + "step": 439 + }, + { + "epoch": 0.24481847266657394, + "grad_norm": 0.33203125, + "learning_rate": 0.00019686528156265175, + "loss": 0.8168, + "step": 440 + }, + { + "epoch": 0.2453748782862707, + "grad_norm": 0.3515625, + "learning_rate": 0.000196850761625055, + "loss": 0.8748, + "step": 441 + }, + { + "epoch": 0.24593128390596744, + "grad_norm": 0.330078125, + "learning_rate": 0.00019683620867508176, + "loss": 0.7686, + "step": 442 + }, + { + "epoch": 0.24648768952566422, + "grad_norm": 0.365234375, + "learning_rate": 0.00019682162271769256, + "loss": 0.9036, + "step": 443 + }, + { + "epoch": 0.24704409514536096, + "grad_norm": 0.357421875, + "learning_rate": 0.0001968070037578591, + "loss": 0.9472, + "step": 444 + }, + { + "epoch": 0.24760050076505774, + "grad_norm": 0.33984375, + "learning_rate": 0.00019679235180056437, + "loss": 0.8882, + "step": 445 + }, + { + "epoch": 0.2481569063847545, + "grad_norm": 0.36328125, + "learning_rate": 0.00019677766685080264, + "loss": 0.9693, + "step": 446 + }, + { + "epoch": 0.24871331200445124, + "grad_norm": 0.357421875, + "learning_rate": 0.00019676294891357937, + "loss": 0.8764, + "step": 447 + }, + { + "epoch": 0.249269717624148, + "grad_norm": 0.38671875, + "learning_rate": 0.00019674819799391129, + "loss": 0.9671, + "step": 448 + }, + { + "epoch": 0.24982612324384476, + "grad_norm": 0.345703125, + "learning_rate": 0.0001967334140968264, + "loss": 0.9398, + "step": 449 + }, + { + "epoch": 0.2503825288635415, + "grad_norm": 0.345703125, + "learning_rate": 0.00019671859722736389, + "loss": 0.9572, + "step": 450 + }, + { + "epoch": 0.2509389344832383, + "grad_norm": 0.357421875, + "learning_rate": 0.0001967037473905742, + "loss": 0.958, + "step": 451 + }, + { + "epoch": 0.25149534010293506, + "grad_norm": 0.35546875, + "learning_rate": 0.00019668886459151908, + "loss": 0.8959, + "step": 452 + }, + { + "epoch": 0.2520517457226318, + "grad_norm": 0.35546875, + "learning_rate": 0.00019667394883527139, + "loss": 0.9529, + "step": 453 + }, + { + "epoch": 0.25260815134232856, + "grad_norm": 0.361328125, + "learning_rate": 0.00019665900012691529, + "loss": 0.9213, + "step": 454 + }, + { + "epoch": 0.25316455696202533, + "grad_norm": 0.359375, + "learning_rate": 0.0001966440184715462, + "loss": 0.952, + "step": 455 + }, + { + "epoch": 0.25372096258172205, + "grad_norm": 0.365234375, + "learning_rate": 0.00019662900387427075, + "loss": 0.9548, + "step": 456 + }, + { + "epoch": 0.25427736820141883, + "grad_norm": 0.3359375, + "learning_rate": 0.00019661395634020673, + "loss": 0.8359, + "step": 457 + }, + { + "epoch": 0.2548337738211156, + "grad_norm": 0.345703125, + "learning_rate": 0.00019659887587448327, + "loss": 0.9204, + "step": 458 + }, + { + "epoch": 0.2553901794408123, + "grad_norm": 0.359375, + "learning_rate": 0.00019658376248224063, + "loss": 0.9216, + "step": 459 + }, + { + "epoch": 0.2559465850605091, + "grad_norm": 0.365234375, + "learning_rate": 0.00019656861616863038, + "loss": 0.9383, + "step": 460 + }, + { + "epoch": 0.2565029906802059, + "grad_norm": 0.35546875, + "learning_rate": 0.00019655343693881526, + "loss": 0.9428, + "step": 461 + }, + { + "epoch": 0.25705939629990265, + "grad_norm": 0.3515625, + "learning_rate": 0.0001965382247979692, + "loss": 0.9357, + "step": 462 + }, + { + "epoch": 0.2576158019195994, + "grad_norm": 0.35546875, + "learning_rate": 0.00019652297975127737, + "loss": 0.967, + "step": 463 + }, + { + "epoch": 0.25817220753929615, + "grad_norm": 0.349609375, + "learning_rate": 0.0001965077018039362, + "loss": 0.8993, + "step": 464 + }, + { + "epoch": 0.2587286131589929, + "grad_norm": 0.359375, + "learning_rate": 0.00019649239096115336, + "loss": 1.0113, + "step": 465 + }, + { + "epoch": 0.25928501877868965, + "grad_norm": 0.33984375, + "learning_rate": 0.00019647704722814755, + "loss": 0.9117, + "step": 466 + }, + { + "epoch": 0.2598414243983864, + "grad_norm": 0.345703125, + "learning_rate": 0.00019646167061014892, + "loss": 0.9318, + "step": 467 + }, + { + "epoch": 0.2603978300180832, + "grad_norm": 0.34375, + "learning_rate": 0.00019644626111239863, + "loss": 0.9237, + "step": 468 + }, + { + "epoch": 0.2609542356377799, + "grad_norm": 0.37109375, + "learning_rate": 0.0001964308187401492, + "loss": 0.9926, + "step": 469 + }, + { + "epoch": 0.2615106412574767, + "grad_norm": 0.337890625, + "learning_rate": 0.0001964153434986643, + "loss": 0.8647, + "step": 470 + }, + { + "epoch": 0.26206704687717347, + "grad_norm": 0.330078125, + "learning_rate": 0.0001963998353932187, + "loss": 0.7973, + "step": 471 + }, + { + "epoch": 0.2626234524968702, + "grad_norm": 0.35546875, + "learning_rate": 0.00019638429442909853, + "loss": 0.8922, + "step": 472 + }, + { + "epoch": 0.26317985811656697, + "grad_norm": 0.3671875, + "learning_rate": 0.00019636872061160109, + "loss": 0.9267, + "step": 473 + }, + { + "epoch": 0.26373626373626374, + "grad_norm": 0.34765625, + "learning_rate": 0.00019635311394603476, + "loss": 0.8743, + "step": 474 + }, + { + "epoch": 0.2642926693559605, + "grad_norm": 0.33203125, + "learning_rate": 0.00019633747443771924, + "loss": 0.7997, + "step": 475 + }, + { + "epoch": 0.26484907497565724, + "grad_norm": 0.36328125, + "learning_rate": 0.00019632180209198542, + "loss": 0.9695, + "step": 476 + }, + { + "epoch": 0.265405480595354, + "grad_norm": 0.3359375, + "learning_rate": 0.00019630609691417527, + "loss": 0.7992, + "step": 477 + }, + { + "epoch": 0.2659618862150508, + "grad_norm": 0.3125, + "learning_rate": 0.00019629035890964208, + "loss": 0.7188, + "step": 478 + }, + { + "epoch": 0.2665182918347475, + "grad_norm": 0.384765625, + "learning_rate": 0.0001962745880837503, + "loss": 0.93, + "step": 479 + }, + { + "epoch": 0.2670746974544443, + "grad_norm": 0.341796875, + "learning_rate": 0.00019625878444187551, + "loss": 0.8121, + "step": 480 + }, + { + "epoch": 0.26763110307414106, + "grad_norm": 0.357421875, + "learning_rate": 0.00019624294798940451, + "loss": 0.9366, + "step": 481 + }, + { + "epoch": 0.2681875086938378, + "grad_norm": 0.359375, + "learning_rate": 0.0001962270787317353, + "loss": 0.94, + "step": 482 + }, + { + "epoch": 0.26874391431353456, + "grad_norm": 0.357421875, + "learning_rate": 0.000196211176674277, + "loss": 0.9317, + "step": 483 + }, + { + "epoch": 0.26930031993323134, + "grad_norm": 0.34375, + "learning_rate": 0.00019619524182245, + "loss": 0.8751, + "step": 484 + }, + { + "epoch": 0.2698567255529281, + "grad_norm": 0.345703125, + "learning_rate": 0.00019617927418168587, + "loss": 0.9093, + "step": 485 + }, + { + "epoch": 0.27041313117262483, + "grad_norm": 0.33984375, + "learning_rate": 0.00019616327375742722, + "loss": 0.8399, + "step": 486 + }, + { + "epoch": 0.2709695367923216, + "grad_norm": 0.353515625, + "learning_rate": 0.00019614724055512799, + "loss": 0.9025, + "step": 487 + }, + { + "epoch": 0.2715259424120184, + "grad_norm": 0.353515625, + "learning_rate": 0.0001961311745802532, + "loss": 0.8661, + "step": 488 + }, + { + "epoch": 0.2720823480317151, + "grad_norm": 0.34375, + "learning_rate": 0.00019611507583827903, + "loss": 0.8429, + "step": 489 + }, + { + "epoch": 0.2726387536514119, + "grad_norm": 0.3515625, + "learning_rate": 0.00019609894433469295, + "loss": 0.9272, + "step": 490 + }, + { + "epoch": 0.27319515927110866, + "grad_norm": 0.353515625, + "learning_rate": 0.0001960827800749935, + "loss": 0.87, + "step": 491 + }, + { + "epoch": 0.2737515648908054, + "grad_norm": 0.373046875, + "learning_rate": 0.00019606658306469034, + "loss": 0.865, + "step": 492 + }, + { + "epoch": 0.27430797051050215, + "grad_norm": 0.328125, + "learning_rate": 0.00019605035330930442, + "loss": 0.8939, + "step": 493 + }, + { + "epoch": 0.27486437613019893, + "grad_norm": 0.3515625, + "learning_rate": 0.00019603409081436773, + "loss": 0.9009, + "step": 494 + }, + { + "epoch": 0.27542078174989565, + "grad_norm": 0.349609375, + "learning_rate": 0.0001960177955854235, + "loss": 0.8828, + "step": 495 + }, + { + "epoch": 0.2759771873695924, + "grad_norm": 0.349609375, + "learning_rate": 0.00019600146762802613, + "loss": 0.9509, + "step": 496 + }, + { + "epoch": 0.2765335929892892, + "grad_norm": 0.333984375, + "learning_rate": 0.00019598510694774108, + "loss": 0.7998, + "step": 497 + }, + { + "epoch": 0.277089998608986, + "grad_norm": 0.349609375, + "learning_rate": 0.00019596871355014506, + "loss": 0.8909, + "step": 498 + }, + { + "epoch": 0.2776464042286827, + "grad_norm": 0.36328125, + "learning_rate": 0.00019595228744082586, + "loss": 0.9643, + "step": 499 + }, + { + "epoch": 0.2782028098483795, + "grad_norm": 0.3515625, + "learning_rate": 0.00019593582862538248, + "loss": 0.9311, + "step": 500 + }, + { + "epoch": 0.27875921546807625, + "grad_norm": 0.33984375, + "learning_rate": 0.00019591933710942503, + "loss": 0.8344, + "step": 501 + }, + { + "epoch": 0.27931562108777297, + "grad_norm": 0.34765625, + "learning_rate": 0.00019590281289857478, + "loss": 0.9109, + "step": 502 + }, + { + "epoch": 0.27987202670746975, + "grad_norm": 0.357421875, + "learning_rate": 0.00019588625599846417, + "loss": 0.994, + "step": 503 + }, + { + "epoch": 0.2804284323271665, + "grad_norm": 0.35546875, + "learning_rate": 0.00019586966641473673, + "loss": 0.8917, + "step": 504 + }, + { + "epoch": 0.28098483794686324, + "grad_norm": 0.33984375, + "learning_rate": 0.00019585304415304711, + "loss": 0.8417, + "step": 505 + }, + { + "epoch": 0.28154124356656, + "grad_norm": 0.341796875, + "learning_rate": 0.00019583638921906124, + "loss": 0.9425, + "step": 506 + }, + { + "epoch": 0.2820976491862568, + "grad_norm": 0.333984375, + "learning_rate": 0.00019581970161845603, + "loss": 0.8792, + "step": 507 + }, + { + "epoch": 0.2826540548059535, + "grad_norm": 0.349609375, + "learning_rate": 0.0001958029813569196, + "loss": 0.9791, + "step": 508 + }, + { + "epoch": 0.2832104604256503, + "grad_norm": 0.330078125, + "learning_rate": 0.00019578622844015117, + "loss": 0.8546, + "step": 509 + }, + { + "epoch": 0.28376686604534707, + "grad_norm": 0.359375, + "learning_rate": 0.00019576944287386113, + "loss": 0.8836, + "step": 510 + }, + { + "epoch": 0.28432327166504384, + "grad_norm": 0.341796875, + "learning_rate": 0.000195752624663771, + "loss": 0.8764, + "step": 511 + }, + { + "epoch": 0.28487967728474056, + "grad_norm": 0.369140625, + "learning_rate": 0.00019573577381561336, + "loss": 0.9411, + "step": 512 + }, + { + "epoch": 0.28543608290443734, + "grad_norm": 0.345703125, + "learning_rate": 0.00019571889033513198, + "loss": 0.9222, + "step": 513 + }, + { + "epoch": 0.2859924885241341, + "grad_norm": 0.365234375, + "learning_rate": 0.00019570197422808173, + "loss": 0.9914, + "step": 514 + }, + { + "epoch": 0.28654889414383083, + "grad_norm": 0.34765625, + "learning_rate": 0.00019568502550022858, + "loss": 0.8788, + "step": 515 + }, + { + "epoch": 0.2871052997635276, + "grad_norm": 0.3359375, + "learning_rate": 0.0001956680441573497, + "loss": 0.811, + "step": 516 + }, + { + "epoch": 0.2876617053832244, + "grad_norm": 0.3515625, + "learning_rate": 0.00019565103020523328, + "loss": 0.9191, + "step": 517 + }, + { + "epoch": 0.2882181110029211, + "grad_norm": 0.345703125, + "learning_rate": 0.00019563398364967868, + "loss": 0.9129, + "step": 518 + }, + { + "epoch": 0.2887745166226179, + "grad_norm": 0.357421875, + "learning_rate": 0.00019561690449649636, + "loss": 0.9591, + "step": 519 + }, + { + "epoch": 0.28933092224231466, + "grad_norm": 0.326171875, + "learning_rate": 0.00019559979275150783, + "loss": 0.8129, + "step": 520 + }, + { + "epoch": 0.2898873278620114, + "grad_norm": 0.349609375, + "learning_rate": 0.00019558264842054585, + "loss": 0.8881, + "step": 521 + }, + { + "epoch": 0.29044373348170816, + "grad_norm": 0.37890625, + "learning_rate": 0.00019556547150945418, + "loss": 0.8941, + "step": 522 + }, + { + "epoch": 0.29100013910140493, + "grad_norm": 0.33984375, + "learning_rate": 0.0001955482620240877, + "loss": 0.8872, + "step": 523 + }, + { + "epoch": 0.2915565447211017, + "grad_norm": 0.3359375, + "learning_rate": 0.0001955310199703124, + "loss": 0.7887, + "step": 524 + }, + { + "epoch": 0.2921129503407984, + "grad_norm": 0.34375, + "learning_rate": 0.00019551374535400543, + "loss": 0.8684, + "step": 525 + }, + { + "epoch": 0.2926693559604952, + "grad_norm": 0.3203125, + "learning_rate": 0.0001954964381810549, + "loss": 0.7739, + "step": 526 + }, + { + "epoch": 0.293225761580192, + "grad_norm": 0.33203125, + "learning_rate": 0.00019547909845736013, + "loss": 0.8568, + "step": 527 + }, + { + "epoch": 0.2937821671998887, + "grad_norm": 0.353515625, + "learning_rate": 0.00019546172618883155, + "loss": 0.9188, + "step": 528 + }, + { + "epoch": 0.2943385728195855, + "grad_norm": 0.330078125, + "learning_rate": 0.00019544432138139062, + "loss": 0.8119, + "step": 529 + }, + { + "epoch": 0.29489497843928225, + "grad_norm": 0.34765625, + "learning_rate": 0.00019542688404096986, + "loss": 0.9192, + "step": 530 + }, + { + "epoch": 0.29545138405897897, + "grad_norm": 0.33984375, + "learning_rate": 0.000195409414173513, + "loss": 0.8616, + "step": 531 + }, + { + "epoch": 0.29600778967867575, + "grad_norm": 0.34375, + "learning_rate": 0.00019539191178497472, + "loss": 0.9191, + "step": 532 + }, + { + "epoch": 0.2965641952983725, + "grad_norm": 0.32421875, + "learning_rate": 0.00019537437688132094, + "loss": 0.7765, + "step": 533 + }, + { + "epoch": 0.2971206009180693, + "grad_norm": 0.34765625, + "learning_rate": 0.00019535680946852852, + "loss": 0.9185, + "step": 534 + }, + { + "epoch": 0.297677006537766, + "grad_norm": 0.353515625, + "learning_rate": 0.00019533920955258546, + "loss": 0.8514, + "step": 535 + }, + { + "epoch": 0.2982334121574628, + "grad_norm": 0.80078125, + "learning_rate": 0.00019532157713949083, + "loss": 1.8071, + "step": 536 + }, + { + "epoch": 0.29878981777715957, + "grad_norm": 0.361328125, + "learning_rate": 0.00019530391223525478, + "loss": 0.9006, + "step": 537 + }, + { + "epoch": 0.2993462233968563, + "grad_norm": 0.357421875, + "learning_rate": 0.00019528621484589859, + "loss": 0.9186, + "step": 538 + }, + { + "epoch": 0.29990262901655307, + "grad_norm": 0.333984375, + "learning_rate": 0.00019526848497745448, + "loss": 0.8292, + "step": 539 + }, + { + "epoch": 0.30045903463624984, + "grad_norm": 0.3359375, + "learning_rate": 0.00019525072263596585, + "loss": 0.7303, + "step": 540 + }, + { + "epoch": 0.30101544025594656, + "grad_norm": 0.357421875, + "learning_rate": 0.0001952329278274872, + "loss": 0.9214, + "step": 541 + }, + { + "epoch": 0.30157184587564334, + "grad_norm": 0.3515625, + "learning_rate": 0.00019521510055808396, + "loss": 0.9173, + "step": 542 + }, + { + "epoch": 0.3021282514953401, + "grad_norm": 0.345703125, + "learning_rate": 0.00019519724083383273, + "loss": 0.8349, + "step": 543 + }, + { + "epoch": 0.30268465711503684, + "grad_norm": 0.353515625, + "learning_rate": 0.00019517934866082116, + "loss": 0.9247, + "step": 544 + }, + { + "epoch": 0.3032410627347336, + "grad_norm": 0.3671875, + "learning_rate": 0.0001951614240451479, + "loss": 0.8788, + "step": 545 + }, + { + "epoch": 0.3037974683544304, + "grad_norm": 0.33203125, + "learning_rate": 0.00019514346699292274, + "loss": 0.8521, + "step": 546 + }, + { + "epoch": 0.30435387397412716, + "grad_norm": 0.34375, + "learning_rate": 0.00019512547751026648, + "loss": 0.8488, + "step": 547 + }, + { + "epoch": 0.3049102795938239, + "grad_norm": 0.3671875, + "learning_rate": 0.00019510745560331097, + "loss": 0.9655, + "step": 548 + }, + { + "epoch": 0.30546668521352066, + "grad_norm": 0.375, + "learning_rate": 0.00019508940127819913, + "loss": 1.0282, + "step": 549 + }, + { + "epoch": 0.30602309083321744, + "grad_norm": 0.37109375, + "learning_rate": 0.00019507131454108497, + "loss": 1.0152, + "step": 550 + }, + { + "epoch": 0.30657949645291416, + "grad_norm": 0.330078125, + "learning_rate": 0.0001950531953981334, + "loss": 0.8472, + "step": 551 + }, + { + "epoch": 0.30713590207261093, + "grad_norm": 0.345703125, + "learning_rate": 0.0001950350438555206, + "loss": 0.8322, + "step": 552 + }, + { + "epoch": 0.3076923076923077, + "grad_norm": 0.357421875, + "learning_rate": 0.00019501685991943357, + "loss": 0.9437, + "step": 553 + }, + { + "epoch": 0.30824871331200443, + "grad_norm": 0.330078125, + "learning_rate": 0.0001949986435960705, + "loss": 0.817, + "step": 554 + }, + { + "epoch": 0.3088051189317012, + "grad_norm": 0.53125, + "learning_rate": 0.00019498039489164058, + "loss": 0.9954, + "step": 555 + }, + { + "epoch": 0.309361524551398, + "grad_norm": 0.34765625, + "learning_rate": 0.00019496211381236401, + "loss": 0.8284, + "step": 556 + }, + { + "epoch": 0.3099179301710947, + "grad_norm": 0.357421875, + "learning_rate": 0.0001949438003644721, + "loss": 0.8699, + "step": 557 + }, + { + "epoch": 0.3104743357907915, + "grad_norm": 0.3515625, + "learning_rate": 0.00019492545455420703, + "loss": 0.8738, + "step": 558 + }, + { + "epoch": 0.31103074141048825, + "grad_norm": 0.34765625, + "learning_rate": 0.0001949070763878222, + "loss": 0.894, + "step": 559 + }, + { + "epoch": 0.31158714703018503, + "grad_norm": 0.33203125, + "learning_rate": 0.00019488866587158196, + "loss": 0.8258, + "step": 560 + }, + { + "epoch": 0.31214355264988175, + "grad_norm": 0.3671875, + "learning_rate": 0.00019487022301176168, + "loss": 0.9347, + "step": 561 + }, + { + "epoch": 0.3126999582695785, + "grad_norm": 0.34765625, + "learning_rate": 0.00019485174781464773, + "loss": 0.9083, + "step": 562 + }, + { + "epoch": 0.3132563638892753, + "grad_norm": 0.330078125, + "learning_rate": 0.00019483324028653754, + "loss": 0.784, + "step": 563 + }, + { + "epoch": 0.313812769508972, + "grad_norm": 0.345703125, + "learning_rate": 0.0001948147004337396, + "loss": 0.8954, + "step": 564 + }, + { + "epoch": 0.3143691751286688, + "grad_norm": 0.345703125, + "learning_rate": 0.0001947961282625733, + "loss": 0.8873, + "step": 565 + }, + { + "epoch": 0.3149255807483656, + "grad_norm": 0.337890625, + "learning_rate": 0.00019477752377936917, + "loss": 0.9064, + "step": 566 + }, + { + "epoch": 0.3154819863680623, + "grad_norm": 0.333984375, + "learning_rate": 0.00019475888699046866, + "loss": 0.8424, + "step": 567 + }, + { + "epoch": 0.31603839198775907, + "grad_norm": 0.384765625, + "learning_rate": 0.0001947402179022243, + "loss": 0.9908, + "step": 568 + }, + { + "epoch": 0.31659479760745585, + "grad_norm": 0.353515625, + "learning_rate": 0.00019472151652099963, + "loss": 0.9116, + "step": 569 + }, + { + "epoch": 0.31715120322715257, + "grad_norm": 0.3359375, + "learning_rate": 0.0001947027828531691, + "loss": 0.9074, + "step": 570 + }, + { + "epoch": 0.31770760884684934, + "grad_norm": 0.357421875, + "learning_rate": 0.00019468401690511832, + "loss": 0.9643, + "step": 571 + }, + { + "epoch": 0.3182640144665461, + "grad_norm": 0.353515625, + "learning_rate": 0.00019466521868324375, + "loss": 0.9002, + "step": 572 + }, + { + "epoch": 0.3188204200862429, + "grad_norm": 0.357421875, + "learning_rate": 0.0001946463881939529, + "loss": 0.9181, + "step": 573 + }, + { + "epoch": 0.3193768257059396, + "grad_norm": 0.34765625, + "learning_rate": 0.00019462752544366436, + "loss": 0.9275, + "step": 574 + }, + { + "epoch": 0.3199332313256364, + "grad_norm": 0.341796875, + "learning_rate": 0.0001946086304388076, + "loss": 0.8565, + "step": 575 + }, + { + "epoch": 0.32048963694533317, + "grad_norm": 0.3671875, + "learning_rate": 0.0001945897031858232, + "loss": 0.9334, + "step": 576 + }, + { + "epoch": 0.3210460425650299, + "grad_norm": 0.35546875, + "learning_rate": 0.0001945707436911626, + "loss": 0.9068, + "step": 577 + }, + { + "epoch": 0.32160244818472666, + "grad_norm": 0.34765625, + "learning_rate": 0.00019455175196128838, + "loss": 0.8463, + "step": 578 + }, + { + "epoch": 0.32215885380442344, + "grad_norm": 0.349609375, + "learning_rate": 0.00019453272800267393, + "loss": 0.8119, + "step": 579 + }, + { + "epoch": 0.32271525942412016, + "grad_norm": 0.349609375, + "learning_rate": 0.00019451367182180378, + "loss": 0.9386, + "step": 580 + }, + { + "epoch": 0.32327166504381694, + "grad_norm": 0.349609375, + "learning_rate": 0.00019449458342517338, + "loss": 0.9527, + "step": 581 + }, + { + "epoch": 0.3238280706635137, + "grad_norm": 0.359375, + "learning_rate": 0.00019447546281928916, + "loss": 0.9448, + "step": 582 + }, + { + "epoch": 0.3243844762832105, + "grad_norm": 0.330078125, + "learning_rate": 0.00019445631001066853, + "loss": 0.8528, + "step": 583 + }, + { + "epoch": 0.3249408819029072, + "grad_norm": 0.353515625, + "learning_rate": 0.00019443712500583993, + "loss": 0.9443, + "step": 584 + }, + { + "epoch": 0.325497287522604, + "grad_norm": 0.361328125, + "learning_rate": 0.00019441790781134265, + "loss": 0.9369, + "step": 585 + }, + { + "epoch": 0.32605369314230076, + "grad_norm": 0.34765625, + "learning_rate": 0.0001943986584337271, + "loss": 0.8631, + "step": 586 + }, + { + "epoch": 0.3266100987619975, + "grad_norm": 0.359375, + "learning_rate": 0.0001943793768795545, + "loss": 0.921, + "step": 587 + }, + { + "epoch": 0.32716650438169426, + "grad_norm": 0.345703125, + "learning_rate": 0.0001943600631553972, + "loss": 0.8827, + "step": 588 + }, + { + "epoch": 0.32772291000139103, + "grad_norm": 0.36328125, + "learning_rate": 0.00019434071726783843, + "loss": 0.9897, + "step": 589 + }, + { + "epoch": 0.32827931562108775, + "grad_norm": 0.333984375, + "learning_rate": 0.0001943213392234724, + "loss": 0.7868, + "step": 590 + }, + { + "epoch": 0.32883572124078453, + "grad_norm": 0.34765625, + "learning_rate": 0.00019430192902890423, + "loss": 0.8787, + "step": 591 + }, + { + "epoch": 0.3293921268604813, + "grad_norm": 0.341796875, + "learning_rate": 0.0001942824866907501, + "loss": 0.861, + "step": 592 + }, + { + "epoch": 0.329948532480178, + "grad_norm": 0.337890625, + "learning_rate": 0.00019426301221563702, + "loss": 0.8415, + "step": 593 + }, + { + "epoch": 0.3305049380998748, + "grad_norm": 0.361328125, + "learning_rate": 0.00019424350561020312, + "loss": 0.9665, + "step": 594 + }, + { + "epoch": 0.3310613437195716, + "grad_norm": 0.353515625, + "learning_rate": 0.0001942239668810973, + "loss": 0.934, + "step": 595 + }, + { + "epoch": 0.33161774933926835, + "grad_norm": 0.353515625, + "learning_rate": 0.00019420439603497956, + "loss": 0.9518, + "step": 596 + }, + { + "epoch": 0.3321741549589651, + "grad_norm": 0.330078125, + "learning_rate": 0.00019418479307852074, + "loss": 0.8112, + "step": 597 + }, + { + "epoch": 0.33273056057866185, + "grad_norm": 0.33203125, + "learning_rate": 0.0001941651580184027, + "loss": 0.7649, + "step": 598 + }, + { + "epoch": 0.3332869661983586, + "grad_norm": 0.3515625, + "learning_rate": 0.00019414549086131815, + "loss": 0.9321, + "step": 599 + }, + { + "epoch": 0.33384337181805535, + "grad_norm": 0.3671875, + "learning_rate": 0.0001941257916139709, + "loss": 0.9537, + "step": 600 + }, + { + "epoch": 0.3343997774377521, + "grad_norm": 0.3515625, + "learning_rate": 0.0001941060602830755, + "loss": 0.961, + "step": 601 + }, + { + "epoch": 0.3349561830574489, + "grad_norm": 0.33984375, + "learning_rate": 0.00019408629687535763, + "loss": 0.9001, + "step": 602 + }, + { + "epoch": 0.3355125886771456, + "grad_norm": 0.326171875, + "learning_rate": 0.00019406650139755375, + "loss": 0.8772, + "step": 603 + }, + { + "epoch": 0.3360689942968424, + "grad_norm": 0.357421875, + "learning_rate": 0.00019404667385641128, + "loss": 0.9444, + "step": 604 + }, + { + "epoch": 0.33662539991653917, + "grad_norm": 0.3359375, + "learning_rate": 0.0001940268142586887, + "loss": 0.8729, + "step": 605 + }, + { + "epoch": 0.3371818055362359, + "grad_norm": 0.333984375, + "learning_rate": 0.0001940069226111552, + "loss": 0.8825, + "step": 606 + }, + { + "epoch": 0.33773821115593267, + "grad_norm": 0.361328125, + "learning_rate": 0.00019398699892059112, + "loss": 0.9166, + "step": 607 + }, + { + "epoch": 0.33829461677562944, + "grad_norm": 0.35546875, + "learning_rate": 0.00019396704319378754, + "loss": 0.9447, + "step": 608 + }, + { + "epoch": 0.3388510223953262, + "grad_norm": 0.3515625, + "learning_rate": 0.0001939470554375466, + "loss": 0.932, + "step": 609 + }, + { + "epoch": 0.33940742801502294, + "grad_norm": 0.345703125, + "learning_rate": 0.0001939270356586812, + "loss": 0.8981, + "step": 610 + }, + { + "epoch": 0.3399638336347197, + "grad_norm": 0.337890625, + "learning_rate": 0.00019390698386401533, + "loss": 0.8561, + "step": 611 + }, + { + "epoch": 0.3405202392544165, + "grad_norm": 0.359375, + "learning_rate": 0.00019388690006038375, + "loss": 0.8992, + "step": 612 + }, + { + "epoch": 0.3410766448741132, + "grad_norm": 0.3515625, + "learning_rate": 0.00019386678425463221, + "loss": 0.952, + "step": 613 + }, + { + "epoch": 0.34163305049381, + "grad_norm": 0.34375, + "learning_rate": 0.00019384663645361736, + "loss": 0.8244, + "step": 614 + }, + { + "epoch": 0.34218945611350676, + "grad_norm": 0.357421875, + "learning_rate": 0.00019382645666420674, + "loss": 0.9347, + "step": 615 + }, + { + "epoch": 0.3427458617332035, + "grad_norm": 0.357421875, + "learning_rate": 0.00019380624489327876, + "loss": 0.9098, + "step": 616 + }, + { + "epoch": 0.34330226735290026, + "grad_norm": 1.4375, + "learning_rate": 0.00019378600114772283, + "loss": 1.9243, + "step": 617 + }, + { + "epoch": 0.34385867297259703, + "grad_norm": 0.326171875, + "learning_rate": 0.00019376572543443915, + "loss": 0.8078, + "step": 618 + }, + { + "epoch": 0.34441507859229376, + "grad_norm": 0.333984375, + "learning_rate": 0.00019374541776033888, + "loss": 0.8494, + "step": 619 + }, + { + "epoch": 0.34497148421199053, + "grad_norm": 0.3515625, + "learning_rate": 0.00019372507813234405, + "loss": 0.8841, + "step": 620 + }, + { + "epoch": 0.3455278898316873, + "grad_norm": 0.33984375, + "learning_rate": 0.0001937047065573876, + "loss": 0.8424, + "step": 621 + }, + { + "epoch": 0.3460842954513841, + "grad_norm": 0.333984375, + "learning_rate": 0.0001936843030424133, + "loss": 0.8028, + "step": 622 + }, + { + "epoch": 0.3466407010710808, + "grad_norm": 0.337890625, + "learning_rate": 0.00019366386759437596, + "loss": 0.8883, + "step": 623 + }, + { + "epoch": 0.3471971066907776, + "grad_norm": 0.328125, + "learning_rate": 0.0001936434002202411, + "loss": 0.7923, + "step": 624 + }, + { + "epoch": 0.34775351231047436, + "grad_norm": 0.357421875, + "learning_rate": 0.00019362290092698515, + "loss": 0.8398, + "step": 625 + }, + { + "epoch": 0.3483099179301711, + "grad_norm": 0.330078125, + "learning_rate": 0.0001936023697215956, + "loss": 0.866, + "step": 626 + }, + { + "epoch": 0.34886632354986785, + "grad_norm": 0.328125, + "learning_rate": 0.00019358180661107051, + "loss": 0.8925, + "step": 627 + }, + { + "epoch": 0.3494227291695646, + "grad_norm": 0.33984375, + "learning_rate": 0.0001935612116024191, + "loss": 0.8689, + "step": 628 + }, + { + "epoch": 0.34997913478926135, + "grad_norm": 0.345703125, + "learning_rate": 0.0001935405847026613, + "loss": 0.8397, + "step": 629 + }, + { + "epoch": 0.3505355404089581, + "grad_norm": 0.34375, + "learning_rate": 0.00019351992591882798, + "loss": 0.8053, + "step": 630 + }, + { + "epoch": 0.3510919460286549, + "grad_norm": 0.3515625, + "learning_rate": 0.00019349923525796084, + "loss": 0.8547, + "step": 631 + }, + { + "epoch": 0.3516483516483517, + "grad_norm": 0.3515625, + "learning_rate": 0.00019347851272711247, + "loss": 0.944, + "step": 632 + }, + { + "epoch": 0.3522047572680484, + "grad_norm": 0.35546875, + "learning_rate": 0.00019345775833334633, + "loss": 0.8962, + "step": 633 + }, + { + "epoch": 0.35276116288774517, + "grad_norm": 0.33203125, + "learning_rate": 0.0001934369720837367, + "loss": 0.8379, + "step": 634 + }, + { + "epoch": 0.35331756850744195, + "grad_norm": 0.345703125, + "learning_rate": 0.00019341615398536877, + "loss": 0.9129, + "step": 635 + }, + { + "epoch": 0.35387397412713867, + "grad_norm": 0.32421875, + "learning_rate": 0.00019339530404533854, + "loss": 0.8244, + "step": 636 + }, + { + "epoch": 0.35443037974683544, + "grad_norm": 0.33203125, + "learning_rate": 0.00019337442227075285, + "loss": 0.8405, + "step": 637 + }, + { + "epoch": 0.3549867853665322, + "grad_norm": 0.33984375, + "learning_rate": 0.0001933535086687295, + "loss": 0.906, + "step": 638 + }, + { + "epoch": 0.35554319098622894, + "grad_norm": 0.34375, + "learning_rate": 0.000193332563246397, + "loss": 0.9176, + "step": 639 + }, + { + "epoch": 0.3560995966059257, + "grad_norm": 0.33203125, + "learning_rate": 0.0001933115860108948, + "loss": 0.8164, + "step": 640 + }, + { + "epoch": 0.3566560022256225, + "grad_norm": 0.34765625, + "learning_rate": 0.00019329057696937317, + "loss": 0.8414, + "step": 641 + }, + { + "epoch": 0.3572124078453192, + "grad_norm": 0.357421875, + "learning_rate": 0.00019326953612899318, + "loss": 0.9027, + "step": 642 + }, + { + "epoch": 0.357768813465016, + "grad_norm": 0.3671875, + "learning_rate": 0.0001932484634969268, + "loss": 0.8986, + "step": 643 + }, + { + "epoch": 0.35832521908471276, + "grad_norm": 0.3515625, + "learning_rate": 0.00019322735908035678, + "loss": 0.891, + "step": 644 + }, + { + "epoch": 0.35888162470440954, + "grad_norm": 0.353515625, + "learning_rate": 0.0001932062228864768, + "loss": 0.9787, + "step": 645 + }, + { + "epoch": 0.35943803032410626, + "grad_norm": 0.33984375, + "learning_rate": 0.0001931850549224912, + "loss": 0.8145, + "step": 646 + }, + { + "epoch": 0.35999443594380304, + "grad_norm": 0.34375, + "learning_rate": 0.00019316385519561538, + "loss": 0.8913, + "step": 647 + }, + { + "epoch": 0.3605508415634998, + "grad_norm": 0.328125, + "learning_rate": 0.00019314262371307532, + "loss": 0.7981, + "step": 648 + }, + { + "epoch": 0.36110724718319653, + "grad_norm": 0.337890625, + "learning_rate": 0.00019312136048210801, + "loss": 0.8772, + "step": 649 + }, + { + "epoch": 0.3616636528028933, + "grad_norm": 0.326171875, + "learning_rate": 0.0001931000655099612, + "loss": 0.8837, + "step": 650 + }, + { + "epoch": 0.3622200584225901, + "grad_norm": 0.345703125, + "learning_rate": 0.00019307873880389345, + "loss": 0.923, + "step": 651 + }, + { + "epoch": 0.3627764640422868, + "grad_norm": 0.337890625, + "learning_rate": 0.00019305738037117412, + "loss": 0.8609, + "step": 652 + }, + { + "epoch": 0.3633328696619836, + "grad_norm": 0.34765625, + "learning_rate": 0.00019303599021908342, + "loss": 0.8668, + "step": 653 + }, + { + "epoch": 0.36388927528168036, + "grad_norm": 0.3515625, + "learning_rate": 0.00019301456835491237, + "loss": 0.9394, + "step": 654 + }, + { + "epoch": 0.3644456809013771, + "grad_norm": 0.328125, + "learning_rate": 0.0001929931147859628, + "loss": 0.7789, + "step": 655 + }, + { + "epoch": 0.36500208652107385, + "grad_norm": 0.359375, + "learning_rate": 0.00019297162951954728, + "loss": 0.9009, + "step": 656 + }, + { + "epoch": 0.36555849214077063, + "grad_norm": 0.333984375, + "learning_rate": 0.00019295011256298929, + "loss": 0.7917, + "step": 657 + }, + { + "epoch": 0.3661148977604674, + "grad_norm": 0.353515625, + "learning_rate": 0.00019292856392362303, + "loss": 0.9266, + "step": 658 + }, + { + "epoch": 0.3666713033801641, + "grad_norm": 0.3359375, + "learning_rate": 0.00019290698360879355, + "loss": 0.849, + "step": 659 + }, + { + "epoch": 0.3672277089998609, + "grad_norm": 0.3359375, + "learning_rate": 0.0001928853716258567, + "loss": 0.8509, + "step": 660 + }, + { + "epoch": 0.3677841146195577, + "grad_norm": 0.326171875, + "learning_rate": 0.00019286372798217905, + "loss": 0.8732, + "step": 661 + }, + { + "epoch": 0.3683405202392544, + "grad_norm": 0.345703125, + "learning_rate": 0.00019284205268513803, + "loss": 0.8759, + "step": 662 + }, + { + "epoch": 0.3688969258589512, + "grad_norm": 1.8046875, + "learning_rate": 0.00019282034574212185, + "loss": 1.8908, + "step": 663 + }, + { + "epoch": 0.36945333147864795, + "grad_norm": 0.33203125, + "learning_rate": 0.00019279860716052947, + "loss": 0.8435, + "step": 664 + }, + { + "epoch": 0.37000973709834467, + "grad_norm": 0.353515625, + "learning_rate": 0.00019277683694777074, + "loss": 0.8151, + "step": 665 + }, + { + "epoch": 0.37056614271804145, + "grad_norm": 0.36328125, + "learning_rate": 0.00019275503511126618, + "loss": 0.9638, + "step": 666 + }, + { + "epoch": 0.3711225483377382, + "grad_norm": 0.337890625, + "learning_rate": 0.00019273320165844707, + "loss": 0.9098, + "step": 667 + }, + { + "epoch": 0.37167895395743494, + "grad_norm": 0.359375, + "learning_rate": 0.00019271133659675563, + "loss": 0.9032, + "step": 668 + }, + { + "epoch": 0.3722353595771317, + "grad_norm": 0.33984375, + "learning_rate": 0.00019268943993364462, + "loss": 0.7979, + "step": 669 + }, + { + "epoch": 0.3727917651968285, + "grad_norm": 0.353515625, + "learning_rate": 0.00019266751167657778, + "loss": 0.9499, + "step": 670 + }, + { + "epoch": 0.37334817081652527, + "grad_norm": 0.3359375, + "learning_rate": 0.0001926455518330295, + "loss": 0.8741, + "step": 671 + }, + { + "epoch": 0.373904576436222, + "grad_norm": 0.349609375, + "learning_rate": 0.000192623560410485, + "loss": 0.9162, + "step": 672 + }, + { + "epoch": 0.37446098205591877, + "grad_norm": 0.33203125, + "learning_rate": 0.00019260153741644021, + "loss": 0.8135, + "step": 673 + }, + { + "epoch": 0.37501738767561554, + "grad_norm": 0.3125, + "learning_rate": 0.00019257948285840188, + "loss": 0.6664, + "step": 674 + }, + { + "epoch": 0.37557379329531226, + "grad_norm": 0.359375, + "learning_rate": 0.00019255739674388746, + "loss": 0.9409, + "step": 675 + }, + { + "epoch": 0.37613019891500904, + "grad_norm": 0.34765625, + "learning_rate": 0.00019253527908042522, + "loss": 0.9325, + "step": 676 + }, + { + "epoch": 0.3766866045347058, + "grad_norm": 0.3359375, + "learning_rate": 0.00019251312987555408, + "loss": 0.8681, + "step": 677 + }, + { + "epoch": 0.37724301015440254, + "grad_norm": 0.3515625, + "learning_rate": 0.00019249094913682382, + "loss": 0.9045, + "step": 678 + }, + { + "epoch": 0.3777994157740993, + "grad_norm": 0.345703125, + "learning_rate": 0.00019246873687179495, + "loss": 0.8807, + "step": 679 + }, + { + "epoch": 0.3783558213937961, + "grad_norm": 0.35546875, + "learning_rate": 0.00019244649308803867, + "loss": 0.9102, + "step": 680 + }, + { + "epoch": 0.37891222701349286, + "grad_norm": 0.333984375, + "learning_rate": 0.00019242421779313692, + "loss": 0.8124, + "step": 681 + }, + { + "epoch": 0.3794686326331896, + "grad_norm": 0.37109375, + "learning_rate": 0.00019240191099468253, + "loss": 0.9337, + "step": 682 + }, + { + "epoch": 0.38002503825288636, + "grad_norm": 0.90625, + "learning_rate": 0.00019237957270027884, + "loss": 1.8244, + "step": 683 + }, + { + "epoch": 0.38058144387258314, + "grad_norm": 0.349609375, + "learning_rate": 0.0001923572029175401, + "loss": 0.8975, + "step": 684 + }, + { + "epoch": 0.38113784949227986, + "grad_norm": 0.34375, + "learning_rate": 0.0001923348016540912, + "loss": 0.8755, + "step": 685 + }, + { + "epoch": 0.38169425511197663, + "grad_norm": 0.328125, + "learning_rate": 0.00019231236891756787, + "loss": 0.8125, + "step": 686 + }, + { + "epoch": 0.3822506607316734, + "grad_norm": 0.337890625, + "learning_rate": 0.00019228990471561636, + "loss": 0.8673, + "step": 687 + }, + { + "epoch": 0.38280706635137013, + "grad_norm": 0.33203125, + "learning_rate": 0.0001922674090558939, + "loss": 0.8462, + "step": 688 + }, + { + "epoch": 0.3833634719710669, + "grad_norm": 0.3671875, + "learning_rate": 0.00019224488194606826, + "loss": 1.0229, + "step": 689 + }, + { + "epoch": 0.3839198775907637, + "grad_norm": 0.333984375, + "learning_rate": 0.00019222232339381802, + "loss": 0.8654, + "step": 690 + }, + { + "epoch": 0.3844762832104604, + "grad_norm": 0.361328125, + "learning_rate": 0.00019219973340683244, + "loss": 0.877, + "step": 691 + }, + { + "epoch": 0.3850326888301572, + "grad_norm": 0.357421875, + "learning_rate": 0.00019217711199281148, + "loss": 0.9084, + "step": 692 + }, + { + "epoch": 0.38558909444985395, + "grad_norm": 0.34765625, + "learning_rate": 0.00019215445915946584, + "loss": 0.9667, + "step": 693 + }, + { + "epoch": 0.38614550006955073, + "grad_norm": 0.32421875, + "learning_rate": 0.00019213177491451692, + "loss": 0.8251, + "step": 694 + }, + { + "epoch": 0.38670190568924745, + "grad_norm": 0.341796875, + "learning_rate": 0.00019210905926569688, + "loss": 0.8764, + "step": 695 + }, + { + "epoch": 0.3872583113089442, + "grad_norm": 0.36328125, + "learning_rate": 0.0001920863122207485, + "loss": 0.8484, + "step": 696 + }, + { + "epoch": 0.387814716928641, + "grad_norm": 0.373046875, + "learning_rate": 0.00019206353378742528, + "loss": 1.0047, + "step": 697 + }, + { + "epoch": 0.3883711225483377, + "grad_norm": 0.337890625, + "learning_rate": 0.00019204072397349144, + "loss": 0.8996, + "step": 698 + }, + { + "epoch": 0.3889275281680345, + "grad_norm": 0.34375, + "learning_rate": 0.00019201788278672193, + "loss": 0.9355, + "step": 699 + }, + { + "epoch": 0.3894839337877313, + "grad_norm": 0.345703125, + "learning_rate": 0.00019199501023490232, + "loss": 0.8971, + "step": 700 + }, + { + "epoch": 0.390040339407428, + "grad_norm": 0.357421875, + "learning_rate": 0.0001919721063258289, + "loss": 0.8437, + "step": 701 + }, + { + "epoch": 0.39059674502712477, + "grad_norm": 0.349609375, + "learning_rate": 0.0001919491710673087, + "loss": 0.9162, + "step": 702 + }, + { + "epoch": 0.39115315064682155, + "grad_norm": 0.349609375, + "learning_rate": 0.00019192620446715933, + "loss": 0.9021, + "step": 703 + }, + { + "epoch": 0.39170955626651827, + "grad_norm": 0.326171875, + "learning_rate": 0.00019190320653320918, + "loss": 0.8099, + "step": 704 + }, + { + "epoch": 0.39226596188621504, + "grad_norm": 0.359375, + "learning_rate": 0.0001918801772732973, + "loss": 0.9569, + "step": 705 + }, + { + "epoch": 0.3928223675059118, + "grad_norm": 0.349609375, + "learning_rate": 0.0001918571166952734, + "loss": 0.9673, + "step": 706 + }, + { + "epoch": 0.3933787731256086, + "grad_norm": 0.337890625, + "learning_rate": 0.0001918340248069978, + "loss": 0.8443, + "step": 707 + }, + { + "epoch": 0.3939351787453053, + "grad_norm": 0.359375, + "learning_rate": 0.00019181090161634167, + "loss": 0.9339, + "step": 708 + }, + { + "epoch": 0.3944915843650021, + "grad_norm": 0.3359375, + "learning_rate": 0.00019178774713118663, + "loss": 0.834, + "step": 709 + }, + { + "epoch": 0.39504798998469887, + "grad_norm": 0.357421875, + "learning_rate": 0.00019176456135942515, + "loss": 0.9933, + "step": 710 + }, + { + "epoch": 0.3956043956043956, + "grad_norm": 0.34765625, + "learning_rate": 0.00019174134430896027, + "loss": 0.924, + "step": 711 + }, + { + "epoch": 0.39616080122409236, + "grad_norm": 0.341796875, + "learning_rate": 0.0001917180959877057, + "loss": 0.8976, + "step": 712 + }, + { + "epoch": 0.39671720684378914, + "grad_norm": 0.33984375, + "learning_rate": 0.00019169481640358583, + "loss": 0.8488, + "step": 713 + }, + { + "epoch": 0.39727361246348586, + "grad_norm": 0.337890625, + "learning_rate": 0.00019167150556453569, + "loss": 0.8426, + "step": 714 + }, + { + "epoch": 0.39783001808318263, + "grad_norm": 0.357421875, + "learning_rate": 0.00019164816347850096, + "loss": 1.0289, + "step": 715 + }, + { + "epoch": 0.3983864237028794, + "grad_norm": 0.341796875, + "learning_rate": 0.00019162479015343803, + "loss": 0.8953, + "step": 716 + }, + { + "epoch": 0.39894282932257613, + "grad_norm": 0.34765625, + "learning_rate": 0.00019160138559731386, + "loss": 0.9065, + "step": 717 + }, + { + "epoch": 0.3994992349422729, + "grad_norm": 0.349609375, + "learning_rate": 0.00019157794981810607, + "loss": 0.908, + "step": 718 + }, + { + "epoch": 0.4000556405619697, + "grad_norm": 0.337890625, + "learning_rate": 0.00019155448282380294, + "loss": 0.796, + "step": 719 + }, + { + "epoch": 0.40061204618166646, + "grad_norm": 0.37109375, + "learning_rate": 0.00019153098462240342, + "loss": 0.9603, + "step": 720 + }, + { + "epoch": 0.4011684518013632, + "grad_norm": 0.341796875, + "learning_rate": 0.00019150745522191701, + "loss": 0.9096, + "step": 721 + }, + { + "epoch": 0.40172485742105996, + "grad_norm": 0.357421875, + "learning_rate": 0.00019148389463036395, + "loss": 1.0184, + "step": 722 + }, + { + "epoch": 0.40228126304075673, + "grad_norm": 0.33203125, + "learning_rate": 0.00019146030285577506, + "loss": 0.8274, + "step": 723 + }, + { + "epoch": 0.40283766866045345, + "grad_norm": 0.353515625, + "learning_rate": 0.00019143667990619177, + "loss": 0.9454, + "step": 724 + }, + { + "epoch": 0.4033940742801502, + "grad_norm": 0.330078125, + "learning_rate": 0.00019141302578966614, + "loss": 0.8516, + "step": 725 + }, + { + "epoch": 0.403950479899847, + "grad_norm": 0.349609375, + "learning_rate": 0.00019138934051426092, + "loss": 0.9234, + "step": 726 + }, + { + "epoch": 0.4045068855195437, + "grad_norm": 0.337890625, + "learning_rate": 0.00019136562408804934, + "loss": 0.8836, + "step": 727 + }, + { + "epoch": 0.4050632911392405, + "grad_norm": 0.349609375, + "learning_rate": 0.00019134187651911546, + "loss": 0.9283, + "step": 728 + }, + { + "epoch": 0.4056196967589373, + "grad_norm": 0.353515625, + "learning_rate": 0.00019131809781555373, + "loss": 0.9731, + "step": 729 + }, + { + "epoch": 0.40617610237863405, + "grad_norm": 0.326171875, + "learning_rate": 0.00019129428798546938, + "loss": 0.7996, + "step": 730 + }, + { + "epoch": 0.4067325079983308, + "grad_norm": 0.33984375, + "learning_rate": 0.00019127044703697816, + "loss": 0.8759, + "step": 731 + }, + { + "epoch": 0.40728891361802755, + "grad_norm": 0.345703125, + "learning_rate": 0.00019124657497820644, + "loss": 0.9137, + "step": 732 + }, + { + "epoch": 0.4078453192377243, + "grad_norm": 0.3515625, + "learning_rate": 0.00019122267181729123, + "loss": 0.8381, + "step": 733 + }, + { + "epoch": 0.40840172485742104, + "grad_norm": 0.365234375, + "learning_rate": 0.0001911987375623801, + "loss": 0.9209, + "step": 734 + }, + { + "epoch": 0.4089581304771178, + "grad_norm": 0.345703125, + "learning_rate": 0.00019117477222163127, + "loss": 0.8937, + "step": 735 + }, + { + "epoch": 0.4095145360968146, + "grad_norm": 0.35546875, + "learning_rate": 0.0001911507758032135, + "loss": 0.9494, + "step": 736 + }, + { + "epoch": 0.4100709417165113, + "grad_norm": 0.353515625, + "learning_rate": 0.00019112674831530616, + "loss": 0.9375, + "step": 737 + }, + { + "epoch": 0.4106273473362081, + "grad_norm": 0.369140625, + "learning_rate": 0.00019110268976609923, + "loss": 0.8973, + "step": 738 + }, + { + "epoch": 0.41118375295590487, + "grad_norm": 0.36328125, + "learning_rate": 0.00019107860016379324, + "loss": 0.9579, + "step": 739 + }, + { + "epoch": 0.4117401585756016, + "grad_norm": 0.328125, + "learning_rate": 0.00019105447951659935, + "loss": 0.7406, + "step": 740 + }, + { + "epoch": 0.41229656419529837, + "grad_norm": 0.365234375, + "learning_rate": 0.0001910303278327393, + "loss": 0.9315, + "step": 741 + }, + { + "epoch": 0.41285296981499514, + "grad_norm": 0.35546875, + "learning_rate": 0.00019100614512044536, + "loss": 0.8731, + "step": 742 + }, + { + "epoch": 0.4134093754346919, + "grad_norm": 0.349609375, + "learning_rate": 0.0001909819313879604, + "loss": 0.8679, + "step": 743 + }, + { + "epoch": 0.41396578105438864, + "grad_norm": 0.345703125, + "learning_rate": 0.00019095768664353785, + "loss": 0.8881, + "step": 744 + }, + { + "epoch": 0.4145221866740854, + "grad_norm": 0.3359375, + "learning_rate": 0.0001909334108954418, + "loss": 0.8616, + "step": 745 + }, + { + "epoch": 0.4150785922937822, + "grad_norm": 0.3359375, + "learning_rate": 0.00019090910415194682, + "loss": 0.8466, + "step": 746 + }, + { + "epoch": 0.4156349979134789, + "grad_norm": 0.359375, + "learning_rate": 0.000190884766421338, + "loss": 0.963, + "step": 747 + }, + { + "epoch": 0.4161914035331757, + "grad_norm": 0.3359375, + "learning_rate": 0.00019086039771191112, + "loss": 0.8879, + "step": 748 + }, + { + "epoch": 0.41674780915287246, + "grad_norm": 0.357421875, + "learning_rate": 0.0001908359980319724, + "loss": 0.9345, + "step": 749 + }, + { + "epoch": 0.4173042147725692, + "grad_norm": 0.3671875, + "learning_rate": 0.00019081156738983874, + "loss": 0.9356, + "step": 750 + }, + { + "epoch": 0.41786062039226596, + "grad_norm": 0.33203125, + "learning_rate": 0.00019078710579383746, + "loss": 0.8295, + "step": 751 + }, + { + "epoch": 0.41841702601196273, + "grad_norm": 0.345703125, + "learning_rate": 0.00019076261325230655, + "loss": 0.9092, + "step": 752 + }, + { + "epoch": 0.41897343163165945, + "grad_norm": 0.32421875, + "learning_rate": 0.00019073808977359446, + "loss": 0.83, + "step": 753 + }, + { + "epoch": 0.41952983725135623, + "grad_norm": 0.318359375, + "learning_rate": 0.0001907135353660602, + "loss": 0.7945, + "step": 754 + }, + { + "epoch": 0.420086242871053, + "grad_norm": 0.353515625, + "learning_rate": 0.00019068895003807339, + "loss": 0.959, + "step": 755 + }, + { + "epoch": 0.4206426484907498, + "grad_norm": 0.361328125, + "learning_rate": 0.0001906643337980141, + "loss": 0.8404, + "step": 756 + }, + { + "epoch": 0.4211990541104465, + "grad_norm": 0.32421875, + "learning_rate": 0.000190639686654273, + "loss": 0.811, + "step": 757 + }, + { + "epoch": 0.4217554597301433, + "grad_norm": 0.34375, + "learning_rate": 0.00019061500861525127, + "loss": 0.9065, + "step": 758 + }, + { + "epoch": 0.42231186534984005, + "grad_norm": 0.357421875, + "learning_rate": 0.00019059029968936062, + "loss": 0.8662, + "step": 759 + }, + { + "epoch": 0.4228682709695368, + "grad_norm": 0.33203125, + "learning_rate": 0.0001905655598850233, + "loss": 0.7609, + "step": 760 + }, + { + "epoch": 0.42342467658923355, + "grad_norm": 0.33984375, + "learning_rate": 0.00019054078921067205, + "loss": 0.8826, + "step": 761 + }, + { + "epoch": 0.4239810822089303, + "grad_norm": 0.328125, + "learning_rate": 0.00019051598767475016, + "loss": 0.8253, + "step": 762 + }, + { + "epoch": 0.42453748782862705, + "grad_norm": 0.3203125, + "learning_rate": 0.00019049115528571147, + "loss": 0.7537, + "step": 763 + }, + { + "epoch": 0.4250938934483238, + "grad_norm": 0.3515625, + "learning_rate": 0.00019046629205202027, + "loss": 0.9363, + "step": 764 + }, + { + "epoch": 0.4256502990680206, + "grad_norm": 0.3359375, + "learning_rate": 0.00019044139798215143, + "loss": 0.8248, + "step": 765 + }, + { + "epoch": 0.4262067046877173, + "grad_norm": 0.322265625, + "learning_rate": 0.00019041647308459022, + "loss": 0.7364, + "step": 766 + }, + { + "epoch": 0.4267631103074141, + "grad_norm": 0.345703125, + "learning_rate": 0.00019039151736783262, + "loss": 0.9075, + "step": 767 + }, + { + "epoch": 0.42731951592711087, + "grad_norm": 0.341796875, + "learning_rate": 0.00019036653084038486, + "loss": 0.966, + "step": 768 + }, + { + "epoch": 0.42787592154680765, + "grad_norm": 0.3515625, + "learning_rate": 0.00019034151351076387, + "loss": 0.8645, + "step": 769 + }, + { + "epoch": 0.42843232716650437, + "grad_norm": 0.3359375, + "learning_rate": 0.00019031646538749707, + "loss": 0.824, + "step": 770 + }, + { + "epoch": 0.42898873278620114, + "grad_norm": 0.33984375, + "learning_rate": 0.00019029138647912218, + "loss": 0.8858, + "step": 771 + }, + { + "epoch": 0.4295451384058979, + "grad_norm": 0.328125, + "learning_rate": 0.00019026627679418767, + "loss": 0.8066, + "step": 772 + }, + { + "epoch": 0.43010154402559464, + "grad_norm": 0.3203125, + "learning_rate": 0.00019024113634125234, + "loss": 0.8256, + "step": 773 + }, + { + "epoch": 0.4306579496452914, + "grad_norm": 0.33984375, + "learning_rate": 0.0001902159651288855, + "loss": 0.8785, + "step": 774 + }, + { + "epoch": 0.4312143552649882, + "grad_norm": 0.337890625, + "learning_rate": 0.000190190763165667, + "loss": 0.8643, + "step": 775 + }, + { + "epoch": 0.4317707608846849, + "grad_norm": 0.34375, + "learning_rate": 0.00019016553046018712, + "loss": 0.9032, + "step": 776 + }, + { + "epoch": 0.4323271665043817, + "grad_norm": 0.34375, + "learning_rate": 0.00019014026702104662, + "loss": 0.8785, + "step": 777 + }, + { + "epoch": 0.43288357212407846, + "grad_norm": 0.33984375, + "learning_rate": 0.00019011497285685678, + "loss": 0.8607, + "step": 778 + }, + { + "epoch": 0.4334399777437752, + "grad_norm": 0.3671875, + "learning_rate": 0.0001900896479762393, + "loss": 0.8355, + "step": 779 + }, + { + "epoch": 0.43399638336347196, + "grad_norm": 0.376953125, + "learning_rate": 0.00019006429238782637, + "loss": 0.9727, + "step": 780 + }, + { + "epoch": 0.43455278898316874, + "grad_norm": 0.34375, + "learning_rate": 0.00019003890610026067, + "loss": 0.8807, + "step": 781 + }, + { + "epoch": 0.4351091946028655, + "grad_norm": 0.33984375, + "learning_rate": 0.00019001348912219533, + "loss": 0.9325, + "step": 782 + }, + { + "epoch": 0.43566560022256223, + "grad_norm": 0.3359375, + "learning_rate": 0.00018998804146229387, + "loss": 0.8559, + "step": 783 + }, + { + "epoch": 0.436222005842259, + "grad_norm": 0.337890625, + "learning_rate": 0.00018996256312923043, + "loss": 0.8304, + "step": 784 + }, + { + "epoch": 0.4367784114619558, + "grad_norm": 0.353515625, + "learning_rate": 0.00018993705413168944, + "loss": 0.9767, + "step": 785 + }, + { + "epoch": 0.4373348170816525, + "grad_norm": 0.33984375, + "learning_rate": 0.0001899115144783659, + "loss": 0.8711, + "step": 786 + }, + { + "epoch": 0.4378912227013493, + "grad_norm": 0.345703125, + "learning_rate": 0.00018988594417796516, + "loss": 0.8525, + "step": 787 + }, + { + "epoch": 0.43844762832104606, + "grad_norm": 0.341796875, + "learning_rate": 0.00018986034323920312, + "loss": 0.83, + "step": 788 + }, + { + "epoch": 0.4390040339407428, + "grad_norm": 0.333984375, + "learning_rate": 0.000189834711670806, + "loss": 0.8088, + "step": 789 + }, + { + "epoch": 0.43956043956043955, + "grad_norm": 0.318359375, + "learning_rate": 0.00018980904948151062, + "loss": 0.784, + "step": 790 + }, + { + "epoch": 0.44011684518013633, + "grad_norm": 0.33984375, + "learning_rate": 0.00018978335668006407, + "loss": 0.9129, + "step": 791 + }, + { + "epoch": 0.4406732507998331, + "grad_norm": 0.36328125, + "learning_rate": 0.000189757633275224, + "loss": 1.0234, + "step": 792 + }, + { + "epoch": 0.4412296564195298, + "grad_norm": 0.349609375, + "learning_rate": 0.0001897318792757584, + "loss": 0.8711, + "step": 793 + }, + { + "epoch": 0.4417860620392266, + "grad_norm": 0.34375, + "learning_rate": 0.00018970609469044577, + "loss": 0.9124, + "step": 794 + }, + { + "epoch": 0.4423424676589234, + "grad_norm": 0.310546875, + "learning_rate": 0.00018968027952807494, + "loss": 0.7081, + "step": 795 + }, + { + "epoch": 0.4428988732786201, + "grad_norm": 0.31640625, + "learning_rate": 0.00018965443379744535, + "loss": 0.7776, + "step": 796 + }, + { + "epoch": 0.4434552788983169, + "grad_norm": 0.32421875, + "learning_rate": 0.00018962855750736656, + "loss": 0.8216, + "step": 797 + }, + { + "epoch": 0.44401168451801365, + "grad_norm": 0.345703125, + "learning_rate": 0.00018960265066665882, + "loss": 0.8406, + "step": 798 + }, + { + "epoch": 0.44456809013771037, + "grad_norm": 0.33984375, + "learning_rate": 0.0001895767132841527, + "loss": 0.9074, + "step": 799 + }, + { + "epoch": 0.44512449575740715, + "grad_norm": 0.37109375, + "learning_rate": 0.00018955074536868913, + "loss": 0.88, + "step": 800 + }, + { + "epoch": 0.4456809013771039, + "grad_norm": 0.361328125, + "learning_rate": 0.0001895247469291195, + "loss": 0.9454, + "step": 801 + }, + { + "epoch": 0.44623730699680064, + "grad_norm": 0.330078125, + "learning_rate": 0.00018949871797430557, + "loss": 0.7475, + "step": 802 + }, + { + "epoch": 0.4467937126164974, + "grad_norm": 0.353515625, + "learning_rate": 0.00018947265851311954, + "loss": 0.9466, + "step": 803 + }, + { + "epoch": 0.4473501182361942, + "grad_norm": 0.328125, + "learning_rate": 0.000189446568554444, + "loss": 0.791, + "step": 804 + }, + { + "epoch": 0.44790652385589097, + "grad_norm": 0.345703125, + "learning_rate": 0.00018942044810717188, + "loss": 0.898, + "step": 805 + }, + { + "epoch": 0.4484629294755877, + "grad_norm": 0.318359375, + "learning_rate": 0.00018939429718020661, + "loss": 0.7666, + "step": 806 + }, + { + "epoch": 0.44901933509528447, + "grad_norm": 0.34375, + "learning_rate": 0.0001893681157824619, + "loss": 0.9343, + "step": 807 + }, + { + "epoch": 0.44957574071498124, + "grad_norm": 0.337890625, + "learning_rate": 0.00018934190392286198, + "loss": 0.8074, + "step": 808 + }, + { + "epoch": 0.45013214633467796, + "grad_norm": 0.333984375, + "learning_rate": 0.00018931566161034126, + "loss": 0.8394, + "step": 809 + }, + { + "epoch": 0.45068855195437474, + "grad_norm": 0.3359375, + "learning_rate": 0.00018928938885384472, + "loss": 0.8655, + "step": 810 + }, + { + "epoch": 0.4512449575740715, + "grad_norm": 0.322265625, + "learning_rate": 0.00018926308566232763, + "loss": 0.792, + "step": 811 + }, + { + "epoch": 0.45180136319376824, + "grad_norm": 0.34765625, + "learning_rate": 0.00018923675204475564, + "loss": 0.8638, + "step": 812 + }, + { + "epoch": 0.452357768813465, + "grad_norm": 0.33984375, + "learning_rate": 0.00018921038801010477, + "loss": 0.8151, + "step": 813 + }, + { + "epoch": 0.4529141744331618, + "grad_norm": 0.3359375, + "learning_rate": 0.00018918399356736146, + "loss": 0.8198, + "step": 814 + }, + { + "epoch": 0.4534705800528585, + "grad_norm": 0.30078125, + "learning_rate": 0.00018915756872552242, + "loss": 0.6904, + "step": 815 + }, + { + "epoch": 0.4540269856725553, + "grad_norm": 0.333984375, + "learning_rate": 0.0001891311134935948, + "loss": 0.8762, + "step": 816 + }, + { + "epoch": 0.45458339129225206, + "grad_norm": 0.3125, + "learning_rate": 0.0001891046278805961, + "loss": 0.7527, + "step": 817 + }, + { + "epoch": 0.45513979691194884, + "grad_norm": 0.33984375, + "learning_rate": 0.00018907811189555412, + "loss": 0.7777, + "step": 818 + }, + { + "epoch": 0.45569620253164556, + "grad_norm": 0.3359375, + "learning_rate": 0.00018905156554750708, + "loss": 0.8556, + "step": 819 + }, + { + "epoch": 0.45625260815134233, + "grad_norm": 0.328125, + "learning_rate": 0.0001890249888455035, + "loss": 0.8273, + "step": 820 + }, + { + "epoch": 0.4568090137710391, + "grad_norm": 0.330078125, + "learning_rate": 0.0001889983817986023, + "loss": 0.8405, + "step": 821 + }, + { + "epoch": 0.45736541939073583, + "grad_norm": 0.341796875, + "learning_rate": 0.00018897174441587266, + "loss": 0.9006, + "step": 822 + }, + { + "epoch": 0.4579218250104326, + "grad_norm": 0.349609375, + "learning_rate": 0.00018894507670639418, + "loss": 0.9191, + "step": 823 + }, + { + "epoch": 0.4584782306301294, + "grad_norm": 0.357421875, + "learning_rate": 0.00018891837867925678, + "loss": 0.9777, + "step": 824 + }, + { + "epoch": 0.4590346362498261, + "grad_norm": 0.34375, + "learning_rate": 0.00018889165034356072, + "loss": 0.892, + "step": 825 + }, + { + "epoch": 0.4595910418695229, + "grad_norm": 0.353515625, + "learning_rate": 0.00018886489170841648, + "loss": 0.8615, + "step": 826 + }, + { + "epoch": 0.46014744748921965, + "grad_norm": 0.341796875, + "learning_rate": 0.00018883810278294503, + "loss": 0.8765, + "step": 827 + }, + { + "epoch": 0.4607038531089164, + "grad_norm": 0.34375, + "learning_rate": 0.00018881128357627763, + "loss": 0.8803, + "step": 828 + }, + { + "epoch": 0.46126025872861315, + "grad_norm": 0.349609375, + "learning_rate": 0.00018878443409755575, + "loss": 0.9017, + "step": 829 + }, + { + "epoch": 0.4618166643483099, + "grad_norm": 0.330078125, + "learning_rate": 0.00018875755435593129, + "loss": 0.8139, + "step": 830 + }, + { + "epoch": 0.4623730699680067, + "grad_norm": 0.34375, + "learning_rate": 0.0001887306443605664, + "loss": 0.9611, + "step": 831 + }, + { + "epoch": 0.4629294755877034, + "grad_norm": 0.34765625, + "learning_rate": 0.00018870370412063365, + "loss": 0.9311, + "step": 832 + }, + { + "epoch": 0.4634858812074002, + "grad_norm": 0.357421875, + "learning_rate": 0.00018867673364531578, + "loss": 0.9108, + "step": 833 + }, + { + "epoch": 0.464042286827097, + "grad_norm": 0.36328125, + "learning_rate": 0.00018864973294380587, + "loss": 1.0142, + "step": 834 + }, + { + "epoch": 0.4645986924467937, + "grad_norm": 0.328125, + "learning_rate": 0.00018862270202530738, + "loss": 0.744, + "step": 835 + }, + { + "epoch": 0.46515509806649047, + "grad_norm": 0.341796875, + "learning_rate": 0.00018859564089903404, + "loss": 0.8419, + "step": 836 + }, + { + "epoch": 0.46571150368618724, + "grad_norm": 0.322265625, + "learning_rate": 0.00018856854957420983, + "loss": 0.7806, + "step": 837 + }, + { + "epoch": 0.46626790930588397, + "grad_norm": 0.3515625, + "learning_rate": 0.00018854142806006902, + "loss": 0.9135, + "step": 838 + }, + { + "epoch": 0.46682431492558074, + "grad_norm": 0.330078125, + "learning_rate": 0.00018851427636585623, + "loss": 0.836, + "step": 839 + }, + { + "epoch": 0.4673807205452775, + "grad_norm": 0.337890625, + "learning_rate": 0.00018848709450082637, + "loss": 0.8926, + "step": 840 + }, + { + "epoch": 0.4679371261649743, + "grad_norm": 0.35546875, + "learning_rate": 0.0001884598824742445, + "loss": 0.8962, + "step": 841 + }, + { + "epoch": 0.468493531784671, + "grad_norm": 0.330078125, + "learning_rate": 0.0001884326402953862, + "loss": 0.8185, + "step": 842 + }, + { + "epoch": 0.4690499374043678, + "grad_norm": 0.35546875, + "learning_rate": 0.0001884053679735371, + "loss": 0.948, + "step": 843 + }, + { + "epoch": 0.46960634302406457, + "grad_norm": 0.365234375, + "learning_rate": 0.0001883780655179932, + "loss": 1.0652, + "step": 844 + }, + { + "epoch": 0.4701627486437613, + "grad_norm": 0.35546875, + "learning_rate": 0.00018835073293806075, + "loss": 0.9291, + "step": 845 + }, + { + "epoch": 0.47071915426345806, + "grad_norm": 0.34765625, + "learning_rate": 0.00018832337024305638, + "loss": 0.9503, + "step": 846 + }, + { + "epoch": 0.47127555988315484, + "grad_norm": 0.337890625, + "learning_rate": 0.00018829597744230678, + "loss": 0.8685, + "step": 847 + }, + { + "epoch": 0.47183196550285156, + "grad_norm": 0.3359375, + "learning_rate": 0.00018826855454514907, + "loss": 0.8796, + "step": 848 + }, + { + "epoch": 0.47238837112254833, + "grad_norm": 0.32421875, + "learning_rate": 0.00018824110156093053, + "loss": 0.7906, + "step": 849 + }, + { + "epoch": 0.4729447767422451, + "grad_norm": 0.34375, + "learning_rate": 0.00018821361849900877, + "loss": 0.9085, + "step": 850 + }, + { + "epoch": 0.47350118236194183, + "grad_norm": 0.330078125, + "learning_rate": 0.00018818610536875164, + "loss": 0.9018, + "step": 851 + }, + { + "epoch": 0.4740575879816386, + "grad_norm": 0.33203125, + "learning_rate": 0.00018815856217953714, + "loss": 0.8632, + "step": 852 + }, + { + "epoch": 0.4746139936013354, + "grad_norm": 0.33984375, + "learning_rate": 0.00018813098894075365, + "loss": 0.9047, + "step": 853 + }, + { + "epoch": 0.47517039922103216, + "grad_norm": 0.34375, + "learning_rate": 0.00018810338566179973, + "loss": 0.9046, + "step": 854 + }, + { + "epoch": 0.4757268048407289, + "grad_norm": 0.328125, + "learning_rate": 0.00018807575235208415, + "loss": 0.8683, + "step": 855 + }, + { + "epoch": 0.47628321046042565, + "grad_norm": 0.330078125, + "learning_rate": 0.00018804808902102597, + "loss": 0.8541, + "step": 856 + }, + { + "epoch": 0.47683961608012243, + "grad_norm": 0.33203125, + "learning_rate": 0.00018802039567805448, + "loss": 0.8484, + "step": 857 + }, + { + "epoch": 0.47739602169981915, + "grad_norm": 0.345703125, + "learning_rate": 0.00018799267233260918, + "loss": 0.8559, + "step": 858 + }, + { + "epoch": 0.4779524273195159, + "grad_norm": 0.3359375, + "learning_rate": 0.00018796491899413976, + "loss": 0.8646, + "step": 859 + }, + { + "epoch": 0.4785088329392127, + "grad_norm": 0.349609375, + "learning_rate": 0.0001879371356721062, + "loss": 0.9625, + "step": 860 + }, + { + "epoch": 0.4790652385589094, + "grad_norm": 0.3203125, + "learning_rate": 0.0001879093223759787, + "loss": 0.7927, + "step": 861 + }, + { + "epoch": 0.4796216441786062, + "grad_norm": 0.32421875, + "learning_rate": 0.00018788147911523762, + "loss": 0.8716, + "step": 862 + }, + { + "epoch": 0.480178049798303, + "grad_norm": 0.3515625, + "learning_rate": 0.00018785360589937354, + "loss": 0.7913, + "step": 863 + }, + { + "epoch": 0.4807344554179997, + "grad_norm": 0.34375, + "learning_rate": 0.00018782570273788736, + "loss": 0.8636, + "step": 864 + }, + { + "epoch": 0.48129086103769647, + "grad_norm": 0.330078125, + "learning_rate": 0.00018779776964029, + "loss": 0.8082, + "step": 865 + }, + { + "epoch": 0.48184726665739325, + "grad_norm": 0.33984375, + "learning_rate": 0.00018776980661610273, + "loss": 0.8549, + "step": 866 + }, + { + "epoch": 0.48240367227709, + "grad_norm": 0.33203125, + "learning_rate": 0.000187741813674857, + "loss": 0.8543, + "step": 867 + }, + { + "epoch": 0.48296007789678674, + "grad_norm": 0.337890625, + "learning_rate": 0.00018771379082609436, + "loss": 0.8735, + "step": 868 + }, + { + "epoch": 0.4835164835164835, + "grad_norm": 0.33984375, + "learning_rate": 0.0001876857380793667, + "loss": 0.8338, + "step": 869 + }, + { + "epoch": 0.4840728891361803, + "grad_norm": 0.333984375, + "learning_rate": 0.00018765765544423606, + "loss": 0.8904, + "step": 870 + }, + { + "epoch": 0.484629294755877, + "grad_norm": 0.326171875, + "learning_rate": 0.00018762954293027453, + "loss": 0.7981, + "step": 871 + }, + { + "epoch": 0.4851857003755738, + "grad_norm": 0.341796875, + "learning_rate": 0.00018760140054706453, + "loss": 0.9093, + "step": 872 + }, + { + "epoch": 0.48574210599527057, + "grad_norm": 0.341796875, + "learning_rate": 0.00018757322830419867, + "loss": 0.9216, + "step": 873 + }, + { + "epoch": 0.4862985116149673, + "grad_norm": 0.31640625, + "learning_rate": 0.00018754502621127966, + "loss": 0.7949, + "step": 874 + }, + { + "epoch": 0.48685491723466406, + "grad_norm": 0.345703125, + "learning_rate": 0.00018751679427792038, + "loss": 0.9682, + "step": 875 + }, + { + "epoch": 0.48741132285436084, + "grad_norm": 0.328125, + "learning_rate": 0.00018748853251374396, + "loss": 0.792, + "step": 876 + }, + { + "epoch": 0.48796772847405756, + "grad_norm": 0.349609375, + "learning_rate": 0.00018746024092838366, + "loss": 0.8901, + "step": 877 + }, + { + "epoch": 0.48852413409375434, + "grad_norm": 0.349609375, + "learning_rate": 0.00018743191953148286, + "loss": 0.9955, + "step": 878 + }, + { + "epoch": 0.4890805397134511, + "grad_norm": 0.34375, + "learning_rate": 0.00018740356833269516, + "loss": 0.9246, + "step": 879 + }, + { + "epoch": 0.4896369453331479, + "grad_norm": 0.34375, + "learning_rate": 0.00018737518734168428, + "loss": 0.9122, + "step": 880 + }, + { + "epoch": 0.4901933509528446, + "grad_norm": 0.3359375, + "learning_rate": 0.0001873467765681242, + "loss": 0.9119, + "step": 881 + }, + { + "epoch": 0.4907497565725414, + "grad_norm": 0.322265625, + "learning_rate": 0.00018731833602169884, + "loss": 0.7738, + "step": 882 + }, + { + "epoch": 0.49130616219223816, + "grad_norm": 0.33203125, + "learning_rate": 0.00018728986571210244, + "loss": 0.8866, + "step": 883 + }, + { + "epoch": 0.4918625678119349, + "grad_norm": 0.333984375, + "learning_rate": 0.00018726136564903938, + "loss": 0.9322, + "step": 884 + }, + { + "epoch": 0.49241897343163166, + "grad_norm": 0.318359375, + "learning_rate": 0.0001872328358422241, + "loss": 0.7841, + "step": 885 + }, + { + "epoch": 0.49297537905132843, + "grad_norm": 0.349609375, + "learning_rate": 0.00018720427630138122, + "loss": 0.9459, + "step": 886 + }, + { + "epoch": 0.49353178467102515, + "grad_norm": 0.330078125, + "learning_rate": 0.0001871756870362455, + "loss": 0.8372, + "step": 887 + }, + { + "epoch": 0.49408819029072193, + "grad_norm": 0.3359375, + "learning_rate": 0.00018714706805656185, + "loss": 0.8801, + "step": 888 + }, + { + "epoch": 0.4946445959104187, + "grad_norm": 0.31640625, + "learning_rate": 0.00018711841937208527, + "loss": 0.7583, + "step": 889 + }, + { + "epoch": 0.4952010015301155, + "grad_norm": 0.333984375, + "learning_rate": 0.00018708974099258086, + "loss": 0.8007, + "step": 890 + }, + { + "epoch": 0.4957574071498122, + "grad_norm": 0.349609375, + "learning_rate": 0.0001870610329278239, + "loss": 0.9087, + "step": 891 + }, + { + "epoch": 0.496313812769509, + "grad_norm": 0.333984375, + "learning_rate": 0.00018703229518759985, + "loss": 0.8809, + "step": 892 + }, + { + "epoch": 0.49687021838920575, + "grad_norm": 0.33203125, + "learning_rate": 0.00018700352778170407, + "loss": 0.8517, + "step": 893 + }, + { + "epoch": 0.4974266240089025, + "grad_norm": 0.345703125, + "learning_rate": 0.00018697473071994225, + "loss": 0.9004, + "step": 894 + }, + { + "epoch": 0.49798302962859925, + "grad_norm": 0.33203125, + "learning_rate": 0.0001869459040121301, + "loss": 0.8786, + "step": 895 + }, + { + "epoch": 0.498539435248296, + "grad_norm": 0.34765625, + "learning_rate": 0.0001869170476680934, + "loss": 0.911, + "step": 896 + }, + { + "epoch": 0.49909584086799275, + "grad_norm": 0.34765625, + "learning_rate": 0.0001868881616976681, + "loss": 0.8994, + "step": 897 + }, + { + "epoch": 0.4996522464876895, + "grad_norm": 0.328125, + "learning_rate": 0.00018685924611070025, + "loss": 0.8162, + "step": 898 + }, + { + "epoch": 0.5002086521073863, + "grad_norm": 0.33984375, + "learning_rate": 0.00018683030091704592, + "loss": 0.891, + "step": 899 + }, + { + "epoch": 0.5002086521073863, + "eval_loss": 0.8054929971694946, + "eval_runtime": 1766.0961, + "eval_samples_per_second": 2.656, + "eval_steps_per_second": 1.328, + "step": 899 + }, + { + "epoch": 0.500765057727083, + "grad_norm": 0.326171875, + "learning_rate": 0.0001868013261265713, + "loss": 0.8493, + "step": 900 + }, + { + "epoch": 0.5013214633467798, + "grad_norm": 0.3515625, + "learning_rate": 0.00018677232174915275, + "loss": 0.8253, + "step": 901 + }, + { + "epoch": 0.5018778689664766, + "grad_norm": 0.341796875, + "learning_rate": 0.0001867432877946766, + "loss": 0.8701, + "step": 902 + }, + { + "epoch": 0.5024342745861733, + "grad_norm": 0.359375, + "learning_rate": 0.0001867142242730393, + "loss": 0.9443, + "step": 903 + }, + { + "epoch": 0.5029906802058701, + "grad_norm": 0.3515625, + "learning_rate": 0.00018668513119414744, + "loss": 0.9508, + "step": 904 + }, + { + "epoch": 0.5035470858255668, + "grad_norm": 0.330078125, + "learning_rate": 0.00018665600856791764, + "loss": 0.7881, + "step": 905 + }, + { + "epoch": 0.5041034914452636, + "grad_norm": 0.345703125, + "learning_rate": 0.00018662685640427653, + "loss": 0.8958, + "step": 906 + }, + { + "epoch": 0.5046598970649604, + "grad_norm": 0.345703125, + "learning_rate": 0.00018659767471316087, + "loss": 0.8597, + "step": 907 + }, + { + "epoch": 0.5052163026846571, + "grad_norm": 0.33984375, + "learning_rate": 0.0001865684635045175, + "loss": 0.8495, + "step": 908 + }, + { + "epoch": 0.5057727083043538, + "grad_norm": 0.33984375, + "learning_rate": 0.00018653922278830335, + "loss": 0.8809, + "step": 909 + }, + { + "epoch": 0.5063291139240507, + "grad_norm": 0.330078125, + "learning_rate": 0.00018650995257448526, + "loss": 0.8809, + "step": 910 + }, + { + "epoch": 0.5068855195437474, + "grad_norm": 0.337890625, + "learning_rate": 0.0001864806528730403, + "loss": 0.894, + "step": 911 + }, + { + "epoch": 0.5074419251634441, + "grad_norm": 0.333984375, + "learning_rate": 0.00018645132369395545, + "loss": 0.8067, + "step": 912 + }, + { + "epoch": 0.5079983307831409, + "grad_norm": 0.34375, + "learning_rate": 0.00018642196504722785, + "loss": 0.8951, + "step": 913 + }, + { + "epoch": 0.5085547364028377, + "grad_norm": 0.33984375, + "learning_rate": 0.00018639257694286462, + "loss": 0.846, + "step": 914 + }, + { + "epoch": 0.5091111420225344, + "grad_norm": 0.326171875, + "learning_rate": 0.00018636315939088292, + "loss": 0.8076, + "step": 915 + }, + { + "epoch": 0.5096675476422312, + "grad_norm": 0.322265625, + "learning_rate": 0.00018633371240131, + "loss": 0.7998, + "step": 916 + }, + { + "epoch": 0.5102239532619279, + "grad_norm": 0.330078125, + "learning_rate": 0.00018630423598418308, + "loss": 0.7803, + "step": 917 + }, + { + "epoch": 0.5107803588816247, + "grad_norm": 0.353515625, + "learning_rate": 0.00018627473014954946, + "loss": 0.9279, + "step": 918 + }, + { + "epoch": 0.5113367645013215, + "grad_norm": 0.341796875, + "learning_rate": 0.0001862451949074664, + "loss": 0.9039, + "step": 919 + }, + { + "epoch": 0.5118931701210182, + "grad_norm": 0.326171875, + "learning_rate": 0.00018621563026800127, + "loss": 0.7871, + "step": 920 + }, + { + "epoch": 0.512449575740715, + "grad_norm": 0.3359375, + "learning_rate": 0.00018618603624123145, + "loss": 0.8431, + "step": 921 + }, + { + "epoch": 0.5130059813604118, + "grad_norm": 0.33203125, + "learning_rate": 0.00018615641283724425, + "loss": 0.8864, + "step": 922 + }, + { + "epoch": 0.5135623869801085, + "grad_norm": 0.3359375, + "learning_rate": 0.00018612676006613706, + "loss": 0.8569, + "step": 923 + }, + { + "epoch": 0.5141187925998053, + "grad_norm": 0.345703125, + "learning_rate": 0.00018609707793801726, + "loss": 0.9234, + "step": 924 + }, + { + "epoch": 0.514675198219502, + "grad_norm": 0.34765625, + "learning_rate": 0.0001860673664630023, + "loss": 0.8782, + "step": 925 + }, + { + "epoch": 0.5152316038391987, + "grad_norm": 0.333984375, + "learning_rate": 0.00018603762565121953, + "loss": 0.8889, + "step": 926 + }, + { + "epoch": 0.5157880094588956, + "grad_norm": 0.341796875, + "learning_rate": 0.00018600785551280637, + "loss": 0.9133, + "step": 927 + }, + { + "epoch": 0.5163444150785923, + "grad_norm": 0.337890625, + "learning_rate": 0.0001859780560579102, + "loss": 0.8214, + "step": 928 + }, + { + "epoch": 0.516900820698289, + "grad_norm": 0.310546875, + "learning_rate": 0.00018594822729668843, + "loss": 0.7542, + "step": 929 + }, + { + "epoch": 0.5174572263179859, + "grad_norm": 0.326171875, + "learning_rate": 0.00018591836923930843, + "loss": 0.8442, + "step": 930 + }, + { + "epoch": 0.5180136319376826, + "grad_norm": 0.359375, + "learning_rate": 0.00018588848189594756, + "loss": 0.9669, + "step": 931 + }, + { + "epoch": 0.5185700375573793, + "grad_norm": 0.33203125, + "learning_rate": 0.00018585856527679316, + "loss": 0.8215, + "step": 932 + }, + { + "epoch": 0.5191264431770761, + "grad_norm": 0.34765625, + "learning_rate": 0.00018582861939204256, + "loss": 0.8499, + "step": 933 + }, + { + "epoch": 0.5196828487967728, + "grad_norm": 0.345703125, + "learning_rate": 0.00018579864425190309, + "loss": 0.8397, + "step": 934 + }, + { + "epoch": 0.5202392544164696, + "grad_norm": 0.326171875, + "learning_rate": 0.00018576863986659198, + "loss": 0.8204, + "step": 935 + }, + { + "epoch": 0.5207956600361664, + "grad_norm": 0.314453125, + "learning_rate": 0.0001857386062463365, + "loss": 0.765, + "step": 936 + }, + { + "epoch": 0.5213520656558631, + "grad_norm": 0.341796875, + "learning_rate": 0.0001857085434013738, + "loss": 0.8809, + "step": 937 + }, + { + "epoch": 0.5219084712755598, + "grad_norm": 0.326171875, + "learning_rate": 0.0001856784513419511, + "loss": 0.8837, + "step": 938 + }, + { + "epoch": 0.5224648768952567, + "grad_norm": 0.3359375, + "learning_rate": 0.00018564833007832555, + "loss": 0.8835, + "step": 939 + }, + { + "epoch": 0.5230212825149534, + "grad_norm": 0.345703125, + "learning_rate": 0.00018561817962076414, + "loss": 0.8856, + "step": 940 + }, + { + "epoch": 0.5235776881346501, + "grad_norm": 0.33203125, + "learning_rate": 0.00018558799997954402, + "loss": 0.8138, + "step": 941 + }, + { + "epoch": 0.5241340937543469, + "grad_norm": 0.34375, + "learning_rate": 0.00018555779116495206, + "loss": 0.8903, + "step": 942 + }, + { + "epoch": 0.5246904993740437, + "grad_norm": 0.32421875, + "learning_rate": 0.00018552755318728523, + "loss": 0.7882, + "step": 943 + }, + { + "epoch": 0.5252469049937404, + "grad_norm": 0.341796875, + "learning_rate": 0.00018549728605685042, + "loss": 0.8532, + "step": 944 + }, + { + "epoch": 0.5258033106134372, + "grad_norm": 0.318359375, + "learning_rate": 0.00018546698978396436, + "loss": 0.7801, + "step": 945 + }, + { + "epoch": 0.5263597162331339, + "grad_norm": 0.328125, + "learning_rate": 0.00018543666437895385, + "loss": 0.8937, + "step": 946 + }, + { + "epoch": 0.5269161218528308, + "grad_norm": 0.341796875, + "learning_rate": 0.00018540630985215548, + "loss": 0.9237, + "step": 947 + }, + { + "epoch": 0.5274725274725275, + "grad_norm": 0.349609375, + "learning_rate": 0.00018537592621391591, + "loss": 0.8728, + "step": 948 + }, + { + "epoch": 0.5280289330922242, + "grad_norm": 0.33203125, + "learning_rate": 0.00018534551347459163, + "loss": 0.8142, + "step": 949 + }, + { + "epoch": 0.528585338711921, + "grad_norm": 0.337890625, + "learning_rate": 0.0001853150716445491, + "loss": 0.9059, + "step": 950 + }, + { + "epoch": 0.5291417443316178, + "grad_norm": 0.337890625, + "learning_rate": 0.00018528460073416456, + "loss": 0.8826, + "step": 951 + }, + { + "epoch": 0.5296981499513145, + "grad_norm": 0.359375, + "learning_rate": 0.00018525410075382438, + "loss": 0.9263, + "step": 952 + }, + { + "epoch": 0.5302545555710113, + "grad_norm": 0.341796875, + "learning_rate": 0.0001852235717139247, + "loss": 0.8751, + "step": 953 + }, + { + "epoch": 0.530810961190708, + "grad_norm": 0.330078125, + "learning_rate": 0.00018519301362487158, + "loss": 0.833, + "step": 954 + }, + { + "epoch": 0.5313673668104048, + "grad_norm": 0.34765625, + "learning_rate": 0.00018516242649708103, + "loss": 0.9195, + "step": 955 + }, + { + "epoch": 0.5319237724301016, + "grad_norm": 0.359375, + "learning_rate": 0.00018513181034097886, + "loss": 0.961, + "step": 956 + }, + { + "epoch": 0.5324801780497983, + "grad_norm": 0.326171875, + "learning_rate": 0.0001851011651670009, + "loss": 0.8375, + "step": 957 + }, + { + "epoch": 0.533036583669495, + "grad_norm": 0.3359375, + "learning_rate": 0.00018507049098559275, + "loss": 0.8556, + "step": 958 + }, + { + "epoch": 0.5335929892891919, + "grad_norm": 0.35546875, + "learning_rate": 0.00018503978780721, + "loss": 0.9757, + "step": 959 + }, + { + "epoch": 0.5341493949088886, + "grad_norm": 0.349609375, + "learning_rate": 0.00018500905564231807, + "loss": 0.9229, + "step": 960 + }, + { + "epoch": 0.5347058005285853, + "grad_norm": 0.330078125, + "learning_rate": 0.0001849782945013923, + "loss": 0.8092, + "step": 961 + }, + { + "epoch": 0.5352622061482821, + "grad_norm": 0.359375, + "learning_rate": 0.00018494750439491782, + "loss": 0.8918, + "step": 962 + }, + { + "epoch": 0.5358186117679788, + "grad_norm": 0.328125, + "learning_rate": 0.00018491668533338973, + "loss": 0.8153, + "step": 963 + }, + { + "epoch": 0.5363750173876756, + "grad_norm": 0.333984375, + "learning_rate": 0.00018488583732731293, + "loss": 0.8872, + "step": 964 + }, + { + "epoch": 0.5369314230073724, + "grad_norm": 0.35546875, + "learning_rate": 0.00018485496038720222, + "loss": 0.9895, + "step": 965 + }, + { + "epoch": 0.5374878286270691, + "grad_norm": 0.34375, + "learning_rate": 0.00018482405452358228, + "loss": 0.8771, + "step": 966 + }, + { + "epoch": 0.5380442342467658, + "grad_norm": 0.353515625, + "learning_rate": 0.00018479311974698765, + "loss": 0.9115, + "step": 967 + }, + { + "epoch": 0.5386006398664627, + "grad_norm": 0.341796875, + "learning_rate": 0.00018476215606796263, + "loss": 0.8777, + "step": 968 + }, + { + "epoch": 0.5391570454861594, + "grad_norm": 0.34375, + "learning_rate": 0.00018473116349706153, + "loss": 0.9589, + "step": 969 + }, + { + "epoch": 0.5397134511058562, + "grad_norm": 0.306640625, + "learning_rate": 0.00018470014204484834, + "loss": 0.7294, + "step": 970 + }, + { + "epoch": 0.5402698567255529, + "grad_norm": 0.333984375, + "learning_rate": 0.000184669091721897, + "loss": 0.8296, + "step": 971 + }, + { + "epoch": 0.5408262623452497, + "grad_norm": 0.3515625, + "learning_rate": 0.00018463801253879128, + "loss": 0.9549, + "step": 972 + }, + { + "epoch": 0.5413826679649465, + "grad_norm": 0.33203125, + "learning_rate": 0.0001846069045061248, + "loss": 0.8513, + "step": 973 + }, + { + "epoch": 0.5419390735846432, + "grad_norm": 0.333984375, + "learning_rate": 0.00018457576763450097, + "loss": 0.879, + "step": 974 + }, + { + "epoch": 0.5424954792043399, + "grad_norm": 0.337890625, + "learning_rate": 0.00018454460193453303, + "loss": 0.8198, + "step": 975 + }, + { + "epoch": 0.5430518848240368, + "grad_norm": 0.333984375, + "learning_rate": 0.00018451340741684408, + "loss": 0.8208, + "step": 976 + }, + { + "epoch": 0.5436082904437335, + "grad_norm": 0.345703125, + "learning_rate": 0.000184482184092067, + "loss": 0.9505, + "step": 977 + }, + { + "epoch": 0.5441646960634302, + "grad_norm": 0.34375, + "learning_rate": 0.00018445093197084456, + "loss": 0.8958, + "step": 978 + }, + { + "epoch": 0.544721101683127, + "grad_norm": 0.33203125, + "learning_rate": 0.00018441965106382927, + "loss": 0.8845, + "step": 979 + }, + { + "epoch": 0.5452775073028238, + "grad_norm": 0.31640625, + "learning_rate": 0.0001843883413816835, + "loss": 0.7969, + "step": 980 + }, + { + "epoch": 0.5458339129225205, + "grad_norm": 0.333984375, + "learning_rate": 0.0001843570029350794, + "loss": 0.8301, + "step": 981 + }, + { + "epoch": 0.5463903185422173, + "grad_norm": 0.306640625, + "learning_rate": 0.00018432563573469895, + "loss": 0.7873, + "step": 982 + }, + { + "epoch": 0.546946724161914, + "grad_norm": 0.34765625, + "learning_rate": 0.0001842942397912339, + "loss": 0.8894, + "step": 983 + }, + { + "epoch": 0.5475031297816108, + "grad_norm": 0.33984375, + "learning_rate": 0.00018426281511538584, + "loss": 0.91, + "step": 984 + }, + { + "epoch": 0.5480595354013076, + "grad_norm": 0.328125, + "learning_rate": 0.00018423136171786611, + "loss": 0.823, + "step": 985 + }, + { + "epoch": 0.5486159410210043, + "grad_norm": 0.318359375, + "learning_rate": 0.00018419987960939586, + "loss": 0.8033, + "step": 986 + }, + { + "epoch": 0.549172346640701, + "grad_norm": 0.34375, + "learning_rate": 0.00018416836880070603, + "loss": 0.8735, + "step": 987 + }, + { + "epoch": 0.5497287522603979, + "grad_norm": 0.3203125, + "learning_rate": 0.00018413682930253734, + "loss": 0.8009, + "step": 988 + }, + { + "epoch": 0.5502851578800946, + "grad_norm": 0.34765625, + "learning_rate": 0.00018410526112564026, + "loss": 0.9164, + "step": 989 + }, + { + "epoch": 0.5508415634997913, + "grad_norm": 0.328125, + "learning_rate": 0.0001840736642807751, + "loss": 0.869, + "step": 990 + }, + { + "epoch": 0.5513979691194881, + "grad_norm": 0.3359375, + "learning_rate": 0.00018404203877871186, + "loss": 0.7933, + "step": 991 + }, + { + "epoch": 0.5519543747391848, + "grad_norm": 0.330078125, + "learning_rate": 0.0001840103846302304, + "loss": 0.8246, + "step": 992 + }, + { + "epoch": 0.5525107803588816, + "grad_norm": 0.337890625, + "learning_rate": 0.0001839787018461203, + "loss": 0.8753, + "step": 993 + }, + { + "epoch": 0.5530671859785784, + "grad_norm": 0.330078125, + "learning_rate": 0.00018394699043718085, + "loss": 0.8591, + "step": 994 + }, + { + "epoch": 0.5536235915982751, + "grad_norm": 0.333984375, + "learning_rate": 0.00018391525041422114, + "loss": 0.8757, + "step": 995 + }, + { + "epoch": 0.554179997217972, + "grad_norm": 0.333984375, + "learning_rate": 0.00018388348178806008, + "loss": 0.8981, + "step": 996 + }, + { + "epoch": 0.5547364028376687, + "grad_norm": 0.326171875, + "learning_rate": 0.00018385168456952621, + "loss": 0.7889, + "step": 997 + }, + { + "epoch": 0.5552928084573654, + "grad_norm": 0.333984375, + "learning_rate": 0.0001838198587694579, + "loss": 0.8176, + "step": 998 + }, + { + "epoch": 0.5558492140770622, + "grad_norm": 0.34765625, + "learning_rate": 0.0001837880043987032, + "loss": 0.9225, + "step": 999 + }, + { + "epoch": 0.556405619696759, + "grad_norm": 0.328125, + "learning_rate": 0.00018375612146812, + "loss": 0.8243, + "step": 1000 + }, + { + "epoch": 0.5569620253164557, + "grad_norm": 0.32421875, + "learning_rate": 0.00018372420998857584, + "loss": 0.7683, + "step": 1001 + }, + { + "epoch": 0.5575184309361525, + "grad_norm": 0.33203125, + "learning_rate": 0.00018369226997094792, + "loss": 0.906, + "step": 1002 + }, + { + "epoch": 0.5580748365558492, + "grad_norm": 0.341796875, + "learning_rate": 0.0001836603014261234, + "loss": 0.8942, + "step": 1003 + }, + { + "epoch": 0.5586312421755459, + "grad_norm": 0.34375, + "learning_rate": 0.00018362830436499886, + "loss": 0.8148, + "step": 1004 + }, + { + "epoch": 0.5591876477952428, + "grad_norm": 0.33984375, + "learning_rate": 0.0001835962787984809, + "loss": 0.9291, + "step": 1005 + }, + { + "epoch": 0.5597440534149395, + "grad_norm": 0.318359375, + "learning_rate": 0.00018356422473748564, + "loss": 0.7764, + "step": 1006 + }, + { + "epoch": 0.5603004590346362, + "grad_norm": 0.3359375, + "learning_rate": 0.00018353214219293896, + "loss": 0.8792, + "step": 1007 + }, + { + "epoch": 0.560856864654333, + "grad_norm": 0.34375, + "learning_rate": 0.0001835000311757765, + "loss": 0.8608, + "step": 1008 + }, + { + "epoch": 0.5614132702740298, + "grad_norm": 0.33984375, + "learning_rate": 0.0001834678916969435, + "loss": 0.927, + "step": 1009 + }, + { + "epoch": 0.5619696758937265, + "grad_norm": 0.314453125, + "learning_rate": 0.00018343572376739505, + "loss": 0.7639, + "step": 1010 + }, + { + "epoch": 0.5625260815134233, + "grad_norm": 0.345703125, + "learning_rate": 0.00018340352739809577, + "loss": 0.8458, + "step": 1011 + }, + { + "epoch": 0.56308248713312, + "grad_norm": 0.341796875, + "learning_rate": 0.0001833713026000201, + "loss": 0.9023, + "step": 1012 + }, + { + "epoch": 0.5636388927528168, + "grad_norm": 0.322265625, + "learning_rate": 0.0001833390493841521, + "loss": 0.8104, + "step": 1013 + }, + { + "epoch": 0.5641952983725136, + "grad_norm": 0.349609375, + "learning_rate": 0.00018330676776148557, + "loss": 0.8363, + "step": 1014 + }, + { + "epoch": 0.5647517039922103, + "grad_norm": 0.33203125, + "learning_rate": 0.00018327445774302397, + "loss": 0.8123, + "step": 1015 + }, + { + "epoch": 0.565308109611907, + "grad_norm": 0.357421875, + "learning_rate": 0.00018324211933978043, + "loss": 0.8886, + "step": 1016 + }, + { + "epoch": 0.5658645152316039, + "grad_norm": 0.31640625, + "learning_rate": 0.0001832097525627777, + "loss": 0.7553, + "step": 1017 + }, + { + "epoch": 0.5664209208513006, + "grad_norm": 0.33984375, + "learning_rate": 0.00018317735742304832, + "loss": 0.8144, + "step": 1018 + }, + { + "epoch": 0.5669773264709974, + "grad_norm": 0.3359375, + "learning_rate": 0.00018314493393163444, + "loss": 0.8803, + "step": 1019 + }, + { + "epoch": 0.5675337320906941, + "grad_norm": 0.330078125, + "learning_rate": 0.0001831124820995879, + "loss": 0.8016, + "step": 1020 + }, + { + "epoch": 0.5680901377103909, + "grad_norm": 0.3203125, + "learning_rate": 0.00018308000193797012, + "loss": 0.7583, + "step": 1021 + }, + { + "epoch": 0.5686465433300877, + "grad_norm": 0.3359375, + "learning_rate": 0.00018304749345785223, + "loss": 0.8315, + "step": 1022 + }, + { + "epoch": 0.5692029489497844, + "grad_norm": 0.322265625, + "learning_rate": 0.00018301495667031505, + "loss": 0.7795, + "step": 1023 + }, + { + "epoch": 0.5697593545694811, + "grad_norm": 0.3203125, + "learning_rate": 0.000182982391586449, + "loss": 0.7642, + "step": 1024 + }, + { + "epoch": 0.570315760189178, + "grad_norm": 0.357421875, + "learning_rate": 0.00018294979821735413, + "loss": 0.8609, + "step": 1025 + }, + { + "epoch": 0.5708721658088747, + "grad_norm": 0.39453125, + "learning_rate": 0.00018291717657414013, + "loss": 0.9518, + "step": 1026 + }, + { + "epoch": 0.5714285714285714, + "grad_norm": 0.33984375, + "learning_rate": 0.00018288452666792644, + "loss": 0.8299, + "step": 1027 + }, + { + "epoch": 0.5719849770482682, + "grad_norm": 0.3515625, + "learning_rate": 0.000182851848509842, + "loss": 0.9225, + "step": 1028 + }, + { + "epoch": 0.572541382667965, + "grad_norm": 0.3203125, + "learning_rate": 0.00018281914211102542, + "loss": 0.7879, + "step": 1029 + }, + { + "epoch": 0.5730977882876617, + "grad_norm": 0.3046875, + "learning_rate": 0.00018278640748262498, + "loss": 0.6934, + "step": 1030 + }, + { + "epoch": 0.5736541939073585, + "grad_norm": 0.33984375, + "learning_rate": 0.00018275364463579853, + "loss": 0.9117, + "step": 1031 + }, + { + "epoch": 0.5742105995270552, + "grad_norm": 0.34375, + "learning_rate": 0.00018272085358171354, + "loss": 0.8398, + "step": 1032 + }, + { + "epoch": 0.5747670051467519, + "grad_norm": 0.3203125, + "learning_rate": 0.00018268803433154712, + "loss": 0.7672, + "step": 1033 + }, + { + "epoch": 0.5753234107664488, + "grad_norm": 0.3203125, + "learning_rate": 0.00018265518689648596, + "loss": 0.7822, + "step": 1034 + }, + { + "epoch": 0.5758798163861455, + "grad_norm": 0.318359375, + "learning_rate": 0.00018262231128772643, + "loss": 0.7824, + "step": 1035 + }, + { + "epoch": 0.5764362220058422, + "grad_norm": 0.357421875, + "learning_rate": 0.00018258940751647443, + "loss": 0.9375, + "step": 1036 + }, + { + "epoch": 0.576992627625539, + "grad_norm": 0.32421875, + "learning_rate": 0.00018255647559394542, + "loss": 0.8067, + "step": 1037 + }, + { + "epoch": 0.5775490332452358, + "grad_norm": 0.345703125, + "learning_rate": 0.0001825235155313646, + "loss": 0.912, + "step": 1038 + }, + { + "epoch": 0.5781054388649325, + "grad_norm": 0.3515625, + "learning_rate": 0.00018249052733996664, + "loss": 0.9291, + "step": 1039 + }, + { + "epoch": 0.5786618444846293, + "grad_norm": 0.341796875, + "learning_rate": 0.0001824575110309958, + "loss": 0.8022, + "step": 1040 + }, + { + "epoch": 0.579218250104326, + "grad_norm": 0.333984375, + "learning_rate": 0.00018242446661570605, + "loss": 0.8948, + "step": 1041 + }, + { + "epoch": 0.5797746557240228, + "grad_norm": 0.37109375, + "learning_rate": 0.00018239139410536077, + "loss": 0.9978, + "step": 1042 + }, + { + "epoch": 0.5803310613437196, + "grad_norm": 0.32421875, + "learning_rate": 0.00018235829351123304, + "loss": 0.8297, + "step": 1043 + }, + { + "epoch": 0.5808874669634163, + "grad_norm": 0.337890625, + "learning_rate": 0.00018232516484460541, + "loss": 0.8944, + "step": 1044 + }, + { + "epoch": 0.5814438725831131, + "grad_norm": 0.3359375, + "learning_rate": 0.00018229200811677016, + "loss": 0.9035, + "step": 1045 + }, + { + "epoch": 0.5820002782028099, + "grad_norm": 0.322265625, + "learning_rate": 0.00018225882333902896, + "loss": 0.7308, + "step": 1046 + }, + { + "epoch": 0.5825566838225066, + "grad_norm": 0.34765625, + "learning_rate": 0.0001822256105226931, + "loss": 0.8482, + "step": 1047 + }, + { + "epoch": 0.5831130894422034, + "grad_norm": 0.31640625, + "learning_rate": 0.00018219236967908348, + "loss": 0.7428, + "step": 1048 + }, + { + "epoch": 0.5836694950619001, + "grad_norm": 0.33984375, + "learning_rate": 0.00018215910081953053, + "loss": 0.8924, + "step": 1049 + }, + { + "epoch": 0.5842259006815969, + "grad_norm": 0.314453125, + "learning_rate": 0.00018212580395537417, + "loss": 0.7806, + "step": 1050 + }, + { + "epoch": 0.5847823063012937, + "grad_norm": 0.333984375, + "learning_rate": 0.0001820924790979639, + "loss": 0.8435, + "step": 1051 + }, + { + "epoch": 0.5853387119209904, + "grad_norm": 0.34375, + "learning_rate": 0.00018205912625865881, + "loss": 0.9288, + "step": 1052 + }, + { + "epoch": 0.5858951175406871, + "grad_norm": 0.32421875, + "learning_rate": 0.0001820257454488275, + "loss": 0.7463, + "step": 1053 + }, + { + "epoch": 0.586451523160384, + "grad_norm": 0.326171875, + "learning_rate": 0.00018199233667984802, + "loss": 0.7873, + "step": 1054 + }, + { + "epoch": 0.5870079287800807, + "grad_norm": 0.341796875, + "learning_rate": 0.00018195889996310813, + "loss": 0.8771, + "step": 1055 + }, + { + "epoch": 0.5875643343997774, + "grad_norm": 0.330078125, + "learning_rate": 0.0001819254353100049, + "loss": 0.8398, + "step": 1056 + }, + { + "epoch": 0.5881207400194742, + "grad_norm": 0.345703125, + "learning_rate": 0.00018189194273194513, + "loss": 0.9242, + "step": 1057 + }, + { + "epoch": 0.588677145639171, + "grad_norm": 0.341796875, + "learning_rate": 0.00018185842224034495, + "loss": 0.889, + "step": 1058 + }, + { + "epoch": 0.5892335512588677, + "grad_norm": 0.33984375, + "learning_rate": 0.00018182487384663017, + "loss": 0.9042, + "step": 1059 + }, + { + "epoch": 0.5897899568785645, + "grad_norm": 0.328125, + "learning_rate": 0.000181791297562236, + "loss": 0.8742, + "step": 1060 + }, + { + "epoch": 0.5903463624982612, + "grad_norm": 0.328125, + "learning_rate": 0.00018175769339860714, + "loss": 0.8162, + "step": 1061 + }, + { + "epoch": 0.5909027681179579, + "grad_norm": 0.345703125, + "learning_rate": 0.00018172406136719794, + "loss": 0.9006, + "step": 1062 + }, + { + "epoch": 0.5914591737376548, + "grad_norm": 0.322265625, + "learning_rate": 0.00018169040147947212, + "loss": 0.7795, + "step": 1063 + }, + { + "epoch": 0.5920155793573515, + "grad_norm": 0.345703125, + "learning_rate": 0.00018165671374690293, + "loss": 0.8475, + "step": 1064 + }, + { + "epoch": 0.5925719849770482, + "grad_norm": 0.322265625, + "learning_rate": 0.00018162299818097303, + "loss": 0.803, + "step": 1065 + }, + { + "epoch": 0.593128390596745, + "grad_norm": 0.3359375, + "learning_rate": 0.0001815892547931748, + "loss": 0.8489, + "step": 1066 + }, + { + "epoch": 0.5936847962164418, + "grad_norm": 0.330078125, + "learning_rate": 0.00018155548359500981, + "loss": 0.8463, + "step": 1067 + }, + { + "epoch": 0.5942412018361386, + "grad_norm": 0.328125, + "learning_rate": 0.00018152168459798935, + "loss": 0.8165, + "step": 1068 + }, + { + "epoch": 0.5947976074558353, + "grad_norm": 0.330078125, + "learning_rate": 0.00018148785781363401, + "loss": 0.8304, + "step": 1069 + }, + { + "epoch": 0.595354013075532, + "grad_norm": 0.306640625, + "learning_rate": 0.000181454003253474, + "loss": 0.7645, + "step": 1070 + }, + { + "epoch": 0.5959104186952289, + "grad_norm": 0.31640625, + "learning_rate": 0.00018142012092904885, + "loss": 0.8276, + "step": 1071 + }, + { + "epoch": 0.5964668243149256, + "grad_norm": 0.337890625, + "learning_rate": 0.00018138621085190767, + "loss": 0.889, + "step": 1072 + }, + { + "epoch": 0.5970232299346223, + "grad_norm": 0.34765625, + "learning_rate": 0.000181352273033609, + "loss": 0.83, + "step": 1073 + }, + { + "epoch": 0.5975796355543191, + "grad_norm": 0.3515625, + "learning_rate": 0.00018131830748572078, + "loss": 0.9765, + "step": 1074 + }, + { + "epoch": 0.5981360411740159, + "grad_norm": 0.337890625, + "learning_rate": 0.00018128431421982047, + "loss": 0.8316, + "step": 1075 + }, + { + "epoch": 0.5986924467937126, + "grad_norm": 0.33984375, + "learning_rate": 0.00018125029324749499, + "loss": 0.867, + "step": 1076 + }, + { + "epoch": 0.5992488524134094, + "grad_norm": 0.314453125, + "learning_rate": 0.00018121624458034062, + "loss": 0.7405, + "step": 1077 + }, + { + "epoch": 0.5998052580331061, + "grad_norm": 0.333984375, + "learning_rate": 0.0001811821682299631, + "loss": 0.9073, + "step": 1078 + }, + { + "epoch": 0.6003616636528029, + "grad_norm": 0.34375, + "learning_rate": 0.0001811480642079777, + "loss": 0.8368, + "step": 1079 + }, + { + "epoch": 0.6009180692724997, + "grad_norm": 0.3125, + "learning_rate": 0.00018111393252600902, + "loss": 0.7902, + "step": 1080 + }, + { + "epoch": 0.6014744748921964, + "grad_norm": 0.330078125, + "learning_rate": 0.00018107977319569115, + "loss": 0.8296, + "step": 1081 + }, + { + "epoch": 0.6020308805118931, + "grad_norm": 0.32421875, + "learning_rate": 0.00018104558622866756, + "loss": 0.8123, + "step": 1082 + }, + { + "epoch": 0.60258728613159, + "grad_norm": 0.33984375, + "learning_rate": 0.00018101137163659115, + "loss": 0.8452, + "step": 1083 + }, + { + "epoch": 0.6031436917512867, + "grad_norm": 0.353515625, + "learning_rate": 0.0001809771294311242, + "loss": 0.8303, + "step": 1084 + }, + { + "epoch": 0.6037000973709834, + "grad_norm": 0.3203125, + "learning_rate": 0.00018094285962393856, + "loss": 0.7941, + "step": 1085 + }, + { + "epoch": 0.6042565029906802, + "grad_norm": 0.345703125, + "learning_rate": 0.0001809085622267153, + "loss": 0.9412, + "step": 1086 + }, + { + "epoch": 0.604812908610377, + "grad_norm": 0.333984375, + "learning_rate": 0.00018087423725114498, + "loss": 0.8406, + "step": 1087 + }, + { + "epoch": 0.6053693142300737, + "grad_norm": 0.34765625, + "learning_rate": 0.00018083988470892756, + "loss": 0.9588, + "step": 1088 + }, + { + "epoch": 0.6059257198497705, + "grad_norm": 0.349609375, + "learning_rate": 0.0001808055046117724, + "loss": 0.9362, + "step": 1089 + }, + { + "epoch": 0.6064821254694672, + "grad_norm": 0.33203125, + "learning_rate": 0.00018077109697139818, + "loss": 0.8893, + "step": 1090 + }, + { + "epoch": 0.607038531089164, + "grad_norm": 0.328125, + "learning_rate": 0.00018073666179953306, + "loss": 0.8135, + "step": 1091 + }, + { + "epoch": 0.6075949367088608, + "grad_norm": 0.337890625, + "learning_rate": 0.00018070219910791457, + "loss": 0.8649, + "step": 1092 + }, + { + "epoch": 0.6081513423285575, + "grad_norm": 0.322265625, + "learning_rate": 0.0001806677089082896, + "loss": 0.8343, + "step": 1093 + }, + { + "epoch": 0.6087077479482543, + "grad_norm": 0.3359375, + "learning_rate": 0.00018063319121241434, + "loss": 0.8991, + "step": 1094 + }, + { + "epoch": 0.609264153567951, + "grad_norm": 0.3359375, + "learning_rate": 0.00018059864603205456, + "loss": 0.8552, + "step": 1095 + }, + { + "epoch": 0.6098205591876478, + "grad_norm": 0.322265625, + "learning_rate": 0.00018056407337898515, + "loss": 0.7787, + "step": 1096 + }, + { + "epoch": 0.6103769648073446, + "grad_norm": 0.369140625, + "learning_rate": 0.00018052947326499056, + "loss": 0.8752, + "step": 1097 + }, + { + "epoch": 0.6109333704270413, + "grad_norm": 0.337890625, + "learning_rate": 0.0001804948457018645, + "loss": 0.851, + "step": 1098 + }, + { + "epoch": 0.611489776046738, + "grad_norm": 0.3203125, + "learning_rate": 0.00018046019070141006, + "loss": 0.8067, + "step": 1099 + }, + { + "epoch": 0.6120461816664349, + "grad_norm": 0.333984375, + "learning_rate": 0.00018042550827543962, + "loss": 0.8336, + "step": 1100 + }, + { + "epoch": 0.6126025872861316, + "grad_norm": 0.33984375, + "learning_rate": 0.00018039079843577507, + "loss": 0.8959, + "step": 1101 + }, + { + "epoch": 0.6131589929058283, + "grad_norm": 0.328125, + "learning_rate": 0.0001803560611942475, + "loss": 0.8092, + "step": 1102 + }, + { + "epoch": 0.6137153985255251, + "grad_norm": 0.33203125, + "learning_rate": 0.00018032129656269738, + "loss": 0.8899, + "step": 1103 + }, + { + "epoch": 0.6142718041452219, + "grad_norm": 0.365234375, + "learning_rate": 0.0001802865045529745, + "loss": 0.9702, + "step": 1104 + }, + { + "epoch": 0.6148282097649186, + "grad_norm": 0.31640625, + "learning_rate": 0.00018025168517693803, + "loss": 0.7923, + "step": 1105 + }, + { + "epoch": 0.6153846153846154, + "grad_norm": 0.328125, + "learning_rate": 0.00018021683844645644, + "loss": 0.9082, + "step": 1106 + }, + { + "epoch": 0.6159410210043121, + "grad_norm": 0.33203125, + "learning_rate": 0.0001801819643734075, + "loss": 0.8512, + "step": 1107 + }, + { + "epoch": 0.6164974266240089, + "grad_norm": 0.333984375, + "learning_rate": 0.00018014706296967838, + "loss": 0.8654, + "step": 1108 + }, + { + "epoch": 0.6170538322437057, + "grad_norm": 0.333984375, + "learning_rate": 0.00018011213424716545, + "loss": 0.9283, + "step": 1109 + }, + { + "epoch": 0.6176102378634024, + "grad_norm": 0.31640625, + "learning_rate": 0.00018007717821777445, + "loss": 0.7825, + "step": 1110 + }, + { + "epoch": 0.6181666434830991, + "grad_norm": 0.33203125, + "learning_rate": 0.00018004219489342046, + "loss": 0.8412, + "step": 1111 + }, + { + "epoch": 0.618723049102796, + "grad_norm": 0.33203125, + "learning_rate": 0.00018000718428602786, + "loss": 0.8454, + "step": 1112 + }, + { + "epoch": 0.6192794547224927, + "grad_norm": 0.337890625, + "learning_rate": 0.00017997214640753023, + "loss": 0.9071, + "step": 1113 + }, + { + "epoch": 0.6198358603421894, + "grad_norm": 0.333984375, + "learning_rate": 0.00017993708126987058, + "loss": 0.9327, + "step": 1114 + }, + { + "epoch": 0.6203922659618862, + "grad_norm": 0.314453125, + "learning_rate": 0.0001799019888850011, + "loss": 0.7861, + "step": 1115 + }, + { + "epoch": 0.620948671581583, + "grad_norm": 0.3203125, + "learning_rate": 0.00017986686926488337, + "loss": 0.8135, + "step": 1116 + }, + { + "epoch": 0.6215050772012798, + "grad_norm": 0.34375, + "learning_rate": 0.00017983172242148814, + "loss": 0.9299, + "step": 1117 + }, + { + "epoch": 0.6220614828209765, + "grad_norm": 0.345703125, + "learning_rate": 0.00017979654836679555, + "loss": 0.8861, + "step": 1118 + }, + { + "epoch": 0.6226178884406732, + "grad_norm": 0.314453125, + "learning_rate": 0.00017976134711279496, + "loss": 0.7301, + "step": 1119 + }, + { + "epoch": 0.6231742940603701, + "grad_norm": 0.33203125, + "learning_rate": 0.00017972611867148497, + "loss": 0.8426, + "step": 1120 + }, + { + "epoch": 0.6237306996800668, + "grad_norm": 0.345703125, + "learning_rate": 0.00017969086305487354, + "loss": 0.9108, + "step": 1121 + }, + { + "epoch": 0.6242871052997635, + "grad_norm": 0.345703125, + "learning_rate": 0.00017965558027497775, + "loss": 0.9061, + "step": 1122 + }, + { + "epoch": 0.6248435109194603, + "grad_norm": 0.326171875, + "learning_rate": 0.00017962027034382411, + "loss": 0.7835, + "step": 1123 + }, + { + "epoch": 0.625399916539157, + "grad_norm": 0.353515625, + "learning_rate": 0.00017958493327344826, + "loss": 0.92, + "step": 1124 + }, + { + "epoch": 0.6259563221588538, + "grad_norm": 0.345703125, + "learning_rate": 0.00017954956907589517, + "loss": 0.9074, + "step": 1125 + }, + { + "epoch": 0.6265127277785506, + "grad_norm": 0.33203125, + "learning_rate": 0.00017951417776321894, + "loss": 0.8155, + "step": 1126 + }, + { + "epoch": 0.6270691333982473, + "grad_norm": 0.337890625, + "learning_rate": 0.00017947875934748306, + "loss": 0.8642, + "step": 1127 + }, + { + "epoch": 0.627625539017944, + "grad_norm": 0.35546875, + "learning_rate": 0.00017944331384076015, + "loss": 0.9221, + "step": 1128 + }, + { + "epoch": 0.6281819446376409, + "grad_norm": 0.330078125, + "learning_rate": 0.00017940784125513212, + "loss": 0.8576, + "step": 1129 + }, + { + "epoch": 0.6287383502573376, + "grad_norm": 0.34375, + "learning_rate": 0.0001793723416026901, + "loss": 0.8981, + "step": 1130 + }, + { + "epoch": 0.6292947558770343, + "grad_norm": 0.359375, + "learning_rate": 0.0001793368148955344, + "loss": 0.8868, + "step": 1131 + }, + { + "epoch": 0.6298511614967311, + "grad_norm": 0.359375, + "learning_rate": 0.00017930126114577464, + "loss": 0.9415, + "step": 1132 + }, + { + "epoch": 0.6304075671164279, + "grad_norm": 0.328125, + "learning_rate": 0.00017926568036552953, + "loss": 0.7919, + "step": 1133 + }, + { + "epoch": 0.6309639727361246, + "grad_norm": 0.357421875, + "learning_rate": 0.00017923007256692713, + "loss": 0.9498, + "step": 1134 + }, + { + "epoch": 0.6315203783558214, + "grad_norm": 0.34375, + "learning_rate": 0.0001791944377621047, + "loss": 0.9427, + "step": 1135 + }, + { + "epoch": 0.6320767839755181, + "grad_norm": 0.3359375, + "learning_rate": 0.00017915877596320855, + "loss": 0.8621, + "step": 1136 + }, + { + "epoch": 0.6326331895952149, + "grad_norm": 0.302734375, + "learning_rate": 0.00017912308718239433, + "loss": 0.7246, + "step": 1137 + }, + { + "epoch": 0.6331895952149117, + "grad_norm": 0.365234375, + "learning_rate": 0.0001790873714318269, + "loss": 0.9493, + "step": 1138 + }, + { + "epoch": 0.6337460008346084, + "grad_norm": 0.328125, + "learning_rate": 0.00017905162872368022, + "loss": 0.8226, + "step": 1139 + }, + { + "epoch": 0.6343024064543051, + "grad_norm": 0.34765625, + "learning_rate": 0.0001790158590701375, + "loss": 0.8976, + "step": 1140 + }, + { + "epoch": 0.634858812074002, + "grad_norm": 0.345703125, + "learning_rate": 0.0001789800624833911, + "loss": 0.9134, + "step": 1141 + }, + { + "epoch": 0.6354152176936987, + "grad_norm": 0.328125, + "learning_rate": 0.0001789442389756426, + "loss": 0.8021, + "step": 1142 + }, + { + "epoch": 0.6359716233133955, + "grad_norm": 0.33984375, + "learning_rate": 0.00017890838855910276, + "loss": 0.8757, + "step": 1143 + }, + { + "epoch": 0.6365280289330922, + "grad_norm": 0.328125, + "learning_rate": 0.00017887251124599143, + "loss": 0.8675, + "step": 1144 + }, + { + "epoch": 0.637084434552789, + "grad_norm": 0.326171875, + "learning_rate": 0.00017883660704853775, + "loss": 0.811, + "step": 1145 + }, + { + "epoch": 0.6376408401724858, + "grad_norm": 0.33984375, + "learning_rate": 0.0001788006759789799, + "loss": 0.8422, + "step": 1146 + }, + { + "epoch": 0.6381972457921825, + "grad_norm": 0.318359375, + "learning_rate": 0.00017876471804956535, + "loss": 0.751, + "step": 1147 + }, + { + "epoch": 0.6387536514118792, + "grad_norm": 0.326171875, + "learning_rate": 0.0001787287332725506, + "loss": 0.793, + "step": 1148 + }, + { + "epoch": 0.6393100570315761, + "grad_norm": 0.326171875, + "learning_rate": 0.00017869272166020134, + "loss": 0.7835, + "step": 1149 + }, + { + "epoch": 0.6398664626512728, + "grad_norm": 0.33203125, + "learning_rate": 0.0001786566832247925, + "loss": 0.8919, + "step": 1150 + }, + { + "epoch": 0.6404228682709695, + "grad_norm": 0.337890625, + "learning_rate": 0.00017862061797860798, + "loss": 0.8714, + "step": 1151 + }, + { + "epoch": 0.6409792738906663, + "grad_norm": 0.33984375, + "learning_rate": 0.000178584525933941, + "loss": 0.9351, + "step": 1152 + }, + { + "epoch": 0.641535679510363, + "grad_norm": 0.33203125, + "learning_rate": 0.0001785484071030938, + "loss": 0.92, + "step": 1153 + }, + { + "epoch": 0.6420920851300598, + "grad_norm": 0.333984375, + "learning_rate": 0.00017851226149837778, + "loss": 0.8925, + "step": 1154 + }, + { + "epoch": 0.6426484907497566, + "grad_norm": 0.33984375, + "learning_rate": 0.00017847608913211342, + "loss": 0.9021, + "step": 1155 + }, + { + "epoch": 0.6432048963694533, + "grad_norm": 0.337890625, + "learning_rate": 0.00017843989001663043, + "loss": 0.8723, + "step": 1156 + }, + { + "epoch": 0.64376130198915, + "grad_norm": 0.33984375, + "learning_rate": 0.00017840366416426755, + "loss": 0.9426, + "step": 1157 + }, + { + "epoch": 0.6443177076088469, + "grad_norm": 0.3515625, + "learning_rate": 0.00017836741158737266, + "loss": 0.8317, + "step": 1158 + }, + { + "epoch": 0.6448741132285436, + "grad_norm": 0.3359375, + "learning_rate": 0.0001783311322983028, + "loss": 0.906, + "step": 1159 + }, + { + "epoch": 0.6454305188482403, + "grad_norm": 0.33203125, + "learning_rate": 0.00017829482630942394, + "loss": 0.8644, + "step": 1160 + }, + { + "epoch": 0.6459869244679372, + "grad_norm": 0.3125, + "learning_rate": 0.0001782584936331114, + "loss": 0.7903, + "step": 1161 + }, + { + "epoch": 0.6465433300876339, + "grad_norm": 0.33203125, + "learning_rate": 0.0001782221342817494, + "loss": 0.8909, + "step": 1162 + }, + { + "epoch": 0.6470997357073306, + "grad_norm": 0.333984375, + "learning_rate": 0.00017818574826773133, + "loss": 0.8217, + "step": 1163 + }, + { + "epoch": 0.6476561413270274, + "grad_norm": 0.31640625, + "learning_rate": 0.0001781493356034597, + "loss": 0.7876, + "step": 1164 + }, + { + "epoch": 0.6482125469467241, + "grad_norm": 0.337890625, + "learning_rate": 0.00017811289630134601, + "loss": 0.8962, + "step": 1165 + }, + { + "epoch": 0.648768952566421, + "grad_norm": 0.32421875, + "learning_rate": 0.00017807643037381094, + "loss": 0.8133, + "step": 1166 + }, + { + "epoch": 0.6493253581861177, + "grad_norm": 0.302734375, + "learning_rate": 0.0001780399378332842, + "loss": 0.6899, + "step": 1167 + }, + { + "epoch": 0.6498817638058144, + "grad_norm": 0.33984375, + "learning_rate": 0.00017800341869220456, + "loss": 0.8604, + "step": 1168 + }, + { + "epoch": 0.6504381694255112, + "grad_norm": 0.328125, + "learning_rate": 0.00017796687296301985, + "loss": 0.7746, + "step": 1169 + }, + { + "epoch": 0.650994575045208, + "grad_norm": 0.330078125, + "learning_rate": 0.00017793030065818699, + "loss": 0.8599, + "step": 1170 + }, + { + "epoch": 0.6515509806649047, + "grad_norm": 0.326171875, + "learning_rate": 0.00017789370179017194, + "loss": 0.8365, + "step": 1171 + }, + { + "epoch": 0.6521073862846015, + "grad_norm": 0.32421875, + "learning_rate": 0.00017785707637144976, + "loss": 0.8358, + "step": 1172 + }, + { + "epoch": 0.6526637919042982, + "grad_norm": 0.3359375, + "learning_rate": 0.0001778204244145045, + "loss": 0.8689, + "step": 1173 + }, + { + "epoch": 0.653220197523995, + "grad_norm": 0.32421875, + "learning_rate": 0.00017778374593182931, + "loss": 0.8538, + "step": 1174 + }, + { + "epoch": 0.6537766031436918, + "grad_norm": 0.337890625, + "learning_rate": 0.00017774704093592632, + "loss": 0.8759, + "step": 1175 + }, + { + "epoch": 0.6543330087633885, + "grad_norm": 0.337890625, + "learning_rate": 0.0001777103094393067, + "loss": 0.902, + "step": 1176 + }, + { + "epoch": 0.6548894143830852, + "grad_norm": 0.3203125, + "learning_rate": 0.00017767355145449073, + "loss": 0.8503, + "step": 1177 + }, + { + "epoch": 0.6554458200027821, + "grad_norm": 0.322265625, + "learning_rate": 0.0001776367669940077, + "loss": 0.8096, + "step": 1178 + }, + { + "epoch": 0.6560022256224788, + "grad_norm": 0.32421875, + "learning_rate": 0.0001775999560703958, + "loss": 0.7987, + "step": 1179 + }, + { + "epoch": 0.6565586312421755, + "grad_norm": 0.333984375, + "learning_rate": 0.0001775631186962024, + "loss": 0.8789, + "step": 1180 + }, + { + "epoch": 0.6571150368618723, + "grad_norm": 0.3359375, + "learning_rate": 0.00017752625488398383, + "loss": 0.8525, + "step": 1181 + }, + { + "epoch": 0.6576714424815691, + "grad_norm": 0.328125, + "learning_rate": 0.0001774893646463054, + "loss": 0.8729, + "step": 1182 + }, + { + "epoch": 0.6582278481012658, + "grad_norm": 0.337890625, + "learning_rate": 0.00017745244799574145, + "loss": 0.9459, + "step": 1183 + }, + { + "epoch": 0.6587842537209626, + "grad_norm": 0.333984375, + "learning_rate": 0.00017741550494487532, + "loss": 0.8822, + "step": 1184 + }, + { + "epoch": 0.6593406593406593, + "grad_norm": 0.35546875, + "learning_rate": 0.00017737853550629936, + "loss": 0.9891, + "step": 1185 + }, + { + "epoch": 0.659897064960356, + "grad_norm": 0.31640625, + "learning_rate": 0.0001773415396926149, + "loss": 0.7885, + "step": 1186 + }, + { + "epoch": 0.6604534705800529, + "grad_norm": 0.337890625, + "learning_rate": 0.00017730451751643228, + "loss": 0.8582, + "step": 1187 + }, + { + "epoch": 0.6610098761997496, + "grad_norm": 0.328125, + "learning_rate": 0.00017726746899037079, + "loss": 0.9027, + "step": 1188 + }, + { + "epoch": 0.6615662818194463, + "grad_norm": 0.32421875, + "learning_rate": 0.00017723039412705877, + "loss": 0.802, + "step": 1189 + }, + { + "epoch": 0.6621226874391432, + "grad_norm": 0.33984375, + "learning_rate": 0.00017719329293913342, + "loss": 0.8545, + "step": 1190 + }, + { + "epoch": 0.6626790930588399, + "grad_norm": 0.326171875, + "learning_rate": 0.00017715616543924104, + "loss": 0.8496, + "step": 1191 + }, + { + "epoch": 0.6632354986785367, + "grad_norm": 0.31640625, + "learning_rate": 0.0001771190116400368, + "loss": 0.7292, + "step": 1192 + }, + { + "epoch": 0.6637919042982334, + "grad_norm": 0.326171875, + "learning_rate": 0.00017708183155418493, + "loss": 0.7995, + "step": 1193 + }, + { + "epoch": 0.6643483099179301, + "grad_norm": 0.328125, + "learning_rate": 0.00017704462519435853, + "loss": 0.8858, + "step": 1194 + }, + { + "epoch": 0.664904715537627, + "grad_norm": 0.318359375, + "learning_rate": 0.0001770073925732397, + "loss": 0.8412, + "step": 1195 + }, + { + "epoch": 0.6654611211573237, + "grad_norm": 0.31640625, + "learning_rate": 0.0001769701337035195, + "loss": 0.8375, + "step": 1196 + }, + { + "epoch": 0.6660175267770204, + "grad_norm": 0.337890625, + "learning_rate": 0.0001769328485978979, + "loss": 0.8676, + "step": 1197 + }, + { + "epoch": 0.6665739323967173, + "grad_norm": 0.361328125, + "learning_rate": 0.00017689553726908387, + "loss": 0.9652, + "step": 1198 + }, + { + "epoch": 0.667130338016414, + "grad_norm": 0.32421875, + "learning_rate": 0.00017685819972979524, + "loss": 0.7994, + "step": 1199 + }, + { + "epoch": 0.6676867436361107, + "grad_norm": 0.333984375, + "learning_rate": 0.0001768208359927588, + "loss": 0.7993, + "step": 1200 + }, + { + "epoch": 0.6682431492558075, + "grad_norm": 0.328125, + "learning_rate": 0.00017678344607071034, + "loss": 0.7557, + "step": 1201 + }, + { + "epoch": 0.6687995548755042, + "grad_norm": 0.322265625, + "learning_rate": 0.0001767460299763945, + "loss": 0.8166, + "step": 1202 + }, + { + "epoch": 0.669355960495201, + "grad_norm": 0.33203125, + "learning_rate": 0.00017670858772256487, + "loss": 0.8687, + "step": 1203 + }, + { + "epoch": 0.6699123661148978, + "grad_norm": 0.328125, + "learning_rate": 0.00017667111932198392, + "loss": 0.8069, + "step": 1204 + }, + { + "epoch": 0.6704687717345945, + "grad_norm": 0.333984375, + "learning_rate": 0.0001766336247874231, + "loss": 0.8764, + "step": 1205 + }, + { + "epoch": 0.6710251773542912, + "grad_norm": 0.310546875, + "learning_rate": 0.00017659610413166269, + "loss": 0.7519, + "step": 1206 + }, + { + "epoch": 0.6715815829739881, + "grad_norm": 0.353515625, + "learning_rate": 0.00017655855736749198, + "loss": 0.945, + "step": 1207 + }, + { + "epoch": 0.6721379885936848, + "grad_norm": 0.328125, + "learning_rate": 0.00017652098450770902, + "loss": 0.8002, + "step": 1208 + }, + { + "epoch": 0.6726943942133815, + "grad_norm": 0.333984375, + "learning_rate": 0.00017648338556512086, + "loss": 0.8359, + "step": 1209 + }, + { + "epoch": 0.6732507998330783, + "grad_norm": 0.337890625, + "learning_rate": 0.00017644576055254337, + "loss": 0.8427, + "step": 1210 + }, + { + "epoch": 0.6738072054527751, + "grad_norm": 0.33984375, + "learning_rate": 0.00017640810948280143, + "loss": 0.8537, + "step": 1211 + }, + { + "epoch": 0.6743636110724718, + "grad_norm": 0.34765625, + "learning_rate": 0.00017637043236872865, + "loss": 0.924, + "step": 1212 + }, + { + "epoch": 0.6749200166921686, + "grad_norm": 0.34765625, + "learning_rate": 0.00017633272922316757, + "loss": 0.9067, + "step": 1213 + }, + { + "epoch": 0.6754764223118653, + "grad_norm": 0.345703125, + "learning_rate": 0.00017629500005896968, + "loss": 0.9031, + "step": 1214 + }, + { + "epoch": 0.6760328279315622, + "grad_norm": 0.328125, + "learning_rate": 0.0001762572448889952, + "loss": 0.7696, + "step": 1215 + }, + { + "epoch": 0.6765892335512589, + "grad_norm": 0.30078125, + "learning_rate": 0.00017621946372611336, + "loss": 0.7062, + "step": 1216 + }, + { + "epoch": 0.6771456391709556, + "grad_norm": 0.337890625, + "learning_rate": 0.00017618165658320214, + "loss": 0.8407, + "step": 1217 + }, + { + "epoch": 0.6777020447906524, + "grad_norm": 0.33203125, + "learning_rate": 0.00017614382347314842, + "loss": 0.8517, + "step": 1218 + }, + { + "epoch": 0.6782584504103492, + "grad_norm": 0.33203125, + "learning_rate": 0.00017610596440884792, + "loss": 0.8495, + "step": 1219 + }, + { + "epoch": 0.6788148560300459, + "grad_norm": 0.33984375, + "learning_rate": 0.00017606807940320524, + "loss": 0.8023, + "step": 1220 + }, + { + "epoch": 0.6793712616497427, + "grad_norm": 0.353515625, + "learning_rate": 0.00017603016846913375, + "loss": 0.8952, + "step": 1221 + }, + { + "epoch": 0.6799276672694394, + "grad_norm": 0.337890625, + "learning_rate": 0.00017599223161955573, + "loss": 0.8634, + "step": 1222 + }, + { + "epoch": 0.6804840728891361, + "grad_norm": 0.3515625, + "learning_rate": 0.0001759542688674023, + "loss": 0.9523, + "step": 1223 + }, + { + "epoch": 0.681040478508833, + "grad_norm": 0.314453125, + "learning_rate": 0.0001759162802256133, + "loss": 0.7981, + "step": 1224 + }, + { + "epoch": 0.6815968841285297, + "grad_norm": 0.33984375, + "learning_rate": 0.00017587826570713752, + "loss": 0.9015, + "step": 1225 + }, + { + "epoch": 0.6821532897482264, + "grad_norm": 0.33203125, + "learning_rate": 0.00017584022532493254, + "loss": 0.8965, + "step": 1226 + }, + { + "epoch": 0.6827096953679233, + "grad_norm": 0.330078125, + "learning_rate": 0.00017580215909196467, + "loss": 0.8798, + "step": 1227 + }, + { + "epoch": 0.68326610098762, + "grad_norm": 0.32421875, + "learning_rate": 0.00017576406702120916, + "loss": 0.8244, + "step": 1228 + }, + { + "epoch": 0.6838225066073167, + "grad_norm": 0.322265625, + "learning_rate": 0.00017572594912564996, + "loss": 0.8822, + "step": 1229 + }, + { + "epoch": 0.6843789122270135, + "grad_norm": 0.328125, + "learning_rate": 0.00017568780541827987, + "loss": 0.8287, + "step": 1230 + }, + { + "epoch": 0.6849353178467102, + "grad_norm": 0.33203125, + "learning_rate": 0.00017564963591210054, + "loss": 0.8288, + "step": 1231 + }, + { + "epoch": 0.685491723466407, + "grad_norm": 0.3515625, + "learning_rate": 0.0001756114406201223, + "loss": 0.9075, + "step": 1232 + }, + { + "epoch": 0.6860481290861038, + "grad_norm": 0.337890625, + "learning_rate": 0.00017557321955536436, + "loss": 0.9005, + "step": 1233 + }, + { + "epoch": 0.6866045347058005, + "grad_norm": 0.322265625, + "learning_rate": 0.00017553497273085466, + "loss": 0.8201, + "step": 1234 + }, + { + "epoch": 0.6871609403254972, + "grad_norm": 0.328125, + "learning_rate": 0.00017549670015962998, + "loss": 0.8144, + "step": 1235 + }, + { + "epoch": 0.6877173459451941, + "grad_norm": 0.337890625, + "learning_rate": 0.00017545840185473582, + "loss": 0.8734, + "step": 1236 + }, + { + "epoch": 0.6882737515648908, + "grad_norm": 0.33984375, + "learning_rate": 0.00017542007782922646, + "loss": 0.8758, + "step": 1237 + }, + { + "epoch": 0.6888301571845875, + "grad_norm": 0.322265625, + "learning_rate": 0.00017538172809616494, + "loss": 0.7767, + "step": 1238 + }, + { + "epoch": 0.6893865628042843, + "grad_norm": 0.314453125, + "learning_rate": 0.00017534335266862315, + "loss": 0.7504, + "step": 1239 + }, + { + "epoch": 0.6899429684239811, + "grad_norm": 0.328125, + "learning_rate": 0.0001753049515596816, + "loss": 0.8237, + "step": 1240 + }, + { + "epoch": 0.6904993740436779, + "grad_norm": 0.3359375, + "learning_rate": 0.00017526652478242966, + "loss": 0.7673, + "step": 1241 + }, + { + "epoch": 0.6910557796633746, + "grad_norm": 0.3203125, + "learning_rate": 0.00017522807234996538, + "loss": 0.8724, + "step": 1242 + }, + { + "epoch": 0.6916121852830713, + "grad_norm": 0.326171875, + "learning_rate": 0.00017518959427539562, + "loss": 0.8233, + "step": 1243 + }, + { + "epoch": 0.6921685909027682, + "grad_norm": 0.306640625, + "learning_rate": 0.00017515109057183596, + "loss": 0.6871, + "step": 1244 + }, + { + "epoch": 0.6927249965224649, + "grad_norm": 0.33203125, + "learning_rate": 0.00017511256125241068, + "loss": 0.7386, + "step": 1245 + }, + { + "epoch": 0.6932814021421616, + "grad_norm": 0.337890625, + "learning_rate": 0.00017507400633025278, + "loss": 0.7724, + "step": 1246 + }, + { + "epoch": 0.6938378077618584, + "grad_norm": 0.32421875, + "learning_rate": 0.00017503542581850408, + "loss": 0.81, + "step": 1247 + }, + { + "epoch": 0.6943942133815552, + "grad_norm": 0.80859375, + "learning_rate": 0.00017499681973031503, + "loss": 1.8259, + "step": 1248 + }, + { + "epoch": 0.6949506190012519, + "grad_norm": 0.326171875, + "learning_rate": 0.00017495818807884486, + "loss": 0.7351, + "step": 1249 + }, + { + "epoch": 0.6955070246209487, + "grad_norm": 0.3515625, + "learning_rate": 0.00017491953087726142, + "loss": 0.9045, + "step": 1250 + }, + { + "epoch": 0.6960634302406454, + "grad_norm": 0.31640625, + "learning_rate": 0.00017488084813874143, + "loss": 0.6979, + "step": 1251 + }, + { + "epoch": 0.6966198358603422, + "grad_norm": 0.353515625, + "learning_rate": 0.00017484213987647014, + "loss": 0.8801, + "step": 1252 + }, + { + "epoch": 0.697176241480039, + "grad_norm": 0.35546875, + "learning_rate": 0.00017480340610364161, + "loss": 0.8701, + "step": 1253 + }, + { + "epoch": 0.6977326470997357, + "grad_norm": 0.341796875, + "learning_rate": 0.0001747646468334586, + "loss": 0.8198, + "step": 1254 + }, + { + "epoch": 0.6982890527194324, + "grad_norm": 0.3359375, + "learning_rate": 0.00017472586207913245, + "loss": 0.8907, + "step": 1255 + }, + { + "epoch": 0.6988454583391293, + "grad_norm": 0.345703125, + "learning_rate": 0.00017468705185388333, + "loss": 0.8712, + "step": 1256 + }, + { + "epoch": 0.699401863958826, + "grad_norm": 0.365234375, + "learning_rate": 0.00017464821617093997, + "loss": 1.0164, + "step": 1257 + }, + { + "epoch": 0.6999582695785227, + "grad_norm": 0.349609375, + "learning_rate": 0.00017460935504353984, + "loss": 0.9493, + "step": 1258 + }, + { + "epoch": 0.7005146751982195, + "grad_norm": 0.322265625, + "learning_rate": 0.0001745704684849291, + "loss": 0.7055, + "step": 1259 + }, + { + "epoch": 0.7010710808179162, + "grad_norm": 0.33203125, + "learning_rate": 0.00017453155650836255, + "loss": 0.8317, + "step": 1260 + }, + { + "epoch": 0.701627486437613, + "grad_norm": 0.326171875, + "learning_rate": 0.00017449261912710364, + "loss": 0.8239, + "step": 1261 + }, + { + "epoch": 0.7021838920573098, + "grad_norm": 0.318359375, + "learning_rate": 0.0001744536563544245, + "loss": 0.771, + "step": 1262 + }, + { + "epoch": 0.7027402976770065, + "grad_norm": 0.345703125, + "learning_rate": 0.00017441466820360596, + "loss": 0.9036, + "step": 1263 + }, + { + "epoch": 0.7032967032967034, + "grad_norm": 0.34765625, + "learning_rate": 0.00017437565468793732, + "loss": 0.7954, + "step": 1264 + }, + { + "epoch": 0.7038531089164001, + "grad_norm": 0.333984375, + "learning_rate": 0.00017433661582071677, + "loss": 0.8681, + "step": 1265 + }, + { + "epoch": 0.7044095145360968, + "grad_norm": 0.33984375, + "learning_rate": 0.00017429755161525105, + "loss": 0.8394, + "step": 1266 + }, + { + "epoch": 0.7049659201557936, + "grad_norm": 0.337890625, + "learning_rate": 0.00017425846208485543, + "loss": 0.8418, + "step": 1267 + }, + { + "epoch": 0.7055223257754903, + "grad_norm": 0.396484375, + "learning_rate": 0.00017421934724285394, + "loss": 0.9183, + "step": 1268 + }, + { + "epoch": 0.7060787313951871, + "grad_norm": 0.337890625, + "learning_rate": 0.00017418020710257916, + "loss": 0.8866, + "step": 1269 + }, + { + "epoch": 0.7066351370148839, + "grad_norm": 0.32421875, + "learning_rate": 0.00017414104167737238, + "loss": 0.8109, + "step": 1270 + }, + { + "epoch": 0.7071915426345806, + "grad_norm": 0.33984375, + "learning_rate": 0.0001741018509805834, + "loss": 0.922, + "step": 1271 + }, + { + "epoch": 0.7077479482542773, + "grad_norm": 0.34375, + "learning_rate": 0.00017406263502557073, + "loss": 0.9158, + "step": 1272 + }, + { + "epoch": 0.7083043538739742, + "grad_norm": 0.337890625, + "learning_rate": 0.00017402339382570147, + "loss": 0.8446, + "step": 1273 + }, + { + "epoch": 0.7088607594936709, + "grad_norm": 0.33984375, + "learning_rate": 0.00017398412739435125, + "loss": 0.881, + "step": 1274 + }, + { + "epoch": 0.7094171651133676, + "grad_norm": 0.337890625, + "learning_rate": 0.00017394483574490437, + "loss": 0.9032, + "step": 1275 + }, + { + "epoch": 0.7099735707330644, + "grad_norm": 0.333984375, + "learning_rate": 0.00017390551889075374, + "loss": 0.8782, + "step": 1276 + }, + { + "epoch": 0.7105299763527612, + "grad_norm": 0.33203125, + "learning_rate": 0.00017386617684530083, + "loss": 0.8744, + "step": 1277 + }, + { + "epoch": 0.7110863819724579, + "grad_norm": 0.330078125, + "learning_rate": 0.00017382680962195566, + "loss": 0.8058, + "step": 1278 + }, + { + "epoch": 0.7116427875921547, + "grad_norm": 0.314453125, + "learning_rate": 0.0001737874172341369, + "loss": 0.6973, + "step": 1279 + }, + { + "epoch": 0.7121991932118514, + "grad_norm": 0.34375, + "learning_rate": 0.00017374799969527174, + "loss": 0.8964, + "step": 1280 + }, + { + "epoch": 0.7127555988315482, + "grad_norm": 0.3203125, + "learning_rate": 0.000173708557018796, + "loss": 0.8244, + "step": 1281 + }, + { + "epoch": 0.713312004451245, + "grad_norm": 0.337890625, + "learning_rate": 0.00017366908921815404, + "loss": 0.8864, + "step": 1282 + }, + { + "epoch": 0.7138684100709417, + "grad_norm": 0.328125, + "learning_rate": 0.0001736295963067988, + "loss": 0.8349, + "step": 1283 + }, + { + "epoch": 0.7144248156906384, + "grad_norm": 0.34375, + "learning_rate": 0.0001735900782981917, + "loss": 0.8759, + "step": 1284 + }, + { + "epoch": 0.7149812213103353, + "grad_norm": 0.306640625, + "learning_rate": 0.00017355053520580284, + "loss": 0.7247, + "step": 1285 + }, + { + "epoch": 0.715537626930032, + "grad_norm": 0.3203125, + "learning_rate": 0.0001735109670431108, + "loss": 0.8187, + "step": 1286 + }, + { + "epoch": 0.7160940325497287, + "grad_norm": 0.328125, + "learning_rate": 0.00017347137382360268, + "loss": 0.8287, + "step": 1287 + }, + { + "epoch": 0.7166504381694255, + "grad_norm": 0.333984375, + "learning_rate": 0.00017343175556077417, + "loss": 0.8427, + "step": 1288 + }, + { + "epoch": 0.7172068437891223, + "grad_norm": 0.318359375, + "learning_rate": 0.00017339211226812947, + "loss": 0.7638, + "step": 1289 + }, + { + "epoch": 0.7177632494088191, + "grad_norm": 0.341796875, + "learning_rate": 0.00017335244395918136, + "loss": 0.8827, + "step": 1290 + }, + { + "epoch": 0.7183196550285158, + "grad_norm": 0.337890625, + "learning_rate": 0.00017331275064745105, + "loss": 0.925, + "step": 1291 + }, + { + "epoch": 0.7188760606482125, + "grad_norm": 0.310546875, + "learning_rate": 0.00017327303234646838, + "loss": 0.7816, + "step": 1292 + }, + { + "epoch": 0.7194324662679094, + "grad_norm": 0.33984375, + "learning_rate": 0.00017323328906977166, + "loss": 0.9068, + "step": 1293 + }, + { + "epoch": 0.7199888718876061, + "grad_norm": 0.322265625, + "learning_rate": 0.00017319352083090765, + "loss": 0.8168, + "step": 1294 + }, + { + "epoch": 0.7205452775073028, + "grad_norm": 0.326171875, + "learning_rate": 0.00017315372764343173, + "loss": 0.8039, + "step": 1295 + }, + { + "epoch": 0.7211016831269996, + "grad_norm": 0.318359375, + "learning_rate": 0.00017311390952090775, + "loss": 0.7918, + "step": 1296 + }, + { + "epoch": 0.7216580887466963, + "grad_norm": 0.330078125, + "learning_rate": 0.000173074066476908, + "loss": 0.8398, + "step": 1297 + }, + { + "epoch": 0.7222144943663931, + "grad_norm": 0.337890625, + "learning_rate": 0.00017303419852501332, + "loss": 0.7868, + "step": 1298 + }, + { + "epoch": 0.7227708999860899, + "grad_norm": 0.318359375, + "learning_rate": 0.00017299430567881306, + "loss": 0.7343, + "step": 1299 + }, + { + "epoch": 0.7233273056057866, + "grad_norm": 0.318359375, + "learning_rate": 0.000172954387951905, + "loss": 0.8459, + "step": 1300 + }, + { + "epoch": 0.7238837112254833, + "grad_norm": 0.341796875, + "learning_rate": 0.0001729144453578954, + "loss": 0.8439, + "step": 1301 + }, + { + "epoch": 0.7244401168451802, + "grad_norm": 0.33984375, + "learning_rate": 0.00017287447791039905, + "loss": 0.8974, + "step": 1302 + }, + { + "epoch": 0.7249965224648769, + "grad_norm": 0.333984375, + "learning_rate": 0.00017283448562303922, + "loss": 0.8415, + "step": 1303 + }, + { + "epoch": 0.7255529280845736, + "grad_norm": 0.310546875, + "learning_rate": 0.00017279446850944753, + "loss": 0.8327, + "step": 1304 + }, + { + "epoch": 0.7261093337042704, + "grad_norm": 0.35546875, + "learning_rate": 0.0001727544265832642, + "loss": 0.9265, + "step": 1305 + }, + { + "epoch": 0.7266657393239672, + "grad_norm": 0.33203125, + "learning_rate": 0.0001727143598581378, + "loss": 0.8761, + "step": 1306 + }, + { + "epoch": 0.7272221449436639, + "grad_norm": 0.3359375, + "learning_rate": 0.00017267426834772546, + "loss": 0.8675, + "step": 1307 + }, + { + "epoch": 0.7277785505633607, + "grad_norm": 0.33203125, + "learning_rate": 0.00017263415206569268, + "loss": 0.8685, + "step": 1308 + }, + { + "epoch": 0.7283349561830574, + "grad_norm": 0.326171875, + "learning_rate": 0.00017259401102571336, + "loss": 0.882, + "step": 1309 + }, + { + "epoch": 0.7288913618027542, + "grad_norm": 0.314453125, + "learning_rate": 0.00017255384524146998, + "loss": 0.7924, + "step": 1310 + }, + { + "epoch": 0.729447767422451, + "grad_norm": 0.337890625, + "learning_rate": 0.0001725136547266534, + "loss": 0.8117, + "step": 1311 + }, + { + "epoch": 0.7300041730421477, + "grad_norm": 0.326171875, + "learning_rate": 0.0001724734394949628, + "loss": 0.8178, + "step": 1312 + }, + { + "epoch": 0.7305605786618445, + "grad_norm": 0.310546875, + "learning_rate": 0.00017243319956010587, + "loss": 0.7766, + "step": 1313 + }, + { + "epoch": 0.7311169842815413, + "grad_norm": 0.328125, + "learning_rate": 0.00017239293493579884, + "loss": 0.8458, + "step": 1314 + }, + { + "epoch": 0.731673389901238, + "grad_norm": 0.328125, + "learning_rate": 0.00017235264563576612, + "loss": 0.8981, + "step": 1315 + }, + { + "epoch": 0.7322297955209348, + "grad_norm": 0.310546875, + "learning_rate": 0.0001723123316737407, + "loss": 0.7909, + "step": 1316 + }, + { + "epoch": 0.7327862011406315, + "grad_norm": 0.33203125, + "learning_rate": 0.00017227199306346392, + "loss": 0.9362, + "step": 1317 + }, + { + "epoch": 0.7333426067603283, + "grad_norm": 0.326171875, + "learning_rate": 0.0001722316298186855, + "loss": 0.8683, + "step": 1318 + }, + { + "epoch": 0.7338990123800251, + "grad_norm": 0.333984375, + "learning_rate": 0.00017219124195316362, + "loss": 0.8612, + "step": 1319 + }, + { + "epoch": 0.7344554179997218, + "grad_norm": 0.322265625, + "learning_rate": 0.0001721508294806648, + "loss": 0.862, + "step": 1320 + }, + { + "epoch": 0.7350118236194185, + "grad_norm": 0.341796875, + "learning_rate": 0.00017211039241496394, + "loss": 0.8873, + "step": 1321 + }, + { + "epoch": 0.7355682292391154, + "grad_norm": 0.328125, + "learning_rate": 0.0001720699307698444, + "loss": 0.9028, + "step": 1322 + }, + { + "epoch": 0.7361246348588121, + "grad_norm": 0.3203125, + "learning_rate": 0.0001720294445590978, + "loss": 0.8096, + "step": 1323 + }, + { + "epoch": 0.7366810404785088, + "grad_norm": 0.328125, + "learning_rate": 0.00017198893379652424, + "loss": 0.8278, + "step": 1324 + }, + { + "epoch": 0.7372374460982056, + "grad_norm": 0.333984375, + "learning_rate": 0.00017194839849593214, + "loss": 0.8649, + "step": 1325 + }, + { + "epoch": 0.7377938517179023, + "grad_norm": 0.326171875, + "learning_rate": 0.00017190783867113827, + "loss": 0.8515, + "step": 1326 + }, + { + "epoch": 0.7383502573375991, + "grad_norm": 0.31640625, + "learning_rate": 0.00017186725433596779, + "loss": 0.7985, + "step": 1327 + }, + { + "epoch": 0.7389066629572959, + "grad_norm": 0.328125, + "learning_rate": 0.00017182664550425422, + "loss": 0.7828, + "step": 1328 + }, + { + "epoch": 0.7394630685769926, + "grad_norm": 0.326171875, + "learning_rate": 0.00017178601218983941, + "loss": 0.7932, + "step": 1329 + }, + { + "epoch": 0.7400194741966893, + "grad_norm": 0.341796875, + "learning_rate": 0.0001717453544065735, + "loss": 0.8869, + "step": 1330 + }, + { + "epoch": 0.7405758798163862, + "grad_norm": 0.33203125, + "learning_rate": 0.00017170467216831513, + "loss": 0.8317, + "step": 1331 + }, + { + "epoch": 0.7411322854360829, + "grad_norm": 0.3125, + "learning_rate": 0.00017166396548893106, + "loss": 0.7774, + "step": 1332 + }, + { + "epoch": 0.7416886910557796, + "grad_norm": 0.357421875, + "learning_rate": 0.00017162323438229662, + "loss": 0.8914, + "step": 1333 + }, + { + "epoch": 0.7422450966754764, + "grad_norm": 0.322265625, + "learning_rate": 0.00017158247886229523, + "loss": 0.8156, + "step": 1334 + }, + { + "epoch": 0.7428015022951732, + "grad_norm": 0.3359375, + "learning_rate": 0.0001715416989428188, + "loss": 0.829, + "step": 1335 + }, + { + "epoch": 0.7433579079148699, + "grad_norm": 0.345703125, + "learning_rate": 0.00017150089463776754, + "loss": 0.8551, + "step": 1336 + }, + { + "epoch": 0.7439143135345667, + "grad_norm": 0.328125, + "learning_rate": 0.00017146006596104982, + "loss": 0.8098, + "step": 1337 + }, + { + "epoch": 0.7444707191542634, + "grad_norm": 0.306640625, + "learning_rate": 0.00017141921292658253, + "loss": 0.702, + "step": 1338 + }, + { + "epoch": 0.7450271247739603, + "grad_norm": 0.314453125, + "learning_rate": 0.0001713783355482907, + "loss": 0.805, + "step": 1339 + }, + { + "epoch": 0.745583530393657, + "grad_norm": 0.31640625, + "learning_rate": 0.00017133743384010778, + "loss": 0.7702, + "step": 1340 + }, + { + "epoch": 0.7461399360133537, + "grad_norm": 0.330078125, + "learning_rate": 0.0001712965078159754, + "loss": 0.8625, + "step": 1341 + }, + { + "epoch": 0.7466963416330505, + "grad_norm": 0.3125, + "learning_rate": 0.00017125555748984353, + "loss": 0.7743, + "step": 1342 + }, + { + "epoch": 0.7472527472527473, + "grad_norm": 0.326171875, + "learning_rate": 0.00017121458287567044, + "loss": 0.8225, + "step": 1343 + }, + { + "epoch": 0.747809152872444, + "grad_norm": 0.353515625, + "learning_rate": 0.00017117358398742268, + "loss": 0.8579, + "step": 1344 + }, + { + "epoch": 0.7483655584921408, + "grad_norm": 0.328125, + "learning_rate": 0.00017113256083907505, + "loss": 0.8478, + "step": 1345 + }, + { + "epoch": 0.7489219641118375, + "grad_norm": 0.33203125, + "learning_rate": 0.00017109151344461062, + "loss": 0.8209, + "step": 1346 + }, + { + "epoch": 0.7494783697315343, + "grad_norm": 0.318359375, + "learning_rate": 0.0001710504418180207, + "loss": 0.7909, + "step": 1347 + }, + { + "epoch": 0.7500347753512311, + "grad_norm": 0.33203125, + "learning_rate": 0.00017100934597330495, + "loss": 0.8893, + "step": 1348 + }, + { + "epoch": 0.7505911809709278, + "grad_norm": 0.33203125, + "learning_rate": 0.0001709682259244712, + "loss": 0.7629, + "step": 1349 + }, + { + "epoch": 0.7511475865906245, + "grad_norm": 0.328125, + "learning_rate": 0.00017092708168553556, + "loss": 0.8391, + "step": 1350 + }, + { + "epoch": 0.7517039922103214, + "grad_norm": 0.333984375, + "learning_rate": 0.00017088591327052237, + "loss": 0.8319, + "step": 1351 + }, + { + "epoch": 0.7522603978300181, + "grad_norm": 0.33203125, + "learning_rate": 0.0001708447206934642, + "loss": 0.8358, + "step": 1352 + }, + { + "epoch": 0.7528168034497148, + "grad_norm": 0.326171875, + "learning_rate": 0.00017080350396840192, + "loss": 0.8513, + "step": 1353 + }, + { + "epoch": 0.7533732090694116, + "grad_norm": 0.326171875, + "learning_rate": 0.0001707622631093846, + "loss": 0.8373, + "step": 1354 + }, + { + "epoch": 0.7539296146891084, + "grad_norm": 0.3359375, + "learning_rate": 0.00017072099813046944, + "loss": 0.874, + "step": 1355 + }, + { + "epoch": 0.7544860203088051, + "grad_norm": 0.31640625, + "learning_rate": 0.00017067970904572206, + "loss": 0.7982, + "step": 1356 + }, + { + "epoch": 0.7550424259285019, + "grad_norm": 0.3359375, + "learning_rate": 0.0001706383958692161, + "loss": 0.8553, + "step": 1357 + }, + { + "epoch": 0.7555988315481986, + "grad_norm": 0.33203125, + "learning_rate": 0.00017059705861503351, + "loss": 0.8443, + "step": 1358 + }, + { + "epoch": 0.7561552371678953, + "grad_norm": 0.3203125, + "learning_rate": 0.0001705556972972645, + "loss": 0.8124, + "step": 1359 + }, + { + "epoch": 0.7567116427875922, + "grad_norm": 0.369140625, + "learning_rate": 0.0001705143119300073, + "loss": 0.9886, + "step": 1360 + }, + { + "epoch": 0.7572680484072889, + "grad_norm": 0.33203125, + "learning_rate": 0.00017047290252736853, + "loss": 0.8099, + "step": 1361 + }, + { + "epoch": 0.7578244540269857, + "grad_norm": 0.322265625, + "learning_rate": 0.00017043146910346294, + "loss": 0.767, + "step": 1362 + }, + { + "epoch": 0.7583808596466824, + "grad_norm": 0.33203125, + "learning_rate": 0.00017039001167241336, + "loss": 0.9191, + "step": 1363 + }, + { + "epoch": 0.7589372652663792, + "grad_norm": 0.333984375, + "learning_rate": 0.000170348530248351, + "loss": 0.8611, + "step": 1364 + }, + { + "epoch": 0.759493670886076, + "grad_norm": 0.306640625, + "learning_rate": 0.00017030702484541505, + "loss": 0.7678, + "step": 1365 + }, + { + "epoch": 0.7600500765057727, + "grad_norm": 0.3359375, + "learning_rate": 0.00017026549547775302, + "loss": 0.8689, + "step": 1366 + }, + { + "epoch": 0.7606064821254694, + "grad_norm": 0.333984375, + "learning_rate": 0.0001702239421595206, + "loss": 0.8887, + "step": 1367 + }, + { + "epoch": 0.7611628877451663, + "grad_norm": 0.3203125, + "learning_rate": 0.00017018236490488144, + "loss": 0.8028, + "step": 1368 + }, + { + "epoch": 0.761719293364863, + "grad_norm": 0.322265625, + "learning_rate": 0.00017014076372800761, + "loss": 0.8483, + "step": 1369 + }, + { + "epoch": 0.7622756989845597, + "grad_norm": 0.310546875, + "learning_rate": 0.0001700991386430791, + "loss": 0.776, + "step": 1370 + }, + { + "epoch": 0.7628321046042565, + "grad_norm": 0.330078125, + "learning_rate": 0.0001700574896642843, + "loss": 0.9116, + "step": 1371 + }, + { + "epoch": 0.7633885102239533, + "grad_norm": 0.3125, + "learning_rate": 0.00017001581680581944, + "loss": 0.7911, + "step": 1372 + }, + { + "epoch": 0.76394491584365, + "grad_norm": 0.328125, + "learning_rate": 0.00016997412008188918, + "loss": 0.828, + "step": 1373 + }, + { + "epoch": 0.7645013214633468, + "grad_norm": 0.30859375, + "learning_rate": 0.00016993239950670614, + "loss": 0.8273, + "step": 1374 + }, + { + "epoch": 0.7650577270830435, + "grad_norm": 0.3203125, + "learning_rate": 0.00016989065509449115, + "loss": 0.789, + "step": 1375 + }, + { + "epoch": 0.7656141327027403, + "grad_norm": 0.333984375, + "learning_rate": 0.00016984888685947312, + "loss": 0.8335, + "step": 1376 + }, + { + "epoch": 0.7661705383224371, + "grad_norm": 0.341796875, + "learning_rate": 0.0001698070948158891, + "loss": 0.9027, + "step": 1377 + }, + { + "epoch": 0.7667269439421338, + "grad_norm": 0.921875, + "learning_rate": 0.00016976527897798424, + "loss": 1.8603, + "step": 1378 + }, + { + "epoch": 0.7672833495618305, + "grad_norm": 0.330078125, + "learning_rate": 0.00016972343936001182, + "loss": 0.8506, + "step": 1379 + }, + { + "epoch": 0.7678397551815274, + "grad_norm": 0.326171875, + "learning_rate": 0.0001696815759762332, + "loss": 0.8211, + "step": 1380 + }, + { + "epoch": 0.7683961608012241, + "grad_norm": 0.326171875, + "learning_rate": 0.0001696396888409179, + "loss": 0.8601, + "step": 1381 + }, + { + "epoch": 0.7689525664209208, + "grad_norm": 0.33203125, + "learning_rate": 0.00016959777796834343, + "loss": 0.8136, + "step": 1382 + }, + { + "epoch": 0.7695089720406176, + "grad_norm": 0.322265625, + "learning_rate": 0.0001695558433727955, + "loss": 0.8032, + "step": 1383 + }, + { + "epoch": 0.7700653776603144, + "grad_norm": 0.35546875, + "learning_rate": 0.00016951388506856785, + "loss": 0.9557, + "step": 1384 + }, + { + "epoch": 0.7706217832800111, + "grad_norm": 0.359375, + "learning_rate": 0.00016947190306996235, + "loss": 0.9532, + "step": 1385 + }, + { + "epoch": 0.7711781888997079, + "grad_norm": 0.333984375, + "learning_rate": 0.00016942989739128885, + "loss": 0.9016, + "step": 1386 + }, + { + "epoch": 0.7717345945194046, + "grad_norm": 0.310546875, + "learning_rate": 0.0001693878680468653, + "loss": 0.7211, + "step": 1387 + }, + { + "epoch": 0.7722910001391015, + "grad_norm": 0.330078125, + "learning_rate": 0.00016934581505101786, + "loss": 0.8251, + "step": 1388 + }, + { + "epoch": 0.7728474057587982, + "grad_norm": 0.333984375, + "learning_rate": 0.00016930373841808055, + "loss": 0.9217, + "step": 1389 + }, + { + "epoch": 0.7734038113784949, + "grad_norm": 0.314453125, + "learning_rate": 0.00016926163816239557, + "loss": 0.798, + "step": 1390 + }, + { + "epoch": 0.7739602169981917, + "grad_norm": 0.361328125, + "learning_rate": 0.0001692195142983131, + "loss": 0.9148, + "step": 1391 + }, + { + "epoch": 0.7745166226178885, + "grad_norm": 0.32421875, + "learning_rate": 0.00016917736684019148, + "loss": 0.7434, + "step": 1392 + }, + { + "epoch": 0.7750730282375852, + "grad_norm": 0.318359375, + "learning_rate": 0.00016913519580239697, + "loss": 0.7526, + "step": 1393 + }, + { + "epoch": 0.775629433857282, + "grad_norm": 0.32421875, + "learning_rate": 0.00016909300119930388, + "loss": 0.7844, + "step": 1394 + }, + { + "epoch": 0.7761858394769787, + "grad_norm": 0.326171875, + "learning_rate": 0.00016905078304529464, + "loss": 0.8256, + "step": 1395 + }, + { + "epoch": 0.7767422450966754, + "grad_norm": 0.337890625, + "learning_rate": 0.00016900854135475966, + "loss": 0.8872, + "step": 1396 + }, + { + "epoch": 0.7772986507163723, + "grad_norm": 0.328125, + "learning_rate": 0.0001689662761420973, + "loss": 0.8541, + "step": 1397 + }, + { + "epoch": 0.777855056336069, + "grad_norm": 0.330078125, + "learning_rate": 0.00016892398742171408, + "loss": 0.8455, + "step": 1398 + }, + { + "epoch": 0.7784114619557657, + "grad_norm": 0.328125, + "learning_rate": 0.0001688816752080244, + "loss": 0.9086, + "step": 1399 + }, + { + "epoch": 0.7789678675754625, + "grad_norm": 0.310546875, + "learning_rate": 0.00016883933951545075, + "loss": 0.7954, + "step": 1400 + }, + { + "epoch": 0.7795242731951593, + "grad_norm": 0.353515625, + "learning_rate": 0.0001687969803584236, + "loss": 0.8604, + "step": 1401 + }, + { + "epoch": 0.780080678814856, + "grad_norm": 0.30859375, + "learning_rate": 0.00016875459775138142, + "loss": 0.7879, + "step": 1402 + }, + { + "epoch": 0.7806370844345528, + "grad_norm": 0.30859375, + "learning_rate": 0.00016871219170877066, + "loss": 0.6737, + "step": 1403 + }, + { + "epoch": 0.7811934900542495, + "grad_norm": 0.326171875, + "learning_rate": 0.00016866976224504576, + "loss": 0.8641, + "step": 1404 + }, + { + "epoch": 0.7817498956739463, + "grad_norm": 0.3125, + "learning_rate": 0.00016862730937466916, + "loss": 0.7594, + "step": 1405 + }, + { + "epoch": 0.7823063012936431, + "grad_norm": 0.306640625, + "learning_rate": 0.0001685848331121113, + "loss": 0.7111, + "step": 1406 + }, + { + "epoch": 0.7828627069133398, + "grad_norm": 0.326171875, + "learning_rate": 0.00016854233347185052, + "loss": 0.7913, + "step": 1407 + }, + { + "epoch": 0.7834191125330365, + "grad_norm": 0.330078125, + "learning_rate": 0.00016849981046837317, + "loss": 0.8784, + "step": 1408 + }, + { + "epoch": 0.7839755181527334, + "grad_norm": 0.31640625, + "learning_rate": 0.0001684572641161736, + "loss": 0.8471, + "step": 1409 + }, + { + "epoch": 0.7845319237724301, + "grad_norm": 0.322265625, + "learning_rate": 0.00016841469442975405, + "loss": 0.8297, + "step": 1410 + }, + { + "epoch": 0.7850883293921269, + "grad_norm": 0.3359375, + "learning_rate": 0.00016837210142362473, + "loss": 0.907, + "step": 1411 + }, + { + "epoch": 0.7856447350118236, + "grad_norm": 0.31640625, + "learning_rate": 0.00016832948511230386, + "loss": 0.7822, + "step": 1412 + }, + { + "epoch": 0.7862011406315204, + "grad_norm": 0.328125, + "learning_rate": 0.00016828684551031757, + "loss": 0.8785, + "step": 1413 + }, + { + "epoch": 0.7867575462512172, + "grad_norm": 0.33984375, + "learning_rate": 0.00016824418263219987, + "loss": 0.9351, + "step": 1414 + }, + { + "epoch": 0.7873139518709139, + "grad_norm": 0.33984375, + "learning_rate": 0.00016820149649249277, + "loss": 0.8879, + "step": 1415 + }, + { + "epoch": 0.7878703574906106, + "grad_norm": 0.33203125, + "learning_rate": 0.00016815878710574618, + "loss": 0.8392, + "step": 1416 + }, + { + "epoch": 0.7884267631103075, + "grad_norm": 0.328125, + "learning_rate": 0.000168116054486518, + "loss": 0.812, + "step": 1417 + }, + { + "epoch": 0.7889831687300042, + "grad_norm": 0.345703125, + "learning_rate": 0.00016807329864937392, + "loss": 0.8525, + "step": 1418 + }, + { + "epoch": 0.7895395743497009, + "grad_norm": 0.341796875, + "learning_rate": 0.00016803051960888766, + "loss": 0.9073, + "step": 1419 + }, + { + "epoch": 0.7900959799693977, + "grad_norm": 0.3203125, + "learning_rate": 0.00016798771737964077, + "loss": 0.7898, + "step": 1420 + }, + { + "epoch": 0.7906523855890945, + "grad_norm": 0.33203125, + "learning_rate": 0.00016794489197622279, + "loss": 0.8647, + "step": 1421 + }, + { + "epoch": 0.7912087912087912, + "grad_norm": 0.322265625, + "learning_rate": 0.0001679020434132311, + "loss": 0.8098, + "step": 1422 + }, + { + "epoch": 0.791765196828488, + "grad_norm": 0.33984375, + "learning_rate": 0.00016785917170527095, + "loss": 0.8732, + "step": 1423 + }, + { + "epoch": 0.7923216024481847, + "grad_norm": 0.353515625, + "learning_rate": 0.00016781627686695552, + "loss": 0.8421, + "step": 1424 + }, + { + "epoch": 0.7928780080678814, + "grad_norm": 0.3359375, + "learning_rate": 0.0001677733589129059, + "loss": 0.9202, + "step": 1425 + }, + { + "epoch": 0.7934344136875783, + "grad_norm": 0.337890625, + "learning_rate": 0.000167730417857751, + "loss": 0.8072, + "step": 1426 + }, + { + "epoch": 0.793990819307275, + "grad_norm": 0.33203125, + "learning_rate": 0.00016768745371612765, + "loss": 0.8307, + "step": 1427 + }, + { + "epoch": 0.7945472249269717, + "grad_norm": 0.326171875, + "learning_rate": 0.00016764446650268053, + "loss": 0.8375, + "step": 1428 + }, + { + "epoch": 0.7951036305466685, + "grad_norm": 0.9453125, + "learning_rate": 0.0001676014562320622, + "loss": 1.8024, + "step": 1429 + }, + { + "epoch": 0.7956600361663653, + "grad_norm": 0.30859375, + "learning_rate": 0.000167558422918933, + "loss": 0.6899, + "step": 1430 + }, + { + "epoch": 0.796216441786062, + "grad_norm": 0.328125, + "learning_rate": 0.00016751536657796123, + "loss": 0.8481, + "step": 1431 + }, + { + "epoch": 0.7967728474057588, + "grad_norm": 0.333984375, + "learning_rate": 0.00016747228722382298, + "loss": 0.8259, + "step": 1432 + }, + { + "epoch": 0.7973292530254555, + "grad_norm": 0.333984375, + "learning_rate": 0.00016742918487120222, + "loss": 0.8527, + "step": 1433 + }, + { + "epoch": 0.7978856586451523, + "grad_norm": 0.306640625, + "learning_rate": 0.00016738605953479075, + "loss": 0.6968, + "step": 1434 + }, + { + "epoch": 0.7984420642648491, + "grad_norm": 0.353515625, + "learning_rate": 0.00016734291122928816, + "loss": 0.9613, + "step": 1435 + }, + { + "epoch": 0.7989984698845458, + "grad_norm": 0.333984375, + "learning_rate": 0.00016729973996940193, + "loss": 0.8458, + "step": 1436 + }, + { + "epoch": 0.7995548755042426, + "grad_norm": 0.326171875, + "learning_rate": 0.0001672565457698473, + "loss": 0.8313, + "step": 1437 + }, + { + "epoch": 0.8001112811239394, + "grad_norm": 0.318359375, + "learning_rate": 0.00016721332864534744, + "loss": 0.7054, + "step": 1438 + }, + { + "epoch": 0.8006676867436361, + "grad_norm": 0.3359375, + "learning_rate": 0.00016717008861063322, + "loss": 0.9309, + "step": 1439 + }, + { + "epoch": 0.8012240923633329, + "grad_norm": 0.3046875, + "learning_rate": 0.00016712682568044334, + "loss": 0.7426, + "step": 1440 + }, + { + "epoch": 0.8017804979830296, + "grad_norm": 0.333984375, + "learning_rate": 0.00016708353986952433, + "loss": 0.8588, + "step": 1441 + }, + { + "epoch": 0.8023369036027264, + "grad_norm": 0.34765625, + "learning_rate": 0.00016704023119263058, + "loss": 0.9493, + "step": 1442 + }, + { + "epoch": 0.8028933092224232, + "grad_norm": 0.326171875, + "learning_rate": 0.0001669968996645241, + "loss": 0.838, + "step": 1443 + }, + { + "epoch": 0.8034497148421199, + "grad_norm": 0.310546875, + "learning_rate": 0.00016695354529997486, + "loss": 0.7359, + "step": 1444 + }, + { + "epoch": 0.8040061204618166, + "grad_norm": 0.353515625, + "learning_rate": 0.00016691016811376058, + "loss": 0.905, + "step": 1445 + }, + { + "epoch": 0.8045625260815135, + "grad_norm": 0.322265625, + "learning_rate": 0.00016686676812066668, + "loss": 0.8376, + "step": 1446 + }, + { + "epoch": 0.8051189317012102, + "grad_norm": 0.302734375, + "learning_rate": 0.00016682334533548642, + "loss": 0.7095, + "step": 1447 + }, + { + "epoch": 0.8056753373209069, + "grad_norm": 0.337890625, + "learning_rate": 0.00016677989977302078, + "loss": 0.8286, + "step": 1448 + }, + { + "epoch": 0.8062317429406037, + "grad_norm": 0.333984375, + "learning_rate": 0.0001667364314480786, + "loss": 0.8728, + "step": 1449 + }, + { + "epoch": 0.8067881485603005, + "grad_norm": 0.337890625, + "learning_rate": 0.00016669294037547635, + "loss": 0.8136, + "step": 1450 + }, + { + "epoch": 0.8073445541799972, + "grad_norm": 0.353515625, + "learning_rate": 0.00016664942657003838, + "loss": 0.8813, + "step": 1451 + }, + { + "epoch": 0.807900959799694, + "grad_norm": 0.310546875, + "learning_rate": 0.00016660589004659663, + "loss": 0.7414, + "step": 1452 + }, + { + "epoch": 0.8084573654193907, + "grad_norm": 0.3203125, + "learning_rate": 0.000166562330819991, + "loss": 0.8392, + "step": 1453 + }, + { + "epoch": 0.8090137710390874, + "grad_norm": 0.337890625, + "learning_rate": 0.00016651874890506897, + "loss": 0.8908, + "step": 1454 + }, + { + "epoch": 0.8095701766587843, + "grad_norm": 0.33984375, + "learning_rate": 0.0001664751443166858, + "loss": 0.8836, + "step": 1455 + }, + { + "epoch": 0.810126582278481, + "grad_norm": 0.328125, + "learning_rate": 0.0001664315170697044, + "loss": 0.8345, + "step": 1456 + }, + { + "epoch": 0.8106829878981777, + "grad_norm": 0.322265625, + "learning_rate": 0.00016638786717899551, + "loss": 0.7562, + "step": 1457 + }, + { + "epoch": 0.8112393935178746, + "grad_norm": 0.333984375, + "learning_rate": 0.0001663441946594376, + "loss": 0.8809, + "step": 1458 + }, + { + "epoch": 0.8117957991375713, + "grad_norm": 0.333984375, + "learning_rate": 0.0001663004995259168, + "loss": 0.8177, + "step": 1459 + }, + { + "epoch": 0.8123522047572681, + "grad_norm": 0.3203125, + "learning_rate": 0.00016625678179332692, + "loss": 0.8005, + "step": 1460 + }, + { + "epoch": 0.8129086103769648, + "grad_norm": 0.32421875, + "learning_rate": 0.00016621304147656948, + "loss": 0.8027, + "step": 1461 + }, + { + "epoch": 0.8134650159966615, + "grad_norm": 0.3203125, + "learning_rate": 0.0001661692785905538, + "loss": 0.8325, + "step": 1462 + }, + { + "epoch": 0.8140214216163584, + "grad_norm": 0.32421875, + "learning_rate": 0.0001661254931501968, + "loss": 0.7392, + "step": 1463 + }, + { + "epoch": 0.8145778272360551, + "grad_norm": 0.326171875, + "learning_rate": 0.00016608168517042305, + "loss": 0.819, + "step": 1464 + }, + { + "epoch": 0.8151342328557518, + "grad_norm": 0.32421875, + "learning_rate": 0.00016603785466616492, + "loss": 0.7979, + "step": 1465 + }, + { + "epoch": 0.8156906384754486, + "grad_norm": 0.3125, + "learning_rate": 0.0001659940016523624, + "loss": 0.7334, + "step": 1466 + }, + { + "epoch": 0.8162470440951454, + "grad_norm": 0.30859375, + "learning_rate": 0.0001659501261439631, + "loss": 0.6871, + "step": 1467 + }, + { + "epoch": 0.8168034497148421, + "grad_norm": 0.3359375, + "learning_rate": 0.00016590622815592236, + "loss": 0.902, + "step": 1468 + }, + { + "epoch": 0.8173598553345389, + "grad_norm": 0.337890625, + "learning_rate": 0.0001658623077032032, + "loss": 0.896, + "step": 1469 + }, + { + "epoch": 0.8179162609542356, + "grad_norm": 0.328125, + "learning_rate": 0.00016581836480077623, + "loss": 0.8791, + "step": 1470 + }, + { + "epoch": 0.8184726665739324, + "grad_norm": 0.431640625, + "learning_rate": 0.00016577439946361978, + "loss": 0.9967, + "step": 1471 + }, + { + "epoch": 0.8190290721936292, + "grad_norm": 0.328125, + "learning_rate": 0.0001657304117067198, + "loss": 0.8007, + "step": 1472 + }, + { + "epoch": 0.8195854778133259, + "grad_norm": 0.357421875, + "learning_rate": 0.00016568640154506983, + "loss": 0.9419, + "step": 1473 + }, + { + "epoch": 0.8201418834330226, + "grad_norm": 0.328125, + "learning_rate": 0.00016564236899367115, + "loss": 0.8375, + "step": 1474 + }, + { + "epoch": 0.8206982890527195, + "grad_norm": 0.322265625, + "learning_rate": 0.00016559831406753257, + "loss": 0.7735, + "step": 1475 + }, + { + "epoch": 0.8212546946724162, + "grad_norm": 0.3203125, + "learning_rate": 0.00016555423678167063, + "loss": 0.7933, + "step": 1476 + }, + { + "epoch": 0.8218111002921129, + "grad_norm": 0.318359375, + "learning_rate": 0.00016551013715110932, + "loss": 0.8085, + "step": 1477 + }, + { + "epoch": 0.8223675059118097, + "grad_norm": 0.3125, + "learning_rate": 0.00016546601519088053, + "loss": 0.7552, + "step": 1478 + }, + { + "epoch": 0.8229239115315065, + "grad_norm": 0.3203125, + "learning_rate": 0.00016542187091602345, + "loss": 0.7744, + "step": 1479 + }, + { + "epoch": 0.8234803171512032, + "grad_norm": 0.322265625, + "learning_rate": 0.0001653777043415851, + "loss": 0.8328, + "step": 1480 + }, + { + "epoch": 0.8240367227709, + "grad_norm": 0.349609375, + "learning_rate": 0.00016533351548261998, + "loss": 0.8444, + "step": 1481 + }, + { + "epoch": 0.8245931283905967, + "grad_norm": 0.337890625, + "learning_rate": 0.0001652893043541902, + "loss": 0.8248, + "step": 1482 + }, + { + "epoch": 0.8251495340102935, + "grad_norm": 0.330078125, + "learning_rate": 0.00016524507097136554, + "loss": 0.8688, + "step": 1483 + }, + { + "epoch": 0.8257059396299903, + "grad_norm": 0.322265625, + "learning_rate": 0.0001652008153492233, + "loss": 0.8503, + "step": 1484 + }, + { + "epoch": 0.826262345249687, + "grad_norm": 0.3125, + "learning_rate": 0.00016515653750284837, + "loss": 0.8137, + "step": 1485 + }, + { + "epoch": 0.8268187508693838, + "grad_norm": 0.306640625, + "learning_rate": 0.0001651122374473332, + "loss": 0.7943, + "step": 1486 + }, + { + "epoch": 0.8273751564890806, + "grad_norm": 0.318359375, + "learning_rate": 0.00016506791519777786, + "loss": 0.7787, + "step": 1487 + }, + { + "epoch": 0.8279315621087773, + "grad_norm": 0.333984375, + "learning_rate": 0.00016502357076928989, + "loss": 0.856, + "step": 1488 + }, + { + "epoch": 0.8284879677284741, + "grad_norm": 0.314453125, + "learning_rate": 0.0001649792041769845, + "loss": 0.7854, + "step": 1489 + }, + { + "epoch": 0.8290443733481708, + "grad_norm": 0.3203125, + "learning_rate": 0.00016493481543598443, + "loss": 0.7728, + "step": 1490 + }, + { + "epoch": 0.8296007789678675, + "grad_norm": 0.318359375, + "learning_rate": 0.00016489040456141988, + "loss": 0.7262, + "step": 1491 + }, + { + "epoch": 0.8301571845875644, + "grad_norm": 0.3125, + "learning_rate": 0.0001648459715684287, + "loss": 0.721, + "step": 1492 + }, + { + "epoch": 0.8307135902072611, + "grad_norm": 0.359375, + "learning_rate": 0.00016480151647215626, + "loss": 0.8713, + "step": 1493 + }, + { + "epoch": 0.8312699958269578, + "grad_norm": 0.3359375, + "learning_rate": 0.00016475703928775542, + "loss": 0.8769, + "step": 1494 + }, + { + "epoch": 0.8318264014466547, + "grad_norm": 0.326171875, + "learning_rate": 0.00016471254003038656, + "loss": 0.8117, + "step": 1495 + }, + { + "epoch": 0.8323828070663514, + "grad_norm": 0.326171875, + "learning_rate": 0.0001646680187152177, + "loss": 0.8064, + "step": 1496 + }, + { + "epoch": 0.8329392126860481, + "grad_norm": 0.3125, + "learning_rate": 0.00016462347535742419, + "loss": 0.783, + "step": 1497 + }, + { + "epoch": 0.8334956183057449, + "grad_norm": 0.33203125, + "learning_rate": 0.00016457890997218907, + "loss": 0.8462, + "step": 1498 + }, + { + "epoch": 0.8340520239254416, + "grad_norm": 0.32421875, + "learning_rate": 0.00016453432257470282, + "loss": 0.8125, + "step": 1499 + }, + { + "epoch": 0.8346084295451384, + "grad_norm": 0.3203125, + "learning_rate": 0.0001644897131801634, + "loss": 0.812, + "step": 1500 + }, + { + "epoch": 0.8351648351648352, + "grad_norm": 0.310546875, + "learning_rate": 0.00016444508180377628, + "loss": 0.722, + "step": 1501 + }, + { + "epoch": 0.8357212407845319, + "grad_norm": 0.318359375, + "learning_rate": 0.00016440042846075446, + "loss": 0.8175, + "step": 1502 + }, + { + "epoch": 0.8362776464042286, + "grad_norm": 0.33984375, + "learning_rate": 0.00016435575316631833, + "loss": 0.9485, + "step": 1503 + }, + { + "epoch": 0.8368340520239255, + "grad_norm": 0.3125, + "learning_rate": 0.0001643110559356959, + "loss": 0.7406, + "step": 1504 + }, + { + "epoch": 0.8373904576436222, + "grad_norm": 0.330078125, + "learning_rate": 0.00016426633678412263, + "loss": 0.7846, + "step": 1505 + }, + { + "epoch": 0.8379468632633189, + "grad_norm": 0.32421875, + "learning_rate": 0.00016422159572684127, + "loss": 0.8702, + "step": 1506 + }, + { + "epoch": 0.8385032688830157, + "grad_norm": 0.3359375, + "learning_rate": 0.00016417683277910228, + "loss": 0.8541, + "step": 1507 + }, + { + "epoch": 0.8390596745027125, + "grad_norm": 0.32421875, + "learning_rate": 0.00016413204795616345, + "loss": 0.8826, + "step": 1508 + }, + { + "epoch": 0.8396160801224093, + "grad_norm": 0.31640625, + "learning_rate": 0.00016408724127329005, + "loss": 0.7689, + "step": 1509 + }, + { + "epoch": 0.840172485742106, + "grad_norm": 0.337890625, + "learning_rate": 0.0001640424127457548, + "loss": 0.9144, + "step": 1510 + }, + { + "epoch": 0.8407288913618027, + "grad_norm": 0.33203125, + "learning_rate": 0.0001639975623888379, + "loss": 0.8451, + "step": 1511 + }, + { + "epoch": 0.8412852969814996, + "grad_norm": 0.33203125, + "learning_rate": 0.00016395269021782696, + "loss": 0.8054, + "step": 1512 + }, + { + "epoch": 0.8418417026011963, + "grad_norm": 0.33984375, + "learning_rate": 0.00016390779624801699, + "loss": 0.8637, + "step": 1513 + }, + { + "epoch": 0.842398108220893, + "grad_norm": 0.333984375, + "learning_rate": 0.00016386288049471048, + "loss": 0.8536, + "step": 1514 + }, + { + "epoch": 0.8429545138405898, + "grad_norm": 0.3359375, + "learning_rate": 0.00016381794297321735, + "loss": 0.8799, + "step": 1515 + }, + { + "epoch": 0.8435109194602866, + "grad_norm": 0.330078125, + "learning_rate": 0.0001637729836988549, + "loss": 0.814, + "step": 1516 + }, + { + "epoch": 0.8440673250799833, + "grad_norm": 0.31640625, + "learning_rate": 0.00016372800268694792, + "loss": 0.7759, + "step": 1517 + }, + { + "epoch": 0.8446237306996801, + "grad_norm": 0.33984375, + "learning_rate": 0.00016368299995282848, + "loss": 0.851, + "step": 1518 + }, + { + "epoch": 0.8451801363193768, + "grad_norm": 0.349609375, + "learning_rate": 0.0001636379755118362, + "loss": 0.8151, + "step": 1519 + }, + { + "epoch": 0.8457365419390735, + "grad_norm": 0.318359375, + "learning_rate": 0.00016359292937931797, + "loss": 0.7879, + "step": 1520 + }, + { + "epoch": 0.8462929475587704, + "grad_norm": 0.326171875, + "learning_rate": 0.00016354786157062818, + "loss": 0.825, + "step": 1521 + }, + { + "epoch": 0.8468493531784671, + "grad_norm": 0.31640625, + "learning_rate": 0.00016350277210112855, + "loss": 0.7414, + "step": 1522 + }, + { + "epoch": 0.8474057587981638, + "grad_norm": 0.31640625, + "learning_rate": 0.0001634576609861882, + "loss": 0.7679, + "step": 1523 + }, + { + "epoch": 0.8479621644178607, + "grad_norm": 0.322265625, + "learning_rate": 0.00016341252824118358, + "loss": 0.8039, + "step": 1524 + }, + { + "epoch": 0.8485185700375574, + "grad_norm": 0.306640625, + "learning_rate": 0.00016336737388149863, + "loss": 0.6974, + "step": 1525 + }, + { + "epoch": 0.8490749756572541, + "grad_norm": 0.33203125, + "learning_rate": 0.0001633221979225245, + "loss": 0.852, + "step": 1526 + }, + { + "epoch": 0.8496313812769509, + "grad_norm": 0.39453125, + "learning_rate": 0.00016327700037965987, + "loss": 0.9108, + "step": 1527 + }, + { + "epoch": 0.8501877868966476, + "grad_norm": 0.35546875, + "learning_rate": 0.00016323178126831066, + "loss": 0.8568, + "step": 1528 + }, + { + "epoch": 0.8507441925163444, + "grad_norm": 0.3359375, + "learning_rate": 0.00016318654060389016, + "loss": 0.8606, + "step": 1529 + }, + { + "epoch": 0.8513005981360412, + "grad_norm": 0.314453125, + "learning_rate": 0.00016314127840181902, + "loss": 0.7187, + "step": 1530 + }, + { + "epoch": 0.8518570037557379, + "grad_norm": 0.3203125, + "learning_rate": 0.00016309599467752522, + "loss": 0.8346, + "step": 1531 + }, + { + "epoch": 0.8524134093754346, + "grad_norm": 0.326171875, + "learning_rate": 0.00016305068944644417, + "loss": 0.9214, + "step": 1532 + }, + { + "epoch": 0.8529698149951315, + "grad_norm": 0.3203125, + "learning_rate": 0.00016300536272401844, + "loss": 0.7717, + "step": 1533 + }, + { + "epoch": 0.8535262206148282, + "grad_norm": 0.330078125, + "learning_rate": 0.00016296001452569808, + "loss": 0.8721, + "step": 1534 + }, + { + "epoch": 0.854082626234525, + "grad_norm": 0.3359375, + "learning_rate": 0.0001629146448669403, + "loss": 0.8538, + "step": 1535 + }, + { + "epoch": 0.8546390318542217, + "grad_norm": 0.345703125, + "learning_rate": 0.00016286925376320986, + "loss": 0.9144, + "step": 1536 + }, + { + "epoch": 0.8551954374739185, + "grad_norm": 0.3359375, + "learning_rate": 0.00016282384122997856, + "loss": 0.8563, + "step": 1537 + }, + { + "epoch": 0.8557518430936153, + "grad_norm": 0.32421875, + "learning_rate": 0.0001627784072827257, + "loss": 0.8297, + "step": 1538 + }, + { + "epoch": 0.856308248713312, + "grad_norm": 0.310546875, + "learning_rate": 0.0001627329519369378, + "loss": 0.7608, + "step": 1539 + }, + { + "epoch": 0.8568646543330087, + "grad_norm": 0.328125, + "learning_rate": 0.00016268747520810869, + "loss": 0.8589, + "step": 1540 + }, + { + "epoch": 0.8574210599527056, + "grad_norm": 0.318359375, + "learning_rate": 0.00016264197711173946, + "loss": 0.8031, + "step": 1541 + }, + { + "epoch": 0.8579774655724023, + "grad_norm": 0.310546875, + "learning_rate": 0.00016259645766333858, + "loss": 0.7716, + "step": 1542 + }, + { + "epoch": 0.858533871192099, + "grad_norm": 0.30859375, + "learning_rate": 0.00016255091687842165, + "loss": 0.7499, + "step": 1543 + }, + { + "epoch": 0.8590902768117958, + "grad_norm": 0.337890625, + "learning_rate": 0.00016250535477251168, + "loss": 0.8882, + "step": 1544 + }, + { + "epoch": 0.8596466824314926, + "grad_norm": 0.322265625, + "learning_rate": 0.00016245977136113883, + "loss": 0.7705, + "step": 1545 + }, + { + "epoch": 0.8602030880511893, + "grad_norm": 0.333984375, + "learning_rate": 0.00016241416665984063, + "loss": 0.8502, + "step": 1546 + }, + { + "epoch": 0.8607594936708861, + "grad_norm": 0.314453125, + "learning_rate": 0.00016236854068416178, + "loss": 0.774, + "step": 1547 + }, + { + "epoch": 0.8613158992905828, + "grad_norm": 0.349609375, + "learning_rate": 0.00016232289344965428, + "loss": 0.9005, + "step": 1548 + }, + { + "epoch": 0.8618723049102796, + "grad_norm": 0.333984375, + "learning_rate": 0.0001622772249718774, + "loss": 0.8641, + "step": 1549 + }, + { + "epoch": 0.8624287105299764, + "grad_norm": 0.349609375, + "learning_rate": 0.00016223153526639754, + "loss": 0.9873, + "step": 1550 + }, + { + "epoch": 0.8629851161496731, + "grad_norm": 0.333984375, + "learning_rate": 0.00016218582434878847, + "loss": 0.8727, + "step": 1551 + }, + { + "epoch": 0.8635415217693698, + "grad_norm": 0.3203125, + "learning_rate": 0.0001621400922346311, + "loss": 0.802, + "step": 1552 + }, + { + "epoch": 0.8640979273890667, + "grad_norm": 0.3203125, + "learning_rate": 0.00016209433893951366, + "loss": 0.8282, + "step": 1553 + }, + { + "epoch": 0.8646543330087634, + "grad_norm": 0.333984375, + "learning_rate": 0.00016204856447903145, + "loss": 0.9093, + "step": 1554 + }, + { + "epoch": 0.8652107386284601, + "grad_norm": 0.345703125, + "learning_rate": 0.0001620027688687871, + "loss": 0.9119, + "step": 1555 + }, + { + "epoch": 0.8657671442481569, + "grad_norm": 0.326171875, + "learning_rate": 0.00016195695212439044, + "loss": 0.8263, + "step": 1556 + }, + { + "epoch": 0.8663235498678536, + "grad_norm": 0.357421875, + "learning_rate": 0.00016191111426145846, + "loss": 0.9783, + "step": 1557 + }, + { + "epoch": 0.8668799554875504, + "grad_norm": 0.330078125, + "learning_rate": 0.00016186525529561537, + "loss": 0.8366, + "step": 1558 + }, + { + "epoch": 0.8674363611072472, + "grad_norm": 0.318359375, + "learning_rate": 0.00016181937524249263, + "loss": 0.7817, + "step": 1559 + }, + { + "epoch": 0.8679927667269439, + "grad_norm": 0.328125, + "learning_rate": 0.00016177347411772876, + "loss": 0.7799, + "step": 1560 + }, + { + "epoch": 0.8685491723466408, + "grad_norm": 0.33203125, + "learning_rate": 0.00016172755193696956, + "loss": 0.8473, + "step": 1561 + }, + { + "epoch": 0.8691055779663375, + "grad_norm": 0.33203125, + "learning_rate": 0.000161681608715868, + "loss": 0.8397, + "step": 1562 + }, + { + "epoch": 0.8696619835860342, + "grad_norm": 0.330078125, + "learning_rate": 0.00016163564447008416, + "loss": 0.8296, + "step": 1563 + }, + { + "epoch": 0.870218389205731, + "grad_norm": 0.32421875, + "learning_rate": 0.00016158965921528537, + "loss": 0.7982, + "step": 1564 + }, + { + "epoch": 0.8707747948254277, + "grad_norm": 0.310546875, + "learning_rate": 0.0001615436529671461, + "loss": 0.7385, + "step": 1565 + }, + { + "epoch": 0.8713312004451245, + "grad_norm": 0.330078125, + "learning_rate": 0.0001614976257413479, + "loss": 0.8126, + "step": 1566 + }, + { + "epoch": 0.8718876060648213, + "grad_norm": 0.306640625, + "learning_rate": 0.00016145157755357957, + "loss": 0.756, + "step": 1567 + }, + { + "epoch": 0.872444011684518, + "grad_norm": 0.326171875, + "learning_rate": 0.00016140550841953698, + "loss": 0.8072, + "step": 1568 + }, + { + "epoch": 0.8730004173042147, + "grad_norm": 0.3515625, + "learning_rate": 0.00016135941835492318, + "loss": 0.9268, + "step": 1569 + }, + { + "epoch": 0.8735568229239116, + "grad_norm": 0.330078125, + "learning_rate": 0.0001613133073754484, + "loss": 0.8406, + "step": 1570 + }, + { + "epoch": 0.8741132285436083, + "grad_norm": 0.341796875, + "learning_rate": 0.0001612671754968299, + "loss": 0.8901, + "step": 1571 + }, + { + "epoch": 0.874669634163305, + "grad_norm": 0.33984375, + "learning_rate": 0.00016122102273479207, + "loss": 0.8701, + "step": 1572 + }, + { + "epoch": 0.8752260397830018, + "grad_norm": 0.322265625, + "learning_rate": 0.00016117484910506652, + "loss": 0.7697, + "step": 1573 + }, + { + "epoch": 0.8757824454026986, + "grad_norm": 0.33984375, + "learning_rate": 0.0001611286546233919, + "loss": 0.9444, + "step": 1574 + }, + { + "epoch": 0.8763388510223953, + "grad_norm": 0.34375, + "learning_rate": 0.00016108243930551392, + "loss": 0.8555, + "step": 1575 + }, + { + "epoch": 0.8768952566420921, + "grad_norm": 0.3203125, + "learning_rate": 0.00016103620316718553, + "loss": 0.7459, + "step": 1576 + }, + { + "epoch": 0.8774516622617888, + "grad_norm": 0.322265625, + "learning_rate": 0.00016098994622416665, + "loss": 0.8395, + "step": 1577 + }, + { + "epoch": 0.8780080678814856, + "grad_norm": 0.3125, + "learning_rate": 0.00016094366849222437, + "loss": 0.799, + "step": 1578 + }, + { + "epoch": 0.8785644735011824, + "grad_norm": 0.337890625, + "learning_rate": 0.0001608973699871328, + "loss": 0.9022, + "step": 1579 + }, + { + "epoch": 0.8791208791208791, + "grad_norm": 0.32421875, + "learning_rate": 0.00016085105072467318, + "loss": 0.7485, + "step": 1580 + }, + { + "epoch": 0.8796772847405758, + "grad_norm": 0.330078125, + "learning_rate": 0.00016080471072063386, + "loss": 0.8096, + "step": 1581 + }, + { + "epoch": 0.8802336903602727, + "grad_norm": 0.328125, + "learning_rate": 0.00016075834999081012, + "loss": 0.8698, + "step": 1582 + }, + { + "epoch": 0.8807900959799694, + "grad_norm": 0.314453125, + "learning_rate": 0.00016071196855100444, + "loss": 0.7957, + "step": 1583 + }, + { + "epoch": 0.8813465015996662, + "grad_norm": 0.328125, + "learning_rate": 0.0001606655664170263, + "loss": 0.8163, + "step": 1584 + }, + { + "epoch": 0.8819029072193629, + "grad_norm": 0.328125, + "learning_rate": 0.00016061914360469227, + "loss": 0.8946, + "step": 1585 + }, + { + "epoch": 0.8824593128390597, + "grad_norm": 0.328125, + "learning_rate": 0.00016057270012982596, + "loss": 0.8084, + "step": 1586 + }, + { + "epoch": 0.8830157184587565, + "grad_norm": 0.341796875, + "learning_rate": 0.00016052623600825797, + "loss": 0.8569, + "step": 1587 + }, + { + "epoch": 0.8835721240784532, + "grad_norm": 0.326171875, + "learning_rate": 0.000160479751255826, + "loss": 0.771, + "step": 1588 + }, + { + "epoch": 0.8841285296981499, + "grad_norm": 0.3125, + "learning_rate": 0.00016043324588837474, + "loss": 0.7227, + "step": 1589 + }, + { + "epoch": 0.8846849353178468, + "grad_norm": 0.322265625, + "learning_rate": 0.00016038671992175594, + "loss": 0.8204, + "step": 1590 + }, + { + "epoch": 0.8852413409375435, + "grad_norm": 0.333984375, + "learning_rate": 0.00016034017337182834, + "loss": 0.9365, + "step": 1591 + }, + { + "epoch": 0.8857977465572402, + "grad_norm": 0.328125, + "learning_rate": 0.00016029360625445773, + "loss": 0.7505, + "step": 1592 + }, + { + "epoch": 0.886354152176937, + "grad_norm": 0.31640625, + "learning_rate": 0.00016024701858551694, + "loss": 0.8051, + "step": 1593 + }, + { + "epoch": 0.8869105577966337, + "grad_norm": 0.333984375, + "learning_rate": 0.00016020041038088568, + "loss": 0.8519, + "step": 1594 + }, + { + "epoch": 0.8874669634163305, + "grad_norm": 0.3359375, + "learning_rate": 0.00016015378165645076, + "loss": 0.9041, + "step": 1595 + }, + { + "epoch": 0.8880233690360273, + "grad_norm": 0.32421875, + "learning_rate": 0.000160107132428106, + "loss": 0.7955, + "step": 1596 + }, + { + "epoch": 0.888579774655724, + "grad_norm": 0.3125, + "learning_rate": 0.0001600604627117521, + "loss": 0.7547, + "step": 1597 + }, + { + "epoch": 0.8891361802754207, + "grad_norm": 0.33203125, + "learning_rate": 0.0001600137725232969, + "loss": 0.8384, + "step": 1598 + }, + { + "epoch": 0.8896925858951176, + "grad_norm": 0.322265625, + "learning_rate": 0.0001599670618786551, + "loss": 0.7619, + "step": 1599 + }, + { + "epoch": 0.8902489915148143, + "grad_norm": 0.328125, + "learning_rate": 0.00015992033079374838, + "loss": 0.8609, + "step": 1600 + }, + { + "epoch": 0.890805397134511, + "grad_norm": 0.34765625, + "learning_rate": 0.00015987357928450542, + "loss": 0.9125, + "step": 1601 + }, + { + "epoch": 0.8913618027542078, + "grad_norm": 0.333984375, + "learning_rate": 0.0001598268073668619, + "loss": 0.8257, + "step": 1602 + }, + { + "epoch": 0.8919182083739046, + "grad_norm": 0.32421875, + "learning_rate": 0.00015978001505676035, + "loss": 0.766, + "step": 1603 + }, + { + "epoch": 0.8924746139936013, + "grad_norm": 0.33984375, + "learning_rate": 0.0001597332023701504, + "loss": 0.8824, + "step": 1604 + }, + { + "epoch": 0.8930310196132981, + "grad_norm": 0.33203125, + "learning_rate": 0.00015968636932298843, + "loss": 0.8271, + "step": 1605 + }, + { + "epoch": 0.8935874252329948, + "grad_norm": 0.32421875, + "learning_rate": 0.00015963951593123797, + "loss": 0.8136, + "step": 1606 + }, + { + "epoch": 0.8941438308526916, + "grad_norm": 0.31640625, + "learning_rate": 0.0001595926422108693, + "loss": 0.7455, + "step": 1607 + }, + { + "epoch": 0.8947002364723884, + "grad_norm": 0.328125, + "learning_rate": 0.00015954574817785973, + "loss": 0.8478, + "step": 1608 + }, + { + "epoch": 0.8952566420920851, + "grad_norm": 0.3203125, + "learning_rate": 0.00015949883384819353, + "loss": 0.7687, + "step": 1609 + }, + { + "epoch": 0.8958130477117819, + "grad_norm": 0.3359375, + "learning_rate": 0.00015945189923786176, + "loss": 0.8229, + "step": 1610 + }, + { + "epoch": 0.8963694533314787, + "grad_norm": 0.3359375, + "learning_rate": 0.00015940494436286252, + "loss": 0.8361, + "step": 1611 + }, + { + "epoch": 0.8969258589511754, + "grad_norm": 0.333984375, + "learning_rate": 0.00015935796923920078, + "loss": 0.7942, + "step": 1612 + }, + { + "epoch": 0.8974822645708722, + "grad_norm": 0.337890625, + "learning_rate": 0.0001593109738828883, + "loss": 0.79, + "step": 1613 + }, + { + "epoch": 0.8980386701905689, + "grad_norm": 0.33203125, + "learning_rate": 0.00015926395830994394, + "loss": 0.8377, + "step": 1614 + }, + { + "epoch": 0.8985950758102657, + "grad_norm": 0.3203125, + "learning_rate": 0.00015921692253639328, + "loss": 0.796, + "step": 1615 + }, + { + "epoch": 0.8991514814299625, + "grad_norm": 0.330078125, + "learning_rate": 0.00015916986657826885, + "loss": 0.7916, + "step": 1616 + }, + { + "epoch": 0.8997078870496592, + "grad_norm": 0.3125, + "learning_rate": 0.0001591227904516101, + "loss": 0.7697, + "step": 1617 + }, + { + "epoch": 0.9002642926693559, + "grad_norm": 0.32421875, + "learning_rate": 0.00015907569417246333, + "loss": 0.8299, + "step": 1618 + }, + { + "epoch": 0.9008206982890528, + "grad_norm": 0.33984375, + "learning_rate": 0.0001590285777568816, + "loss": 0.7568, + "step": 1619 + }, + { + "epoch": 0.9013771039087495, + "grad_norm": 0.333984375, + "learning_rate": 0.00015898144122092505, + "loss": 0.8744, + "step": 1620 + }, + { + "epoch": 0.9019335095284462, + "grad_norm": 0.3125, + "learning_rate": 0.00015893428458066042, + "loss": 0.7819, + "step": 1621 + }, + { + "epoch": 0.902489915148143, + "grad_norm": 0.328125, + "learning_rate": 0.00015888710785216153, + "loss": 0.8183, + "step": 1622 + }, + { + "epoch": 0.9030463207678397, + "grad_norm": 0.322265625, + "learning_rate": 0.0001588399110515089, + "loss": 0.7904, + "step": 1623 + }, + { + "epoch": 0.9036027263875365, + "grad_norm": 0.3359375, + "learning_rate": 0.00015879269419479002, + "loss": 0.8461, + "step": 1624 + }, + { + "epoch": 0.9041591320072333, + "grad_norm": 0.33984375, + "learning_rate": 0.00015874545729809906, + "loss": 0.8309, + "step": 1625 + }, + { + "epoch": 0.90471553762693, + "grad_norm": 0.310546875, + "learning_rate": 0.00015869820037753716, + "loss": 0.6568, + "step": 1626 + }, + { + "epoch": 0.9052719432466267, + "grad_norm": 0.333984375, + "learning_rate": 0.0001586509234492122, + "loss": 0.8832, + "step": 1627 + }, + { + "epoch": 0.9058283488663236, + "grad_norm": 0.328125, + "learning_rate": 0.00015860362652923894, + "loss": 0.7721, + "step": 1628 + }, + { + "epoch": 0.9063847544860203, + "grad_norm": 0.328125, + "learning_rate": 0.0001585563096337389, + "loss": 0.8272, + "step": 1629 + }, + { + "epoch": 0.906941160105717, + "grad_norm": 0.3359375, + "learning_rate": 0.00015850897277884044, + "loss": 0.9139, + "step": 1630 + }, + { + "epoch": 0.9074975657254138, + "grad_norm": 0.322265625, + "learning_rate": 0.0001584616159806787, + "loss": 0.8265, + "step": 1631 + }, + { + "epoch": 0.9080539713451106, + "grad_norm": 0.3125, + "learning_rate": 0.00015841423925539564, + "loss": 0.7456, + "step": 1632 + }, + { + "epoch": 0.9086103769648074, + "grad_norm": 0.3125, + "learning_rate": 0.00015836684261914, + "loss": 0.7247, + "step": 1633 + }, + { + "epoch": 0.9091667825845041, + "grad_norm": 0.3203125, + "learning_rate": 0.00015831942608806734, + "loss": 0.7684, + "step": 1634 + }, + { + "epoch": 0.9097231882042008, + "grad_norm": 0.31640625, + "learning_rate": 0.00015827198967833993, + "loss": 0.7793, + "step": 1635 + }, + { + "epoch": 0.9102795938238977, + "grad_norm": 0.33984375, + "learning_rate": 0.00015822453340612693, + "loss": 0.8197, + "step": 1636 + }, + { + "epoch": 0.9108359994435944, + "grad_norm": 0.341796875, + "learning_rate": 0.0001581770572876041, + "loss": 0.8746, + "step": 1637 + }, + { + "epoch": 0.9113924050632911, + "grad_norm": 0.333984375, + "learning_rate": 0.00015812956133895412, + "loss": 0.8732, + "step": 1638 + }, + { + "epoch": 0.9119488106829879, + "grad_norm": 0.33203125, + "learning_rate": 0.00015808204557636642, + "loss": 0.8501, + "step": 1639 + }, + { + "epoch": 0.9125052163026847, + "grad_norm": 0.322265625, + "learning_rate": 0.00015803451001603706, + "loss": 0.7846, + "step": 1640 + }, + { + "epoch": 0.9130616219223814, + "grad_norm": 0.328125, + "learning_rate": 0.00015798695467416897, + "loss": 0.871, + "step": 1641 + }, + { + "epoch": 0.9136180275420782, + "grad_norm": 0.318359375, + "learning_rate": 0.00015793937956697175, + "loss": 0.7881, + "step": 1642 + }, + { + "epoch": 0.9141744331617749, + "grad_norm": 0.310546875, + "learning_rate": 0.00015789178471066178, + "loss": 0.7453, + "step": 1643 + }, + { + "epoch": 0.9147308387814717, + "grad_norm": 0.32421875, + "learning_rate": 0.00015784417012146217, + "loss": 0.8814, + "step": 1644 + }, + { + "epoch": 0.9152872444011685, + "grad_norm": 0.35546875, + "learning_rate": 0.0001577965358156027, + "loss": 0.8565, + "step": 1645 + }, + { + "epoch": 0.9158436500208652, + "grad_norm": 0.330078125, + "learning_rate": 0.00015774888180931993, + "loss": 0.8185, + "step": 1646 + }, + { + "epoch": 0.9164000556405619, + "grad_norm": 0.345703125, + "learning_rate": 0.0001577012081188572, + "loss": 0.8507, + "step": 1647 + }, + { + "epoch": 0.9169564612602588, + "grad_norm": 0.32421875, + "learning_rate": 0.00015765351476046435, + "loss": 0.721, + "step": 1648 + }, + { + "epoch": 0.9175128668799555, + "grad_norm": 0.333984375, + "learning_rate": 0.00015760580175039812, + "loss": 0.8975, + "step": 1649 + }, + { + "epoch": 0.9180692724996522, + "grad_norm": 0.306640625, + "learning_rate": 0.00015755806910492187, + "loss": 0.7393, + "step": 1650 + }, + { + "epoch": 0.918625678119349, + "grad_norm": 0.330078125, + "learning_rate": 0.00015751031684030568, + "loss": 0.8464, + "step": 1651 + }, + { + "epoch": 0.9191820837390458, + "grad_norm": 0.32421875, + "learning_rate": 0.0001574625449728263, + "loss": 0.7609, + "step": 1652 + }, + { + "epoch": 0.9197384893587425, + "grad_norm": 0.328125, + "learning_rate": 0.00015741475351876717, + "loss": 0.8405, + "step": 1653 + }, + { + "epoch": 0.9202948949784393, + "grad_norm": 0.337890625, + "learning_rate": 0.00015736694249441832, + "loss": 0.8624, + "step": 1654 + }, + { + "epoch": 0.920851300598136, + "grad_norm": 0.3203125, + "learning_rate": 0.00015731911191607661, + "loss": 0.7953, + "step": 1655 + }, + { + "epoch": 0.9214077062178327, + "grad_norm": 0.318359375, + "learning_rate": 0.0001572712618000455, + "loss": 0.7722, + "step": 1656 + }, + { + "epoch": 0.9219641118375296, + "grad_norm": 0.380859375, + "learning_rate": 0.00015722339216263503, + "loss": 0.8173, + "step": 1657 + }, + { + "epoch": 0.9225205174572263, + "grad_norm": 0.314453125, + "learning_rate": 0.00015717550302016197, + "loss": 0.7558, + "step": 1658 + }, + { + "epoch": 0.9230769230769231, + "grad_norm": 0.33203125, + "learning_rate": 0.00015712759438894978, + "loss": 0.83, + "step": 1659 + }, + { + "epoch": 0.9236333286966198, + "grad_norm": 0.326171875, + "learning_rate": 0.00015707966628532847, + "loss": 0.8345, + "step": 1660 + }, + { + "epoch": 0.9241897343163166, + "grad_norm": 0.322265625, + "learning_rate": 0.00015703171872563472, + "loss": 0.7731, + "step": 1661 + }, + { + "epoch": 0.9247461399360134, + "grad_norm": 0.306640625, + "learning_rate": 0.00015698375172621188, + "loss": 0.7117, + "step": 1662 + }, + { + "epoch": 0.9253025455557101, + "grad_norm": 0.318359375, + "learning_rate": 0.00015693576530340987, + "loss": 0.7655, + "step": 1663 + }, + { + "epoch": 0.9258589511754068, + "grad_norm": 0.3203125, + "learning_rate": 0.00015688775947358524, + "loss": 0.8557, + "step": 1664 + }, + { + "epoch": 0.9264153567951037, + "grad_norm": 0.30859375, + "learning_rate": 0.0001568397342531012, + "loss": 0.7155, + "step": 1665 + }, + { + "epoch": 0.9269717624148004, + "grad_norm": 0.3203125, + "learning_rate": 0.0001567916896583275, + "loss": 0.7699, + "step": 1666 + }, + { + "epoch": 0.9275281680344971, + "grad_norm": 0.34375, + "learning_rate": 0.00015674362570564063, + "loss": 0.9217, + "step": 1667 + }, + { + "epoch": 0.928084573654194, + "grad_norm": 0.330078125, + "learning_rate": 0.0001566955424114235, + "loss": 0.8761, + "step": 1668 + }, + { + "epoch": 0.9286409792738907, + "grad_norm": 0.330078125, + "learning_rate": 0.00015664743979206574, + "loss": 0.854, + "step": 1669 + }, + { + "epoch": 0.9291973848935874, + "grad_norm": 0.361328125, + "learning_rate": 0.00015659931786396345, + "loss": 0.9302, + "step": 1670 + }, + { + "epoch": 0.9297537905132842, + "grad_norm": 0.310546875, + "learning_rate": 0.00015655117664351945, + "loss": 0.6646, + "step": 1671 + }, + { + "epoch": 0.9303101961329809, + "grad_norm": 0.30078125, + "learning_rate": 0.00015650301614714303, + "loss": 0.6723, + "step": 1672 + }, + { + "epoch": 0.9308666017526777, + "grad_norm": 0.333984375, + "learning_rate": 0.00015645483639125014, + "loss": 0.8624, + "step": 1673 + }, + { + "epoch": 0.9314230073723745, + "grad_norm": 0.33984375, + "learning_rate": 0.0001564066373922632, + "loss": 0.8976, + "step": 1674 + }, + { + "epoch": 0.9319794129920712, + "grad_norm": 0.33203125, + "learning_rate": 0.00015635841916661127, + "loss": 0.8596, + "step": 1675 + }, + { + "epoch": 0.9325358186117679, + "grad_norm": 0.341796875, + "learning_rate": 0.00015631018173072988, + "loss": 0.9422, + "step": 1676 + }, + { + "epoch": 0.9330922242314648, + "grad_norm": 0.32421875, + "learning_rate": 0.00015626192510106123, + "loss": 0.7931, + "step": 1677 + }, + { + "epoch": 0.9336486298511615, + "grad_norm": 0.337890625, + "learning_rate": 0.0001562136492940539, + "loss": 0.8655, + "step": 1678 + }, + { + "epoch": 0.9342050354708582, + "grad_norm": 0.298828125, + "learning_rate": 0.00015616535432616316, + "loss": 0.7388, + "step": 1679 + }, + { + "epoch": 0.934761441090555, + "grad_norm": 0.3359375, + "learning_rate": 0.0001561170402138507, + "loss": 0.8805, + "step": 1680 + }, + { + "epoch": 0.9353178467102518, + "grad_norm": 0.318359375, + "learning_rate": 0.00015606870697358483, + "loss": 0.7576, + "step": 1681 + }, + { + "epoch": 0.9358742523299486, + "grad_norm": 0.33984375, + "learning_rate": 0.0001560203546218403, + "loss": 0.8602, + "step": 1682 + }, + { + "epoch": 0.9364306579496453, + "grad_norm": 0.330078125, + "learning_rate": 0.00015597198317509838, + "loss": 0.8555, + "step": 1683 + }, + { + "epoch": 0.936987063569342, + "grad_norm": 0.328125, + "learning_rate": 0.00015592359264984693, + "loss": 0.8234, + "step": 1684 + }, + { + "epoch": 0.9375434691890389, + "grad_norm": 0.326171875, + "learning_rate": 0.00015587518306258024, + "loss": 0.8587, + "step": 1685 + }, + { + "epoch": 0.9380998748087356, + "grad_norm": 0.330078125, + "learning_rate": 0.0001558267544297991, + "loss": 0.7904, + "step": 1686 + }, + { + "epoch": 0.9386562804284323, + "grad_norm": 0.33984375, + "learning_rate": 0.00015577830676801079, + "loss": 0.8887, + "step": 1687 + }, + { + "epoch": 0.9392126860481291, + "grad_norm": 0.341796875, + "learning_rate": 0.00015572984009372916, + "loss": 0.8488, + "step": 1688 + }, + { + "epoch": 0.9397690916678259, + "grad_norm": 0.41015625, + "learning_rate": 0.0001556813544234744, + "loss": 0.7988, + "step": 1689 + }, + { + "epoch": 0.9403254972875226, + "grad_norm": 0.314453125, + "learning_rate": 0.0001556328497737733, + "loss": 0.8036, + "step": 1690 + }, + { + "epoch": 0.9408819029072194, + "grad_norm": 0.330078125, + "learning_rate": 0.00015558432616115901, + "loss": 0.8355, + "step": 1691 + }, + { + "epoch": 0.9414383085269161, + "grad_norm": 0.326171875, + "learning_rate": 0.00015553578360217126, + "loss": 0.8544, + "step": 1692 + }, + { + "epoch": 0.9419947141466128, + "grad_norm": 0.341796875, + "learning_rate": 0.0001554872221133562, + "loss": 0.8712, + "step": 1693 + }, + { + "epoch": 0.9425511197663097, + "grad_norm": 0.306640625, + "learning_rate": 0.00015543864171126633, + "loss": 0.696, + "step": 1694 + }, + { + "epoch": 0.9431075253860064, + "grad_norm": 0.32421875, + "learning_rate": 0.0001553900424124607, + "loss": 0.8413, + "step": 1695 + }, + { + "epoch": 0.9436639310057031, + "grad_norm": 0.328125, + "learning_rate": 0.00015534142423350483, + "loss": 0.8712, + "step": 1696 + }, + { + "epoch": 0.9442203366254, + "grad_norm": 0.31640625, + "learning_rate": 0.00015529278719097057, + "loss": 0.8315, + "step": 1697 + }, + { + "epoch": 0.9447767422450967, + "grad_norm": 0.3203125, + "learning_rate": 0.00015524413130143632, + "loss": 0.8684, + "step": 1698 + }, + { + "epoch": 0.9453331478647934, + "grad_norm": 0.99609375, + "learning_rate": 0.00015519545658148674, + "loss": 1.8628, + "step": 1699 + }, + { + "epoch": 0.9458895534844902, + "grad_norm": 0.322265625, + "learning_rate": 0.0001551467630477131, + "loss": 0.8537, + "step": 1700 + }, + { + "epoch": 0.9464459591041869, + "grad_norm": 0.333984375, + "learning_rate": 0.00015509805071671292, + "loss": 0.8349, + "step": 1701 + }, + { + "epoch": 0.9470023647238837, + "grad_norm": 0.3203125, + "learning_rate": 0.0001550493196050903, + "loss": 0.7872, + "step": 1702 + }, + { + "epoch": 0.9475587703435805, + "grad_norm": 0.34375, + "learning_rate": 0.0001550005697294555, + "loss": 0.9317, + "step": 1703 + }, + { + "epoch": 0.9481151759632772, + "grad_norm": 0.337890625, + "learning_rate": 0.00015495180110642543, + "loss": 0.909, + "step": 1704 + }, + { + "epoch": 0.9486715815829739, + "grad_norm": 0.330078125, + "learning_rate": 0.0001549030137526232, + "loss": 0.8417, + "step": 1705 + }, + { + "epoch": 0.9492279872026708, + "grad_norm": 0.31640625, + "learning_rate": 0.00015485420768467843, + "loss": 0.7273, + "step": 1706 + }, + { + "epoch": 0.9497843928223675, + "grad_norm": 0.3203125, + "learning_rate": 0.00015480538291922703, + "loss": 0.8405, + "step": 1707 + }, + { + "epoch": 0.9503407984420643, + "grad_norm": 0.3515625, + "learning_rate": 0.00015475653947291136, + "loss": 0.9213, + "step": 1708 + }, + { + "epoch": 0.950897204061761, + "grad_norm": 0.326171875, + "learning_rate": 0.00015470767736238007, + "loss": 0.8166, + "step": 1709 + }, + { + "epoch": 0.9514536096814578, + "grad_norm": 0.3203125, + "learning_rate": 0.0001546587966042883, + "loss": 0.8303, + "step": 1710 + }, + { + "epoch": 0.9520100153011546, + "grad_norm": 0.3359375, + "learning_rate": 0.0001546098972152973, + "loss": 0.8412, + "step": 1711 + }, + { + "epoch": 0.9525664209208513, + "grad_norm": 0.310546875, + "learning_rate": 0.00015456097921207496, + "loss": 0.7055, + "step": 1712 + }, + { + "epoch": 0.953122826540548, + "grad_norm": 0.361328125, + "learning_rate": 0.00015451204261129534, + "loss": 0.925, + "step": 1713 + }, + { + "epoch": 0.9536792321602449, + "grad_norm": 0.353515625, + "learning_rate": 0.0001544630874296389, + "loss": 0.8671, + "step": 1714 + }, + { + "epoch": 0.9542356377799416, + "grad_norm": 0.326171875, + "learning_rate": 0.00015441411368379235, + "loss": 0.7963, + "step": 1715 + }, + { + "epoch": 0.9547920433996383, + "grad_norm": 0.3125, + "learning_rate": 0.00015436512139044886, + "loss": 0.7544, + "step": 1716 + }, + { + "epoch": 0.9553484490193351, + "grad_norm": 0.318359375, + "learning_rate": 0.00015431611056630784, + "loss": 0.7664, + "step": 1717 + }, + { + "epoch": 0.9559048546390319, + "grad_norm": 0.328125, + "learning_rate": 0.000154267081228075, + "loss": 0.8951, + "step": 1718 + }, + { + "epoch": 0.9564612602587286, + "grad_norm": 0.33203125, + "learning_rate": 0.00015421803339246245, + "loss": 0.8817, + "step": 1719 + }, + { + "epoch": 0.9570176658784254, + "grad_norm": 0.326171875, + "learning_rate": 0.00015416896707618852, + "loss": 0.8008, + "step": 1720 + }, + { + "epoch": 0.9575740714981221, + "grad_norm": 0.357421875, + "learning_rate": 0.00015411988229597782, + "loss": 0.9524, + "step": 1721 + }, + { + "epoch": 0.9581304771178188, + "grad_norm": 0.330078125, + "learning_rate": 0.00015407077906856137, + "loss": 0.7774, + "step": 1722 + }, + { + "epoch": 0.9586868827375157, + "grad_norm": 0.318359375, + "learning_rate": 0.0001540216574106764, + "loss": 0.7962, + "step": 1723 + }, + { + "epoch": 0.9592432883572124, + "grad_norm": 0.33984375, + "learning_rate": 0.00015397251733906636, + "loss": 0.8501, + "step": 1724 + }, + { + "epoch": 0.9597996939769091, + "grad_norm": 0.322265625, + "learning_rate": 0.00015392335887048114, + "loss": 0.7948, + "step": 1725 + }, + { + "epoch": 0.960356099596606, + "grad_norm": 0.33984375, + "learning_rate": 0.00015387418202167673, + "loss": 0.8632, + "step": 1726 + }, + { + "epoch": 0.9609125052163027, + "grad_norm": 0.35546875, + "learning_rate": 0.0001538249868094155, + "loss": 0.9704, + "step": 1727 + }, + { + "epoch": 0.9614689108359994, + "grad_norm": 0.30859375, + "learning_rate": 0.00015377577325046606, + "loss": 0.7435, + "step": 1728 + }, + { + "epoch": 0.9620253164556962, + "grad_norm": 0.330078125, + "learning_rate": 0.00015372654136160322, + "loss": 0.8626, + "step": 1729 + }, + { + "epoch": 0.9625817220753929, + "grad_norm": 0.30859375, + "learning_rate": 0.0001536772911596081, + "loss": 0.7785, + "step": 1730 + }, + { + "epoch": 0.9631381276950898, + "grad_norm": 0.318359375, + "learning_rate": 0.000153628022661268, + "loss": 0.7472, + "step": 1731 + }, + { + "epoch": 0.9636945333147865, + "grad_norm": 0.326171875, + "learning_rate": 0.0001535787358833765, + "loss": 0.854, + "step": 1732 + }, + { + "epoch": 0.9642509389344832, + "grad_norm": 0.330078125, + "learning_rate": 0.00015352943084273344, + "loss": 0.8813, + "step": 1733 + }, + { + "epoch": 0.96480734455418, + "grad_norm": 0.35546875, + "learning_rate": 0.0001534801075561448, + "loss": 0.9251, + "step": 1734 + }, + { + "epoch": 0.9653637501738768, + "grad_norm": 0.306640625, + "learning_rate": 0.00015343076604042285, + "loss": 0.7338, + "step": 1735 + }, + { + "epoch": 0.9659201557935735, + "grad_norm": 0.32421875, + "learning_rate": 0.00015338140631238604, + "loss": 0.761, + "step": 1736 + }, + { + "epoch": 0.9664765614132703, + "grad_norm": 0.33203125, + "learning_rate": 0.00015333202838885908, + "loss": 0.8448, + "step": 1737 + }, + { + "epoch": 0.967032967032967, + "grad_norm": 0.32421875, + "learning_rate": 0.0001532826322866728, + "loss": 0.7478, + "step": 1738 + }, + { + "epoch": 0.9675893726526638, + "grad_norm": 0.328125, + "learning_rate": 0.00015323321802266428, + "loss": 0.8116, + "step": 1739 + }, + { + "epoch": 0.9681457782723606, + "grad_norm": 0.32421875, + "learning_rate": 0.00015318378561367672, + "loss": 0.8851, + "step": 1740 + }, + { + "epoch": 0.9687021838920573, + "grad_norm": 0.333984375, + "learning_rate": 0.00015313433507655968, + "loss": 0.8524, + "step": 1741 + }, + { + "epoch": 0.969258589511754, + "grad_norm": 0.314453125, + "learning_rate": 0.00015308486642816869, + "loss": 0.7722, + "step": 1742 + }, + { + "epoch": 0.9698149951314509, + "grad_norm": 0.33984375, + "learning_rate": 0.00015303537968536557, + "loss": 0.9203, + "step": 1743 + }, + { + "epoch": 0.9703714007511476, + "grad_norm": 0.333984375, + "learning_rate": 0.00015298587486501828, + "loss": 0.8607, + "step": 1744 + }, + { + "epoch": 0.9709278063708443, + "grad_norm": 0.32421875, + "learning_rate": 0.00015293635198400095, + "loss": 0.7585, + "step": 1745 + }, + { + "epoch": 0.9714842119905411, + "grad_norm": 0.318359375, + "learning_rate": 0.00015288681105919385, + "loss": 0.7475, + "step": 1746 + }, + { + "epoch": 0.9720406176102379, + "grad_norm": 0.322265625, + "learning_rate": 0.0001528372521074834, + "loss": 0.7884, + "step": 1747 + }, + { + "epoch": 0.9725970232299346, + "grad_norm": 0.328125, + "learning_rate": 0.00015278767514576223, + "loss": 0.7938, + "step": 1748 + }, + { + "epoch": 0.9731534288496314, + "grad_norm": 0.326171875, + "learning_rate": 0.00015273808019092896, + "loss": 0.8201, + "step": 1749 + }, + { + "epoch": 0.9737098344693281, + "grad_norm": 0.3125, + "learning_rate": 0.00015268846725988855, + "loss": 0.7848, + "step": 1750 + }, + { + "epoch": 0.9742662400890248, + "grad_norm": 0.328125, + "learning_rate": 0.0001526388363695519, + "loss": 0.8647, + "step": 1751 + }, + { + "epoch": 0.9748226457087217, + "grad_norm": 0.34765625, + "learning_rate": 0.00015258918753683612, + "loss": 0.8837, + "step": 1752 + }, + { + "epoch": 0.9753790513284184, + "grad_norm": 0.302734375, + "learning_rate": 0.00015253952077866437, + "loss": 0.72, + "step": 1753 + }, + { + "epoch": 0.9759354569481151, + "grad_norm": 0.33984375, + "learning_rate": 0.0001524898361119661, + "loss": 0.9329, + "step": 1754 + }, + { + "epoch": 0.976491862567812, + "grad_norm": 0.30859375, + "learning_rate": 0.00015244013355367663, + "loss": 0.7853, + "step": 1755 + }, + { + "epoch": 0.9770482681875087, + "grad_norm": 0.333984375, + "learning_rate": 0.00015239041312073754, + "loss": 0.7884, + "step": 1756 + }, + { + "epoch": 0.9776046738072055, + "grad_norm": 0.328125, + "learning_rate": 0.0001523406748300964, + "loss": 0.8844, + "step": 1757 + }, + { + "epoch": 0.9781610794269022, + "grad_norm": 0.34375, + "learning_rate": 0.00015229091869870697, + "loss": 0.8389, + "step": 1758 + }, + { + "epoch": 0.978717485046599, + "grad_norm": 0.34375, + "learning_rate": 0.00015224114474352897, + "loss": 0.9325, + "step": 1759 + }, + { + "epoch": 0.9792738906662958, + "grad_norm": 0.3359375, + "learning_rate": 0.00015219135298152835, + "loss": 0.8748, + "step": 1760 + }, + { + "epoch": 0.9798302962859925, + "grad_norm": 0.3359375, + "learning_rate": 0.00015214154342967691, + "loss": 0.8733, + "step": 1761 + }, + { + "epoch": 0.9803867019056892, + "grad_norm": 0.326171875, + "learning_rate": 0.00015209171610495277, + "loss": 0.8178, + "step": 1762 + }, + { + "epoch": 0.980943107525386, + "grad_norm": 0.322265625, + "learning_rate": 0.0001520418710243399, + "loss": 0.7557, + "step": 1763 + }, + { + "epoch": 0.9814995131450828, + "grad_norm": 0.337890625, + "learning_rate": 0.0001519920082048285, + "loss": 0.8423, + "step": 1764 + }, + { + "epoch": 0.9820559187647795, + "grad_norm": 0.333984375, + "learning_rate": 0.00015194212766341458, + "loss": 0.8604, + "step": 1765 + }, + { + "epoch": 0.9826123243844763, + "grad_norm": 0.326171875, + "learning_rate": 0.00015189222941710043, + "loss": 0.849, + "step": 1766 + }, + { + "epoch": 0.983168730004173, + "grad_norm": 0.322265625, + "learning_rate": 0.0001518423134828943, + "loss": 0.8249, + "step": 1767 + }, + { + "epoch": 0.9837251356238698, + "grad_norm": 0.32421875, + "learning_rate": 0.00015179237987781037, + "loss": 0.8117, + "step": 1768 + }, + { + "epoch": 0.9842815412435666, + "grad_norm": 0.326171875, + "learning_rate": 0.00015174242861886892, + "loss": 0.7628, + "step": 1769 + }, + { + "epoch": 0.9848379468632633, + "grad_norm": 0.298828125, + "learning_rate": 0.00015169245972309633, + "loss": 0.7333, + "step": 1770 + }, + { + "epoch": 0.98539435248296, + "grad_norm": 0.333984375, + "learning_rate": 0.00015164247320752488, + "loss": 0.8816, + "step": 1771 + }, + { + "epoch": 0.9859507581026569, + "grad_norm": 0.3359375, + "learning_rate": 0.0001515924690891928, + "loss": 0.814, + "step": 1772 + }, + { + "epoch": 0.9865071637223536, + "grad_norm": 0.318359375, + "learning_rate": 0.00015154244738514448, + "loss": 0.7076, + "step": 1773 + }, + { + "epoch": 0.9870635693420503, + "grad_norm": 0.326171875, + "learning_rate": 0.00015149240811243022, + "loss": 0.8046, + "step": 1774 + }, + { + "epoch": 0.9876199749617471, + "grad_norm": 0.318359375, + "learning_rate": 0.0001514423512881063, + "loss": 0.7358, + "step": 1775 + }, + { + "epoch": 0.9881763805814439, + "grad_norm": 0.314453125, + "learning_rate": 0.000151392276929235, + "loss": 0.7509, + "step": 1776 + }, + { + "epoch": 0.9887327862011406, + "grad_norm": 0.3125, + "learning_rate": 0.00015134218505288458, + "loss": 0.719, + "step": 1777 + }, + { + "epoch": 0.9892891918208374, + "grad_norm": 0.365234375, + "learning_rate": 0.00015129207567612932, + "loss": 0.945, + "step": 1778 + }, + { + "epoch": 0.9898455974405341, + "grad_norm": 0.328125, + "learning_rate": 0.00015124194881604929, + "loss": 0.8144, + "step": 1779 + }, + { + "epoch": 0.990402003060231, + "grad_norm": 0.330078125, + "learning_rate": 0.00015119180448973073, + "loss": 0.7926, + "step": 1780 + }, + { + "epoch": 0.9909584086799277, + "grad_norm": 0.337890625, + "learning_rate": 0.0001511416427142658, + "loss": 0.9045, + "step": 1781 + }, + { + "epoch": 0.9915148142996244, + "grad_norm": 0.318359375, + "learning_rate": 0.00015109146350675236, + "loss": 0.7689, + "step": 1782 + }, + { + "epoch": 0.9920712199193212, + "grad_norm": 0.33203125, + "learning_rate": 0.00015104126688429462, + "loss": 0.8457, + "step": 1783 + }, + { + "epoch": 0.992627625539018, + "grad_norm": 0.341796875, + "learning_rate": 0.00015099105286400238, + "loss": 0.8154, + "step": 1784 + }, + { + "epoch": 0.9931840311587147, + "grad_norm": 0.326171875, + "learning_rate": 0.00015094082146299153, + "loss": 0.7393, + "step": 1785 + }, + { + "epoch": 0.9937404367784115, + "grad_norm": 0.306640625, + "learning_rate": 0.0001508905726983839, + "loss": 0.7191, + "step": 1786 + }, + { + "epoch": 0.9942968423981082, + "grad_norm": 0.314453125, + "learning_rate": 0.0001508403065873071, + "loss": 0.7611, + "step": 1787 + }, + { + "epoch": 0.994853248017805, + "grad_norm": 0.330078125, + "learning_rate": 0.00015079002314689484, + "loss": 0.7972, + "step": 1788 + }, + { + "epoch": 0.9954096536375018, + "grad_norm": 0.3203125, + "learning_rate": 0.0001507397223942866, + "loss": 0.7863, + "step": 1789 + }, + { + "epoch": 0.9959660592571985, + "grad_norm": 0.318359375, + "learning_rate": 0.00015068940434662776, + "loss": 0.801, + "step": 1790 + }, + { + "epoch": 0.9965224648768952, + "grad_norm": 0.3046875, + "learning_rate": 0.00015063906902106975, + "loss": 0.7198, + "step": 1791 + }, + { + "epoch": 0.997078870496592, + "grad_norm": 0.302734375, + "learning_rate": 0.0001505887164347697, + "loss": 0.6595, + "step": 1792 + }, + { + "epoch": 0.9976352761162888, + "grad_norm": 0.318359375, + "learning_rate": 0.00015053834660489073, + "loss": 0.8524, + "step": 1793 + }, + { + "epoch": 0.9981916817359855, + "grad_norm": 0.310546875, + "learning_rate": 0.00015048795954860172, + "loss": 0.8229, + "step": 1794 + }, + { + "epoch": 0.9987480873556823, + "grad_norm": 0.310546875, + "learning_rate": 0.00015043755528307767, + "loss": 0.8114, + "step": 1795 + }, + { + "epoch": 0.999304492975379, + "grad_norm": 0.328125, + "learning_rate": 0.0001503871338254992, + "loss": 0.868, + "step": 1796 + }, + { + "epoch": 0.9998608985950758, + "grad_norm": 0.3203125, + "learning_rate": 0.00015033669519305286, + "loss": 0.771, + "step": 1797 + }, + { + "epoch": 1.0004173042147726, + "grad_norm": 0.318359375, + "learning_rate": 0.0001502862394029311, + "loss": 0.7104, + "step": 1798 + }, + { + "epoch": 1.0004173042147726, + "eval_loss": 0.7584885954856873, + "eval_runtime": 1763.8079, + "eval_samples_per_second": 2.66, + "eval_steps_per_second": 1.33, + "step": 1798 + }, + { + "epoch": 1.0009737098344693, + "grad_norm": 0.330078125, + "learning_rate": 0.00015023576647233216, + "loss": 0.8586, + "step": 1799 + }, + { + "epoch": 1.001530115454166, + "grad_norm": 0.326171875, + "learning_rate": 0.00015018527641846022, + "loss": 0.7438, + "step": 1800 + }, + { + "epoch": 1.0020865210738628, + "grad_norm": 0.322265625, + "learning_rate": 0.0001501347692585252, + "loss": 0.8, + "step": 1801 + }, + { + "epoch": 1.0026429266935597, + "grad_norm": 0.314453125, + "learning_rate": 0.00015008424500974284, + "loss": 0.783, + "step": 1802 + }, + { + "epoch": 1.0031993323132564, + "grad_norm": 0.306640625, + "learning_rate": 0.00015003370368933477, + "loss": 0.7659, + "step": 1803 + }, + { + "epoch": 1.0037557379329531, + "grad_norm": 0.3125, + "learning_rate": 0.0001499831453145284, + "loss": 0.7565, + "step": 1804 + }, + { + "epoch": 1.0043121435526499, + "grad_norm": 0.34375, + "learning_rate": 0.00014993256990255703, + "loss": 0.8467, + "step": 1805 + }, + { + "epoch": 1.0048685491723466, + "grad_norm": 0.318359375, + "learning_rate": 0.0001498819774706596, + "loss": 0.8653, + "step": 1806 + }, + { + "epoch": 1.0054249547920433, + "grad_norm": 0.333984375, + "learning_rate": 0.00014983136803608103, + "loss": 0.8532, + "step": 1807 + }, + { + "epoch": 1.0059813604117402, + "grad_norm": 0.310546875, + "learning_rate": 0.00014978074161607196, + "loss": 0.8039, + "step": 1808 + }, + { + "epoch": 1.006537766031437, + "grad_norm": 0.34375, + "learning_rate": 0.00014973009822788876, + "loss": 0.8743, + "step": 1809 + }, + { + "epoch": 1.0070941716511337, + "grad_norm": 0.3203125, + "learning_rate": 0.00014967943788879372, + "loss": 0.8014, + "step": 1810 + }, + { + "epoch": 1.0076505772708304, + "grad_norm": 0.322265625, + "learning_rate": 0.0001496287606160548, + "loss": 0.7635, + "step": 1811 + }, + { + "epoch": 1.0082069828905271, + "grad_norm": 0.3125, + "learning_rate": 0.00014957806642694577, + "loss": 0.7415, + "step": 1812 + }, + { + "epoch": 1.0087633885102238, + "grad_norm": 0.330078125, + "learning_rate": 0.00014952735533874613, + "loss": 0.8067, + "step": 1813 + }, + { + "epoch": 1.0093197941299208, + "grad_norm": 0.32421875, + "learning_rate": 0.00014947662736874116, + "loss": 0.8203, + "step": 1814 + }, + { + "epoch": 1.0098761997496175, + "grad_norm": 0.32421875, + "learning_rate": 0.00014942588253422204, + "loss": 0.7599, + "step": 1815 + }, + { + "epoch": 1.0104326053693142, + "grad_norm": 0.31640625, + "learning_rate": 0.0001493751208524854, + "loss": 0.75, + "step": 1816 + }, + { + "epoch": 1.010989010989011, + "grad_norm": 0.3125, + "learning_rate": 0.00014932434234083384, + "loss": 0.7379, + "step": 1817 + }, + { + "epoch": 1.0115454166087077, + "grad_norm": 0.328125, + "learning_rate": 0.00014927354701657562, + "loss": 0.869, + "step": 1818 + }, + { + "epoch": 1.0121018222284046, + "grad_norm": 0.322265625, + "learning_rate": 0.00014922273489702473, + "loss": 0.8448, + "step": 1819 + }, + { + "epoch": 1.0126582278481013, + "grad_norm": 0.294921875, + "learning_rate": 0.00014917190599950098, + "loss": 0.6718, + "step": 1820 + }, + { + "epoch": 1.013214633467798, + "grad_norm": 0.3203125, + "learning_rate": 0.00014912106034132972, + "loss": 0.8731, + "step": 1821 + }, + { + "epoch": 1.0137710390874948, + "grad_norm": 0.31640625, + "learning_rate": 0.00014907019793984215, + "loss": 0.8285, + "step": 1822 + }, + { + "epoch": 1.0143274447071915, + "grad_norm": 0.318359375, + "learning_rate": 0.0001490193188123751, + "loss": 0.8005, + "step": 1823 + }, + { + "epoch": 1.0148838503268882, + "grad_norm": 0.322265625, + "learning_rate": 0.0001489684229762712, + "loss": 0.8463, + "step": 1824 + }, + { + "epoch": 1.0154402559465852, + "grad_norm": 0.333984375, + "learning_rate": 0.0001489175104488787, + "loss": 0.7568, + "step": 1825 + }, + { + "epoch": 1.0159966615662819, + "grad_norm": 0.32421875, + "learning_rate": 0.0001488665812475515, + "loss": 0.8092, + "step": 1826 + }, + { + "epoch": 1.0165530671859786, + "grad_norm": 0.314453125, + "learning_rate": 0.0001488156353896493, + "loss": 0.7904, + "step": 1827 + }, + { + "epoch": 1.0171094728056753, + "grad_norm": 0.31640625, + "learning_rate": 0.00014876467289253736, + "loss": 0.7924, + "step": 1828 + }, + { + "epoch": 1.0004173042147726, + "grad_norm": 0.337890625, + "learning_rate": 0.00014871369377358673, + "loss": 0.765, + "step": 1829 + }, + { + "epoch": 1.0009737098344693, + "grad_norm": 0.291015625, + "learning_rate": 0.000148662698050174, + "loss": 0.6763, + "step": 1830 + }, + { + "epoch": 1.001530115454166, + "grad_norm": 0.330078125, + "learning_rate": 0.0001486116857396815, + "loss": 0.7167, + "step": 1831 + }, + { + "epoch": 1.0020865210738628, + "grad_norm": 0.306640625, + "learning_rate": 0.0001485606568594972, + "loss": 0.7272, + "step": 1832 + }, + { + "epoch": 1.0026429266935597, + "grad_norm": 0.306640625, + "learning_rate": 0.0001485096114270147, + "loss": 0.646, + "step": 1833 + }, + { + "epoch": 1.0031993323132564, + "grad_norm": 0.326171875, + "learning_rate": 0.00014845854945963328, + "loss": 0.7645, + "step": 1834 + }, + { + "epoch": 1.0037557379329531, + "grad_norm": 0.318359375, + "learning_rate": 0.0001484074709747578, + "loss": 0.6898, + "step": 1835 + }, + { + "epoch": 1.0043121435526499, + "grad_norm": 0.314453125, + "learning_rate": 0.0001483563759897988, + "loss": 0.7485, + "step": 1836 + }, + { + "epoch": 1.0048685491723466, + "grad_norm": 0.32421875, + "learning_rate": 0.0001483052645221724, + "loss": 0.7517, + "step": 1837 + }, + { + "epoch": 1.0054249547920433, + "grad_norm": 0.330078125, + "learning_rate": 0.0001482541365893004, + "loss": 0.8237, + "step": 1838 + }, + { + "epoch": 1.0059813604117402, + "grad_norm": 0.322265625, + "learning_rate": 0.00014820299220861014, + "loss": 0.7087, + "step": 1839 + }, + { + "epoch": 1.006537766031437, + "grad_norm": 0.314453125, + "learning_rate": 0.00014815183139753462, + "loss": 0.7483, + "step": 1840 + }, + { + "epoch": 1.0070941716511337, + "grad_norm": 0.326171875, + "learning_rate": 0.00014810065417351245, + "loss": 0.7024, + "step": 1841 + }, + { + "epoch": 1.0076505772708304, + "grad_norm": 0.32421875, + "learning_rate": 0.00014804946055398777, + "loss": 0.747, + "step": 1842 + }, + { + "epoch": 1.0082069828905271, + "grad_norm": 0.322265625, + "learning_rate": 0.00014799825055641032, + "loss": 0.7064, + "step": 1843 + }, + { + "epoch": 1.0087633885102238, + "grad_norm": 0.34375, + "learning_rate": 0.0001479470241982355, + "loss": 0.7605, + "step": 1844 + }, + { + "epoch": 1.0093197941299208, + "grad_norm": 0.322265625, + "learning_rate": 0.00014789578149692424, + "loss": 0.7385, + "step": 1845 + }, + { + "epoch": 1.0098761997496175, + "grad_norm": 0.3203125, + "learning_rate": 0.000147844522469943, + "loss": 0.7465, + "step": 1846 + }, + { + "epoch": 1.0104326053693142, + "grad_norm": 0.314453125, + "learning_rate": 0.0001477932471347638, + "loss": 0.6345, + "step": 1847 + }, + { + "epoch": 1.010989010989011, + "grad_norm": 0.328125, + "learning_rate": 0.00014774195550886437, + "loss": 0.7599, + "step": 1848 + }, + { + "epoch": 1.0115454166087077, + "grad_norm": 0.330078125, + "learning_rate": 0.0001476906476097278, + "loss": 0.7338, + "step": 1849 + }, + { + "epoch": 1.0121018222284046, + "grad_norm": 0.310546875, + "learning_rate": 0.00014763932345484283, + "loss": 0.7355, + "step": 1850 + }, + { + "epoch": 1.0126582278481013, + "grad_norm": 0.314453125, + "learning_rate": 0.00014758798306170368, + "loss": 0.6857, + "step": 1851 + }, + { + "epoch": 1.013214633467798, + "grad_norm": 0.31640625, + "learning_rate": 0.00014753662644781022, + "loss": 0.7606, + "step": 1852 + }, + { + "epoch": 1.0137710390874948, + "grad_norm": 0.345703125, + "learning_rate": 0.00014748525363066773, + "loss": 0.7229, + "step": 1853 + }, + { + "epoch": 1.0143274447071915, + "grad_norm": 0.287109375, + "learning_rate": 0.00014743386462778702, + "loss": 0.6471, + "step": 1854 + }, + { + "epoch": 1.0148838503268882, + "grad_norm": 0.306640625, + "learning_rate": 0.00014738245945668455, + "loss": 0.6658, + "step": 1855 + }, + { + "epoch": 1.0154402559465852, + "grad_norm": 0.314453125, + "learning_rate": 0.00014733103813488208, + "loss": 0.6872, + "step": 1856 + }, + { + "epoch": 1.0159966615662819, + "grad_norm": 0.314453125, + "learning_rate": 0.0001472796006799071, + "loss": 0.7166, + "step": 1857 + }, + { + "epoch": 1.0165530671859786, + "grad_norm": 0.306640625, + "learning_rate": 0.0001472281471092924, + "loss": 0.6374, + "step": 1858 + }, + { + "epoch": 1.0171094728056753, + "grad_norm": 0.30859375, + "learning_rate": 0.00014717667744057643, + "loss": 0.6689, + "step": 1859 + }, + { + "epoch": 1.017665878425372, + "grad_norm": 0.328125, + "learning_rate": 0.00014712519169130298, + "loss": 0.7737, + "step": 1860 + }, + { + "epoch": 1.0182222840450688, + "grad_norm": 0.310546875, + "learning_rate": 0.0001470736898790214, + "loss": 0.6229, + "step": 1861 + }, + { + "epoch": 1.0187786896647657, + "grad_norm": 0.34375, + "learning_rate": 0.00014702217202128655, + "loss": 0.7961, + "step": 1862 + }, + { + "epoch": 1.0193350952844624, + "grad_norm": 0.345703125, + "learning_rate": 0.00014697063813565866, + "loss": 0.8566, + "step": 1863 + }, + { + "epoch": 1.0198915009041591, + "grad_norm": 0.298828125, + "learning_rate": 0.0001469190882397035, + "loss": 0.6613, + "step": 1864 + }, + { + "epoch": 1.0204479065238559, + "grad_norm": 0.30859375, + "learning_rate": 0.00014686752235099228, + "loss": 0.6908, + "step": 1865 + }, + { + "epoch": 1.0210043121435526, + "grad_norm": 0.34375, + "learning_rate": 0.00014681594048710165, + "loss": 0.7401, + "step": 1866 + }, + { + "epoch": 1.0215607177632493, + "grad_norm": 0.306640625, + "learning_rate": 0.00014676434266561373, + "loss": 0.6622, + "step": 1867 + }, + { + "epoch": 1.0221171233829462, + "grad_norm": 0.328125, + "learning_rate": 0.00014671272890411603, + "loss": 0.7471, + "step": 1868 + }, + { + "epoch": 1.022673529002643, + "grad_norm": 0.3125, + "learning_rate": 0.00014666109922020155, + "loss": 0.6932, + "step": 1869 + }, + { + "epoch": 1.0232299346223397, + "grad_norm": 0.314453125, + "learning_rate": 0.0001466094536314687, + "loss": 0.7359, + "step": 1870 + }, + { + "epoch": 1.0237863402420364, + "grad_norm": 0.3359375, + "learning_rate": 0.00014655779215552126, + "loss": 0.8022, + "step": 1871 + }, + { + "epoch": 1.0243427458617331, + "grad_norm": 0.296875, + "learning_rate": 0.00014650611480996844, + "loss": 0.6607, + "step": 1872 + }, + { + "epoch": 1.02489915148143, + "grad_norm": 0.357421875, + "learning_rate": 0.00014645442161242503, + "loss": 0.6856, + "step": 1873 + }, + { + "epoch": 1.0254555571011268, + "grad_norm": 0.31640625, + "learning_rate": 0.00014640271258051094, + "loss": 0.7469, + "step": 1874 + }, + { + "epoch": 1.0260119627208235, + "grad_norm": 0.32421875, + "learning_rate": 0.0001463509877318517, + "loss": 0.7641, + "step": 1875 + }, + { + "epoch": 1.0265683683405202, + "grad_norm": 0.31640625, + "learning_rate": 0.00014629924708407807, + "loss": 0.6705, + "step": 1876 + }, + { + "epoch": 1.027124773960217, + "grad_norm": 0.31640625, + "learning_rate": 0.00014624749065482635, + "loss": 0.6511, + "step": 1877 + }, + { + "epoch": 1.0276811795799137, + "grad_norm": 0.310546875, + "learning_rate": 0.0001461957184617381, + "loss": 0.7156, + "step": 1878 + }, + { + "epoch": 1.0282375851996106, + "grad_norm": 0.322265625, + "learning_rate": 0.00014614393052246033, + "loss": 0.7368, + "step": 1879 + }, + { + "epoch": 1.0287939908193073, + "grad_norm": 0.322265625, + "learning_rate": 0.00014609212685464531, + "loss": 0.7679, + "step": 1880 + }, + { + "epoch": 1.029350396439004, + "grad_norm": 0.330078125, + "learning_rate": 0.00014604030747595082, + "loss": 0.7735, + "step": 1881 + }, + { + "epoch": 1.0299068020587008, + "grad_norm": 0.314453125, + "learning_rate": 0.00014598847240403996, + "loss": 0.7116, + "step": 1882 + }, + { + "epoch": 1.0304632076783975, + "grad_norm": 0.306640625, + "learning_rate": 0.000145936621656581, + "loss": 0.6661, + "step": 1883 + }, + { + "epoch": 1.0310196132980942, + "grad_norm": 0.341796875, + "learning_rate": 0.0001458847552512478, + "loss": 0.8101, + "step": 1884 + }, + { + "epoch": 1.0315760189177912, + "grad_norm": 0.3125, + "learning_rate": 0.0001458328732057194, + "loss": 0.669, + "step": 1885 + }, + { + "epoch": 1.0321324245374879, + "grad_norm": 0.328125, + "learning_rate": 0.00014578097553768026, + "loss": 0.7535, + "step": 1886 + }, + { + "epoch": 1.0326888301571846, + "grad_norm": 0.3203125, + "learning_rate": 0.00014572906226482013, + "loss": 0.7153, + "step": 1887 + }, + { + "epoch": 1.0332452357768813, + "grad_norm": 0.314453125, + "learning_rate": 0.000145677133404834, + "loss": 0.7, + "step": 1888 + }, + { + "epoch": 1.033801641396578, + "grad_norm": 0.3125, + "learning_rate": 0.00014562518897542236, + "loss": 0.642, + "step": 1889 + }, + { + "epoch": 1.0343580470162748, + "grad_norm": 0.310546875, + "learning_rate": 0.0001455732289942908, + "loss": 0.695, + "step": 1890 + }, + { + "epoch": 1.0349144526359717, + "grad_norm": 0.3359375, + "learning_rate": 0.00014552125347915037, + "loss": 0.7717, + "step": 1891 + }, + { + "epoch": 1.0354708582556684, + "grad_norm": 0.30859375, + "learning_rate": 0.00014546926244771727, + "loss": 0.6881, + "step": 1892 + }, + { + "epoch": 1.0360272638753651, + "grad_norm": 0.3203125, + "learning_rate": 0.0001454172559177131, + "loss": 0.7123, + "step": 1893 + }, + { + "epoch": 1.0365836694950619, + "grad_norm": 0.314453125, + "learning_rate": 0.0001453652339068648, + "loss": 0.7429, + "step": 1894 + }, + { + "epoch": 1.0371400751147586, + "grad_norm": 0.341796875, + "learning_rate": 0.00014531319643290442, + "loss": 0.7497, + "step": 1895 + }, + { + "epoch": 1.0376964807344553, + "grad_norm": 0.330078125, + "learning_rate": 0.00014526114351356935, + "loss": 0.7219, + "step": 1896 + }, + { + "epoch": 1.0382528863541522, + "grad_norm": 0.306640625, + "learning_rate": 0.00014520907516660228, + "loss": 0.6278, + "step": 1897 + }, + { + "epoch": 1.038809291973849, + "grad_norm": 0.333984375, + "learning_rate": 0.00014515699140975113, + "loss": 0.7991, + "step": 1898 + }, + { + "epoch": 1.0393656975935457, + "grad_norm": 0.328125, + "learning_rate": 0.00014510489226076908, + "loss": 0.774, + "step": 1899 + }, + { + "epoch": 1.0399221032132424, + "grad_norm": 0.326171875, + "learning_rate": 0.00014505277773741454, + "loss": 0.7075, + "step": 1900 + }, + { + "epoch": 1.0404785088329391, + "grad_norm": 0.33984375, + "learning_rate": 0.00014500064785745117, + "loss": 0.7846, + "step": 1901 + }, + { + "epoch": 1.041034914452636, + "grad_norm": 0.318359375, + "learning_rate": 0.0001449485026386479, + "loss": 0.6826, + "step": 1902 + }, + { + "epoch": 1.0415913200723328, + "grad_norm": 0.275390625, + "learning_rate": 0.00014489634209877882, + "loss": 0.5808, + "step": 1903 + }, + { + "epoch": 1.0421477256920295, + "grad_norm": 0.33203125, + "learning_rate": 0.00014484416625562332, + "loss": 0.7508, + "step": 1904 + }, + { + "epoch": 1.0427041313117262, + "grad_norm": 0.32421875, + "learning_rate": 0.00014479197512696587, + "loss": 0.7421, + "step": 1905 + }, + { + "epoch": 1.043260536931423, + "grad_norm": 0.322265625, + "learning_rate": 0.00014473976873059636, + "loss": 0.6702, + "step": 1906 + }, + { + "epoch": 1.0438169425511197, + "grad_norm": 0.330078125, + "learning_rate": 0.0001446875470843097, + "loss": 0.7361, + "step": 1907 + }, + { + "epoch": 1.0443733481708166, + "grad_norm": 0.310546875, + "learning_rate": 0.0001446353102059061, + "loss": 0.6914, + "step": 1908 + }, + { + "epoch": 1.0449297537905133, + "grad_norm": 0.34375, + "learning_rate": 0.00014458305811319088, + "loss": 0.8452, + "step": 1909 + }, + { + "epoch": 1.04548615941021, + "grad_norm": 0.337890625, + "learning_rate": 0.00014453079082397464, + "loss": 0.7407, + "step": 1910 + }, + { + "epoch": 1.0460425650299068, + "grad_norm": 0.322265625, + "learning_rate": 0.0001444785083560731, + "loss": 0.686, + "step": 1911 + }, + { + "epoch": 1.0465989706496035, + "grad_norm": 0.33203125, + "learning_rate": 0.00014442621072730717, + "loss": 0.7235, + "step": 1912 + }, + { + "epoch": 1.0471553762693002, + "grad_norm": 0.32421875, + "learning_rate": 0.00014437389795550292, + "loss": 0.7048, + "step": 1913 + }, + { + "epoch": 1.0477117818889972, + "grad_norm": 0.322265625, + "learning_rate": 0.0001443215700584916, + "loss": 0.7442, + "step": 1914 + }, + { + "epoch": 1.0482681875086939, + "grad_norm": 0.33203125, + "learning_rate": 0.0001442692270541096, + "loss": 0.7133, + "step": 1915 + }, + { + "epoch": 1.0488245931283906, + "grad_norm": 0.33203125, + "learning_rate": 0.00014421686896019846, + "loss": 0.8122, + "step": 1916 + }, + { + "epoch": 1.0493809987480873, + "grad_norm": 0.330078125, + "learning_rate": 0.00014416449579460484, + "loss": 0.7727, + "step": 1917 + }, + { + "epoch": 1.049937404367784, + "grad_norm": 0.314453125, + "learning_rate": 0.00014411210757518062, + "loss": 0.6545, + "step": 1918 + }, + { + "epoch": 1.050493809987481, + "grad_norm": 0.3359375, + "learning_rate": 0.00014405970431978268, + "loss": 0.7588, + "step": 1919 + }, + { + "epoch": 1.0510502156071777, + "grad_norm": 0.318359375, + "learning_rate": 0.00014400728604627315, + "loss": 0.7321, + "step": 1920 + }, + { + "epoch": 1.0516066212268744, + "grad_norm": 0.310546875, + "learning_rate": 0.00014395485277251922, + "loss": 0.6813, + "step": 1921 + }, + { + "epoch": 1.0521630268465711, + "grad_norm": 0.31640625, + "learning_rate": 0.0001439024045163932, + "loss": 0.7022, + "step": 1922 + }, + { + "epoch": 1.0527194324662679, + "grad_norm": 0.318359375, + "learning_rate": 0.0001438499412957725, + "loss": 0.7135, + "step": 1923 + }, + { + "epoch": 1.0532758380859646, + "grad_norm": 0.328125, + "learning_rate": 0.00014379746312853967, + "loss": 0.6793, + "step": 1924 + }, + { + "epoch": 1.0538322437056615, + "grad_norm": 0.3203125, + "learning_rate": 0.00014374497003258232, + "loss": 0.6833, + "step": 1925 + }, + { + "epoch": 1.0543886493253583, + "grad_norm": 0.34765625, + "learning_rate": 0.00014369246202579306, + "loss": 0.8443, + "step": 1926 + }, + { + "epoch": 1.054945054945055, + "grad_norm": 0.3203125, + "learning_rate": 0.00014363993912606978, + "loss": 0.6652, + "step": 1927 + }, + { + "epoch": 1.0555014605647517, + "grad_norm": 0.322265625, + "learning_rate": 0.00014358740135131533, + "loss": 0.7908, + "step": 1928 + }, + { + "epoch": 1.0560578661844484, + "grad_norm": 0.326171875, + "learning_rate": 0.00014353484871943762, + "loss": 0.7398, + "step": 1929 + }, + { + "epoch": 1.0566142718041451, + "grad_norm": 0.322265625, + "learning_rate": 0.00014348228124834962, + "loss": 0.6743, + "step": 1930 + }, + { + "epoch": 1.057170677423842, + "grad_norm": 0.33203125, + "learning_rate": 0.00014342969895596942, + "loss": 0.7547, + "step": 1931 + }, + { + "epoch": 1.0577270830435388, + "grad_norm": 0.318359375, + "learning_rate": 0.00014337710186022013, + "loss": 0.7303, + "step": 1932 + }, + { + "epoch": 1.0582834886632355, + "grad_norm": 0.330078125, + "learning_rate": 0.00014332448997902988, + "loss": 0.7287, + "step": 1933 + }, + { + "epoch": 1.0588398942829322, + "grad_norm": 0.32421875, + "learning_rate": 0.00014327186333033186, + "loss": 0.7541, + "step": 1934 + }, + { + "epoch": 1.059396299902629, + "grad_norm": 0.333984375, + "learning_rate": 0.0001432192219320643, + "loss": 0.7516, + "step": 1935 + }, + { + "epoch": 1.0599527055223257, + "grad_norm": 0.326171875, + "learning_rate": 0.00014316656580217046, + "loss": 0.778, + "step": 1936 + }, + { + "epoch": 1.0605091111420226, + "grad_norm": 0.345703125, + "learning_rate": 0.00014311389495859863, + "loss": 0.7498, + "step": 1937 + }, + { + "epoch": 1.0610655167617193, + "grad_norm": 0.328125, + "learning_rate": 0.000143061209419302, + "loss": 0.7228, + "step": 1938 + }, + { + "epoch": 1.061621922381416, + "grad_norm": 0.328125, + "learning_rate": 0.00014300850920223902, + "loss": 0.7578, + "step": 1939 + }, + { + "epoch": 1.0621783280011128, + "grad_norm": 0.3125, + "learning_rate": 0.0001429557943253729, + "loss": 0.6644, + "step": 1940 + }, + { + "epoch": 1.0627347336208095, + "grad_norm": 0.3125, + "learning_rate": 0.00014290306480667196, + "loss": 0.7054, + "step": 1941 + }, + { + "epoch": 1.0632911392405062, + "grad_norm": 0.302734375, + "learning_rate": 0.00014285032066410947, + "loss": 0.6552, + "step": 1942 + }, + { + "epoch": 1.0638475448602032, + "grad_norm": 0.326171875, + "learning_rate": 0.00014279756191566372, + "loss": 0.6962, + "step": 1943 + }, + { + "epoch": 1.0644039504798999, + "grad_norm": 0.3125, + "learning_rate": 0.00014274478857931796, + "loss": 0.6713, + "step": 1944 + }, + { + "epoch": 1.0649603560995966, + "grad_norm": 0.31640625, + "learning_rate": 0.00014269200067306044, + "loss": 0.7131, + "step": 1945 + }, + { + "epoch": 1.0655167617192933, + "grad_norm": 0.326171875, + "learning_rate": 0.0001426391982148843, + "loss": 0.7393, + "step": 1946 + }, + { + "epoch": 1.06607316733899, + "grad_norm": 0.302734375, + "learning_rate": 0.0001425863812227877, + "loss": 0.6188, + "step": 1947 + }, + { + "epoch": 1.066629572958687, + "grad_norm": 0.3203125, + "learning_rate": 0.0001425335497147738, + "loss": 0.707, + "step": 1948 + }, + { + "epoch": 1.0671859785783837, + "grad_norm": 0.322265625, + "learning_rate": 0.0001424807037088506, + "loss": 0.6628, + "step": 1949 + }, + { + "epoch": 1.0677423841980804, + "grad_norm": 0.314453125, + "learning_rate": 0.0001424278432230311, + "loss": 0.644, + "step": 1950 + }, + { + "epoch": 1.0682987898177772, + "grad_norm": 0.318359375, + "learning_rate": 0.00014237496827533325, + "loss": 0.7098, + "step": 1951 + }, + { + "epoch": 1.0688551954374739, + "grad_norm": 0.31640625, + "learning_rate": 0.0001423220788837799, + "loss": 0.7083, + "step": 1952 + }, + { + "epoch": 1.0694116010571706, + "grad_norm": 0.306640625, + "learning_rate": 0.00014226917506639884, + "loss": 0.7144, + "step": 1953 + }, + { + "epoch": 1.0699680066768675, + "grad_norm": 0.341796875, + "learning_rate": 0.0001422162568412227, + "loss": 0.8027, + "step": 1954 + }, + { + "epoch": 1.0705244122965643, + "grad_norm": 0.30859375, + "learning_rate": 0.0001421633242262892, + "loss": 0.7031, + "step": 1955 + }, + { + "epoch": 1.071080817916261, + "grad_norm": 0.330078125, + "learning_rate": 0.00014211037723964082, + "loss": 0.7143, + "step": 1956 + }, + { + "epoch": 1.0716372235359577, + "grad_norm": 0.31640625, + "learning_rate": 0.00014205741589932495, + "loss": 0.6698, + "step": 1957 + }, + { + "epoch": 1.0721936291556544, + "grad_norm": 0.330078125, + "learning_rate": 0.0001420044402233939, + "loss": 0.6744, + "step": 1958 + }, + { + "epoch": 1.0727500347753511, + "grad_norm": 0.3125, + "learning_rate": 0.0001419514502299049, + "loss": 0.6424, + "step": 1959 + }, + { + "epoch": 1.073306440395048, + "grad_norm": 0.322265625, + "learning_rate": 0.00014189844593691996, + "loss": 0.7172, + "step": 1960 + }, + { + "epoch": 1.0738628460147448, + "grad_norm": 0.3203125, + "learning_rate": 0.0001418454273625061, + "loss": 0.6733, + "step": 1961 + }, + { + "epoch": 1.0744192516344415, + "grad_norm": 0.3203125, + "learning_rate": 0.0001417923945247351, + "loss": 0.6761, + "step": 1962 + }, + { + "epoch": 1.0749756572541382, + "grad_norm": 0.33984375, + "learning_rate": 0.00014173934744168365, + "loss": 0.7845, + "step": 1963 + }, + { + "epoch": 1.075532062873835, + "grad_norm": 0.326171875, + "learning_rate": 0.00014168628613143324, + "loss": 0.7739, + "step": 1964 + }, + { + "epoch": 1.076088468493532, + "grad_norm": 0.33984375, + "learning_rate": 0.00014163321061207032, + "loss": 0.7574, + "step": 1965 + }, + { + "epoch": 1.0766448741132286, + "grad_norm": 0.32421875, + "learning_rate": 0.00014158012090168605, + "loss": 0.7267, + "step": 1966 + }, + { + "epoch": 1.0772012797329253, + "grad_norm": 0.33203125, + "learning_rate": 0.00014152701701837652, + "loss": 0.7494, + "step": 1967 + }, + { + "epoch": 1.077757685352622, + "grad_norm": 0.310546875, + "learning_rate": 0.00014147389898024266, + "loss": 0.7303, + "step": 1968 + }, + { + "epoch": 1.0783140909723188, + "grad_norm": 0.3125, + "learning_rate": 0.00014142076680539014, + "loss": 0.6702, + "step": 1969 + }, + { + "epoch": 1.0788704965920155, + "grad_norm": 0.34765625, + "learning_rate": 0.00014136762051192954, + "loss": 0.8128, + "step": 1970 + }, + { + "epoch": 1.0794269022117122, + "grad_norm": 0.349609375, + "learning_rate": 0.00014131446011797613, + "loss": 0.8153, + "step": 1971 + }, + { + "epoch": 1.0799833078314092, + "grad_norm": 0.32421875, + "learning_rate": 0.00014126128564165017, + "loss": 0.7381, + "step": 1972 + }, + { + "epoch": 1.0805397134511059, + "grad_norm": 0.337890625, + "learning_rate": 0.00014120809710107656, + "loss": 0.7896, + "step": 1973 + }, + { + "epoch": 1.0810961190708026, + "grad_norm": 0.322265625, + "learning_rate": 0.00014115489451438506, + "loss": 0.7006, + "step": 1974 + }, + { + "epoch": 1.0816525246904993, + "grad_norm": 0.31640625, + "learning_rate": 0.0001411016778997102, + "loss": 0.6817, + "step": 1975 + }, + { + "epoch": 1.082208930310196, + "grad_norm": 0.337890625, + "learning_rate": 0.00014104844727519134, + "loss": 0.7037, + "step": 1976 + }, + { + "epoch": 1.082765335929893, + "grad_norm": 0.32421875, + "learning_rate": 0.0001409952026589725, + "loss": 0.6873, + "step": 1977 + }, + { + "epoch": 1.0833217415495897, + "grad_norm": 0.3125, + "learning_rate": 0.0001409419440692026, + "loss": 0.6717, + "step": 1978 + }, + { + "epoch": 1.0838781471692864, + "grad_norm": 0.341796875, + "learning_rate": 0.00014088867152403525, + "loss": 0.8306, + "step": 1979 + }, + { + "epoch": 1.0844345527889832, + "grad_norm": 0.341796875, + "learning_rate": 0.00014083538504162883, + "loss": 0.7378, + "step": 1980 + }, + { + "epoch": 1.0849909584086799, + "grad_norm": 0.3359375, + "learning_rate": 0.0001407820846401465, + "loss": 0.7589, + "step": 1981 + }, + { + "epoch": 1.0855473640283766, + "grad_norm": 0.330078125, + "learning_rate": 0.0001407287703377561, + "loss": 0.7173, + "step": 1982 + }, + { + "epoch": 1.0861037696480735, + "grad_norm": 0.33203125, + "learning_rate": 0.00014067544215263025, + "loss": 0.7238, + "step": 1983 + }, + { + "epoch": 1.0866601752677703, + "grad_norm": 0.333984375, + "learning_rate": 0.00014062210010294635, + "loss": 0.7191, + "step": 1984 + }, + { + "epoch": 1.087216580887467, + "grad_norm": 0.328125, + "learning_rate": 0.00014056874420688645, + "loss": 0.7664, + "step": 1985 + }, + { + "epoch": 1.0877729865071637, + "grad_norm": 0.3359375, + "learning_rate": 0.00014051537448263733, + "loss": 0.7795, + "step": 1986 + }, + { + "epoch": 1.0883293921268604, + "grad_norm": 0.32421875, + "learning_rate": 0.0001404619909483905, + "loss": 0.7233, + "step": 1987 + }, + { + "epoch": 1.0888857977465571, + "grad_norm": 0.326171875, + "learning_rate": 0.0001404085936223422, + "loss": 0.772, + "step": 1988 + }, + { + "epoch": 1.089442203366254, + "grad_norm": 0.337890625, + "learning_rate": 0.0001403551825226933, + "loss": 0.8023, + "step": 1989 + }, + { + "epoch": 1.0899986089859508, + "grad_norm": 0.3125, + "learning_rate": 0.00014030175766764944, + "loss": 0.6747, + "step": 1990 + }, + { + "epoch": 1.0905550146056475, + "grad_norm": 0.34375, + "learning_rate": 0.00014024831907542094, + "loss": 0.8103, + "step": 1991 + }, + { + "epoch": 1.0911114202253442, + "grad_norm": 0.30859375, + "learning_rate": 0.00014019486676422274, + "loss": 0.5994, + "step": 1992 + }, + { + "epoch": 1.091667825845041, + "grad_norm": 0.322265625, + "learning_rate": 0.00014014140075227453, + "loss": 0.7553, + "step": 1993 + }, + { + "epoch": 1.092224231464738, + "grad_norm": 0.314453125, + "learning_rate": 0.00014008792105780064, + "loss": 0.646, + "step": 1994 + }, + { + "epoch": 1.0927806370844346, + "grad_norm": 0.328125, + "learning_rate": 0.00014003442769903, + "loss": 0.7307, + "step": 1995 + }, + { + "epoch": 1.0933370427041313, + "grad_norm": 0.32421875, + "learning_rate": 0.00013998092069419636, + "loss": 0.7022, + "step": 1996 + }, + { + "epoch": 1.093893448323828, + "grad_norm": 0.333984375, + "learning_rate": 0.00013992740006153796, + "loss": 0.8088, + "step": 1997 + }, + { + "epoch": 1.0944498539435248, + "grad_norm": 0.31640625, + "learning_rate": 0.0001398738658192978, + "loss": 0.652, + "step": 1998 + }, + { + "epoch": 1.0950062595632215, + "grad_norm": 0.330078125, + "learning_rate": 0.00013982031798572336, + "loss": 0.756, + "step": 1999 + }, + { + "epoch": 1.0955626651829184, + "grad_norm": 0.322265625, + "learning_rate": 0.00013976675657906694, + "loss": 0.6802, + "step": 2000 + }, + { + "epoch": 1.0961190708026152, + "grad_norm": 0.3046875, + "learning_rate": 0.0001397131816175854, + "loss": 0.6098, + "step": 2001 + }, + { + "epoch": 1.096675476422312, + "grad_norm": 0.3125, + "learning_rate": 0.00013965959311954012, + "loss": 0.7077, + "step": 2002 + }, + { + "epoch": 1.0972318820420086, + "grad_norm": 0.328125, + "learning_rate": 0.00013960599110319726, + "loss": 0.7186, + "step": 2003 + }, + { + "epoch": 1.0977882876617053, + "grad_norm": 0.326171875, + "learning_rate": 0.0001395523755868275, + "loss": 0.6666, + "step": 2004 + }, + { + "epoch": 1.098344693281402, + "grad_norm": 0.322265625, + "learning_rate": 0.00013949874658870608, + "loss": 0.7174, + "step": 2005 + }, + { + "epoch": 1.098901098901099, + "grad_norm": 0.326171875, + "learning_rate": 0.00013944510412711294, + "loss": 0.7461, + "step": 2006 + }, + { + "epoch": 1.0994575045207957, + "grad_norm": 0.33203125, + "learning_rate": 0.00013939144822033253, + "loss": 0.701, + "step": 2007 + }, + { + "epoch": 1.1000139101404924, + "grad_norm": 0.306640625, + "learning_rate": 0.00013933777888665383, + "loss": 0.6507, + "step": 2008 + }, + { + "epoch": 1.1005703157601892, + "grad_norm": 0.33203125, + "learning_rate": 0.0001392840961443706, + "loss": 0.6878, + "step": 2009 + }, + { + "epoch": 1.1011267213798859, + "grad_norm": 0.337890625, + "learning_rate": 0.00013923040001178098, + "loss": 0.7909, + "step": 2010 + }, + { + "epoch": 1.1016831269995826, + "grad_norm": 0.326171875, + "learning_rate": 0.00013917669050718775, + "loss": 0.7602, + "step": 2011 + }, + { + "epoch": 1.1022395326192795, + "grad_norm": 0.33203125, + "learning_rate": 0.0001391229676488982, + "loss": 0.7296, + "step": 2012 + }, + { + "epoch": 1.1027959382389763, + "grad_norm": 0.314453125, + "learning_rate": 0.00013906923145522426, + "loss": 0.7202, + "step": 2013 + }, + { + "epoch": 1.103352343858673, + "grad_norm": 0.30859375, + "learning_rate": 0.0001390154819444823, + "loss": 0.6599, + "step": 2014 + }, + { + "epoch": 1.1039087494783697, + "grad_norm": 0.3125, + "learning_rate": 0.00013896171913499336, + "loss": 0.663, + "step": 2015 + }, + { + "epoch": 1.1044651550980664, + "grad_norm": 0.3125, + "learning_rate": 0.00013890794304508282, + "loss": 0.6925, + "step": 2016 + }, + { + "epoch": 1.1050215607177631, + "grad_norm": 0.3359375, + "learning_rate": 0.0001388541536930808, + "loss": 0.7559, + "step": 2017 + }, + { + "epoch": 1.10557796633746, + "grad_norm": 0.337890625, + "learning_rate": 0.0001388003510973218, + "loss": 0.8143, + "step": 2018 + }, + { + "epoch": 1.1061343719571568, + "grad_norm": 0.330078125, + "learning_rate": 0.00013874653527614492, + "loss": 0.6369, + "step": 2019 + }, + { + "epoch": 1.1066907775768535, + "grad_norm": 0.322265625, + "learning_rate": 0.00013869270624789366, + "loss": 0.6989, + "step": 2020 + }, + { + "epoch": 1.1072471831965502, + "grad_norm": 0.33984375, + "learning_rate": 0.0001386388640309161, + "loss": 0.7568, + "step": 2021 + }, + { + "epoch": 1.107803588816247, + "grad_norm": 0.322265625, + "learning_rate": 0.00013858500864356483, + "loss": 0.7158, + "step": 2022 + }, + { + "epoch": 1.108359994435944, + "grad_norm": 0.328125, + "learning_rate": 0.00013853114010419689, + "loss": 0.6169, + "step": 2023 + }, + { + "epoch": 1.1089164000556406, + "grad_norm": 1.09375, + "learning_rate": 0.00013847725843117375, + "loss": 1.592, + "step": 2024 + }, + { + "epoch": 1.1094728056753373, + "grad_norm": 0.328125, + "learning_rate": 0.00013842336364286155, + "loss": 0.7245, + "step": 2025 + }, + { + "epoch": 1.110029211295034, + "grad_norm": 0.31640625, + "learning_rate": 0.00013836945575763067, + "loss": 0.648, + "step": 2026 + }, + { + "epoch": 1.1105856169147308, + "grad_norm": 0.365234375, + "learning_rate": 0.0001383155347938561, + "loss": 0.8117, + "step": 2027 + }, + { + "epoch": 1.1111420225344275, + "grad_norm": 0.33203125, + "learning_rate": 0.0001382616007699172, + "loss": 0.7578, + "step": 2028 + }, + { + "epoch": 1.1116984281541245, + "grad_norm": 0.314453125, + "learning_rate": 0.00013820765370419787, + "loss": 0.6699, + "step": 2029 + }, + { + "epoch": 1.1122548337738212, + "grad_norm": 0.30859375, + "learning_rate": 0.00013815369361508636, + "loss": 0.6344, + "step": 2030 + }, + { + "epoch": 1.112811239393518, + "grad_norm": 0.33984375, + "learning_rate": 0.0001380997205209755, + "loss": 0.7591, + "step": 2031 + }, + { + "epoch": 1.1133676450132146, + "grad_norm": 0.32421875, + "learning_rate": 0.00013804573444026237, + "loss": 0.7, + "step": 2032 + }, + { + "epoch": 1.1139240506329113, + "grad_norm": 0.333984375, + "learning_rate": 0.0001379917353913486, + "loss": 0.7596, + "step": 2033 + }, + { + "epoch": 1.114480456252608, + "grad_norm": 0.328125, + "learning_rate": 0.0001379377233926402, + "loss": 0.7626, + "step": 2034 + }, + { + "epoch": 1.115036861872305, + "grad_norm": 0.35546875, + "learning_rate": 0.00013788369846254764, + "loss": 0.8207, + "step": 2035 + }, + { + "epoch": 1.1155932674920017, + "grad_norm": 0.345703125, + "learning_rate": 0.0001378296606194857, + "loss": 0.8149, + "step": 2036 + }, + { + "epoch": 1.1161496731116984, + "grad_norm": 0.328125, + "learning_rate": 0.00013777560988187363, + "loss": 0.7041, + "step": 2037 + }, + { + "epoch": 1.1167060787313952, + "grad_norm": 0.33984375, + "learning_rate": 0.00013772154626813513, + "loss": 0.7434, + "step": 2038 + }, + { + "epoch": 1.1172624843510919, + "grad_norm": 0.328125, + "learning_rate": 0.0001376674697966982, + "loss": 0.7522, + "step": 2039 + }, + { + "epoch": 1.1178188899707888, + "grad_norm": 0.34375, + "learning_rate": 0.00013761338048599525, + "loss": 0.771, + "step": 2040 + }, + { + "epoch": 1.1183752955904855, + "grad_norm": 0.330078125, + "learning_rate": 0.00013755927835446298, + "loss": 0.761, + "step": 2041 + }, + { + "epoch": 1.1189317012101823, + "grad_norm": 0.3359375, + "learning_rate": 0.00013750516342054265, + "loss": 0.7439, + "step": 2042 + }, + { + "epoch": 1.119488106829879, + "grad_norm": 0.33984375, + "learning_rate": 0.00013745103570267977, + "loss": 0.7558, + "step": 2043 + }, + { + "epoch": 1.1200445124495757, + "grad_norm": 0.330078125, + "learning_rate": 0.00013739689521932418, + "loss": 0.7231, + "step": 2044 + }, + { + "epoch": 1.1206009180692724, + "grad_norm": 0.3359375, + "learning_rate": 0.00013734274198893008, + "loss": 0.7869, + "step": 2045 + }, + { + "epoch": 1.1211573236889691, + "grad_norm": 0.3203125, + "learning_rate": 0.00013728857602995613, + "loss": 0.6867, + "step": 2046 + }, + { + "epoch": 1.121713729308666, + "grad_norm": 0.333984375, + "learning_rate": 0.0001372343973608652, + "loss": 0.744, + "step": 2047 + }, + { + "epoch": 1.1222701349283628, + "grad_norm": 0.337890625, + "learning_rate": 0.00013718020600012452, + "loss": 0.7727, + "step": 2048 + }, + { + "epoch": 1.1228265405480595, + "grad_norm": 0.310546875, + "learning_rate": 0.00013712600196620566, + "loss": 0.6544, + "step": 2049 + }, + { + "epoch": 1.1233829461677562, + "grad_norm": 0.328125, + "learning_rate": 0.00013707178527758452, + "loss": 0.725, + "step": 2050 + }, + { + "epoch": 1.123939351787453, + "grad_norm": 0.298828125, + "learning_rate": 0.0001370175559527413, + "loss": 0.5613, + "step": 2051 + }, + { + "epoch": 1.12449575740715, + "grad_norm": 0.34375, + "learning_rate": 0.00013696331401016052, + "loss": 0.8063, + "step": 2052 + }, + { + "epoch": 1.1250521630268466, + "grad_norm": 0.310546875, + "learning_rate": 0.00013690905946833098, + "loss": 0.6506, + "step": 2053 + }, + { + "epoch": 1.1256085686465434, + "grad_norm": 0.328125, + "learning_rate": 0.00013685479234574577, + "loss": 0.7368, + "step": 2054 + }, + { + "epoch": 1.12616497426624, + "grad_norm": 0.310546875, + "learning_rate": 0.0001368005126609023, + "loss": 0.7023, + "step": 2055 + }, + { + "epoch": 1.1267213798859368, + "grad_norm": 0.32421875, + "learning_rate": 0.00013674622043230228, + "loss": 0.7071, + "step": 2056 + }, + { + "epoch": 1.1272777855056335, + "grad_norm": 0.310546875, + "learning_rate": 0.00013669191567845158, + "loss": 0.6503, + "step": 2057 + }, + { + "epoch": 1.1278341911253305, + "grad_norm": 0.330078125, + "learning_rate": 0.0001366375984178605, + "loss": 0.6353, + "step": 2058 + }, + { + "epoch": 1.1283905967450272, + "grad_norm": 0.337890625, + "learning_rate": 0.0001365832686690435, + "loss": 0.7651, + "step": 2059 + }, + { + "epoch": 1.128947002364724, + "grad_norm": 0.30859375, + "learning_rate": 0.0001365289264505193, + "loss": 0.6411, + "step": 2060 + }, + { + "epoch": 1.1295034079844206, + "grad_norm": 0.291015625, + "learning_rate": 0.0001364745717808109, + "loss": 0.593, + "step": 2061 + }, + { + "epoch": 1.1300598136041173, + "grad_norm": 0.310546875, + "learning_rate": 0.00013642020467844555, + "loss": 0.657, + "step": 2062 + }, + { + "epoch": 1.130616219223814, + "grad_norm": 0.326171875, + "learning_rate": 0.00013636582516195472, + "loss": 0.704, + "step": 2063 + }, + { + "epoch": 1.131172624843511, + "grad_norm": 0.349609375, + "learning_rate": 0.0001363114332498741, + "loss": 0.8242, + "step": 2064 + }, + { + "epoch": 1.1317290304632077, + "grad_norm": 0.333984375, + "learning_rate": 0.00013625702896074361, + "loss": 0.749, + "step": 2065 + }, + { + "epoch": 1.1322854360829044, + "grad_norm": 0.341796875, + "learning_rate": 0.00013620261231310743, + "loss": 0.7413, + "step": 2066 + }, + { + "epoch": 1.1328418417026012, + "grad_norm": 0.337890625, + "learning_rate": 0.0001361481833255139, + "loss": 0.7288, + "step": 2067 + }, + { + "epoch": 1.1333982473222979, + "grad_norm": 0.333984375, + "learning_rate": 0.0001360937420165156, + "loss": 0.7728, + "step": 2068 + }, + { + "epoch": 1.1339546529419948, + "grad_norm": 0.34375, + "learning_rate": 0.00013603928840466928, + "loss": 0.7667, + "step": 2069 + }, + { + "epoch": 1.1345110585616915, + "grad_norm": 0.337890625, + "learning_rate": 0.00013598482250853588, + "loss": 0.7391, + "step": 2070 + }, + { + "epoch": 1.1350674641813883, + "grad_norm": 0.314453125, + "learning_rate": 0.00013593034434668063, + "loss": 0.6652, + "step": 2071 + }, + { + "epoch": 1.135623869801085, + "grad_norm": 0.330078125, + "learning_rate": 0.00013587585393767278, + "loss": 0.7477, + "step": 2072 + }, + { + "epoch": 1.1361802754207817, + "grad_norm": 0.330078125, + "learning_rate": 0.00013582135130008584, + "loss": 0.7554, + "step": 2073 + }, + { + "epoch": 1.1367366810404784, + "grad_norm": 0.34765625, + "learning_rate": 0.0001357668364524975, + "loss": 0.8427, + "step": 2074 + }, + { + "epoch": 1.1372930866601751, + "grad_norm": 0.330078125, + "learning_rate": 0.0001357123094134896, + "loss": 0.7391, + "step": 2075 + }, + { + "epoch": 1.137849492279872, + "grad_norm": 0.32421875, + "learning_rate": 0.00013565777020164814, + "loss": 0.7641, + "step": 2076 + }, + { + "epoch": 1.1384058978995688, + "grad_norm": 0.3125, + "learning_rate": 0.0001356032188355632, + "loss": 0.6439, + "step": 2077 + }, + { + "epoch": 1.1389623035192655, + "grad_norm": 0.3046875, + "learning_rate": 0.00013554865533382915, + "loss": 0.5995, + "step": 2078 + }, + { + "epoch": 1.1395187091389622, + "grad_norm": 0.330078125, + "learning_rate": 0.0001354940797150443, + "loss": 0.7362, + "step": 2079 + }, + { + "epoch": 1.140075114758659, + "grad_norm": 0.3359375, + "learning_rate": 0.0001354394919978113, + "loss": 0.7123, + "step": 2080 + }, + { + "epoch": 1.140631520378356, + "grad_norm": 0.345703125, + "learning_rate": 0.00013538489220073677, + "loss": 0.7947, + "step": 2081 + }, + { + "epoch": 1.1411879259980526, + "grad_norm": 0.314453125, + "learning_rate": 0.00013533028034243145, + "loss": 0.7087, + "step": 2082 + }, + { + "epoch": 1.1417443316177494, + "grad_norm": 0.33984375, + "learning_rate": 0.00013527565644151033, + "loss": 0.7273, + "step": 2083 + }, + { + "epoch": 1.142300737237446, + "grad_norm": 0.337890625, + "learning_rate": 0.00013522102051659234, + "loss": 0.8265, + "step": 2084 + }, + { + "epoch": 1.1428571428571428, + "grad_norm": 0.322265625, + "learning_rate": 0.00013516637258630067, + "loss": 0.696, + "step": 2085 + }, + { + "epoch": 1.1434135484768397, + "grad_norm": 0.326171875, + "learning_rate": 0.00013511171266926243, + "loss": 0.708, + "step": 2086 + }, + { + "epoch": 1.1439699540965365, + "grad_norm": 0.326171875, + "learning_rate": 0.00013505704078410898, + "loss": 0.7611, + "step": 2087 + }, + { + "epoch": 1.1445263597162332, + "grad_norm": 0.330078125, + "learning_rate": 0.00013500235694947557, + "loss": 0.7434, + "step": 2088 + }, + { + "epoch": 1.14508276533593, + "grad_norm": 0.3359375, + "learning_rate": 0.00013494766118400176, + "loss": 0.7102, + "step": 2089 + }, + { + "epoch": 1.1456391709556266, + "grad_norm": 0.328125, + "learning_rate": 0.00013489295350633095, + "loss": 0.6826, + "step": 2090 + }, + { + "epoch": 1.1461955765753233, + "grad_norm": 0.318359375, + "learning_rate": 0.00013483823393511076, + "loss": 0.6801, + "step": 2091 + }, + { + "epoch": 1.14675198219502, + "grad_norm": 0.330078125, + "learning_rate": 0.0001347835024889928, + "loss": 0.7401, + "step": 2092 + }, + { + "epoch": 1.147308387814717, + "grad_norm": 0.3125, + "learning_rate": 0.00013472875918663272, + "loss": 0.6547, + "step": 2093 + }, + { + "epoch": 1.1478647934344137, + "grad_norm": 0.34765625, + "learning_rate": 0.00013467400404669017, + "loss": 0.7809, + "step": 2094 + }, + { + "epoch": 1.1484211990541104, + "grad_norm": 0.322265625, + "learning_rate": 0.00013461923708782896, + "loss": 0.7175, + "step": 2095 + }, + { + "epoch": 1.1489776046738072, + "grad_norm": 0.322265625, + "learning_rate": 0.00013456445832871688, + "loss": 0.7707, + "step": 2096 + }, + { + "epoch": 1.1495340102935039, + "grad_norm": 0.314453125, + "learning_rate": 0.00013450966778802564, + "loss": 0.6497, + "step": 2097 + }, + { + "epoch": 1.1500904159132008, + "grad_norm": 0.306640625, + "learning_rate": 0.00013445486548443108, + "loss": 0.6757, + "step": 2098 + }, + { + "epoch": 1.1506468215328975, + "grad_norm": 0.33984375, + "learning_rate": 0.00013440005143661304, + "loss": 0.7418, + "step": 2099 + }, + { + "epoch": 1.1512032271525943, + "grad_norm": 0.326171875, + "learning_rate": 0.0001343452256632553, + "loss": 0.7288, + "step": 2100 + }, + { + "epoch": 1.151759632772291, + "grad_norm": 0.341796875, + "learning_rate": 0.0001342903881830457, + "loss": 0.7445, + "step": 2101 + }, + { + "epoch": 1.1523160383919877, + "grad_norm": 0.32421875, + "learning_rate": 0.00013423553901467598, + "loss": 0.7464, + "step": 2102 + }, + { + "epoch": 1.1528724440116844, + "grad_norm": 0.30859375, + "learning_rate": 0.00013418067817684204, + "loss": 0.6812, + "step": 2103 + }, + { + "epoch": 1.1534288496313814, + "grad_norm": 0.337890625, + "learning_rate": 0.00013412580568824357, + "loss": 0.6988, + "step": 2104 + }, + { + "epoch": 1.153985255251078, + "grad_norm": 0.337890625, + "learning_rate": 0.00013407092156758434, + "loss": 0.7441, + "step": 2105 + }, + { + "epoch": 1.1545416608707748, + "grad_norm": 0.3203125, + "learning_rate": 0.00013401602583357206, + "loss": 0.762, + "step": 2106 + }, + { + "epoch": 1.1550980664904715, + "grad_norm": 0.333984375, + "learning_rate": 0.00013396111850491832, + "loss": 0.7637, + "step": 2107 + }, + { + "epoch": 1.1556544721101683, + "grad_norm": 0.3203125, + "learning_rate": 0.00013390619960033882, + "loss": 0.7193, + "step": 2108 + }, + { + "epoch": 1.156210877729865, + "grad_norm": 0.322265625, + "learning_rate": 0.00013385126913855307, + "loss": 0.7295, + "step": 2109 + }, + { + "epoch": 1.156767283349562, + "grad_norm": 0.326171875, + "learning_rate": 0.00013379632713828458, + "loss": 0.7448, + "step": 2110 + }, + { + "epoch": 1.1573236889692586, + "grad_norm": 0.33203125, + "learning_rate": 0.00013374137361826075, + "loss": 0.7631, + "step": 2111 + }, + { + "epoch": 1.1578800945889554, + "grad_norm": 0.318359375, + "learning_rate": 0.00013368640859721302, + "loss": 0.7028, + "step": 2112 + }, + { + "epoch": 1.158436500208652, + "grad_norm": 0.318359375, + "learning_rate": 0.0001336314320938766, + "loss": 0.7438, + "step": 2113 + }, + { + "epoch": 1.1589929058283488, + "grad_norm": 0.3125, + "learning_rate": 0.00013357644412699071, + "loss": 0.6364, + "step": 2114 + }, + { + "epoch": 1.1595493114480457, + "grad_norm": 0.326171875, + "learning_rate": 0.0001335214447152984, + "loss": 0.7231, + "step": 2115 + }, + { + "epoch": 1.1601057170677425, + "grad_norm": 0.30859375, + "learning_rate": 0.0001334664338775468, + "loss": 0.6903, + "step": 2116 + }, + { + "epoch": 1.1606621226874392, + "grad_norm": 2.078125, + "learning_rate": 0.00013341141163248664, + "loss": 1.7995, + "step": 2117 + }, + { + "epoch": 1.161218528307136, + "grad_norm": 0.328125, + "learning_rate": 0.00013335637799887282, + "loss": 0.7381, + "step": 2118 + }, + { + "epoch": 1.1617749339268326, + "grad_norm": 0.330078125, + "learning_rate": 0.00013330133299546394, + "loss": 0.698, + "step": 2119 + }, + { + "epoch": 1.1623313395465293, + "grad_norm": 0.302734375, + "learning_rate": 0.00013324627664102257, + "loss": 0.6403, + "step": 2120 + }, + { + "epoch": 1.162887745166226, + "grad_norm": 0.33984375, + "learning_rate": 0.00013319120895431517, + "loss": 0.7535, + "step": 2121 + }, + { + "epoch": 1.163444150785923, + "grad_norm": 0.337890625, + "learning_rate": 0.0001331361299541119, + "loss": 0.7806, + "step": 2122 + }, + { + "epoch": 1.1640005564056197, + "grad_norm": 0.3046875, + "learning_rate": 0.00013308103965918694, + "loss": 0.6097, + "step": 2123 + }, + { + "epoch": 1.1645569620253164, + "grad_norm": 0.3359375, + "learning_rate": 0.0001330259380883183, + "loss": 0.7872, + "step": 2124 + }, + { + "epoch": 1.1651133676450132, + "grad_norm": 0.32421875, + "learning_rate": 0.00013297082526028778, + "loss": 0.7253, + "step": 2125 + }, + { + "epoch": 1.1656697732647099, + "grad_norm": 0.326171875, + "learning_rate": 0.00013291570119388106, + "loss": 0.7474, + "step": 2126 + }, + { + "epoch": 1.1662261788844068, + "grad_norm": 0.333984375, + "learning_rate": 0.00013286056590788755, + "loss": 0.7563, + "step": 2127 + }, + { + "epoch": 1.1667825845041035, + "grad_norm": 0.33203125, + "learning_rate": 0.0001328054194211007, + "loss": 0.7449, + "step": 2128 + }, + { + "epoch": 1.1673389901238003, + "grad_norm": 0.326171875, + "learning_rate": 0.00013275026175231752, + "loss": 0.7136, + "step": 2129 + }, + { + "epoch": 1.167895395743497, + "grad_norm": 0.30859375, + "learning_rate": 0.00013269509292033901, + "loss": 0.6464, + "step": 2130 + }, + { + "epoch": 1.1684518013631937, + "grad_norm": 0.3203125, + "learning_rate": 0.0001326399129439699, + "loss": 0.6748, + "step": 2131 + }, + { + "epoch": 1.1690082069828907, + "grad_norm": 0.32421875, + "learning_rate": 0.00013258472184201877, + "loss": 0.7183, + "step": 2132 + }, + { + "epoch": 1.1695646126025874, + "grad_norm": 0.337890625, + "learning_rate": 0.0001325295196332979, + "loss": 0.7433, + "step": 2133 + }, + { + "epoch": 1.170121018222284, + "grad_norm": 0.349609375, + "learning_rate": 0.0001324743063366235, + "loss": 0.8233, + "step": 2134 + }, + { + "epoch": 1.1706774238419808, + "grad_norm": 0.318359375, + "learning_rate": 0.00013241908197081538, + "loss": 0.6533, + "step": 2135 + }, + { + "epoch": 1.1712338294616775, + "grad_norm": 0.326171875, + "learning_rate": 0.00013236384655469728, + "loss": 0.6984, + "step": 2136 + }, + { + "epoch": 1.1717902350813743, + "grad_norm": 0.326171875, + "learning_rate": 0.00013230860010709665, + "loss": 0.6996, + "step": 2137 + }, + { + "epoch": 1.172346640701071, + "grad_norm": 0.32421875, + "learning_rate": 0.00013225334264684466, + "loss": 0.6575, + "step": 2138 + }, + { + "epoch": 1.172903046320768, + "grad_norm": 0.3359375, + "learning_rate": 0.00013219807419277626, + "loss": 0.8134, + "step": 2139 + }, + { + "epoch": 1.1734594519404646, + "grad_norm": 0.318359375, + "learning_rate": 0.0001321427947637302, + "loss": 0.6798, + "step": 2140 + }, + { + "epoch": 1.1740158575601614, + "grad_norm": 0.3515625, + "learning_rate": 0.0001320875043785489, + "loss": 0.776, + "step": 2141 + }, + { + "epoch": 1.174572263179858, + "grad_norm": 0.3203125, + "learning_rate": 0.00013203220305607854, + "loss": 0.7181, + "step": 2142 + }, + { + "epoch": 1.1751286687995548, + "grad_norm": 0.328125, + "learning_rate": 0.00013197689081516906, + "loss": 0.7336, + "step": 2143 + }, + { + "epoch": 1.1756850744192517, + "grad_norm": 0.333984375, + "learning_rate": 0.00013192156767467405, + "loss": 0.721, + "step": 2144 + }, + { + "epoch": 1.1762414800389485, + "grad_norm": 0.3046875, + "learning_rate": 0.00013186623365345088, + "loss": 0.615, + "step": 2145 + }, + { + "epoch": 1.1767978856586452, + "grad_norm": 0.318359375, + "learning_rate": 0.00013181088877036058, + "loss": 0.6619, + "step": 2146 + }, + { + "epoch": 1.177354291278342, + "grad_norm": 0.337890625, + "learning_rate": 0.00013175553304426796, + "loss": 0.6747, + "step": 2147 + }, + { + "epoch": 1.1779106968980386, + "grad_norm": 0.306640625, + "learning_rate": 0.0001317001664940414, + "loss": 0.6285, + "step": 2148 + }, + { + "epoch": 1.1784671025177353, + "grad_norm": 0.3046875, + "learning_rate": 0.0001316447891385531, + "loss": 0.6129, + "step": 2149 + }, + { + "epoch": 1.179023508137432, + "grad_norm": 0.322265625, + "learning_rate": 0.00013158940099667889, + "loss": 0.6575, + "step": 2150 + }, + { + "epoch": 1.179579913757129, + "grad_norm": 0.3359375, + "learning_rate": 0.00013153400208729823, + "loss": 0.7835, + "step": 2151 + }, + { + "epoch": 1.1801363193768257, + "grad_norm": 0.35546875, + "learning_rate": 0.00013147859242929432, + "loss": 0.8569, + "step": 2152 + }, + { + "epoch": 1.1806927249965224, + "grad_norm": 0.34375, + "learning_rate": 0.00013142317204155397, + "loss": 0.7776, + "step": 2153 + }, + { + "epoch": 1.1812491306162192, + "grad_norm": 0.32421875, + "learning_rate": 0.0001313677409429677, + "loss": 0.6162, + "step": 2154 + }, + { + "epoch": 1.1818055362359159, + "grad_norm": 0.322265625, + "learning_rate": 0.00013131229915242965, + "loss": 0.7546, + "step": 2155 + }, + { + "epoch": 1.1823619418556128, + "grad_norm": 0.330078125, + "learning_rate": 0.00013125684668883762, + "loss": 0.7137, + "step": 2156 + }, + { + "epoch": 1.1829183474753096, + "grad_norm": 0.3203125, + "learning_rate": 0.00013120138357109303, + "loss": 0.7243, + "step": 2157 + }, + { + "epoch": 1.1834747530950063, + "grad_norm": 0.3203125, + "learning_rate": 0.0001311459098181009, + "loss": 0.7115, + "step": 2158 + }, + { + "epoch": 1.184031158714703, + "grad_norm": 0.33203125, + "learning_rate": 0.00013109042544876996, + "loss": 0.7396, + "step": 2159 + }, + { + "epoch": 1.1845875643343997, + "grad_norm": 0.341796875, + "learning_rate": 0.00013103493048201246, + "loss": 0.7101, + "step": 2160 + }, + { + "epoch": 1.1851439699540967, + "grad_norm": 0.341796875, + "learning_rate": 0.0001309794249367444, + "loss": 0.8095, + "step": 2161 + }, + { + "epoch": 1.1857003755737934, + "grad_norm": 0.3125, + "learning_rate": 0.00013092390883188526, + "loss": 0.6576, + "step": 2162 + }, + { + "epoch": 1.18625678119349, + "grad_norm": 0.330078125, + "learning_rate": 0.00013086838218635813, + "loss": 0.7822, + "step": 2163 + }, + { + "epoch": 1.1868131868131868, + "grad_norm": 0.322265625, + "learning_rate": 0.00013081284501908976, + "loss": 0.6946, + "step": 2164 + }, + { + "epoch": 1.1873695924328835, + "grad_norm": 0.31640625, + "learning_rate": 0.00013075729734901044, + "loss": 0.6942, + "step": 2165 + }, + { + "epoch": 1.1879259980525803, + "grad_norm": 0.330078125, + "learning_rate": 0.000130701739195054, + "loss": 0.7864, + "step": 2166 + }, + { + "epoch": 1.188482403672277, + "grad_norm": 0.3125, + "learning_rate": 0.000130646170576158, + "loss": 0.6483, + "step": 2167 + }, + { + "epoch": 1.189038809291974, + "grad_norm": 0.31640625, + "learning_rate": 0.00013059059151126335, + "loss": 0.674, + "step": 2168 + }, + { + "epoch": 1.1895952149116706, + "grad_norm": 0.322265625, + "learning_rate": 0.00013053500201931468, + "loss": 0.7591, + "step": 2169 + }, + { + "epoch": 1.1901516205313674, + "grad_norm": 0.318359375, + "learning_rate": 0.00013047940211926014, + "loss": 0.7159, + "step": 2170 + }, + { + "epoch": 1.190708026151064, + "grad_norm": 0.34765625, + "learning_rate": 0.00013042379183005135, + "loss": 0.7877, + "step": 2171 + }, + { + "epoch": 1.1912644317707608, + "grad_norm": 0.326171875, + "learning_rate": 0.00013036817117064361, + "loss": 0.7207, + "step": 2172 + }, + { + "epoch": 1.1918208373904577, + "grad_norm": 0.33203125, + "learning_rate": 0.00013031254015999564, + "loss": 0.7245, + "step": 2173 + }, + { + "epoch": 1.1923772430101545, + "grad_norm": 0.326171875, + "learning_rate": 0.00013025689881706972, + "loss": 0.7071, + "step": 2174 + }, + { + "epoch": 1.1929336486298512, + "grad_norm": 0.34375, + "learning_rate": 0.0001302012471608317, + "loss": 0.7959, + "step": 2175 + }, + { + "epoch": 1.193490054249548, + "grad_norm": 0.333984375, + "learning_rate": 0.0001301455852102508, + "loss": 0.7475, + "step": 2176 + }, + { + "epoch": 1.1940464598692446, + "grad_norm": 0.306640625, + "learning_rate": 0.00013008991298429999, + "loss": 0.6675, + "step": 2177 + }, + { + "epoch": 1.1946028654889413, + "grad_norm": 0.326171875, + "learning_rate": 0.0001300342305019555, + "loss": 0.693, + "step": 2178 + }, + { + "epoch": 1.1951592711086383, + "grad_norm": 0.337890625, + "learning_rate": 0.00012997853778219721, + "loss": 0.7605, + "step": 2179 + }, + { + "epoch": 1.195715676728335, + "grad_norm": 0.322265625, + "learning_rate": 0.00012992283484400847, + "loss": 0.7103, + "step": 2180 + }, + { + "epoch": 1.1962720823480317, + "grad_norm": 0.3203125, + "learning_rate": 0.00012986712170637597, + "loss": 0.6464, + "step": 2181 + }, + { + "epoch": 1.1968284879677284, + "grad_norm": 0.322265625, + "learning_rate": 0.0001298113983882901, + "loss": 0.7043, + "step": 2182 + }, + { + "epoch": 1.1973848935874252, + "grad_norm": 0.3203125, + "learning_rate": 0.00012975566490874458, + "loss": 0.7208, + "step": 2183 + }, + { + "epoch": 1.197941299207122, + "grad_norm": 0.314453125, + "learning_rate": 0.00012969992128673662, + "loss": 0.6785, + "step": 2184 + }, + { + "epoch": 1.1984977048268188, + "grad_norm": 0.34765625, + "learning_rate": 0.00012964416754126685, + "loss": 0.7649, + "step": 2185 + }, + { + "epoch": 1.1990541104465156, + "grad_norm": 0.326171875, + "learning_rate": 0.00012958840369133945, + "loss": 0.7386, + "step": 2186 + }, + { + "epoch": 1.1996105160662123, + "grad_norm": 0.31640625, + "learning_rate": 0.00012953262975596197, + "loss": 0.6977, + "step": 2187 + }, + { + "epoch": 1.200166921685909, + "grad_norm": 0.3515625, + "learning_rate": 0.0001294768457541454, + "loss": 0.806, + "step": 2188 + }, + { + "epoch": 1.2007233273056057, + "grad_norm": 0.35546875, + "learning_rate": 0.00012942105170490417, + "loss": 0.8397, + "step": 2189 + }, + { + "epoch": 1.2012797329253027, + "grad_norm": 0.30859375, + "learning_rate": 0.00012936524762725618, + "loss": 0.6611, + "step": 2190 + }, + { + "epoch": 1.2018361385449994, + "grad_norm": 0.3359375, + "learning_rate": 0.00012930943354022263, + "loss": 0.787, + "step": 2191 + }, + { + "epoch": 1.202392544164696, + "grad_norm": 0.34765625, + "learning_rate": 0.0001292536094628283, + "loss": 0.7341, + "step": 2192 + }, + { + "epoch": 1.2029489497843928, + "grad_norm": 0.3359375, + "learning_rate": 0.00012919777541410126, + "loss": 0.8072, + "step": 2193 + }, + { + "epoch": 1.2035053554040895, + "grad_norm": 0.333984375, + "learning_rate": 0.000129141931413073, + "loss": 0.7513, + "step": 2194 + }, + { + "epoch": 1.2040617610237863, + "grad_norm": 0.349609375, + "learning_rate": 0.00012908607747877832, + "loss": 0.7565, + "step": 2195 + }, + { + "epoch": 1.204618166643483, + "grad_norm": 0.318359375, + "learning_rate": 0.0001290302136302556, + "loss": 0.7509, + "step": 2196 + }, + { + "epoch": 1.20517457226318, + "grad_norm": 0.318359375, + "learning_rate": 0.00012897433988654644, + "loss": 0.6535, + "step": 2197 + }, + { + "epoch": 1.2057309778828766, + "grad_norm": 0.322265625, + "learning_rate": 0.0001289184562666959, + "loss": 0.6429, + "step": 2198 + }, + { + "epoch": 1.2062873835025734, + "grad_norm": 0.32421875, + "learning_rate": 0.00012886256278975231, + "loss": 0.735, + "step": 2199 + }, + { + "epoch": 1.20684378912227, + "grad_norm": 0.34765625, + "learning_rate": 0.00012880665947476752, + "loss": 0.8336, + "step": 2200 + }, + { + "epoch": 1.2074001947419668, + "grad_norm": 0.3203125, + "learning_rate": 0.00012875074634079652, + "loss": 0.6905, + "step": 2201 + }, + { + "epoch": 1.2079566003616637, + "grad_norm": 0.3359375, + "learning_rate": 0.0001286948234068978, + "loss": 0.7417, + "step": 2202 + }, + { + "epoch": 1.2085130059813605, + "grad_norm": 0.322265625, + "learning_rate": 0.00012863889069213323, + "loss": 0.6903, + "step": 2203 + }, + { + "epoch": 1.2090694116010572, + "grad_norm": 0.33203125, + "learning_rate": 0.00012858294821556786, + "loss": 0.7601, + "step": 2204 + }, + { + "epoch": 1.209625817220754, + "grad_norm": 0.32421875, + "learning_rate": 0.0001285269959962701, + "loss": 0.7129, + "step": 2205 + }, + { + "epoch": 1.2101822228404506, + "grad_norm": 0.318359375, + "learning_rate": 0.00012847103405331178, + "loss": 0.6655, + "step": 2206 + }, + { + "epoch": 1.2107386284601476, + "grad_norm": 0.32421875, + "learning_rate": 0.00012841506240576802, + "loss": 0.722, + "step": 2207 + }, + { + "epoch": 1.2112950340798443, + "grad_norm": 0.83203125, + "learning_rate": 0.00012835908107271715, + "loss": 0.8153, + "step": 2208 + }, + { + "epoch": 1.211851439699541, + "grad_norm": 0.3359375, + "learning_rate": 0.00012830309007324088, + "loss": 0.7669, + "step": 2209 + }, + { + "epoch": 1.2124078453192377, + "grad_norm": 0.70703125, + "learning_rate": 0.00012824708942642423, + "loss": 1.6954, + "step": 2210 + }, + { + "epoch": 1.2129642509389345, + "grad_norm": 0.318359375, + "learning_rate": 0.00012819107915135547, + "loss": 0.6916, + "step": 2211 + }, + { + "epoch": 1.2135206565586312, + "grad_norm": 0.345703125, + "learning_rate": 0.00012813505926712615, + "loss": 0.778, + "step": 2212 + }, + { + "epoch": 1.214077062178328, + "grad_norm": 0.326171875, + "learning_rate": 0.00012807902979283112, + "loss": 0.7034, + "step": 2213 + }, + { + "epoch": 1.2146334677980248, + "grad_norm": 0.328125, + "learning_rate": 0.00012802299074756842, + "loss": 0.7273, + "step": 2214 + }, + { + "epoch": 1.2151898734177216, + "grad_norm": 0.3515625, + "learning_rate": 0.0001279669421504395, + "loss": 0.8011, + "step": 2215 + }, + { + "epoch": 1.2157462790374183, + "grad_norm": 0.318359375, + "learning_rate": 0.000127910884020549, + "loss": 0.6606, + "step": 2216 + }, + { + "epoch": 1.216302684657115, + "grad_norm": 0.306640625, + "learning_rate": 0.00012785481637700473, + "loss": 0.6426, + "step": 2217 + }, + { + "epoch": 1.2168590902768117, + "grad_norm": 0.326171875, + "learning_rate": 0.0001277987392389178, + "loss": 0.6805, + "step": 2218 + }, + { + "epoch": 1.2174154958965087, + "grad_norm": 0.333984375, + "learning_rate": 0.00012774265262540264, + "loss": 0.7354, + "step": 2219 + }, + { + "epoch": 1.2179719015162054, + "grad_norm": 0.318359375, + "learning_rate": 0.00012768655655557675, + "loss": 0.6821, + "step": 2220 + }, + { + "epoch": 1.218528307135902, + "grad_norm": 0.3359375, + "learning_rate": 0.000127630451048561, + "loss": 0.7067, + "step": 2221 + }, + { + "epoch": 1.2190847127555988, + "grad_norm": 0.333984375, + "learning_rate": 0.00012757433612347937, + "loss": 0.7222, + "step": 2222 + }, + { + "epoch": 1.2196411183752955, + "grad_norm": 0.3203125, + "learning_rate": 0.0001275182117994591, + "loss": 0.7561, + "step": 2223 + }, + { + "epoch": 1.2201975239949923, + "grad_norm": 0.333984375, + "learning_rate": 0.00012746207809563064, + "loss": 0.7554, + "step": 2224 + }, + { + "epoch": 1.2207539296146892, + "grad_norm": 0.353515625, + "learning_rate": 0.00012740593503112762, + "loss": 0.7669, + "step": 2225 + }, + { + "epoch": 1.221310335234386, + "grad_norm": 0.32421875, + "learning_rate": 0.00012734978262508682, + "loss": 0.6811, + "step": 2226 + }, + { + "epoch": 1.2218667408540826, + "grad_norm": 0.314453125, + "learning_rate": 0.00012729362089664828, + "loss": 0.6752, + "step": 2227 + }, + { + "epoch": 1.2224231464737794, + "grad_norm": 0.33203125, + "learning_rate": 0.00012723744986495525, + "loss": 0.7698, + "step": 2228 + }, + { + "epoch": 1.222979552093476, + "grad_norm": 0.333984375, + "learning_rate": 0.000127181269549154, + "loss": 0.7404, + "step": 2229 + }, + { + "epoch": 1.2235359577131728, + "grad_norm": 0.32421875, + "learning_rate": 0.00012712507996839404, + "loss": 0.7231, + "step": 2230 + }, + { + "epoch": 1.2240923633328697, + "grad_norm": 0.3125, + "learning_rate": 0.00012706888114182813, + "loss": 0.6616, + "step": 2231 + }, + { + "epoch": 1.2246487689525665, + "grad_norm": 0.33203125, + "learning_rate": 0.00012701267308861208, + "loss": 0.7432, + "step": 2232 + }, + { + "epoch": 1.2252051745722632, + "grad_norm": 0.326171875, + "learning_rate": 0.0001269564558279048, + "loss": 0.6797, + "step": 2233 + }, + { + "epoch": 1.22576158019196, + "grad_norm": 0.333984375, + "learning_rate": 0.0001269002293788685, + "loss": 0.7114, + "step": 2234 + }, + { + "epoch": 1.2263179858116566, + "grad_norm": 0.322265625, + "learning_rate": 0.00012684399376066834, + "loss": 0.7546, + "step": 2235 + }, + { + "epoch": 1.2268743914313536, + "grad_norm": 0.3203125, + "learning_rate": 0.00012678774899247273, + "loss": 0.7231, + "step": 2236 + }, + { + "epoch": 1.2274307970510503, + "grad_norm": 0.337890625, + "learning_rate": 0.00012673149509345318, + "loss": 0.6955, + "step": 2237 + }, + { + "epoch": 1.227987202670747, + "grad_norm": 0.326171875, + "learning_rate": 0.00012667523208278425, + "loss": 0.6672, + "step": 2238 + }, + { + "epoch": 1.2285436082904437, + "grad_norm": 0.326171875, + "learning_rate": 0.00012661895997964368, + "loss": 0.7162, + "step": 2239 + }, + { + "epoch": 1.2291000139101405, + "grad_norm": 0.326171875, + "learning_rate": 0.00012656267880321227, + "loss": 0.6993, + "step": 2240 + }, + { + "epoch": 1.2296564195298372, + "grad_norm": 0.318359375, + "learning_rate": 0.00012650638857267393, + "loss": 0.6886, + "step": 2241 + }, + { + "epoch": 1.230212825149534, + "grad_norm": 0.32421875, + "learning_rate": 0.00012645008930721563, + "loss": 0.7082, + "step": 2242 + }, + { + "epoch": 1.2307692307692308, + "grad_norm": 0.31640625, + "learning_rate": 0.00012639378102602744, + "loss": 0.6476, + "step": 2243 + }, + { + "epoch": 1.2313256363889276, + "grad_norm": 0.33203125, + "learning_rate": 0.0001263374637483025, + "loss": 0.8026, + "step": 2244 + }, + { + "epoch": 1.2318820420086243, + "grad_norm": 0.328125, + "learning_rate": 0.00012628113749323702, + "loss": 0.7189, + "step": 2245 + }, + { + "epoch": 1.232438447628321, + "grad_norm": 0.330078125, + "learning_rate": 0.00012622480228003025, + "loss": 0.6837, + "step": 2246 + }, + { + "epoch": 1.2329948532480177, + "grad_norm": 0.33203125, + "learning_rate": 0.00012616845812788455, + "loss": 0.7705, + "step": 2247 + }, + { + "epoch": 1.2335512588677147, + "grad_norm": 0.345703125, + "learning_rate": 0.00012611210505600524, + "loss": 0.7483, + "step": 2248 + }, + { + "epoch": 1.2341076644874114, + "grad_norm": 0.314453125, + "learning_rate": 0.00012605574308360078, + "loss": 0.6461, + "step": 2249 + }, + { + "epoch": 1.234664070107108, + "grad_norm": 0.33203125, + "learning_rate": 0.00012599937222988253, + "loss": 0.7263, + "step": 2250 + }, + { + "epoch": 1.2352204757268048, + "grad_norm": 0.330078125, + "learning_rate": 0.00012594299251406503, + "loss": 0.7418, + "step": 2251 + }, + { + "epoch": 1.2357768813465015, + "grad_norm": 0.34375, + "learning_rate": 0.00012588660395536575, + "loss": 0.7898, + "step": 2252 + }, + { + "epoch": 1.2363332869661983, + "grad_norm": 0.337890625, + "learning_rate": 0.00012583020657300518, + "loss": 0.8093, + "step": 2253 + }, + { + "epoch": 1.2368896925858952, + "grad_norm": 0.296875, + "learning_rate": 0.00012577380038620686, + "loss": 0.6003, + "step": 2254 + }, + { + "epoch": 1.237446098205592, + "grad_norm": 0.32421875, + "learning_rate": 0.0001257173854141972, + "loss": 0.7314, + "step": 2255 + }, + { + "epoch": 1.2380025038252886, + "grad_norm": 0.34375, + "learning_rate": 0.00012566096167620586, + "loss": 0.84, + "step": 2256 + }, + { + "epoch": 1.2385589094449854, + "grad_norm": 0.330078125, + "learning_rate": 0.00012560452919146522, + "loss": 0.7701, + "step": 2257 + }, + { + "epoch": 1.239115315064682, + "grad_norm": 0.318359375, + "learning_rate": 0.00012554808797921084, + "loss": 0.7251, + "step": 2258 + }, + { + "epoch": 1.2396717206843788, + "grad_norm": 0.30078125, + "learning_rate": 0.00012549163805868107, + "loss": 0.6146, + "step": 2259 + }, + { + "epoch": 1.2402281263040758, + "grad_norm": 0.314453125, + "learning_rate": 0.0001254351794491174, + "loss": 0.6198, + "step": 2260 + }, + { + "epoch": 1.2407845319237725, + "grad_norm": 0.328125, + "learning_rate": 0.00012537871216976423, + "loss": 0.7025, + "step": 2261 + }, + { + "epoch": 1.2413409375434692, + "grad_norm": 0.3203125, + "learning_rate": 0.00012532223623986882, + "loss": 0.7221, + "step": 2262 + }, + { + "epoch": 1.241897343163166, + "grad_norm": 0.337890625, + "learning_rate": 0.00012526575167868146, + "loss": 0.7848, + "step": 2263 + }, + { + "epoch": 1.2424537487828626, + "grad_norm": 0.341796875, + "learning_rate": 0.00012520925850545547, + "loss": 0.738, + "step": 2264 + }, + { + "epoch": 1.2430101544025596, + "grad_norm": 0.326171875, + "learning_rate": 0.00012515275673944691, + "loss": 0.7456, + "step": 2265 + }, + { + "epoch": 1.2435665600222563, + "grad_norm": 0.32421875, + "learning_rate": 0.00012509624639991493, + "loss": 0.7308, + "step": 2266 + }, + { + "epoch": 1.244122965641953, + "grad_norm": 0.328125, + "learning_rate": 0.00012503972750612153, + "loss": 0.727, + "step": 2267 + }, + { + "epoch": 1.2446793712616497, + "grad_norm": 0.318359375, + "learning_rate": 0.0001249832000773316, + "loss": 0.688, + "step": 2268 + }, + { + "epoch": 1.2452357768813465, + "grad_norm": 0.453125, + "learning_rate": 0.000124926664132813, + "loss": 0.7602, + "step": 2269 + }, + { + "epoch": 1.2457921825010432, + "grad_norm": 0.3203125, + "learning_rate": 0.00012487011969183654, + "loss": 0.694, + "step": 2270 + }, + { + "epoch": 1.24634858812074, + "grad_norm": 0.322265625, + "learning_rate": 0.00012481356677367572, + "loss": 0.6497, + "step": 2271 + }, + { + "epoch": 1.2469049937404368, + "grad_norm": 0.330078125, + "learning_rate": 0.0001247570053976072, + "loss": 0.7003, + "step": 2272 + }, + { + "epoch": 1.2474613993601336, + "grad_norm": 0.32421875, + "learning_rate": 0.0001247004355829103, + "loss": 0.6852, + "step": 2273 + }, + { + "epoch": 1.2480178049798303, + "grad_norm": 0.33203125, + "learning_rate": 0.00012464385734886735, + "loss": 0.6751, + "step": 2274 + }, + { + "epoch": 1.248574210599527, + "grad_norm": 0.337890625, + "learning_rate": 0.00012458727071476346, + "loss": 0.7585, + "step": 2275 + }, + { + "epoch": 1.2491306162192237, + "grad_norm": 0.3125, + "learning_rate": 0.00012453067569988672, + "loss": 0.6429, + "step": 2276 + }, + { + "epoch": 1.2496870218389207, + "grad_norm": 0.328125, + "learning_rate": 0.00012447407232352796, + "loss": 0.7197, + "step": 2277 + }, + { + "epoch": 1.2502434274586174, + "grad_norm": 0.337890625, + "learning_rate": 0.0001244174606049809, + "loss": 0.7096, + "step": 2278 + }, + { + "epoch": 1.250799833078314, + "grad_norm": 0.33203125, + "learning_rate": 0.00012436084056354212, + "loss": 0.6943, + "step": 2279 + }, + { + "epoch": 1.2513562386980108, + "grad_norm": 0.3359375, + "learning_rate": 0.00012430421221851101, + "loss": 0.7557, + "step": 2280 + }, + { + "epoch": 1.2519126443177075, + "grad_norm": 0.33203125, + "learning_rate": 0.00012424757558918983, + "loss": 0.7623, + "step": 2281 + }, + { + "epoch": 1.2524690499374045, + "grad_norm": 0.3203125, + "learning_rate": 0.00012419093069488364, + "loss": 0.7014, + "step": 2282 + }, + { + "epoch": 1.2530254555571012, + "grad_norm": 0.31640625, + "learning_rate": 0.00012413427755490026, + "loss": 0.6064, + "step": 2283 + }, + { + "epoch": 1.253581861176798, + "grad_norm": 0.3125, + "learning_rate": 0.00012407761618855046, + "loss": 0.7065, + "step": 2284 + }, + { + "epoch": 1.2541382667964946, + "grad_norm": 0.330078125, + "learning_rate": 0.0001240209466151477, + "loss": 0.7589, + "step": 2285 + }, + { + "epoch": 1.2546946724161914, + "grad_norm": 0.298828125, + "learning_rate": 0.00012396426885400824, + "loss": 0.6115, + "step": 2286 + }, + { + "epoch": 1.255251078035888, + "grad_norm": 0.349609375, + "learning_rate": 0.0001239075829244512, + "loss": 0.6757, + "step": 2287 + }, + { + "epoch": 1.2558074836555848, + "grad_norm": 0.337890625, + "learning_rate": 0.0001238508888457984, + "loss": 0.71, + "step": 2288 + }, + { + "epoch": 1.2563638892752818, + "grad_norm": 0.333984375, + "learning_rate": 0.00012379418663737454, + "loss": 0.7367, + "step": 2289 + }, + { + "epoch": 1.2569202948949785, + "grad_norm": 0.349609375, + "learning_rate": 0.00012373747631850697, + "loss": 0.7003, + "step": 2290 + }, + { + "epoch": 1.2574767005146752, + "grad_norm": 0.34765625, + "learning_rate": 0.00012368075790852596, + "loss": 0.8202, + "step": 2291 + }, + { + "epoch": 1.258033106134372, + "grad_norm": 0.3125, + "learning_rate": 0.00012362403142676432, + "loss": 0.6014, + "step": 2292 + }, + { + "epoch": 1.2585895117540686, + "grad_norm": 0.328125, + "learning_rate": 0.0001235672968925578, + "loss": 0.7264, + "step": 2293 + }, + { + "epoch": 1.2591459173737656, + "grad_norm": 0.3359375, + "learning_rate": 0.00012351055432524487, + "loss": 0.7665, + "step": 2294 + }, + { + "epoch": 1.2597023229934623, + "grad_norm": 0.326171875, + "learning_rate": 0.00012345380374416667, + "loss": 0.7192, + "step": 2295 + }, + { + "epoch": 1.260258728613159, + "grad_norm": 0.302734375, + "learning_rate": 0.0001233970451686671, + "loss": 0.534, + "step": 2296 + }, + { + "epoch": 1.2608151342328557, + "grad_norm": 0.302734375, + "learning_rate": 0.00012334027861809273, + "loss": 0.6783, + "step": 2297 + }, + { + "epoch": 1.2613715398525525, + "grad_norm": 0.298828125, + "learning_rate": 0.000123283504111793, + "loss": 0.6038, + "step": 2298 + }, + { + "epoch": 1.2619279454722494, + "grad_norm": 0.33203125, + "learning_rate": 0.00012322672166911988, + "loss": 0.7733, + "step": 2299 + }, + { + "epoch": 1.262484351091946, + "grad_norm": 0.314453125, + "learning_rate": 0.00012316993130942816, + "loss": 0.6894, + "step": 2300 + }, + { + "epoch": 1.2630407567116428, + "grad_norm": 0.33203125, + "learning_rate": 0.00012311313305207535, + "loss": 0.7264, + "step": 2301 + }, + { + "epoch": 1.2635971623313396, + "grad_norm": 0.314453125, + "learning_rate": 0.00012305632691642155, + "loss": 0.6621, + "step": 2302 + }, + { + "epoch": 1.2641535679510363, + "grad_norm": 0.353515625, + "learning_rate": 0.0001229995129218296, + "loss": 0.8092, + "step": 2303 + }, + { + "epoch": 1.264709973570733, + "grad_norm": 0.32421875, + "learning_rate": 0.00012294269108766503, + "loss": 0.6325, + "step": 2304 + }, + { + "epoch": 1.2652663791904297, + "grad_norm": 0.318359375, + "learning_rate": 0.00012288586143329605, + "loss": 0.6751, + "step": 2305 + }, + { + "epoch": 1.2658227848101267, + "grad_norm": 0.314453125, + "learning_rate": 0.00012282902397809344, + "loss": 0.6358, + "step": 2306 + }, + { + "epoch": 1.2663791904298234, + "grad_norm": 0.3125, + "learning_rate": 0.00012277217874143078, + "loss": 0.669, + "step": 2307 + }, + { + "epoch": 1.26693559604952, + "grad_norm": 0.3203125, + "learning_rate": 0.00012271532574268423, + "loss": 0.6807, + "step": 2308 + }, + { + "epoch": 1.2674920016692168, + "grad_norm": 0.3203125, + "learning_rate": 0.0001226584650012326, + "loss": 0.7123, + "step": 2309 + }, + { + "epoch": 1.2680484072889135, + "grad_norm": 0.33203125, + "learning_rate": 0.0001226015965364573, + "loss": 0.6789, + "step": 2310 + }, + { + "epoch": 1.2686048129086105, + "grad_norm": 0.33203125, + "learning_rate": 0.00012254472036774244, + "loss": 0.8336, + "step": 2311 + }, + { + "epoch": 1.2691612185283072, + "grad_norm": 0.341796875, + "learning_rate": 0.0001224878365144747, + "loss": 0.7687, + "step": 2312 + }, + { + "epoch": 1.269717624148004, + "grad_norm": 0.35546875, + "learning_rate": 0.00012243094499604347, + "loss": 0.8454, + "step": 2313 + }, + { + "epoch": 1.2702740297677007, + "grad_norm": 0.32421875, + "learning_rate": 0.00012237404583184067, + "loss": 0.682, + "step": 2314 + }, + { + "epoch": 1.2708304353873974, + "grad_norm": 0.33203125, + "learning_rate": 0.00012231713904126083, + "loss": 0.7317, + "step": 2315 + }, + { + "epoch": 1.271386841007094, + "grad_norm": 0.3984375, + "learning_rate": 0.00012226022464370106, + "loss": 0.7626, + "step": 2316 + }, + { + "epoch": 1.2719432466267908, + "grad_norm": 0.31640625, + "learning_rate": 0.00012220330265856117, + "loss": 0.6676, + "step": 2317 + }, + { + "epoch": 1.2724996522464878, + "grad_norm": 0.322265625, + "learning_rate": 0.00012214637310524346, + "loss": 0.6812, + "step": 2318 + }, + { + "epoch": 1.2730560578661845, + "grad_norm": 0.318359375, + "learning_rate": 0.0001220894360031528, + "loss": 0.6828, + "step": 2319 + }, + { + "epoch": 1.2736124634858812, + "grad_norm": 0.333984375, + "learning_rate": 0.00012203249137169672, + "loss": 0.7557, + "step": 2320 + }, + { + "epoch": 1.274168869105578, + "grad_norm": 0.322265625, + "learning_rate": 0.00012197553923028524, + "loss": 0.7348, + "step": 2321 + }, + { + "epoch": 1.2747252747252746, + "grad_norm": 0.31640625, + "learning_rate": 0.00012191857959833097, + "loss": 0.704, + "step": 2322 + }, + { + "epoch": 1.2752816803449716, + "grad_norm": 0.32421875, + "learning_rate": 0.00012186161249524908, + "loss": 0.7566, + "step": 2323 + }, + { + "epoch": 1.2758380859646683, + "grad_norm": 0.318359375, + "learning_rate": 0.00012180463794045724, + "loss": 0.6965, + "step": 2324 + }, + { + "epoch": 1.276394491584365, + "grad_norm": 0.333984375, + "learning_rate": 0.00012174765595337572, + "loss": 0.7508, + "step": 2325 + }, + { + "epoch": 1.2769508972040617, + "grad_norm": 0.337890625, + "learning_rate": 0.00012169066655342729, + "loss": 0.7656, + "step": 2326 + }, + { + "epoch": 1.2775073028237585, + "grad_norm": 0.33203125, + "learning_rate": 0.00012163366976003727, + "loss": 0.7445, + "step": 2327 + }, + { + "epoch": 1.2780637084434554, + "grad_norm": 0.333984375, + "learning_rate": 0.00012157666559263345, + "loss": 0.6856, + "step": 2328 + }, + { + "epoch": 1.278620114063152, + "grad_norm": 0.33984375, + "learning_rate": 0.00012151965407064618, + "loss": 0.731, + "step": 2329 + }, + { + "epoch": 1.2791765196828488, + "grad_norm": 0.333984375, + "learning_rate": 0.00012146263521350831, + "loss": 0.6502, + "step": 2330 + }, + { + "epoch": 1.2797329253025456, + "grad_norm": 0.328125, + "learning_rate": 0.00012140560904065521, + "loss": 0.733, + "step": 2331 + }, + { + "epoch": 1.2802893309222423, + "grad_norm": 0.314453125, + "learning_rate": 0.0001213485755715247, + "loss": 0.6601, + "step": 2332 + }, + { + "epoch": 1.280845736541939, + "grad_norm": 0.345703125, + "learning_rate": 0.00012129153482555708, + "loss": 0.8235, + "step": 2333 + }, + { + "epoch": 1.2814021421616357, + "grad_norm": 0.330078125, + "learning_rate": 0.0001212344868221952, + "loss": 0.7273, + "step": 2334 + }, + { + "epoch": 1.2819585477813327, + "grad_norm": 0.328125, + "learning_rate": 0.00012117743158088431, + "loss": 0.7548, + "step": 2335 + }, + { + "epoch": 1.2825149534010294, + "grad_norm": 0.32421875, + "learning_rate": 0.00012112036912107216, + "loss": 0.7111, + "step": 2336 + }, + { + "epoch": 1.283071359020726, + "grad_norm": 0.33984375, + "learning_rate": 0.00012106329946220898, + "loss": 0.7864, + "step": 2337 + }, + { + "epoch": 1.2836277646404228, + "grad_norm": 0.322265625, + "learning_rate": 0.00012100622262374741, + "loss": 0.6583, + "step": 2338 + }, + { + "epoch": 1.2841841702601196, + "grad_norm": 0.33984375, + "learning_rate": 0.00012094913862514257, + "loss": 0.7486, + "step": 2339 + }, + { + "epoch": 1.2847405758798165, + "grad_norm": 0.318359375, + "learning_rate": 0.00012089204748585197, + "loss": 0.6553, + "step": 2340 + }, + { + "epoch": 1.2852969814995132, + "grad_norm": 0.3359375, + "learning_rate": 0.00012083494922533562, + "loss": 0.7516, + "step": 2341 + }, + { + "epoch": 1.28585338711921, + "grad_norm": 0.328125, + "learning_rate": 0.00012077784386305597, + "loss": 0.7478, + "step": 2342 + }, + { + "epoch": 1.2864097927389067, + "grad_norm": 0.34765625, + "learning_rate": 0.00012072073141847782, + "loss": 0.8059, + "step": 2343 + }, + { + "epoch": 1.2869661983586034, + "grad_norm": 0.32421875, + "learning_rate": 0.00012066361191106841, + "loss": 0.7272, + "step": 2344 + }, + { + "epoch": 1.2875226039783003, + "grad_norm": 0.337890625, + "learning_rate": 0.00012060648536029735, + "loss": 0.7071, + "step": 2345 + }, + { + "epoch": 1.2880790095979968, + "grad_norm": 0.314453125, + "learning_rate": 0.00012054935178563678, + "loss": 0.7131, + "step": 2346 + }, + { + "epoch": 1.2886354152176938, + "grad_norm": 0.33203125, + "learning_rate": 0.0001204922112065611, + "loss": 0.691, + "step": 2347 + }, + { + "epoch": 1.2891918208373905, + "grad_norm": 0.326171875, + "learning_rate": 0.00012043506364254715, + "loss": 0.761, + "step": 2348 + }, + { + "epoch": 1.2897482264570872, + "grad_norm": 0.337890625, + "learning_rate": 0.00012037790911307412, + "loss": 0.7459, + "step": 2349 + }, + { + "epoch": 1.290304632076784, + "grad_norm": 0.318359375, + "learning_rate": 0.00012032074763762367, + "loss": 0.7152, + "step": 2350 + }, + { + "epoch": 1.2908610376964806, + "grad_norm": 0.3359375, + "learning_rate": 0.00012026357923567968, + "loss": 0.7651, + "step": 2351 + }, + { + "epoch": 1.2914174433161776, + "grad_norm": 0.318359375, + "learning_rate": 0.00012020640392672853, + "loss": 0.708, + "step": 2352 + }, + { + "epoch": 1.2919738489358743, + "grad_norm": 0.328125, + "learning_rate": 0.00012014922173025885, + "loss": 0.7738, + "step": 2353 + }, + { + "epoch": 1.292530254555571, + "grad_norm": 0.330078125, + "learning_rate": 0.0001200920326657617, + "loss": 0.6697, + "step": 2354 + }, + { + "epoch": 1.2930866601752677, + "grad_norm": 0.326171875, + "learning_rate": 0.0001200348367527304, + "loss": 0.7044, + "step": 2355 + }, + { + "epoch": 1.2936430657949645, + "grad_norm": 0.330078125, + "learning_rate": 0.00011997763401066067, + "loss": 0.6637, + "step": 2356 + }, + { + "epoch": 1.2941994714146614, + "grad_norm": 0.3359375, + "learning_rate": 0.0001199204244590505, + "loss": 0.7396, + "step": 2357 + }, + { + "epoch": 1.2947558770343581, + "grad_norm": 0.341796875, + "learning_rate": 0.00011986320811740028, + "loss": 0.7384, + "step": 2358 + }, + { + "epoch": 1.2953122826540548, + "grad_norm": 0.330078125, + "learning_rate": 0.00011980598500521267, + "loss": 0.6891, + "step": 2359 + }, + { + "epoch": 1.2958686882737516, + "grad_norm": 0.328125, + "learning_rate": 0.0001197487551419926, + "loss": 0.6816, + "step": 2360 + }, + { + "epoch": 1.2964250938934483, + "grad_norm": 0.34375, + "learning_rate": 0.00011969151854724737, + "loss": 0.7689, + "step": 2361 + }, + { + "epoch": 1.296981499513145, + "grad_norm": 0.328125, + "learning_rate": 0.00011963427524048652, + "loss": 0.7527, + "step": 2362 + }, + { + "epoch": 1.2975379051328417, + "grad_norm": 0.3203125, + "learning_rate": 0.00011957702524122192, + "loss": 0.665, + "step": 2363 + }, + { + "epoch": 1.2980943107525387, + "grad_norm": 0.33203125, + "learning_rate": 0.0001195197685689677, + "loss": 0.7169, + "step": 2364 + }, + { + "epoch": 1.2986507163722354, + "grad_norm": 0.345703125, + "learning_rate": 0.00011946250524324023, + "loss": 0.7796, + "step": 2365 + }, + { + "epoch": 1.2992071219919321, + "grad_norm": 0.30859375, + "learning_rate": 0.00011940523528355825, + "loss": 0.6301, + "step": 2366 + }, + { + "epoch": 1.2997635276116288, + "grad_norm": 0.322265625, + "learning_rate": 0.00011934795870944264, + "loss": 0.7309, + "step": 2367 + }, + { + "epoch": 1.3003199332313256, + "grad_norm": 0.333984375, + "learning_rate": 0.00011929067554041662, + "loss": 0.7462, + "step": 2368 + }, + { + "epoch": 1.3008763388510225, + "grad_norm": 0.314453125, + "learning_rate": 0.00011923338579600563, + "loss": 0.6727, + "step": 2369 + }, + { + "epoch": 1.3014327444707192, + "grad_norm": 0.32421875, + "learning_rate": 0.00011917608949573729, + "loss": 0.7083, + "step": 2370 + }, + { + "epoch": 1.301989150090416, + "grad_norm": 0.326171875, + "learning_rate": 0.00011911878665914158, + "loss": 0.7418, + "step": 2371 + }, + { + "epoch": 1.3025455557101127, + "grad_norm": 0.330078125, + "learning_rate": 0.00011906147730575063, + "loss": 0.7167, + "step": 2372 + }, + { + "epoch": 1.3031019613298094, + "grad_norm": 0.328125, + "learning_rate": 0.0001190041614550988, + "loss": 0.7078, + "step": 2373 + }, + { + "epoch": 1.3036583669495063, + "grad_norm": 0.330078125, + "learning_rate": 0.00011894683912672267, + "loss": 0.7599, + "step": 2374 + }, + { + "epoch": 1.3042147725692028, + "grad_norm": 0.328125, + "learning_rate": 0.000118889510340161, + "loss": 0.6686, + "step": 2375 + }, + { + "epoch": 1.3047711781888998, + "grad_norm": 0.337890625, + "learning_rate": 0.00011883217511495481, + "loss": 0.7635, + "step": 2376 + }, + { + "epoch": 1.3053275838085965, + "grad_norm": 0.32421875, + "learning_rate": 0.00011877483347064726, + "loss": 0.6949, + "step": 2377 + }, + { + "epoch": 1.3058839894282932, + "grad_norm": 0.337890625, + "learning_rate": 0.00011871748542678371, + "loss": 0.7317, + "step": 2378 + }, + { + "epoch": 1.30644039504799, + "grad_norm": 0.3359375, + "learning_rate": 0.00011866013100291176, + "loss": 0.8254, + "step": 2379 + }, + { + "epoch": 1.3069968006676866, + "grad_norm": 0.33984375, + "learning_rate": 0.00011860277021858111, + "loss": 0.7949, + "step": 2380 + }, + { + "epoch": 1.3075532062873836, + "grad_norm": 0.333984375, + "learning_rate": 0.00011854540309334363, + "loss": 0.6833, + "step": 2381 + }, + { + "epoch": 1.3081096119070803, + "grad_norm": 0.341796875, + "learning_rate": 0.0001184880296467534, + "loss": 0.8344, + "step": 2382 + }, + { + "epoch": 1.308666017526777, + "grad_norm": 0.33203125, + "learning_rate": 0.00011843064989836662, + "loss": 0.7631, + "step": 2383 + }, + { + "epoch": 1.3092224231464737, + "grad_norm": 0.32421875, + "learning_rate": 0.00011837326386774168, + "loss": 0.7569, + "step": 2384 + }, + { + "epoch": 1.3097788287661705, + "grad_norm": 0.322265625, + "learning_rate": 0.00011831587157443901, + "loss": 0.7218, + "step": 2385 + }, + { + "epoch": 1.3103352343858674, + "grad_norm": 0.341796875, + "learning_rate": 0.00011825847303802129, + "loss": 0.8403, + "step": 2386 + }, + { + "epoch": 1.3108916400055641, + "grad_norm": 0.328125, + "learning_rate": 0.00011820106827805328, + "loss": 0.6815, + "step": 2387 + }, + { + "epoch": 1.3114480456252608, + "grad_norm": 0.330078125, + "learning_rate": 0.00011814365731410188, + "loss": 0.7281, + "step": 2388 + }, + { + "epoch": 1.3120044512449576, + "grad_norm": 0.3359375, + "learning_rate": 0.00011808624016573606, + "loss": 0.7367, + "step": 2389 + }, + { + "epoch": 1.3125608568646543, + "grad_norm": 0.32421875, + "learning_rate": 0.00011802881685252692, + "loss": 0.7303, + "step": 2390 + }, + { + "epoch": 1.313117262484351, + "grad_norm": 0.33203125, + "learning_rate": 0.00011797138739404767, + "loss": 0.7195, + "step": 2391 + }, + { + "epoch": 1.3136736681040477, + "grad_norm": 0.31640625, + "learning_rate": 0.00011791395180987365, + "loss": 0.6809, + "step": 2392 + }, + { + "epoch": 1.3142300737237447, + "grad_norm": 0.306640625, + "learning_rate": 0.00011785651011958223, + "loss": 0.6217, + "step": 2393 + }, + { + "epoch": 1.3147864793434414, + "grad_norm": 0.53125, + "learning_rate": 0.00011779906234275287, + "loss": 0.7547, + "step": 2394 + }, + { + "epoch": 1.3153428849631381, + "grad_norm": 0.314453125, + "learning_rate": 0.00011774160849896707, + "loss": 0.6672, + "step": 2395 + }, + { + "epoch": 1.3158992905828348, + "grad_norm": 0.326171875, + "learning_rate": 0.00011768414860780856, + "loss": 0.7261, + "step": 2396 + }, + { + "epoch": 1.3164556962025316, + "grad_norm": 0.35546875, + "learning_rate": 0.00011762668268886295, + "loss": 0.8287, + "step": 2397 + }, + { + "epoch": 1.3170121018222285, + "grad_norm": 0.32421875, + "learning_rate": 0.00011756921076171795, + "loss": 0.7152, + "step": 2398 + }, + { + "epoch": 1.3175685074419252, + "grad_norm": 0.3359375, + "learning_rate": 0.00011751173284596331, + "loss": 0.7445, + "step": 2399 + }, + { + "epoch": 1.318124913061622, + "grad_norm": 0.333984375, + "learning_rate": 0.00011745424896119095, + "loss": 0.7196, + "step": 2400 + }, + { + "epoch": 1.3186813186813187, + "grad_norm": 0.330078125, + "learning_rate": 0.00011739675912699466, + "loss": 0.708, + "step": 2401 + }, + { + "epoch": 1.3192377243010154, + "grad_norm": 0.322265625, + "learning_rate": 0.00011733926336297032, + "loss": 0.6392, + "step": 2402 + }, + { + "epoch": 1.3197941299207123, + "grad_norm": 0.326171875, + "learning_rate": 0.00011728176168871583, + "loss": 0.7001, + "step": 2403 + }, + { + "epoch": 1.3203505355404088, + "grad_norm": 0.328125, + "learning_rate": 0.00011722425412383115, + "loss": 0.7052, + "step": 2404 + }, + { + "epoch": 1.3209069411601058, + "grad_norm": 0.326171875, + "learning_rate": 0.00011716674068791818, + "loss": 0.6777, + "step": 2405 + }, + { + "epoch": 1.3214633467798025, + "grad_norm": 0.33984375, + "learning_rate": 0.00011710922140058083, + "loss": 0.7806, + "step": 2406 + }, + { + "epoch": 1.3220197523994992, + "grad_norm": 0.310546875, + "learning_rate": 0.000117051696281425, + "loss": 0.6538, + "step": 2407 + }, + { + "epoch": 1.322576158019196, + "grad_norm": 0.33203125, + "learning_rate": 0.00011699416535005869, + "loss": 0.7291, + "step": 2408 + }, + { + "epoch": 1.3231325636388926, + "grad_norm": 0.345703125, + "learning_rate": 0.00011693662862609171, + "loss": 0.8131, + "step": 2409 + }, + { + "epoch": 1.3236889692585896, + "grad_norm": 0.3125, + "learning_rate": 0.00011687908612913594, + "loss": 0.7145, + "step": 2410 + }, + { + "epoch": 1.3242453748782863, + "grad_norm": 0.318359375, + "learning_rate": 0.00011682153787880523, + "loss": 0.715, + "step": 2411 + }, + { + "epoch": 1.324801780497983, + "grad_norm": 0.322265625, + "learning_rate": 0.00011676398389471537, + "loss": 0.7063, + "step": 2412 + }, + { + "epoch": 1.3253581861176797, + "grad_norm": 0.330078125, + "learning_rate": 0.00011670642419648408, + "loss": 0.7678, + "step": 2413 + }, + { + "epoch": 1.3259145917373765, + "grad_norm": 0.349609375, + "learning_rate": 0.00011664885880373109, + "loss": 0.7466, + "step": 2414 + }, + { + "epoch": 1.3264709973570734, + "grad_norm": 0.34375, + "learning_rate": 0.00011659128773607799, + "loss": 0.7625, + "step": 2415 + }, + { + "epoch": 1.3270274029767701, + "grad_norm": 0.337890625, + "learning_rate": 0.0001165337110131484, + "loss": 0.7488, + "step": 2416 + }, + { + "epoch": 1.3275838085964669, + "grad_norm": 0.328125, + "learning_rate": 0.00011647612865456781, + "loss": 0.7182, + "step": 2417 + }, + { + "epoch": 1.3281402142161636, + "grad_norm": 0.345703125, + "learning_rate": 0.0001164185406799636, + "loss": 0.8061, + "step": 2418 + }, + { + "epoch": 1.3286966198358603, + "grad_norm": 0.322265625, + "learning_rate": 0.00011636094710896515, + "loss": 0.6612, + "step": 2419 + }, + { + "epoch": 1.3292530254555572, + "grad_norm": 0.3359375, + "learning_rate": 0.00011630334796120367, + "loss": 0.7305, + "step": 2420 + }, + { + "epoch": 1.3298094310752537, + "grad_norm": 0.3046875, + "learning_rate": 0.00011624574325631234, + "loss": 0.6024, + "step": 2421 + }, + { + "epoch": 1.3303658366949507, + "grad_norm": 0.3359375, + "learning_rate": 0.00011618813301392619, + "loss": 0.7148, + "step": 2422 + }, + { + "epoch": 1.3309222423146474, + "grad_norm": 0.328125, + "learning_rate": 0.00011613051725368207, + "loss": 0.7477, + "step": 2423 + }, + { + "epoch": 1.3314786479343441, + "grad_norm": 0.32421875, + "learning_rate": 0.00011607289599521887, + "loss": 0.7515, + "step": 2424 + }, + { + "epoch": 1.3320350535540408, + "grad_norm": 0.353515625, + "learning_rate": 0.00011601526925817723, + "loss": 0.8687, + "step": 2425 + }, + { + "epoch": 1.3325914591737376, + "grad_norm": 0.328125, + "learning_rate": 0.00011595763706219975, + "loss": 0.6806, + "step": 2426 + }, + { + "epoch": 1.3331478647934345, + "grad_norm": 0.34765625, + "learning_rate": 0.00011589999942693072, + "loss": 0.7935, + "step": 2427 + }, + { + "epoch": 1.3337042704131312, + "grad_norm": 0.314453125, + "learning_rate": 0.0001158423563720165, + "loss": 0.6766, + "step": 2428 + }, + { + "epoch": 1.334260676032828, + "grad_norm": 0.33203125, + "learning_rate": 0.00011578470791710519, + "loss": 0.7671, + "step": 2429 + }, + { + "epoch": 1.3348170816525247, + "grad_norm": 1.0625, + "learning_rate": 0.00011572705408184669, + "loss": 1.6101, + "step": 2430 + }, + { + "epoch": 1.3353734872722214, + "grad_norm": 0.330078125, + "learning_rate": 0.00011566939488589283, + "loss": 0.6784, + "step": 2431 + }, + { + "epoch": 1.3359298928919183, + "grad_norm": 0.337890625, + "learning_rate": 0.00011561173034889717, + "loss": 0.7173, + "step": 2432 + }, + { + "epoch": 1.336486298511615, + "grad_norm": 0.31640625, + "learning_rate": 0.0001155540604905152, + "loss": 0.6968, + "step": 2433 + }, + { + "epoch": 1.3370427041313118, + "grad_norm": 0.361328125, + "learning_rate": 0.00011549638533040412, + "loss": 0.8715, + "step": 2434 + }, + { + "epoch": 1.3375991097510085, + "grad_norm": 0.322265625, + "learning_rate": 0.00011543870488822302, + "loss": 0.7, + "step": 2435 + }, + { + "epoch": 1.3381555153707052, + "grad_norm": 0.3203125, + "learning_rate": 0.0001153810191836327, + "loss": 0.7332, + "step": 2436 + }, + { + "epoch": 1.338711920990402, + "grad_norm": 0.328125, + "learning_rate": 0.00011532332823629583, + "loss": 0.6859, + "step": 2437 + }, + { + "epoch": 1.3392683266100986, + "grad_norm": 0.3359375, + "learning_rate": 0.00011526563206587688, + "loss": 0.6342, + "step": 2438 + }, + { + "epoch": 1.3398247322297956, + "grad_norm": 0.333984375, + "learning_rate": 0.00011520793069204201, + "loss": 0.7854, + "step": 2439 + }, + { + "epoch": 1.3403811378494923, + "grad_norm": 0.3359375, + "learning_rate": 0.00011515022413445924, + "loss": 0.7628, + "step": 2440 + }, + { + "epoch": 1.340937543469189, + "grad_norm": 0.345703125, + "learning_rate": 0.0001150925124127983, + "loss": 0.7843, + "step": 2441 + }, + { + "epoch": 1.3414939490888858, + "grad_norm": 0.322265625, + "learning_rate": 0.00011503479554673072, + "loss": 0.7167, + "step": 2442 + }, + { + "epoch": 1.3420503547085825, + "grad_norm": 0.32421875, + "learning_rate": 0.00011497707355592975, + "loss": 0.7199, + "step": 2443 + }, + { + "epoch": 1.3426067603282794, + "grad_norm": 0.34765625, + "learning_rate": 0.0001149193464600704, + "loss": 0.7374, + "step": 2444 + }, + { + "epoch": 1.3431631659479761, + "grad_norm": 0.3203125, + "learning_rate": 0.00011486161427882946, + "loss": 0.6563, + "step": 2445 + }, + { + "epoch": 1.3437195715676729, + "grad_norm": 0.31640625, + "learning_rate": 0.0001148038770318854, + "loss": 0.699, + "step": 2446 + }, + { + "epoch": 1.3442759771873696, + "grad_norm": 0.310546875, + "learning_rate": 0.00011474613473891842, + "loss": 0.5989, + "step": 2447 + }, + { + "epoch": 1.3448323828070663, + "grad_norm": 0.345703125, + "learning_rate": 0.00011468838741961045, + "loss": 0.6989, + "step": 2448 + }, + { + "epoch": 1.3453887884267632, + "grad_norm": 0.322265625, + "learning_rate": 0.00011463063509364517, + "loss": 0.6952, + "step": 2449 + }, + { + "epoch": 1.3459451940464597, + "grad_norm": 0.333984375, + "learning_rate": 0.00011457287778070793, + "loss": 0.7021, + "step": 2450 + }, + { + "epoch": 1.3465015996661567, + "grad_norm": 0.33203125, + "learning_rate": 0.00011451511550048577, + "loss": 0.7343, + "step": 2451 + }, + { + "epoch": 1.3470580052858534, + "grad_norm": 0.345703125, + "learning_rate": 0.00011445734827266738, + "loss": 0.8033, + "step": 2452 + }, + { + "epoch": 1.3476144109055501, + "grad_norm": 0.31640625, + "learning_rate": 0.00011439957611694332, + "loss": 0.6597, + "step": 2453 + }, + { + "epoch": 1.3481708165252468, + "grad_norm": 0.337890625, + "learning_rate": 0.00011434179905300559, + "loss": 0.7599, + "step": 2454 + }, + { + "epoch": 1.3487272221449436, + "grad_norm": 0.328125, + "learning_rate": 0.00011428401710054806, + "loss": 0.7468, + "step": 2455 + }, + { + "epoch": 1.3492836277646405, + "grad_norm": 0.33203125, + "learning_rate": 0.00011422623027926612, + "loss": 0.712, + "step": 2456 + }, + { + "epoch": 1.3498400333843372, + "grad_norm": 0.3203125, + "learning_rate": 0.00011416843860885688, + "loss": 0.6616, + "step": 2457 + }, + { + "epoch": 1.350396439004034, + "grad_norm": 0.322265625, + "learning_rate": 0.00011411064210901916, + "loss": 0.7084, + "step": 2458 + }, + { + "epoch": 1.3509528446237307, + "grad_norm": 0.333984375, + "learning_rate": 0.00011405284079945335, + "loss": 0.8171, + "step": 2459 + }, + { + "epoch": 1.3515092502434274, + "grad_norm": 0.35546875, + "learning_rate": 0.00011399503469986144, + "loss": 0.8589, + "step": 2460 + }, + { + "epoch": 1.3520656558631243, + "grad_norm": 0.318359375, + "learning_rate": 0.0001139372238299472, + "loss": 0.6904, + "step": 2461 + }, + { + "epoch": 1.352622061482821, + "grad_norm": 0.337890625, + "learning_rate": 0.0001138794082094159, + "loss": 0.7881, + "step": 2462 + }, + { + "epoch": 1.3531784671025178, + "grad_norm": 0.33984375, + "learning_rate": 0.00011382158785797448, + "loss": 0.7395, + "step": 2463 + }, + { + "epoch": 1.3537348727222145, + "grad_norm": 0.33203125, + "learning_rate": 0.00011376376279533143, + "loss": 0.7064, + "step": 2464 + }, + { + "epoch": 1.3542912783419112, + "grad_norm": 0.322265625, + "learning_rate": 0.000113705933041197, + "loss": 0.7418, + "step": 2465 + }, + { + "epoch": 1.354847683961608, + "grad_norm": 0.298828125, + "learning_rate": 0.00011364809861528283, + "loss": 0.585, + "step": 2466 + }, + { + "epoch": 1.3554040895813046, + "grad_norm": 0.330078125, + "learning_rate": 0.00011359025953730231, + "loss": 0.7384, + "step": 2467 + }, + { + "epoch": 1.3559604952010016, + "grad_norm": 0.330078125, + "learning_rate": 0.00011353241582697039, + "loss": 0.6822, + "step": 2468 + }, + { + "epoch": 1.3565169008206983, + "grad_norm": 0.328125, + "learning_rate": 0.00011347456750400351, + "loss": 0.7019, + "step": 2469 + }, + { + "epoch": 1.357073306440395, + "grad_norm": 0.333984375, + "learning_rate": 0.00011341671458811978, + "loss": 0.7273, + "step": 2470 + }, + { + "epoch": 1.3576297120600918, + "grad_norm": 0.3203125, + "learning_rate": 0.00011335885709903885, + "loss": 0.7323, + "step": 2471 + }, + { + "epoch": 1.3581861176797885, + "grad_norm": 0.330078125, + "learning_rate": 0.0001133009950564819, + "loss": 0.704, + "step": 2472 + }, + { + "epoch": 1.3587425232994854, + "grad_norm": 0.302734375, + "learning_rate": 0.00011324312848017166, + "loss": 0.6183, + "step": 2473 + }, + { + "epoch": 1.3592989289191821, + "grad_norm": 0.3359375, + "learning_rate": 0.0001131852573898325, + "loss": 0.7105, + "step": 2474 + }, + { + "epoch": 1.3598553345388789, + "grad_norm": 0.322265625, + "learning_rate": 0.00011312738180519018, + "loss": 0.6982, + "step": 2475 + }, + { + "epoch": 1.3604117401585756, + "grad_norm": 0.29296875, + "learning_rate": 0.00011306950174597213, + "loss": 0.5876, + "step": 2476 + }, + { + "epoch": 1.3609681457782723, + "grad_norm": 0.32421875, + "learning_rate": 0.00011301161723190717, + "loss": 0.7142, + "step": 2477 + }, + { + "epoch": 1.3615245513979692, + "grad_norm": 0.333984375, + "learning_rate": 0.00011295372828272577, + "loss": 0.7409, + "step": 2478 + }, + { + "epoch": 1.362080957017666, + "grad_norm": 0.33203125, + "learning_rate": 0.00011289583491815985, + "loss": 0.7466, + "step": 2479 + }, + { + "epoch": 1.3626373626373627, + "grad_norm": 0.3125, + "learning_rate": 0.00011283793715794282, + "loss": 0.6091, + "step": 2480 + }, + { + "epoch": 1.3631937682570594, + "grad_norm": 0.32421875, + "learning_rate": 0.0001127800350218096, + "loss": 0.7041, + "step": 2481 + }, + { + "epoch": 1.3637501738767561, + "grad_norm": 0.31640625, + "learning_rate": 0.00011272212852949664, + "loss": 0.6647, + "step": 2482 + }, + { + "epoch": 1.3643065794964528, + "grad_norm": 0.33203125, + "learning_rate": 0.0001126642177007418, + "loss": 0.7769, + "step": 2483 + }, + { + "epoch": 1.3648629851161496, + "grad_norm": 0.30859375, + "learning_rate": 0.00011260630255528451, + "loss": 0.6792, + "step": 2484 + }, + { + "epoch": 1.3654193907358465, + "grad_norm": 0.34375, + "learning_rate": 0.00011254838311286558, + "loss": 0.8129, + "step": 2485 + }, + { + "epoch": 1.3659757963555432, + "grad_norm": 0.337890625, + "learning_rate": 0.00011249045939322736, + "loss": 0.7495, + "step": 2486 + }, + { + "epoch": 1.36653220197524, + "grad_norm": 0.328125, + "learning_rate": 0.00011243253141611358, + "loss": 0.7184, + "step": 2487 + }, + { + "epoch": 1.3670886075949367, + "grad_norm": 0.318359375, + "learning_rate": 0.00011237459920126951, + "loss": 0.6775, + "step": 2488 + }, + { + "epoch": 1.3676450132146334, + "grad_norm": 0.3359375, + "learning_rate": 0.00011231666276844177, + "loss": 0.6989, + "step": 2489 + }, + { + "epoch": 1.3682014188343303, + "grad_norm": 0.322265625, + "learning_rate": 0.00011225872213737854, + "loss": 0.6135, + "step": 2490 + }, + { + "epoch": 1.368757824454027, + "grad_norm": 0.3359375, + "learning_rate": 0.00011220077732782932, + "loss": 0.782, + "step": 2491 + }, + { + "epoch": 1.3693142300737238, + "grad_norm": 0.314453125, + "learning_rate": 0.00011214282835954509, + "loss": 0.6541, + "step": 2492 + }, + { + "epoch": 1.3698706356934205, + "grad_norm": 0.322265625, + "learning_rate": 0.00011208487525227816, + "loss": 0.6114, + "step": 2493 + }, + { + "epoch": 1.3704270413131172, + "grad_norm": 0.32421875, + "learning_rate": 0.00011202691802578243, + "loss": 0.7009, + "step": 2494 + }, + { + "epoch": 1.3709834469328142, + "grad_norm": 0.3203125, + "learning_rate": 0.00011196895669981302, + "loss": 0.6033, + "step": 2495 + }, + { + "epoch": 1.3715398525525107, + "grad_norm": 0.33203125, + "learning_rate": 0.00011191099129412657, + "loss": 0.7491, + "step": 2496 + }, + { + "epoch": 1.3720962581722076, + "grad_norm": 0.326171875, + "learning_rate": 0.00011185302182848101, + "loss": 0.7113, + "step": 2497 + }, + { + "epoch": 1.3726526637919043, + "grad_norm": 0.330078125, + "learning_rate": 0.00011179504832263577, + "loss": 0.6992, + "step": 2498 + }, + { + "epoch": 1.373209069411601, + "grad_norm": 0.33203125, + "learning_rate": 0.00011173707079635156, + "loss": 0.7619, + "step": 2499 + }, + { + "epoch": 1.3737654750312978, + "grad_norm": 0.33984375, + "learning_rate": 0.0001116790892693905, + "loss": 0.7508, + "step": 2500 + }, + { + "epoch": 1.3743218806509945, + "grad_norm": 0.328125, + "learning_rate": 0.00011162110376151604, + "loss": 0.7367, + "step": 2501 + }, + { + "epoch": 1.3748782862706914, + "grad_norm": 0.33203125, + "learning_rate": 0.00011156311429249306, + "loss": 0.7643, + "step": 2502 + }, + { + "epoch": 1.3754346918903881, + "grad_norm": 0.345703125, + "learning_rate": 0.00011150512088208776, + "loss": 0.793, + "step": 2503 + }, + { + "epoch": 1.3759910975100849, + "grad_norm": 0.33203125, + "learning_rate": 0.00011144712355006762, + "loss": 0.7647, + "step": 2504 + }, + { + "epoch": 1.3765475031297816, + "grad_norm": 0.3203125, + "learning_rate": 0.00011138912231620153, + "loss": 0.672, + "step": 2505 + }, + { + "epoch": 1.3771039087494783, + "grad_norm": 0.32421875, + "learning_rate": 0.00011133111720025966, + "loss": 0.7259, + "step": 2506 + }, + { + "epoch": 1.3776603143691752, + "grad_norm": 0.3046875, + "learning_rate": 0.00011127310822201358, + "loss": 0.5997, + "step": 2507 + }, + { + "epoch": 1.378216719988872, + "grad_norm": 0.322265625, + "learning_rate": 0.00011121509540123609, + "loss": 0.7158, + "step": 2508 + }, + { + "epoch": 1.3787731256085687, + "grad_norm": 0.333984375, + "learning_rate": 0.00011115707875770134, + "loss": 0.7551, + "step": 2509 + }, + { + "epoch": 1.3793295312282654, + "grad_norm": 0.326171875, + "learning_rate": 0.0001110990583111848, + "loss": 0.7265, + "step": 2510 + }, + { + "epoch": 1.3798859368479621, + "grad_norm": 0.34765625, + "learning_rate": 0.00011104103408146318, + "loss": 0.8133, + "step": 2511 + }, + { + "epoch": 1.3804423424676588, + "grad_norm": 0.349609375, + "learning_rate": 0.00011098300608831454, + "loss": 0.7369, + "step": 2512 + }, + { + "epoch": 1.3809987480873556, + "grad_norm": 0.302734375, + "learning_rate": 0.00011092497435151818, + "loss": 0.5932, + "step": 2513 + }, + { + "epoch": 1.3815551537070525, + "grad_norm": 0.322265625, + "learning_rate": 0.00011086693889085473, + "loss": 0.6975, + "step": 2514 + }, + { + "epoch": 1.3821115593267492, + "grad_norm": 0.328125, + "learning_rate": 0.00011080889972610602, + "loss": 0.6966, + "step": 2515 + }, + { + "epoch": 1.382667964946446, + "grad_norm": 0.3359375, + "learning_rate": 0.00011075085687705517, + "loss": 0.7353, + "step": 2516 + }, + { + "epoch": 1.3832243705661427, + "grad_norm": 0.3125, + "learning_rate": 0.0001106928103634866, + "loss": 0.6437, + "step": 2517 + }, + { + "epoch": 1.3837807761858394, + "grad_norm": 0.31640625, + "learning_rate": 0.00011063476020518588, + "loss": 0.661, + "step": 2518 + }, + { + "epoch": 1.3843371818055363, + "grad_norm": 0.34375, + "learning_rate": 0.00011057670642193994, + "loss": 0.7933, + "step": 2519 + }, + { + "epoch": 1.384893587425233, + "grad_norm": 0.318359375, + "learning_rate": 0.00011051864903353686, + "loss": 0.6854, + "step": 2520 + }, + { + "epoch": 1.3854499930449298, + "grad_norm": 0.333984375, + "learning_rate": 0.00011046058805976599, + "loss": 0.7005, + "step": 2521 + }, + { + "epoch": 1.3860063986646265, + "grad_norm": 0.326171875, + "learning_rate": 0.00011040252352041782, + "loss": 0.6934, + "step": 2522 + }, + { + "epoch": 1.3865628042843232, + "grad_norm": 0.32421875, + "learning_rate": 0.00011034445543528424, + "loss": 0.6935, + "step": 2523 + }, + { + "epoch": 1.3871192099040202, + "grad_norm": 0.333984375, + "learning_rate": 0.00011028638382415819, + "loss": 0.7408, + "step": 2524 + }, + { + "epoch": 1.3876756155237167, + "grad_norm": 0.330078125, + "learning_rate": 0.00011022830870683382, + "loss": 0.7131, + "step": 2525 + }, + { + "epoch": 1.3882320211434136, + "grad_norm": 0.341796875, + "learning_rate": 0.00011017023010310654, + "loss": 0.7135, + "step": 2526 + }, + { + "epoch": 1.3887884267631103, + "grad_norm": 0.330078125, + "learning_rate": 0.00011011214803277292, + "loss": 0.7208, + "step": 2527 + }, + { + "epoch": 1.389344832382807, + "grad_norm": 0.33203125, + "learning_rate": 0.00011005406251563068, + "loss": 0.7367, + "step": 2528 + }, + { + "epoch": 1.3899012380025038, + "grad_norm": 0.306640625, + "learning_rate": 0.0001099959735714788, + "loss": 0.6008, + "step": 2529 + }, + { + "epoch": 1.3904576436222005, + "grad_norm": 0.30859375, + "learning_rate": 0.0001099378812201173, + "loss": 0.6261, + "step": 2530 + }, + { + "epoch": 1.3910140492418974, + "grad_norm": 0.337890625, + "learning_rate": 0.0001098797854813475, + "loss": 0.7666, + "step": 2531 + }, + { + "epoch": 1.3915704548615941, + "grad_norm": 0.326171875, + "learning_rate": 0.00010982168637497177, + "loss": 0.6456, + "step": 2532 + }, + { + "epoch": 1.3921268604812909, + "grad_norm": 0.345703125, + "learning_rate": 0.00010976358392079369, + "loss": 0.7901, + "step": 2533 + }, + { + "epoch": 1.3926832661009876, + "grad_norm": 0.326171875, + "learning_rate": 0.00010970547813861794, + "loss": 0.7, + "step": 2534 + }, + { + "epoch": 1.3932396717206843, + "grad_norm": 0.3125, + "learning_rate": 0.00010964736904825035, + "loss": 0.6445, + "step": 2535 + }, + { + "epoch": 1.3937960773403812, + "grad_norm": 0.326171875, + "learning_rate": 0.00010958925666949791, + "loss": 0.6889, + "step": 2536 + }, + { + "epoch": 1.394352482960078, + "grad_norm": 0.33203125, + "learning_rate": 0.00010953114102216868, + "loss": 0.687, + "step": 2537 + }, + { + "epoch": 1.3949088885797747, + "grad_norm": 0.322265625, + "learning_rate": 0.00010947302212607186, + "loss": 0.658, + "step": 2538 + }, + { + "epoch": 1.3954652941994714, + "grad_norm": 0.330078125, + "learning_rate": 0.00010941490000101773, + "loss": 0.7327, + "step": 2539 + }, + { + "epoch": 1.3960216998191681, + "grad_norm": 0.32421875, + "learning_rate": 0.00010935677466681773, + "loss": 0.6706, + "step": 2540 + }, + { + "epoch": 1.396578105438865, + "grad_norm": 0.333984375, + "learning_rate": 0.00010929864614328435, + "loss": 0.7794, + "step": 2541 + }, + { + "epoch": 1.3971345110585616, + "grad_norm": 0.33984375, + "learning_rate": 0.00010924051445023116, + "loss": 0.75, + "step": 2542 + }, + { + "epoch": 1.3976909166782585, + "grad_norm": 0.322265625, + "learning_rate": 0.00010918237960747281, + "loss": 0.7045, + "step": 2543 + }, + { + "epoch": 1.3982473222979552, + "grad_norm": 0.322265625, + "learning_rate": 0.00010912424163482506, + "loss": 0.6658, + "step": 2544 + }, + { + "epoch": 1.398803727917652, + "grad_norm": 0.34765625, + "learning_rate": 0.00010906610055210473, + "loss": 0.8038, + "step": 2545 + }, + { + "epoch": 1.3993601335373487, + "grad_norm": 0.32421875, + "learning_rate": 0.00010900795637912967, + "loss": 0.6971, + "step": 2546 + }, + { + "epoch": 1.3999165391570454, + "grad_norm": 0.337890625, + "learning_rate": 0.00010894980913571874, + "loss": 0.7513, + "step": 2547 + }, + { + "epoch": 1.4004729447767423, + "grad_norm": 0.341796875, + "learning_rate": 0.00010889165884169199, + "loss": 0.7808, + "step": 2548 + }, + { + "epoch": 1.401029350396439, + "grad_norm": 0.322265625, + "learning_rate": 0.00010883350551687042, + "loss": 0.668, + "step": 2549 + }, + { + "epoch": 1.4015857560161358, + "grad_norm": 0.330078125, + "learning_rate": 0.00010877534918107601, + "loss": 0.6883, + "step": 2550 + }, + { + "epoch": 1.4021421616358325, + "grad_norm": 0.333984375, + "learning_rate": 0.00010871718985413184, + "loss": 0.7308, + "step": 2551 + }, + { + "epoch": 1.4026985672555292, + "grad_norm": 0.326171875, + "learning_rate": 0.00010865902755586204, + "loss": 0.7401, + "step": 2552 + }, + { + "epoch": 1.4032549728752262, + "grad_norm": 0.328125, + "learning_rate": 0.00010860086230609168, + "loss": 0.6671, + "step": 2553 + }, + { + "epoch": 1.4038113784949229, + "grad_norm": 0.349609375, + "learning_rate": 0.00010854269412464683, + "loss": 0.787, + "step": 2554 + }, + { + "epoch": 1.4043677841146196, + "grad_norm": 0.328125, + "learning_rate": 0.00010848452303135464, + "loss": 0.7039, + "step": 2555 + }, + { + "epoch": 1.4049241897343163, + "grad_norm": 0.330078125, + "learning_rate": 0.00010842634904604319, + "loss": 0.7197, + "step": 2556 + }, + { + "epoch": 1.405480595354013, + "grad_norm": 0.3203125, + "learning_rate": 0.00010836817218854155, + "loss": 0.6888, + "step": 2557 + }, + { + "epoch": 1.4060370009737098, + "grad_norm": 0.34375, + "learning_rate": 0.00010830999247867976, + "loss": 0.7686, + "step": 2558 + }, + { + "epoch": 1.4065934065934065, + "grad_norm": 0.33984375, + "learning_rate": 0.00010825180993628887, + "loss": 0.7198, + "step": 2559 + }, + { + "epoch": 1.4071498122131034, + "grad_norm": 0.3359375, + "learning_rate": 0.0001081936245812009, + "loss": 0.7277, + "step": 2560 + }, + { + "epoch": 1.4077062178328001, + "grad_norm": 0.337890625, + "learning_rate": 0.00010813543643324879, + "loss": 0.7303, + "step": 2561 + }, + { + "epoch": 1.4082626234524969, + "grad_norm": 0.32421875, + "learning_rate": 0.00010807724551226643, + "loss": 0.6663, + "step": 2562 + }, + { + "epoch": 1.4088190290721936, + "grad_norm": 0.341796875, + "learning_rate": 0.0001080190518380887, + "loss": 0.7747, + "step": 2563 + }, + { + "epoch": 1.4093754346918903, + "grad_norm": 0.328125, + "learning_rate": 0.00010796085543055136, + "loss": 0.7035, + "step": 2564 + }, + { + "epoch": 1.4099318403115872, + "grad_norm": 0.32421875, + "learning_rate": 0.00010790265630949117, + "loss": 0.6838, + "step": 2565 + }, + { + "epoch": 1.410488245931284, + "grad_norm": 0.328125, + "learning_rate": 0.00010784445449474576, + "loss": 0.725, + "step": 2566 + }, + { + "epoch": 1.4110446515509807, + "grad_norm": 0.328125, + "learning_rate": 0.00010778625000615365, + "loss": 0.7595, + "step": 2567 + }, + { + "epoch": 1.4116010571706774, + "grad_norm": 0.337890625, + "learning_rate": 0.0001077280428635544, + "loss": 0.6964, + "step": 2568 + }, + { + "epoch": 1.4121574627903741, + "grad_norm": 0.330078125, + "learning_rate": 0.00010766983308678837, + "loss": 0.7032, + "step": 2569 + }, + { + "epoch": 1.412713868410071, + "grad_norm": 0.3359375, + "learning_rate": 0.00010761162069569681, + "loss": 0.7463, + "step": 2570 + }, + { + "epoch": 1.4132702740297676, + "grad_norm": 0.322265625, + "learning_rate": 0.00010755340571012189, + "loss": 0.6686, + "step": 2571 + }, + { + "epoch": 1.4138266796494645, + "grad_norm": 0.330078125, + "learning_rate": 0.00010749518814990669, + "loss": 0.7457, + "step": 2572 + }, + { + "epoch": 1.4143830852691612, + "grad_norm": 0.33203125, + "learning_rate": 0.00010743696803489514, + "loss": 0.6991, + "step": 2573 + }, + { + "epoch": 1.414939490888858, + "grad_norm": 0.341796875, + "learning_rate": 0.00010737874538493203, + "loss": 0.7904, + "step": 2574 + }, + { + "epoch": 1.4154958965085547, + "grad_norm": 0.31640625, + "learning_rate": 0.00010732052021986303, + "loss": 0.6694, + "step": 2575 + }, + { + "epoch": 1.4160523021282514, + "grad_norm": 0.330078125, + "learning_rate": 0.00010726229255953467, + "loss": 0.7395, + "step": 2576 + }, + { + "epoch": 1.4166087077479483, + "grad_norm": 0.34375, + "learning_rate": 0.00010720406242379432, + "loss": 0.7902, + "step": 2577 + }, + { + "epoch": 1.417165113367645, + "grad_norm": 0.33984375, + "learning_rate": 0.00010714582983249022, + "loss": 0.7543, + "step": 2578 + }, + { + "epoch": 1.4177215189873418, + "grad_norm": 0.34765625, + "learning_rate": 0.0001070875948054714, + "loss": 0.8445, + "step": 2579 + }, + { + "epoch": 1.4182779246070385, + "grad_norm": 0.32421875, + "learning_rate": 0.00010702935736258774, + "loss": 0.6908, + "step": 2580 + }, + { + "epoch": 1.4188343302267352, + "grad_norm": 0.326171875, + "learning_rate": 0.00010697111752368997, + "loss": 0.7227, + "step": 2581 + }, + { + "epoch": 1.4193907358464322, + "grad_norm": 0.3125, + "learning_rate": 0.00010691287530862965, + "loss": 0.6442, + "step": 2582 + }, + { + "epoch": 1.4199471414661289, + "grad_norm": 0.306640625, + "learning_rate": 0.00010685463073725906, + "loss": 0.6823, + "step": 2583 + }, + { + "epoch": 1.4205035470858256, + "grad_norm": 0.328125, + "learning_rate": 0.00010679638382943136, + "loss": 0.6991, + "step": 2584 + }, + { + "epoch": 1.4210599527055223, + "grad_norm": 0.296875, + "learning_rate": 0.00010673813460500048, + "loss": 0.5955, + "step": 2585 + }, + { + "epoch": 1.421616358325219, + "grad_norm": 0.337890625, + "learning_rate": 0.00010667988308382117, + "loss": 0.7899, + "step": 2586 + }, + { + "epoch": 1.4221727639449158, + "grad_norm": 0.31640625, + "learning_rate": 0.00010662162928574892, + "loss": 0.6611, + "step": 2587 + }, + { + "epoch": 1.4227291695646125, + "grad_norm": 0.33984375, + "learning_rate": 0.00010656337323063999, + "loss": 0.7835, + "step": 2588 + }, + { + "epoch": 1.4232855751843094, + "grad_norm": 0.31640625, + "learning_rate": 0.00010650511493835149, + "loss": 0.6867, + "step": 2589 + }, + { + "epoch": 1.4238419808040061, + "grad_norm": 0.33203125, + "learning_rate": 0.00010644685442874118, + "loss": 0.7198, + "step": 2590 + }, + { + "epoch": 1.4243983864237029, + "grad_norm": 0.333984375, + "learning_rate": 0.00010638859172166769, + "loss": 0.6852, + "step": 2591 + }, + { + "epoch": 1.4249547920433996, + "grad_norm": 0.3203125, + "learning_rate": 0.00010633032683699029, + "loss": 0.6569, + "step": 2592 + }, + { + "epoch": 1.4255111976630963, + "grad_norm": 0.333984375, + "learning_rate": 0.00010627205979456906, + "loss": 0.6837, + "step": 2593 + }, + { + "epoch": 1.4260676032827933, + "grad_norm": 1.5234375, + "learning_rate": 0.00010621379061426485, + "loss": 1.7361, + "step": 2594 + }, + { + "epoch": 1.42662400890249, + "grad_norm": 0.3203125, + "learning_rate": 0.0001061555193159391, + "loss": 0.6271, + "step": 2595 + }, + { + "epoch": 1.4271804145221867, + "grad_norm": 0.322265625, + "learning_rate": 0.0001060972459194541, + "loss": 0.6884, + "step": 2596 + }, + { + "epoch": 1.4277368201418834, + "grad_norm": 0.3125, + "learning_rate": 0.00010603897044467284, + "loss": 0.6821, + "step": 2597 + }, + { + "epoch": 1.4282932257615801, + "grad_norm": 0.33984375, + "learning_rate": 0.00010598069291145896, + "loss": 0.7808, + "step": 2598 + }, + { + "epoch": 1.428849631381277, + "grad_norm": 0.333984375, + "learning_rate": 0.00010592241333967685, + "loss": 0.7096, + "step": 2599 + }, + { + "epoch": 1.4294060370009736, + "grad_norm": 0.33203125, + "learning_rate": 0.00010586413174919158, + "loss": 0.6982, + "step": 2600 + }, + { + "epoch": 1.4299624426206705, + "grad_norm": 0.310546875, + "learning_rate": 0.00010580584815986891, + "loss": 0.6833, + "step": 2601 + }, + { + "epoch": 1.4305188482403672, + "grad_norm": 0.337890625, + "learning_rate": 0.00010574756259157526, + "loss": 0.7476, + "step": 2602 + }, + { + "epoch": 1.431075253860064, + "grad_norm": 0.34765625, + "learning_rate": 0.00010568927506417779, + "loss": 0.7539, + "step": 2603 + }, + { + "epoch": 1.4316316594797607, + "grad_norm": 0.33203125, + "learning_rate": 0.00010563098559754419, + "loss": 0.7264, + "step": 2604 + }, + { + "epoch": 1.4321880650994574, + "grad_norm": 0.326171875, + "learning_rate": 0.00010557269421154301, + "loss": 0.7798, + "step": 2605 + }, + { + "epoch": 1.4327444707191543, + "grad_norm": 0.357421875, + "learning_rate": 0.00010551440092604328, + "loss": 0.7869, + "step": 2606 + }, + { + "epoch": 1.433300876338851, + "grad_norm": 0.330078125, + "learning_rate": 0.00010545610576091478, + "loss": 0.6997, + "step": 2607 + }, + { + "epoch": 1.4338572819585478, + "grad_norm": 0.34375, + "learning_rate": 0.00010539780873602785, + "loss": 0.7397, + "step": 2608 + }, + { + "epoch": 1.4344136875782445, + "grad_norm": 0.3203125, + "learning_rate": 0.00010533950987125356, + "loss": 0.6898, + "step": 2609 + }, + { + "epoch": 1.4349700931979412, + "grad_norm": 0.326171875, + "learning_rate": 0.00010528120918646353, + "loss": 0.6294, + "step": 2610 + }, + { + "epoch": 1.4355264988176382, + "grad_norm": 0.306640625, + "learning_rate": 0.00010522290670153005, + "loss": 0.604, + "step": 2611 + }, + { + "epoch": 1.4360829044373349, + "grad_norm": 0.326171875, + "learning_rate": 0.00010516460243632597, + "loss": 0.7194, + "step": 2612 + }, + { + "epoch": 1.4366393100570316, + "grad_norm": 0.34765625, + "learning_rate": 0.0001051062964107248, + "loss": 0.7471, + "step": 2613 + }, + { + "epoch": 1.4371957156767283, + "grad_norm": 0.341796875, + "learning_rate": 0.00010504798864460062, + "loss": 0.7736, + "step": 2614 + }, + { + "epoch": 1.437752121296425, + "grad_norm": 0.341796875, + "learning_rate": 0.00010498967915782809, + "loss": 0.7772, + "step": 2615 + }, + { + "epoch": 1.438308526916122, + "grad_norm": 0.34765625, + "learning_rate": 0.00010493136797028249, + "loss": 0.7654, + "step": 2616 + }, + { + "epoch": 1.4388649325358185, + "grad_norm": 0.33203125, + "learning_rate": 0.00010487305510183967, + "loss": 0.7171, + "step": 2617 + }, + { + "epoch": 1.4394213381555154, + "grad_norm": 0.326171875, + "learning_rate": 0.00010481474057237604, + "loss": 0.7437, + "step": 2618 + }, + { + "epoch": 1.4399777437752121, + "grad_norm": 0.337890625, + "learning_rate": 0.00010475642440176861, + "loss": 0.6849, + "step": 2619 + }, + { + "epoch": 1.4405341493949089, + "grad_norm": 0.3125, + "learning_rate": 0.00010469810660989492, + "loss": 0.6591, + "step": 2620 + }, + { + "epoch": 1.4410905550146056, + "grad_norm": 0.353515625, + "learning_rate": 0.00010463978721663301, + "loss": 0.7568, + "step": 2621 + }, + { + "epoch": 1.4416469606343023, + "grad_norm": 0.3203125, + "learning_rate": 0.00010458146624186163, + "loss": 0.6783, + "step": 2622 + }, + { + "epoch": 1.4422033662539993, + "grad_norm": 0.326171875, + "learning_rate": 0.00010452314370545987, + "loss": 0.7343, + "step": 2623 + }, + { + "epoch": 1.442759771873696, + "grad_norm": 0.3203125, + "learning_rate": 0.00010446481962730749, + "loss": 0.6946, + "step": 2624 + }, + { + "epoch": 1.4433161774933927, + "grad_norm": 0.333984375, + "learning_rate": 0.0001044064940272847, + "loss": 0.7157, + "step": 2625 + }, + { + "epoch": 1.4438725831130894, + "grad_norm": 0.330078125, + "learning_rate": 0.0001043481669252723, + "loss": 0.7112, + "step": 2626 + }, + { + "epoch": 1.4444289887327861, + "grad_norm": 0.337890625, + "learning_rate": 0.00010428983834115153, + "loss": 0.7378, + "step": 2627 + }, + { + "epoch": 1.444985394352483, + "grad_norm": 0.333984375, + "learning_rate": 0.0001042315082948042, + "loss": 0.6738, + "step": 2628 + }, + { + "epoch": 1.4455417999721798, + "grad_norm": 0.330078125, + "learning_rate": 0.00010417317680611254, + "loss": 0.692, + "step": 2629 + }, + { + "epoch": 1.4460982055918765, + "grad_norm": 0.32421875, + "learning_rate": 0.00010411484389495934, + "loss": 0.6794, + "step": 2630 + }, + { + "epoch": 1.4466546112115732, + "grad_norm": 0.34765625, + "learning_rate": 0.00010405650958122786, + "loss": 0.8288, + "step": 2631 + }, + { + "epoch": 1.44721101683127, + "grad_norm": 0.31640625, + "learning_rate": 0.0001039981738848018, + "loss": 0.6557, + "step": 2632 + }, + { + "epoch": 1.4477674224509667, + "grad_norm": 0.318359375, + "learning_rate": 0.00010393983682556537, + "loss": 0.6802, + "step": 2633 + }, + { + "epoch": 1.4483238280706634, + "grad_norm": 0.3359375, + "learning_rate": 0.00010388149842340325, + "loss": 0.7585, + "step": 2634 + }, + { + "epoch": 1.4488802336903603, + "grad_norm": 0.333984375, + "learning_rate": 0.00010382315869820056, + "loss": 0.7667, + "step": 2635 + }, + { + "epoch": 1.449436639310057, + "grad_norm": 0.333984375, + "learning_rate": 0.00010376481766984288, + "loss": 0.7618, + "step": 2636 + }, + { + "epoch": 1.4499930449297538, + "grad_norm": 0.330078125, + "learning_rate": 0.00010370647535821618, + "loss": 0.7417, + "step": 2637 + }, + { + "epoch": 1.4505494505494505, + "grad_norm": 0.326171875, + "learning_rate": 0.00010364813178320698, + "loss": 0.7083, + "step": 2638 + }, + { + "epoch": 1.4511058561691472, + "grad_norm": 0.326171875, + "learning_rate": 0.00010358978696470214, + "loss": 0.7089, + "step": 2639 + }, + { + "epoch": 1.4516622617888442, + "grad_norm": 0.333984375, + "learning_rate": 0.00010353144092258898, + "loss": 0.7021, + "step": 2640 + }, + { + "epoch": 1.4522186674085409, + "grad_norm": 0.322265625, + "learning_rate": 0.00010347309367675517, + "loss": 0.7072, + "step": 2641 + }, + { + "epoch": 1.4527750730282376, + "grad_norm": 0.33984375, + "learning_rate": 0.00010341474524708896, + "loss": 0.7622, + "step": 2642 + }, + { + "epoch": 1.4533314786479343, + "grad_norm": 0.31640625, + "learning_rate": 0.00010335639565347882, + "loss": 0.6601, + "step": 2643 + }, + { + "epoch": 1.453887884267631, + "grad_norm": 0.33984375, + "learning_rate": 0.00010329804491581365, + "loss": 0.7328, + "step": 2644 + }, + { + "epoch": 1.454444289887328, + "grad_norm": 0.32421875, + "learning_rate": 0.00010323969305398282, + "loss": 0.7097, + "step": 2645 + }, + { + "epoch": 1.4550006955070245, + "grad_norm": 0.3203125, + "learning_rate": 0.00010318134008787606, + "loss": 0.6524, + "step": 2646 + }, + { + "epoch": 1.4555571011267214, + "grad_norm": 0.31640625, + "learning_rate": 0.00010312298603738342, + "loss": 0.6303, + "step": 2647 + }, + { + "epoch": 1.4561135067464182, + "grad_norm": 0.326171875, + "learning_rate": 0.0001030646309223954, + "loss": 0.708, + "step": 2648 + }, + { + "epoch": 1.4566699123661149, + "grad_norm": 0.3359375, + "learning_rate": 0.00010300627476280278, + "loss": 0.7317, + "step": 2649 + }, + { + "epoch": 1.4572263179858116, + "grad_norm": 0.322265625, + "learning_rate": 0.00010294791757849673, + "loss": 0.703, + "step": 2650 + }, + { + "epoch": 1.4577827236055083, + "grad_norm": 0.353515625, + "learning_rate": 0.00010288955938936882, + "loss": 0.8404, + "step": 2651 + }, + { + "epoch": 1.4583391292252053, + "grad_norm": 0.3515625, + "learning_rate": 0.00010283120021531087, + "loss": 0.7021, + "step": 2652 + }, + { + "epoch": 1.458895534844902, + "grad_norm": 0.3203125, + "learning_rate": 0.00010277284007621513, + "loss": 0.6848, + "step": 2653 + }, + { + "epoch": 1.4594519404645987, + "grad_norm": 0.33203125, + "learning_rate": 0.00010271447899197409, + "loss": 0.7605, + "step": 2654 + }, + { + "epoch": 1.4600083460842954, + "grad_norm": 0.322265625, + "learning_rate": 0.00010265611698248062, + "loss": 0.6646, + "step": 2655 + }, + { + "epoch": 1.4605647517039921, + "grad_norm": 0.333984375, + "learning_rate": 0.0001025977540676279, + "loss": 0.7369, + "step": 2656 + }, + { + "epoch": 1.461121157323689, + "grad_norm": 0.32421875, + "learning_rate": 0.00010253939026730938, + "loss": 0.7267, + "step": 2657 + }, + { + "epoch": 1.4616775629433858, + "grad_norm": 0.330078125, + "learning_rate": 0.0001024810256014189, + "loss": 0.7249, + "step": 2658 + }, + { + "epoch": 1.4622339685630825, + "grad_norm": 0.337890625, + "learning_rate": 0.00010242266008985046, + "loss": 0.7515, + "step": 2659 + }, + { + "epoch": 1.4627903741827792, + "grad_norm": 0.357421875, + "learning_rate": 0.00010236429375249848, + "loss": 0.8316, + "step": 2660 + }, + { + "epoch": 1.463346779802476, + "grad_norm": 0.328125, + "learning_rate": 0.00010230592660925758, + "loss": 0.671, + "step": 2661 + }, + { + "epoch": 1.4639031854221727, + "grad_norm": 0.326171875, + "learning_rate": 0.00010224755868002265, + "loss": 0.7033, + "step": 2662 + }, + { + "epoch": 1.4644595910418694, + "grad_norm": 0.328125, + "learning_rate": 0.00010218918998468892, + "loss": 0.7256, + "step": 2663 + }, + { + "epoch": 1.4650159966615663, + "grad_norm": 0.333984375, + "learning_rate": 0.00010213082054315184, + "loss": 0.7519, + "step": 2664 + }, + { + "epoch": 1.465572402281263, + "grad_norm": 0.33984375, + "learning_rate": 0.00010207245037530711, + "loss": 0.7352, + "step": 2665 + }, + { + "epoch": 1.4661288079009598, + "grad_norm": 0.318359375, + "learning_rate": 0.00010201407950105064, + "loss": 0.6068, + "step": 2666 + }, + { + "epoch": 1.4666852135206565, + "grad_norm": 0.330078125, + "learning_rate": 0.00010195570794027866, + "loss": 0.6952, + "step": 2667 + }, + { + "epoch": 1.4672416191403532, + "grad_norm": 0.333984375, + "learning_rate": 0.00010189733571288761, + "loss": 0.7763, + "step": 2668 + }, + { + "epoch": 1.4677980247600502, + "grad_norm": 0.333984375, + "learning_rate": 0.00010183896283877412, + "loss": 0.776, + "step": 2669 + }, + { + "epoch": 1.4683544303797469, + "grad_norm": 0.33203125, + "learning_rate": 0.00010178058933783503, + "loss": 0.7237, + "step": 2670 + }, + { + "epoch": 1.4689108359994436, + "grad_norm": 0.328125, + "learning_rate": 0.0001017222152299675, + "loss": 0.7179, + "step": 2671 + }, + { + "epoch": 1.4694672416191403, + "grad_norm": 0.32421875, + "learning_rate": 0.00010166384053506878, + "loss": 0.6504, + "step": 2672 + }, + { + "epoch": 1.470023647238837, + "grad_norm": 0.32421875, + "learning_rate": 0.00010160546527303635, + "loss": 0.6798, + "step": 2673 + }, + { + "epoch": 1.470580052858534, + "grad_norm": 0.3359375, + "learning_rate": 0.00010154708946376789, + "loss": 0.7528, + "step": 2674 + }, + { + "epoch": 1.4711364584782307, + "grad_norm": 0.3125, + "learning_rate": 0.00010148871312716133, + "loss": 0.6758, + "step": 2675 + }, + { + "epoch": 1.4716928640979274, + "grad_norm": 0.3359375, + "learning_rate": 0.00010143033628311469, + "loss": 0.728, + "step": 2676 + }, + { + "epoch": 1.4722492697176242, + "grad_norm": 0.33203125, + "learning_rate": 0.00010137195895152618, + "loss": 0.7429, + "step": 2677 + }, + { + "epoch": 1.4728056753373209, + "grad_norm": 0.306640625, + "learning_rate": 0.00010131358115229421, + "loss": 0.6451, + "step": 2678 + }, + { + "epoch": 1.4733620809570176, + "grad_norm": 0.322265625, + "learning_rate": 0.00010125520290531734, + "loss": 0.714, + "step": 2679 + }, + { + "epoch": 1.4739184865767143, + "grad_norm": 0.318359375, + "learning_rate": 0.00010119682423049425, + "loss": 0.6505, + "step": 2680 + }, + { + "epoch": 1.4744748921964113, + "grad_norm": 0.34375, + "learning_rate": 0.00010113844514772384, + "loss": 0.7832, + "step": 2681 + }, + { + "epoch": 1.475031297816108, + "grad_norm": 0.3125, + "learning_rate": 0.000101080065676905, + "loss": 0.6569, + "step": 2682 + }, + { + "epoch": 1.4755877034358047, + "grad_norm": 0.31640625, + "learning_rate": 0.00010102168583793696, + "loss": 0.6913, + "step": 2683 + }, + { + "epoch": 1.4761441090555014, + "grad_norm": 0.30859375, + "learning_rate": 0.00010096330565071892, + "loss": 0.6483, + "step": 2684 + }, + { + "epoch": 1.4767005146751981, + "grad_norm": 0.31640625, + "learning_rate": 0.00010090492513515026, + "loss": 0.636, + "step": 2685 + }, + { + "epoch": 1.477256920294895, + "grad_norm": 0.34765625, + "learning_rate": 0.00010084654431113043, + "loss": 0.7358, + "step": 2686 + }, + { + "epoch": 1.4778133259145918, + "grad_norm": 0.32421875, + "learning_rate": 0.00010078816319855903, + "loss": 0.7291, + "step": 2687 + }, + { + "epoch": 1.4783697315342885, + "grad_norm": 0.32421875, + "learning_rate": 0.00010072978181733574, + "loss": 0.695, + "step": 2688 + }, + { + "epoch": 1.4789261371539852, + "grad_norm": 0.345703125, + "learning_rate": 0.00010067140018736033, + "loss": 0.817, + "step": 2689 + }, + { + "epoch": 1.479482542773682, + "grad_norm": 0.330078125, + "learning_rate": 0.00010061301832853267, + "loss": 0.7234, + "step": 2690 + }, + { + "epoch": 1.480038948393379, + "grad_norm": 0.345703125, + "learning_rate": 0.00010055463626075265, + "loss": 0.7691, + "step": 2691 + }, + { + "epoch": 1.4805953540130754, + "grad_norm": 0.337890625, + "learning_rate": 0.00010049625400392034, + "loss": 0.8378, + "step": 2692 + }, + { + "epoch": 1.4811517596327723, + "grad_norm": 0.330078125, + "learning_rate": 0.00010043787157793579, + "loss": 0.694, + "step": 2693 + }, + { + "epoch": 1.481708165252469, + "grad_norm": 0.318359375, + "learning_rate": 0.00010037948900269909, + "loss": 0.6789, + "step": 2694 + }, + { + "epoch": 1.4822645708721658, + "grad_norm": 0.32421875, + "learning_rate": 0.00010032110629811044, + "loss": 0.7308, + "step": 2695 + }, + { + "epoch": 1.4828209764918625, + "grad_norm": 0.322265625, + "learning_rate": 0.00010026272348407011, + "loss": 0.7267, + "step": 2696 + }, + { + "epoch": 1.4833773821115592, + "grad_norm": 0.34375, + "learning_rate": 0.00010020434058047833, + "loss": 0.788, + "step": 2697 + }, + { + "epoch": 1.4833773821115592, + "eval_loss": 0.738091230392456, + "eval_runtime": 1753.0991, + "eval_samples_per_second": 2.676, + "eval_steps_per_second": 1.338, + "step": 2697 + }, + { + "epoch": 1.4839337877312562, + "grad_norm": 0.341796875, + "learning_rate": 0.00010014595760723538, + "loss": 0.7836, + "step": 2698 + }, + { + "epoch": 1.484490193350953, + "grad_norm": 0.3203125, + "learning_rate": 0.00010008757458424156, + "loss": 0.7232, + "step": 2699 + }, + { + "epoch": 1.4850465989706496, + "grad_norm": 0.326171875, + "learning_rate": 0.00010002919153139724, + "loss": 0.7384, + "step": 2700 + }, + { + "epoch": 1.4856030045903463, + "grad_norm": 0.32421875, + "learning_rate": 9.997080846860278e-05, + "loss": 0.7317, + "step": 2701 + }, + { + "epoch": 1.486159410210043, + "grad_norm": 0.322265625, + "learning_rate": 9.991242541575846e-05, + "loss": 0.6792, + "step": 2702 + }, + { + "epoch": 1.48671581582974, + "grad_norm": 0.3203125, + "learning_rate": 9.985404239276467e-05, + "loss": 0.6993, + "step": 2703 + }, + { + "epoch": 1.4872722214494367, + "grad_norm": 0.34375, + "learning_rate": 9.979565941952172e-05, + "loss": 0.7193, + "step": 2704 + }, + { + "epoch": 1.4878286270691334, + "grad_norm": 0.353515625, + "learning_rate": 9.973727651592987e-05, + "loss": 0.7984, + "step": 2705 + }, + { + "epoch": 1.4883850326888302, + "grad_norm": 0.310546875, + "learning_rate": 9.967889370188956e-05, + "loss": 0.6505, + "step": 2706 + }, + { + "epoch": 1.4889414383085269, + "grad_norm": 0.32421875, + "learning_rate": 9.962051099730094e-05, + "loss": 0.6908, + "step": 2707 + }, + { + "epoch": 1.4894978439282236, + "grad_norm": 0.3359375, + "learning_rate": 9.956212842206425e-05, + "loss": 0.7171, + "step": 2708 + }, + { + "epoch": 1.4900542495479203, + "grad_norm": 0.328125, + "learning_rate": 9.950374599607969e-05, + "loss": 0.7713, + "step": 2709 + }, + { + "epoch": 1.4906106551676173, + "grad_norm": 0.326171875, + "learning_rate": 9.944536373924737e-05, + "loss": 0.7031, + "step": 2710 + }, + { + "epoch": 1.491167060787314, + "grad_norm": 0.349609375, + "learning_rate": 9.938698167146738e-05, + "loss": 0.8217, + "step": 2711 + }, + { + "epoch": 1.4917234664070107, + "grad_norm": 0.3515625, + "learning_rate": 9.932859981263971e-05, + "loss": 0.7784, + "step": 2712 + }, + { + "epoch": 1.4922798720267074, + "grad_norm": 0.32421875, + "learning_rate": 9.927021818266428e-05, + "loss": 0.6783, + "step": 2713 + }, + { + "epoch": 1.4928362776464041, + "grad_norm": 0.30859375, + "learning_rate": 9.921183680144099e-05, + "loss": 0.621, + "step": 2714 + }, + { + "epoch": 1.493392683266101, + "grad_norm": 0.337890625, + "learning_rate": 9.915345568886959e-05, + "loss": 0.7907, + "step": 2715 + }, + { + "epoch": 1.4939490888857978, + "grad_norm": 0.33203125, + "learning_rate": 9.909507486484977e-05, + "loss": 0.6956, + "step": 2716 + }, + { + "epoch": 1.4945054945054945, + "grad_norm": 0.333984375, + "learning_rate": 9.90366943492811e-05, + "loss": 0.8112, + "step": 2717 + }, + { + "epoch": 1.4950619001251912, + "grad_norm": 0.3125, + "learning_rate": 9.897831416206306e-05, + "loss": 0.6216, + "step": 2718 + }, + { + "epoch": 1.495618305744888, + "grad_norm": 0.3359375, + "learning_rate": 9.891993432309501e-05, + "loss": 0.708, + "step": 2719 + }, + { + "epoch": 1.496174711364585, + "grad_norm": 0.3203125, + "learning_rate": 9.886155485227621e-05, + "loss": 0.6999, + "step": 2720 + }, + { + "epoch": 1.4967311169842814, + "grad_norm": 0.3203125, + "learning_rate": 9.880317576950578e-05, + "loss": 0.6709, + "step": 2721 + }, + { + "epoch": 1.4972875226039783, + "grad_norm": 0.33203125, + "learning_rate": 9.874479709468266e-05, + "loss": 0.7497, + "step": 2722 + }, + { + "epoch": 1.497843928223675, + "grad_norm": 0.345703125, + "learning_rate": 9.86864188477058e-05, + "loss": 0.7879, + "step": 2723 + }, + { + "epoch": 1.4984003338433718, + "grad_norm": 0.318359375, + "learning_rate": 9.862804104847383e-05, + "loss": 0.6814, + "step": 2724 + }, + { + "epoch": 1.4989567394630685, + "grad_norm": 0.326171875, + "learning_rate": 9.856966371688532e-05, + "loss": 0.679, + "step": 2725 + }, + { + "epoch": 1.4995131450827652, + "grad_norm": 0.3359375, + "learning_rate": 9.85112868728387e-05, + "loss": 0.7994, + "step": 2726 + }, + { + "epoch": 1.5000695507024622, + "grad_norm": 0.341796875, + "learning_rate": 9.845291053623212e-05, + "loss": 0.7785, + "step": 2727 + }, + { + "epoch": 1.500625956322159, + "grad_norm": 0.337890625, + "learning_rate": 9.83945347269637e-05, + "loss": 0.7471, + "step": 2728 + }, + { + "epoch": 1.5011823619418556, + "grad_norm": 0.3203125, + "learning_rate": 9.833615946493127e-05, + "loss": 0.6778, + "step": 2729 + }, + { + "epoch": 1.5017387675615523, + "grad_norm": 0.318359375, + "learning_rate": 9.827778477003251e-05, + "loss": 0.6972, + "step": 2730 + }, + { + "epoch": 1.502295173181249, + "grad_norm": 0.349609375, + "learning_rate": 9.821941066216496e-05, + "loss": 0.7778, + "step": 2731 + }, + { + "epoch": 1.502851578800946, + "grad_norm": 0.328125, + "learning_rate": 9.81610371612259e-05, + "loss": 0.7179, + "step": 2732 + }, + { + "epoch": 1.5034079844206425, + "grad_norm": 0.326171875, + "learning_rate": 9.810266428711241e-05, + "loss": 0.7209, + "step": 2733 + }, + { + "epoch": 1.5039643900403394, + "grad_norm": 0.33203125, + "learning_rate": 9.804429205972135e-05, + "loss": 0.7074, + "step": 2734 + }, + { + "epoch": 1.5045207956600362, + "grad_norm": 0.314453125, + "learning_rate": 9.798592049894938e-05, + "loss": 0.6736, + "step": 2735 + }, + { + "epoch": 1.5050772012797329, + "grad_norm": 0.34765625, + "learning_rate": 9.792754962469294e-05, + "loss": 0.7606, + "step": 2736 + }, + { + "epoch": 1.5056336068994298, + "grad_norm": 0.328125, + "learning_rate": 9.786917945684819e-05, + "loss": 0.7024, + "step": 2737 + }, + { + "epoch": 1.5061900125191263, + "grad_norm": 0.326171875, + "learning_rate": 9.781081001531107e-05, + "loss": 0.7051, + "step": 2738 + }, + { + "epoch": 1.5067464181388233, + "grad_norm": 0.310546875, + "learning_rate": 9.775244131997736e-05, + "loss": 0.64, + "step": 2739 + }, + { + "epoch": 1.50730282375852, + "grad_norm": 0.33203125, + "learning_rate": 9.769407339074245e-05, + "loss": 0.6996, + "step": 2740 + }, + { + "epoch": 1.5078592293782167, + "grad_norm": 0.34375, + "learning_rate": 9.763570624750154e-05, + "loss": 0.7905, + "step": 2741 + }, + { + "epoch": 1.5084156349979134, + "grad_norm": 0.326171875, + "learning_rate": 9.757733991014955e-05, + "loss": 0.7254, + "step": 2742 + }, + { + "epoch": 1.5089720406176101, + "grad_norm": 0.337890625, + "learning_rate": 9.751897439858113e-05, + "loss": 0.7522, + "step": 2743 + }, + { + "epoch": 1.509528446237307, + "grad_norm": 0.328125, + "learning_rate": 9.746060973269063e-05, + "loss": 0.7335, + "step": 2744 + }, + { + "epoch": 1.5100848518570038, + "grad_norm": 0.337890625, + "learning_rate": 9.740224593237212e-05, + "loss": 0.7064, + "step": 2745 + }, + { + "epoch": 1.5106412574767005, + "grad_norm": 0.31640625, + "learning_rate": 9.73438830175194e-05, + "loss": 0.6917, + "step": 2746 + }, + { + "epoch": 1.5111976630963972, + "grad_norm": 0.326171875, + "learning_rate": 9.728552100802592e-05, + "loss": 0.7373, + "step": 2747 + }, + { + "epoch": 1.511754068716094, + "grad_norm": 0.314453125, + "learning_rate": 9.722715992378489e-05, + "loss": 0.6352, + "step": 2748 + }, + { + "epoch": 1.512310474335791, + "grad_norm": 0.330078125, + "learning_rate": 9.716879978468915e-05, + "loss": 0.7285, + "step": 2749 + }, + { + "epoch": 1.5128668799554874, + "grad_norm": 0.33203125, + "learning_rate": 9.711044061063122e-05, + "loss": 0.7706, + "step": 2750 + }, + { + "epoch": 1.5134232855751844, + "grad_norm": 0.328125, + "learning_rate": 9.70520824215033e-05, + "loss": 0.7315, + "step": 2751 + }, + { + "epoch": 1.513979691194881, + "grad_norm": 0.7578125, + "learning_rate": 9.699372523719727e-05, + "loss": 1.7202, + "step": 2752 + }, + { + "epoch": 1.5145360968145778, + "grad_norm": 0.328125, + "learning_rate": 9.693536907760465e-05, + "loss": 0.7514, + "step": 2753 + }, + { + "epoch": 1.5150925024342747, + "grad_norm": 0.33984375, + "learning_rate": 9.687701396261659e-05, + "loss": 0.7345, + "step": 2754 + }, + { + "epoch": 1.5156489080539712, + "grad_norm": 0.3359375, + "learning_rate": 9.681865991212395e-05, + "loss": 0.7016, + "step": 2755 + }, + { + "epoch": 1.5162053136736682, + "grad_norm": 0.310546875, + "learning_rate": 9.676030694601719e-05, + "loss": 0.6468, + "step": 2756 + }, + { + "epoch": 1.516761719293365, + "grad_norm": 0.3125, + "learning_rate": 9.670195508418637e-05, + "loss": 0.6537, + "step": 2757 + }, + { + "epoch": 1.5173181249130616, + "grad_norm": 0.322265625, + "learning_rate": 9.664360434652122e-05, + "loss": 0.6927, + "step": 2758 + }, + { + "epoch": 1.5178745305327583, + "grad_norm": 0.32421875, + "learning_rate": 9.658525475291105e-05, + "loss": 0.7005, + "step": 2759 + }, + { + "epoch": 1.518430936152455, + "grad_norm": 0.328125, + "learning_rate": 9.652690632324484e-05, + "loss": 0.7236, + "step": 2760 + }, + { + "epoch": 1.518987341772152, + "grad_norm": 0.3359375, + "learning_rate": 9.646855907741107e-05, + "loss": 0.7185, + "step": 2761 + }, + { + "epoch": 1.5195437473918485, + "grad_norm": 0.337890625, + "learning_rate": 9.64102130352979e-05, + "loss": 0.7771, + "step": 2762 + }, + { + "epoch": 1.5201001530115454, + "grad_norm": 0.318359375, + "learning_rate": 9.635186821679301e-05, + "loss": 0.6811, + "step": 2763 + }, + { + "epoch": 1.5206565586312422, + "grad_norm": 0.33203125, + "learning_rate": 9.629352464178383e-05, + "loss": 0.6896, + "step": 2764 + }, + { + "epoch": 1.5212129642509389, + "grad_norm": 0.326171875, + "learning_rate": 9.623518233015715e-05, + "loss": 0.6704, + "step": 2765 + }, + { + "epoch": 1.5217693698706358, + "grad_norm": 0.333984375, + "learning_rate": 9.617684130179947e-05, + "loss": 0.7421, + "step": 2766 + }, + { + "epoch": 1.5223257754903323, + "grad_norm": 0.330078125, + "learning_rate": 9.611850157659677e-05, + "loss": 0.7039, + "step": 2767 + }, + { + "epoch": 1.5228821811100293, + "grad_norm": 0.33203125, + "learning_rate": 9.606016317443467e-05, + "loss": 0.7274, + "step": 2768 + }, + { + "epoch": 1.523438586729726, + "grad_norm": 0.33984375, + "learning_rate": 9.600182611519824e-05, + "loss": 0.7613, + "step": 2769 + }, + { + "epoch": 1.5239949923494227, + "grad_norm": 0.34375, + "learning_rate": 9.59434904187722e-05, + "loss": 0.7661, + "step": 2770 + }, + { + "epoch": 1.5245513979691196, + "grad_norm": 0.3203125, + "learning_rate": 9.588515610504069e-05, + "loss": 0.6761, + "step": 2771 + }, + { + "epoch": 1.5251078035888161, + "grad_norm": 0.322265625, + "learning_rate": 9.582682319388749e-05, + "loss": 0.7102, + "step": 2772 + }, + { + "epoch": 1.525664209208513, + "grad_norm": 0.33203125, + "learning_rate": 9.576849170519583e-05, + "loss": 0.7238, + "step": 2773 + }, + { + "epoch": 1.5262206148282098, + "grad_norm": 0.322265625, + "learning_rate": 9.571016165884847e-05, + "loss": 0.6984, + "step": 2774 + }, + { + "epoch": 1.5267770204479065, + "grad_norm": 0.3359375, + "learning_rate": 9.565183307472771e-05, + "loss": 0.6523, + "step": 2775 + }, + { + "epoch": 1.5273334260676033, + "grad_norm": 0.33984375, + "learning_rate": 9.559350597271531e-05, + "loss": 0.7308, + "step": 2776 + }, + { + "epoch": 1.5278898316873, + "grad_norm": 0.33203125, + "learning_rate": 9.553518037269254e-05, + "loss": 0.6982, + "step": 2777 + }, + { + "epoch": 1.528446237306997, + "grad_norm": 0.3203125, + "learning_rate": 9.547685629454018e-05, + "loss": 0.6757, + "step": 2778 + }, + { + "epoch": 1.5290026429266934, + "grad_norm": 0.341796875, + "learning_rate": 9.541853375813837e-05, + "loss": 0.7805, + "step": 2779 + }, + { + "epoch": 1.5295590485463904, + "grad_norm": 0.326171875, + "learning_rate": 9.536021278336697e-05, + "loss": 0.6971, + "step": 2780 + }, + { + "epoch": 1.530115454166087, + "grad_norm": 0.328125, + "learning_rate": 9.53018933901051e-05, + "loss": 0.6538, + "step": 2781 + }, + { + "epoch": 1.5306718597857838, + "grad_norm": 0.337890625, + "learning_rate": 9.52435755982314e-05, + "loss": 0.7626, + "step": 2782 + }, + { + "epoch": 1.5312282654054807, + "grad_norm": 0.328125, + "learning_rate": 9.518525942762397e-05, + "loss": 0.7124, + "step": 2783 + }, + { + "epoch": 1.5317846710251772, + "grad_norm": 0.33203125, + "learning_rate": 9.512694489816036e-05, + "loss": 0.7088, + "step": 2784 + }, + { + "epoch": 1.5323410766448742, + "grad_norm": 0.32421875, + "learning_rate": 9.506863202971755e-05, + "loss": 0.676, + "step": 2785 + }, + { + "epoch": 1.532897482264571, + "grad_norm": 0.353515625, + "learning_rate": 9.501032084217196e-05, + "loss": 0.8049, + "step": 2786 + }, + { + "epoch": 1.5334538878842676, + "grad_norm": 0.33984375, + "learning_rate": 9.49520113553994e-05, + "loss": 0.7593, + "step": 2787 + }, + { + "epoch": 1.5340102935039643, + "grad_norm": 0.345703125, + "learning_rate": 9.489370358927522e-05, + "loss": 0.7928, + "step": 2788 + }, + { + "epoch": 1.534566699123661, + "grad_norm": 0.322265625, + "learning_rate": 9.483539756367405e-05, + "loss": 0.6924, + "step": 2789 + }, + { + "epoch": 1.535123104743358, + "grad_norm": 0.341796875, + "learning_rate": 9.477709329846996e-05, + "loss": 0.7534, + "step": 2790 + }, + { + "epoch": 1.5356795103630545, + "grad_norm": 0.32421875, + "learning_rate": 9.471879081353648e-05, + "loss": 0.6929, + "step": 2791 + }, + { + "epoch": 1.5362359159827514, + "grad_norm": 0.3359375, + "learning_rate": 9.466049012874645e-05, + "loss": 0.7124, + "step": 2792 + }, + { + "epoch": 1.5367923216024482, + "grad_norm": 0.328125, + "learning_rate": 9.460219126397218e-05, + "loss": 0.7566, + "step": 2793 + }, + { + "epoch": 1.5373487272221449, + "grad_norm": 0.318359375, + "learning_rate": 9.454389423908527e-05, + "loss": 0.6552, + "step": 2794 + }, + { + "epoch": 1.5379051328418418, + "grad_norm": 0.3203125, + "learning_rate": 9.448559907395677e-05, + "loss": 0.7103, + "step": 2795 + }, + { + "epoch": 1.5384615384615383, + "grad_norm": 0.33984375, + "learning_rate": 9.4427305788457e-05, + "loss": 0.7541, + "step": 2796 + }, + { + "epoch": 1.5390179440812353, + "grad_norm": 0.318359375, + "learning_rate": 9.436901440245581e-05, + "loss": 0.6623, + "step": 2797 + }, + { + "epoch": 1.539574349700932, + "grad_norm": 0.330078125, + "learning_rate": 9.431072493582225e-05, + "loss": 0.7121, + "step": 2798 + }, + { + "epoch": 1.5401307553206287, + "grad_norm": 0.322265625, + "learning_rate": 9.425243740842475e-05, + "loss": 0.7047, + "step": 2799 + }, + { + "epoch": 1.5406871609403257, + "grad_norm": 0.330078125, + "learning_rate": 9.419415184013111e-05, + "loss": 0.6929, + "step": 2800 + }, + { + "epoch": 1.5412435665600221, + "grad_norm": 0.34375, + "learning_rate": 9.413586825080844e-05, + "loss": 0.8071, + "step": 2801 + }, + { + "epoch": 1.541799972179719, + "grad_norm": 0.333984375, + "learning_rate": 9.407758666032317e-05, + "loss": 0.6932, + "step": 2802 + }, + { + "epoch": 1.5423563777994158, + "grad_norm": 0.3359375, + "learning_rate": 9.401930708854106e-05, + "loss": 0.7572, + "step": 2803 + }, + { + "epoch": 1.5429127834191125, + "grad_norm": 0.30078125, + "learning_rate": 9.396102955532717e-05, + "loss": 0.5738, + "step": 2804 + }, + { + "epoch": 1.5434691890388093, + "grad_norm": 0.34765625, + "learning_rate": 9.390275408054591e-05, + "loss": 0.739, + "step": 2805 + }, + { + "epoch": 1.544025594658506, + "grad_norm": 0.35546875, + "learning_rate": 9.384448068406092e-05, + "loss": 0.7646, + "step": 2806 + }, + { + "epoch": 1.544582000278203, + "grad_norm": 0.322265625, + "learning_rate": 9.378620938573519e-05, + "loss": 0.6922, + "step": 2807 + }, + { + "epoch": 1.5451384058978994, + "grad_norm": 0.33203125, + "learning_rate": 9.372794020543096e-05, + "loss": 0.7647, + "step": 2808 + }, + { + "epoch": 1.5456948115175964, + "grad_norm": 0.3125, + "learning_rate": 9.366967316300974e-05, + "loss": 0.6548, + "step": 2809 + }, + { + "epoch": 1.546251217137293, + "grad_norm": 0.32421875, + "learning_rate": 9.361140827833235e-05, + "loss": 0.7338, + "step": 2810 + }, + { + "epoch": 1.5468076227569898, + "grad_norm": 0.32421875, + "learning_rate": 9.355314557125885e-05, + "loss": 0.7016, + "step": 2811 + }, + { + "epoch": 1.5473640283766867, + "grad_norm": 0.302734375, + "learning_rate": 9.349488506164852e-05, + "loss": 0.6144, + "step": 2812 + }, + { + "epoch": 1.5479204339963832, + "grad_norm": 0.341796875, + "learning_rate": 9.343662676936001e-05, + "loss": 0.7292, + "step": 2813 + }, + { + "epoch": 1.5484768396160802, + "grad_norm": 0.314453125, + "learning_rate": 9.33783707142511e-05, + "loss": 0.6513, + "step": 2814 + }, + { + "epoch": 1.549033245235777, + "grad_norm": 0.30859375, + "learning_rate": 9.332011691617884e-05, + "loss": 0.6298, + "step": 2815 + }, + { + "epoch": 1.5495896508554736, + "grad_norm": 0.333984375, + "learning_rate": 9.326186539499953e-05, + "loss": 0.7042, + "step": 2816 + }, + { + "epoch": 1.5501460564751706, + "grad_norm": 0.3359375, + "learning_rate": 9.320361617056865e-05, + "loss": 0.7797, + "step": 2817 + }, + { + "epoch": 1.550702462094867, + "grad_norm": 0.310546875, + "learning_rate": 9.314536926274097e-05, + "loss": 0.6486, + "step": 2818 + }, + { + "epoch": 1.551258867714564, + "grad_norm": 0.400390625, + "learning_rate": 9.308712469137037e-05, + "loss": 0.7521, + "step": 2819 + }, + { + "epoch": 1.5518152733342607, + "grad_norm": 0.3359375, + "learning_rate": 9.302888247631002e-05, + "loss": 0.7332, + "step": 2820 + }, + { + "epoch": 1.5523716789539574, + "grad_norm": 0.330078125, + "learning_rate": 9.297064263741226e-05, + "loss": 0.7702, + "step": 2821 + }, + { + "epoch": 1.5529280845736542, + "grad_norm": 0.326171875, + "learning_rate": 9.291240519452863e-05, + "loss": 0.6543, + "step": 2822 + }, + { + "epoch": 1.5534844901933509, + "grad_norm": 0.31640625, + "learning_rate": 9.285417016750982e-05, + "loss": 0.666, + "step": 2823 + }, + { + "epoch": 1.5540408958130478, + "grad_norm": 0.314453125, + "learning_rate": 9.27959375762057e-05, + "loss": 0.6736, + "step": 2824 + }, + { + "epoch": 1.5545973014327443, + "grad_norm": 0.33984375, + "learning_rate": 9.273770744046536e-05, + "loss": 0.7375, + "step": 2825 + }, + { + "epoch": 1.5551537070524413, + "grad_norm": 0.322265625, + "learning_rate": 9.2679479780137e-05, + "loss": 0.658, + "step": 2826 + }, + { + "epoch": 1.555710112672138, + "grad_norm": 0.3359375, + "learning_rate": 9.262125461506801e-05, + "loss": 0.7522, + "step": 2827 + }, + { + "epoch": 1.5562665182918347, + "grad_norm": 0.328125, + "learning_rate": 9.25630319651049e-05, + "loss": 0.6961, + "step": 2828 + }, + { + "epoch": 1.5568229239115317, + "grad_norm": 0.3203125, + "learning_rate": 9.250481185009332e-05, + "loss": 0.6611, + "step": 2829 + }, + { + "epoch": 1.5573793295312282, + "grad_norm": 0.34375, + "learning_rate": 9.244659428987813e-05, + "loss": 0.7516, + "step": 2830 + }, + { + "epoch": 1.557935735150925, + "grad_norm": 0.345703125, + "learning_rate": 9.238837930430321e-05, + "loss": 0.748, + "step": 2831 + }, + { + "epoch": 1.5584921407706218, + "grad_norm": 0.333984375, + "learning_rate": 9.233016691321165e-05, + "loss": 0.7353, + "step": 2832 + }, + { + "epoch": 1.5590485463903185, + "grad_norm": 0.349609375, + "learning_rate": 9.22719571364456e-05, + "loss": 0.8053, + "step": 2833 + }, + { + "epoch": 1.5596049520100153, + "grad_norm": 0.330078125, + "learning_rate": 9.221374999384636e-05, + "loss": 0.7697, + "step": 2834 + }, + { + "epoch": 1.560161357629712, + "grad_norm": 0.326171875, + "learning_rate": 9.215554550525429e-05, + "loss": 0.6776, + "step": 2835 + }, + { + "epoch": 1.560717763249409, + "grad_norm": 0.33203125, + "learning_rate": 9.209734369050887e-05, + "loss": 0.7335, + "step": 2836 + }, + { + "epoch": 1.5612741688691054, + "grad_norm": 0.318359375, + "learning_rate": 9.203914456944863e-05, + "loss": 0.6751, + "step": 2837 + }, + { + "epoch": 1.5618305744888024, + "grad_norm": 0.318359375, + "learning_rate": 9.198094816191131e-05, + "loss": 0.6719, + "step": 2838 + }, + { + "epoch": 1.562386980108499, + "grad_norm": 0.330078125, + "learning_rate": 9.192275448773358e-05, + "loss": 0.7057, + "step": 2839 + }, + { + "epoch": 1.5629433857281958, + "grad_norm": 0.306640625, + "learning_rate": 9.186456356675124e-05, + "loss": 0.6172, + "step": 2840 + }, + { + "epoch": 1.5634997913478927, + "grad_norm": 0.314453125, + "learning_rate": 9.180637541879912e-05, + "loss": 0.6384, + "step": 2841 + }, + { + "epoch": 1.5640561969675892, + "grad_norm": 0.337890625, + "learning_rate": 9.174819006371115e-05, + "loss": 0.7106, + "step": 2842 + }, + { + "epoch": 1.5646126025872862, + "grad_norm": 0.357421875, + "learning_rate": 9.169000752132029e-05, + "loss": 0.8161, + "step": 2843 + }, + { + "epoch": 1.565169008206983, + "grad_norm": 0.30078125, + "learning_rate": 9.163182781145852e-05, + "loss": 0.5613, + "step": 2844 + }, + { + "epoch": 1.5657254138266796, + "grad_norm": 0.35546875, + "learning_rate": 9.157365095395685e-05, + "loss": 0.8183, + "step": 2845 + }, + { + "epoch": 1.5662818194463766, + "grad_norm": 0.330078125, + "learning_rate": 9.151547696864537e-05, + "loss": 0.7441, + "step": 2846 + }, + { + "epoch": 1.566838225066073, + "grad_norm": 0.330078125, + "learning_rate": 9.145730587535318e-05, + "loss": 0.7141, + "step": 2847 + }, + { + "epoch": 1.56739463068577, + "grad_norm": 0.353515625, + "learning_rate": 9.139913769390836e-05, + "loss": 0.7938, + "step": 2848 + }, + { + "epoch": 1.5679510363054667, + "grad_norm": 0.341796875, + "learning_rate": 9.134097244413798e-05, + "loss": 0.7654, + "step": 2849 + }, + { + "epoch": 1.5685074419251634, + "grad_norm": 0.34375, + "learning_rate": 9.128281014586818e-05, + "loss": 0.7077, + "step": 2850 + }, + { + "epoch": 1.5690638475448602, + "grad_norm": 0.337890625, + "learning_rate": 9.122465081892404e-05, + "loss": 0.7241, + "step": 2851 + }, + { + "epoch": 1.5696202531645569, + "grad_norm": 0.333984375, + "learning_rate": 9.116649448312964e-05, + "loss": 0.705, + "step": 2852 + }, + { + "epoch": 1.5701766587842538, + "grad_norm": 0.322265625, + "learning_rate": 9.1108341158308e-05, + "loss": 0.6932, + "step": 2853 + }, + { + "epoch": 1.5707330644039503, + "grad_norm": 0.328125, + "learning_rate": 9.105019086428127e-05, + "loss": 0.7234, + "step": 2854 + }, + { + "epoch": 1.5712894700236473, + "grad_norm": 0.333984375, + "learning_rate": 9.099204362087037e-05, + "loss": 0.7096, + "step": 2855 + }, + { + "epoch": 1.571845875643344, + "grad_norm": 0.34765625, + "learning_rate": 9.093389944789529e-05, + "loss": 0.7347, + "step": 2856 + }, + { + "epoch": 1.5724022812630407, + "grad_norm": 0.337890625, + "learning_rate": 9.087575836517495e-05, + "loss": 0.7595, + "step": 2857 + }, + { + "epoch": 1.5729586868827377, + "grad_norm": 0.322265625, + "learning_rate": 9.081762039252721e-05, + "loss": 0.6612, + "step": 2858 + }, + { + "epoch": 1.5735150925024342, + "grad_norm": 0.3125, + "learning_rate": 9.075948554976887e-05, + "loss": 0.6194, + "step": 2859 + }, + { + "epoch": 1.574071498122131, + "grad_norm": 0.322265625, + "learning_rate": 9.070135385671568e-05, + "loss": 0.6603, + "step": 2860 + }, + { + "epoch": 1.5746279037418278, + "grad_norm": 0.3359375, + "learning_rate": 9.064322533318225e-05, + "loss": 0.723, + "step": 2861 + }, + { + "epoch": 1.5751843093615245, + "grad_norm": 0.318359375, + "learning_rate": 9.058509999898225e-05, + "loss": 0.7042, + "step": 2862 + }, + { + "epoch": 1.5757407149812213, + "grad_norm": 0.310546875, + "learning_rate": 9.052697787392816e-05, + "loss": 0.6453, + "step": 2863 + }, + { + "epoch": 1.576297120600918, + "grad_norm": 0.361328125, + "learning_rate": 9.046885897783133e-05, + "loss": 0.8275, + "step": 2864 + }, + { + "epoch": 1.576853526220615, + "grad_norm": 0.345703125, + "learning_rate": 9.04107433305021e-05, + "loss": 0.8241, + "step": 2865 + }, + { + "epoch": 1.5774099318403116, + "grad_norm": 0.3125, + "learning_rate": 9.035263095174966e-05, + "loss": 0.6421, + "step": 2866 + }, + { + "epoch": 1.5779663374600084, + "grad_norm": 0.314453125, + "learning_rate": 9.02945218613821e-05, + "loss": 0.6377, + "step": 2867 + }, + { + "epoch": 1.578522743079705, + "grad_norm": 0.333984375, + "learning_rate": 9.023641607920635e-05, + "loss": 0.7202, + "step": 2868 + }, + { + "epoch": 1.5790791486994018, + "grad_norm": 0.32421875, + "learning_rate": 9.017831362502828e-05, + "loss": 0.7256, + "step": 2869 + }, + { + "epoch": 1.5796355543190987, + "grad_norm": 0.328125, + "learning_rate": 9.012021451865251e-05, + "loss": 0.7499, + "step": 2870 + }, + { + "epoch": 1.5801919599387952, + "grad_norm": 0.33203125, + "learning_rate": 9.00621187798827e-05, + "loss": 0.7419, + "step": 2871 + }, + { + "epoch": 1.5807483655584922, + "grad_norm": 0.333984375, + "learning_rate": 9.000402642852122e-05, + "loss": 0.6842, + "step": 2872 + }, + { + "epoch": 1.581304771178189, + "grad_norm": 0.3515625, + "learning_rate": 8.994593748436933e-05, + "loss": 0.807, + "step": 2873 + }, + { + "epoch": 1.5818611767978856, + "grad_norm": 0.310546875, + "learning_rate": 8.988785196722712e-05, + "loss": 0.6686, + "step": 2874 + }, + { + "epoch": 1.5824175824175826, + "grad_norm": 0.328125, + "learning_rate": 8.982976989689348e-05, + "loss": 0.7019, + "step": 2875 + }, + { + "epoch": 1.582973988037279, + "grad_norm": 0.333984375, + "learning_rate": 8.97716912931662e-05, + "loss": 0.7445, + "step": 2876 + }, + { + "epoch": 1.583530393656976, + "grad_norm": 0.341796875, + "learning_rate": 8.971361617584184e-05, + "loss": 0.7618, + "step": 2877 + }, + { + "epoch": 1.5840867992766727, + "grad_norm": 0.33203125, + "learning_rate": 8.965554456471576e-05, + "loss": 0.7228, + "step": 2878 + }, + { + "epoch": 1.5846432048963695, + "grad_norm": 0.318359375, + "learning_rate": 8.959747647958217e-05, + "loss": 0.6708, + "step": 2879 + }, + { + "epoch": 1.5851996105160662, + "grad_norm": 0.32421875, + "learning_rate": 8.953941194023405e-05, + "loss": 0.6951, + "step": 2880 + }, + { + "epoch": 1.585756016135763, + "grad_norm": 0.33984375, + "learning_rate": 8.948135096646317e-05, + "loss": 0.7609, + "step": 2881 + }, + { + "epoch": 1.5863124217554598, + "grad_norm": 0.3359375, + "learning_rate": 8.94232935780601e-05, + "loss": 0.7917, + "step": 2882 + }, + { + "epoch": 1.5868688273751563, + "grad_norm": 0.314453125, + "learning_rate": 8.936523979481415e-05, + "loss": 0.6449, + "step": 2883 + }, + { + "epoch": 1.5874252329948533, + "grad_norm": 0.349609375, + "learning_rate": 8.930718963651345e-05, + "loss": 0.797, + "step": 2884 + }, + { + "epoch": 1.58798163861455, + "grad_norm": 0.361328125, + "learning_rate": 8.924914312294485e-05, + "loss": 0.8216, + "step": 2885 + }, + { + "epoch": 1.5885380442342467, + "grad_norm": 0.314453125, + "learning_rate": 8.919110027389399e-05, + "loss": 0.6568, + "step": 2886 + }, + { + "epoch": 1.5890944498539437, + "grad_norm": 0.349609375, + "learning_rate": 8.913306110914529e-05, + "loss": 0.8175, + "step": 2887 + }, + { + "epoch": 1.5896508554736402, + "grad_norm": 0.32421875, + "learning_rate": 8.907502564848183e-05, + "loss": 0.6791, + "step": 2888 + }, + { + "epoch": 1.590207261093337, + "grad_norm": 0.34375, + "learning_rate": 8.901699391168549e-05, + "loss": 0.7591, + "step": 2889 + }, + { + "epoch": 1.5907636667130338, + "grad_norm": 0.345703125, + "learning_rate": 8.895896591853686e-05, + "loss": 0.756, + "step": 2890 + }, + { + "epoch": 1.5913200723327305, + "grad_norm": 0.30078125, + "learning_rate": 8.890094168881524e-05, + "loss": 0.6052, + "step": 2891 + }, + { + "epoch": 1.5918764779524275, + "grad_norm": 0.3203125, + "learning_rate": 8.884292124229869e-05, + "loss": 0.666, + "step": 2892 + }, + { + "epoch": 1.592432883572124, + "grad_norm": 0.328125, + "learning_rate": 8.878490459876396e-05, + "loss": 0.7015, + "step": 2893 + }, + { + "epoch": 1.592989289191821, + "grad_norm": 0.322265625, + "learning_rate": 8.872689177798643e-05, + "loss": 0.676, + "step": 2894 + }, + { + "epoch": 1.5935456948115176, + "grad_norm": 0.33984375, + "learning_rate": 8.866888279974034e-05, + "loss": 0.759, + "step": 2895 + }, + { + "epoch": 1.5941021004312144, + "grad_norm": 0.337890625, + "learning_rate": 8.86108776837985e-05, + "loss": 0.7105, + "step": 2896 + }, + { + "epoch": 1.594658506050911, + "grad_norm": 0.328125, + "learning_rate": 8.855287644993241e-05, + "loss": 0.7363, + "step": 2897 + }, + { + "epoch": 1.5952149116706078, + "grad_norm": 0.3046875, + "learning_rate": 8.849487911791228e-05, + "loss": 0.6814, + "step": 2898 + }, + { + "epoch": 1.5957713172903047, + "grad_norm": 0.357421875, + "learning_rate": 8.843688570750696e-05, + "loss": 0.7803, + "step": 2899 + }, + { + "epoch": 1.5963277229100012, + "grad_norm": 0.33203125, + "learning_rate": 8.8378896238484e-05, + "loss": 0.7239, + "step": 2900 + }, + { + "epoch": 1.5968841285296982, + "grad_norm": 0.341796875, + "learning_rate": 8.832091073060955e-05, + "loss": 0.8145, + "step": 2901 + }, + { + "epoch": 1.597440534149395, + "grad_norm": 0.34375, + "learning_rate": 8.826292920364849e-05, + "loss": 0.7593, + "step": 2902 + }, + { + "epoch": 1.5979969397690916, + "grad_norm": 0.328125, + "learning_rate": 8.820495167736424e-05, + "loss": 0.7041, + "step": 2903 + }, + { + "epoch": 1.5985533453887886, + "grad_norm": 0.345703125, + "learning_rate": 8.8146978171519e-05, + "loss": 0.7853, + "step": 2904 + }, + { + "epoch": 1.599109751008485, + "grad_norm": 0.3203125, + "learning_rate": 8.808900870587347e-05, + "loss": 0.6522, + "step": 2905 + }, + { + "epoch": 1.599666156628182, + "grad_norm": 0.31640625, + "learning_rate": 8.8031043300187e-05, + "loss": 0.6584, + "step": 2906 + }, + { + "epoch": 1.6002225622478787, + "grad_norm": 0.345703125, + "learning_rate": 8.797308197421761e-05, + "loss": 0.7991, + "step": 2907 + }, + { + "epoch": 1.6007789678675755, + "grad_norm": 0.33203125, + "learning_rate": 8.791512474772186e-05, + "loss": 0.7312, + "step": 2908 + }, + { + "epoch": 1.6013353734872722, + "grad_norm": 0.328125, + "learning_rate": 8.785717164045498e-05, + "loss": 0.7392, + "step": 2909 + }, + { + "epoch": 1.601891779106969, + "grad_norm": 0.34765625, + "learning_rate": 8.779922267217072e-05, + "loss": 0.7615, + "step": 2910 + }, + { + "epoch": 1.6024481847266658, + "grad_norm": 0.326171875, + "learning_rate": 8.774127786262145e-05, + "loss": 0.6892, + "step": 2911 + }, + { + "epoch": 1.6030045903463623, + "grad_norm": 0.337890625, + "learning_rate": 8.768333723155821e-05, + "loss": 0.7566, + "step": 2912 + }, + { + "epoch": 1.6035609959660593, + "grad_norm": 0.333984375, + "learning_rate": 8.762540079873052e-05, + "loss": 0.7436, + "step": 2913 + }, + { + "epoch": 1.604117401585756, + "grad_norm": 0.333984375, + "learning_rate": 8.756746858388644e-05, + "loss": 0.7157, + "step": 2914 + }, + { + "epoch": 1.6046738072054527, + "grad_norm": 0.32421875, + "learning_rate": 8.750954060677268e-05, + "loss": 0.684, + "step": 2915 + }, + { + "epoch": 1.6052302128251497, + "grad_norm": 0.3359375, + "learning_rate": 8.745161688713443e-05, + "loss": 0.715, + "step": 2916 + }, + { + "epoch": 1.6057866184448462, + "grad_norm": 0.328125, + "learning_rate": 8.739369744471551e-05, + "loss": 0.7095, + "step": 2917 + }, + { + "epoch": 1.606343024064543, + "grad_norm": 0.34765625, + "learning_rate": 8.733578229925821e-05, + "loss": 0.7868, + "step": 2918 + }, + { + "epoch": 1.6068994296842398, + "grad_norm": 0.333984375, + "learning_rate": 8.727787147050337e-05, + "loss": 0.7046, + "step": 2919 + }, + { + "epoch": 1.6074558353039365, + "grad_norm": 0.33984375, + "learning_rate": 8.72199649781904e-05, + "loss": 0.7415, + "step": 2920 + }, + { + "epoch": 1.6080122409236335, + "grad_norm": 0.32421875, + "learning_rate": 8.716206284205719e-05, + "loss": 0.7006, + "step": 2921 + }, + { + "epoch": 1.60856864654333, + "grad_norm": 0.330078125, + "learning_rate": 8.710416508184016e-05, + "loss": 0.7574, + "step": 2922 + }, + { + "epoch": 1.609125052163027, + "grad_norm": 0.330078125, + "learning_rate": 8.704627171727424e-05, + "loss": 0.7135, + "step": 2923 + }, + { + "epoch": 1.6096814577827236, + "grad_norm": 0.3203125, + "learning_rate": 8.698838276809286e-05, + "loss": 0.7248, + "step": 2924 + }, + { + "epoch": 1.6102378634024204, + "grad_norm": 0.29296875, + "learning_rate": 8.693049825402793e-05, + "loss": 0.5508, + "step": 2925 + }, + { + "epoch": 1.610794269022117, + "grad_norm": 0.328125, + "learning_rate": 8.687261819480985e-05, + "loss": 0.6829, + "step": 2926 + }, + { + "epoch": 1.6113506746418138, + "grad_norm": 0.318359375, + "learning_rate": 8.68147426101675e-05, + "loss": 0.6935, + "step": 2927 + }, + { + "epoch": 1.6119070802615107, + "grad_norm": 0.33203125, + "learning_rate": 8.675687151982833e-05, + "loss": 0.7266, + "step": 2928 + }, + { + "epoch": 1.6124634858812072, + "grad_norm": 0.32421875, + "learning_rate": 8.669900494351812e-05, + "loss": 0.6239, + "step": 2929 + }, + { + "epoch": 1.6130198915009042, + "grad_norm": 1.0625, + "learning_rate": 8.664114290096116e-05, + "loss": 1.5934, + "step": 2930 + }, + { + "epoch": 1.613576297120601, + "grad_norm": 0.361328125, + "learning_rate": 8.658328541188023e-05, + "loss": 0.7803, + "step": 2931 + }, + { + "epoch": 1.6141327027402976, + "grad_norm": 0.345703125, + "learning_rate": 8.652543249599651e-05, + "loss": 0.8155, + "step": 2932 + }, + { + "epoch": 1.6146891083599946, + "grad_norm": 0.328125, + "learning_rate": 8.646758417302965e-05, + "loss": 0.7464, + "step": 2933 + }, + { + "epoch": 1.615245513979691, + "grad_norm": 0.3203125, + "learning_rate": 8.64097404626977e-05, + "loss": 0.697, + "step": 2934 + }, + { + "epoch": 1.615801919599388, + "grad_norm": 0.322265625, + "learning_rate": 8.635190138471717e-05, + "loss": 0.6445, + "step": 2935 + }, + { + "epoch": 1.6163583252190847, + "grad_norm": 0.33203125, + "learning_rate": 8.629406695880302e-05, + "loss": 0.718, + "step": 2936 + }, + { + "epoch": 1.6169147308387815, + "grad_norm": 0.330078125, + "learning_rate": 8.623623720466855e-05, + "loss": 0.7053, + "step": 2937 + }, + { + "epoch": 1.6174711364584782, + "grad_norm": 0.341796875, + "learning_rate": 8.617841214202555e-05, + "loss": 0.725, + "step": 2938 + }, + { + "epoch": 1.618027542078175, + "grad_norm": 0.3046875, + "learning_rate": 8.612059179058412e-05, + "loss": 0.6323, + "step": 2939 + }, + { + "epoch": 1.6185839476978718, + "grad_norm": 0.314453125, + "learning_rate": 8.606277617005282e-05, + "loss": 0.6929, + "step": 2940 + }, + { + "epoch": 1.6191403533175686, + "grad_norm": 0.322265625, + "learning_rate": 8.600496530013857e-05, + "loss": 0.7401, + "step": 2941 + }, + { + "epoch": 1.6196967589372653, + "grad_norm": 0.345703125, + "learning_rate": 8.59471592005467e-05, + "loss": 0.7672, + "step": 2942 + }, + { + "epoch": 1.620253164556962, + "grad_norm": 0.328125, + "learning_rate": 8.588935789098087e-05, + "loss": 0.727, + "step": 2943 + }, + { + "epoch": 1.6208095701766587, + "grad_norm": 0.298828125, + "learning_rate": 8.58315613911431e-05, + "loss": 0.5882, + "step": 2944 + }, + { + "epoch": 1.6213659757963557, + "grad_norm": 0.322265625, + "learning_rate": 8.577376972073388e-05, + "loss": 0.7213, + "step": 2945 + }, + { + "epoch": 1.6219223814160522, + "grad_norm": 0.330078125, + "learning_rate": 8.571598289945195e-05, + "loss": 0.6805, + "step": 2946 + }, + { + "epoch": 1.622478787035749, + "grad_norm": 0.33203125, + "learning_rate": 8.565820094699442e-05, + "loss": 0.7019, + "step": 2947 + }, + { + "epoch": 1.6230351926554458, + "grad_norm": 0.3203125, + "learning_rate": 8.560042388305671e-05, + "loss": 0.6233, + "step": 2948 + }, + { + "epoch": 1.6235915982751425, + "grad_norm": 0.314453125, + "learning_rate": 8.554265172733263e-05, + "loss": 0.6029, + "step": 2949 + }, + { + "epoch": 1.6241480038948395, + "grad_norm": 0.328125, + "learning_rate": 8.548488449951429e-05, + "loss": 0.6847, + "step": 2950 + }, + { + "epoch": 1.624704409514536, + "grad_norm": 0.31640625, + "learning_rate": 8.542712221929212e-05, + "loss": 0.6027, + "step": 2951 + }, + { + "epoch": 1.625260815134233, + "grad_norm": 0.326171875, + "learning_rate": 8.536936490635482e-05, + "loss": 0.692, + "step": 2952 + }, + { + "epoch": 1.6258172207539296, + "grad_norm": 0.33203125, + "learning_rate": 8.531161258038954e-05, + "loss": 0.6799, + "step": 2953 + }, + { + "epoch": 1.6263736263736264, + "grad_norm": 0.3203125, + "learning_rate": 8.52538652610816e-05, + "loss": 0.6631, + "step": 2954 + }, + { + "epoch": 1.626930031993323, + "grad_norm": 0.35546875, + "learning_rate": 8.519612296811462e-05, + "loss": 0.7291, + "step": 2955 + }, + { + "epoch": 1.6274864376130198, + "grad_norm": 0.349609375, + "learning_rate": 8.513838572117056e-05, + "loss": 0.7482, + "step": 2956 + }, + { + "epoch": 1.6280428432327168, + "grad_norm": 0.3515625, + "learning_rate": 8.508065353992962e-05, + "loss": 0.7863, + "step": 2957 + }, + { + "epoch": 1.6285992488524133, + "grad_norm": 0.33203125, + "learning_rate": 8.502292644407029e-05, + "loss": 0.7251, + "step": 2958 + }, + { + "epoch": 1.6291556544721102, + "grad_norm": 0.34765625, + "learning_rate": 8.496520445326933e-05, + "loss": 0.7886, + "step": 2959 + }, + { + "epoch": 1.629712060091807, + "grad_norm": 0.333984375, + "learning_rate": 8.490748758720173e-05, + "loss": 0.7144, + "step": 2960 + }, + { + "epoch": 1.6302684657115036, + "grad_norm": 0.31640625, + "learning_rate": 8.484977586554079e-05, + "loss": 0.648, + "step": 2961 + }, + { + "epoch": 1.6308248713312006, + "grad_norm": 0.34375, + "learning_rate": 8.479206930795801e-05, + "loss": 0.7811, + "step": 2962 + }, + { + "epoch": 1.631381276950897, + "grad_norm": 0.314453125, + "learning_rate": 8.473436793412314e-05, + "loss": 0.6209, + "step": 2963 + }, + { + "epoch": 1.631937682570594, + "grad_norm": 0.314453125, + "learning_rate": 8.467667176370418e-05, + "loss": 0.669, + "step": 2964 + }, + { + "epoch": 1.6324940881902907, + "grad_norm": 0.32421875, + "learning_rate": 8.461898081636733e-05, + "loss": 0.6646, + "step": 2965 + }, + { + "epoch": 1.6330504938099875, + "grad_norm": 0.314453125, + "learning_rate": 8.456129511177702e-05, + "loss": 0.6645, + "step": 2966 + }, + { + "epoch": 1.6336068994296844, + "grad_norm": 0.326171875, + "learning_rate": 8.450361466959592e-05, + "loss": 0.7527, + "step": 2967 + }, + { + "epoch": 1.634163305049381, + "grad_norm": 0.330078125, + "learning_rate": 8.444593950948481e-05, + "loss": 0.7361, + "step": 2968 + }, + { + "epoch": 1.6347197106690778, + "grad_norm": 0.3203125, + "learning_rate": 8.438826965110283e-05, + "loss": 0.6279, + "step": 2969 + }, + { + "epoch": 1.6352761162887746, + "grad_norm": 0.330078125, + "learning_rate": 8.433060511410719e-05, + "loss": 0.6921, + "step": 2970 + }, + { + "epoch": 1.6358325219084713, + "grad_norm": 0.3359375, + "learning_rate": 8.427294591815333e-05, + "loss": 0.7315, + "step": 2971 + }, + { + "epoch": 1.636388927528168, + "grad_norm": 0.345703125, + "learning_rate": 8.421529208289484e-05, + "loss": 0.8319, + "step": 2972 + }, + { + "epoch": 1.6369453331478647, + "grad_norm": 0.31640625, + "learning_rate": 8.415764362798351e-05, + "loss": 0.6628, + "step": 2973 + }, + { + "epoch": 1.6375017387675617, + "grad_norm": 0.345703125, + "learning_rate": 8.41000005730693e-05, + "loss": 0.8388, + "step": 2974 + }, + { + "epoch": 1.6380581443872582, + "grad_norm": 0.3203125, + "learning_rate": 8.40423629378003e-05, + "loss": 0.6801, + "step": 2975 + }, + { + "epoch": 1.638614550006955, + "grad_norm": 0.349609375, + "learning_rate": 8.398473074182278e-05, + "loss": 0.7476, + "step": 2976 + }, + { + "epoch": 1.6391709556266518, + "grad_norm": 0.328125, + "learning_rate": 8.392710400478113e-05, + "loss": 0.7123, + "step": 2977 + }, + { + "epoch": 1.6397273612463485, + "grad_norm": 0.32421875, + "learning_rate": 8.386948274631793e-05, + "loss": 0.6786, + "step": 2978 + }, + { + "epoch": 1.6402837668660455, + "grad_norm": 0.322265625, + "learning_rate": 8.381186698607384e-05, + "loss": 0.7191, + "step": 2979 + }, + { + "epoch": 1.640840172485742, + "grad_norm": 0.32421875, + "learning_rate": 8.375425674368767e-05, + "loss": 0.7247, + "step": 2980 + }, + { + "epoch": 1.641396578105439, + "grad_norm": 0.322265625, + "learning_rate": 8.369665203879633e-05, + "loss": 0.6686, + "step": 2981 + }, + { + "epoch": 1.6419529837251357, + "grad_norm": 0.322265625, + "learning_rate": 8.363905289103486e-05, + "loss": 0.7129, + "step": 2982 + }, + { + "epoch": 1.6425093893448324, + "grad_norm": 0.3125, + "learning_rate": 8.358145932003643e-05, + "loss": 0.6306, + "step": 2983 + }, + { + "epoch": 1.643065794964529, + "grad_norm": 0.328125, + "learning_rate": 8.352387134543224e-05, + "loss": 0.7094, + "step": 2984 + }, + { + "epoch": 1.6436222005842258, + "grad_norm": 0.318359375, + "learning_rate": 8.346628898685161e-05, + "loss": 0.681, + "step": 2985 + }, + { + "epoch": 1.6441786062039228, + "grad_norm": 0.33203125, + "learning_rate": 8.340871226392202e-05, + "loss": 0.7378, + "step": 2986 + }, + { + "epoch": 1.6447350118236193, + "grad_norm": 0.337890625, + "learning_rate": 8.335114119626893e-05, + "loss": 0.7647, + "step": 2987 + }, + { + "epoch": 1.6452914174433162, + "grad_norm": 0.33203125, + "learning_rate": 8.329357580351594e-05, + "loss": 0.7211, + "step": 2988 + }, + { + "epoch": 1.645847823063013, + "grad_norm": 0.33203125, + "learning_rate": 8.323601610528467e-05, + "loss": 0.7693, + "step": 2989 + }, + { + "epoch": 1.6464042286827096, + "grad_norm": 0.33984375, + "learning_rate": 8.31784621211948e-05, + "loss": 0.751, + "step": 2990 + }, + { + "epoch": 1.6469606343024066, + "grad_norm": 0.328125, + "learning_rate": 8.312091387086406e-05, + "loss": 0.7263, + "step": 2991 + }, + { + "epoch": 1.647517039922103, + "grad_norm": 0.337890625, + "learning_rate": 8.306337137390831e-05, + "loss": 0.7473, + "step": 2992 + }, + { + "epoch": 1.6480734455418, + "grad_norm": 0.314453125, + "learning_rate": 8.300583464994132e-05, + "loss": 0.6768, + "step": 2993 + }, + { + "epoch": 1.6486298511614967, + "grad_norm": 0.3125, + "learning_rate": 8.294830371857498e-05, + "loss": 0.6799, + "step": 2994 + }, + { + "epoch": 1.6491862567811935, + "grad_norm": 0.322265625, + "learning_rate": 8.289077859941919e-05, + "loss": 0.6828, + "step": 2995 + }, + { + "epoch": 1.6497426624008904, + "grad_norm": 0.33984375, + "learning_rate": 8.283325931208184e-05, + "loss": 0.7569, + "step": 2996 + }, + { + "epoch": 1.650299068020587, + "grad_norm": 0.337890625, + "learning_rate": 8.277574587616887e-05, + "loss": 0.7867, + "step": 2997 + }, + { + "epoch": 1.6508554736402838, + "grad_norm": 0.32421875, + "learning_rate": 8.271823831128418e-05, + "loss": 0.6443, + "step": 2998 + }, + { + "epoch": 1.6514118792599806, + "grad_norm": 0.3125, + "learning_rate": 8.266073663702972e-05, + "loss": 0.625, + "step": 2999 + }, + { + "epoch": 1.6519682848796773, + "grad_norm": 0.318359375, + "learning_rate": 8.260324087300539e-05, + "loss": 0.6984, + "step": 3000 + }, + { + "epoch": 1.652524690499374, + "grad_norm": 0.318359375, + "learning_rate": 8.254575103880906e-05, + "loss": 0.6621, + "step": 3001 + }, + { + "epoch": 1.6530810961190707, + "grad_norm": 0.34375, + "learning_rate": 8.248826715403669e-05, + "loss": 0.7133, + "step": 3002 + }, + { + "epoch": 1.6536375017387677, + "grad_norm": 0.326171875, + "learning_rate": 8.243078923828209e-05, + "loss": 0.6627, + "step": 3003 + }, + { + "epoch": 1.6541939073584642, + "grad_norm": 0.322265625, + "learning_rate": 8.237331731113709e-05, + "loss": 0.7144, + "step": 3004 + }, + { + "epoch": 1.654750312978161, + "grad_norm": 0.328125, + "learning_rate": 8.231585139219146e-05, + "loss": 0.6648, + "step": 3005 + }, + { + "epoch": 1.6553067185978578, + "grad_norm": 0.333984375, + "learning_rate": 8.225839150103294e-05, + "loss": 0.6821, + "step": 3006 + }, + { + "epoch": 1.6558631242175545, + "grad_norm": 0.322265625, + "learning_rate": 8.22009376572472e-05, + "loss": 0.6986, + "step": 3007 + }, + { + "epoch": 1.6564195298372515, + "grad_norm": 0.3125, + "learning_rate": 8.214348988041782e-05, + "loss": 0.6301, + "step": 3008 + }, + { + "epoch": 1.656975935456948, + "grad_norm": 0.306640625, + "learning_rate": 8.208604819012638e-05, + "loss": 0.5965, + "step": 3009 + }, + { + "epoch": 1.657532341076645, + "grad_norm": 0.341796875, + "learning_rate": 8.202861260595232e-05, + "loss": 0.7793, + "step": 3010 + }, + { + "epoch": 1.6580887466963417, + "grad_norm": 0.318359375, + "learning_rate": 8.197118314747309e-05, + "loss": 0.6792, + "step": 3011 + }, + { + "epoch": 1.6586451523160384, + "grad_norm": 0.328125, + "learning_rate": 8.191375983426398e-05, + "loss": 0.7053, + "step": 3012 + }, + { + "epoch": 1.6592015579357353, + "grad_norm": 0.31640625, + "learning_rate": 8.185634268589814e-05, + "loss": 0.6604, + "step": 3013 + }, + { + "epoch": 1.6597579635554318, + "grad_norm": 0.361328125, + "learning_rate": 8.179893172194674e-05, + "loss": 0.7893, + "step": 3014 + }, + { + "epoch": 1.6603143691751288, + "grad_norm": 0.30859375, + "learning_rate": 8.174152696197873e-05, + "loss": 0.6425, + "step": 3015 + }, + { + "epoch": 1.6608707747948255, + "grad_norm": 0.318359375, + "learning_rate": 8.168412842556102e-05, + "loss": 0.7134, + "step": 3016 + }, + { + "epoch": 1.6614271804145222, + "grad_norm": 0.330078125, + "learning_rate": 8.162673613225837e-05, + "loss": 0.7172, + "step": 3017 + }, + { + "epoch": 1.661983586034219, + "grad_norm": 0.328125, + "learning_rate": 8.156935010163338e-05, + "loss": 0.7394, + "step": 3018 + }, + { + "epoch": 1.6625399916539156, + "grad_norm": 0.3203125, + "learning_rate": 8.151197035324662e-05, + "loss": 0.646, + "step": 3019 + }, + { + "epoch": 1.6630963972736126, + "grad_norm": 0.326171875, + "learning_rate": 8.145459690665638e-05, + "loss": 0.7177, + "step": 3020 + }, + { + "epoch": 1.663652802893309, + "grad_norm": 0.34375, + "learning_rate": 8.139722978141892e-05, + "loss": 0.7873, + "step": 3021 + }, + { + "epoch": 1.664209208513006, + "grad_norm": 0.341796875, + "learning_rate": 8.133986899708825e-05, + "loss": 0.767, + "step": 3022 + }, + { + "epoch": 1.6647656141327027, + "grad_norm": 0.322265625, + "learning_rate": 8.12825145732163e-05, + "loss": 0.6911, + "step": 3023 + }, + { + "epoch": 1.6653220197523995, + "grad_norm": 0.318359375, + "learning_rate": 8.122516652935278e-05, + "loss": 0.6797, + "step": 3024 + }, + { + "epoch": 1.6658784253720964, + "grad_norm": 0.333984375, + "learning_rate": 8.116782488504524e-05, + "loss": 0.7333, + "step": 3025 + }, + { + "epoch": 1.666434830991793, + "grad_norm": 0.32421875, + "learning_rate": 8.1110489659839e-05, + "loss": 0.7368, + "step": 3026 + }, + { + "epoch": 1.6669912366114898, + "grad_norm": 0.318359375, + "learning_rate": 8.105316087327734e-05, + "loss": 0.6546, + "step": 3027 + }, + { + "epoch": 1.6675476422311866, + "grad_norm": 0.3125, + "learning_rate": 8.09958385449012e-05, + "loss": 0.6336, + "step": 3028 + }, + { + "epoch": 1.6681040478508833, + "grad_norm": 0.3359375, + "learning_rate": 8.093852269424938e-05, + "loss": 0.7057, + "step": 3029 + }, + { + "epoch": 1.66866045347058, + "grad_norm": 0.33203125, + "learning_rate": 8.088121334085843e-05, + "loss": 0.7545, + "step": 3030 + }, + { + "epoch": 1.6692168590902767, + "grad_norm": 0.30859375, + "learning_rate": 8.082391050426273e-05, + "loss": 0.6032, + "step": 3031 + }, + { + "epoch": 1.6697732647099737, + "grad_norm": 0.341796875, + "learning_rate": 8.076661420399443e-05, + "loss": 0.796, + "step": 3032 + }, + { + "epoch": 1.6703296703296702, + "grad_norm": 0.341796875, + "learning_rate": 8.070932445958343e-05, + "loss": 0.7505, + "step": 3033 + }, + { + "epoch": 1.6708860759493671, + "grad_norm": 0.33203125, + "learning_rate": 8.065204129055739e-05, + "loss": 0.7076, + "step": 3034 + }, + { + "epoch": 1.6714424815690638, + "grad_norm": 0.341796875, + "learning_rate": 8.059476471644178e-05, + "loss": 0.8116, + "step": 3035 + }, + { + "epoch": 1.6719988871887606, + "grad_norm": 0.337890625, + "learning_rate": 8.053749475675978e-05, + "loss": 0.7695, + "step": 3036 + }, + { + "epoch": 1.6725552928084575, + "grad_norm": 0.375, + "learning_rate": 8.048023143103234e-05, + "loss": 0.7014, + "step": 3037 + }, + { + "epoch": 1.673111698428154, + "grad_norm": 0.33203125, + "learning_rate": 8.042297475877811e-05, + "loss": 0.7452, + "step": 3038 + }, + { + "epoch": 1.673668104047851, + "grad_norm": 0.314453125, + "learning_rate": 8.036572475951352e-05, + "loss": 0.6748, + "step": 3039 + }, + { + "epoch": 1.6742245096675477, + "grad_norm": 0.337890625, + "learning_rate": 8.030848145275267e-05, + "loss": 0.7851, + "step": 3040 + }, + { + "epoch": 1.6747809152872444, + "grad_norm": 0.337890625, + "learning_rate": 8.025124485800743e-05, + "loss": 0.772, + "step": 3041 + }, + { + "epoch": 1.6753373209069413, + "grad_norm": 0.333984375, + "learning_rate": 8.019401499478734e-05, + "loss": 0.7544, + "step": 3042 + }, + { + "epoch": 1.6758937265266378, + "grad_norm": 0.318359375, + "learning_rate": 8.013679188259973e-05, + "loss": 0.6766, + "step": 3043 + }, + { + "epoch": 1.6764501321463348, + "grad_norm": 0.314453125, + "learning_rate": 8.007957554094951e-05, + "loss": 0.6393, + "step": 3044 + }, + { + "epoch": 1.6770065377660315, + "grad_norm": 0.330078125, + "learning_rate": 8.002236598933937e-05, + "loss": 0.6625, + "step": 3045 + }, + { + "epoch": 1.6775629433857282, + "grad_norm": 0.326171875, + "learning_rate": 7.996516324726964e-05, + "loss": 0.6738, + "step": 3046 + }, + { + "epoch": 1.678119349005425, + "grad_norm": 0.3359375, + "learning_rate": 7.990796733423834e-05, + "loss": 0.686, + "step": 3047 + }, + { + "epoch": 1.6786757546251216, + "grad_norm": 0.345703125, + "learning_rate": 7.985077826974116e-05, + "loss": 0.8127, + "step": 3048 + }, + { + "epoch": 1.6792321602448186, + "grad_norm": 0.34375, + "learning_rate": 7.97935960732715e-05, + "loss": 0.7389, + "step": 3049 + }, + { + "epoch": 1.679788565864515, + "grad_norm": 0.341796875, + "learning_rate": 7.973642076432033e-05, + "loss": 0.7938, + "step": 3050 + }, + { + "epoch": 1.680344971484212, + "grad_norm": 0.330078125, + "learning_rate": 7.967925236237634e-05, + "loss": 0.7465, + "step": 3051 + }, + { + "epoch": 1.6809013771039087, + "grad_norm": 0.32421875, + "learning_rate": 7.962209088692587e-05, + "loss": 0.6959, + "step": 3052 + }, + { + "epoch": 1.6814577827236055, + "grad_norm": 0.33984375, + "learning_rate": 7.956493635745286e-05, + "loss": 0.8103, + "step": 3053 + }, + { + "epoch": 1.6820141883433024, + "grad_norm": 0.34375, + "learning_rate": 7.950778879343891e-05, + "loss": 0.7522, + "step": 3054 + }, + { + "epoch": 1.682570593962999, + "grad_norm": 0.333984375, + "learning_rate": 7.945064821436324e-05, + "loss": 0.6949, + "step": 3055 + }, + { + "epoch": 1.6831269995826958, + "grad_norm": 0.337890625, + "learning_rate": 7.939351463970266e-05, + "loss": 0.7312, + "step": 3056 + }, + { + "epoch": 1.6836834052023926, + "grad_norm": 0.345703125, + "learning_rate": 7.933638808893165e-05, + "loss": 0.7179, + "step": 3057 + }, + { + "epoch": 1.6842398108220893, + "grad_norm": 0.318359375, + "learning_rate": 7.927926858152223e-05, + "loss": 0.638, + "step": 3058 + }, + { + "epoch": 1.684796216441786, + "grad_norm": 0.326171875, + "learning_rate": 7.922215613694403e-05, + "loss": 0.678, + "step": 3059 + }, + { + "epoch": 1.6853526220614827, + "grad_norm": 0.330078125, + "learning_rate": 7.916505077466436e-05, + "loss": 0.7344, + "step": 3060 + }, + { + "epoch": 1.6859090276811797, + "grad_norm": 0.310546875, + "learning_rate": 7.910795251414804e-05, + "loss": 0.614, + "step": 3061 + }, + { + "epoch": 1.6864654333008764, + "grad_norm": 0.34765625, + "learning_rate": 7.905086137485747e-05, + "loss": 0.7833, + "step": 3062 + }, + { + "epoch": 1.6870218389205731, + "grad_norm": 0.357421875, + "learning_rate": 7.899377737625262e-05, + "loss": 0.722, + "step": 3063 + }, + { + "epoch": 1.6875782445402698, + "grad_norm": 0.32421875, + "learning_rate": 7.893670053779104e-05, + "loss": 0.6724, + "step": 3064 + }, + { + "epoch": 1.6881346501599666, + "grad_norm": 0.35546875, + "learning_rate": 7.887963087892788e-05, + "loss": 0.791, + "step": 3065 + }, + { + "epoch": 1.6886910557796635, + "grad_norm": 0.3125, + "learning_rate": 7.882256841911574e-05, + "loss": 0.6306, + "step": 3066 + }, + { + "epoch": 1.68924746139936, + "grad_norm": 0.33984375, + "learning_rate": 7.876551317780481e-05, + "loss": 0.7223, + "step": 3067 + }, + { + "epoch": 1.689803867019057, + "grad_norm": 0.333984375, + "learning_rate": 7.870846517444293e-05, + "loss": 0.7811, + "step": 3068 + }, + { + "epoch": 1.6903602726387537, + "grad_norm": 0.337890625, + "learning_rate": 7.865142442847532e-05, + "loss": 0.7057, + "step": 3069 + }, + { + "epoch": 1.6909166782584504, + "grad_norm": 0.330078125, + "learning_rate": 7.85943909593448e-05, + "loss": 0.7355, + "step": 3070 + }, + { + "epoch": 1.6914730838781473, + "grad_norm": 0.33203125, + "learning_rate": 7.85373647864917e-05, + "loss": 0.7619, + "step": 3071 + }, + { + "epoch": 1.6920294894978438, + "grad_norm": 0.31640625, + "learning_rate": 7.848034592935385e-05, + "loss": 0.6392, + "step": 3072 + }, + { + "epoch": 1.6925858951175408, + "grad_norm": 0.318359375, + "learning_rate": 7.842333440736659e-05, + "loss": 0.6825, + "step": 3073 + }, + { + "epoch": 1.6931423007372375, + "grad_norm": 0.326171875, + "learning_rate": 7.836633023996278e-05, + "loss": 0.7056, + "step": 3074 + }, + { + "epoch": 1.6936987063569342, + "grad_norm": 0.314453125, + "learning_rate": 7.830933344657272e-05, + "loss": 0.711, + "step": 3075 + }, + { + "epoch": 1.694255111976631, + "grad_norm": 0.337890625, + "learning_rate": 7.825234404662429e-05, + "loss": 0.7537, + "step": 3076 + }, + { + "epoch": 1.6948115175963276, + "grad_norm": 0.296875, + "learning_rate": 7.819536205954278e-05, + "loss": 0.5464, + "step": 3077 + }, + { + "epoch": 1.6953679232160246, + "grad_norm": 0.326171875, + "learning_rate": 7.813838750475095e-05, + "loss": 0.7123, + "step": 3078 + }, + { + "epoch": 1.695924328835721, + "grad_norm": 0.34375, + "learning_rate": 7.808142040166904e-05, + "loss": 0.7296, + "step": 3079 + }, + { + "epoch": 1.696480734455418, + "grad_norm": 0.3359375, + "learning_rate": 7.802446076971477e-05, + "loss": 0.7502, + "step": 3080 + }, + { + "epoch": 1.6970371400751147, + "grad_norm": 0.328125, + "learning_rate": 7.79675086283033e-05, + "loss": 0.6857, + "step": 3081 + }, + { + "epoch": 1.6975935456948115, + "grad_norm": 0.333984375, + "learning_rate": 7.791056399684722e-05, + "loss": 0.7681, + "step": 3082 + }, + { + "epoch": 1.6981499513145084, + "grad_norm": 0.314453125, + "learning_rate": 7.785362689475658e-05, + "loss": 0.6666, + "step": 3083 + }, + { + "epoch": 1.698706356934205, + "grad_norm": 0.33203125, + "learning_rate": 7.779669734143882e-05, + "loss": 0.73, + "step": 3084 + }, + { + "epoch": 1.6992627625539019, + "grad_norm": 0.333984375, + "learning_rate": 7.773977535629894e-05, + "loss": 0.7452, + "step": 3085 + }, + { + "epoch": 1.6998191681735986, + "grad_norm": 0.353515625, + "learning_rate": 7.76828609587392e-05, + "loss": 0.7709, + "step": 3086 + }, + { + "epoch": 1.7003755737932953, + "grad_norm": 0.328125, + "learning_rate": 7.762595416815934e-05, + "loss": 0.714, + "step": 3087 + }, + { + "epoch": 1.7009319794129922, + "grad_norm": 0.33984375, + "learning_rate": 7.756905500395654e-05, + "loss": 0.7206, + "step": 3088 + }, + { + "epoch": 1.7014883850326887, + "grad_norm": 0.341796875, + "learning_rate": 7.75121634855253e-05, + "loss": 0.7826, + "step": 3089 + }, + { + "epoch": 1.7020447906523857, + "grad_norm": 0.326171875, + "learning_rate": 7.74552796322576e-05, + "loss": 0.7041, + "step": 3090 + }, + { + "epoch": 1.7026011962720824, + "grad_norm": 0.341796875, + "learning_rate": 7.739840346354275e-05, + "loss": 0.7233, + "step": 3091 + }, + { + "epoch": 1.7031576018917791, + "grad_norm": 0.328125, + "learning_rate": 7.734153499876745e-05, + "loss": 0.7218, + "step": 3092 + }, + { + "epoch": 1.7037140075114758, + "grad_norm": 0.32421875, + "learning_rate": 7.728467425731578e-05, + "loss": 0.7136, + "step": 3093 + }, + { + "epoch": 1.7042704131311726, + "grad_norm": 0.326171875, + "learning_rate": 7.722782125856923e-05, + "loss": 0.7551, + "step": 3094 + }, + { + "epoch": 1.7048268187508695, + "grad_norm": 0.291015625, + "learning_rate": 7.717097602190659e-05, + "loss": 0.5817, + "step": 3095 + }, + { + "epoch": 1.705383224370566, + "grad_norm": 0.310546875, + "learning_rate": 7.7114138566704e-05, + "loss": 0.6164, + "step": 3096 + }, + { + "epoch": 1.705939629990263, + "grad_norm": 0.31640625, + "learning_rate": 7.7057308912335e-05, + "loss": 0.6545, + "step": 3097 + }, + { + "epoch": 1.7064960356099597, + "grad_norm": 0.33203125, + "learning_rate": 7.700048707817044e-05, + "loss": 0.6774, + "step": 3098 + }, + { + "epoch": 1.7070524412296564, + "grad_norm": 0.3125, + "learning_rate": 7.69436730835785e-05, + "loss": 0.6516, + "step": 3099 + }, + { + "epoch": 1.7076088468493533, + "grad_norm": 0.33203125, + "learning_rate": 7.688686694792465e-05, + "loss": 0.7391, + "step": 3100 + }, + { + "epoch": 1.7081652524690498, + "grad_norm": 0.318359375, + "learning_rate": 7.683006869057183e-05, + "loss": 0.718, + "step": 3101 + }, + { + "epoch": 1.7087216580887468, + "grad_norm": 0.3125, + "learning_rate": 7.677327833088014e-05, + "loss": 0.6579, + "step": 3102 + }, + { + "epoch": 1.7092780637084435, + "grad_norm": 0.322265625, + "learning_rate": 7.671649588820705e-05, + "loss": 0.7098, + "step": 3103 + }, + { + "epoch": 1.7098344693281402, + "grad_norm": 0.326171875, + "learning_rate": 7.665972138190729e-05, + "loss": 0.6756, + "step": 3104 + }, + { + "epoch": 1.710390874947837, + "grad_norm": 0.380859375, + "learning_rate": 7.660295483133295e-05, + "loss": 0.7559, + "step": 3105 + }, + { + "epoch": 1.7109472805675336, + "grad_norm": 0.322265625, + "learning_rate": 7.654619625583334e-05, + "loss": 0.6492, + "step": 3106 + }, + { + "epoch": 1.7115036861872306, + "grad_norm": 0.33203125, + "learning_rate": 7.648944567475514e-05, + "loss": 0.7314, + "step": 3107 + }, + { + "epoch": 1.712060091806927, + "grad_norm": 0.349609375, + "learning_rate": 7.643270310744217e-05, + "loss": 0.7397, + "step": 3108 + }, + { + "epoch": 1.712616497426624, + "grad_norm": 0.34375, + "learning_rate": 7.637596857323568e-05, + "loss": 0.757, + "step": 3109 + }, + { + "epoch": 1.7131729030463207, + "grad_norm": 0.337890625, + "learning_rate": 7.631924209147408e-05, + "loss": 0.7677, + "step": 3110 + }, + { + "epoch": 1.7137293086660175, + "grad_norm": 0.349609375, + "learning_rate": 7.626252368149302e-05, + "loss": 0.7734, + "step": 3111 + }, + { + "epoch": 1.7142857142857144, + "grad_norm": 0.310546875, + "learning_rate": 7.62058133626255e-05, + "loss": 0.6734, + "step": 3112 + }, + { + "epoch": 1.714842119905411, + "grad_norm": 0.322265625, + "learning_rate": 7.614911115420163e-05, + "loss": 0.7024, + "step": 3113 + }, + { + "epoch": 1.7153985255251079, + "grad_norm": 0.341796875, + "learning_rate": 7.609241707554884e-05, + "loss": 0.7851, + "step": 3114 + }, + { + "epoch": 1.7159549311448046, + "grad_norm": 0.380859375, + "learning_rate": 7.603573114599181e-05, + "loss": 0.7338, + "step": 3115 + }, + { + "epoch": 1.7165113367645013, + "grad_norm": 0.306640625, + "learning_rate": 7.597905338485236e-05, + "loss": 0.6673, + "step": 3116 + }, + { + "epoch": 1.7170677423841982, + "grad_norm": 0.341796875, + "learning_rate": 7.592238381144955e-05, + "loss": 0.7961, + "step": 3117 + }, + { + "epoch": 1.7176241480038947, + "grad_norm": 0.310546875, + "learning_rate": 7.586572244509973e-05, + "loss": 0.679, + "step": 3118 + }, + { + "epoch": 1.7181805536235917, + "grad_norm": 0.353515625, + "learning_rate": 7.580906930511639e-05, + "loss": 0.7772, + "step": 3119 + }, + { + "epoch": 1.7187369592432884, + "grad_norm": 0.318359375, + "learning_rate": 7.57524244108102e-05, + "loss": 0.6743, + "step": 3120 + }, + { + "epoch": 1.7192933648629851, + "grad_norm": 0.33984375, + "learning_rate": 7.569578778148901e-05, + "loss": 0.7183, + "step": 3121 + }, + { + "epoch": 1.7198497704826818, + "grad_norm": 0.298828125, + "learning_rate": 7.563915943645792e-05, + "loss": 0.5856, + "step": 3122 + }, + { + "epoch": 1.7204061761023786, + "grad_norm": 0.3359375, + "learning_rate": 7.558253939501914e-05, + "loss": 0.7203, + "step": 3123 + }, + { + "epoch": 1.7209625817220755, + "grad_norm": 0.33984375, + "learning_rate": 7.552592767647208e-05, + "loss": 0.7436, + "step": 3124 + }, + { + "epoch": 1.721518987341772, + "grad_norm": 0.337890625, + "learning_rate": 7.546932430011329e-05, + "loss": 0.7516, + "step": 3125 + }, + { + "epoch": 1.722075392961469, + "grad_norm": 0.34375, + "learning_rate": 7.541272928523653e-05, + "loss": 0.7454, + "step": 3126 + }, + { + "epoch": 1.7226317985811657, + "grad_norm": 0.34375, + "learning_rate": 7.535614265113266e-05, + "loss": 0.787, + "step": 3127 + }, + { + "epoch": 1.7231882042008624, + "grad_norm": 0.3671875, + "learning_rate": 7.529956441708972e-05, + "loss": 0.7091, + "step": 3128 + }, + { + "epoch": 1.7237446098205593, + "grad_norm": 0.3359375, + "learning_rate": 7.524299460239282e-05, + "loss": 0.7175, + "step": 3129 + }, + { + "epoch": 1.7243010154402558, + "grad_norm": 0.328125, + "learning_rate": 7.518643322632428e-05, + "loss": 0.742, + "step": 3130 + }, + { + "epoch": 1.7248574210599528, + "grad_norm": 0.34765625, + "learning_rate": 7.51298803081635e-05, + "loss": 0.7534, + "step": 3131 + }, + { + "epoch": 1.7254138266796495, + "grad_norm": 0.31640625, + "learning_rate": 7.507333586718701e-05, + "loss": 0.6486, + "step": 3132 + }, + { + "epoch": 1.7259702322993462, + "grad_norm": 0.318359375, + "learning_rate": 7.501679992266842e-05, + "loss": 0.6582, + "step": 3133 + }, + { + "epoch": 1.726526637919043, + "grad_norm": 0.3359375, + "learning_rate": 7.496027249387851e-05, + "loss": 0.7183, + "step": 3134 + }, + { + "epoch": 1.7270830435387396, + "grad_norm": 0.3515625, + "learning_rate": 7.49037536000851e-05, + "loss": 0.8283, + "step": 3135 + }, + { + "epoch": 1.7276394491584366, + "grad_norm": 0.330078125, + "learning_rate": 7.48472432605531e-05, + "loss": 0.7372, + "step": 3136 + }, + { + "epoch": 1.7281958547781333, + "grad_norm": 0.3359375, + "learning_rate": 7.479074149454456e-05, + "loss": 0.7601, + "step": 3137 + }, + { + "epoch": 1.72875226039783, + "grad_norm": 0.318359375, + "learning_rate": 7.473424832131855e-05, + "loss": 0.6519, + "step": 3138 + }, + { + "epoch": 1.7293086660175268, + "grad_norm": 0.3203125, + "learning_rate": 7.467776376013122e-05, + "loss": 0.6702, + "step": 3139 + }, + { + "epoch": 1.7298650716372235, + "grad_norm": 0.318359375, + "learning_rate": 7.462128783023583e-05, + "loss": 0.6905, + "step": 3140 + }, + { + "epoch": 1.7304214772569204, + "grad_norm": 0.33203125, + "learning_rate": 7.456482055088258e-05, + "loss": 0.738, + "step": 3141 + }, + { + "epoch": 1.730977882876617, + "grad_norm": 0.330078125, + "learning_rate": 7.450836194131893e-05, + "loss": 0.7107, + "step": 3142 + }, + { + "epoch": 1.7315342884963139, + "grad_norm": 0.33203125, + "learning_rate": 7.445191202078918e-05, + "loss": 0.7509, + "step": 3143 + }, + { + "epoch": 1.7320906941160106, + "grad_norm": 0.32421875, + "learning_rate": 7.439547080853477e-05, + "loss": 0.6999, + "step": 3144 + }, + { + "epoch": 1.7326470997357073, + "grad_norm": 0.33984375, + "learning_rate": 7.433903832379416e-05, + "loss": 0.7028, + "step": 3145 + }, + { + "epoch": 1.7332035053554042, + "grad_norm": 0.3359375, + "learning_rate": 7.42826145858028e-05, + "loss": 0.7517, + "step": 3146 + }, + { + "epoch": 1.7337599109751007, + "grad_norm": 0.337890625, + "learning_rate": 7.42261996137932e-05, + "loss": 0.7304, + "step": 3147 + }, + { + "epoch": 1.7343163165947977, + "grad_norm": 0.33984375, + "learning_rate": 7.416979342699486e-05, + "loss": 0.7556, + "step": 3148 + }, + { + "epoch": 1.7348727222144944, + "grad_norm": 0.3203125, + "learning_rate": 7.411339604463426e-05, + "loss": 0.6988, + "step": 3149 + }, + { + "epoch": 1.7354291278341911, + "grad_norm": 0.3125, + "learning_rate": 7.405700748593498e-05, + "loss": 0.6057, + "step": 3150 + }, + { + "epoch": 1.7359855334538878, + "grad_norm": 0.3359375, + "learning_rate": 7.400062777011747e-05, + "loss": 0.772, + "step": 3151 + }, + { + "epoch": 1.7365419390735846, + "grad_norm": 0.345703125, + "learning_rate": 7.394425691639926e-05, + "loss": 0.7567, + "step": 3152 + }, + { + "epoch": 1.7370983446932815, + "grad_norm": 0.3359375, + "learning_rate": 7.388789494399477e-05, + "loss": 0.7537, + "step": 3153 + }, + { + "epoch": 1.737654750312978, + "grad_norm": 0.3203125, + "learning_rate": 7.383154187211547e-05, + "loss": 0.6661, + "step": 3154 + }, + { + "epoch": 1.738211155932675, + "grad_norm": 0.32421875, + "learning_rate": 7.377519771996977e-05, + "loss": 0.6761, + "step": 3155 + }, + { + "epoch": 1.7387675615523717, + "grad_norm": 0.3359375, + "learning_rate": 7.371886250676302e-05, + "loss": 0.7068, + "step": 3156 + }, + { + "epoch": 1.7393239671720684, + "grad_norm": 0.326171875, + "learning_rate": 7.366253625169755e-05, + "loss": 0.636, + "step": 3157 + }, + { + "epoch": 1.7398803727917653, + "grad_norm": 0.34765625, + "learning_rate": 7.360621897397257e-05, + "loss": 0.76, + "step": 3158 + }, + { + "epoch": 1.7404367784114618, + "grad_norm": 0.328125, + "learning_rate": 7.354991069278439e-05, + "loss": 0.7092, + "step": 3159 + }, + { + "epoch": 1.7409931840311588, + "grad_norm": 0.32421875, + "learning_rate": 7.349361142732608e-05, + "loss": 0.6917, + "step": 3160 + }, + { + "epoch": 1.7415495896508555, + "grad_norm": 0.32421875, + "learning_rate": 7.343732119678774e-05, + "loss": 0.6703, + "step": 3161 + }, + { + "epoch": 1.7421059952705522, + "grad_norm": 0.31640625, + "learning_rate": 7.338104002035633e-05, + "loss": 0.6398, + "step": 3162 + }, + { + "epoch": 1.7426624008902492, + "grad_norm": 0.349609375, + "learning_rate": 7.332476791721577e-05, + "loss": 0.7847, + "step": 3163 + }, + { + "epoch": 1.7432188065099457, + "grad_norm": 0.318359375, + "learning_rate": 7.326850490654685e-05, + "loss": 0.6727, + "step": 3164 + }, + { + "epoch": 1.7437752121296426, + "grad_norm": 0.326171875, + "learning_rate": 7.32122510075273e-05, + "loss": 0.6705, + "step": 3165 + }, + { + "epoch": 1.7443316177493393, + "grad_norm": 0.345703125, + "learning_rate": 7.315600623933167e-05, + "loss": 0.7465, + "step": 3166 + }, + { + "epoch": 1.744888023369036, + "grad_norm": 0.345703125, + "learning_rate": 7.309977062113154e-05, + "loss": 0.7312, + "step": 3167 + }, + { + "epoch": 1.7454444289887328, + "grad_norm": 0.314453125, + "learning_rate": 7.30435441720952e-05, + "loss": 0.5896, + "step": 3168 + }, + { + "epoch": 1.7460008346084295, + "grad_norm": 0.3203125, + "learning_rate": 7.298732691138796e-05, + "loss": 0.6629, + "step": 3169 + }, + { + "epoch": 1.7465572402281264, + "grad_norm": 0.34375, + "learning_rate": 7.293111885817188e-05, + "loss": 0.8049, + "step": 3170 + }, + { + "epoch": 1.747113645847823, + "grad_norm": 0.330078125, + "learning_rate": 7.287492003160597e-05, + "loss": 0.6748, + "step": 3171 + }, + { + "epoch": 1.7476700514675199, + "grad_norm": 0.341796875, + "learning_rate": 7.281873045084606e-05, + "loss": 0.759, + "step": 3172 + }, + { + "epoch": 1.7482264570872166, + "grad_norm": 0.33203125, + "learning_rate": 7.276255013504481e-05, + "loss": 0.7165, + "step": 3173 + }, + { + "epoch": 1.7487828627069133, + "grad_norm": 0.33984375, + "learning_rate": 7.27063791033517e-05, + "loss": 0.7724, + "step": 3174 + }, + { + "epoch": 1.7493392683266102, + "grad_norm": 0.32421875, + "learning_rate": 7.26502173749132e-05, + "loss": 0.6862, + "step": 3175 + }, + { + "epoch": 1.7498956739463067, + "grad_norm": 0.3359375, + "learning_rate": 7.259406496887242e-05, + "loss": 0.69, + "step": 3176 + }, + { + "epoch": 1.7504520795660037, + "grad_norm": 0.3671875, + "learning_rate": 7.253792190436939e-05, + "loss": 0.8408, + "step": 3177 + }, + { + "epoch": 1.7510084851857004, + "grad_norm": 0.326171875, + "learning_rate": 7.248178820054093e-05, + "loss": 0.7029, + "step": 3178 + }, + { + "epoch": 1.7515648908053971, + "grad_norm": 0.318359375, + "learning_rate": 7.242566387652065e-05, + "loss": 0.6992, + "step": 3179 + }, + { + "epoch": 1.7521212964250938, + "grad_norm": 0.337890625, + "learning_rate": 7.236954895143902e-05, + "loss": 0.7389, + "step": 3180 + }, + { + "epoch": 1.7526777020447906, + "grad_norm": 0.3125, + "learning_rate": 7.231344344442325e-05, + "loss": 0.6238, + "step": 3181 + }, + { + "epoch": 1.7532341076644875, + "grad_norm": 0.3203125, + "learning_rate": 7.225734737459736e-05, + "loss": 0.6545, + "step": 3182 + }, + { + "epoch": 1.753790513284184, + "grad_norm": 0.341796875, + "learning_rate": 7.220126076108218e-05, + "loss": 0.762, + "step": 3183 + }, + { + "epoch": 1.754346918903881, + "grad_norm": 0.333984375, + "learning_rate": 7.214518362299529e-05, + "loss": 0.6942, + "step": 3184 + }, + { + "epoch": 1.7549033245235777, + "grad_norm": 0.333984375, + "learning_rate": 7.208911597945103e-05, + "loss": 0.7132, + "step": 3185 + }, + { + "epoch": 1.7554597301432744, + "grad_norm": 0.34765625, + "learning_rate": 7.20330578495605e-05, + "loss": 0.7813, + "step": 3186 + }, + { + "epoch": 1.7560161357629713, + "grad_norm": 0.310546875, + "learning_rate": 7.19770092524316e-05, + "loss": 0.6511, + "step": 3187 + }, + { + "epoch": 1.7565725413826678, + "grad_norm": 0.314453125, + "learning_rate": 7.192097020716895e-05, + "loss": 0.5954, + "step": 3188 + }, + { + "epoch": 1.7571289470023648, + "grad_norm": 0.3203125, + "learning_rate": 7.186494073287391e-05, + "loss": 0.666, + "step": 3189 + }, + { + "epoch": 1.7576853526220615, + "grad_norm": 0.337890625, + "learning_rate": 7.180892084864458e-05, + "loss": 0.7496, + "step": 3190 + }, + { + "epoch": 1.7582417582417582, + "grad_norm": 0.31640625, + "learning_rate": 7.175291057357577e-05, + "loss": 0.6791, + "step": 3191 + }, + { + "epoch": 1.7587981638614552, + "grad_norm": 0.326171875, + "learning_rate": 7.169690992675914e-05, + "loss": 0.7171, + "step": 3192 + }, + { + "epoch": 1.7593545694811517, + "grad_norm": 0.337890625, + "learning_rate": 7.164091892728288e-05, + "loss": 0.6987, + "step": 3193 + }, + { + "epoch": 1.7599109751008486, + "grad_norm": 0.328125, + "learning_rate": 7.158493759423201e-05, + "loss": 0.712, + "step": 3194 + }, + { + "epoch": 1.7604673807205453, + "grad_norm": 0.318359375, + "learning_rate": 7.152896594668823e-05, + "loss": 0.7056, + "step": 3195 + }, + { + "epoch": 1.761023786340242, + "grad_norm": 0.3125, + "learning_rate": 7.147300400372994e-05, + "loss": 0.6694, + "step": 3196 + }, + { + "epoch": 1.7615801919599388, + "grad_norm": 0.32421875, + "learning_rate": 7.141705178443221e-05, + "loss": 0.66, + "step": 3197 + }, + { + "epoch": 1.7621365975796355, + "grad_norm": 0.349609375, + "learning_rate": 7.136110930786682e-05, + "loss": 0.7498, + "step": 3198 + }, + { + "epoch": 1.7626930031993324, + "grad_norm": 0.34375, + "learning_rate": 7.130517659310216e-05, + "loss": 0.7275, + "step": 3199 + }, + { + "epoch": 1.763249408819029, + "grad_norm": 0.302734375, + "learning_rate": 7.124925365920347e-05, + "loss": 0.5852, + "step": 3200 + }, + { + "epoch": 1.7638058144387259, + "grad_norm": 0.328125, + "learning_rate": 7.11933405252325e-05, + "loss": 0.7147, + "step": 3201 + }, + { + "epoch": 1.7643622200584226, + "grad_norm": 0.328125, + "learning_rate": 7.113743721024768e-05, + "loss": 0.7254, + "step": 3202 + }, + { + "epoch": 1.7649186256781193, + "grad_norm": 0.34375, + "learning_rate": 7.108154373330413e-05, + "loss": 0.7856, + "step": 3203 + }, + { + "epoch": 1.7654750312978162, + "grad_norm": 0.337890625, + "learning_rate": 7.102566011345357e-05, + "loss": 0.7396, + "step": 3204 + }, + { + "epoch": 1.7660314369175127, + "grad_norm": 0.32421875, + "learning_rate": 7.096978636974443e-05, + "loss": 0.7022, + "step": 3205 + }, + { + "epoch": 1.7665878425372097, + "grad_norm": 0.326171875, + "learning_rate": 7.091392252122172e-05, + "loss": 0.6869, + "step": 3206 + }, + { + "epoch": 1.7671442481569064, + "grad_norm": 0.318359375, + "learning_rate": 7.085806858692705e-05, + "loss": 0.6662, + "step": 3207 + }, + { + "epoch": 1.7677006537766031, + "grad_norm": 0.326171875, + "learning_rate": 7.080222458589875e-05, + "loss": 0.703, + "step": 3208 + }, + { + "epoch": 1.7682570593963, + "grad_norm": 0.322265625, + "learning_rate": 7.07463905371717e-05, + "loss": 0.7245, + "step": 3209 + }, + { + "epoch": 1.7688134650159966, + "grad_norm": 0.318359375, + "learning_rate": 7.069056645977736e-05, + "loss": 0.6679, + "step": 3210 + }, + { + "epoch": 1.7693698706356935, + "grad_norm": 0.3046875, + "learning_rate": 7.063475237274385e-05, + "loss": 0.6326, + "step": 3211 + }, + { + "epoch": 1.7699262762553902, + "grad_norm": 0.33203125, + "learning_rate": 7.057894829509585e-05, + "loss": 0.6385, + "step": 3212 + }, + { + "epoch": 1.770482681875087, + "grad_norm": 0.326171875, + "learning_rate": 7.052315424585464e-05, + "loss": 0.6695, + "step": 3213 + }, + { + "epoch": 1.7710390874947837, + "grad_norm": 0.3125, + "learning_rate": 7.046737024403807e-05, + "loss": 0.665, + "step": 3214 + }, + { + "epoch": 1.7715954931144804, + "grad_norm": 0.337890625, + "learning_rate": 7.041159630866055e-05, + "loss": 0.7247, + "step": 3215 + }, + { + "epoch": 1.7721518987341773, + "grad_norm": 0.333984375, + "learning_rate": 7.035583245873314e-05, + "loss": 0.7378, + "step": 3216 + }, + { + "epoch": 1.7727083043538738, + "grad_norm": 0.33203125, + "learning_rate": 7.030007871326341e-05, + "loss": 0.7304, + "step": 3217 + }, + { + "epoch": 1.7732647099735708, + "grad_norm": 0.333984375, + "learning_rate": 7.024433509125543e-05, + "loss": 0.7267, + "step": 3218 + }, + { + "epoch": 1.7738211155932675, + "grad_norm": 0.326171875, + "learning_rate": 7.01886016117099e-05, + "loss": 0.7158, + "step": 3219 + }, + { + "epoch": 1.7743775212129642, + "grad_norm": 0.32421875, + "learning_rate": 7.013287829362405e-05, + "loss": 0.67, + "step": 3220 + }, + { + "epoch": 1.7749339268326612, + "grad_norm": 0.3203125, + "learning_rate": 7.007716515599158e-05, + "loss": 0.6976, + "step": 3221 + }, + { + "epoch": 1.7754903324523577, + "grad_norm": 0.330078125, + "learning_rate": 7.002146221780281e-05, + "loss": 0.7439, + "step": 3222 + }, + { + "epoch": 1.7760467380720546, + "grad_norm": 0.314453125, + "learning_rate": 6.996576949804449e-05, + "loss": 0.6182, + "step": 3223 + }, + { + "epoch": 1.7766031436917513, + "grad_norm": 0.326171875, + "learning_rate": 6.991008701570004e-05, + "loss": 0.7231, + "step": 3224 + }, + { + "epoch": 1.777159549311448, + "grad_norm": 0.3359375, + "learning_rate": 6.98544147897492e-05, + "loss": 0.743, + "step": 3225 + }, + { + "epoch": 1.7777159549311448, + "grad_norm": 0.3203125, + "learning_rate": 6.979875283916835e-05, + "loss": 0.6778, + "step": 3226 + }, + { + "epoch": 1.7782723605508415, + "grad_norm": 0.33984375, + "learning_rate": 6.97431011829303e-05, + "loss": 0.7176, + "step": 3227 + }, + { + "epoch": 1.7788287661705384, + "grad_norm": 0.34375, + "learning_rate": 6.968745984000438e-05, + "loss": 0.7246, + "step": 3228 + }, + { + "epoch": 1.779385171790235, + "grad_norm": 0.330078125, + "learning_rate": 6.963182882935641e-05, + "loss": 0.6826, + "step": 3229 + }, + { + "epoch": 1.7799415774099319, + "grad_norm": 0.33203125, + "learning_rate": 6.957620816994867e-05, + "loss": 0.7414, + "step": 3230 + }, + { + "epoch": 1.7804979830296286, + "grad_norm": 0.337890625, + "learning_rate": 6.952059788073992e-05, + "loss": 0.7728, + "step": 3231 + }, + { + "epoch": 1.7810543886493253, + "grad_norm": 0.33203125, + "learning_rate": 6.946499798068533e-05, + "loss": 0.7267, + "step": 3232 + }, + { + "epoch": 1.7816107942690222, + "grad_norm": 0.322265625, + "learning_rate": 6.940940848873666e-05, + "loss": 0.6468, + "step": 3233 + }, + { + "epoch": 1.7821671998887187, + "grad_norm": 0.33203125, + "learning_rate": 6.935382942384202e-05, + "loss": 0.6919, + "step": 3234 + }, + { + "epoch": 1.7827236055084157, + "grad_norm": 0.333984375, + "learning_rate": 6.929826080494598e-05, + "loss": 0.7716, + "step": 3235 + }, + { + "epoch": 1.7832800111281124, + "grad_norm": 0.322265625, + "learning_rate": 6.924270265098958e-05, + "loss": 0.6758, + "step": 3236 + }, + { + "epoch": 1.7838364167478091, + "grad_norm": 0.322265625, + "learning_rate": 6.918715498091027e-05, + "loss": 0.6879, + "step": 3237 + }, + { + "epoch": 1.784392822367506, + "grad_norm": 0.330078125, + "learning_rate": 6.913161781364188e-05, + "loss": 0.6379, + "step": 3238 + }, + { + "epoch": 1.7849492279872026, + "grad_norm": 0.333984375, + "learning_rate": 6.907609116811477e-05, + "loss": 0.7262, + "step": 3239 + }, + { + "epoch": 1.7855056336068995, + "grad_norm": 0.33203125, + "learning_rate": 6.902057506325559e-05, + "loss": 0.6914, + "step": 3240 + }, + { + "epoch": 1.7860620392265962, + "grad_norm": 0.333984375, + "learning_rate": 6.896506951798752e-05, + "loss": 0.7587, + "step": 3241 + }, + { + "epoch": 1.786618444846293, + "grad_norm": 0.32421875, + "learning_rate": 6.890957455123007e-05, + "loss": 0.6734, + "step": 3242 + }, + { + "epoch": 1.7871748504659897, + "grad_norm": 0.318359375, + "learning_rate": 6.885409018189913e-05, + "loss": 0.6342, + "step": 3243 + }, + { + "epoch": 1.7877312560856864, + "grad_norm": 0.322265625, + "learning_rate": 6.879861642890701e-05, + "loss": 0.6495, + "step": 3244 + }, + { + "epoch": 1.7882876617053833, + "grad_norm": 0.33984375, + "learning_rate": 6.87431533111624e-05, + "loss": 0.7334, + "step": 3245 + }, + { + "epoch": 1.7888440673250798, + "grad_norm": 0.33203125, + "learning_rate": 6.868770084757037e-05, + "loss": 0.7495, + "step": 3246 + }, + { + "epoch": 1.7894004729447768, + "grad_norm": 0.337890625, + "learning_rate": 6.863225905703233e-05, + "loss": 0.7151, + "step": 3247 + }, + { + "epoch": 1.7899568785644735, + "grad_norm": 0.328125, + "learning_rate": 6.857682795844603e-05, + "loss": 0.6749, + "step": 3248 + }, + { + "epoch": 1.7905132841841702, + "grad_norm": 0.318359375, + "learning_rate": 6.852140757070569e-05, + "loss": 0.666, + "step": 3249 + }, + { + "epoch": 1.7910696898038672, + "grad_norm": 0.34765625, + "learning_rate": 6.846599791270178e-05, + "loss": 0.8306, + "step": 3250 + }, + { + "epoch": 1.7916260954235637, + "grad_norm": 0.32421875, + "learning_rate": 6.841059900332112e-05, + "loss": 0.7596, + "step": 3251 + }, + { + "epoch": 1.7921825010432606, + "grad_norm": 0.326171875, + "learning_rate": 6.835521086144692e-05, + "loss": 0.7302, + "step": 3252 + }, + { + "epoch": 1.7927389066629573, + "grad_norm": 0.330078125, + "learning_rate": 6.829983350595861e-05, + "loss": 0.6916, + "step": 3253 + }, + { + "epoch": 1.793295312282654, + "grad_norm": 0.302734375, + "learning_rate": 6.824446695573208e-05, + "loss": 0.5942, + "step": 3254 + }, + { + "epoch": 1.7938517179023508, + "grad_norm": 0.3125, + "learning_rate": 6.818911122963945e-05, + "loss": 0.6869, + "step": 3255 + }, + { + "epoch": 1.7944081235220475, + "grad_norm": 0.30859375, + "learning_rate": 6.813376634654913e-05, + "loss": 0.6252, + "step": 3256 + }, + { + "epoch": 1.7949645291417444, + "grad_norm": 0.64453125, + "learning_rate": 6.807843232532596e-05, + "loss": 1.6073, + "step": 3257 + }, + { + "epoch": 1.7955209347614411, + "grad_norm": 0.31640625, + "learning_rate": 6.802310918483094e-05, + "loss": 0.6278, + "step": 3258 + }, + { + "epoch": 1.7960773403811379, + "grad_norm": 0.3125, + "learning_rate": 6.796779694392145e-05, + "loss": 0.6261, + "step": 3259 + }, + { + "epoch": 1.7966337460008346, + "grad_norm": 0.322265625, + "learning_rate": 6.791249562145112e-05, + "loss": 0.6858, + "step": 3260 + }, + { + "epoch": 1.7971901516205313, + "grad_norm": 0.33984375, + "learning_rate": 6.785720523626983e-05, + "loss": 0.7385, + "step": 3261 + }, + { + "epoch": 1.7977465572402282, + "grad_norm": 0.34765625, + "learning_rate": 6.780192580722377e-05, + "loss": 0.7482, + "step": 3262 + }, + { + "epoch": 1.7983029628599247, + "grad_norm": 0.330078125, + "learning_rate": 6.77466573531554e-05, + "loss": 0.7308, + "step": 3263 + }, + { + "epoch": 1.7988593684796217, + "grad_norm": 0.328125, + "learning_rate": 6.76913998929034e-05, + "loss": 0.7119, + "step": 3264 + }, + { + "epoch": 1.7994157740993184, + "grad_norm": 0.3359375, + "learning_rate": 6.763615344530274e-05, + "loss": 0.7659, + "step": 3265 + }, + { + "epoch": 1.7999721797190151, + "grad_norm": 0.322265625, + "learning_rate": 6.758091802918465e-05, + "loss": 0.7055, + "step": 3266 + }, + { + "epoch": 1.800528585338712, + "grad_norm": 0.326171875, + "learning_rate": 6.752569366337655e-05, + "loss": 0.679, + "step": 3267 + }, + { + "epoch": 1.8010849909584086, + "grad_norm": 0.3359375, + "learning_rate": 6.747048036670212e-05, + "loss": 0.7704, + "step": 3268 + }, + { + "epoch": 1.8016413965781055, + "grad_norm": 0.330078125, + "learning_rate": 6.741527815798127e-05, + "loss": 0.7181, + "step": 3269 + }, + { + "epoch": 1.8021978021978022, + "grad_norm": 0.328125, + "learning_rate": 6.736008705603013e-05, + "loss": 0.726, + "step": 3270 + }, + { + "epoch": 1.802754207817499, + "grad_norm": 0.3359375, + "learning_rate": 6.730490707966104e-05, + "loss": 0.7282, + "step": 3271 + }, + { + "epoch": 1.8033106134371957, + "grad_norm": 0.314453125, + "learning_rate": 6.724973824768253e-05, + "loss": 0.6759, + "step": 3272 + }, + { + "epoch": 1.8038670190568924, + "grad_norm": 0.333984375, + "learning_rate": 6.719458057889931e-05, + "loss": 0.7553, + "step": 3273 + }, + { + "epoch": 1.8044234246765893, + "grad_norm": 0.3359375, + "learning_rate": 6.713943409211243e-05, + "loss": 0.7402, + "step": 3274 + }, + { + "epoch": 1.8049798302962858, + "grad_norm": 0.68359375, + "learning_rate": 6.708429880611896e-05, + "loss": 1.5996, + "step": 3275 + }, + { + "epoch": 1.8055362359159828, + "grad_norm": 0.326171875, + "learning_rate": 6.702917473971221e-05, + "loss": 0.75, + "step": 3276 + }, + { + "epoch": 1.8060926415356795, + "grad_norm": 0.330078125, + "learning_rate": 6.697406191168171e-05, + "loss": 0.7314, + "step": 3277 + }, + { + "epoch": 1.8066490471553762, + "grad_norm": 0.318359375, + "learning_rate": 6.691896034081307e-05, + "loss": 0.6474, + "step": 3278 + }, + { + "epoch": 1.8072054527750732, + "grad_norm": 0.328125, + "learning_rate": 6.686387004588815e-05, + "loss": 0.7095, + "step": 3279 + }, + { + "epoch": 1.8077618583947697, + "grad_norm": 0.341796875, + "learning_rate": 6.680879104568489e-05, + "loss": 0.7352, + "step": 3280 + }, + { + "epoch": 1.8083182640144666, + "grad_norm": 0.314453125, + "learning_rate": 6.675372335897743e-05, + "loss": 0.6489, + "step": 3281 + }, + { + "epoch": 1.8088746696341633, + "grad_norm": 0.306640625, + "learning_rate": 6.669866700453607e-05, + "loss": 0.6261, + "step": 3282 + }, + { + "epoch": 1.80943107525386, + "grad_norm": 0.33203125, + "learning_rate": 6.66436220011272e-05, + "loss": 0.7173, + "step": 3283 + }, + { + "epoch": 1.809987480873557, + "grad_norm": 0.333984375, + "learning_rate": 6.658858836751337e-05, + "loss": 0.7458, + "step": 3284 + }, + { + "epoch": 1.8105438864932535, + "grad_norm": 0.3203125, + "learning_rate": 6.653356612245325e-05, + "loss": 0.5983, + "step": 3285 + }, + { + "epoch": 1.8111002921129504, + "grad_norm": 0.337890625, + "learning_rate": 6.64785552847016e-05, + "loss": 0.7382, + "step": 3286 + }, + { + "epoch": 1.8116566977326471, + "grad_norm": 0.30859375, + "learning_rate": 6.642355587300932e-05, + "loss": 0.6262, + "step": 3287 + }, + { + "epoch": 1.8122131033523439, + "grad_norm": 0.345703125, + "learning_rate": 6.636856790612344e-05, + "loss": 0.7902, + "step": 3288 + }, + { + "epoch": 1.8127695089720406, + "grad_norm": 0.314453125, + "learning_rate": 6.631359140278697e-05, + "loss": 0.6979, + "step": 3289 + }, + { + "epoch": 1.8133259145917373, + "grad_norm": 0.333984375, + "learning_rate": 6.625862638173924e-05, + "loss": 0.751, + "step": 3290 + }, + { + "epoch": 1.8138823202114343, + "grad_norm": 0.318359375, + "learning_rate": 6.620367286171544e-05, + "loss": 0.6716, + "step": 3291 + }, + { + "epoch": 1.8144387258311307, + "grad_norm": 0.341796875, + "learning_rate": 6.614873086144696e-05, + "loss": 0.7716, + "step": 3292 + }, + { + "epoch": 1.8149951314508277, + "grad_norm": 0.35546875, + "learning_rate": 6.609380039966121e-05, + "loss": 0.7985, + "step": 3293 + }, + { + "epoch": 1.8155515370705244, + "grad_norm": 0.3359375, + "learning_rate": 6.60388814950817e-05, + "loss": 0.7319, + "step": 3294 + }, + { + "epoch": 1.8161079426902211, + "grad_norm": 0.3125, + "learning_rate": 6.598397416642798e-05, + "loss": 0.6392, + "step": 3295 + }, + { + "epoch": 1.816664348309918, + "grad_norm": 0.326171875, + "learning_rate": 6.592907843241568e-05, + "loss": 0.6527, + "step": 3296 + }, + { + "epoch": 1.8172207539296146, + "grad_norm": 0.357421875, + "learning_rate": 6.587419431175645e-05, + "loss": 0.7794, + "step": 3297 + }, + { + "epoch": 1.8177771595493115, + "grad_norm": 0.3203125, + "learning_rate": 6.581932182315797e-05, + "loss": 0.6956, + "step": 3298 + }, + { + "epoch": 1.8183335651690082, + "grad_norm": 0.357421875, + "learning_rate": 6.5764460985324e-05, + "loss": 0.8092, + "step": 3299 + }, + { + "epoch": 1.818889970788705, + "grad_norm": 0.32421875, + "learning_rate": 6.570961181695435e-05, + "loss": 0.7234, + "step": 3300 + }, + { + "epoch": 1.8194463764084017, + "grad_norm": 0.306640625, + "learning_rate": 6.565477433674473e-05, + "loss": 0.624, + "step": 3301 + }, + { + "epoch": 1.8200027820280984, + "grad_norm": 0.330078125, + "learning_rate": 6.5599948563387e-05, + "loss": 0.6949, + "step": 3302 + }, + { + "epoch": 1.8205591876477953, + "grad_norm": 0.333984375, + "learning_rate": 6.554513451556894e-05, + "loss": 0.7474, + "step": 3303 + }, + { + "epoch": 1.8211155932674918, + "grad_norm": 0.318359375, + "learning_rate": 6.549033221197439e-05, + "loss": 0.6641, + "step": 3304 + }, + { + "epoch": 1.8216719988871888, + "grad_norm": 0.3359375, + "learning_rate": 6.543554167128316e-05, + "loss": 0.7217, + "step": 3305 + }, + { + "epoch": 1.8222284045068855, + "grad_norm": 0.318359375, + "learning_rate": 6.5380762912171e-05, + "loss": 0.6955, + "step": 3306 + }, + { + "epoch": 1.8227848101265822, + "grad_norm": 0.328125, + "learning_rate": 6.532599595330982e-05, + "loss": 0.7, + "step": 3307 + }, + { + "epoch": 1.8233412157462792, + "grad_norm": 0.318359375, + "learning_rate": 6.527124081336731e-05, + "loss": 0.6261, + "step": 3308 + }, + { + "epoch": 1.8238976213659757, + "grad_norm": 0.34375, + "learning_rate": 6.521649751100721e-05, + "loss": 0.7717, + "step": 3309 + }, + { + "epoch": 1.8244540269856726, + "grad_norm": 0.3359375, + "learning_rate": 6.516176606488924e-05, + "loss": 0.7757, + "step": 3310 + }, + { + "epoch": 1.8250104326053693, + "grad_norm": 0.33984375, + "learning_rate": 6.510704649366907e-05, + "loss": 0.689, + "step": 3311 + }, + { + "epoch": 1.825566838225066, + "grad_norm": 0.3203125, + "learning_rate": 6.505233881599829e-05, + "loss": 0.6868, + "step": 3312 + }, + { + "epoch": 1.826123243844763, + "grad_norm": 0.337890625, + "learning_rate": 6.499764305052445e-05, + "loss": 0.7566, + "step": 3313 + }, + { + "epoch": 1.8266796494644595, + "grad_norm": 0.322265625, + "learning_rate": 6.494295921589105e-05, + "loss": 0.6813, + "step": 3314 + }, + { + "epoch": 1.8272360550841564, + "grad_norm": 0.314453125, + "learning_rate": 6.488828733073758e-05, + "loss": 0.6462, + "step": 3315 + }, + { + "epoch": 1.8277924607038532, + "grad_norm": 0.326171875, + "learning_rate": 6.483362741369934e-05, + "loss": 0.6924, + "step": 3316 + }, + { + "epoch": 1.8283488663235499, + "grad_norm": 0.330078125, + "learning_rate": 6.477897948340765e-05, + "loss": 0.7112, + "step": 3317 + }, + { + "epoch": 1.8289052719432466, + "grad_norm": 0.32421875, + "learning_rate": 6.472434355848971e-05, + "loss": 0.6333, + "step": 3318 + }, + { + "epoch": 1.8294616775629433, + "grad_norm": 0.341796875, + "learning_rate": 6.466971965756858e-05, + "loss": 0.7648, + "step": 3319 + }, + { + "epoch": 1.8300180831826403, + "grad_norm": 0.341796875, + "learning_rate": 6.46151077992633e-05, + "loss": 0.7868, + "step": 3320 + }, + { + "epoch": 1.8305744888023368, + "grad_norm": 0.34375, + "learning_rate": 6.456050800218874e-05, + "loss": 0.7442, + "step": 3321 + }, + { + "epoch": 1.8311308944220337, + "grad_norm": 0.33203125, + "learning_rate": 6.45059202849557e-05, + "loss": 0.6999, + "step": 3322 + }, + { + "epoch": 1.8316873000417304, + "grad_norm": 0.333984375, + "learning_rate": 6.445134466617089e-05, + "loss": 0.7836, + "step": 3323 + }, + { + "epoch": 1.8322437056614271, + "grad_norm": 0.337890625, + "learning_rate": 6.43967811644368e-05, + "loss": 0.7049, + "step": 3324 + }, + { + "epoch": 1.832800111281124, + "grad_norm": 0.330078125, + "learning_rate": 6.43422297983519e-05, + "loss": 0.7271, + "step": 3325 + }, + { + "epoch": 1.8333565169008206, + "grad_norm": 0.33203125, + "learning_rate": 6.428769058651041e-05, + "loss": 0.7317, + "step": 3326 + }, + { + "epoch": 1.8339129225205175, + "grad_norm": 0.306640625, + "learning_rate": 6.423316354750253e-05, + "loss": 0.645, + "step": 3327 + }, + { + "epoch": 1.8344693281402142, + "grad_norm": 0.330078125, + "learning_rate": 6.41786486999142e-05, + "loss": 0.7404, + "step": 3328 + }, + { + "epoch": 1.835025733759911, + "grad_norm": 0.34765625, + "learning_rate": 6.412414606232728e-05, + "loss": 0.7605, + "step": 3329 + }, + { + "epoch": 1.8355821393796077, + "grad_norm": 0.326171875, + "learning_rate": 6.40696556533194e-05, + "loss": 0.6615, + "step": 3330 + }, + { + "epoch": 1.8361385449993044, + "grad_norm": 0.328125, + "learning_rate": 6.401517749146412e-05, + "loss": 0.7039, + "step": 3331 + }, + { + "epoch": 1.8366949506190013, + "grad_norm": 0.333984375, + "learning_rate": 6.396071159533075e-05, + "loss": 0.7691, + "step": 3332 + }, + { + "epoch": 1.837251356238698, + "grad_norm": 0.3125, + "learning_rate": 6.390625798348444e-05, + "loss": 0.6287, + "step": 3333 + }, + { + "epoch": 1.8378077618583948, + "grad_norm": 0.330078125, + "learning_rate": 6.385181667448612e-05, + "loss": 0.6919, + "step": 3334 + }, + { + "epoch": 1.8383641674780915, + "grad_norm": 0.32421875, + "learning_rate": 6.379738768689261e-05, + "loss": 0.7444, + "step": 3335 + }, + { + "epoch": 1.8389205730977882, + "grad_norm": 0.31640625, + "learning_rate": 6.374297103925641e-05, + "loss": 0.674, + "step": 3336 + }, + { + "epoch": 1.8394769787174852, + "grad_norm": 0.33203125, + "learning_rate": 6.368856675012595e-05, + "loss": 0.7459, + "step": 3337 + }, + { + "epoch": 1.8400333843371817, + "grad_norm": 0.3359375, + "learning_rate": 6.363417483804532e-05, + "loss": 0.749, + "step": 3338 + }, + { + "epoch": 1.8405897899568786, + "grad_norm": 0.32421875, + "learning_rate": 6.357979532155447e-05, + "loss": 0.6839, + "step": 3339 + }, + { + "epoch": 1.8411461955765753, + "grad_norm": 0.341796875, + "learning_rate": 6.352542821918912e-05, + "loss": 0.7087, + "step": 3340 + }, + { + "epoch": 1.841702601196272, + "grad_norm": 0.330078125, + "learning_rate": 6.347107354948072e-05, + "loss": 0.7029, + "step": 3341 + }, + { + "epoch": 1.842259006815969, + "grad_norm": 0.302734375, + "learning_rate": 6.341673133095653e-05, + "loss": 0.592, + "step": 3342 + }, + { + "epoch": 1.8428154124356655, + "grad_norm": 0.328125, + "learning_rate": 6.336240158213953e-05, + "loss": 0.72, + "step": 3343 + }, + { + "epoch": 1.8433718180553624, + "grad_norm": 0.32421875, + "learning_rate": 6.330808432154844e-05, + "loss": 0.705, + "step": 3344 + }, + { + "epoch": 1.8439282236750592, + "grad_norm": 0.32421875, + "learning_rate": 6.325377956769777e-05, + "loss": 0.7179, + "step": 3345 + }, + { + "epoch": 1.8444846292947559, + "grad_norm": 0.318359375, + "learning_rate": 6.319948733909774e-05, + "loss": 0.6467, + "step": 3346 + }, + { + "epoch": 1.8450410349144526, + "grad_norm": 0.330078125, + "learning_rate": 6.314520765425424e-05, + "loss": 0.6794, + "step": 3347 + }, + { + "epoch": 1.8455974405341493, + "grad_norm": 0.3125, + "learning_rate": 6.309094053166903e-05, + "loss": 0.6588, + "step": 3348 + }, + { + "epoch": 1.8461538461538463, + "grad_norm": 0.30859375, + "learning_rate": 6.303668598983949e-05, + "loss": 0.6287, + "step": 3349 + }, + { + "epoch": 1.8467102517735428, + "grad_norm": 0.37890625, + "learning_rate": 6.298244404725872e-05, + "loss": 0.8185, + "step": 3350 + }, + { + "epoch": 1.8472666573932397, + "grad_norm": 0.330078125, + "learning_rate": 6.29282147224155e-05, + "loss": 0.7748, + "step": 3351 + }, + { + "epoch": 1.8478230630129364, + "grad_norm": 0.318359375, + "learning_rate": 6.287399803379435e-05, + "loss": 0.6746, + "step": 3352 + }, + { + "epoch": 1.8483794686326331, + "grad_norm": 0.32421875, + "learning_rate": 6.28197939998755e-05, + "loss": 0.7148, + "step": 3353 + }, + { + "epoch": 1.84893587425233, + "grad_norm": 0.326171875, + "learning_rate": 6.276560263913483e-05, + "loss": 0.6878, + "step": 3354 + }, + { + "epoch": 1.8494922798720266, + "grad_norm": 0.326171875, + "learning_rate": 6.271142397004386e-05, + "loss": 0.6446, + "step": 3355 + }, + { + "epoch": 1.8500486854917235, + "grad_norm": 0.32421875, + "learning_rate": 6.26572580110699e-05, + "loss": 0.6817, + "step": 3356 + }, + { + "epoch": 1.8506050911114202, + "grad_norm": 0.32421875, + "learning_rate": 6.260310478067583e-05, + "loss": 0.6624, + "step": 3357 + }, + { + "epoch": 1.851161496731117, + "grad_norm": 0.314453125, + "learning_rate": 6.254896429732026e-05, + "loss": 0.656, + "step": 3358 + }, + { + "epoch": 1.851717902350814, + "grad_norm": 0.3203125, + "learning_rate": 6.249483657945736e-05, + "loss": 0.6688, + "step": 3359 + }, + { + "epoch": 1.8522743079705104, + "grad_norm": 0.3359375, + "learning_rate": 6.244072164553705e-05, + "loss": 0.7349, + "step": 3360 + }, + { + "epoch": 1.8528307135902073, + "grad_norm": 0.328125, + "learning_rate": 6.238661951400482e-05, + "loss": 0.7017, + "step": 3361 + }, + { + "epoch": 1.853387119209904, + "grad_norm": 0.33203125, + "learning_rate": 6.233253020330185e-05, + "loss": 0.7051, + "step": 3362 + }, + { + "epoch": 1.8539435248296008, + "grad_norm": 0.32421875, + "learning_rate": 6.227845373186487e-05, + "loss": 0.6676, + "step": 3363 + }, + { + "epoch": 1.8544999304492975, + "grad_norm": 0.33203125, + "learning_rate": 6.222439011812635e-05, + "loss": 0.7275, + "step": 3364 + }, + { + "epoch": 1.8550563360689942, + "grad_norm": 0.328125, + "learning_rate": 6.217033938051433e-05, + "loss": 0.7192, + "step": 3365 + }, + { + "epoch": 1.8556127416886912, + "grad_norm": 0.34375, + "learning_rate": 6.211630153745241e-05, + "loss": 0.7779, + "step": 3366 + }, + { + "epoch": 1.8561691473083877, + "grad_norm": 0.318359375, + "learning_rate": 6.206227660735981e-05, + "loss": 0.6455, + "step": 3367 + }, + { + "epoch": 1.8567255529280846, + "grad_norm": 0.330078125, + "learning_rate": 6.200826460865143e-05, + "loss": 0.721, + "step": 3368 + }, + { + "epoch": 1.8572819585477813, + "grad_norm": 0.345703125, + "learning_rate": 6.195426555973766e-05, + "loss": 0.756, + "step": 3369 + }, + { + "epoch": 1.857838364167478, + "grad_norm": 0.328125, + "learning_rate": 6.190027947902455e-05, + "loss": 0.6996, + "step": 3370 + }, + { + "epoch": 1.858394769787175, + "grad_norm": 0.330078125, + "learning_rate": 6.184630638491366e-05, + "loss": 0.7374, + "step": 3371 + }, + { + "epoch": 1.8589511754068715, + "grad_norm": 0.341796875, + "learning_rate": 6.179234629580214e-05, + "loss": 0.7902, + "step": 3372 + }, + { + "epoch": 1.8595075810265684, + "grad_norm": 0.3359375, + "learning_rate": 6.173839923008281e-05, + "loss": 0.7347, + "step": 3373 + }, + { + "epoch": 1.8600639866462652, + "grad_norm": 0.328125, + "learning_rate": 6.168446520614393e-05, + "loss": 0.6949, + "step": 3374 + }, + { + "epoch": 1.8606203922659619, + "grad_norm": 0.31640625, + "learning_rate": 6.163054424236934e-05, + "loss": 0.6839, + "step": 3375 + }, + { + "epoch": 1.8611767978856586, + "grad_norm": 0.322265625, + "learning_rate": 6.157663635713847e-05, + "loss": 0.704, + "step": 3376 + }, + { + "epoch": 1.8617332035053553, + "grad_norm": 0.34375, + "learning_rate": 6.152274156882624e-05, + "loss": 0.7211, + "step": 3377 + }, + { + "epoch": 1.8622896091250523, + "grad_norm": 0.328125, + "learning_rate": 6.146885989580315e-05, + "loss": 0.7162, + "step": 3378 + }, + { + "epoch": 1.8628460147447488, + "grad_norm": 0.32421875, + "learning_rate": 6.141499135643521e-05, + "loss": 0.6816, + "step": 3379 + }, + { + "epoch": 1.8634024203644457, + "grad_norm": 0.345703125, + "learning_rate": 6.136113596908392e-05, + "loss": 0.7894, + "step": 3380 + }, + { + "epoch": 1.8639588259841424, + "grad_norm": 0.326171875, + "learning_rate": 6.130729375210638e-05, + "loss": 0.71, + "step": 3381 + }, + { + "epoch": 1.8645152316038391, + "grad_norm": 0.33203125, + "learning_rate": 6.125346472385512e-05, + "loss": 0.64, + "step": 3382 + }, + { + "epoch": 1.865071637223536, + "grad_norm": 0.3828125, + "learning_rate": 6.11996489026782e-05, + "loss": 0.7395, + "step": 3383 + }, + { + "epoch": 1.8656280428432326, + "grad_norm": 0.32421875, + "learning_rate": 6.114584630691922e-05, + "loss": 0.6855, + "step": 3384 + }, + { + "epoch": 1.8661844484629295, + "grad_norm": 0.337890625, + "learning_rate": 6.10920569549172e-05, + "loss": 0.8022, + "step": 3385 + }, + { + "epoch": 1.8667408540826262, + "grad_norm": 0.30859375, + "learning_rate": 6.103828086500669e-05, + "loss": 0.6426, + "step": 3386 + }, + { + "epoch": 1.867297259702323, + "grad_norm": 0.5703125, + "learning_rate": 6.0984518055517725e-05, + "loss": 1.639, + "step": 3387 + }, + { + "epoch": 1.86785366532202, + "grad_norm": 0.3515625, + "learning_rate": 6.0930768544775754e-05, + "loss": 0.7888, + "step": 3388 + }, + { + "epoch": 1.8684100709417164, + "grad_norm": 0.34375, + "learning_rate": 6.0877032351101805e-05, + "loss": 0.7912, + "step": 3389 + }, + { + "epoch": 1.8689664765614133, + "grad_norm": 0.3203125, + "learning_rate": 6.0823309492812275e-05, + "loss": 0.6784, + "step": 3390 + }, + { + "epoch": 1.86952288218111, + "grad_norm": 0.326171875, + "learning_rate": 6.076959998821903e-05, + "loss": 0.6824, + "step": 3391 + }, + { + "epoch": 1.8700792878008068, + "grad_norm": 0.31640625, + "learning_rate": 6.071590385562942e-05, + "loss": 0.6199, + "step": 3392 + }, + { + "epoch": 1.8706356934205035, + "grad_norm": 0.330078125, + "learning_rate": 6.066222111334618e-05, + "loss": 0.6577, + "step": 3393 + }, + { + "epoch": 1.8711920990402002, + "grad_norm": 0.330078125, + "learning_rate": 6.060855177966753e-05, + "loss": 0.7511, + "step": 3394 + }, + { + "epoch": 1.8717485046598972, + "grad_norm": 0.333984375, + "learning_rate": 6.0554895872887095e-05, + "loss": 0.7647, + "step": 3395 + }, + { + "epoch": 1.8723049102795937, + "grad_norm": 0.31640625, + "learning_rate": 6.050125341129391e-05, + "loss": 0.6199, + "step": 3396 + }, + { + "epoch": 1.8728613158992906, + "grad_norm": 0.294921875, + "learning_rate": 6.044762441317251e-05, + "loss": 0.5352, + "step": 3397 + }, + { + "epoch": 1.8734177215189873, + "grad_norm": 0.31640625, + "learning_rate": 6.039400889680273e-05, + "loss": 0.6586, + "step": 3398 + }, + { + "epoch": 1.873974127138684, + "grad_norm": 0.32421875, + "learning_rate": 6.034040688045989e-05, + "loss": 0.7554, + "step": 3399 + }, + { + "epoch": 1.874530532758381, + "grad_norm": 0.328125, + "learning_rate": 6.0286818382414636e-05, + "loss": 0.7082, + "step": 3400 + }, + { + "epoch": 1.8750869383780775, + "grad_norm": 0.349609375, + "learning_rate": 6.0233243420933086e-05, + "loss": 0.7585, + "step": 3401 + }, + { + "epoch": 1.8756433439977744, + "grad_norm": 0.400390625, + "learning_rate": 6.017968201427667e-05, + "loss": 0.7286, + "step": 3402 + }, + { + "epoch": 1.8761997496174712, + "grad_norm": 0.318359375, + "learning_rate": 6.012613418070226e-05, + "loss": 0.6677, + "step": 3403 + }, + { + "epoch": 1.8767561552371679, + "grad_norm": 0.30859375, + "learning_rate": 6.0072599938462024e-05, + "loss": 0.6098, + "step": 3404 + }, + { + "epoch": 1.8773125608568648, + "grad_norm": 0.326171875, + "learning_rate": 6.001907930580364e-05, + "loss": 0.7346, + "step": 3405 + }, + { + "epoch": 1.8778689664765613, + "grad_norm": 0.322265625, + "learning_rate": 5.996557230096999e-05, + "loss": 0.7008, + "step": 3406 + }, + { + "epoch": 1.8784253720962583, + "grad_norm": 0.34375, + "learning_rate": 5.991207894219939e-05, + "loss": 0.763, + "step": 3407 + }, + { + "epoch": 1.878981777715955, + "grad_norm": 0.333984375, + "learning_rate": 5.9858599247725486e-05, + "loss": 0.6815, + "step": 3408 + }, + { + "epoch": 1.8795381833356517, + "grad_norm": 0.33984375, + "learning_rate": 5.980513323577728e-05, + "loss": 0.7103, + "step": 3409 + }, + { + "epoch": 1.8800945889553484, + "grad_norm": 0.328125, + "learning_rate": 5.975168092457909e-05, + "loss": 0.6802, + "step": 3410 + }, + { + "epoch": 1.8806509945750451, + "grad_norm": 0.330078125, + "learning_rate": 5.9698242332350574e-05, + "loss": 0.6619, + "step": 3411 + }, + { + "epoch": 1.881207400194742, + "grad_norm": 0.322265625, + "learning_rate": 5.964481747730673e-05, + "loss": 0.6355, + "step": 3412 + }, + { + "epoch": 1.8817638058144386, + "grad_norm": 0.326171875, + "learning_rate": 5.959140637765782e-05, + "loss": 0.6509, + "step": 3413 + }, + { + "epoch": 1.8823202114341355, + "grad_norm": 0.337890625, + "learning_rate": 5.953800905160951e-05, + "loss": 0.7293, + "step": 3414 + }, + { + "epoch": 1.8828766170538322, + "grad_norm": 0.318359375, + "learning_rate": 5.948462551736268e-05, + "loss": 0.6123, + "step": 3415 + }, + { + "epoch": 1.883433022673529, + "grad_norm": 0.33203125, + "learning_rate": 5.943125579311356e-05, + "loss": 0.7313, + "step": 3416 + }, + { + "epoch": 1.883989428293226, + "grad_norm": 0.3359375, + "learning_rate": 5.937789989705365e-05, + "loss": 0.7579, + "step": 3417 + }, + { + "epoch": 1.8845458339129224, + "grad_norm": 0.333984375, + "learning_rate": 5.932455784736974e-05, + "loss": 0.7487, + "step": 3418 + }, + { + "epoch": 1.8851022395326194, + "grad_norm": 0.31640625, + "learning_rate": 5.927122966224393e-05, + "loss": 0.6194, + "step": 3419 + }, + { + "epoch": 1.885658645152316, + "grad_norm": 0.33203125, + "learning_rate": 5.921791535985355e-05, + "loss": 0.7643, + "step": 3420 + }, + { + "epoch": 1.8862150507720128, + "grad_norm": 0.328125, + "learning_rate": 5.916461495837117e-05, + "loss": 0.7069, + "step": 3421 + }, + { + "epoch": 1.8867714563917095, + "grad_norm": 0.333984375, + "learning_rate": 5.911132847596477e-05, + "loss": 0.6694, + "step": 3422 + }, + { + "epoch": 1.8873278620114062, + "grad_norm": 0.330078125, + "learning_rate": 5.905805593079742e-05, + "loss": 0.6906, + "step": 3423 + }, + { + "epoch": 1.8878842676311032, + "grad_norm": 0.318359375, + "learning_rate": 5.900479734102752e-05, + "loss": 0.6751, + "step": 3424 + }, + { + "epoch": 1.8884406732507997, + "grad_norm": 0.345703125, + "learning_rate": 5.8951552724808704e-05, + "loss": 0.7788, + "step": 3425 + }, + { + "epoch": 1.8889970788704966, + "grad_norm": 0.34765625, + "learning_rate": 5.8898322100289804e-05, + "loss": 0.7784, + "step": 3426 + }, + { + "epoch": 1.8895534844901933, + "grad_norm": 0.328125, + "learning_rate": 5.8845105485614956e-05, + "loss": 0.6916, + "step": 3427 + }, + { + "epoch": 1.89010989010989, + "grad_norm": 0.34765625, + "learning_rate": 5.8791902898923454e-05, + "loss": 0.786, + "step": 3428 + }, + { + "epoch": 1.890666295729587, + "grad_norm": 0.33203125, + "learning_rate": 5.873871435834983e-05, + "loss": 0.7214, + "step": 3429 + }, + { + "epoch": 1.8912227013492835, + "grad_norm": 0.33203125, + "learning_rate": 5.868553988202386e-05, + "loss": 0.7162, + "step": 3430 + }, + { + "epoch": 1.8917791069689804, + "grad_norm": 0.318359375, + "learning_rate": 5.86323794880705e-05, + "loss": 0.6719, + "step": 3431 + }, + { + "epoch": 1.8923355125886772, + "grad_norm": 0.32421875, + "learning_rate": 5.857923319460987e-05, + "loss": 0.6797, + "step": 3432 + }, + { + "epoch": 1.8928919182083739, + "grad_norm": 0.337890625, + "learning_rate": 5.852610101975736e-05, + "loss": 0.7795, + "step": 3433 + }, + { + "epoch": 1.8934483238280708, + "grad_norm": 0.318359375, + "learning_rate": 5.847298298162349e-05, + "loss": 0.6841, + "step": 3434 + }, + { + "epoch": 1.8940047294477673, + "grad_norm": 0.328125, + "learning_rate": 5.841987909831399e-05, + "loss": 0.6571, + "step": 3435 + }, + { + "epoch": 1.8945611350674643, + "grad_norm": 0.33203125, + "learning_rate": 5.8366789387929734e-05, + "loss": 0.7283, + "step": 3436 + }, + { + "epoch": 1.895117540687161, + "grad_norm": 0.330078125, + "learning_rate": 5.831371386856677e-05, + "loss": 0.6967, + "step": 3437 + }, + { + "epoch": 1.8956739463068577, + "grad_norm": 0.328125, + "learning_rate": 5.8260652558316365e-05, + "loss": 0.7038, + "step": 3438 + }, + { + "epoch": 1.8962303519265544, + "grad_norm": 0.359375, + "learning_rate": 5.820760547526492e-05, + "loss": 0.7008, + "step": 3439 + }, + { + "epoch": 1.8967867575462511, + "grad_norm": 0.32421875, + "learning_rate": 5.8154572637493887e-05, + "loss": 0.6641, + "step": 3440 + }, + { + "epoch": 1.897343163165948, + "grad_norm": 0.294921875, + "learning_rate": 5.810155406308004e-05, + "loss": 0.5793, + "step": 3441 + }, + { + "epoch": 1.8978995687856446, + "grad_norm": 0.318359375, + "learning_rate": 5.804854977009511e-05, + "loss": 0.6426, + "step": 3442 + }, + { + "epoch": 1.8984559744053415, + "grad_norm": 0.32421875, + "learning_rate": 5.7995559776606114e-05, + "loss": 0.7162, + "step": 3443 + }, + { + "epoch": 1.8990123800250382, + "grad_norm": 0.3359375, + "learning_rate": 5.7942584100675056e-05, + "loss": 0.73, + "step": 3444 + }, + { + "epoch": 1.899568785644735, + "grad_norm": 0.32421875, + "learning_rate": 5.7889622760359206e-05, + "loss": 0.6438, + "step": 3445 + }, + { + "epoch": 1.900125191264432, + "grad_norm": 0.337890625, + "learning_rate": 5.7836675773710814e-05, + "loss": 0.7466, + "step": 3446 + }, + { + "epoch": 1.9006815968841284, + "grad_norm": 0.345703125, + "learning_rate": 5.778374315877728e-05, + "loss": 0.7691, + "step": 3447 + }, + { + "epoch": 1.9012380025038254, + "grad_norm": 0.318359375, + "learning_rate": 5.7730824933601224e-05, + "loss": 0.6923, + "step": 3448 + }, + { + "epoch": 1.901794408123522, + "grad_norm": 0.322265625, + "learning_rate": 5.7677921116220127e-05, + "loss": 0.7179, + "step": 3449 + }, + { + "epoch": 1.9023508137432188, + "grad_norm": 0.333984375, + "learning_rate": 5.762503172466679e-05, + "loss": 0.7088, + "step": 3450 + }, + { + "epoch": 1.9029072193629155, + "grad_norm": 0.32421875, + "learning_rate": 5.757215677696892e-05, + "loss": 0.6678, + "step": 3451 + }, + { + "epoch": 1.9034636249826122, + "grad_norm": 0.349609375, + "learning_rate": 5.751929629114945e-05, + "loss": 0.7415, + "step": 3452 + }, + { + "epoch": 1.9040200306023092, + "grad_norm": 0.337890625, + "learning_rate": 5.7466450285226235e-05, + "loss": 0.74, + "step": 3453 + }, + { + "epoch": 1.904576436222006, + "grad_norm": 0.333984375, + "learning_rate": 5.74136187772123e-05, + "loss": 0.7213, + "step": 3454 + }, + { + "epoch": 1.9051328418417026, + "grad_norm": 0.322265625, + "learning_rate": 5.736080178511569e-05, + "loss": 0.7096, + "step": 3455 + }, + { + "epoch": 1.9056892474613993, + "grad_norm": 0.333984375, + "learning_rate": 5.7307999326939584e-05, + "loss": 0.6926, + "step": 3456 + }, + { + "epoch": 1.906245653081096, + "grad_norm": 0.365234375, + "learning_rate": 5.7255211420682034e-05, + "loss": 0.7527, + "step": 3457 + }, + { + "epoch": 1.906802058700793, + "grad_norm": 0.341796875, + "learning_rate": 5.7202438084336296e-05, + "loss": 0.724, + "step": 3458 + }, + { + "epoch": 1.9073584643204895, + "grad_norm": 0.33203125, + "learning_rate": 5.714967933589054e-05, + "loss": 0.7235, + "step": 3459 + }, + { + "epoch": 1.9079148699401864, + "grad_norm": 0.306640625, + "learning_rate": 5.709693519332808e-05, + "loss": 0.6135, + "step": 3460 + }, + { + "epoch": 1.9084712755598832, + "grad_norm": 0.33203125, + "learning_rate": 5.7044205674627115e-05, + "loss": 0.7181, + "step": 3461 + }, + { + "epoch": 1.9090276811795799, + "grad_norm": 0.328125, + "learning_rate": 5.6991490797760984e-05, + "loss": 0.6753, + "step": 3462 + }, + { + "epoch": 1.9095840867992768, + "grad_norm": 0.306640625, + "learning_rate": 5.6938790580697996e-05, + "loss": 0.6165, + "step": 3463 + }, + { + "epoch": 1.9101404924189733, + "grad_norm": 0.328125, + "learning_rate": 5.6886105041401416e-05, + "loss": 0.6911, + "step": 3464 + }, + { + "epoch": 1.9106968980386703, + "grad_norm": 0.330078125, + "learning_rate": 5.683343419782958e-05, + "loss": 0.7027, + "step": 3465 + }, + { + "epoch": 1.911253303658367, + "grad_norm": 0.318359375, + "learning_rate": 5.678077806793571e-05, + "loss": 0.6894, + "step": 3466 + }, + { + "epoch": 1.9118097092780637, + "grad_norm": 0.33984375, + "learning_rate": 5.672813666966819e-05, + "loss": 0.7583, + "step": 3467 + }, + { + "epoch": 1.9123661148977604, + "grad_norm": 0.326171875, + "learning_rate": 5.6675510020970154e-05, + "loss": 0.6529, + "step": 3468 + }, + { + "epoch": 1.9129225205174571, + "grad_norm": 0.353515625, + "learning_rate": 5.662289813977992e-05, + "loss": 0.8184, + "step": 3469 + }, + { + "epoch": 1.913478926137154, + "grad_norm": 0.34375, + "learning_rate": 5.657030104403056e-05, + "loss": 0.7597, + "step": 3470 + }, + { + "epoch": 1.9140353317568506, + "grad_norm": 0.34375, + "learning_rate": 5.651771875165038e-05, + "loss": 0.7448, + "step": 3471 + }, + { + "epoch": 1.9145917373765475, + "grad_norm": 0.3359375, + "learning_rate": 5.646515128056238e-05, + "loss": 0.7357, + "step": 3472 + }, + { + "epoch": 1.9151481429962443, + "grad_norm": 0.341796875, + "learning_rate": 5.6412598648684676e-05, + "loss": 0.7623, + "step": 3473 + }, + { + "epoch": 1.915704548615941, + "grad_norm": 0.337890625, + "learning_rate": 5.63600608739302e-05, + "loss": 0.7732, + "step": 3474 + }, + { + "epoch": 1.916260954235638, + "grad_norm": 0.337890625, + "learning_rate": 5.6307537974206945e-05, + "loss": 0.7417, + "step": 3475 + }, + { + "epoch": 1.9168173598553344, + "grad_norm": 0.33203125, + "learning_rate": 5.625502996741772e-05, + "loss": 0.6915, + "step": 3476 + }, + { + "epoch": 1.9173737654750314, + "grad_norm": 0.349609375, + "learning_rate": 5.6202536871460355e-05, + "loss": 0.7983, + "step": 3477 + }, + { + "epoch": 1.917930171094728, + "grad_norm": 0.32421875, + "learning_rate": 5.615005870422749e-05, + "loss": 0.6911, + "step": 3478 + }, + { + "epoch": 1.9184865767144248, + "grad_norm": 0.326171875, + "learning_rate": 5.609759548360679e-05, + "loss": 0.6519, + "step": 3479 + }, + { + "epoch": 1.9190429823341217, + "grad_norm": 0.333984375, + "learning_rate": 5.60451472274808e-05, + "loss": 0.6688, + "step": 3480 + }, + { + "epoch": 1.9195993879538182, + "grad_norm": 0.34375, + "learning_rate": 5.599271395372685e-05, + "loss": 0.7917, + "step": 3481 + }, + { + "epoch": 1.9201557935735152, + "grad_norm": 0.31640625, + "learning_rate": 5.594029568021735e-05, + "loss": 0.6868, + "step": 3482 + }, + { + "epoch": 1.920712199193212, + "grad_norm": 0.361328125, + "learning_rate": 5.588789242481942e-05, + "loss": 0.7817, + "step": 3483 + }, + { + "epoch": 1.9212686048129086, + "grad_norm": 0.31640625, + "learning_rate": 5.583550420539519e-05, + "loss": 0.6174, + "step": 3484 + }, + { + "epoch": 1.9218250104326053, + "grad_norm": 0.314453125, + "learning_rate": 5.578313103980157e-05, + "loss": 0.6561, + "step": 3485 + }, + { + "epoch": 1.922381416052302, + "grad_norm": 0.3203125, + "learning_rate": 5.573077294589044e-05, + "loss": 0.6711, + "step": 3486 + }, + { + "epoch": 1.922937821671999, + "grad_norm": 0.33203125, + "learning_rate": 5.567842994150838e-05, + "loss": 0.7462, + "step": 3487 + }, + { + "epoch": 1.9234942272916955, + "grad_norm": 0.326171875, + "learning_rate": 5.562610204449708e-05, + "loss": 0.6818, + "step": 3488 + }, + { + "epoch": 1.9240506329113924, + "grad_norm": 0.31640625, + "learning_rate": 5.557378927269282e-05, + "loss": 0.6861, + "step": 3489 + }, + { + "epoch": 1.9246070385310892, + "grad_norm": 0.314453125, + "learning_rate": 5.552149164392692e-05, + "loss": 0.6297, + "step": 3490 + }, + { + "epoch": 1.9251634441507859, + "grad_norm": 0.330078125, + "learning_rate": 5.546920917602536e-05, + "loss": 0.7244, + "step": 3491 + }, + { + "epoch": 1.9257198497704828, + "grad_norm": 0.333984375, + "learning_rate": 5.541694188680915e-05, + "loss": 0.7663, + "step": 3492 + }, + { + "epoch": 1.9262762553901793, + "grad_norm": 0.333984375, + "learning_rate": 5.536468979409397e-05, + "loss": 0.6797, + "step": 3493 + }, + { + "epoch": 1.9268326610098763, + "grad_norm": 0.31640625, + "learning_rate": 5.531245291569034e-05, + "loss": 0.6774, + "step": 3494 + }, + { + "epoch": 1.927389066629573, + "grad_norm": 0.349609375, + "learning_rate": 5.526023126940366e-05, + "loss": 0.7999, + "step": 3495 + }, + { + "epoch": 1.9279454722492697, + "grad_norm": 0.3359375, + "learning_rate": 5.520802487303412e-05, + "loss": 0.7356, + "step": 3496 + }, + { + "epoch": 1.9285018778689664, + "grad_norm": 0.326171875, + "learning_rate": 5.515583374437673e-05, + "loss": 0.6731, + "step": 3497 + }, + { + "epoch": 1.9290582834886632, + "grad_norm": 0.330078125, + "learning_rate": 5.5103657901221183e-05, + "loss": 0.6694, + "step": 3498 + }, + { + "epoch": 1.92961468910836, + "grad_norm": 0.322265625, + "learning_rate": 5.5051497361352134e-05, + "loss": 0.6748, + "step": 3499 + }, + { + "epoch": 1.9301710947280566, + "grad_norm": 0.31640625, + "learning_rate": 5.499935214254883e-05, + "loss": 0.6626, + "step": 3500 + }, + { + "epoch": 1.9307275003477535, + "grad_norm": 0.33203125, + "learning_rate": 5.494722226258551e-05, + "loss": 0.6737, + "step": 3501 + }, + { + "epoch": 1.9312839059674503, + "grad_norm": 0.34375, + "learning_rate": 5.4895107739230944e-05, + "loss": 0.7086, + "step": 3502 + }, + { + "epoch": 1.931840311587147, + "grad_norm": 0.345703125, + "learning_rate": 5.484300859024888e-05, + "loss": 0.7379, + "step": 3503 + }, + { + "epoch": 1.932396717206844, + "grad_norm": 0.337890625, + "learning_rate": 5.479092483339775e-05, + "loss": 0.7199, + "step": 3504 + }, + { + "epoch": 1.9329531228265404, + "grad_norm": 0.3125, + "learning_rate": 5.473885648643067e-05, + "loss": 0.6758, + "step": 3505 + }, + { + "epoch": 1.9335095284462374, + "grad_norm": 0.345703125, + "learning_rate": 5.4686803567095634e-05, + "loss": 0.6889, + "step": 3506 + }, + { + "epoch": 1.934065934065934, + "grad_norm": 0.330078125, + "learning_rate": 5.463476609313522e-05, + "loss": 0.7026, + "step": 3507 + }, + { + "epoch": 1.9346223396856308, + "grad_norm": 0.330078125, + "learning_rate": 5.4582744082286915e-05, + "loss": 0.6789, + "step": 3508 + }, + { + "epoch": 1.9351787453053277, + "grad_norm": 0.322265625, + "learning_rate": 5.453073755228277e-05, + "loss": 0.6913, + "step": 3509 + }, + { + "epoch": 1.9357351509250242, + "grad_norm": 0.33984375, + "learning_rate": 5.4478746520849724e-05, + "loss": 0.7839, + "step": 3510 + }, + { + "epoch": 1.9362915565447212, + "grad_norm": 0.330078125, + "learning_rate": 5.44267710057092e-05, + "loss": 0.6969, + "step": 3511 + }, + { + "epoch": 1.936847962164418, + "grad_norm": 0.32421875, + "learning_rate": 5.437481102457767e-05, + "loss": 0.7012, + "step": 3512 + }, + { + "epoch": 1.9374043677841146, + "grad_norm": 0.337890625, + "learning_rate": 5.432286659516597e-05, + "loss": 0.6831, + "step": 3513 + }, + { + "epoch": 1.9379607734038113, + "grad_norm": 0.361328125, + "learning_rate": 5.427093773517989e-05, + "loss": 0.8023, + "step": 3514 + }, + { + "epoch": 1.938517179023508, + "grad_norm": 0.328125, + "learning_rate": 5.421902446231972e-05, + "loss": 0.6966, + "step": 3515 + }, + { + "epoch": 1.939073584643205, + "grad_norm": 0.33984375, + "learning_rate": 5.416712679428061e-05, + "loss": 0.7335, + "step": 3516 + }, + { + "epoch": 1.9396299902629015, + "grad_norm": 0.337890625, + "learning_rate": 5.41152447487522e-05, + "loss": 0.7449, + "step": 3517 + }, + { + "epoch": 1.9401863958825984, + "grad_norm": 0.32421875, + "learning_rate": 5.4063378343419034e-05, + "loss": 0.7172, + "step": 3518 + }, + { + "epoch": 1.9407428015022952, + "grad_norm": 0.322265625, + "learning_rate": 5.4011527595960086e-05, + "loss": 0.6315, + "step": 3519 + }, + { + "epoch": 1.9412992071219919, + "grad_norm": 0.3359375, + "learning_rate": 5.3959692524049157e-05, + "loss": 0.6937, + "step": 3520 + }, + { + "epoch": 1.9418556127416888, + "grad_norm": 0.310546875, + "learning_rate": 5.39078731453547e-05, + "loss": 0.6152, + "step": 3521 + }, + { + "epoch": 1.9424120183613853, + "grad_norm": 0.328125, + "learning_rate": 5.3856069477539696e-05, + "loss": 0.6915, + "step": 3522 + }, + { + "epoch": 1.9429684239810823, + "grad_norm": 0.322265625, + "learning_rate": 5.380428153826194e-05, + "loss": 0.6513, + "step": 3523 + }, + { + "epoch": 1.943524829600779, + "grad_norm": 0.341796875, + "learning_rate": 5.375250934517367e-05, + "loss": 0.76, + "step": 3524 + }, + { + "epoch": 1.9440812352204757, + "grad_norm": 0.328125, + "learning_rate": 5.370075291592196e-05, + "loss": 0.7397, + "step": 3525 + }, + { + "epoch": 1.9446376408401724, + "grad_norm": 0.322265625, + "learning_rate": 5.364901226814834e-05, + "loss": 0.6783, + "step": 3526 + }, + { + "epoch": 1.9451940464598692, + "grad_norm": 0.328125, + "learning_rate": 5.35972874194891e-05, + "loss": 0.6934, + "step": 3527 + }, + { + "epoch": 1.945750452079566, + "grad_norm": 0.3359375, + "learning_rate": 5.3545578387574966e-05, + "loss": 0.7227, + "step": 3528 + }, + { + "epoch": 1.9463068576992628, + "grad_norm": 0.328125, + "learning_rate": 5.3493885190031535e-05, + "loss": 0.6903, + "step": 3529 + }, + { + "epoch": 1.9468632633189595, + "grad_norm": 0.328125, + "learning_rate": 5.344220784447875e-05, + "loss": 0.6746, + "step": 3530 + }, + { + "epoch": 1.9474196689386563, + "grad_norm": 0.314453125, + "learning_rate": 5.339054636853134e-05, + "loss": 0.6808, + "step": 3531 + }, + { + "epoch": 1.947976074558353, + "grad_norm": 0.328125, + "learning_rate": 5.333890077979845e-05, + "loss": 0.7267, + "step": 3532 + }, + { + "epoch": 1.94853248017805, + "grad_norm": 0.33203125, + "learning_rate": 5.328727109588398e-05, + "loss": 0.7169, + "step": 3533 + }, + { + "epoch": 1.9490888857977464, + "grad_norm": 0.330078125, + "learning_rate": 5.323565733438629e-05, + "loss": 0.7544, + "step": 3534 + }, + { + "epoch": 1.9496452914174434, + "grad_norm": 0.306640625, + "learning_rate": 5.318405951289838e-05, + "loss": 0.5667, + "step": 3535 + }, + { + "epoch": 1.95020169703714, + "grad_norm": 0.326171875, + "learning_rate": 5.313247764900775e-05, + "loss": 0.6999, + "step": 3536 + }, + { + "epoch": 1.9507581026568368, + "grad_norm": 0.318359375, + "learning_rate": 5.3080911760296515e-05, + "loss": 0.6345, + "step": 3537 + }, + { + "epoch": 1.9513145082765337, + "grad_norm": 0.353515625, + "learning_rate": 5.302936186434138e-05, + "loss": 0.7883, + "step": 3538 + }, + { + "epoch": 1.9518709138962302, + "grad_norm": 0.326171875, + "learning_rate": 5.297782797871349e-05, + "loss": 0.7205, + "step": 3539 + }, + { + "epoch": 1.9524273195159272, + "grad_norm": 0.30859375, + "learning_rate": 5.292631012097864e-05, + "loss": 0.6068, + "step": 3540 + }, + { + "epoch": 1.952983725135624, + "grad_norm": 0.322265625, + "learning_rate": 5.287480830869706e-05, + "loss": 0.7076, + "step": 3541 + }, + { + "epoch": 1.9535401307553206, + "grad_norm": 0.322265625, + "learning_rate": 5.2823322559423615e-05, + "loss": 0.668, + "step": 3542 + }, + { + "epoch": 1.9540965363750173, + "grad_norm": 0.306640625, + "learning_rate": 5.277185289070761e-05, + "loss": 0.6219, + "step": 3543 + }, + { + "epoch": 1.954652941994714, + "grad_norm": 0.306640625, + "learning_rate": 5.2720399320092905e-05, + "loss": 0.5702, + "step": 3544 + }, + { + "epoch": 1.955209347614411, + "grad_norm": 0.349609375, + "learning_rate": 5.2668961865117894e-05, + "loss": 0.7888, + "step": 3545 + }, + { + "epoch": 1.9557657532341075, + "grad_norm": 0.328125, + "learning_rate": 5.2617540543315466e-05, + "loss": 0.6866, + "step": 3546 + }, + { + "epoch": 1.9563221588538044, + "grad_norm": 0.333984375, + "learning_rate": 5.2566135372212956e-05, + "loss": 0.7442, + "step": 3547 + }, + { + "epoch": 1.9568785644735012, + "grad_norm": 0.330078125, + "learning_rate": 5.251474636933231e-05, + "loss": 0.7, + "step": 3548 + }, + { + "epoch": 1.957434970093198, + "grad_norm": 0.349609375, + "learning_rate": 5.24633735521898e-05, + "loss": 0.7032, + "step": 3549 + }, + { + "epoch": 1.9579913757128948, + "grad_norm": 0.3359375, + "learning_rate": 5.241201693829634e-05, + "loss": 0.7412, + "step": 3550 + }, + { + "epoch": 1.9585477813325913, + "grad_norm": 0.328125, + "learning_rate": 5.2360676545157204e-05, + "loss": 0.6876, + "step": 3551 + }, + { + "epoch": 1.9591041869522883, + "grad_norm": 0.30859375, + "learning_rate": 5.2309352390272246e-05, + "loss": 0.6344, + "step": 3552 + }, + { + "epoch": 1.959660592571985, + "grad_norm": 0.33203125, + "learning_rate": 5.225804449113565e-05, + "loss": 0.68, + "step": 3553 + }, + { + "epoch": 1.9602169981916817, + "grad_norm": 0.32421875, + "learning_rate": 5.2206752865236185e-05, + "loss": 0.6309, + "step": 3554 + }, + { + "epoch": 1.9607734038113787, + "grad_norm": 0.3203125, + "learning_rate": 5.215547753005705e-05, + "loss": 0.6127, + "step": 3555 + }, + { + "epoch": 1.9613298094310752, + "grad_norm": 0.337890625, + "learning_rate": 5.2104218503075775e-05, + "loss": 0.7644, + "step": 3556 + }, + { + "epoch": 1.961886215050772, + "grad_norm": 0.337890625, + "learning_rate": 5.2052975801764514e-05, + "loss": 0.7308, + "step": 3557 + }, + { + "epoch": 1.9624426206704688, + "grad_norm": 0.318359375, + "learning_rate": 5.2001749443589685e-05, + "loss": 0.6296, + "step": 3558 + }, + { + "epoch": 1.9629990262901655, + "grad_norm": 0.314453125, + "learning_rate": 5.195053944601227e-05, + "loss": 0.6291, + "step": 3559 + }, + { + "epoch": 1.9635554319098623, + "grad_norm": 0.33203125, + "learning_rate": 5.189934582648757e-05, + "loss": 0.6677, + "step": 3560 + }, + { + "epoch": 1.964111837529559, + "grad_norm": 0.322265625, + "learning_rate": 5.184816860246537e-05, + "loss": 0.7093, + "step": 3561 + }, + { + "epoch": 1.964668243149256, + "grad_norm": 0.3359375, + "learning_rate": 5.179700779138989e-05, + "loss": 0.7041, + "step": 3562 + }, + { + "epoch": 1.9652246487689524, + "grad_norm": 0.34765625, + "learning_rate": 5.1745863410699626e-05, + "loss": 0.7722, + "step": 3563 + }, + { + "epoch": 1.9657810543886494, + "grad_norm": 0.369140625, + "learning_rate": 5.169473547782764e-05, + "loss": 0.7212, + "step": 3564 + }, + { + "epoch": 1.966337460008346, + "grad_norm": 0.345703125, + "learning_rate": 5.164362401020124e-05, + "loss": 0.7581, + "step": 3565 + }, + { + "epoch": 1.9668938656280428, + "grad_norm": 0.302734375, + "learning_rate": 5.1592529025242257e-05, + "loss": 0.5711, + "step": 3566 + }, + { + "epoch": 1.9674502712477397, + "grad_norm": 0.341796875, + "learning_rate": 5.154145054036677e-05, + "loss": 0.742, + "step": 3567 + }, + { + "epoch": 1.9680066768674362, + "grad_norm": 0.337890625, + "learning_rate": 5.149038857298536e-05, + "loss": 0.7484, + "step": 3568 + }, + { + "epoch": 1.9685630824871332, + "grad_norm": 0.326171875, + "learning_rate": 5.143934314050281e-05, + "loss": 0.717, + "step": 3569 + }, + { + "epoch": 1.96911948810683, + "grad_norm": 0.32421875, + "learning_rate": 5.138831426031853e-05, + "loss": 0.6858, + "step": 3570 + }, + { + "epoch": 1.9696758937265266, + "grad_norm": 0.29296875, + "learning_rate": 5.133730194982601e-05, + "loss": 0.4857, + "step": 3571 + }, + { + "epoch": 1.9702322993462233, + "grad_norm": 0.330078125, + "learning_rate": 5.12863062264133e-05, + "loss": 0.6772, + "step": 3572 + }, + { + "epoch": 1.97078870496592, + "grad_norm": 0.333984375, + "learning_rate": 5.123532710746264e-05, + "loss": 0.6981, + "step": 3573 + }, + { + "epoch": 1.971345110585617, + "grad_norm": 0.337890625, + "learning_rate": 5.118436461035073e-05, + "loss": 0.6969, + "step": 3574 + }, + { + "epoch": 1.9719015162053135, + "grad_norm": 0.341796875, + "learning_rate": 5.11334187524485e-05, + "loss": 0.774, + "step": 3575 + }, + { + "epoch": 1.9724579218250105, + "grad_norm": 0.341796875, + "learning_rate": 5.108248955112135e-05, + "loss": 0.7688, + "step": 3576 + }, + { + "epoch": 1.9730143274447072, + "grad_norm": 0.333984375, + "learning_rate": 5.103157702372881e-05, + "loss": 0.7283, + "step": 3577 + }, + { + "epoch": 1.973570733064404, + "grad_norm": 0.328125, + "learning_rate": 5.09806811876249e-05, + "loss": 0.6848, + "step": 3578 + }, + { + "epoch": 1.9741271386841008, + "grad_norm": 0.32421875, + "learning_rate": 5.0929802060157907e-05, + "loss": 0.7131, + "step": 3579 + }, + { + "epoch": 1.9746835443037973, + "grad_norm": 0.326171875, + "learning_rate": 5.087893965867032e-05, + "loss": 0.6419, + "step": 3580 + }, + { + "epoch": 1.9752399499234943, + "grad_norm": 0.3046875, + "learning_rate": 5.0828094000499084e-05, + "loss": 0.5633, + "step": 3581 + }, + { + "epoch": 1.975796355543191, + "grad_norm": 0.314453125, + "learning_rate": 5.077726510297528e-05, + "loss": 0.667, + "step": 3582 + }, + { + "epoch": 1.9763527611628877, + "grad_norm": 0.326171875, + "learning_rate": 5.072645298342445e-05, + "loss": 0.6718, + "step": 3583 + }, + { + "epoch": 1.9769091667825847, + "grad_norm": 0.3359375, + "learning_rate": 5.067565765916621e-05, + "loss": 0.7026, + "step": 3584 + }, + { + "epoch": 1.9774655724022812, + "grad_norm": 0.34375, + "learning_rate": 5.0624879147514636e-05, + "loss": 0.7568, + "step": 3585 + }, + { + "epoch": 1.978021978021978, + "grad_norm": 0.3203125, + "learning_rate": 5.0574117465777984e-05, + "loss": 0.62, + "step": 3586 + }, + { + "epoch": 1.9785783836416748, + "grad_norm": 0.341796875, + "learning_rate": 5.0523372631258814e-05, + "loss": 0.7078, + "step": 3587 + }, + { + "epoch": 1.9791347892613715, + "grad_norm": 0.330078125, + "learning_rate": 5.047264466125388e-05, + "loss": 0.6778, + "step": 3588 + }, + { + "epoch": 1.9796911948810683, + "grad_norm": 0.337890625, + "learning_rate": 5.042193357305427e-05, + "loss": 0.6849, + "step": 3589 + }, + { + "epoch": 1.980247600500765, + "grad_norm": 0.31640625, + "learning_rate": 5.03712393839452e-05, + "loss": 0.6501, + "step": 3590 + }, + { + "epoch": 1.980804006120462, + "grad_norm": 0.337890625, + "learning_rate": 5.03205621112063e-05, + "loss": 0.7426, + "step": 3591 + }, + { + "epoch": 1.9813604117401584, + "grad_norm": 0.353515625, + "learning_rate": 5.026990177211124e-05, + "loss": 0.7118, + "step": 3592 + }, + { + "epoch": 1.9819168173598554, + "grad_norm": 0.33203125, + "learning_rate": 5.021925838392809e-05, + "loss": 0.7003, + "step": 3593 + }, + { + "epoch": 1.982473222979552, + "grad_norm": 0.326171875, + "learning_rate": 5.0168631963918985e-05, + "loss": 0.6924, + "step": 3594 + }, + { + "epoch": 1.9830296285992488, + "grad_norm": 0.3125, + "learning_rate": 5.01180225293404e-05, + "loss": 0.6514, + "step": 3595 + }, + { + "epoch": 1.9835860342189457, + "grad_norm": 0.318359375, + "learning_rate": 5.006743009744301e-05, + "loss": 0.6604, + "step": 3596 + }, + { + "epoch": 1.9835860342189457, + "eval_loss": 0.7215954661369324, + "eval_runtime": 1750.3467, + "eval_samples_per_second": 2.68, + "eval_steps_per_second": 1.34, + "step": 3596 + } + ], + "logging_steps": 1, + "max_steps": 5391, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 899, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.498498600311718e+18, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}