diff --git "a/checkpoint-2398/trainer_state.json" "b/checkpoint-2398/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-2398/trainer_state.json" @@ -0,0 +1,16867 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.99812382739212, + "eval_steps": 400, + "global_step": 2398, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0008338544923910778, + "grad_norm": 0.35252463817596436, + "learning_rate": 2e-08, + "loss": 0.8273, + "step": 1 + }, + { + "epoch": 0.0008338544923910778, + "eval_loss": 0.861491858959198, + "eval_runtime": 319.044, + "eval_samples_per_second": 16.872, + "eval_steps_per_second": 2.815, + "step": 1 + }, + { + "epoch": 0.0016677089847821555, + "grad_norm": 0.3189675807952881, + "learning_rate": 4e-08, + "loss": 0.8572, + "step": 2 + }, + { + "epoch": 0.0025015634771732333, + "grad_norm": 0.3477325141429901, + "learning_rate": 6e-08, + "loss": 0.9059, + "step": 3 + }, + { + "epoch": 0.003335417969564311, + "grad_norm": 0.32423585653305054, + "learning_rate": 8e-08, + "loss": 0.8386, + "step": 4 + }, + { + "epoch": 0.004169272461955389, + "grad_norm": 0.32164493203163147, + "learning_rate": 1e-07, + "loss": 0.8753, + "step": 5 + }, + { + "epoch": 0.0050031269543464665, + "grad_norm": 0.33231157064437866, + "learning_rate": 1.2e-07, + "loss": 0.8665, + "step": 6 + }, + { + "epoch": 0.005836981446737544, + "grad_norm": 0.31154102087020874, + "learning_rate": 1.4e-07, + "loss": 0.8813, + "step": 7 + }, + { + "epoch": 0.006670835939128622, + "grad_norm": 0.3176712393760681, + "learning_rate": 1.6e-07, + "loss": 0.8683, + "step": 8 + }, + { + "epoch": 0.0075046904315197, + "grad_norm": 0.3419969975948334, + "learning_rate": 1.8e-07, + "loss": 0.849, + "step": 9 + }, + { + "epoch": 0.008338544923910778, + "grad_norm": 0.33408281207084656, + "learning_rate": 2e-07, + "loss": 0.8778, + "step": 10 + }, + { + "epoch": 0.009172399416301855, + "grad_norm": 0.3125899136066437, + "learning_rate": 2.1999999999999998e-07, + "loss": 0.8051, + "step": 11 + }, + { + "epoch": 0.010006253908692933, + "grad_norm": 0.32648083567619324, + "learning_rate": 2.4e-07, + "loss": 0.8297, + "step": 12 + }, + { + "epoch": 0.01084010840108401, + "grad_norm": 0.3143678605556488, + "learning_rate": 2.6e-07, + "loss": 0.8401, + "step": 13 + }, + { + "epoch": 0.011673962893475089, + "grad_norm": 0.3291407525539398, + "learning_rate": 2.8e-07, + "loss": 0.8668, + "step": 14 + }, + { + "epoch": 0.012507817385866166, + "grad_norm": 0.3211856484413147, + "learning_rate": 3e-07, + "loss": 0.84, + "step": 15 + }, + { + "epoch": 0.013341671878257244, + "grad_norm": 0.31169289350509644, + "learning_rate": 3.2e-07, + "loss": 0.8659, + "step": 16 + }, + { + "epoch": 0.014175526370648322, + "grad_norm": 0.3279392719268799, + "learning_rate": 3.4000000000000003e-07, + "loss": 0.8317, + "step": 17 + }, + { + "epoch": 0.0150093808630394, + "grad_norm": 0.3028814494609833, + "learning_rate": 3.6e-07, + "loss": 0.8246, + "step": 18 + }, + { + "epoch": 0.015843235355430477, + "grad_norm": 0.307066410779953, + "learning_rate": 3.7999999999999996e-07, + "loss": 0.8025, + "step": 19 + }, + { + "epoch": 0.016677089847821555, + "grad_norm": 0.3293515741825104, + "learning_rate": 4e-07, + "loss": 0.8674, + "step": 20 + }, + { + "epoch": 0.017510944340212633, + "grad_norm": 0.3215535581111908, + "learning_rate": 4.1999999999999995e-07, + "loss": 0.8533, + "step": 21 + }, + { + "epoch": 0.01834479883260371, + "grad_norm": 0.3236245810985565, + "learning_rate": 4.3999999999999997e-07, + "loss": 0.8957, + "step": 22 + }, + { + "epoch": 0.01917865332499479, + "grad_norm": 0.33956846594810486, + "learning_rate": 4.6e-07, + "loss": 0.8293, + "step": 23 + }, + { + "epoch": 0.020012507817385866, + "grad_norm": 0.33766457438468933, + "learning_rate": 4.8e-07, + "loss": 0.8381, + "step": 24 + }, + { + "epoch": 0.020846362309776944, + "grad_norm": 0.3336811065673828, + "learning_rate": 5e-07, + "loss": 0.8475, + "step": 25 + }, + { + "epoch": 0.02168021680216802, + "grad_norm": 0.307309627532959, + "learning_rate": 5.2e-07, + "loss": 0.852, + "step": 26 + }, + { + "epoch": 0.0225140712945591, + "grad_norm": 0.29773110151290894, + "learning_rate": 5.4e-07, + "loss": 0.8042, + "step": 27 + }, + { + "epoch": 0.023347925786950177, + "grad_norm": 0.3210948407649994, + "learning_rate": 5.6e-07, + "loss": 0.9099, + "step": 28 + }, + { + "epoch": 0.024181780279341255, + "grad_norm": 0.3149116337299347, + "learning_rate": 5.8e-07, + "loss": 0.8, + "step": 29 + }, + { + "epoch": 0.025015634771732333, + "grad_norm": 0.29307302832603455, + "learning_rate": 6e-07, + "loss": 0.8583, + "step": 30 + }, + { + "epoch": 0.02584948926412341, + "grad_norm": 0.3131772577762604, + "learning_rate": 6.2e-07, + "loss": 0.8432, + "step": 31 + }, + { + "epoch": 0.026683343756514488, + "grad_norm": 0.28058937191963196, + "learning_rate": 6.4e-07, + "loss": 0.8099, + "step": 32 + }, + { + "epoch": 0.027517198248905566, + "grad_norm": 0.2649301290512085, + "learning_rate": 6.6e-07, + "loss": 0.8715, + "step": 33 + }, + { + "epoch": 0.028351052741296644, + "grad_norm": 0.2659442126750946, + "learning_rate": 6.800000000000001e-07, + "loss": 0.7432, + "step": 34 + }, + { + "epoch": 0.02918490723368772, + "grad_norm": 0.24875527620315552, + "learning_rate": 7e-07, + "loss": 0.8502, + "step": 35 + }, + { + "epoch": 0.0300187617260788, + "grad_norm": 0.21730971336364746, + "learning_rate": 7.2e-07, + "loss": 0.813, + "step": 36 + }, + { + "epoch": 0.030852616218469877, + "grad_norm": 0.19625824689865112, + "learning_rate": 7.4e-07, + "loss": 0.7839, + "step": 37 + }, + { + "epoch": 0.031686470710860955, + "grad_norm": 0.19287347793579102, + "learning_rate": 7.599999999999999e-07, + "loss": 0.7949, + "step": 38 + }, + { + "epoch": 0.032520325203252036, + "grad_norm": 0.1588095873594284, + "learning_rate": 7.799999999999999e-07, + "loss": 0.7137, + "step": 39 + }, + { + "epoch": 0.03335417969564311, + "grad_norm": 0.17520244419574738, + "learning_rate": 8e-07, + "loss": 0.8112, + "step": 40 + }, + { + "epoch": 0.03418803418803419, + "grad_norm": 0.16394659876823425, + "learning_rate": 8.199999999999999e-07, + "loss": 0.8096, + "step": 41 + }, + { + "epoch": 0.035021888680425266, + "grad_norm": 0.1522587239742279, + "learning_rate": 8.399999999999999e-07, + "loss": 0.708, + "step": 42 + }, + { + "epoch": 0.03585574317281635, + "grad_norm": 0.16340652108192444, + "learning_rate": 8.599999999999999e-07, + "loss": 0.7816, + "step": 43 + }, + { + "epoch": 0.03668959766520742, + "grad_norm": 0.15452717244625092, + "learning_rate": 8.799999999999999e-07, + "loss": 0.7946, + "step": 44 + }, + { + "epoch": 0.0375234521575985, + "grad_norm": 0.1545872688293457, + "learning_rate": 9e-07, + "loss": 0.7908, + "step": 45 + }, + { + "epoch": 0.03835730664998958, + "grad_norm": 0.13799017667770386, + "learning_rate": 9.2e-07, + "loss": 0.7056, + "step": 46 + }, + { + "epoch": 0.03919116114238066, + "grad_norm": 0.14599646627902985, + "learning_rate": 9.399999999999999e-07, + "loss": 0.8098, + "step": 47 + }, + { + "epoch": 0.04002501563477173, + "grad_norm": 0.13951805233955383, + "learning_rate": 9.6e-07, + "loss": 0.7938, + "step": 48 + }, + { + "epoch": 0.040858870127162814, + "grad_norm": 0.14058919250965118, + "learning_rate": 9.8e-07, + "loss": 0.7806, + "step": 49 + }, + { + "epoch": 0.04169272461955389, + "grad_norm": 0.14300380647182465, + "learning_rate": 1e-06, + "loss": 0.8192, + "step": 50 + }, + { + "epoch": 0.04252657911194497, + "grad_norm": 0.13719584047794342, + "learning_rate": 1.02e-06, + "loss": 0.7939, + "step": 51 + }, + { + "epoch": 0.04336043360433604, + "grad_norm": 0.14002980291843414, + "learning_rate": 1.04e-06, + "loss": 0.8163, + "step": 52 + }, + { + "epoch": 0.044194288096727125, + "grad_norm": 0.13815677165985107, + "learning_rate": 1.06e-06, + "loss": 0.8137, + "step": 53 + }, + { + "epoch": 0.0450281425891182, + "grad_norm": 0.13772878050804138, + "learning_rate": 1.08e-06, + "loss": 0.771, + "step": 54 + }, + { + "epoch": 0.04586199708150928, + "grad_norm": 0.1414576917886734, + "learning_rate": 1.1e-06, + "loss": 0.7989, + "step": 55 + }, + { + "epoch": 0.046695851573900354, + "grad_norm": 0.11953306943178177, + "learning_rate": 1.12e-06, + "loss": 0.7074, + "step": 56 + }, + { + "epoch": 0.047529706066291436, + "grad_norm": 0.13051201403141022, + "learning_rate": 1.1399999999999999e-06, + "loss": 0.8034, + "step": 57 + }, + { + "epoch": 0.04836356055868251, + "grad_norm": 0.12802985310554504, + "learning_rate": 1.16e-06, + "loss": 0.7581, + "step": 58 + }, + { + "epoch": 0.04919741505107359, + "grad_norm": 0.13593435287475586, + "learning_rate": 1.18e-06, + "loss": 0.7978, + "step": 59 + }, + { + "epoch": 0.050031269543464665, + "grad_norm": 0.12595658004283905, + "learning_rate": 1.2e-06, + "loss": 0.7997, + "step": 60 + }, + { + "epoch": 0.05086512403585575, + "grad_norm": 0.1257220059633255, + "learning_rate": 1.22e-06, + "loss": 0.7194, + "step": 61 + }, + { + "epoch": 0.05169897852824682, + "grad_norm": 0.12169249355792999, + "learning_rate": 1.24e-06, + "loss": 0.7412, + "step": 62 + }, + { + "epoch": 0.0525328330206379, + "grad_norm": 0.12391054630279541, + "learning_rate": 1.26e-06, + "loss": 0.7849, + "step": 63 + }, + { + "epoch": 0.053366687513028976, + "grad_norm": 0.12134575098752975, + "learning_rate": 1.28e-06, + "loss": 0.6859, + "step": 64 + }, + { + "epoch": 0.05420054200542006, + "grad_norm": 0.11605742573738098, + "learning_rate": 1.3e-06, + "loss": 0.8112, + "step": 65 + }, + { + "epoch": 0.05503439649781113, + "grad_norm": 0.11593819409608841, + "learning_rate": 1.32e-06, + "loss": 0.6993, + "step": 66 + }, + { + "epoch": 0.05586825099020221, + "grad_norm": 0.12103109061717987, + "learning_rate": 1.34e-06, + "loss": 0.7595, + "step": 67 + }, + { + "epoch": 0.05670210548259329, + "grad_norm": 0.12391602247953415, + "learning_rate": 1.3600000000000001e-06, + "loss": 0.7676, + "step": 68 + }, + { + "epoch": 0.05753595997498437, + "grad_norm": 0.10895267128944397, + "learning_rate": 1.38e-06, + "loss": 0.7256, + "step": 69 + }, + { + "epoch": 0.05836981446737544, + "grad_norm": 0.1129627376794815, + "learning_rate": 1.4e-06, + "loss": 0.7097, + "step": 70 + }, + { + "epoch": 0.059203668959766524, + "grad_norm": 0.1078682616353035, + "learning_rate": 1.42e-06, + "loss": 0.7052, + "step": 71 + }, + { + "epoch": 0.0600375234521576, + "grad_norm": 0.1140650063753128, + "learning_rate": 1.44e-06, + "loss": 0.8067, + "step": 72 + }, + { + "epoch": 0.06087137794454868, + "grad_norm": 0.10528790205717087, + "learning_rate": 1.46e-06, + "loss": 0.7241, + "step": 73 + }, + { + "epoch": 0.061705232436939754, + "grad_norm": 0.11815836280584335, + "learning_rate": 1.48e-06, + "loss": 0.7344, + "step": 74 + }, + { + "epoch": 0.06253908692933083, + "grad_norm": 0.11140462756156921, + "learning_rate": 1.5e-06, + "loss": 0.7664, + "step": 75 + }, + { + "epoch": 0.06337294142172191, + "grad_norm": 0.11665944010019302, + "learning_rate": 1.5199999999999998e-06, + "loss": 0.7178, + "step": 76 + }, + { + "epoch": 0.06420679591411299, + "grad_norm": 0.10751160234212875, + "learning_rate": 1.5399999999999999e-06, + "loss": 0.7269, + "step": 77 + }, + { + "epoch": 0.06504065040650407, + "grad_norm": 0.09768356382846832, + "learning_rate": 1.5599999999999999e-06, + "loss": 0.7449, + "step": 78 + }, + { + "epoch": 0.06587450489889514, + "grad_norm": 0.09686972200870514, + "learning_rate": 1.58e-06, + "loss": 0.7097, + "step": 79 + }, + { + "epoch": 0.06670835939128622, + "grad_norm": 0.09527275711297989, + "learning_rate": 1.6e-06, + "loss": 0.7301, + "step": 80 + }, + { + "epoch": 0.0675422138836773, + "grad_norm": 0.0931195393204689, + "learning_rate": 1.62e-06, + "loss": 0.7524, + "step": 81 + }, + { + "epoch": 0.06837606837606838, + "grad_norm": 0.0892946794629097, + "learning_rate": 1.6399999999999998e-06, + "loss": 0.7156, + "step": 82 + }, + { + "epoch": 0.06920992286845945, + "grad_norm": 0.08845791220664978, + "learning_rate": 1.6599999999999998e-06, + "loss": 0.7349, + "step": 83 + }, + { + "epoch": 0.07004377736085053, + "grad_norm": 0.0896732360124588, + "learning_rate": 1.6799999999999998e-06, + "loss": 0.7145, + "step": 84 + }, + { + "epoch": 0.07087763185324161, + "grad_norm": 0.09997335076332092, + "learning_rate": 1.6999999999999998e-06, + "loss": 0.7103, + "step": 85 + }, + { + "epoch": 0.0717114863456327, + "grad_norm": 0.08821084350347519, + "learning_rate": 1.7199999999999998e-06, + "loss": 0.725, + "step": 86 + }, + { + "epoch": 0.07254534083802376, + "grad_norm": 0.08937980234622955, + "learning_rate": 1.7399999999999999e-06, + "loss": 0.7194, + "step": 87 + }, + { + "epoch": 0.07337919533041484, + "grad_norm": 0.09307857602834702, + "learning_rate": 1.7599999999999999e-06, + "loss": 0.7595, + "step": 88 + }, + { + "epoch": 0.07421304982280592, + "grad_norm": 0.08865829557180405, + "learning_rate": 1.78e-06, + "loss": 0.7464, + "step": 89 + }, + { + "epoch": 0.075046904315197, + "grad_norm": 0.09038495272397995, + "learning_rate": 1.8e-06, + "loss": 0.6698, + "step": 90 + }, + { + "epoch": 0.07588075880758807, + "grad_norm": 0.08254078030586243, + "learning_rate": 1.82e-06, + "loss": 0.7426, + "step": 91 + }, + { + "epoch": 0.07671461329997915, + "grad_norm": 0.08807505667209625, + "learning_rate": 1.84e-06, + "loss": 0.6789, + "step": 92 + }, + { + "epoch": 0.07754846779237023, + "grad_norm": 0.09523889422416687, + "learning_rate": 1.86e-06, + "loss": 0.7218, + "step": 93 + }, + { + "epoch": 0.07838232228476132, + "grad_norm": 0.08782277256250381, + "learning_rate": 1.8799999999999998e-06, + "loss": 0.7599, + "step": 94 + }, + { + "epoch": 0.07921617677715238, + "grad_norm": 0.08968065679073334, + "learning_rate": 1.8999999999999998e-06, + "loss": 0.7971, + "step": 95 + }, + { + "epoch": 0.08005003126954346, + "grad_norm": 0.08655782043933868, + "learning_rate": 1.92e-06, + "loss": 0.7479, + "step": 96 + }, + { + "epoch": 0.08088388576193455, + "grad_norm": 0.0854155421257019, + "learning_rate": 1.94e-06, + "loss": 0.7327, + "step": 97 + }, + { + "epoch": 0.08171774025432563, + "grad_norm": 0.08121508359909058, + "learning_rate": 1.96e-06, + "loss": 0.6749, + "step": 98 + }, + { + "epoch": 0.0825515947467167, + "grad_norm": 0.08429264277219772, + "learning_rate": 1.98e-06, + "loss": 0.6402, + "step": 99 + }, + { + "epoch": 0.08338544923910778, + "grad_norm": 0.08390086144208908, + "learning_rate": 2e-06, + "loss": 0.7434, + "step": 100 + }, + { + "epoch": 0.08421930373149886, + "grad_norm": 0.08662284910678864, + "learning_rate": 1.9999995964675577e-06, + "loss": 0.6681, + "step": 101 + }, + { + "epoch": 0.08505315822388994, + "grad_norm": 0.08700251579284668, + "learning_rate": 1.9999983858705566e-06, + "loss": 0.763, + "step": 102 + }, + { + "epoch": 0.085887012716281, + "grad_norm": 0.08504343777894974, + "learning_rate": 1.9999963682099734e-06, + "loss": 0.77, + "step": 103 + }, + { + "epoch": 0.08672086720867209, + "grad_norm": 0.08222745358943939, + "learning_rate": 1.999993543487437e-06, + "loss": 0.7103, + "step": 104 + }, + { + "epoch": 0.08755472170106317, + "grad_norm": 0.08124719560146332, + "learning_rate": 1.9999899117052263e-06, + "loss": 0.6913, + "step": 105 + }, + { + "epoch": 0.08838857619345425, + "grad_norm": 0.09033836424350739, + "learning_rate": 1.9999854728662734e-06, + "loss": 0.7065, + "step": 106 + }, + { + "epoch": 0.08922243068584532, + "grad_norm": 0.09045730531215668, + "learning_rate": 1.99998022697416e-06, + "loss": 0.7624, + "step": 107 + }, + { + "epoch": 0.0900562851782364, + "grad_norm": 0.0832991823554039, + "learning_rate": 1.9999741740331203e-06, + "loss": 0.7399, + "step": 108 + }, + { + "epoch": 0.09089013967062748, + "grad_norm": 0.07666011154651642, + "learning_rate": 1.9999673140480388e-06, + "loss": 0.6477, + "step": 109 + }, + { + "epoch": 0.09172399416301856, + "grad_norm": 0.08195007592439651, + "learning_rate": 1.9999596470244527e-06, + "loss": 0.7068, + "step": 110 + }, + { + "epoch": 0.09255784865540963, + "grad_norm": 0.07755164802074432, + "learning_rate": 1.99995117296855e-06, + "loss": 0.7021, + "step": 111 + }, + { + "epoch": 0.09339170314780071, + "grad_norm": 0.0782230794429779, + "learning_rate": 1.9999418918871685e-06, + "loss": 0.7039, + "step": 112 + }, + { + "epoch": 0.09422555764019179, + "grad_norm": 0.08311248570680618, + "learning_rate": 1.9999318037877995e-06, + "loss": 0.7238, + "step": 113 + }, + { + "epoch": 0.09505941213258287, + "grad_norm": 0.08455490320920944, + "learning_rate": 1.999920908678585e-06, + "loss": 0.7184, + "step": 114 + }, + { + "epoch": 0.09589326662497394, + "grad_norm": 0.08463476598262787, + "learning_rate": 1.9999092065683174e-06, + "loss": 0.6162, + "step": 115 + }, + { + "epoch": 0.09672712111736502, + "grad_norm": 0.07902689278125763, + "learning_rate": 1.999896697466442e-06, + "loss": 0.6649, + "step": 116 + }, + { + "epoch": 0.0975609756097561, + "grad_norm": 0.08383440971374512, + "learning_rate": 1.999883381383053e-06, + "loss": 0.703, + "step": 117 + }, + { + "epoch": 0.09839483010214718, + "grad_norm": 0.07288578897714615, + "learning_rate": 1.999869258328899e-06, + "loss": 0.7286, + "step": 118 + }, + { + "epoch": 0.09922868459453825, + "grad_norm": 0.08177065849304199, + "learning_rate": 1.999854328315377e-06, + "loss": 0.6899, + "step": 119 + }, + { + "epoch": 0.10006253908692933, + "grad_norm": 0.07529062032699585, + "learning_rate": 1.999838591354537e-06, + "loss": 0.6781, + "step": 120 + }, + { + "epoch": 0.10089639357932041, + "grad_norm": 0.08230678737163544, + "learning_rate": 1.9998220474590795e-06, + "loss": 0.7059, + "step": 121 + }, + { + "epoch": 0.1017302480717115, + "grad_norm": 0.08000528067350388, + "learning_rate": 1.9998046966423567e-06, + "loss": 0.6561, + "step": 122 + }, + { + "epoch": 0.10256410256410256, + "grad_norm": 0.07669328153133392, + "learning_rate": 1.9997865389183717e-06, + "loss": 0.6452, + "step": 123 + }, + { + "epoch": 0.10339795705649364, + "grad_norm": 0.07872146368026733, + "learning_rate": 1.9997675743017794e-06, + "loss": 0.7337, + "step": 124 + }, + { + "epoch": 0.10423181154888472, + "grad_norm": 0.07842138409614563, + "learning_rate": 1.9997478028078853e-06, + "loss": 0.7126, + "step": 125 + }, + { + "epoch": 0.1050656660412758, + "grad_norm": 0.07890679687261581, + "learning_rate": 1.9997272244526453e-06, + "loss": 0.7049, + "step": 126 + }, + { + "epoch": 0.10589952053366687, + "grad_norm": 0.07983887195587158, + "learning_rate": 1.999705839252669e-06, + "loss": 0.7578, + "step": 127 + }, + { + "epoch": 0.10673337502605795, + "grad_norm": 0.07721222192049026, + "learning_rate": 1.9996836472252144e-06, + "loss": 0.678, + "step": 128 + }, + { + "epoch": 0.10756722951844903, + "grad_norm": 0.07825150340795517, + "learning_rate": 1.9996606483881927e-06, + "loss": 0.6763, + "step": 129 + }, + { + "epoch": 0.10840108401084012, + "grad_norm": 0.08063995838165283, + "learning_rate": 1.999636842760165e-06, + "loss": 0.758, + "step": 130 + }, + { + "epoch": 0.10923493850323118, + "grad_norm": 0.0724928081035614, + "learning_rate": 1.9996122303603443e-06, + "loss": 0.7312, + "step": 131 + }, + { + "epoch": 0.11006879299562226, + "grad_norm": 0.08180397748947144, + "learning_rate": 1.999586811208594e-06, + "loss": 0.6794, + "step": 132 + }, + { + "epoch": 0.11090264748801335, + "grad_norm": 0.0856177881360054, + "learning_rate": 1.99956058532543e-06, + "loss": 0.6885, + "step": 133 + }, + { + "epoch": 0.11173650198040443, + "grad_norm": 0.0763324424624443, + "learning_rate": 1.999533552732017e-06, + "loss": 0.6821, + "step": 134 + }, + { + "epoch": 0.1125703564727955, + "grad_norm": 0.07964632660150528, + "learning_rate": 1.9995057134501725e-06, + "loss": 0.7362, + "step": 135 + }, + { + "epoch": 0.11340421096518657, + "grad_norm": 0.07633720338344574, + "learning_rate": 1.999477067502365e-06, + "loss": 0.7396, + "step": 136 + }, + { + "epoch": 0.11423806545757766, + "grad_norm": 0.07655656337738037, + "learning_rate": 1.9994476149117132e-06, + "loss": 0.728, + "step": 137 + }, + { + "epoch": 0.11507191994996874, + "grad_norm": 0.07770884782075882, + "learning_rate": 1.9994173557019877e-06, + "loss": 0.6878, + "step": 138 + }, + { + "epoch": 0.1159057744423598, + "grad_norm": 0.07332731038331985, + "learning_rate": 1.999386289897609e-06, + "loss": 0.7115, + "step": 139 + }, + { + "epoch": 0.11673962893475089, + "grad_norm": 0.0784095972776413, + "learning_rate": 1.9993544175236497e-06, + "loss": 0.6653, + "step": 140 + }, + { + "epoch": 0.11757348342714197, + "grad_norm": 0.08311771601438522, + "learning_rate": 1.9993217386058326e-06, + "loss": 0.6504, + "step": 141 + }, + { + "epoch": 0.11840733791953305, + "grad_norm": 0.07529395818710327, + "learning_rate": 1.999288253170532e-06, + "loss": 0.7083, + "step": 142 + }, + { + "epoch": 0.11924119241192412, + "grad_norm": 0.07608146965503693, + "learning_rate": 1.999253961244773e-06, + "loss": 0.6856, + "step": 143 + }, + { + "epoch": 0.1200750469043152, + "grad_norm": 0.07125360518693924, + "learning_rate": 1.9992188628562303e-06, + "loss": 0.6095, + "step": 144 + }, + { + "epoch": 0.12090890139670628, + "grad_norm": 0.07715660333633423, + "learning_rate": 1.999182958033232e-06, + "loss": 0.7337, + "step": 145 + }, + { + "epoch": 0.12174275588909736, + "grad_norm": 0.07756403833627701, + "learning_rate": 1.999146246804755e-06, + "loss": 0.6929, + "step": 146 + }, + { + "epoch": 0.12257661038148843, + "grad_norm": 0.07667449861764908, + "learning_rate": 1.9991087292004273e-06, + "loss": 0.67, + "step": 147 + }, + { + "epoch": 0.12341046487387951, + "grad_norm": 0.08285968005657196, + "learning_rate": 1.9990704052505284e-06, + "loss": 0.7483, + "step": 148 + }, + { + "epoch": 0.12424431936627059, + "grad_norm": 0.07476554065942764, + "learning_rate": 1.9990312749859887e-06, + "loss": 0.6846, + "step": 149 + }, + { + "epoch": 0.12507817385866166, + "grad_norm": 0.0818118005990982, + "learning_rate": 1.998991338438388e-06, + "loss": 0.6709, + "step": 150 + }, + { + "epoch": 0.12591202835105275, + "grad_norm": 0.08350253850221634, + "learning_rate": 1.998950595639958e-06, + "loss": 0.7379, + "step": 151 + }, + { + "epoch": 0.12674588284344382, + "grad_norm": 0.07916589826345444, + "learning_rate": 1.9989090466235806e-06, + "loss": 0.7206, + "step": 152 + }, + { + "epoch": 0.1275797373358349, + "grad_norm": 0.07860454171895981, + "learning_rate": 1.998866691422789e-06, + "loss": 0.6777, + "step": 153 + }, + { + "epoch": 0.12841359182822598, + "grad_norm": 0.07861131429672241, + "learning_rate": 1.998823530071766e-06, + "loss": 0.6938, + "step": 154 + }, + { + "epoch": 0.12924744632061705, + "grad_norm": 0.07819496840238571, + "learning_rate": 1.9987795626053465e-06, + "loss": 0.7193, + "step": 155 + }, + { + "epoch": 0.13008130081300814, + "grad_norm": 0.07828415185213089, + "learning_rate": 1.9987347890590144e-06, + "loss": 0.6809, + "step": 156 + }, + { + "epoch": 0.1309151553053992, + "grad_norm": 0.07755222916603088, + "learning_rate": 1.998689209468905e-06, + "loss": 0.7213, + "step": 157 + }, + { + "epoch": 0.13174900979779028, + "grad_norm": 0.07672612369060516, + "learning_rate": 1.998642823871804e-06, + "loss": 0.7196, + "step": 158 + }, + { + "epoch": 0.13258286429018137, + "grad_norm": 0.07963719964027405, + "learning_rate": 1.9985956323051475e-06, + "loss": 0.7465, + "step": 159 + }, + { + "epoch": 0.13341671878257244, + "grad_norm": 0.0795946940779686, + "learning_rate": 1.9985476348070223e-06, + "loss": 0.717, + "step": 160 + }, + { + "epoch": 0.1342505732749635, + "grad_norm": 0.07344063371419907, + "learning_rate": 1.9984988314161656e-06, + "loss": 0.6861, + "step": 161 + }, + { + "epoch": 0.1350844277673546, + "grad_norm": 0.07722171396017075, + "learning_rate": 1.9984492221719645e-06, + "loss": 0.6825, + "step": 162 + }, + { + "epoch": 0.13591828225974567, + "grad_norm": 0.0812341496348381, + "learning_rate": 1.9983988071144572e-06, + "loss": 0.7245, + "step": 163 + }, + { + "epoch": 0.13675213675213677, + "grad_norm": 0.08472032845020294, + "learning_rate": 1.998347586284332e-06, + "loss": 0.7315, + "step": 164 + }, + { + "epoch": 0.13758599124452783, + "grad_norm": 0.07785464823246002, + "learning_rate": 1.9982955597229275e-06, + "loss": 0.6906, + "step": 165 + }, + { + "epoch": 0.1384198457369189, + "grad_norm": 0.07382892072200775, + "learning_rate": 1.998242727472232e-06, + "loss": 0.6488, + "step": 166 + }, + { + "epoch": 0.13925370022931, + "grad_norm": 0.07445723563432693, + "learning_rate": 1.9981890895748845e-06, + "loss": 0.692, + "step": 167 + }, + { + "epoch": 0.14008755472170106, + "grad_norm": 0.07889281213283539, + "learning_rate": 1.998134646074175e-06, + "loss": 0.6438, + "step": 168 + }, + { + "epoch": 0.14092140921409213, + "grad_norm": 0.07664698362350464, + "learning_rate": 1.9980793970140426e-06, + "loss": 0.6993, + "step": 169 + }, + { + "epoch": 0.14175526370648323, + "grad_norm": 0.07807335257530212, + "learning_rate": 1.9980233424390773e-06, + "loss": 0.7334, + "step": 170 + }, + { + "epoch": 0.1425891181988743, + "grad_norm": 0.0683249682188034, + "learning_rate": 1.9979664823945174e-06, + "loss": 0.6325, + "step": 171 + }, + { + "epoch": 0.1434229726912654, + "grad_norm": 0.08339341729879379, + "learning_rate": 1.997908816926254e-06, + "loss": 0.697, + "step": 172 + }, + { + "epoch": 0.14425682718365646, + "grad_norm": 0.07781082391738892, + "learning_rate": 1.997850346080827e-06, + "loss": 0.6748, + "step": 173 + }, + { + "epoch": 0.14509068167604752, + "grad_norm": 0.07718750089406967, + "learning_rate": 1.9977910699054247e-06, + "loss": 0.7002, + "step": 174 + }, + { + "epoch": 0.14592453616843862, + "grad_norm": 0.07779830694198608, + "learning_rate": 1.997730988447888e-06, + "loss": 0.723, + "step": 175 + }, + { + "epoch": 0.14675839066082969, + "grad_norm": 0.07958701252937317, + "learning_rate": 1.997670101756706e-06, + "loss": 0.6651, + "step": 176 + }, + { + "epoch": 0.14759224515322075, + "grad_norm": 0.08142640441656113, + "learning_rate": 1.997608409881019e-06, + "loss": 0.6652, + "step": 177 + }, + { + "epoch": 0.14842609964561185, + "grad_norm": 0.07898704707622528, + "learning_rate": 1.9975459128706152e-06, + "loss": 0.6842, + "step": 178 + }, + { + "epoch": 0.14925995413800291, + "grad_norm": 0.07870691269636154, + "learning_rate": 1.997482610775935e-06, + "loss": 0.7167, + "step": 179 + }, + { + "epoch": 0.150093808630394, + "grad_norm": 0.07766727358102798, + "learning_rate": 1.9974185036480658e-06, + "loss": 0.6145, + "step": 180 + }, + { + "epoch": 0.15092766312278508, + "grad_norm": 0.07913653552532196, + "learning_rate": 1.997353591538748e-06, + "loss": 0.7598, + "step": 181 + }, + { + "epoch": 0.15176151761517614, + "grad_norm": 0.07850881665945053, + "learning_rate": 1.9972878745003684e-06, + "loss": 0.7492, + "step": 182 + }, + { + "epoch": 0.15259537210756724, + "grad_norm": 0.07599500566720963, + "learning_rate": 1.9972213525859656e-06, + "loss": 0.6802, + "step": 183 + }, + { + "epoch": 0.1534292265999583, + "grad_norm": 0.08165914565324783, + "learning_rate": 1.997154025849227e-06, + "loss": 0.6343, + "step": 184 + }, + { + "epoch": 0.15426308109234937, + "grad_norm": 0.0763566642999649, + "learning_rate": 1.9970858943444893e-06, + "loss": 0.7214, + "step": 185 + }, + { + "epoch": 0.15509693558474047, + "grad_norm": 0.07993835210800171, + "learning_rate": 1.99701695812674e-06, + "loss": 0.6975, + "step": 186 + }, + { + "epoch": 0.15593079007713154, + "grad_norm": 0.07139826565980911, + "learning_rate": 1.996947217251614e-06, + "loss": 0.6714, + "step": 187 + }, + { + "epoch": 0.15676464456952263, + "grad_norm": 0.0777943804860115, + "learning_rate": 1.996876671775397e-06, + "loss": 0.6684, + "step": 188 + }, + { + "epoch": 0.1575984990619137, + "grad_norm": 0.07822203636169434, + "learning_rate": 1.9968053217550242e-06, + "loss": 0.7161, + "step": 189 + }, + { + "epoch": 0.15843235355430477, + "grad_norm": 0.07645302265882492, + "learning_rate": 1.9967331672480794e-06, + "loss": 0.647, + "step": 190 + }, + { + "epoch": 0.15926620804669586, + "grad_norm": 0.07744666188955307, + "learning_rate": 1.996660208312796e-06, + "loss": 0.6887, + "step": 191 + }, + { + "epoch": 0.16010006253908693, + "grad_norm": 0.07616739720106125, + "learning_rate": 1.9965864450080565e-06, + "loss": 0.7209, + "step": 192 + }, + { + "epoch": 0.160933917031478, + "grad_norm": 0.0808212161064148, + "learning_rate": 1.996511877393393e-06, + "loss": 0.7229, + "step": 193 + }, + { + "epoch": 0.1617677715238691, + "grad_norm": 0.07879424095153809, + "learning_rate": 1.996436505528986e-06, + "loss": 0.6753, + "step": 194 + }, + { + "epoch": 0.16260162601626016, + "grad_norm": 0.07811419665813446, + "learning_rate": 1.9963603294756654e-06, + "loss": 0.6427, + "step": 195 + }, + { + "epoch": 0.16343548050865125, + "grad_norm": 0.08347994834184647, + "learning_rate": 1.996283349294911e-06, + "loss": 0.6575, + "step": 196 + }, + { + "epoch": 0.16426933500104232, + "grad_norm": 0.08415097743272781, + "learning_rate": 1.99620556504885e-06, + "loss": 0.6859, + "step": 197 + }, + { + "epoch": 0.1651031894934334, + "grad_norm": 0.08112502098083496, + "learning_rate": 1.9961269768002596e-06, + "loss": 0.6601, + "step": 198 + }, + { + "epoch": 0.16593704398582448, + "grad_norm": 0.07908611744642258, + "learning_rate": 1.9960475846125656e-06, + "loss": 0.6697, + "step": 199 + }, + { + "epoch": 0.16677089847821555, + "grad_norm": 0.07735829800367355, + "learning_rate": 1.995967388549843e-06, + "loss": 0.6631, + "step": 200 + }, + { + "epoch": 0.16760475297060662, + "grad_norm": 0.07564505934715271, + "learning_rate": 1.9958863886768145e-06, + "loss": 0.7004, + "step": 201 + }, + { + "epoch": 0.1684386074629977, + "grad_norm": 0.08117896318435669, + "learning_rate": 1.9958045850588527e-06, + "loss": 0.62, + "step": 202 + }, + { + "epoch": 0.16927246195538878, + "grad_norm": 0.0761483758687973, + "learning_rate": 1.9957219777619784e-06, + "loss": 0.7145, + "step": 203 + }, + { + "epoch": 0.17010631644777988, + "grad_norm": 0.08249073475599289, + "learning_rate": 1.995638566852861e-06, + "loss": 0.703, + "step": 204 + }, + { + "epoch": 0.17094017094017094, + "grad_norm": 0.08197219669818878, + "learning_rate": 1.9955543523988187e-06, + "loss": 0.695, + "step": 205 + }, + { + "epoch": 0.171774025432562, + "grad_norm": 0.07933253049850464, + "learning_rate": 1.995469334467818e-06, + "loss": 0.6575, + "step": 206 + }, + { + "epoch": 0.1726078799249531, + "grad_norm": 0.07888033986091614, + "learning_rate": 1.9953835131284735e-06, + "loss": 0.6504, + "step": 207 + }, + { + "epoch": 0.17344173441734417, + "grad_norm": 0.080783411860466, + "learning_rate": 1.995296888450049e-06, + "loss": 0.6586, + "step": 208 + }, + { + "epoch": 0.17427558890973524, + "grad_norm": 0.07863239198923111, + "learning_rate": 1.995209460502456e-06, + "loss": 0.7204, + "step": 209 + }, + { + "epoch": 0.17510944340212634, + "grad_norm": 0.07864314317703247, + "learning_rate": 1.9951212293562544e-06, + "loss": 0.6867, + "step": 210 + }, + { + "epoch": 0.1759432978945174, + "grad_norm": 0.07574496418237686, + "learning_rate": 1.995032195082653e-06, + "loss": 0.6806, + "step": 211 + }, + { + "epoch": 0.1767771523869085, + "grad_norm": 0.08079942315816879, + "learning_rate": 1.9949423577535078e-06, + "loss": 0.6929, + "step": 212 + }, + { + "epoch": 0.17761100687929957, + "grad_norm": 0.08361431211233139, + "learning_rate": 1.9948517174413235e-06, + "loss": 0.6948, + "step": 213 + }, + { + "epoch": 0.17844486137169063, + "grad_norm": 0.07974692434072495, + "learning_rate": 1.994760274219253e-06, + "loss": 0.7269, + "step": 214 + }, + { + "epoch": 0.17927871586408173, + "grad_norm": 0.08275672793388367, + "learning_rate": 1.9946680281610957e-06, + "loss": 0.6268, + "step": 215 + }, + { + "epoch": 0.1801125703564728, + "grad_norm": 0.08361789584159851, + "learning_rate": 1.9945749793413017e-06, + "loss": 0.6145, + "step": 216 + }, + { + "epoch": 0.18094642484886386, + "grad_norm": 0.0829552635550499, + "learning_rate": 1.9944811278349665e-06, + "loss": 0.7203, + "step": 217 + }, + { + "epoch": 0.18178027934125496, + "grad_norm": 0.07961665093898773, + "learning_rate": 1.994386473717835e-06, + "loss": 0.6598, + "step": 218 + }, + { + "epoch": 0.18261413383364603, + "grad_norm": 0.08015939593315125, + "learning_rate": 1.9942910170662986e-06, + "loss": 0.6465, + "step": 219 + }, + { + "epoch": 0.18344798832603712, + "grad_norm": 0.08055105805397034, + "learning_rate": 1.994194757957397e-06, + "loss": 0.6543, + "step": 220 + }, + { + "epoch": 0.1842818428184282, + "grad_norm": 0.07763723284006119, + "learning_rate": 1.994097696468818e-06, + "loss": 0.6984, + "step": 221 + }, + { + "epoch": 0.18511569731081925, + "grad_norm": 0.07882793247699738, + "learning_rate": 1.9939998326788966e-06, + "loss": 0.6672, + "step": 222 + }, + { + "epoch": 0.18594955180321035, + "grad_norm": 0.08098764717578888, + "learning_rate": 1.993901166666615e-06, + "loss": 0.65, + "step": 223 + }, + { + "epoch": 0.18678340629560142, + "grad_norm": 0.07791118323802948, + "learning_rate": 1.9938016985116027e-06, + "loss": 0.6954, + "step": 224 + }, + { + "epoch": 0.18761726078799248, + "grad_norm": 0.07814016193151474, + "learning_rate": 1.993701428294137e-06, + "loss": 0.6788, + "step": 225 + }, + { + "epoch": 0.18845111528038358, + "grad_norm": 0.07884248346090317, + "learning_rate": 1.9936003560951433e-06, + "loss": 0.6577, + "step": 226 + }, + { + "epoch": 0.18928496977277465, + "grad_norm": 0.07807621359825134, + "learning_rate": 1.9934984819961924e-06, + "loss": 0.6424, + "step": 227 + }, + { + "epoch": 0.19011882426516574, + "grad_norm": 0.08191139996051788, + "learning_rate": 1.9933958060795043e-06, + "loss": 0.6809, + "step": 228 + }, + { + "epoch": 0.1909526787575568, + "grad_norm": 0.07597807049751282, + "learning_rate": 1.9932923284279445e-06, + "loss": 0.7109, + "step": 229 + }, + { + "epoch": 0.19178653324994788, + "grad_norm": 0.08211328089237213, + "learning_rate": 1.9931880491250263e-06, + "loss": 0.6916, + "step": 230 + }, + { + "epoch": 0.19262038774233897, + "grad_norm": 0.07775567471981049, + "learning_rate": 1.993082968254909e-06, + "loss": 0.6458, + "step": 231 + }, + { + "epoch": 0.19345424223473004, + "grad_norm": 0.08438362181186676, + "learning_rate": 1.992977085902402e-06, + "loss": 0.6946, + "step": 232 + }, + { + "epoch": 0.1942880967271211, + "grad_norm": 0.08631843328475952, + "learning_rate": 1.9928704021529564e-06, + "loss": 0.6643, + "step": 233 + }, + { + "epoch": 0.1951219512195122, + "grad_norm": 0.07766929268836975, + "learning_rate": 1.992762917092675e-06, + "loss": 0.6753, + "step": 234 + }, + { + "epoch": 0.19595580571190327, + "grad_norm": 0.07574877887964249, + "learning_rate": 1.9926546308083044e-06, + "loss": 0.6346, + "step": 235 + }, + { + "epoch": 0.19678966020429436, + "grad_norm": 0.07815206050872803, + "learning_rate": 1.9925455433872386e-06, + "loss": 0.6946, + "step": 236 + }, + { + "epoch": 0.19762351469668543, + "grad_norm": 0.07457253336906433, + "learning_rate": 1.9924356549175185e-06, + "loss": 0.7108, + "step": 237 + }, + { + "epoch": 0.1984573691890765, + "grad_norm": 0.07688765227794647, + "learning_rate": 1.992324965487831e-06, + "loss": 0.6764, + "step": 238 + }, + { + "epoch": 0.1992912236814676, + "grad_norm": 0.08050046861171722, + "learning_rate": 1.99221347518751e-06, + "loss": 0.6263, + "step": 239 + }, + { + "epoch": 0.20012507817385866, + "grad_norm": 0.07763144373893738, + "learning_rate": 1.9921011841065347e-06, + "loss": 0.6618, + "step": 240 + }, + { + "epoch": 0.20095893266624973, + "grad_norm": 0.07744456082582474, + "learning_rate": 1.991988092335532e-06, + "loss": 0.6922, + "step": 241 + }, + { + "epoch": 0.20179278715864082, + "grad_norm": 0.07596256583929062, + "learning_rate": 1.9918741999657743e-06, + "loss": 0.7267, + "step": 242 + }, + { + "epoch": 0.2026266416510319, + "grad_norm": 0.08003415167331696, + "learning_rate": 1.9917595070891793e-06, + "loss": 0.6796, + "step": 243 + }, + { + "epoch": 0.203460496143423, + "grad_norm": 0.08418845385313034, + "learning_rate": 1.9916440137983125e-06, + "loss": 0.6815, + "step": 244 + }, + { + "epoch": 0.20429435063581405, + "grad_norm": 0.0794699415564537, + "learning_rate": 1.991527720186384e-06, + "loss": 0.6962, + "step": 245 + }, + { + "epoch": 0.20512820512820512, + "grad_norm": 0.07894127070903778, + "learning_rate": 1.991410626347251e-06, + "loss": 0.6898, + "step": 246 + }, + { + "epoch": 0.20596205962059622, + "grad_norm": 0.0819648951292038, + "learning_rate": 1.9912927323754145e-06, + "loss": 0.726, + "step": 247 + }, + { + "epoch": 0.20679591411298728, + "grad_norm": 0.08042652159929276, + "learning_rate": 1.991174038366024e-06, + "loss": 0.6398, + "step": 248 + }, + { + "epoch": 0.20762976860537835, + "grad_norm": 0.07848809659481049, + "learning_rate": 1.9910545444148722e-06, + "loss": 0.6562, + "step": 249 + }, + { + "epoch": 0.20846362309776945, + "grad_norm": 0.08295590430498123, + "learning_rate": 1.9909342506183987e-06, + "loss": 0.6767, + "step": 250 + }, + { + "epoch": 0.2092974775901605, + "grad_norm": 0.0840587317943573, + "learning_rate": 1.9908131570736887e-06, + "loss": 0.7274, + "step": 251 + }, + { + "epoch": 0.2101313320825516, + "grad_norm": 0.08375398814678192, + "learning_rate": 1.9906912638784727e-06, + "loss": 0.6716, + "step": 252 + }, + { + "epoch": 0.21096518657494268, + "grad_norm": 0.08229872584342957, + "learning_rate": 1.990568571131126e-06, + "loss": 0.7385, + "step": 253 + }, + { + "epoch": 0.21179904106733374, + "grad_norm": 0.07736154645681381, + "learning_rate": 1.99044507893067e-06, + "loss": 0.71, + "step": 254 + }, + { + "epoch": 0.21263289555972484, + "grad_norm": 0.08812221139669418, + "learning_rate": 1.9903207873767704e-06, + "loss": 0.7045, + "step": 255 + }, + { + "epoch": 0.2134667500521159, + "grad_norm": 0.08660853654146194, + "learning_rate": 1.9901956965697386e-06, + "loss": 0.7218, + "step": 256 + }, + { + "epoch": 0.21430060454450697, + "grad_norm": 0.08562801033258438, + "learning_rate": 1.9900698066105318e-06, + "loss": 0.6376, + "step": 257 + }, + { + "epoch": 0.21513445903689807, + "grad_norm": 0.07764414697885513, + "learning_rate": 1.9899431176007503e-06, + "loss": 0.6629, + "step": 258 + }, + { + "epoch": 0.21596831352928914, + "grad_norm": 0.07547292113304138, + "learning_rate": 1.989815629642641e-06, + "loss": 0.6887, + "step": 259 + }, + { + "epoch": 0.21680216802168023, + "grad_norm": 0.08463111519813538, + "learning_rate": 1.9896873428390948e-06, + "loss": 0.6169, + "step": 260 + }, + { + "epoch": 0.2176360225140713, + "grad_norm": 0.08064530789852142, + "learning_rate": 1.9895582572936473e-06, + "loss": 0.6905, + "step": 261 + }, + { + "epoch": 0.21846987700646237, + "grad_norm": 0.07929642498493195, + "learning_rate": 1.9894283731104796e-06, + "loss": 0.6415, + "step": 262 + }, + { + "epoch": 0.21930373149885346, + "grad_norm": 0.07858303934335709, + "learning_rate": 1.9892976903944154e-06, + "loss": 0.6955, + "step": 263 + }, + { + "epoch": 0.22013758599124453, + "grad_norm": 0.07967871427536011, + "learning_rate": 1.9891662092509255e-06, + "loss": 0.7004, + "step": 264 + }, + { + "epoch": 0.2209714404836356, + "grad_norm": 0.08009492605924606, + "learning_rate": 1.989033929786123e-06, + "loss": 0.6206, + "step": 265 + }, + { + "epoch": 0.2218052949760267, + "grad_norm": 0.08245829492807388, + "learning_rate": 1.988900852106766e-06, + "loss": 0.7182, + "step": 266 + }, + { + "epoch": 0.22263914946841776, + "grad_norm": 0.07844138890504837, + "learning_rate": 1.9887669763202568e-06, + "loss": 0.6372, + "step": 267 + }, + { + "epoch": 0.22347300396080885, + "grad_norm": 0.0783441960811615, + "learning_rate": 1.9886323025346417e-06, + "loss": 0.6182, + "step": 268 + }, + { + "epoch": 0.22430685845319992, + "grad_norm": 0.0803346335887909, + "learning_rate": 1.9884968308586117e-06, + "loss": 0.6436, + "step": 269 + }, + { + "epoch": 0.225140712945591, + "grad_norm": 0.0813342034816742, + "learning_rate": 1.9883605614015015e-06, + "loss": 0.6472, + "step": 270 + }, + { + "epoch": 0.22597456743798208, + "grad_norm": 0.07891476154327393, + "learning_rate": 1.988223494273288e-06, + "loss": 0.6625, + "step": 271 + }, + { + "epoch": 0.22680842193037315, + "grad_norm": 0.08087664842605591, + "learning_rate": 1.9880856295845945e-06, + "loss": 0.7226, + "step": 272 + }, + { + "epoch": 0.22764227642276422, + "grad_norm": 0.08254389464855194, + "learning_rate": 1.9879469674466865e-06, + "loss": 0.6992, + "step": 273 + }, + { + "epoch": 0.2284761309151553, + "grad_norm": 0.08207474648952484, + "learning_rate": 1.987807507971473e-06, + "loss": 0.6827, + "step": 274 + }, + { + "epoch": 0.22930998540754638, + "grad_norm": 0.08394856005907059, + "learning_rate": 1.9876672512715075e-06, + "loss": 0.7445, + "step": 275 + }, + { + "epoch": 0.23014383989993747, + "grad_norm": 0.07934897392988205, + "learning_rate": 1.9875261974599856e-06, + "loss": 0.6988, + "step": 276 + }, + { + "epoch": 0.23097769439232854, + "grad_norm": 0.08756977319717407, + "learning_rate": 1.987384346650747e-06, + "loss": 0.6588, + "step": 277 + }, + { + "epoch": 0.2318115488847196, + "grad_norm": 0.08198294788599014, + "learning_rate": 1.9872416989582747e-06, + "loss": 0.6729, + "step": 278 + }, + { + "epoch": 0.2326454033771107, + "grad_norm": 0.07671192288398743, + "learning_rate": 1.9870982544976948e-06, + "loss": 0.6685, + "step": 279 + }, + { + "epoch": 0.23347925786950177, + "grad_norm": 0.07943634688854218, + "learning_rate": 1.986954013384776e-06, + "loss": 0.7352, + "step": 280 + }, + { + "epoch": 0.23431311236189284, + "grad_norm": 0.0829697921872139, + "learning_rate": 1.98680897573593e-06, + "loss": 0.6645, + "step": 281 + }, + { + "epoch": 0.23514696685428393, + "grad_norm": 0.08583538234233856, + "learning_rate": 1.986663141668212e-06, + "loss": 0.713, + "step": 282 + }, + { + "epoch": 0.235980821346675, + "grad_norm": 0.08206543326377869, + "learning_rate": 1.9865165112993192e-06, + "loss": 0.7211, + "step": 283 + }, + { + "epoch": 0.2368146758390661, + "grad_norm": 0.07899665087461472, + "learning_rate": 1.9863690847475924e-06, + "loss": 0.6757, + "step": 284 + }, + { + "epoch": 0.23764853033145716, + "grad_norm": 0.0813852995634079, + "learning_rate": 1.986220862132014e-06, + "loss": 0.692, + "step": 285 + }, + { + "epoch": 0.23848238482384823, + "grad_norm": 0.09079395979642868, + "learning_rate": 1.9860718435722093e-06, + "loss": 0.6756, + "step": 286 + }, + { + "epoch": 0.23931623931623933, + "grad_norm": 0.07518535107374191, + "learning_rate": 1.985922029188446e-06, + "loss": 0.626, + "step": 287 + }, + { + "epoch": 0.2401500938086304, + "grad_norm": 0.08228084444999695, + "learning_rate": 1.9857714191016337e-06, + "loss": 0.7394, + "step": 288 + }, + { + "epoch": 0.24098394830102146, + "grad_norm": 0.08036590367555618, + "learning_rate": 1.985620013433325e-06, + "loss": 0.6012, + "step": 289 + }, + { + "epoch": 0.24181780279341256, + "grad_norm": 0.08429574221372604, + "learning_rate": 1.985467812305714e-06, + "loss": 0.6881, + "step": 290 + }, + { + "epoch": 0.24265165728580362, + "grad_norm": 0.07866602391004562, + "learning_rate": 1.9853148158416364e-06, + "loss": 0.6708, + "step": 291 + }, + { + "epoch": 0.24348551177819472, + "grad_norm": 0.0796622484922409, + "learning_rate": 1.9851610241645714e-06, + "loss": 0.7129, + "step": 292 + }, + { + "epoch": 0.24431936627058579, + "grad_norm": 0.08136799931526184, + "learning_rate": 1.9850064373986374e-06, + "loss": 0.6806, + "step": 293 + }, + { + "epoch": 0.24515322076297685, + "grad_norm": 0.07808342576026917, + "learning_rate": 1.9848510556685973e-06, + "loss": 0.6217, + "step": 294 + }, + { + "epoch": 0.24598707525536795, + "grad_norm": 0.07293733209371567, + "learning_rate": 1.984694879099853e-06, + "loss": 0.662, + "step": 295 + }, + { + "epoch": 0.24682092974775902, + "grad_norm": 0.07730741053819656, + "learning_rate": 1.98453790781845e-06, + "loss": 0.6386, + "step": 296 + }, + { + "epoch": 0.24765478424015008, + "grad_norm": 0.08601437509059906, + "learning_rate": 1.9843801419510743e-06, + "loss": 0.6639, + "step": 297 + }, + { + "epoch": 0.24848863873254118, + "grad_norm": 0.0815718024969101, + "learning_rate": 1.9842215816250525e-06, + "loss": 0.6891, + "step": 298 + }, + { + "epoch": 0.24932249322493225, + "grad_norm": 0.08793067187070847, + "learning_rate": 1.9840622269683537e-06, + "loss": 0.684, + "step": 299 + }, + { + "epoch": 0.2501563477173233, + "grad_norm": 0.07922165095806122, + "learning_rate": 1.9839020781095873e-06, + "loss": 0.6913, + "step": 300 + }, + { + "epoch": 0.2509902022097144, + "grad_norm": 0.0789838656783104, + "learning_rate": 1.9837411351780035e-06, + "loss": 0.6556, + "step": 301 + }, + { + "epoch": 0.2518240567021055, + "grad_norm": 0.07958388328552246, + "learning_rate": 1.9835793983034944e-06, + "loss": 0.6301, + "step": 302 + }, + { + "epoch": 0.25265791119449654, + "grad_norm": 0.08036024868488312, + "learning_rate": 1.9834168676165915e-06, + "loss": 0.592, + "step": 303 + }, + { + "epoch": 0.25349176568688764, + "grad_norm": 0.08134333044290543, + "learning_rate": 1.9832535432484677e-06, + "loss": 0.7142, + "step": 304 + }, + { + "epoch": 0.25432562017927873, + "grad_norm": 0.08333199471235275, + "learning_rate": 1.983089425330937e-06, + "loss": 0.7021, + "step": 305 + }, + { + "epoch": 0.2551594746716698, + "grad_norm": 0.08486991375684738, + "learning_rate": 1.982924513996452e-06, + "loss": 0.6724, + "step": 306 + }, + { + "epoch": 0.25599332916406087, + "grad_norm": 0.08286510407924652, + "learning_rate": 1.982758809378108e-06, + "loss": 0.6856, + "step": 307 + }, + { + "epoch": 0.25682718365645196, + "grad_norm": 0.08175141364336014, + "learning_rate": 1.982592311609639e-06, + "loss": 0.6605, + "step": 308 + }, + { + "epoch": 0.257661038148843, + "grad_norm": 0.08094872534275055, + "learning_rate": 1.982425020825419e-06, + "loss": 0.7136, + "step": 309 + }, + { + "epoch": 0.2584948926412341, + "grad_norm": 0.07797659933567047, + "learning_rate": 1.9822569371604632e-06, + "loss": 0.6975, + "step": 310 + }, + { + "epoch": 0.2593287471336252, + "grad_norm": 0.08523422479629517, + "learning_rate": 1.982088060750426e-06, + "loss": 0.6487, + "step": 311 + }, + { + "epoch": 0.2601626016260163, + "grad_norm": 0.09064790606498718, + "learning_rate": 1.981918391731601e-06, + "loss": 0.7006, + "step": 312 + }, + { + "epoch": 0.2609964561184073, + "grad_norm": 0.08136545121669769, + "learning_rate": 1.9817479302409225e-06, + "loss": 0.67, + "step": 313 + }, + { + "epoch": 0.2618303106107984, + "grad_norm": 0.07847245037555695, + "learning_rate": 1.9815766764159642e-06, + "loss": 0.6692, + "step": 314 + }, + { + "epoch": 0.2626641651031895, + "grad_norm": 0.07690361142158508, + "learning_rate": 1.9814046303949387e-06, + "loss": 0.5999, + "step": 315 + }, + { + "epoch": 0.26349801959558056, + "grad_norm": 0.08166715502738953, + "learning_rate": 1.9812317923166987e-06, + "loss": 0.6666, + "step": 316 + }, + { + "epoch": 0.26433187408797165, + "grad_norm": 0.08074827492237091, + "learning_rate": 1.981058162320735e-06, + "loss": 0.657, + "step": 317 + }, + { + "epoch": 0.26516572858036275, + "grad_norm": 0.08311966806650162, + "learning_rate": 1.980883740547179e-06, + "loss": 0.6679, + "step": 318 + }, + { + "epoch": 0.2659995830727538, + "grad_norm": 0.08272409439086914, + "learning_rate": 1.9807085271368e-06, + "loss": 0.6467, + "step": 319 + }, + { + "epoch": 0.2668334375651449, + "grad_norm": 0.08190209418535233, + "learning_rate": 1.980532522231007e-06, + "loss": 0.623, + "step": 320 + }, + { + "epoch": 0.267667292057536, + "grad_norm": 0.0819978415966034, + "learning_rate": 1.980355725971847e-06, + "loss": 0.677, + "step": 321 + }, + { + "epoch": 0.268501146549927, + "grad_norm": 0.07838036119937897, + "learning_rate": 1.980178138502006e-06, + "loss": 0.6545, + "step": 322 + }, + { + "epoch": 0.2693350010423181, + "grad_norm": 0.08094768226146698, + "learning_rate": 1.979999759964809e-06, + "loss": 0.6429, + "step": 323 + }, + { + "epoch": 0.2701688555347092, + "grad_norm": 0.0796407088637352, + "learning_rate": 1.9798205905042184e-06, + "loss": 0.7114, + "step": 324 + }, + { + "epoch": 0.27100271002710025, + "grad_norm": 0.07667295634746552, + "learning_rate": 1.9796406302648366e-06, + "loss": 0.7059, + "step": 325 + }, + { + "epoch": 0.27183656451949134, + "grad_norm": 0.07456690818071365, + "learning_rate": 1.9794598793919023e-06, + "loss": 0.5275, + "step": 326 + }, + { + "epoch": 0.27267041901188244, + "grad_norm": 0.08084844797849655, + "learning_rate": 1.9792783380312935e-06, + "loss": 0.6449, + "step": 327 + }, + { + "epoch": 0.27350427350427353, + "grad_norm": 0.08692745119333267, + "learning_rate": 1.9790960063295254e-06, + "loss": 0.6831, + "step": 328 + }, + { + "epoch": 0.27433812799666457, + "grad_norm": 0.0852961540222168, + "learning_rate": 1.9789128844337524e-06, + "loss": 0.6518, + "step": 329 + }, + { + "epoch": 0.27517198248905567, + "grad_norm": 0.08140264451503754, + "learning_rate": 1.978728972491765e-06, + "loss": 0.606, + "step": 330 + }, + { + "epoch": 0.27600583698144676, + "grad_norm": 0.08339618146419525, + "learning_rate": 1.9785442706519927e-06, + "loss": 0.6792, + "step": 331 + }, + { + "epoch": 0.2768396914738378, + "grad_norm": 0.08711510896682739, + "learning_rate": 1.9783587790635017e-06, + "loss": 0.6903, + "step": 332 + }, + { + "epoch": 0.2776735459662289, + "grad_norm": 0.08612138777971268, + "learning_rate": 1.978172497875995e-06, + "loss": 0.6998, + "step": 333 + }, + { + "epoch": 0.27850740045862, + "grad_norm": 0.07922092080116272, + "learning_rate": 1.9779854272398146e-06, + "loss": 0.654, + "step": 334 + }, + { + "epoch": 0.27934125495101103, + "grad_norm": 0.08098642528057098, + "learning_rate": 1.9777975673059382e-06, + "loss": 0.6706, + "step": 335 + }, + { + "epoch": 0.2801751094434021, + "grad_norm": 0.0873970165848732, + "learning_rate": 1.977608918225981e-06, + "loss": 0.6856, + "step": 336 + }, + { + "epoch": 0.2810089639357932, + "grad_norm": 0.082435242831707, + "learning_rate": 1.9774194801521947e-06, + "loss": 0.6375, + "step": 337 + }, + { + "epoch": 0.28184281842818426, + "grad_norm": 0.08199653774499893, + "learning_rate": 1.9772292532374687e-06, + "loss": 0.6785, + "step": 338 + }, + { + "epoch": 0.28267667292057536, + "grad_norm": 0.08817990124225616, + "learning_rate": 1.9770382376353285e-06, + "loss": 0.6642, + "step": 339 + }, + { + "epoch": 0.28351052741296645, + "grad_norm": 0.08359472453594208, + "learning_rate": 1.976846433499935e-06, + "loss": 0.6555, + "step": 340 + }, + { + "epoch": 0.2843443819053575, + "grad_norm": 0.08499454706907272, + "learning_rate": 1.976653840986088e-06, + "loss": 0.687, + "step": 341 + }, + { + "epoch": 0.2851782363977486, + "grad_norm": 0.08286130428314209, + "learning_rate": 1.976460460249222e-06, + "loss": 0.6904, + "step": 342 + }, + { + "epoch": 0.2860120908901397, + "grad_norm": 0.08244482427835464, + "learning_rate": 1.976266291445406e-06, + "loss": 0.6546, + "step": 343 + }, + { + "epoch": 0.2868459453825308, + "grad_norm": 0.07950432598590851, + "learning_rate": 1.976071334731349e-06, + "loss": 0.6683, + "step": 344 + }, + { + "epoch": 0.2876797998749218, + "grad_norm": 0.0854944959282875, + "learning_rate": 1.975875590264393e-06, + "loss": 0.6947, + "step": 345 + }, + { + "epoch": 0.2885136543673129, + "grad_norm": 0.08143036812543869, + "learning_rate": 1.975679058202516e-06, + "loss": 0.6422, + "step": 346 + }, + { + "epoch": 0.289347508859704, + "grad_norm": 0.0797310546040535, + "learning_rate": 1.9754817387043325e-06, + "loss": 0.6939, + "step": 347 + }, + { + "epoch": 0.29018136335209505, + "grad_norm": 0.08961952477693558, + "learning_rate": 1.9752836319290926e-06, + "loss": 0.6848, + "step": 348 + }, + { + "epoch": 0.29101521784448614, + "grad_norm": 0.08825525641441345, + "learning_rate": 1.9750847380366806e-06, + "loss": 0.7088, + "step": 349 + }, + { + "epoch": 0.29184907233687724, + "grad_norm": 0.08601095527410507, + "learning_rate": 1.9748850571876168e-06, + "loss": 0.6283, + "step": 350 + }, + { + "epoch": 0.2926829268292683, + "grad_norm": 0.08371152728796005, + "learning_rate": 1.9746845895430568e-06, + "loss": 0.6755, + "step": 351 + }, + { + "epoch": 0.29351678132165937, + "grad_norm": 0.08786465227603912, + "learning_rate": 1.974483335264791e-06, + "loss": 0.6149, + "step": 352 + }, + { + "epoch": 0.29435063581405047, + "grad_norm": 0.083786241710186, + "learning_rate": 1.974281294515245e-06, + "loss": 0.7104, + "step": 353 + }, + { + "epoch": 0.2951844903064415, + "grad_norm": 0.0846792683005333, + "learning_rate": 1.9740784674574785e-06, + "loss": 0.7032, + "step": 354 + }, + { + "epoch": 0.2960183447988326, + "grad_norm": 0.08477747440338135, + "learning_rate": 1.973874854255186e-06, + "loss": 0.6244, + "step": 355 + }, + { + "epoch": 0.2968521992912237, + "grad_norm": 0.09119998663663864, + "learning_rate": 1.9736704550726965e-06, + "loss": 0.6373, + "step": 356 + }, + { + "epoch": 0.29768605378361473, + "grad_norm": 0.08474856615066528, + "learning_rate": 1.9734652700749733e-06, + "loss": 0.736, + "step": 357 + }, + { + "epoch": 0.29851990827600583, + "grad_norm": 0.08070901036262512, + "learning_rate": 1.9732592994276145e-06, + "loss": 0.6442, + "step": 358 + }, + { + "epoch": 0.2993537627683969, + "grad_norm": 0.08306030929088593, + "learning_rate": 1.9730525432968517e-06, + "loss": 0.6863, + "step": 359 + }, + { + "epoch": 0.300187617260788, + "grad_norm": 0.08554356545209885, + "learning_rate": 1.97284500184955e-06, + "loss": 0.6782, + "step": 360 + }, + { + "epoch": 0.30102147175317906, + "grad_norm": 0.08350827544927597, + "learning_rate": 1.9726366752532096e-06, + "loss": 0.6656, + "step": 361 + }, + { + "epoch": 0.30185532624557015, + "grad_norm": 0.08501608669757843, + "learning_rate": 1.972427563675963e-06, + "loss": 0.6651, + "step": 362 + }, + { + "epoch": 0.30268918073796125, + "grad_norm": 0.08401606976985931, + "learning_rate": 1.972217667286577e-06, + "loss": 0.7153, + "step": 363 + }, + { + "epoch": 0.3035230352303523, + "grad_norm": 0.08713418990373611, + "learning_rate": 1.9720069862544513e-06, + "loss": 0.6623, + "step": 364 + }, + { + "epoch": 0.3043568897227434, + "grad_norm": 0.0871342197060585, + "learning_rate": 1.9717955207496197e-06, + "loss": 0.6571, + "step": 365 + }, + { + "epoch": 0.3051907442151345, + "grad_norm": 0.08501847088336945, + "learning_rate": 1.9715832709427477e-06, + "loss": 0.6714, + "step": 366 + }, + { + "epoch": 0.3060245987075255, + "grad_norm": 0.07894369214773178, + "learning_rate": 1.971370237005136e-06, + "loss": 0.6737, + "step": 367 + }, + { + "epoch": 0.3068584531999166, + "grad_norm": 0.08392394334077835, + "learning_rate": 1.9711564191087157e-06, + "loss": 0.6807, + "step": 368 + }, + { + "epoch": 0.3076923076923077, + "grad_norm": 0.08396563678979874, + "learning_rate": 1.970941817426052e-06, + "loss": 0.6766, + "step": 369 + }, + { + "epoch": 0.30852616218469875, + "grad_norm": 0.07790801674127579, + "learning_rate": 1.9707264321303425e-06, + "loss": 0.6293, + "step": 370 + }, + { + "epoch": 0.30936001667708984, + "grad_norm": 0.08699779212474823, + "learning_rate": 1.970510263395417e-06, + "loss": 0.681, + "step": 371 + }, + { + "epoch": 0.31019387116948094, + "grad_norm": 0.08176835626363754, + "learning_rate": 1.9702933113957384e-06, + "loss": 0.6931, + "step": 372 + }, + { + "epoch": 0.311027725661872, + "grad_norm": 0.08946527540683746, + "learning_rate": 1.9700755763063997e-06, + "loss": 0.6062, + "step": 373 + }, + { + "epoch": 0.3118615801542631, + "grad_norm": 0.08260657638311386, + "learning_rate": 1.9698570583031284e-06, + "loss": 0.6466, + "step": 374 + }, + { + "epoch": 0.31269543464665417, + "grad_norm": 0.08951833099126816, + "learning_rate": 1.969637757562282e-06, + "loss": 0.6458, + "step": 375 + }, + { + "epoch": 0.31352928913904526, + "grad_norm": 0.08605215698480606, + "learning_rate": 1.9694176742608507e-06, + "loss": 0.6919, + "step": 376 + }, + { + "epoch": 0.3143631436314363, + "grad_norm": 0.08389750868082047, + "learning_rate": 1.969196808576456e-06, + "loss": 0.6941, + "step": 377 + }, + { + "epoch": 0.3151969981238274, + "grad_norm": 0.08429732173681259, + "learning_rate": 1.968975160687351e-06, + "loss": 0.622, + "step": 378 + }, + { + "epoch": 0.3160308526162185, + "grad_norm": 0.08689334988594055, + "learning_rate": 1.9687527307724195e-06, + "loss": 0.6235, + "step": 379 + }, + { + "epoch": 0.31686470710860953, + "grad_norm": 0.08494516462087631, + "learning_rate": 1.9685295190111774e-06, + "loss": 0.6491, + "step": 380 + }, + { + "epoch": 0.31769856160100063, + "grad_norm": 0.0890749841928482, + "learning_rate": 1.9683055255837708e-06, + "loss": 0.6791, + "step": 381 + }, + { + "epoch": 0.3185324160933917, + "grad_norm": 0.07894483208656311, + "learning_rate": 1.968080750670977e-06, + "loss": 0.7075, + "step": 382 + }, + { + "epoch": 0.31936627058578276, + "grad_norm": 0.08575093746185303, + "learning_rate": 1.9678551944542033e-06, + "loss": 0.7027, + "step": 383 + }, + { + "epoch": 0.32020012507817386, + "grad_norm": 0.08360203355550766, + "learning_rate": 1.9676288571154895e-06, + "loss": 0.5894, + "step": 384 + }, + { + "epoch": 0.32103397957056495, + "grad_norm": 0.08341158181428909, + "learning_rate": 1.9674017388375036e-06, + "loss": 0.658, + "step": 385 + }, + { + "epoch": 0.321867834062956, + "grad_norm": 0.0844089537858963, + "learning_rate": 1.9671738398035446e-06, + "loss": 0.6691, + "step": 386 + }, + { + "epoch": 0.3227016885553471, + "grad_norm": 0.08446256816387177, + "learning_rate": 1.9669451601975426e-06, + "loss": 0.7211, + "step": 387 + }, + { + "epoch": 0.3235355430477382, + "grad_norm": 0.08008047193288803, + "learning_rate": 1.9667157002040565e-06, + "loss": 0.5843, + "step": 388 + }, + { + "epoch": 0.3243693975401292, + "grad_norm": 0.08821803331375122, + "learning_rate": 1.9664854600082754e-06, + "loss": 0.6444, + "step": 389 + }, + { + "epoch": 0.3252032520325203, + "grad_norm": 0.08548033237457275, + "learning_rate": 1.966254439796018e-06, + "loss": 0.6643, + "step": 390 + }, + { + "epoch": 0.3260371065249114, + "grad_norm": 0.08261854946613312, + "learning_rate": 1.9660226397537322e-06, + "loss": 0.6469, + "step": 391 + }, + { + "epoch": 0.3268709610173025, + "grad_norm": 0.07960055768489838, + "learning_rate": 1.965790060068497e-06, + "loss": 0.6149, + "step": 392 + }, + { + "epoch": 0.32770481550969355, + "grad_norm": 0.0834561213850975, + "learning_rate": 1.9655567009280177e-06, + "loss": 0.6798, + "step": 393 + }, + { + "epoch": 0.32853867000208464, + "grad_norm": 0.08524155616760254, + "learning_rate": 1.9653225625206313e-06, + "loss": 0.6999, + "step": 394 + }, + { + "epoch": 0.32937252449447574, + "grad_norm": 0.08503536134958267, + "learning_rate": 1.965087645035302e-06, + "loss": 0.6674, + "step": 395 + }, + { + "epoch": 0.3302063789868668, + "grad_norm": 0.09243112802505493, + "learning_rate": 1.964851948661624e-06, + "loss": 0.6391, + "step": 396 + }, + { + "epoch": 0.3310402334792579, + "grad_norm": 0.08015627413988113, + "learning_rate": 1.96461547358982e-06, + "loss": 0.611, + "step": 397 + }, + { + "epoch": 0.33187408797164897, + "grad_norm": 0.08681736886501312, + "learning_rate": 1.9643782200107395e-06, + "loss": 0.6967, + "step": 398 + }, + { + "epoch": 0.33270794246404, + "grad_norm": 0.08576110005378723, + "learning_rate": 1.9641401881158623e-06, + "loss": 0.6753, + "step": 399 + }, + { + "epoch": 0.3335417969564311, + "grad_norm": 0.08432309329509735, + "learning_rate": 1.9639013780972954e-06, + "loss": 0.6312, + "step": 400 + }, + { + "epoch": 0.3335417969564311, + "eval_loss": 0.6676867008209229, + "eval_runtime": 321.9965, + "eval_samples_per_second": 16.718, + "eval_steps_per_second": 2.789, + "step": 400 + }, + { + "epoch": 0.3343756514488222, + "grad_norm": 0.08781838417053223, + "learning_rate": 1.9636617901477742e-06, + "loss": 0.698, + "step": 401 + }, + { + "epoch": 0.33520950594121324, + "grad_norm": 0.08723490685224533, + "learning_rate": 1.9634214244606613e-06, + "loss": 0.6642, + "step": 402 + }, + { + "epoch": 0.33604336043360433, + "grad_norm": 0.08249269425868988, + "learning_rate": 1.963180281229948e-06, + "loss": 0.705, + "step": 403 + }, + { + "epoch": 0.3368772149259954, + "grad_norm": 0.08193965256214142, + "learning_rate": 1.962938360650252e-06, + "loss": 0.6813, + "step": 404 + }, + { + "epoch": 0.33771106941838647, + "grad_norm": 0.08125729858875275, + "learning_rate": 1.962695662916819e-06, + "loss": 0.6571, + "step": 405 + }, + { + "epoch": 0.33854492391077756, + "grad_norm": 0.08508846163749695, + "learning_rate": 1.962452188225522e-06, + "loss": 0.6212, + "step": 406 + }, + { + "epoch": 0.33937877840316866, + "grad_norm": 0.0881478488445282, + "learning_rate": 1.962207936772861e-06, + "loss": 0.6582, + "step": 407 + }, + { + "epoch": 0.34021263289555975, + "grad_norm": 0.08948707580566406, + "learning_rate": 1.9619629087559622e-06, + "loss": 0.6924, + "step": 408 + }, + { + "epoch": 0.3410464873879508, + "grad_norm": 0.08214866369962692, + "learning_rate": 1.9617171043725797e-06, + "loss": 0.6266, + "step": 409 + }, + { + "epoch": 0.3418803418803419, + "grad_norm": 0.09406815469264984, + "learning_rate": 1.961470523821093e-06, + "loss": 0.6754, + "step": 410 + }, + { + "epoch": 0.342714196372733, + "grad_norm": 0.08819446712732315, + "learning_rate": 1.961223167300509e-06, + "loss": 0.6858, + "step": 411 + }, + { + "epoch": 0.343548050865124, + "grad_norm": 0.0902935117483139, + "learning_rate": 1.9609750350104606e-06, + "loss": 0.6654, + "step": 412 + }, + { + "epoch": 0.3443819053575151, + "grad_norm": 0.0852714329957962, + "learning_rate": 1.9607261271512065e-06, + "loss": 0.6604, + "step": 413 + }, + { + "epoch": 0.3452157598499062, + "grad_norm": 0.08480080962181091, + "learning_rate": 1.9604764439236313e-06, + "loss": 0.7252, + "step": 414 + }, + { + "epoch": 0.34604961434229725, + "grad_norm": 0.08467400074005127, + "learning_rate": 1.960225985529246e-06, + "loss": 0.6182, + "step": 415 + }, + { + "epoch": 0.34688346883468835, + "grad_norm": 0.09190120548009872, + "learning_rate": 1.9599747521701865e-06, + "loss": 0.6605, + "step": 416 + }, + { + "epoch": 0.34771732332707944, + "grad_norm": 0.08994212001562119, + "learning_rate": 1.9597227440492143e-06, + "loss": 0.6674, + "step": 417 + }, + { + "epoch": 0.3485511778194705, + "grad_norm": 0.08530712872743607, + "learning_rate": 1.959469961369716e-06, + "loss": 0.673, + "step": 418 + }, + { + "epoch": 0.3493850323118616, + "grad_norm": 0.0814647376537323, + "learning_rate": 1.9592164043357046e-06, + "loss": 0.6224, + "step": 419 + }, + { + "epoch": 0.35021888680425267, + "grad_norm": 0.08769946545362473, + "learning_rate": 1.9589620731518164e-06, + "loss": 0.6446, + "step": 420 + }, + { + "epoch": 0.3510527412966437, + "grad_norm": 0.0855315700173378, + "learning_rate": 1.958706968023313e-06, + "loss": 0.6665, + "step": 421 + }, + { + "epoch": 0.3518865957890348, + "grad_norm": 0.08764835447072983, + "learning_rate": 1.958451089156082e-06, + "loss": 0.6888, + "step": 422 + }, + { + "epoch": 0.3527204502814259, + "grad_norm": 0.08674878627061844, + "learning_rate": 1.9581944367566324e-06, + "loss": 0.6894, + "step": 423 + }, + { + "epoch": 0.353554304773817, + "grad_norm": 0.07654520124197006, + "learning_rate": 1.9579370110321005e-06, + "loss": 0.5972, + "step": 424 + }, + { + "epoch": 0.35438815926620804, + "grad_norm": 0.08671235293149948, + "learning_rate": 1.9576788121902454e-06, + "loss": 0.6705, + "step": 425 + }, + { + "epoch": 0.35522201375859913, + "grad_norm": 0.08799296617507935, + "learning_rate": 1.9574198404394505e-06, + "loss": 0.66, + "step": 426 + }, + { + "epoch": 0.3560558682509902, + "grad_norm": 0.08908937871456146, + "learning_rate": 1.9571600959887224e-06, + "loss": 0.6687, + "step": 427 + }, + { + "epoch": 0.35688972274338127, + "grad_norm": 0.08397683501243591, + "learning_rate": 1.9568995790476915e-06, + "loss": 0.5915, + "step": 428 + }, + { + "epoch": 0.35772357723577236, + "grad_norm": 0.08805875480175018, + "learning_rate": 1.9566382898266126e-06, + "loss": 0.6706, + "step": 429 + }, + { + "epoch": 0.35855743172816346, + "grad_norm": 0.08281008154153824, + "learning_rate": 1.956376228536363e-06, + "loss": 0.5692, + "step": 430 + }, + { + "epoch": 0.3593912862205545, + "grad_norm": 0.08514556288719177, + "learning_rate": 1.9561133953884427e-06, + "loss": 0.5938, + "step": 431 + }, + { + "epoch": 0.3602251407129456, + "grad_norm": 0.09112891554832458, + "learning_rate": 1.955849790594975e-06, + "loss": 0.7044, + "step": 432 + }, + { + "epoch": 0.3610589952053367, + "grad_norm": 0.08578234165906906, + "learning_rate": 1.9555854143687064e-06, + "loss": 0.6901, + "step": 433 + }, + { + "epoch": 0.3618928496977277, + "grad_norm": 0.08405105024576187, + "learning_rate": 1.955320266923006e-06, + "loss": 0.6307, + "step": 434 + }, + { + "epoch": 0.3627267041901188, + "grad_norm": 0.085452601313591, + "learning_rate": 1.9550543484718644e-06, + "loss": 0.6147, + "step": 435 + }, + { + "epoch": 0.3635605586825099, + "grad_norm": 0.0871538370847702, + "learning_rate": 1.9547876592298955e-06, + "loss": 0.6415, + "step": 436 + }, + { + "epoch": 0.36439441317490096, + "grad_norm": 0.08855723589658737, + "learning_rate": 1.954520199412334e-06, + "loss": 0.7026, + "step": 437 + }, + { + "epoch": 0.36522826766729205, + "grad_norm": 0.08221881091594696, + "learning_rate": 1.9542519692350387e-06, + "loss": 0.6813, + "step": 438 + }, + { + "epoch": 0.36606212215968315, + "grad_norm": 0.08595414459705353, + "learning_rate": 1.9539829689144876e-06, + "loss": 0.6809, + "step": 439 + }, + { + "epoch": 0.36689597665207424, + "grad_norm": 0.08485814183950424, + "learning_rate": 1.9537131986677816e-06, + "loss": 0.6352, + "step": 440 + }, + { + "epoch": 0.3677298311444653, + "grad_norm": 0.08620929718017578, + "learning_rate": 1.9534426587126433e-06, + "loss": 0.7072, + "step": 441 + }, + { + "epoch": 0.3685636856368564, + "grad_norm": 0.0919066071510315, + "learning_rate": 1.9531713492674154e-06, + "loss": 0.6809, + "step": 442 + }, + { + "epoch": 0.36939754012924747, + "grad_norm": 0.08376951515674591, + "learning_rate": 1.9528992705510628e-06, + "loss": 0.6486, + "step": 443 + }, + { + "epoch": 0.3702313946216385, + "grad_norm": 0.088445745408535, + "learning_rate": 1.95262642278317e-06, + "loss": 0.709, + "step": 444 + }, + { + "epoch": 0.3710652491140296, + "grad_norm": 0.08183503895998001, + "learning_rate": 1.952352806183943e-06, + "loss": 0.6277, + "step": 445 + }, + { + "epoch": 0.3718991036064207, + "grad_norm": 0.08638439327478409, + "learning_rate": 1.9520784209742094e-06, + "loss": 0.6665, + "step": 446 + }, + { + "epoch": 0.37273295809881174, + "grad_norm": 0.08837208896875381, + "learning_rate": 1.951803267375414e-06, + "loss": 0.6103, + "step": 447 + }, + { + "epoch": 0.37356681259120283, + "grad_norm": 0.08010973036289215, + "learning_rate": 1.9515273456096247e-06, + "loss": 0.662, + "step": 448 + }, + { + "epoch": 0.37440066708359393, + "grad_norm": 0.08395659178495407, + "learning_rate": 1.9512506558995283e-06, + "loss": 0.687, + "step": 449 + }, + { + "epoch": 0.37523452157598497, + "grad_norm": 0.08506524562835693, + "learning_rate": 1.950973198468431e-06, + "loss": 0.6695, + "step": 450 + }, + { + "epoch": 0.37606837606837606, + "grad_norm": 0.08150207251310349, + "learning_rate": 1.950694973540259e-06, + "loss": 0.6258, + "step": 451 + }, + { + "epoch": 0.37690223056076716, + "grad_norm": 0.08509568125009537, + "learning_rate": 1.9504159813395576e-06, + "loss": 0.5774, + "step": 452 + }, + { + "epoch": 0.3777360850531582, + "grad_norm": 0.08139500021934509, + "learning_rate": 1.950136222091492e-06, + "loss": 0.6652, + "step": 453 + }, + { + "epoch": 0.3785699395455493, + "grad_norm": 0.08891302347183228, + "learning_rate": 1.949855696021846e-06, + "loss": 0.6602, + "step": 454 + }, + { + "epoch": 0.3794037940379404, + "grad_norm": 0.08729292452335358, + "learning_rate": 1.949574403357022e-06, + "loss": 0.6232, + "step": 455 + }, + { + "epoch": 0.3802376485303315, + "grad_norm": 0.0855722650885582, + "learning_rate": 1.9492923443240425e-06, + "loss": 0.7249, + "step": 456 + }, + { + "epoch": 0.3810715030227225, + "grad_norm": 0.09225167334079742, + "learning_rate": 1.949009519150546e-06, + "loss": 0.7061, + "step": 457 + }, + { + "epoch": 0.3819053575151136, + "grad_norm": 0.08833880722522736, + "learning_rate": 1.9487259280647918e-06, + "loss": 0.6683, + "step": 458 + }, + { + "epoch": 0.3827392120075047, + "grad_norm": 0.08648336678743362, + "learning_rate": 1.948441571295656e-06, + "loss": 0.6716, + "step": 459 + }, + { + "epoch": 0.38357306649989575, + "grad_norm": 0.08678124845027924, + "learning_rate": 1.9481564490726324e-06, + "loss": 0.6869, + "step": 460 + }, + { + "epoch": 0.38440692099228685, + "grad_norm": 0.0866696834564209, + "learning_rate": 1.9478705616258344e-06, + "loss": 0.6346, + "step": 461 + }, + { + "epoch": 0.38524077548467794, + "grad_norm": 0.09009408950805664, + "learning_rate": 1.9475839091859904e-06, + "loss": 0.6507, + "step": 462 + }, + { + "epoch": 0.386074629977069, + "grad_norm": 0.09393912553787231, + "learning_rate": 1.9472964919844484e-06, + "loss": 0.6624, + "step": 463 + }, + { + "epoch": 0.3869084844694601, + "grad_norm": 0.09663153439760208, + "learning_rate": 1.9470083102531722e-06, + "loss": 0.6378, + "step": 464 + }, + { + "epoch": 0.3877423389618512, + "grad_norm": 0.08551555126905441, + "learning_rate": 1.9467193642247435e-06, + "loss": 0.6717, + "step": 465 + }, + { + "epoch": 0.3885761934542422, + "grad_norm": 0.08986209332942963, + "learning_rate": 1.9464296541323602e-06, + "loss": 0.7313, + "step": 466 + }, + { + "epoch": 0.3894100479466333, + "grad_norm": 0.08851300179958344, + "learning_rate": 1.9461391802098373e-06, + "loss": 0.6783, + "step": 467 + }, + { + "epoch": 0.3902439024390244, + "grad_norm": 0.09088604897260666, + "learning_rate": 1.9458479426916068e-06, + "loss": 0.5952, + "step": 468 + }, + { + "epoch": 0.39107775693141544, + "grad_norm": 0.08506208658218384, + "learning_rate": 1.9455559418127144e-06, + "loss": 0.6912, + "step": 469 + }, + { + "epoch": 0.39191161142380654, + "grad_norm": 0.087041936814785, + "learning_rate": 1.945263177808826e-06, + "loss": 0.6722, + "step": 470 + }, + { + "epoch": 0.39274546591619763, + "grad_norm": 0.09510099142789841, + "learning_rate": 1.9449696509162193e-06, + "loss": 0.697, + "step": 471 + }, + { + "epoch": 0.39357932040858873, + "grad_norm": 0.0864386260509491, + "learning_rate": 1.944675361371791e-06, + "loss": 0.6882, + "step": 472 + }, + { + "epoch": 0.39441317490097977, + "grad_norm": 0.08964356780052185, + "learning_rate": 1.944380309413051e-06, + "loss": 0.6403, + "step": 473 + }, + { + "epoch": 0.39524702939337086, + "grad_norm": 0.09786627441644669, + "learning_rate": 1.9440844952781253e-06, + "loss": 0.7037, + "step": 474 + }, + { + "epoch": 0.39608088388576196, + "grad_norm": 0.08709276467561722, + "learning_rate": 1.9437879192057556e-06, + "loss": 0.6619, + "step": 475 + }, + { + "epoch": 0.396914738378153, + "grad_norm": 0.08677015453577042, + "learning_rate": 1.9434905814352976e-06, + "loss": 0.6609, + "step": 476 + }, + { + "epoch": 0.3977485928705441, + "grad_norm": 0.08886487782001495, + "learning_rate": 1.9431924822067225e-06, + "loss": 0.6045, + "step": 477 + }, + { + "epoch": 0.3985824473629352, + "grad_norm": 0.08575446903705597, + "learning_rate": 1.942893621760616e-06, + "loss": 0.6886, + "step": 478 + }, + { + "epoch": 0.39941630185532623, + "grad_norm": 0.08386967331171036, + "learning_rate": 1.9425940003381767e-06, + "loss": 0.6721, + "step": 479 + }, + { + "epoch": 0.4002501563477173, + "grad_norm": 0.08702324330806732, + "learning_rate": 1.9422936181812197e-06, + "loss": 0.6959, + "step": 480 + }, + { + "epoch": 0.4010840108401084, + "grad_norm": 0.09090246260166168, + "learning_rate": 1.9419924755321727e-06, + "loss": 0.7094, + "step": 481 + }, + { + "epoch": 0.40191786533249946, + "grad_norm": 0.09213659167289734, + "learning_rate": 1.9416905726340767e-06, + "loss": 0.6446, + "step": 482 + }, + { + "epoch": 0.40275171982489055, + "grad_norm": 0.09240783751010895, + "learning_rate": 1.9413879097305878e-06, + "loss": 0.6401, + "step": 483 + }, + { + "epoch": 0.40358557431728165, + "grad_norm": 0.09058364480733871, + "learning_rate": 1.941084487065974e-06, + "loss": 0.7035, + "step": 484 + }, + { + "epoch": 0.4044194288096727, + "grad_norm": 0.0906047448515892, + "learning_rate": 1.940780304885117e-06, + "loss": 0.6519, + "step": 485 + }, + { + "epoch": 0.4052532833020638, + "grad_norm": 0.09101078659296036, + "learning_rate": 1.940475363433512e-06, + "loss": 0.6934, + "step": 486 + }, + { + "epoch": 0.4060871377944549, + "grad_norm": 0.08729363977909088, + "learning_rate": 1.940169662957266e-06, + "loss": 0.5822, + "step": 487 + }, + { + "epoch": 0.406920992286846, + "grad_norm": 0.0946895033121109, + "learning_rate": 1.9398632037031003e-06, + "loss": 0.6534, + "step": 488 + }, + { + "epoch": 0.407754846779237, + "grad_norm": 0.09069128334522247, + "learning_rate": 1.9395559859183463e-06, + "loss": 0.6706, + "step": 489 + }, + { + "epoch": 0.4085887012716281, + "grad_norm": 0.08318338543176651, + "learning_rate": 1.9392480098509483e-06, + "loss": 0.6368, + "step": 490 + }, + { + "epoch": 0.4094225557640192, + "grad_norm": 0.0826449766755104, + "learning_rate": 1.9389392757494645e-06, + "loss": 0.667, + "step": 491 + }, + { + "epoch": 0.41025641025641024, + "grad_norm": 0.09276958554983139, + "learning_rate": 1.938629783863062e-06, + "loss": 0.6709, + "step": 492 + }, + { + "epoch": 0.41109026474880134, + "grad_norm": 0.08281219005584717, + "learning_rate": 1.9383195344415215e-06, + "loss": 0.584, + "step": 493 + }, + { + "epoch": 0.41192411924119243, + "grad_norm": 0.08632051944732666, + "learning_rate": 1.938008527735234e-06, + "loss": 0.6743, + "step": 494 + }, + { + "epoch": 0.41275797373358347, + "grad_norm": 0.08970591425895691, + "learning_rate": 1.937696763995202e-06, + "loss": 0.6611, + "step": 495 + }, + { + "epoch": 0.41359182822597457, + "grad_norm": 0.0902903825044632, + "learning_rate": 1.93738424347304e-06, + "loss": 0.6455, + "step": 496 + }, + { + "epoch": 0.41442568271836566, + "grad_norm": 0.08964331448078156, + "learning_rate": 1.9370709664209715e-06, + "loss": 0.6561, + "step": 497 + }, + { + "epoch": 0.4152595372107567, + "grad_norm": 0.08478616923093796, + "learning_rate": 1.9367569330918317e-06, + "loss": 0.595, + "step": 498 + }, + { + "epoch": 0.4160933917031478, + "grad_norm": 0.09221872687339783, + "learning_rate": 1.9364421437390658e-06, + "loss": 0.6363, + "step": 499 + }, + { + "epoch": 0.4169272461955389, + "grad_norm": 0.09467694163322449, + "learning_rate": 1.936126598616729e-06, + "loss": 0.6696, + "step": 500 + }, + { + "epoch": 0.41776110068792993, + "grad_norm": 0.09464087337255478, + "learning_rate": 1.935810297979487e-06, + "loss": 0.6224, + "step": 501 + }, + { + "epoch": 0.418594955180321, + "grad_norm": 0.08994987607002258, + "learning_rate": 1.9354932420826147e-06, + "loss": 0.6383, + "step": 502 + }, + { + "epoch": 0.4194288096727121, + "grad_norm": 0.0921832025051117, + "learning_rate": 1.9351754311819974e-06, + "loss": 0.685, + "step": 503 + }, + { + "epoch": 0.4202626641651032, + "grad_norm": 0.08833765983581543, + "learning_rate": 1.934856865534129e-06, + "loss": 0.6233, + "step": 504 + }, + { + "epoch": 0.42109651865749426, + "grad_norm": 0.08863073587417603, + "learning_rate": 1.934537545396111e-06, + "loss": 0.6756, + "step": 505 + }, + { + "epoch": 0.42193037314988535, + "grad_norm": 0.08703204244375229, + "learning_rate": 1.934217471025658e-06, + "loss": 0.6372, + "step": 506 + }, + { + "epoch": 0.42276422764227645, + "grad_norm": 0.09010464698076248, + "learning_rate": 1.9338966426810887e-06, + "loss": 0.6677, + "step": 507 + }, + { + "epoch": 0.4235980821346675, + "grad_norm": 0.09014829248189926, + "learning_rate": 1.9335750606213336e-06, + "loss": 0.6839, + "step": 508 + }, + { + "epoch": 0.4244319366270586, + "grad_norm": 0.08871164172887802, + "learning_rate": 1.93325272510593e-06, + "loss": 0.7012, + "step": 509 + }, + { + "epoch": 0.4252657911194497, + "grad_norm": 0.08788250386714935, + "learning_rate": 1.9329296363950236e-06, + "loss": 0.6169, + "step": 510 + }, + { + "epoch": 0.4260996456118407, + "grad_norm": 0.08845999836921692, + "learning_rate": 1.9326057947493675e-06, + "loss": 0.6424, + "step": 511 + }, + { + "epoch": 0.4269335001042318, + "grad_norm": 0.0858997106552124, + "learning_rate": 1.9322812004303238e-06, + "loss": 0.6389, + "step": 512 + }, + { + "epoch": 0.4277673545966229, + "grad_norm": 0.08850055187940598, + "learning_rate": 1.9319558536998603e-06, + "loss": 0.6975, + "step": 513 + }, + { + "epoch": 0.42860120908901395, + "grad_norm": 0.09360900521278381, + "learning_rate": 1.9316297548205534e-06, + "loss": 0.686, + "step": 514 + }, + { + "epoch": 0.42943506358140504, + "grad_norm": 0.08473866432905197, + "learning_rate": 1.931302904055586e-06, + "loss": 0.6517, + "step": 515 + }, + { + "epoch": 0.43026891807379614, + "grad_norm": 0.08656331151723862, + "learning_rate": 1.9309753016687477e-06, + "loss": 0.7003, + "step": 516 + }, + { + "epoch": 0.4311027725661872, + "grad_norm": 0.08407936990261078, + "learning_rate": 1.9306469479244347e-06, + "loss": 0.6611, + "step": 517 + }, + { + "epoch": 0.43193662705857827, + "grad_norm": 0.09026416391134262, + "learning_rate": 1.930317843087651e-06, + "loss": 0.6749, + "step": 518 + }, + { + "epoch": 0.43277048155096937, + "grad_norm": 0.10031198710203171, + "learning_rate": 1.929987987424004e-06, + "loss": 0.6403, + "step": 519 + }, + { + "epoch": 0.43360433604336046, + "grad_norm": 0.09739411622285843, + "learning_rate": 1.9296573811997086e-06, + "loss": 0.6893, + "step": 520 + }, + { + "epoch": 0.4344381905357515, + "grad_norm": 0.08792299032211304, + "learning_rate": 1.929326024681587e-06, + "loss": 0.6268, + "step": 521 + }, + { + "epoch": 0.4352720450281426, + "grad_norm": 0.0874318853020668, + "learning_rate": 1.9289939181370646e-06, + "loss": 0.6659, + "step": 522 + }, + { + "epoch": 0.4361058995205337, + "grad_norm": 0.09086845815181732, + "learning_rate": 1.9286610618341724e-06, + "loss": 0.6399, + "step": 523 + }, + { + "epoch": 0.43693975401292473, + "grad_norm": 0.09616074711084366, + "learning_rate": 1.9283274560415477e-06, + "loss": 0.6325, + "step": 524 + }, + { + "epoch": 0.4377736085053158, + "grad_norm": 0.08674857765436172, + "learning_rate": 1.9279931010284322e-06, + "loss": 0.6487, + "step": 525 + }, + { + "epoch": 0.4386074629977069, + "grad_norm": 0.09142550826072693, + "learning_rate": 1.927657997064671e-06, + "loss": 0.6509, + "step": 526 + }, + { + "epoch": 0.43944131749009796, + "grad_norm": 0.0838690921664238, + "learning_rate": 1.9273221444207158e-06, + "loss": 0.6087, + "step": 527 + }, + { + "epoch": 0.44027517198248906, + "grad_norm": 0.0922529399394989, + "learning_rate": 1.9269855433676213e-06, + "loss": 0.6815, + "step": 528 + }, + { + "epoch": 0.44110902647488015, + "grad_norm": 0.08945809304714203, + "learning_rate": 1.926648194177046e-06, + "loss": 0.623, + "step": 529 + }, + { + "epoch": 0.4419428809672712, + "grad_norm": 0.09663840383291245, + "learning_rate": 1.926310097121253e-06, + "loss": 0.6244, + "step": 530 + }, + { + "epoch": 0.4427767354596623, + "grad_norm": 0.0872710794210434, + "learning_rate": 1.9259712524731083e-06, + "loss": 0.6877, + "step": 531 + }, + { + "epoch": 0.4436105899520534, + "grad_norm": 0.0860145092010498, + "learning_rate": 1.9256316605060813e-06, + "loss": 0.6458, + "step": 532 + }, + { + "epoch": 0.4444444444444444, + "grad_norm": 0.08691411465406418, + "learning_rate": 1.9252913214942456e-06, + "loss": 0.6286, + "step": 533 + }, + { + "epoch": 0.4452782989368355, + "grad_norm": 0.0899488553404808, + "learning_rate": 1.9249502357122757e-06, + "loss": 0.6154, + "step": 534 + }, + { + "epoch": 0.4461121534292266, + "grad_norm": 0.08670518547296524, + "learning_rate": 1.924608403435451e-06, + "loss": 0.6189, + "step": 535 + }, + { + "epoch": 0.4469460079216177, + "grad_norm": 0.08927959948778152, + "learning_rate": 1.9242658249396514e-06, + "loss": 0.6591, + "step": 536 + }, + { + "epoch": 0.44777986241400874, + "grad_norm": 0.09092804044485092, + "learning_rate": 1.9239225005013607e-06, + "loss": 0.6756, + "step": 537 + }, + { + "epoch": 0.44861371690639984, + "grad_norm": 0.08876843005418777, + "learning_rate": 1.923578430397664e-06, + "loss": 0.6221, + "step": 538 + }, + { + "epoch": 0.44944757139879093, + "grad_norm": 0.09023214131593704, + "learning_rate": 1.923233614906248e-06, + "loss": 0.6478, + "step": 539 + }, + { + "epoch": 0.450281425891182, + "grad_norm": 0.09291035681962967, + "learning_rate": 1.9228880543054006e-06, + "loss": 0.6568, + "step": 540 + }, + { + "epoch": 0.45111528038357307, + "grad_norm": 0.09169748425483704, + "learning_rate": 1.9225417488740127e-06, + "loss": 0.6966, + "step": 541 + }, + { + "epoch": 0.45194913487596416, + "grad_norm": 0.0960812196135521, + "learning_rate": 1.922194698891574e-06, + "loss": 0.6426, + "step": 542 + }, + { + "epoch": 0.4527829893683552, + "grad_norm": 0.09381034225225449, + "learning_rate": 1.9218469046381778e-06, + "loss": 0.7138, + "step": 543 + }, + { + "epoch": 0.4536168438607463, + "grad_norm": 0.09605712443590164, + "learning_rate": 1.9214983663945157e-06, + "loss": 0.6677, + "step": 544 + }, + { + "epoch": 0.4544506983531374, + "grad_norm": 0.09246627241373062, + "learning_rate": 1.921149084441881e-06, + "loss": 0.5995, + "step": 545 + }, + { + "epoch": 0.45528455284552843, + "grad_norm": 0.09615227580070496, + "learning_rate": 1.9207990590621663e-06, + "loss": 0.6543, + "step": 546 + }, + { + "epoch": 0.45611840733791953, + "grad_norm": 0.09620420634746552, + "learning_rate": 1.9204482905378654e-06, + "loss": 0.5895, + "step": 547 + }, + { + "epoch": 0.4569522618303106, + "grad_norm": 0.08878765255212784, + "learning_rate": 1.920096779152071e-06, + "loss": 0.599, + "step": 548 + }, + { + "epoch": 0.45778611632270166, + "grad_norm": 0.09067462384700775, + "learning_rate": 1.9197445251884763e-06, + "loss": 0.6897, + "step": 549 + }, + { + "epoch": 0.45861997081509276, + "grad_norm": 0.09180627018213272, + "learning_rate": 1.9193915289313724e-06, + "loss": 0.6647, + "step": 550 + }, + { + "epoch": 0.45945382530748385, + "grad_norm": 0.08824802190065384, + "learning_rate": 1.91903779066565e-06, + "loss": 0.6376, + "step": 551 + }, + { + "epoch": 0.46028767979987495, + "grad_norm": 0.08835819363594055, + "learning_rate": 1.918683310676799e-06, + "loss": 0.6922, + "step": 552 + }, + { + "epoch": 0.461121534292266, + "grad_norm": 0.09096652269363403, + "learning_rate": 1.918328089250908e-06, + "loss": 0.7169, + "step": 553 + }, + { + "epoch": 0.4619553887846571, + "grad_norm": 0.09415201842784882, + "learning_rate": 1.917972126674664e-06, + "loss": 0.659, + "step": 554 + }, + { + "epoch": 0.4627892432770482, + "grad_norm": 0.08964891731739044, + "learning_rate": 1.9176154232353513e-06, + "loss": 0.6869, + "step": 555 + }, + { + "epoch": 0.4636230977694392, + "grad_norm": 0.08688930422067642, + "learning_rate": 1.917257979220853e-06, + "loss": 0.627, + "step": 556 + }, + { + "epoch": 0.4644569522618303, + "grad_norm": 0.09115028381347656, + "learning_rate": 1.9168997949196496e-06, + "loss": 0.64, + "step": 557 + }, + { + "epoch": 0.4652908067542214, + "grad_norm": 0.09430664777755737, + "learning_rate": 1.9165408706208184e-06, + "loss": 0.6646, + "step": 558 + }, + { + "epoch": 0.46612466124661245, + "grad_norm": 0.09062688797712326, + "learning_rate": 1.916181206614036e-06, + "loss": 0.6392, + "step": 559 + }, + { + "epoch": 0.46695851573900354, + "grad_norm": 0.09099205583333969, + "learning_rate": 1.9158208031895737e-06, + "loss": 0.7133, + "step": 560 + }, + { + "epoch": 0.46779237023139464, + "grad_norm": 0.09944413602352142, + "learning_rate": 1.9154596606383002e-06, + "loss": 0.6958, + "step": 561 + }, + { + "epoch": 0.4686262247237857, + "grad_norm": 0.09643740206956863, + "learning_rate": 1.9150977792516816e-06, + "loss": 0.6813, + "step": 562 + }, + { + "epoch": 0.4694600792161768, + "grad_norm": 0.09532240033149719, + "learning_rate": 1.914735159321779e-06, + "loss": 0.6528, + "step": 563 + }, + { + "epoch": 0.47029393370856787, + "grad_norm": 0.08927876502275467, + "learning_rate": 1.914371801141251e-06, + "loss": 0.639, + "step": 564 + }, + { + "epoch": 0.4711277882009589, + "grad_norm": 0.09561195224523544, + "learning_rate": 1.914007705003351e-06, + "loss": 0.6795, + "step": 565 + }, + { + "epoch": 0.47196164269335, + "grad_norm": 0.0882195308804512, + "learning_rate": 1.9136428712019275e-06, + "loss": 0.6437, + "step": 566 + }, + { + "epoch": 0.4727954971857411, + "grad_norm": 0.08940508961677551, + "learning_rate": 1.913277300031426e-06, + "loss": 0.6485, + "step": 567 + }, + { + "epoch": 0.4736293516781322, + "grad_norm": 0.0933694839477539, + "learning_rate": 1.912910991786886e-06, + "loss": 0.6876, + "step": 568 + }, + { + "epoch": 0.47446320617052323, + "grad_norm": 0.09212841838598251, + "learning_rate": 1.9125439467639414e-06, + "loss": 0.5669, + "step": 569 + }, + { + "epoch": 0.47529706066291433, + "grad_norm": 0.09037458896636963, + "learning_rate": 1.9121761652588214e-06, + "loss": 0.6489, + "step": 570 + }, + { + "epoch": 0.4761309151553054, + "grad_norm": 0.08847213536500931, + "learning_rate": 1.9118076475683506e-06, + "loss": 0.6456, + "step": 571 + }, + { + "epoch": 0.47696476964769646, + "grad_norm": 0.08496098965406418, + "learning_rate": 1.9114383939899455e-06, + "loss": 0.6174, + "step": 572 + }, + { + "epoch": 0.47779862414008756, + "grad_norm": 0.09430284053087234, + "learning_rate": 1.911068404821618e-06, + "loss": 0.7102, + "step": 573 + }, + { + "epoch": 0.47863247863247865, + "grad_norm": 0.09211906790733337, + "learning_rate": 1.910697680361974e-06, + "loss": 0.6161, + "step": 574 + }, + { + "epoch": 0.4794663331248697, + "grad_norm": 0.09092090278863907, + "learning_rate": 1.910326220910211e-06, + "loss": 0.6595, + "step": 575 + }, + { + "epoch": 0.4803001876172608, + "grad_norm": 0.09773701429367065, + "learning_rate": 1.909954026766122e-06, + "loss": 0.6303, + "step": 576 + }, + { + "epoch": 0.4811340421096519, + "grad_norm": 0.09343478083610535, + "learning_rate": 1.909581098230091e-06, + "loss": 0.6578, + "step": 577 + }, + { + "epoch": 0.4819678966020429, + "grad_norm": 0.09686136245727539, + "learning_rate": 1.9092074356030966e-06, + "loss": 0.6504, + "step": 578 + }, + { + "epoch": 0.482801751094434, + "grad_norm": 0.09207551926374435, + "learning_rate": 1.9088330391867076e-06, + "loss": 0.6092, + "step": 579 + }, + { + "epoch": 0.4836356055868251, + "grad_norm": 0.093394935131073, + "learning_rate": 1.908457909283087e-06, + "loss": 0.645, + "step": 580 + }, + { + "epoch": 0.48446946007921615, + "grad_norm": 0.10280878096818924, + "learning_rate": 1.9080820461949886e-06, + "loss": 0.7288, + "step": 581 + }, + { + "epoch": 0.48530331457160725, + "grad_norm": 0.09714383631944656, + "learning_rate": 1.9077054502257585e-06, + "loss": 0.6213, + "step": 582 + }, + { + "epoch": 0.48613716906399834, + "grad_norm": 0.09004565328359604, + "learning_rate": 1.9073281216793337e-06, + "loss": 0.6459, + "step": 583 + }, + { + "epoch": 0.48697102355638944, + "grad_norm": 0.12440581619739532, + "learning_rate": 1.906950060860243e-06, + "loss": 0.6823, + "step": 584 + }, + { + "epoch": 0.4878048780487805, + "grad_norm": 0.09314379096031189, + "learning_rate": 1.9065712680736066e-06, + "loss": 0.6475, + "step": 585 + }, + { + "epoch": 0.48863873254117157, + "grad_norm": 0.09512604027986526, + "learning_rate": 1.906191743625134e-06, + "loss": 0.5932, + "step": 586 + }, + { + "epoch": 0.48947258703356267, + "grad_norm": 0.08977750688791275, + "learning_rate": 1.9058114878211265e-06, + "loss": 0.6546, + "step": 587 + }, + { + "epoch": 0.4903064415259537, + "grad_norm": 0.09166161715984344, + "learning_rate": 1.905430500968475e-06, + "loss": 0.6772, + "step": 588 + }, + { + "epoch": 0.4911402960183448, + "grad_norm": 0.10014794021844864, + "learning_rate": 1.9050487833746609e-06, + "loss": 0.6945, + "step": 589 + }, + { + "epoch": 0.4919741505107359, + "grad_norm": 0.09376100450754166, + "learning_rate": 1.9046663353477548e-06, + "loss": 0.6581, + "step": 590 + }, + { + "epoch": 0.49280800500312694, + "grad_norm": 0.09505487233400345, + "learning_rate": 1.9042831571964171e-06, + "loss": 0.6785, + "step": 591 + }, + { + "epoch": 0.49364185949551803, + "grad_norm": 0.09371201694011688, + "learning_rate": 1.9038992492298975e-06, + "loss": 0.6331, + "step": 592 + }, + { + "epoch": 0.4944757139879091, + "grad_norm": 0.09442009776830673, + "learning_rate": 1.9035146117580348e-06, + "loss": 0.5728, + "step": 593 + }, + { + "epoch": 0.49530956848030017, + "grad_norm": 0.09772226214408875, + "learning_rate": 1.903129245091256e-06, + "loss": 0.6635, + "step": 594 + }, + { + "epoch": 0.49614342297269126, + "grad_norm": 0.09484317898750305, + "learning_rate": 1.9027431495405776e-06, + "loss": 0.6327, + "step": 595 + }, + { + "epoch": 0.49697727746508236, + "grad_norm": 0.09658980369567871, + "learning_rate": 1.9023563254176032e-06, + "loss": 0.6608, + "step": 596 + }, + { + "epoch": 0.4978111319574734, + "grad_norm": 0.09717069566249847, + "learning_rate": 1.9019687730345251e-06, + "loss": 0.6605, + "step": 597 + }, + { + "epoch": 0.4986449864498645, + "grad_norm": 0.09019803255796432, + "learning_rate": 1.9015804927041233e-06, + "loss": 0.6867, + "step": 598 + }, + { + "epoch": 0.4994788409422556, + "grad_norm": 0.09377589821815491, + "learning_rate": 1.9011914847397653e-06, + "loss": 0.6419, + "step": 599 + }, + { + "epoch": 0.5003126954346466, + "grad_norm": 0.09151824563741684, + "learning_rate": 1.9008017494554055e-06, + "loss": 0.7184, + "step": 600 + }, + { + "epoch": 0.5011465499270378, + "grad_norm": 0.0963352620601654, + "learning_rate": 1.900411287165586e-06, + "loss": 0.6897, + "step": 601 + }, + { + "epoch": 0.5019804044194288, + "grad_norm": 0.08974531292915344, + "learning_rate": 1.9000200981854347e-06, + "loss": 0.5847, + "step": 602 + }, + { + "epoch": 0.5028142589118199, + "grad_norm": 0.09220883995294571, + "learning_rate": 1.8996281828306665e-06, + "loss": 0.6652, + "step": 603 + }, + { + "epoch": 0.503648113404211, + "grad_norm": 0.09491855651140213, + "learning_rate": 1.899235541417583e-06, + "loss": 0.595, + "step": 604 + }, + { + "epoch": 0.504481967896602, + "grad_norm": 0.09616294503211975, + "learning_rate": 1.898842174263071e-06, + "loss": 0.6378, + "step": 605 + }, + { + "epoch": 0.5053158223889931, + "grad_norm": 0.08942185342311859, + "learning_rate": 1.8984480816846035e-06, + "loss": 0.6459, + "step": 606 + }, + { + "epoch": 0.5061496768813842, + "grad_norm": 0.09059803187847137, + "learning_rate": 1.8980532640002388e-06, + "loss": 0.6019, + "step": 607 + }, + { + "epoch": 0.5069835313737753, + "grad_norm": 0.09913185238838196, + "learning_rate": 1.89765772152862e-06, + "loss": 0.669, + "step": 608 + }, + { + "epoch": 0.5078173858661663, + "grad_norm": 0.10223888605833054, + "learning_rate": 1.8972614545889756e-06, + "loss": 0.6454, + "step": 609 + }, + { + "epoch": 0.5086512403585575, + "grad_norm": 0.09271295368671417, + "learning_rate": 1.896864463501119e-06, + "loss": 0.6467, + "step": 610 + }, + { + "epoch": 0.5094850948509485, + "grad_norm": 0.08862827718257904, + "learning_rate": 1.896466748585448e-06, + "loss": 0.6343, + "step": 611 + }, + { + "epoch": 0.5103189493433395, + "grad_norm": 0.09303930401802063, + "learning_rate": 1.8960683101629435e-06, + "loss": 0.6389, + "step": 612 + }, + { + "epoch": 0.5111528038357307, + "grad_norm": 0.09331949055194855, + "learning_rate": 1.8956691485551719e-06, + "loss": 0.6507, + "step": 613 + }, + { + "epoch": 0.5119866583281217, + "grad_norm": 0.0993635281920433, + "learning_rate": 1.8952692640842825e-06, + "loss": 0.63, + "step": 614 + }, + { + "epoch": 0.5128205128205128, + "grad_norm": 0.09348301589488983, + "learning_rate": 1.8948686570730074e-06, + "loss": 0.676, + "step": 615 + }, + { + "epoch": 0.5136543673129039, + "grad_norm": 0.0950765460729599, + "learning_rate": 1.894467327844663e-06, + "loss": 0.7035, + "step": 616 + }, + { + "epoch": 0.514488221805295, + "grad_norm": 0.08676007390022278, + "learning_rate": 1.8940652767231483e-06, + "loss": 0.6067, + "step": 617 + }, + { + "epoch": 0.515322076297686, + "grad_norm": 0.09285329282283783, + "learning_rate": 1.8936625040329438e-06, + "loss": 0.6441, + "step": 618 + }, + { + "epoch": 0.5161559307900772, + "grad_norm": 0.09361077845096588, + "learning_rate": 1.8932590100991136e-06, + "loss": 0.6767, + "step": 619 + }, + { + "epoch": 0.5169897852824682, + "grad_norm": 0.09173277020454407, + "learning_rate": 1.8928547952473033e-06, + "loss": 0.6395, + "step": 620 + }, + { + "epoch": 0.5178236397748592, + "grad_norm": 0.09198316186666489, + "learning_rate": 1.892449859803741e-06, + "loss": 0.7006, + "step": 621 + }, + { + "epoch": 0.5186574942672504, + "grad_norm": 0.09348779171705246, + "learning_rate": 1.8920442040952356e-06, + "loss": 0.6641, + "step": 622 + }, + { + "epoch": 0.5194913487596414, + "grad_norm": 0.08798382431268692, + "learning_rate": 1.8916378284491774e-06, + "loss": 0.6478, + "step": 623 + }, + { + "epoch": 0.5203252032520326, + "grad_norm": 0.09442038089036942, + "learning_rate": 1.8912307331935383e-06, + "loss": 0.6201, + "step": 624 + }, + { + "epoch": 0.5211590577444236, + "grad_norm": 0.09519796818494797, + "learning_rate": 1.8908229186568705e-06, + "loss": 0.5917, + "step": 625 + }, + { + "epoch": 0.5219929122368147, + "grad_norm": 0.09593921154737473, + "learning_rate": 1.8904143851683062e-06, + "loss": 0.6859, + "step": 626 + }, + { + "epoch": 0.5228267667292058, + "grad_norm": 0.09510339796543121, + "learning_rate": 1.890005133057559e-06, + "loss": 0.6679, + "step": 627 + }, + { + "epoch": 0.5236606212215968, + "grad_norm": 0.09473798424005508, + "learning_rate": 1.8895951626549222e-06, + "loss": 0.6108, + "step": 628 + }, + { + "epoch": 0.5244944757139879, + "grad_norm": 0.10544616729021072, + "learning_rate": 1.8891844742912678e-06, + "loss": 0.6007, + "step": 629 + }, + { + "epoch": 0.525328330206379, + "grad_norm": 0.09652596712112427, + "learning_rate": 1.8887730682980482e-06, + "loss": 0.6858, + "step": 630 + }, + { + "epoch": 0.5261621846987701, + "grad_norm": 0.09385030716657639, + "learning_rate": 1.888360945007295e-06, + "loss": 0.692, + "step": 631 + }, + { + "epoch": 0.5269960391911611, + "grad_norm": 0.10086268931627274, + "learning_rate": 1.887948104751618e-06, + "loss": 0.7302, + "step": 632 + }, + { + "epoch": 0.5278298936835523, + "grad_norm": 0.09164178371429443, + "learning_rate": 1.8875345478642065e-06, + "loss": 0.5793, + "step": 633 + }, + { + "epoch": 0.5286637481759433, + "grad_norm": 0.08686941117048264, + "learning_rate": 1.8871202746788274e-06, + "loss": 0.6304, + "step": 634 + }, + { + "epoch": 0.5294976026683343, + "grad_norm": 0.09579890966415405, + "learning_rate": 1.8867052855298264e-06, + "loss": 0.6663, + "step": 635 + }, + { + "epoch": 0.5303314571607255, + "grad_norm": 0.09054259210824966, + "learning_rate": 1.8862895807521264e-06, + "loss": 0.6395, + "step": 636 + }, + { + "epoch": 0.5311653116531165, + "grad_norm": 0.0913679376244545, + "learning_rate": 1.8858731606812284e-06, + "loss": 0.6481, + "step": 637 + }, + { + "epoch": 0.5319991661455076, + "grad_norm": 0.09573955833911896, + "learning_rate": 1.8854560256532098e-06, + "loss": 0.6543, + "step": 638 + }, + { + "epoch": 0.5328330206378987, + "grad_norm": 0.09667269140481949, + "learning_rate": 1.8850381760047262e-06, + "loss": 0.6532, + "step": 639 + }, + { + "epoch": 0.5336668751302898, + "grad_norm": 0.09536033123731613, + "learning_rate": 1.8846196120730093e-06, + "loss": 0.6429, + "step": 640 + }, + { + "epoch": 0.5345007296226808, + "grad_norm": 0.09058564901351929, + "learning_rate": 1.8842003341958673e-06, + "loss": 0.579, + "step": 641 + }, + { + "epoch": 0.535334584115072, + "grad_norm": 0.09930071234703064, + "learning_rate": 1.8837803427116844e-06, + "loss": 0.5868, + "step": 642 + }, + { + "epoch": 0.536168438607463, + "grad_norm": 0.0978417843580246, + "learning_rate": 1.8833596379594211e-06, + "loss": 0.6822, + "step": 643 + }, + { + "epoch": 0.537002293099854, + "grad_norm": 0.09346287697553635, + "learning_rate": 1.882938220278614e-06, + "loss": 0.6116, + "step": 644 + }, + { + "epoch": 0.5378361475922452, + "grad_norm": 0.09151905030012131, + "learning_rate": 1.8825160900093738e-06, + "loss": 0.6496, + "step": 645 + }, + { + "epoch": 0.5386700020846362, + "grad_norm": 0.0970907062292099, + "learning_rate": 1.8820932474923871e-06, + "loss": 0.6504, + "step": 646 + }, + { + "epoch": 0.5395038565770273, + "grad_norm": 0.09408881515264511, + "learning_rate": 1.8816696930689159e-06, + "loss": 0.6829, + "step": 647 + }, + { + "epoch": 0.5403377110694184, + "grad_norm": 0.09966862946748734, + "learning_rate": 1.8812454270807951e-06, + "loss": 0.5898, + "step": 648 + }, + { + "epoch": 0.5411715655618095, + "grad_norm": 0.0900130420923233, + "learning_rate": 1.8808204498704356e-06, + "loss": 0.6042, + "step": 649 + }, + { + "epoch": 0.5420054200542005, + "grad_norm": 0.09694640338420868, + "learning_rate": 1.8803947617808214e-06, + "loss": 0.6727, + "step": 650 + }, + { + "epoch": 0.5428392745465916, + "grad_norm": 0.09688873589038849, + "learning_rate": 1.8799683631555103e-06, + "loss": 0.6137, + "step": 651 + }, + { + "epoch": 0.5436731290389827, + "grad_norm": 0.09676310420036316, + "learning_rate": 1.8795412543386337e-06, + "loss": 0.612, + "step": 652 + }, + { + "epoch": 0.5445069835313737, + "grad_norm": 0.09581848233938217, + "learning_rate": 1.8791134356748962e-06, + "loss": 0.6632, + "step": 653 + }, + { + "epoch": 0.5453408380237649, + "grad_norm": 0.099350206553936, + "learning_rate": 1.8786849075095753e-06, + "loss": 0.6652, + "step": 654 + }, + { + "epoch": 0.5461746925161559, + "grad_norm": 0.09312117099761963, + "learning_rate": 1.878255670188521e-06, + "loss": 0.6184, + "step": 655 + }, + { + "epoch": 0.5470085470085471, + "grad_norm": 0.09312383830547333, + "learning_rate": 1.8778257240581554e-06, + "loss": 0.6595, + "step": 656 + }, + { + "epoch": 0.5478424015009381, + "grad_norm": 0.09485237300395966, + "learning_rate": 1.8773950694654732e-06, + "loss": 0.716, + "step": 657 + }, + { + "epoch": 0.5486762559933291, + "grad_norm": 0.09585438668727875, + "learning_rate": 1.8769637067580405e-06, + "loss": 0.7058, + "step": 658 + }, + { + "epoch": 0.5495101104857203, + "grad_norm": 0.09024535119533539, + "learning_rate": 1.876531636283995e-06, + "loss": 0.594, + "step": 659 + }, + { + "epoch": 0.5503439649781113, + "grad_norm": 0.09285992383956909, + "learning_rate": 1.8760988583920456e-06, + "loss": 0.648, + "step": 660 + }, + { + "epoch": 0.5511778194705024, + "grad_norm": 0.08939266204833984, + "learning_rate": 1.8756653734314722e-06, + "loss": 0.5866, + "step": 661 + }, + { + "epoch": 0.5520116739628935, + "grad_norm": 0.10212475061416626, + "learning_rate": 1.8752311817521254e-06, + "loss": 0.5485, + "step": 662 + }, + { + "epoch": 0.5528455284552846, + "grad_norm": 0.0938698872923851, + "learning_rate": 1.8747962837044256e-06, + "loss": 0.692, + "step": 663 + }, + { + "epoch": 0.5536793829476756, + "grad_norm": 0.09794861823320389, + "learning_rate": 1.8743606796393644e-06, + "loss": 0.6449, + "step": 664 + }, + { + "epoch": 0.5545132374400668, + "grad_norm": 0.09574401378631592, + "learning_rate": 1.8739243699085019e-06, + "loss": 0.5878, + "step": 665 + }, + { + "epoch": 0.5553470919324578, + "grad_norm": 0.10169863700866699, + "learning_rate": 1.8734873548639687e-06, + "loss": 0.6968, + "step": 666 + }, + { + "epoch": 0.5561809464248488, + "grad_norm": 0.09560302644968033, + "learning_rate": 1.8730496348584642e-06, + "loss": 0.6448, + "step": 667 + }, + { + "epoch": 0.55701480091724, + "grad_norm": 0.10035623610019684, + "learning_rate": 1.8726112102452572e-06, + "loss": 0.6893, + "step": 668 + }, + { + "epoch": 0.557848655409631, + "grad_norm": 0.093927301466465, + "learning_rate": 1.872172081378184e-06, + "loss": 0.6541, + "step": 669 + }, + { + "epoch": 0.5586825099020221, + "grad_norm": 0.0979999378323555, + "learning_rate": 1.8717322486116508e-06, + "loss": 0.6651, + "step": 670 + }, + { + "epoch": 0.5595163643944132, + "grad_norm": 0.10020069032907486, + "learning_rate": 1.8712917123006312e-06, + "loss": 0.6289, + "step": 671 + }, + { + "epoch": 0.5603502188868043, + "grad_norm": 0.10121606290340424, + "learning_rate": 1.8708504728006664e-06, + "loss": 0.6547, + "step": 672 + }, + { + "epoch": 0.5611840733791953, + "grad_norm": 0.09832029789686203, + "learning_rate": 1.870408530467865e-06, + "loss": 0.651, + "step": 673 + }, + { + "epoch": 0.5620179278715864, + "grad_norm": 0.0978146642446518, + "learning_rate": 1.8699658856589037e-06, + "loss": 0.6453, + "step": 674 + }, + { + "epoch": 0.5628517823639775, + "grad_norm": 0.10008279979228973, + "learning_rate": 1.8695225387310249e-06, + "loss": 0.6945, + "step": 675 + }, + { + "epoch": 0.5636856368563685, + "grad_norm": 0.09760496020317078, + "learning_rate": 1.8690784900420388e-06, + "loss": 0.6222, + "step": 676 + }, + { + "epoch": 0.5645194913487597, + "grad_norm": 0.09725689142942429, + "learning_rate": 1.8686337399503215e-06, + "loss": 0.6464, + "step": 677 + }, + { + "epoch": 0.5653533458411507, + "grad_norm": 0.09347565472126007, + "learning_rate": 1.8681882888148152e-06, + "loss": 0.6282, + "step": 678 + }, + { + "epoch": 0.5661872003335418, + "grad_norm": 0.09704628586769104, + "learning_rate": 1.867742136995028e-06, + "loss": 0.6529, + "step": 679 + }, + { + "epoch": 0.5670210548259329, + "grad_norm": 0.09918136149644852, + "learning_rate": 1.8672952848510328e-06, + "loss": 0.739, + "step": 680 + }, + { + "epoch": 0.5678549093183239, + "grad_norm": 0.09772541373968124, + "learning_rate": 1.8668477327434686e-06, + "loss": 0.6383, + "step": 681 + }, + { + "epoch": 0.568688763810715, + "grad_norm": 0.09866555780172348, + "learning_rate": 1.866399481033539e-06, + "loss": 0.6721, + "step": 682 + }, + { + "epoch": 0.5695226183031061, + "grad_norm": 0.09793423861265182, + "learning_rate": 1.8659505300830123e-06, + "loss": 0.6055, + "step": 683 + }, + { + "epoch": 0.5703564727954972, + "grad_norm": 0.10061061382293701, + "learning_rate": 1.865500880254221e-06, + "loss": 0.6544, + "step": 684 + }, + { + "epoch": 0.5711903272878882, + "grad_norm": 0.09208130091428757, + "learning_rate": 1.8650505319100617e-06, + "loss": 0.6014, + "step": 685 + }, + { + "epoch": 0.5720241817802794, + "grad_norm": 0.0995427742600441, + "learning_rate": 1.8645994854139948e-06, + "loss": 0.6965, + "step": 686 + }, + { + "epoch": 0.5728580362726704, + "grad_norm": 0.09244947880506516, + "learning_rate": 1.8641477411300438e-06, + "loss": 0.6272, + "step": 687 + }, + { + "epoch": 0.5736918907650616, + "grad_norm": 0.09937774389982224, + "learning_rate": 1.863695299422796e-06, + "loss": 0.6648, + "step": 688 + }, + { + "epoch": 0.5745257452574526, + "grad_norm": 0.09111293405294418, + "learning_rate": 1.8632421606574007e-06, + "loss": 0.5476, + "step": 689 + }, + { + "epoch": 0.5753595997498436, + "grad_norm": 0.10068730264902115, + "learning_rate": 1.8627883251995708e-06, + "loss": 0.6535, + "step": 690 + }, + { + "epoch": 0.5761934542422348, + "grad_norm": 0.09240923076868057, + "learning_rate": 1.862333793415581e-06, + "loss": 0.575, + "step": 691 + }, + { + "epoch": 0.5770273087346258, + "grad_norm": 0.09765233844518661, + "learning_rate": 1.8618785656722676e-06, + "loss": 0.6323, + "step": 692 + }, + { + "epoch": 0.5778611632270169, + "grad_norm": 0.09996625781059265, + "learning_rate": 1.861422642337029e-06, + "loss": 0.639, + "step": 693 + }, + { + "epoch": 0.578695017719408, + "grad_norm": 0.10148533433675766, + "learning_rate": 1.860966023777825e-06, + "loss": 0.7196, + "step": 694 + }, + { + "epoch": 0.579528872211799, + "grad_norm": 0.09370430558919907, + "learning_rate": 1.8605087103631764e-06, + "loss": 0.6385, + "step": 695 + }, + { + "epoch": 0.5803627267041901, + "grad_norm": 0.09228570014238358, + "learning_rate": 1.8600507024621648e-06, + "loss": 0.6242, + "step": 696 + }, + { + "epoch": 0.5811965811965812, + "grad_norm": 0.09689746052026749, + "learning_rate": 1.8595920004444324e-06, + "loss": 0.6135, + "step": 697 + }, + { + "epoch": 0.5820304356889723, + "grad_norm": 0.09802035987377167, + "learning_rate": 1.8591326046801812e-06, + "loss": 0.5789, + "step": 698 + }, + { + "epoch": 0.5828642901813633, + "grad_norm": 0.10172217339277267, + "learning_rate": 1.8586725155401735e-06, + "loss": 0.5921, + "step": 699 + }, + { + "epoch": 0.5836981446737545, + "grad_norm": 0.10300835222005844, + "learning_rate": 1.858211733395731e-06, + "loss": 0.6647, + "step": 700 + }, + { + "epoch": 0.5845319991661455, + "grad_norm": 0.09017598628997803, + "learning_rate": 1.8577502586187353e-06, + "loss": 0.6409, + "step": 701 + }, + { + "epoch": 0.5853658536585366, + "grad_norm": 0.10058888792991638, + "learning_rate": 1.8572880915816259e-06, + "loss": 0.5902, + "step": 702 + }, + { + "epoch": 0.5861997081509277, + "grad_norm": 0.10125189274549484, + "learning_rate": 1.856825232657402e-06, + "loss": 0.6392, + "step": 703 + }, + { + "epoch": 0.5870335626433187, + "grad_norm": 0.0983993262052536, + "learning_rate": 1.8563616822196204e-06, + "loss": 0.5915, + "step": 704 + }, + { + "epoch": 0.5878674171357098, + "grad_norm": 0.09559126943349838, + "learning_rate": 1.8558974406423965e-06, + "loss": 0.6386, + "step": 705 + }, + { + "epoch": 0.5887012716281009, + "grad_norm": 0.09088745713233948, + "learning_rate": 1.8554325083004034e-06, + "loss": 0.6483, + "step": 706 + }, + { + "epoch": 0.589535126120492, + "grad_norm": 0.09516370296478271, + "learning_rate": 1.854966885568872e-06, + "loss": 0.6556, + "step": 707 + }, + { + "epoch": 0.590368980612883, + "grad_norm": 0.09569456428289413, + "learning_rate": 1.8545005728235896e-06, + "loss": 0.6394, + "step": 708 + }, + { + "epoch": 0.5912028351052742, + "grad_norm": 0.0980261042714119, + "learning_rate": 1.854033570440901e-06, + "loss": 0.6415, + "step": 709 + }, + { + "epoch": 0.5920366895976652, + "grad_norm": 0.09796703606843948, + "learning_rate": 1.8535658787977075e-06, + "loss": 0.6736, + "step": 710 + }, + { + "epoch": 0.5928705440900562, + "grad_norm": 0.09802737832069397, + "learning_rate": 1.8530974982714664e-06, + "loss": 0.5802, + "step": 711 + }, + { + "epoch": 0.5937043985824474, + "grad_norm": 0.09888147562742233, + "learning_rate": 1.8526284292401914e-06, + "loss": 0.6648, + "step": 712 + }, + { + "epoch": 0.5945382530748384, + "grad_norm": 0.10190586000680923, + "learning_rate": 1.8521586720824515e-06, + "loss": 0.7032, + "step": 713 + }, + { + "epoch": 0.5953721075672295, + "grad_norm": 0.09422854334115982, + "learning_rate": 1.8516882271773712e-06, + "loss": 0.6551, + "step": 714 + }, + { + "epoch": 0.5962059620596206, + "grad_norm": 0.09025522321462631, + "learning_rate": 1.8512170949046305e-06, + "loss": 0.6231, + "step": 715 + }, + { + "epoch": 0.5970398165520117, + "grad_norm": 0.102678582072258, + "learning_rate": 1.850745275644463e-06, + "loss": 0.6924, + "step": 716 + }, + { + "epoch": 0.5978736710444027, + "grad_norm": 0.09951562434434891, + "learning_rate": 1.8502727697776579e-06, + "loss": 0.5763, + "step": 717 + }, + { + "epoch": 0.5987075255367938, + "grad_norm": 0.10048934072256088, + "learning_rate": 1.8497995776855579e-06, + "loss": 0.6369, + "step": 718 + }, + { + "epoch": 0.5995413800291849, + "grad_norm": 0.11139318346977234, + "learning_rate": 1.84932569975006e-06, + "loss": 0.6086, + "step": 719 + }, + { + "epoch": 0.600375234521576, + "grad_norm": 0.10424143821001053, + "learning_rate": 1.8488511363536138e-06, + "loss": 0.6394, + "step": 720 + }, + { + "epoch": 0.6012090890139671, + "grad_norm": 0.09213969111442566, + "learning_rate": 1.8483758878792236e-06, + "loss": 0.6554, + "step": 721 + }, + { + "epoch": 0.6020429435063581, + "grad_norm": 0.0998111367225647, + "learning_rate": 1.8478999547104452e-06, + "loss": 0.6717, + "step": 722 + }, + { + "epoch": 0.6028767979987493, + "grad_norm": 0.09211792796850204, + "learning_rate": 1.8474233372313877e-06, + "loss": 0.5694, + "step": 723 + }, + { + "epoch": 0.6037106524911403, + "grad_norm": 0.10054522007703781, + "learning_rate": 1.8469460358267127e-06, + "loss": 0.6501, + "step": 724 + }, + { + "epoch": 0.6045445069835313, + "grad_norm": 0.09979543089866638, + "learning_rate": 1.8464680508816323e-06, + "loss": 0.6056, + "step": 725 + }, + { + "epoch": 0.6053783614759225, + "grad_norm": 0.1000729575753212, + "learning_rate": 1.8459893827819126e-06, + "loss": 0.6302, + "step": 726 + }, + { + "epoch": 0.6062122159683135, + "grad_norm": 0.09924089163541794, + "learning_rate": 1.8455100319138692e-06, + "loss": 0.6432, + "step": 727 + }, + { + "epoch": 0.6070460704607046, + "grad_norm": 0.0920461118221283, + "learning_rate": 1.8450299986643695e-06, + "loss": 0.6081, + "step": 728 + }, + { + "epoch": 0.6078799249530957, + "grad_norm": 0.10430511087179184, + "learning_rate": 1.8445492834208316e-06, + "loss": 0.6586, + "step": 729 + }, + { + "epoch": 0.6087137794454868, + "grad_norm": 0.09923563152551651, + "learning_rate": 1.8440678865712236e-06, + "loss": 0.6166, + "step": 730 + }, + { + "epoch": 0.6095476339378778, + "grad_norm": 0.1012277603149414, + "learning_rate": 1.843585808504064e-06, + "loss": 0.6843, + "step": 731 + }, + { + "epoch": 0.610381488430269, + "grad_norm": 0.09866297990083694, + "learning_rate": 1.8431030496084214e-06, + "loss": 0.6556, + "step": 732 + }, + { + "epoch": 0.61121534292266, + "grad_norm": 0.09874875843524933, + "learning_rate": 1.8426196102739132e-06, + "loss": 0.5993, + "step": 733 + }, + { + "epoch": 0.612049197415051, + "grad_norm": 0.09275837987661362, + "learning_rate": 1.8421354908907067e-06, + "loss": 0.6326, + "step": 734 + }, + { + "epoch": 0.6128830519074422, + "grad_norm": 0.09502032399177551, + "learning_rate": 1.8416506918495173e-06, + "loss": 0.6077, + "step": 735 + }, + { + "epoch": 0.6137169063998332, + "grad_norm": 0.09185368567705154, + "learning_rate": 1.8411652135416093e-06, + "loss": 0.6064, + "step": 736 + }, + { + "epoch": 0.6145507608922243, + "grad_norm": 0.09815992414951324, + "learning_rate": 1.8406790563587956e-06, + "loss": 0.6313, + "step": 737 + }, + { + "epoch": 0.6153846153846154, + "grad_norm": 0.09629546850919724, + "learning_rate": 1.8401922206934361e-06, + "loss": 0.6106, + "step": 738 + }, + { + "epoch": 0.6162184698770065, + "grad_norm": 0.09635171294212341, + "learning_rate": 1.839704706938439e-06, + "loss": 0.6523, + "step": 739 + }, + { + "epoch": 0.6170523243693975, + "grad_norm": 0.10179682075977325, + "learning_rate": 1.8392165154872595e-06, + "loss": 0.5873, + "step": 740 + }, + { + "epoch": 0.6178861788617886, + "grad_norm": 0.10270754247903824, + "learning_rate": 1.8387276467338996e-06, + "loss": 0.673, + "step": 741 + }, + { + "epoch": 0.6187200333541797, + "grad_norm": 0.09746810793876648, + "learning_rate": 1.8382381010729086e-06, + "loss": 0.6564, + "step": 742 + }, + { + "epoch": 0.6195538878465707, + "grad_norm": 0.0962262824177742, + "learning_rate": 1.8377478788993813e-06, + "loss": 0.5738, + "step": 743 + }, + { + "epoch": 0.6203877423389619, + "grad_norm": 0.09518938511610031, + "learning_rate": 1.8372569806089586e-06, + "loss": 0.6345, + "step": 744 + }, + { + "epoch": 0.6212215968313529, + "grad_norm": 0.09188467264175415, + "learning_rate": 1.8367654065978276e-06, + "loss": 0.6112, + "step": 745 + }, + { + "epoch": 0.622055451323744, + "grad_norm": 0.09746599197387695, + "learning_rate": 1.8362731572627202e-06, + "loss": 0.6349, + "step": 746 + }, + { + "epoch": 0.6228893058161351, + "grad_norm": 0.10009641200304031, + "learning_rate": 1.8357802330009136e-06, + "loss": 0.6502, + "step": 747 + }, + { + "epoch": 0.6237231603085261, + "grad_norm": 0.10499203950166702, + "learning_rate": 1.8352866342102296e-06, + "loss": 0.6812, + "step": 748 + }, + { + "epoch": 0.6245570148009172, + "grad_norm": 0.09757328033447266, + "learning_rate": 1.8347923612890349e-06, + "loss": 0.6025, + "step": 749 + }, + { + "epoch": 0.6253908692933083, + "grad_norm": 0.10616346448659897, + "learning_rate": 1.8342974146362394e-06, + "loss": 0.6638, + "step": 750 + }, + { + "epoch": 0.6262247237856994, + "grad_norm": 0.10131774097681046, + "learning_rate": 1.8338017946512968e-06, + "loss": 0.5456, + "step": 751 + }, + { + "epoch": 0.6270585782780905, + "grad_norm": 0.10244927555322647, + "learning_rate": 1.8333055017342054e-06, + "loss": 0.6751, + "step": 752 + }, + { + "epoch": 0.6278924327704816, + "grad_norm": 0.10338141769170761, + "learning_rate": 1.8328085362855055e-06, + "loss": 0.6647, + "step": 753 + }, + { + "epoch": 0.6287262872628726, + "grad_norm": 0.10026909410953522, + "learning_rate": 1.8323108987062802e-06, + "loss": 0.6543, + "step": 754 + }, + { + "epoch": 0.6295601417552638, + "grad_norm": 0.10641954839229584, + "learning_rate": 1.8318125893981553e-06, + "loss": 0.6425, + "step": 755 + }, + { + "epoch": 0.6303939962476548, + "grad_norm": 0.09939148277044296, + "learning_rate": 1.8313136087632995e-06, + "loss": 0.6531, + "step": 756 + }, + { + "epoch": 0.6312278507400458, + "grad_norm": 0.10437260568141937, + "learning_rate": 1.8308139572044215e-06, + "loss": 0.6783, + "step": 757 + }, + { + "epoch": 0.632061705232437, + "grad_norm": 0.09760341793298721, + "learning_rate": 1.830313635124773e-06, + "loss": 0.6734, + "step": 758 + }, + { + "epoch": 0.632895559724828, + "grad_norm": 0.1017068475484848, + "learning_rate": 1.8298126429281467e-06, + "loss": 0.6171, + "step": 759 + }, + { + "epoch": 0.6337294142172191, + "grad_norm": 0.09881062060594559, + "learning_rate": 1.8293109810188754e-06, + "loss": 0.6309, + "step": 760 + }, + { + "epoch": 0.6345632687096102, + "grad_norm": 0.11102797091007233, + "learning_rate": 1.8288086498018326e-06, + "loss": 0.6361, + "step": 761 + }, + { + "epoch": 0.6353971232020013, + "grad_norm": 0.09473263472318649, + "learning_rate": 1.828305649682433e-06, + "loss": 0.5787, + "step": 762 + }, + { + "epoch": 0.6362309776943923, + "grad_norm": 0.10742990672588348, + "learning_rate": 1.8278019810666293e-06, + "loss": 0.6638, + "step": 763 + }, + { + "epoch": 0.6370648321867834, + "grad_norm": 0.09987856447696686, + "learning_rate": 1.8272976443609156e-06, + "loss": 0.6054, + "step": 764 + }, + { + "epoch": 0.6378986866791745, + "grad_norm": 0.09839458018541336, + "learning_rate": 1.8267926399723238e-06, + "loss": 0.6258, + "step": 765 + }, + { + "epoch": 0.6387325411715655, + "grad_norm": 0.09883236885070801, + "learning_rate": 1.8262869683084258e-06, + "loss": 0.6695, + "step": 766 + }, + { + "epoch": 0.6395663956639567, + "grad_norm": 0.10263945162296295, + "learning_rate": 1.825780629777331e-06, + "loss": 0.6032, + "step": 767 + }, + { + "epoch": 0.6404002501563477, + "grad_norm": 0.09623769670724869, + "learning_rate": 1.8252736247876874e-06, + "loss": 0.6794, + "step": 768 + }, + { + "epoch": 0.6412341046487388, + "grad_norm": 0.10218116641044617, + "learning_rate": 1.8247659537486811e-06, + "loss": 0.5969, + "step": 769 + }, + { + "epoch": 0.6420679591411299, + "grad_norm": 0.09405802190303802, + "learning_rate": 1.8242576170700352e-06, + "loss": 0.6493, + "step": 770 + }, + { + "epoch": 0.642901813633521, + "grad_norm": 0.09709781408309937, + "learning_rate": 1.8237486151620112e-06, + "loss": 0.7054, + "step": 771 + }, + { + "epoch": 0.643735668125912, + "grad_norm": 0.09584866464138031, + "learning_rate": 1.823238948435406e-06, + "loss": 0.6299, + "step": 772 + }, + { + "epoch": 0.6445695226183031, + "grad_norm": 0.09731820970773697, + "learning_rate": 1.8227286173015538e-06, + "loss": 0.6633, + "step": 773 + }, + { + "epoch": 0.6454033771106942, + "grad_norm": 0.09664048999547958, + "learning_rate": 1.822217622172325e-06, + "loss": 0.6075, + "step": 774 + }, + { + "epoch": 0.6462372316030852, + "grad_norm": 0.0990557000041008, + "learning_rate": 1.8217059634601259e-06, + "loss": 0.6755, + "step": 775 + }, + { + "epoch": 0.6470710860954764, + "grad_norm": 0.10071469098329544, + "learning_rate": 1.8211936415778983e-06, + "loss": 0.6197, + "step": 776 + }, + { + "epoch": 0.6479049405878674, + "grad_norm": 0.10386556386947632, + "learning_rate": 1.820680656939119e-06, + "loss": 0.6195, + "step": 777 + }, + { + "epoch": 0.6487387950802584, + "grad_norm": 0.09966234862804413, + "learning_rate": 1.8201670099578001e-06, + "loss": 0.6654, + "step": 778 + }, + { + "epoch": 0.6495726495726496, + "grad_norm": 0.09882418066263199, + "learning_rate": 1.819652701048488e-06, + "loss": 0.6525, + "step": 779 + }, + { + "epoch": 0.6504065040650406, + "grad_norm": 0.09504549950361252, + "learning_rate": 1.8191377306262633e-06, + "loss": 0.6326, + "step": 780 + }, + { + "epoch": 0.6512403585574317, + "grad_norm": 0.09834848344326019, + "learning_rate": 1.8186220991067404e-06, + "loss": 0.6417, + "step": 781 + }, + { + "epoch": 0.6520742130498228, + "grad_norm": 0.09583116322755814, + "learning_rate": 1.8181058069060676e-06, + "loss": 0.6644, + "step": 782 + }, + { + "epoch": 0.6529080675422139, + "grad_norm": 0.09287475794553757, + "learning_rate": 1.8175888544409264e-06, + "loss": 0.6385, + "step": 783 + }, + { + "epoch": 0.653741922034605, + "grad_norm": 0.10364864766597748, + "learning_rate": 1.8170712421285305e-06, + "loss": 0.6996, + "step": 784 + }, + { + "epoch": 0.6545757765269961, + "grad_norm": 0.09863609820604324, + "learning_rate": 1.816552970386627e-06, + "loss": 0.6272, + "step": 785 + }, + { + "epoch": 0.6554096310193871, + "grad_norm": 0.10557069629430771, + "learning_rate": 1.8160340396334949e-06, + "loss": 0.6318, + "step": 786 + }, + { + "epoch": 0.6562434855117782, + "grad_norm": 0.10506842285394669, + "learning_rate": 1.8155144502879443e-06, + "loss": 0.6557, + "step": 787 + }, + { + "epoch": 0.6570773400041693, + "grad_norm": 0.09928935766220093, + "learning_rate": 1.8149942027693182e-06, + "loss": 0.6832, + "step": 788 + }, + { + "epoch": 0.6579111944965603, + "grad_norm": 0.09821849316358566, + "learning_rate": 1.8144732974974902e-06, + "loss": 0.6342, + "step": 789 + }, + { + "epoch": 0.6587450489889515, + "grad_norm": 0.09966670721769333, + "learning_rate": 1.8139517348928637e-06, + "loss": 0.6321, + "step": 790 + }, + { + "epoch": 0.6595789034813425, + "grad_norm": 0.10439924150705338, + "learning_rate": 1.8134295153763745e-06, + "loss": 0.6552, + "step": 791 + }, + { + "epoch": 0.6604127579737336, + "grad_norm": 0.09848225861787796, + "learning_rate": 1.8129066393694877e-06, + "loss": 0.593, + "step": 792 + }, + { + "epoch": 0.6612466124661247, + "grad_norm": 0.09793344885110855, + "learning_rate": 1.8123831072941974e-06, + "loss": 0.6406, + "step": 793 + }, + { + "epoch": 0.6620804669585157, + "grad_norm": 0.0945219025015831, + "learning_rate": 1.8118589195730285e-06, + "loss": 0.6118, + "step": 794 + }, + { + "epoch": 0.6629143214509068, + "grad_norm": 0.10230748355388641, + "learning_rate": 1.8113340766290343e-06, + "loss": 0.65, + "step": 795 + }, + { + "epoch": 0.6637481759432979, + "grad_norm": 0.10141695290803909, + "learning_rate": 1.8108085788857973e-06, + "loss": 0.6109, + "step": 796 + }, + { + "epoch": 0.664582030435689, + "grad_norm": 0.11011820286512375, + "learning_rate": 1.8102824267674282e-06, + "loss": 0.6531, + "step": 797 + }, + { + "epoch": 0.66541588492808, + "grad_norm": 0.09687966108322144, + "learning_rate": 1.8097556206985658e-06, + "loss": 0.649, + "step": 798 + }, + { + "epoch": 0.6662497394204712, + "grad_norm": 0.09653454273939133, + "learning_rate": 1.8092281611043764e-06, + "loss": 0.6487, + "step": 799 + }, + { + "epoch": 0.6670835939128622, + "grad_norm": 0.10373052954673767, + "learning_rate": 1.808700048410555e-06, + "loss": 0.6221, + "step": 800 + }, + { + "epoch": 0.6670835939128622, + "eval_loss": 0.6415942907333374, + "eval_runtime": 321.6008, + "eval_samples_per_second": 16.738, + "eval_steps_per_second": 2.792, + "step": 800 + }, + { + "epoch": 0.6679174484052532, + "grad_norm": 0.10485277324914932, + "learning_rate": 1.8081712830433224e-06, + "loss": 0.6176, + "step": 801 + }, + { + "epoch": 0.6687513028976444, + "grad_norm": 0.10105489194393158, + "learning_rate": 1.8076418654294266e-06, + "loss": 0.6708, + "step": 802 + }, + { + "epoch": 0.6695851573900354, + "grad_norm": 0.1096423864364624, + "learning_rate": 1.8071117959961416e-06, + "loss": 0.601, + "step": 803 + }, + { + "epoch": 0.6704190118824265, + "grad_norm": 0.10762202739715576, + "learning_rate": 1.8065810751712682e-06, + "loss": 0.6567, + "step": 804 + }, + { + "epoch": 0.6712528663748176, + "grad_norm": 0.1107725277543068, + "learning_rate": 1.8060497033831324e-06, + "loss": 0.5889, + "step": 805 + }, + { + "epoch": 0.6720867208672087, + "grad_norm": 0.10468114167451859, + "learning_rate": 1.8055176810605857e-06, + "loss": 0.5847, + "step": 806 + }, + { + "epoch": 0.6729205753595997, + "grad_norm": 0.11013887077569962, + "learning_rate": 1.8049850086330048e-06, + "loss": 0.6008, + "step": 807 + }, + { + "epoch": 0.6737544298519909, + "grad_norm": 0.1043848916888237, + "learning_rate": 1.8044516865302908e-06, + "loss": 0.6368, + "step": 808 + }, + { + "epoch": 0.6745882843443819, + "grad_norm": 0.09962385892868042, + "learning_rate": 1.8039177151828691e-06, + "loss": 0.5842, + "step": 809 + }, + { + "epoch": 0.6754221388367729, + "grad_norm": 0.1098175197839737, + "learning_rate": 1.8033830950216897e-06, + "loss": 0.6397, + "step": 810 + }, + { + "epoch": 0.6762559933291641, + "grad_norm": 0.10420270264148712, + "learning_rate": 1.8028478264782252e-06, + "loss": 0.6131, + "step": 811 + }, + { + "epoch": 0.6770898478215551, + "grad_norm": 0.10396002978086472, + "learning_rate": 1.8023119099844723e-06, + "loss": 0.6499, + "step": 812 + }, + { + "epoch": 0.6779237023139462, + "grad_norm": 0.09836006909608841, + "learning_rate": 1.8017753459729504e-06, + "loss": 0.5962, + "step": 813 + }, + { + "epoch": 0.6787575568063373, + "grad_norm": 0.11005699634552002, + "learning_rate": 1.8012381348767014e-06, + "loss": 0.6202, + "step": 814 + }, + { + "epoch": 0.6795914112987284, + "grad_norm": 0.10937905311584473, + "learning_rate": 1.8007002771292896e-06, + "loss": 0.6053, + "step": 815 + }, + { + "epoch": 0.6804252657911195, + "grad_norm": 0.10817151516675949, + "learning_rate": 1.800161773164801e-06, + "loss": 0.6713, + "step": 816 + }, + { + "epoch": 0.6812591202835105, + "grad_norm": 0.10142417997121811, + "learning_rate": 1.7996226234178434e-06, + "loss": 0.6166, + "step": 817 + }, + { + "epoch": 0.6820929747759016, + "grad_norm": 0.10415297746658325, + "learning_rate": 1.799082828323545e-06, + "loss": 0.6912, + "step": 818 + }, + { + "epoch": 0.6829268292682927, + "grad_norm": 0.09985463321208954, + "learning_rate": 1.7985423883175565e-06, + "loss": 0.6694, + "step": 819 + }, + { + "epoch": 0.6837606837606838, + "grad_norm": 0.1025400459766388, + "learning_rate": 1.7980013038360475e-06, + "loss": 0.6464, + "step": 820 + }, + { + "epoch": 0.6845945382530748, + "grad_norm": 0.09324987977743149, + "learning_rate": 1.7974595753157082e-06, + "loss": 0.6251, + "step": 821 + }, + { + "epoch": 0.685428392745466, + "grad_norm": 0.1065581887960434, + "learning_rate": 1.796917203193749e-06, + "loss": 0.6427, + "step": 822 + }, + { + "epoch": 0.686262247237857, + "grad_norm": 0.1113988533616066, + "learning_rate": 1.7963741879078985e-06, + "loss": 0.6652, + "step": 823 + }, + { + "epoch": 0.687096101730248, + "grad_norm": 0.09690185636281967, + "learning_rate": 1.7958305298964066e-06, + "loss": 0.6294, + "step": 824 + }, + { + "epoch": 0.6879299562226392, + "grad_norm": 0.10385391116142273, + "learning_rate": 1.7952862295980397e-06, + "loss": 0.611, + "step": 825 + }, + { + "epoch": 0.6887638107150302, + "grad_norm": 0.10155371576547623, + "learning_rate": 1.794741287452084e-06, + "loss": 0.6268, + "step": 826 + }, + { + "epoch": 0.6895976652074213, + "grad_norm": 0.10731323808431625, + "learning_rate": 1.7941957038983425e-06, + "loss": 0.6294, + "step": 827 + }, + { + "epoch": 0.6904315196998124, + "grad_norm": 0.0989241898059845, + "learning_rate": 1.7936494793771369e-06, + "loss": 0.6739, + "step": 828 + }, + { + "epoch": 0.6912653741922035, + "grad_norm": 0.10223259776830673, + "learning_rate": 1.7931026143293059e-06, + "loss": 0.5857, + "step": 829 + }, + { + "epoch": 0.6920992286845945, + "grad_norm": 0.10547658056020737, + "learning_rate": 1.792555109196205e-06, + "loss": 0.6864, + "step": 830 + }, + { + "epoch": 0.6929330831769857, + "grad_norm": 0.1027214452624321, + "learning_rate": 1.7920069644197063e-06, + "loss": 0.5857, + "step": 831 + }, + { + "epoch": 0.6937669376693767, + "grad_norm": 0.10017300397157669, + "learning_rate": 1.7914581804421983e-06, + "loss": 0.6466, + "step": 832 + }, + { + "epoch": 0.6946007921617677, + "grad_norm": 0.09962292015552521, + "learning_rate": 1.7909087577065851e-06, + "loss": 0.5917, + "step": 833 + }, + { + "epoch": 0.6954346466541589, + "grad_norm": 0.10357961803674698, + "learning_rate": 1.7903586966562868e-06, + "loss": 0.6382, + "step": 834 + }, + { + "epoch": 0.6962685011465499, + "grad_norm": 0.1033301129937172, + "learning_rate": 1.789807997735238e-06, + "loss": 0.624, + "step": 835 + }, + { + "epoch": 0.697102355638941, + "grad_norm": 0.09935612231492996, + "learning_rate": 1.789256661387889e-06, + "loss": 0.6475, + "step": 836 + }, + { + "epoch": 0.6979362101313321, + "grad_norm": 0.09738706797361374, + "learning_rate": 1.7887046880592033e-06, + "loss": 0.652, + "step": 837 + }, + { + "epoch": 0.6987700646237232, + "grad_norm": 0.10389488190412521, + "learning_rate": 1.7881520781946597e-06, + "loss": 0.6755, + "step": 838 + }, + { + "epoch": 0.6996039191161142, + "grad_norm": 0.09787558019161224, + "learning_rate": 1.7875988322402501e-06, + "loss": 0.618, + "step": 839 + }, + { + "epoch": 0.7004377736085053, + "grad_norm": 0.09852594137191772, + "learning_rate": 1.78704495064248e-06, + "loss": 0.618, + "step": 840 + }, + { + "epoch": 0.7012716281008964, + "grad_norm": 0.09819158911705017, + "learning_rate": 1.7864904338483675e-06, + "loss": 0.5994, + "step": 841 + }, + { + "epoch": 0.7021054825932874, + "grad_norm": 0.10942944139242172, + "learning_rate": 1.7859352823054437e-06, + "loss": 0.6405, + "step": 842 + }, + { + "epoch": 0.7029393370856786, + "grad_norm": 0.10699477791786194, + "learning_rate": 1.7853794964617522e-06, + "loss": 0.6444, + "step": 843 + }, + { + "epoch": 0.7037731915780696, + "grad_norm": 0.10151588916778564, + "learning_rate": 1.784823076765848e-06, + "loss": 0.6233, + "step": 844 + }, + { + "epoch": 0.7046070460704607, + "grad_norm": 0.10147272795438766, + "learning_rate": 1.7842660236667974e-06, + "loss": 0.6606, + "step": 845 + }, + { + "epoch": 0.7054409005628518, + "grad_norm": 0.1031784638762474, + "learning_rate": 1.7837083376141796e-06, + "loss": 0.6332, + "step": 846 + }, + { + "epoch": 0.7062747550552428, + "grad_norm": 0.10318465530872345, + "learning_rate": 1.7831500190580823e-06, + "loss": 0.5944, + "step": 847 + }, + { + "epoch": 0.707108609547634, + "grad_norm": 0.10107962787151337, + "learning_rate": 1.7825910684491054e-06, + "loss": 0.686, + "step": 848 + }, + { + "epoch": 0.707942464040025, + "grad_norm": 0.1008550301194191, + "learning_rate": 1.7820314862383584e-06, + "loss": 0.6356, + "step": 849 + }, + { + "epoch": 0.7087763185324161, + "grad_norm": 0.1049947738647461, + "learning_rate": 1.7814712728774598e-06, + "loss": 0.6436, + "step": 850 + }, + { + "epoch": 0.7096101730248072, + "grad_norm": 0.10559200495481491, + "learning_rate": 1.7809104288185389e-06, + "loss": 0.6522, + "step": 851 + }, + { + "epoch": 0.7104440275171983, + "grad_norm": 0.09807837009429932, + "learning_rate": 1.7803489545142325e-06, + "loss": 0.6252, + "step": 852 + }, + { + "epoch": 0.7112778820095893, + "grad_norm": 0.1072913408279419, + "learning_rate": 1.7797868504176874e-06, + "loss": 0.6483, + "step": 853 + }, + { + "epoch": 0.7121117365019805, + "grad_norm": 0.09987325221300125, + "learning_rate": 1.7792241169825579e-06, + "loss": 0.5955, + "step": 854 + }, + { + "epoch": 0.7129455909943715, + "grad_norm": 0.10160518437623978, + "learning_rate": 1.778660754663006e-06, + "loss": 0.6006, + "step": 855 + }, + { + "epoch": 0.7137794454867625, + "grad_norm": 0.10053478926420212, + "learning_rate": 1.7780967639137025e-06, + "loss": 0.6583, + "step": 856 + }, + { + "epoch": 0.7146132999791537, + "grad_norm": 0.10520876199007034, + "learning_rate": 1.777532145189824e-06, + "loss": 0.6061, + "step": 857 + }, + { + "epoch": 0.7154471544715447, + "grad_norm": 0.10797824710607529, + "learning_rate": 1.776966898947054e-06, + "loss": 0.6764, + "step": 858 + }, + { + "epoch": 0.7162810089639358, + "grad_norm": 0.10396280139684677, + "learning_rate": 1.7764010256415837e-06, + "loss": 0.6302, + "step": 859 + }, + { + "epoch": 0.7171148634563269, + "grad_norm": 0.10683812946081161, + "learning_rate": 1.7758345257301094e-06, + "loss": 0.6618, + "step": 860 + }, + { + "epoch": 0.717948717948718, + "grad_norm": 0.09601815789937973, + "learning_rate": 1.7752673996698326e-06, + "loss": 0.6167, + "step": 861 + }, + { + "epoch": 0.718782572441109, + "grad_norm": 0.1079709529876709, + "learning_rate": 1.774699647918462e-06, + "loss": 0.6085, + "step": 862 + }, + { + "epoch": 0.7196164269335001, + "grad_norm": 0.10923435539007187, + "learning_rate": 1.774131270934209e-06, + "loss": 0.6492, + "step": 863 + }, + { + "epoch": 0.7204502814258912, + "grad_norm": 0.10494917631149292, + "learning_rate": 1.7735622691757912e-06, + "loss": 0.6281, + "step": 864 + }, + { + "epoch": 0.7212841359182822, + "grad_norm": 0.10305362194776535, + "learning_rate": 1.7729926431024301e-06, + "loss": 0.6381, + "step": 865 + }, + { + "epoch": 0.7221179904106734, + "grad_norm": 0.1005750373005867, + "learning_rate": 1.7724223931738505e-06, + "loss": 0.667, + "step": 866 + }, + { + "epoch": 0.7229518449030644, + "grad_norm": 0.10202402621507645, + "learning_rate": 1.7718515198502812e-06, + "loss": 0.6685, + "step": 867 + }, + { + "epoch": 0.7237856993954555, + "grad_norm": 0.10256760567426682, + "learning_rate": 1.7712800235924546e-06, + "loss": 0.6728, + "step": 868 + }, + { + "epoch": 0.7246195538878466, + "grad_norm": 0.1020650714635849, + "learning_rate": 1.7707079048616046e-06, + "loss": 0.6144, + "step": 869 + }, + { + "epoch": 0.7254534083802376, + "grad_norm": 0.10101509839296341, + "learning_rate": 1.770135164119468e-06, + "loss": 0.6244, + "step": 870 + }, + { + "epoch": 0.7262872628726287, + "grad_norm": 0.09737613797187805, + "learning_rate": 1.769561801828284e-06, + "loss": 0.6507, + "step": 871 + }, + { + "epoch": 0.7271211173650198, + "grad_norm": 0.09866712987422943, + "learning_rate": 1.7689878184507937e-06, + "loss": 0.5875, + "step": 872 + }, + { + "epoch": 0.7279549718574109, + "grad_norm": 0.0963241383433342, + "learning_rate": 1.7684132144502382e-06, + "loss": 0.6237, + "step": 873 + }, + { + "epoch": 0.7287888263498019, + "grad_norm": 0.09706535190343857, + "learning_rate": 1.7678379902903603e-06, + "loss": 0.5665, + "step": 874 + }, + { + "epoch": 0.7296226808421931, + "grad_norm": 0.09960032999515533, + "learning_rate": 1.7672621464354034e-06, + "loss": 0.6288, + "step": 875 + }, + { + "epoch": 0.7304565353345841, + "grad_norm": 0.10458897799253464, + "learning_rate": 1.766685683350111e-06, + "loss": 0.6148, + "step": 876 + }, + { + "epoch": 0.7312903898269751, + "grad_norm": 0.10443190485239029, + "learning_rate": 1.766108601499726e-06, + "loss": 0.6248, + "step": 877 + }, + { + "epoch": 0.7321242443193663, + "grad_norm": 0.10145988315343857, + "learning_rate": 1.765530901349991e-06, + "loss": 0.6117, + "step": 878 + }, + { + "epoch": 0.7329580988117573, + "grad_norm": 0.10847010463476181, + "learning_rate": 1.7649525833671474e-06, + "loss": 0.7046, + "step": 879 + }, + { + "epoch": 0.7337919533041485, + "grad_norm": 0.10503443330526352, + "learning_rate": 1.7643736480179352e-06, + "loss": 0.6176, + "step": 880 + }, + { + "epoch": 0.7346258077965395, + "grad_norm": 0.10929647833108902, + "learning_rate": 1.7637940957695934e-06, + "loss": 0.6434, + "step": 881 + }, + { + "epoch": 0.7354596622889306, + "grad_norm": 0.10153260827064514, + "learning_rate": 1.7632139270898576e-06, + "loss": 0.638, + "step": 882 + }, + { + "epoch": 0.7362935167813217, + "grad_norm": 0.10493889451026917, + "learning_rate": 1.7626331424469615e-06, + "loss": 0.6517, + "step": 883 + }, + { + "epoch": 0.7371273712737128, + "grad_norm": 0.09808988124132156, + "learning_rate": 1.7620517423096368e-06, + "loss": 0.6179, + "step": 884 + }, + { + "epoch": 0.7379612257661038, + "grad_norm": 0.10479886084794998, + "learning_rate": 1.7614697271471103e-06, + "loss": 0.6419, + "step": 885 + }, + { + "epoch": 0.7387950802584949, + "grad_norm": 0.09941962361335754, + "learning_rate": 1.7608870974291065e-06, + "loss": 0.6128, + "step": 886 + }, + { + "epoch": 0.739628934750886, + "grad_norm": 0.10169458389282227, + "learning_rate": 1.7603038536258453e-06, + "loss": 0.609, + "step": 887 + }, + { + "epoch": 0.740462789243277, + "grad_norm": 0.09726240485906601, + "learning_rate": 1.7597199962080423e-06, + "loss": 0.6371, + "step": 888 + }, + { + "epoch": 0.7412966437356682, + "grad_norm": 0.10218486189842224, + "learning_rate": 1.759135525646908e-06, + "loss": 0.5768, + "step": 889 + }, + { + "epoch": 0.7421304982280592, + "grad_norm": 0.10321173816919327, + "learning_rate": 1.7585504424141483e-06, + "loss": 0.6007, + "step": 890 + }, + { + "epoch": 0.7429643527204502, + "grad_norm": 0.10717494040727615, + "learning_rate": 1.7579647469819631e-06, + "loss": 0.6353, + "step": 891 + }, + { + "epoch": 0.7437982072128414, + "grad_norm": 0.1051764264702797, + "learning_rate": 1.7573784398230474e-06, + "loss": 0.6098, + "step": 892 + }, + { + "epoch": 0.7446320617052324, + "grad_norm": 0.10321137309074402, + "learning_rate": 1.7567915214105881e-06, + "loss": 0.5986, + "step": 893 + }, + { + "epoch": 0.7454659161976235, + "grad_norm": 0.10979589819908142, + "learning_rate": 1.7562039922182671e-06, + "loss": 0.6518, + "step": 894 + }, + { + "epoch": 0.7462997706900146, + "grad_norm": 0.10153292864561081, + "learning_rate": 1.7556158527202585e-06, + "loss": 0.6358, + "step": 895 + }, + { + "epoch": 0.7471336251824057, + "grad_norm": 0.10513276606798172, + "learning_rate": 1.7550271033912287e-06, + "loss": 0.6085, + "step": 896 + }, + { + "epoch": 0.7479674796747967, + "grad_norm": 0.10275143384933472, + "learning_rate": 1.7544377447063372e-06, + "loss": 0.5748, + "step": 897 + }, + { + "epoch": 0.7488013341671879, + "grad_norm": 0.10438092797994614, + "learning_rate": 1.7538477771412339e-06, + "loss": 0.5943, + "step": 898 + }, + { + "epoch": 0.7496351886595789, + "grad_norm": 0.1010277271270752, + "learning_rate": 1.7532572011720617e-06, + "loss": 0.6112, + "step": 899 + }, + { + "epoch": 0.7504690431519699, + "grad_norm": 0.10229716449975967, + "learning_rate": 1.752666017275453e-06, + "loss": 0.5861, + "step": 900 + }, + { + "epoch": 0.7513028976443611, + "grad_norm": 0.10199355334043503, + "learning_rate": 1.7520742259285323e-06, + "loss": 0.6388, + "step": 901 + }, + { + "epoch": 0.7521367521367521, + "grad_norm": 0.10268741846084595, + "learning_rate": 1.7514818276089128e-06, + "loss": 0.6875, + "step": 902 + }, + { + "epoch": 0.7529706066291432, + "grad_norm": 0.0992840975522995, + "learning_rate": 1.7508888227946992e-06, + "loss": 0.6756, + "step": 903 + }, + { + "epoch": 0.7538044611215343, + "grad_norm": 0.09615278989076614, + "learning_rate": 1.7502952119644845e-06, + "loss": 0.6474, + "step": 904 + }, + { + "epoch": 0.7546383156139254, + "grad_norm": 0.10315293818712234, + "learning_rate": 1.749700995597351e-06, + "loss": 0.6127, + "step": 905 + }, + { + "epoch": 0.7554721701063164, + "grad_norm": 0.10551444441080093, + "learning_rate": 1.7491061741728702e-06, + "loss": 0.6357, + "step": 906 + }, + { + "epoch": 0.7563060245987075, + "grad_norm": 0.10280577838420868, + "learning_rate": 1.748510748171101e-06, + "loss": 0.6045, + "step": 907 + }, + { + "epoch": 0.7571398790910986, + "grad_norm": 0.09516436606645584, + "learning_rate": 1.7479147180725912e-06, + "loss": 0.6368, + "step": 908 + }, + { + "epoch": 0.7579737335834896, + "grad_norm": 0.10799692571163177, + "learning_rate": 1.7473180843583762e-06, + "loss": 0.6669, + "step": 909 + }, + { + "epoch": 0.7588075880758808, + "grad_norm": 0.1064273938536644, + "learning_rate": 1.7467208475099774e-06, + "loss": 0.6481, + "step": 910 + }, + { + "epoch": 0.7596414425682718, + "grad_norm": 0.1054486557841301, + "learning_rate": 1.7461230080094043e-06, + "loss": 0.6108, + "step": 911 + }, + { + "epoch": 0.760475297060663, + "grad_norm": 0.10060249269008636, + "learning_rate": 1.7455245663391516e-06, + "loss": 0.5604, + "step": 912 + }, + { + "epoch": 0.761309151553054, + "grad_norm": 0.1015784963965416, + "learning_rate": 1.7449255229822008e-06, + "loss": 0.6248, + "step": 913 + }, + { + "epoch": 0.762143006045445, + "grad_norm": 0.10114108771085739, + "learning_rate": 1.744325878422019e-06, + "loss": 0.6683, + "step": 914 + }, + { + "epoch": 0.7629768605378362, + "grad_norm": 0.10236384719610214, + "learning_rate": 1.7437256331425576e-06, + "loss": 0.657, + "step": 915 + }, + { + "epoch": 0.7638107150302272, + "grad_norm": 0.10892149806022644, + "learning_rate": 1.743124787628254e-06, + "loss": 0.6434, + "step": 916 + }, + { + "epoch": 0.7646445695226183, + "grad_norm": 0.10639214515686035, + "learning_rate": 1.7425233423640298e-06, + "loss": 0.5859, + "step": 917 + }, + { + "epoch": 0.7654784240150094, + "grad_norm": 0.10279032588005066, + "learning_rate": 1.7419212978352898e-06, + "loss": 0.6779, + "step": 918 + }, + { + "epoch": 0.7663122785074005, + "grad_norm": 0.11178586632013321, + "learning_rate": 1.741318654527923e-06, + "loss": 0.6827, + "step": 919 + }, + { + "epoch": 0.7671461329997915, + "grad_norm": 0.1018877848982811, + "learning_rate": 1.7407154129283019e-06, + "loss": 0.6423, + "step": 920 + }, + { + "epoch": 0.7679799874921827, + "grad_norm": 0.1052117571234703, + "learning_rate": 1.7401115735232817e-06, + "loss": 0.6871, + "step": 921 + }, + { + "epoch": 0.7688138419845737, + "grad_norm": 0.10656667500734329, + "learning_rate": 1.7395071368001997e-06, + "loss": 0.6384, + "step": 922 + }, + { + "epoch": 0.7696476964769647, + "grad_norm": 0.09973619878292084, + "learning_rate": 1.738902103246876e-06, + "loss": 0.5659, + "step": 923 + }, + { + "epoch": 0.7704815509693559, + "grad_norm": 0.10072916746139526, + "learning_rate": 1.7382964733516112e-06, + "loss": 0.6315, + "step": 924 + }, + { + "epoch": 0.7713154054617469, + "grad_norm": 0.11351827532052994, + "learning_rate": 1.7376902476031882e-06, + "loss": 0.7005, + "step": 925 + }, + { + "epoch": 0.772149259954138, + "grad_norm": 0.1084694042801857, + "learning_rate": 1.7370834264908711e-06, + "loss": 0.6077, + "step": 926 + }, + { + "epoch": 0.7729831144465291, + "grad_norm": 0.10465063899755478, + "learning_rate": 1.7364760105044033e-06, + "loss": 0.5937, + "step": 927 + }, + { + "epoch": 0.7738169689389202, + "grad_norm": 0.09937559068202972, + "learning_rate": 1.7358680001340092e-06, + "loss": 0.5854, + "step": 928 + }, + { + "epoch": 0.7746508234313112, + "grad_norm": 0.11517240107059479, + "learning_rate": 1.7352593958703921e-06, + "loss": 0.6738, + "step": 929 + }, + { + "epoch": 0.7754846779237023, + "grad_norm": 0.10875809192657471, + "learning_rate": 1.734650198204736e-06, + "loss": 0.6487, + "step": 930 + }, + { + "epoch": 0.7763185324160934, + "grad_norm": 0.10209451615810394, + "learning_rate": 1.7340404076287021e-06, + "loss": 0.6277, + "step": 931 + }, + { + "epoch": 0.7771523869084844, + "grad_norm": 0.11190709471702576, + "learning_rate": 1.7334300246344316e-06, + "loss": 0.6393, + "step": 932 + }, + { + "epoch": 0.7779862414008756, + "grad_norm": 0.10622856765985489, + "learning_rate": 1.7328190497145427e-06, + "loss": 0.6665, + "step": 933 + }, + { + "epoch": 0.7788200958932666, + "grad_norm": 0.11038866639137268, + "learning_rate": 1.7322074833621318e-06, + "loss": 0.6092, + "step": 934 + }, + { + "epoch": 0.7796539503856577, + "grad_norm": 0.1060342788696289, + "learning_rate": 1.7315953260707735e-06, + "loss": 0.6427, + "step": 935 + }, + { + "epoch": 0.7804878048780488, + "grad_norm": 0.10829740762710571, + "learning_rate": 1.7309825783345174e-06, + "loss": 0.6851, + "step": 936 + }, + { + "epoch": 0.7813216593704398, + "grad_norm": 0.1001635491847992, + "learning_rate": 1.7303692406478908e-06, + "loss": 0.5645, + "step": 937 + }, + { + "epoch": 0.7821555138628309, + "grad_norm": 0.09842494130134583, + "learning_rate": 1.7297553135058978e-06, + "loss": 0.613, + "step": 938 + }, + { + "epoch": 0.782989368355222, + "grad_norm": 0.10335814207792282, + "learning_rate": 1.7291407974040167e-06, + "loss": 0.6583, + "step": 939 + }, + { + "epoch": 0.7838232228476131, + "grad_norm": 0.10193807631731033, + "learning_rate": 1.728525692838202e-06, + "loss": 0.6242, + "step": 940 + }, + { + "epoch": 0.7846570773400041, + "grad_norm": 0.1064397543668747, + "learning_rate": 1.7279100003048832e-06, + "loss": 0.7043, + "step": 941 + }, + { + "epoch": 0.7854909318323953, + "grad_norm": 0.09790313243865967, + "learning_rate": 1.7272937203009642e-06, + "loss": 0.5869, + "step": 942 + }, + { + "epoch": 0.7863247863247863, + "grad_norm": 0.10425697267055511, + "learning_rate": 1.7266768533238225e-06, + "loss": 0.6282, + "step": 943 + }, + { + "epoch": 0.7871586408171775, + "grad_norm": 0.10319948196411133, + "learning_rate": 1.72605939987131e-06, + "loss": 0.5899, + "step": 944 + }, + { + "epoch": 0.7879924953095685, + "grad_norm": 0.10117416828870773, + "learning_rate": 1.725441360441752e-06, + "loss": 0.5734, + "step": 945 + }, + { + "epoch": 0.7888263498019595, + "grad_norm": 0.10920300334692001, + "learning_rate": 1.7248227355339458e-06, + "loss": 0.612, + "step": 946 + }, + { + "epoch": 0.7896602042943507, + "grad_norm": 0.0993722602725029, + "learning_rate": 1.7242035256471623e-06, + "loss": 0.5778, + "step": 947 + }, + { + "epoch": 0.7904940587867417, + "grad_norm": 0.12078487873077393, + "learning_rate": 1.723583731281144e-06, + "loss": 0.7115, + "step": 948 + }, + { + "epoch": 0.7913279132791328, + "grad_norm": 0.10817496478557587, + "learning_rate": 1.7229633529361051e-06, + "loss": 0.6699, + "step": 949 + }, + { + "epoch": 0.7921617677715239, + "grad_norm": 0.10841819643974304, + "learning_rate": 1.7223423911127313e-06, + "loss": 0.6089, + "step": 950 + }, + { + "epoch": 0.792995622263915, + "grad_norm": 0.10630793124437332, + "learning_rate": 1.7217208463121788e-06, + "loss": 0.6615, + "step": 951 + }, + { + "epoch": 0.793829476756306, + "grad_norm": 0.1000799611210823, + "learning_rate": 1.721098719036075e-06, + "loss": 0.6231, + "step": 952 + }, + { + "epoch": 0.7946633312486971, + "grad_norm": 0.11009353399276733, + "learning_rate": 1.7204760097865167e-06, + "loss": 0.6197, + "step": 953 + }, + { + "epoch": 0.7954971857410882, + "grad_norm": 0.10809972882270813, + "learning_rate": 1.7198527190660706e-06, + "loss": 0.5975, + "step": 954 + }, + { + "epoch": 0.7963310402334792, + "grad_norm": 0.10163717716932297, + "learning_rate": 1.719228847377773e-06, + "loss": 0.7125, + "step": 955 + }, + { + "epoch": 0.7971648947258704, + "grad_norm": 0.1121022030711174, + "learning_rate": 1.7186043952251286e-06, + "loss": 0.6821, + "step": 956 + }, + { + "epoch": 0.7979987492182614, + "grad_norm": 0.11300718039274216, + "learning_rate": 1.7179793631121106e-06, + "loss": 0.6137, + "step": 957 + }, + { + "epoch": 0.7988326037106525, + "grad_norm": 0.10873489826917648, + "learning_rate": 1.717353751543161e-06, + "loss": 0.5621, + "step": 958 + }, + { + "epoch": 0.7996664582030436, + "grad_norm": 0.11031791567802429, + "learning_rate": 1.7167275610231887e-06, + "loss": 0.6363, + "step": 959 + }, + { + "epoch": 0.8005003126954346, + "grad_norm": 0.10201506316661835, + "learning_rate": 1.7161007920575704e-06, + "loss": 0.5696, + "step": 960 + }, + { + "epoch": 0.8013341671878257, + "grad_norm": 0.09877403825521469, + "learning_rate": 1.7154734451521486e-06, + "loss": 0.6027, + "step": 961 + }, + { + "epoch": 0.8021680216802168, + "grad_norm": 0.1055438295006752, + "learning_rate": 1.7148455208132334e-06, + "loss": 0.6204, + "step": 962 + }, + { + "epoch": 0.8030018761726079, + "grad_norm": 0.11029175668954849, + "learning_rate": 1.7142170195476005e-06, + "loss": 0.5826, + "step": 963 + }, + { + "epoch": 0.8038357306649989, + "grad_norm": 0.1105756014585495, + "learning_rate": 1.7135879418624913e-06, + "loss": 0.6526, + "step": 964 + }, + { + "epoch": 0.8046695851573901, + "grad_norm": 0.1054423451423645, + "learning_rate": 1.712958288265612e-06, + "loss": 0.6254, + "step": 965 + }, + { + "epoch": 0.8055034396497811, + "grad_norm": 0.10613211989402771, + "learning_rate": 1.712328059265134e-06, + "loss": 0.5974, + "step": 966 + }, + { + "epoch": 0.8063372941421721, + "grad_norm": 0.09556613117456436, + "learning_rate": 1.7116972553696932e-06, + "loss": 0.6144, + "step": 967 + }, + { + "epoch": 0.8071711486345633, + "grad_norm": 0.10485559701919556, + "learning_rate": 1.711065877088389e-06, + "loss": 0.5998, + "step": 968 + }, + { + "epoch": 0.8080050031269543, + "grad_norm": 0.10885662585496902, + "learning_rate": 1.7104339249307848e-06, + "loss": 0.6298, + "step": 969 + }, + { + "epoch": 0.8088388576193454, + "grad_norm": 0.10140252858400345, + "learning_rate": 1.709801399406907e-06, + "loss": 0.5747, + "step": 970 + }, + { + "epoch": 0.8096727121117365, + "grad_norm": 0.09874554723501205, + "learning_rate": 1.7091683010272446e-06, + "loss": 0.5676, + "step": 971 + }, + { + "epoch": 0.8105065666041276, + "grad_norm": 0.10777262598276138, + "learning_rate": 1.7085346303027493e-06, + "loss": 0.6326, + "step": 972 + }, + { + "epoch": 0.8113404210965186, + "grad_norm": 0.09974364936351776, + "learning_rate": 1.7079003877448344e-06, + "loss": 0.5638, + "step": 973 + }, + { + "epoch": 0.8121742755889098, + "grad_norm": 0.10661293566226959, + "learning_rate": 1.7072655738653745e-06, + "loss": 0.6641, + "step": 974 + }, + { + "epoch": 0.8130081300813008, + "grad_norm": 0.10669861733913422, + "learning_rate": 1.7066301891767061e-06, + "loss": 0.6478, + "step": 975 + }, + { + "epoch": 0.813841984573692, + "grad_norm": 0.10368253290653229, + "learning_rate": 1.7059942341916256e-06, + "loss": 0.6198, + "step": 976 + }, + { + "epoch": 0.814675839066083, + "grad_norm": 0.10855797678232193, + "learning_rate": 1.7053577094233897e-06, + "loss": 0.6453, + "step": 977 + }, + { + "epoch": 0.815509693558474, + "grad_norm": 0.11007635295391083, + "learning_rate": 1.7047206153857156e-06, + "loss": 0.5748, + "step": 978 + }, + { + "epoch": 0.8163435480508652, + "grad_norm": 0.10489246994256973, + "learning_rate": 1.7040829525927796e-06, + "loss": 0.579, + "step": 979 + }, + { + "epoch": 0.8171774025432562, + "grad_norm": 0.11224298179149628, + "learning_rate": 1.7034447215592164e-06, + "loss": 0.6484, + "step": 980 + }, + { + "epoch": 0.8180112570356473, + "grad_norm": 0.10526615381240845, + "learning_rate": 1.7028059228001204e-06, + "loss": 0.6746, + "step": 981 + }, + { + "epoch": 0.8188451115280384, + "grad_norm": 0.10187442600727081, + "learning_rate": 1.7021665568310435e-06, + "loss": 0.6462, + "step": 982 + }, + { + "epoch": 0.8196789660204294, + "grad_norm": 0.10846269875764847, + "learning_rate": 1.7015266241679952e-06, + "loss": 0.6329, + "step": 983 + }, + { + "epoch": 0.8205128205128205, + "grad_norm": 0.110533706843853, + "learning_rate": 1.7008861253274429e-06, + "loss": 0.658, + "step": 984 + }, + { + "epoch": 0.8213466750052116, + "grad_norm": 0.10688811540603638, + "learning_rate": 1.7002450608263107e-06, + "loss": 0.5996, + "step": 985 + }, + { + "epoch": 0.8221805294976027, + "grad_norm": 0.10414768755435944, + "learning_rate": 1.6996034311819796e-06, + "loss": 0.6333, + "step": 986 + }, + { + "epoch": 0.8230143839899937, + "grad_norm": 0.0999814048409462, + "learning_rate": 1.698961236912286e-06, + "loss": 0.6609, + "step": 987 + }, + { + "epoch": 0.8238482384823849, + "grad_norm": 0.10203441977500916, + "learning_rate": 1.6983184785355222e-06, + "loss": 0.6095, + "step": 988 + }, + { + "epoch": 0.8246820929747759, + "grad_norm": 0.10571661591529846, + "learning_rate": 1.6976751565704362e-06, + "loss": 0.6125, + "step": 989 + }, + { + "epoch": 0.8255159474671669, + "grad_norm": 0.10715372860431671, + "learning_rate": 1.6970312715362304e-06, + "loss": 0.689, + "step": 990 + }, + { + "epoch": 0.8263498019595581, + "grad_norm": 0.10807687044143677, + "learning_rate": 1.696386823952562e-06, + "loss": 0.611, + "step": 991 + }, + { + "epoch": 0.8271836564519491, + "grad_norm": 0.10625305026769638, + "learning_rate": 1.6957418143395418e-06, + "loss": 0.6616, + "step": 992 + }, + { + "epoch": 0.8280175109443402, + "grad_norm": 0.10512302070856094, + "learning_rate": 1.6950962432177348e-06, + "loss": 0.6147, + "step": 993 + }, + { + "epoch": 0.8288513654367313, + "grad_norm": 0.10524202138185501, + "learning_rate": 1.6944501111081579e-06, + "loss": 0.6094, + "step": 994 + }, + { + "epoch": 0.8296852199291224, + "grad_norm": 0.10839590430259705, + "learning_rate": 1.6938034185322828e-06, + "loss": 0.5816, + "step": 995 + }, + { + "epoch": 0.8305190744215134, + "grad_norm": 0.10930492728948593, + "learning_rate": 1.6931561660120312e-06, + "loss": 0.6621, + "step": 996 + }, + { + "epoch": 0.8313529289139046, + "grad_norm": 0.10111601650714874, + "learning_rate": 1.692508354069779e-06, + "loss": 0.6083, + "step": 997 + }, + { + "epoch": 0.8321867834062956, + "grad_norm": 0.11010827124118805, + "learning_rate": 1.691859983228352e-06, + "loss": 0.6689, + "step": 998 + }, + { + "epoch": 0.8330206378986866, + "grad_norm": 0.10622293502092361, + "learning_rate": 1.6912110540110272e-06, + "loss": 0.6142, + "step": 999 + }, + { + "epoch": 0.8338544923910778, + "grad_norm": 0.10322947055101395, + "learning_rate": 1.6905615669415325e-06, + "loss": 0.5797, + "step": 1000 + }, + { + "epoch": 0.8346883468834688, + "grad_norm": 0.11376485228538513, + "learning_rate": 1.689911522544047e-06, + "loss": 0.6323, + "step": 1001 + }, + { + "epoch": 0.8355222013758599, + "grad_norm": 0.10336299985647202, + "learning_rate": 1.6892609213431981e-06, + "loss": 0.5907, + "step": 1002 + }, + { + "epoch": 0.836356055868251, + "grad_norm": 0.10549134016036987, + "learning_rate": 1.6886097638640631e-06, + "loss": 0.6483, + "step": 1003 + }, + { + "epoch": 0.837189910360642, + "grad_norm": 0.10979744791984558, + "learning_rate": 1.6879580506321687e-06, + "loss": 0.5829, + "step": 1004 + }, + { + "epoch": 0.8380237648530331, + "grad_norm": 0.10647737979888916, + "learning_rate": 1.6873057821734894e-06, + "loss": 0.5714, + "step": 1005 + }, + { + "epoch": 0.8388576193454242, + "grad_norm": 0.11350049823522568, + "learning_rate": 1.6866529590144485e-06, + "loss": 0.6756, + "step": 1006 + }, + { + "epoch": 0.8396914738378153, + "grad_norm": 0.10672125220298767, + "learning_rate": 1.6859995816819167e-06, + "loss": 0.6459, + "step": 1007 + }, + { + "epoch": 0.8405253283302064, + "grad_norm": 0.11561840772628784, + "learning_rate": 1.6853456507032112e-06, + "loss": 0.6912, + "step": 1008 + }, + { + "epoch": 0.8413591828225975, + "grad_norm": 0.11554259806871414, + "learning_rate": 1.6846911666060973e-06, + "loss": 0.6032, + "step": 1009 + }, + { + "epoch": 0.8421930373149885, + "grad_norm": 0.10536365956068039, + "learning_rate": 1.6840361299187857e-06, + "loss": 0.6182, + "step": 1010 + }, + { + "epoch": 0.8430268918073797, + "grad_norm": 0.1125800758600235, + "learning_rate": 1.6833805411699344e-06, + "loss": 0.6519, + "step": 1011 + }, + { + "epoch": 0.8438607462997707, + "grad_norm": 0.10594160109758377, + "learning_rate": 1.6827244008886453e-06, + "loss": 0.6407, + "step": 1012 + }, + { + "epoch": 0.8446946007921617, + "grad_norm": 0.10480530560016632, + "learning_rate": 1.6820677096044667e-06, + "loss": 0.6398, + "step": 1013 + }, + { + "epoch": 0.8455284552845529, + "grad_norm": 0.10666308552026749, + "learning_rate": 1.6814104678473905e-06, + "loss": 0.612, + "step": 1014 + }, + { + "epoch": 0.8463623097769439, + "grad_norm": 0.10556510835886002, + "learning_rate": 1.6807526761478533e-06, + "loss": 0.5729, + "step": 1015 + }, + { + "epoch": 0.847196164269335, + "grad_norm": 0.11039448529481888, + "learning_rate": 1.6800943350367368e-06, + "loss": 0.6323, + "step": 1016 + }, + { + "epoch": 0.8480300187617261, + "grad_norm": 0.10304141789674759, + "learning_rate": 1.6794354450453638e-06, + "loss": 0.6426, + "step": 1017 + }, + { + "epoch": 0.8488638732541172, + "grad_norm": 0.10969959199428558, + "learning_rate": 1.678776006705502e-06, + "loss": 0.558, + "step": 1018 + }, + { + "epoch": 0.8496977277465082, + "grad_norm": 0.10663289576768875, + "learning_rate": 1.6781160205493605e-06, + "loss": 0.593, + "step": 1019 + }, + { + "epoch": 0.8505315822388994, + "grad_norm": 0.10291384905576706, + "learning_rate": 1.6774554871095915e-06, + "loss": 0.6281, + "step": 1020 + }, + { + "epoch": 0.8513654367312904, + "grad_norm": 0.10907971113920212, + "learning_rate": 1.6767944069192876e-06, + "loss": 0.6158, + "step": 1021 + }, + { + "epoch": 0.8521992912236814, + "grad_norm": 0.11564616858959198, + "learning_rate": 1.6761327805119838e-06, + "loss": 0.6662, + "step": 1022 + }, + { + "epoch": 0.8530331457160726, + "grad_norm": 0.10590541362762451, + "learning_rate": 1.6754706084216555e-06, + "loss": 0.6323, + "step": 1023 + }, + { + "epoch": 0.8538670002084636, + "grad_norm": 0.10117157548666, + "learning_rate": 1.6748078911827187e-06, + "loss": 0.6112, + "step": 1024 + }, + { + "epoch": 0.8547008547008547, + "grad_norm": 0.11531368643045425, + "learning_rate": 1.674144629330029e-06, + "loss": 0.6413, + "step": 1025 + }, + { + "epoch": 0.8555347091932458, + "grad_norm": 0.10621378570795059, + "learning_rate": 1.673480823398882e-06, + "loss": 0.6394, + "step": 1026 + }, + { + "epoch": 0.8563685636856369, + "grad_norm": 0.11612808704376221, + "learning_rate": 1.672816473925012e-06, + "loss": 0.637, + "step": 1027 + }, + { + "epoch": 0.8572024181780279, + "grad_norm": 0.10893040150403976, + "learning_rate": 1.672151581444592e-06, + "loss": 0.5923, + "step": 1028 + }, + { + "epoch": 0.858036272670419, + "grad_norm": 0.10822536796331406, + "learning_rate": 1.6714861464942333e-06, + "loss": 0.6766, + "step": 1029 + }, + { + "epoch": 0.8588701271628101, + "grad_norm": 0.10714786499738693, + "learning_rate": 1.6708201696109856e-06, + "loss": 0.5649, + "step": 1030 + }, + { + "epoch": 0.8597039816552011, + "grad_norm": 0.10966484993696213, + "learning_rate": 1.6701536513323349e-06, + "loss": 0.6219, + "step": 1031 + }, + { + "epoch": 0.8605378361475923, + "grad_norm": 0.1037144884467125, + "learning_rate": 1.669486592196205e-06, + "loss": 0.6352, + "step": 1032 + }, + { + "epoch": 0.8613716906399833, + "grad_norm": 0.11299126595258713, + "learning_rate": 1.668818992740956e-06, + "loss": 0.6734, + "step": 1033 + }, + { + "epoch": 0.8622055451323744, + "grad_norm": 0.11029834300279617, + "learning_rate": 1.6681508535053834e-06, + "loss": 0.6579, + "step": 1034 + }, + { + "epoch": 0.8630393996247655, + "grad_norm": 0.10519642382860184, + "learning_rate": 1.6674821750287197e-06, + "loss": 0.6072, + "step": 1035 + }, + { + "epoch": 0.8638732541171565, + "grad_norm": 0.11065202951431274, + "learning_rate": 1.6668129578506313e-06, + "loss": 0.6562, + "step": 1036 + }, + { + "epoch": 0.8647071086095476, + "grad_norm": 0.11024871468544006, + "learning_rate": 1.66614320251122e-06, + "loss": 0.6, + "step": 1037 + }, + { + "epoch": 0.8655409631019387, + "grad_norm": 0.10754359513521194, + "learning_rate": 1.6654729095510219e-06, + "loss": 0.6028, + "step": 1038 + }, + { + "epoch": 0.8663748175943298, + "grad_norm": 0.11345981806516647, + "learning_rate": 1.6648020795110069e-06, + "loss": 0.7018, + "step": 1039 + }, + { + "epoch": 0.8672086720867209, + "grad_norm": 0.10346484929323196, + "learning_rate": 1.6641307129325783e-06, + "loss": 0.6866, + "step": 1040 + }, + { + "epoch": 0.868042526579112, + "grad_norm": 0.10640288889408112, + "learning_rate": 1.6634588103575723e-06, + "loss": 0.5851, + "step": 1041 + }, + { + "epoch": 0.868876381071503, + "grad_norm": 0.11161711812019348, + "learning_rate": 1.6627863723282584e-06, + "loss": 0.6182, + "step": 1042 + }, + { + "epoch": 0.8697102355638942, + "grad_norm": 0.10999694466590881, + "learning_rate": 1.662113399387337e-06, + "loss": 0.5887, + "step": 1043 + }, + { + "epoch": 0.8705440900562852, + "grad_norm": 0.10471642017364502, + "learning_rate": 1.6614398920779418e-06, + "loss": 0.6854, + "step": 1044 + }, + { + "epoch": 0.8713779445486762, + "grad_norm": 0.10635983943939209, + "learning_rate": 1.660765850943636e-06, + "loss": 0.5945, + "step": 1045 + }, + { + "epoch": 0.8722117990410674, + "grad_norm": 0.10250985622406006, + "learning_rate": 1.6600912765284153e-06, + "loss": 0.6283, + "step": 1046 + }, + { + "epoch": 0.8730456535334584, + "grad_norm": 0.10342993587255478, + "learning_rate": 1.6594161693767046e-06, + "loss": 0.5878, + "step": 1047 + }, + { + "epoch": 0.8738795080258495, + "grad_norm": 0.10388782620429993, + "learning_rate": 1.6587405300333593e-06, + "loss": 0.6197, + "step": 1048 + }, + { + "epoch": 0.8747133625182406, + "grad_norm": 0.10757216066122055, + "learning_rate": 1.658064359043664e-06, + "loss": 0.6643, + "step": 1049 + }, + { + "epoch": 0.8755472170106317, + "grad_norm": 0.10185371339321136, + "learning_rate": 1.657387656953333e-06, + "loss": 0.6129, + "step": 1050 + }, + { + "epoch": 0.8763810715030227, + "grad_norm": 0.10625848174095154, + "learning_rate": 1.6567104243085081e-06, + "loss": 0.6389, + "step": 1051 + }, + { + "epoch": 0.8772149259954138, + "grad_norm": 0.1068265438079834, + "learning_rate": 1.6560326616557605e-06, + "loss": 0.6474, + "step": 1052 + }, + { + "epoch": 0.8780487804878049, + "grad_norm": 0.10008926689624786, + "learning_rate": 1.6553543695420888e-06, + "loss": 0.592, + "step": 1053 + }, + { + "epoch": 0.8788826349801959, + "grad_norm": 0.11250849813222885, + "learning_rate": 1.6546755485149181e-06, + "loss": 0.6612, + "step": 1054 + }, + { + "epoch": 0.8797164894725871, + "grad_norm": 0.10780669748783112, + "learning_rate": 1.6539961991221017e-06, + "loss": 0.5717, + "step": 1055 + }, + { + "epoch": 0.8805503439649781, + "grad_norm": 0.10522449761629105, + "learning_rate": 1.6533163219119181e-06, + "loss": 0.6377, + "step": 1056 + }, + { + "epoch": 0.8813841984573692, + "grad_norm": 0.11719299852848053, + "learning_rate": 1.6526359174330727e-06, + "loss": 0.6148, + "step": 1057 + }, + { + "epoch": 0.8822180529497603, + "grad_norm": 0.10418464988470078, + "learning_rate": 1.6519549862346959e-06, + "loss": 0.547, + "step": 1058 + }, + { + "epoch": 0.8830519074421513, + "grad_norm": 0.1149667277932167, + "learning_rate": 1.6512735288663433e-06, + "loss": 0.5509, + "step": 1059 + }, + { + "epoch": 0.8838857619345424, + "grad_norm": 0.11353281885385513, + "learning_rate": 1.6505915458779954e-06, + "loss": 0.6679, + "step": 1060 + }, + { + "epoch": 0.8847196164269335, + "grad_norm": 0.10084139555692673, + "learning_rate": 1.6499090378200564e-06, + "loss": 0.6315, + "step": 1061 + }, + { + "epoch": 0.8855534709193246, + "grad_norm": 0.10753504931926727, + "learning_rate": 1.6492260052433551e-06, + "loss": 0.6441, + "step": 1062 + }, + { + "epoch": 0.8863873254117156, + "grad_norm": 0.10944227129220963, + "learning_rate": 1.6485424486991427e-06, + "loss": 0.6378, + "step": 1063 + }, + { + "epoch": 0.8872211799041068, + "grad_norm": 0.10542403161525726, + "learning_rate": 1.6478583687390937e-06, + "loss": 0.5715, + "step": 1064 + }, + { + "epoch": 0.8880550343964978, + "grad_norm": 0.10192226618528366, + "learning_rate": 1.6471737659153054e-06, + "loss": 0.6035, + "step": 1065 + }, + { + "epoch": 0.8888888888888888, + "grad_norm": 0.11085714399814606, + "learning_rate": 1.6464886407802958e-06, + "loss": 0.6705, + "step": 1066 + }, + { + "epoch": 0.88972274338128, + "grad_norm": 0.10521090775728226, + "learning_rate": 1.645802993887006e-06, + "loss": 0.6487, + "step": 1067 + }, + { + "epoch": 0.890556597873671, + "grad_norm": 0.11182847619056702, + "learning_rate": 1.645116825788798e-06, + "loss": 0.5908, + "step": 1068 + }, + { + "epoch": 0.8913904523660621, + "grad_norm": 0.10744032263755798, + "learning_rate": 1.6444301370394533e-06, + "loss": 0.6978, + "step": 1069 + }, + { + "epoch": 0.8922243068584532, + "grad_norm": 0.11645323783159256, + "learning_rate": 1.6437429281931742e-06, + "loss": 0.6552, + "step": 1070 + }, + { + "epoch": 0.8930581613508443, + "grad_norm": 0.10556191205978394, + "learning_rate": 1.6430551998045833e-06, + "loss": 0.6485, + "step": 1071 + }, + { + "epoch": 0.8938920158432354, + "grad_norm": 0.11244357377290726, + "learning_rate": 1.6423669524287216e-06, + "loss": 0.6352, + "step": 1072 + }, + { + "epoch": 0.8947258703356264, + "grad_norm": 0.11015337705612183, + "learning_rate": 1.6416781866210494e-06, + "loss": 0.5928, + "step": 1073 + }, + { + "epoch": 0.8955597248280175, + "grad_norm": 0.11401670426130295, + "learning_rate": 1.6409889029374457e-06, + "loss": 0.591, + "step": 1074 + }, + { + "epoch": 0.8963935793204086, + "grad_norm": 0.10178755223751068, + "learning_rate": 1.6402991019342073e-06, + "loss": 0.5772, + "step": 1075 + }, + { + "epoch": 0.8972274338127997, + "grad_norm": 0.10722572356462479, + "learning_rate": 1.6396087841680478e-06, + "loss": 0.616, + "step": 1076 + }, + { + "epoch": 0.8980612883051907, + "grad_norm": 0.10824041068553925, + "learning_rate": 1.6389179501960987e-06, + "loss": 0.6289, + "step": 1077 + }, + { + "epoch": 0.8988951427975819, + "grad_norm": 0.10898349434137344, + "learning_rate": 1.638226600575908e-06, + "loss": 0.6197, + "step": 1078 + }, + { + "epoch": 0.8997289972899729, + "grad_norm": 0.10828059166669846, + "learning_rate": 1.6375347358654397e-06, + "loss": 0.6335, + "step": 1079 + }, + { + "epoch": 0.900562851782364, + "grad_norm": 0.11353094130754471, + "learning_rate": 1.6368423566230728e-06, + "loss": 0.6306, + "step": 1080 + }, + { + "epoch": 0.9013967062747551, + "grad_norm": 0.1095949038863182, + "learning_rate": 1.6361494634076033e-06, + "loss": 0.6579, + "step": 1081 + }, + { + "epoch": 0.9022305607671461, + "grad_norm": 0.10779423266649246, + "learning_rate": 1.6354560567782406e-06, + "loss": 0.6268, + "step": 1082 + }, + { + "epoch": 0.9030644152595372, + "grad_norm": 0.10539865493774414, + "learning_rate": 1.6347621372946088e-06, + "loss": 0.6318, + "step": 1083 + }, + { + "epoch": 0.9038982697519283, + "grad_norm": 0.11474636197090149, + "learning_rate": 1.6340677055167458e-06, + "loss": 0.6843, + "step": 1084 + }, + { + "epoch": 0.9047321242443194, + "grad_norm": 0.11284147948026657, + "learning_rate": 1.633372762005103e-06, + "loss": 0.6167, + "step": 1085 + }, + { + "epoch": 0.9055659787367104, + "grad_norm": 0.10842804610729218, + "learning_rate": 1.632677307320545e-06, + "loss": 0.5851, + "step": 1086 + }, + { + "epoch": 0.9063998332291016, + "grad_norm": 0.10843487083911896, + "learning_rate": 1.6319813420243495e-06, + "loss": 0.6337, + "step": 1087 + }, + { + "epoch": 0.9072336877214926, + "grad_norm": 0.10792160034179688, + "learning_rate": 1.6312848666782048e-06, + "loss": 0.6015, + "step": 1088 + }, + { + "epoch": 0.9080675422138836, + "grad_norm": 0.10741087794303894, + "learning_rate": 1.6305878818442122e-06, + "loss": 0.6285, + "step": 1089 + }, + { + "epoch": 0.9089013967062748, + "grad_norm": 0.10371655970811844, + "learning_rate": 1.6298903880848834e-06, + "loss": 0.5982, + "step": 1090 + }, + { + "epoch": 0.9097352511986658, + "grad_norm": 0.10612435638904572, + "learning_rate": 1.6291923859631415e-06, + "loss": 0.5962, + "step": 1091 + }, + { + "epoch": 0.9105691056910569, + "grad_norm": 0.10705637186765671, + "learning_rate": 1.6284938760423188e-06, + "loss": 0.5926, + "step": 1092 + }, + { + "epoch": 0.911402960183448, + "grad_norm": 0.10797090083360672, + "learning_rate": 1.627794858886159e-06, + "loss": 0.632, + "step": 1093 + }, + { + "epoch": 0.9122368146758391, + "grad_norm": 0.1132771223783493, + "learning_rate": 1.6270953350588137e-06, + "loss": 0.647, + "step": 1094 + }, + { + "epoch": 0.9130706691682301, + "grad_norm": 0.10793612152338028, + "learning_rate": 1.626395305124844e-06, + "loss": 0.6367, + "step": 1095 + }, + { + "epoch": 0.9139045236606212, + "grad_norm": 0.10706419497728348, + "learning_rate": 1.6256947696492196e-06, + "loss": 0.6472, + "step": 1096 + }, + { + "epoch": 0.9147383781530123, + "grad_norm": 0.10842972993850708, + "learning_rate": 1.6249937291973184e-06, + "loss": 0.6185, + "step": 1097 + }, + { + "epoch": 0.9155722326454033, + "grad_norm": 0.11361444741487503, + "learning_rate": 1.6242921843349252e-06, + "loss": 0.617, + "step": 1098 + }, + { + "epoch": 0.9164060871377945, + "grad_norm": 0.11315753310918808, + "learning_rate": 1.6235901356282322e-06, + "loss": 0.6478, + "step": 1099 + }, + { + "epoch": 0.9172399416301855, + "grad_norm": 0.12495766580104828, + "learning_rate": 1.6228875836438385e-06, + "loss": 0.6252, + "step": 1100 + }, + { + "epoch": 0.9180737961225766, + "grad_norm": 0.11157601326704025, + "learning_rate": 1.622184528948749e-06, + "loss": 0.6214, + "step": 1101 + }, + { + "epoch": 0.9189076506149677, + "grad_norm": 0.10694071650505066, + "learning_rate": 1.6214809721103744e-06, + "loss": 0.619, + "step": 1102 + }, + { + "epoch": 0.9197415051073587, + "grad_norm": 0.1112818494439125, + "learning_rate": 1.6207769136965307e-06, + "loss": 0.6229, + "step": 1103 + }, + { + "epoch": 0.9205753595997499, + "grad_norm": 0.10422814637422562, + "learning_rate": 1.6200723542754389e-06, + "loss": 0.5837, + "step": 1104 + }, + { + "epoch": 0.9214092140921409, + "grad_norm": 0.11240324378013611, + "learning_rate": 1.619367294415724e-06, + "loss": 0.7071, + "step": 1105 + }, + { + "epoch": 0.922243068584532, + "grad_norm": 0.10819468647241592, + "learning_rate": 1.6186617346864151e-06, + "loss": 0.5728, + "step": 1106 + }, + { + "epoch": 0.9230769230769231, + "grad_norm": 0.1053207591176033, + "learning_rate": 1.6179556756569448e-06, + "loss": 0.5915, + "step": 1107 + }, + { + "epoch": 0.9239107775693142, + "grad_norm": 0.1075674444437027, + "learning_rate": 1.6172491178971482e-06, + "loss": 0.5935, + "step": 1108 + }, + { + "epoch": 0.9247446320617052, + "grad_norm": 0.11159633845090866, + "learning_rate": 1.6165420619772635e-06, + "loss": 0.6493, + "step": 1109 + }, + { + "epoch": 0.9255784865540964, + "grad_norm": 0.11671009659767151, + "learning_rate": 1.6158345084679307e-06, + "loss": 0.6661, + "step": 1110 + }, + { + "epoch": 0.9264123410464874, + "grad_norm": 0.11934785544872284, + "learning_rate": 1.6151264579401917e-06, + "loss": 0.6335, + "step": 1111 + }, + { + "epoch": 0.9272461955388784, + "grad_norm": 0.11291314661502838, + "learning_rate": 1.6144179109654887e-06, + "loss": 0.6413, + "step": 1112 + }, + { + "epoch": 0.9280800500312696, + "grad_norm": 0.10902924090623856, + "learning_rate": 1.6137088681156654e-06, + "loss": 0.6713, + "step": 1113 + }, + { + "epoch": 0.9289139045236606, + "grad_norm": 0.10608460754156113, + "learning_rate": 1.6129993299629651e-06, + "loss": 0.6024, + "step": 1114 + }, + { + "epoch": 0.9297477590160517, + "grad_norm": 0.10624940693378448, + "learning_rate": 1.6122892970800317e-06, + "loss": 0.6128, + "step": 1115 + }, + { + "epoch": 0.9305816135084428, + "grad_norm": 0.11591742187738419, + "learning_rate": 1.6115787700399071e-06, + "loss": 0.6242, + "step": 1116 + }, + { + "epoch": 0.9314154680008339, + "grad_norm": 0.11706274002790451, + "learning_rate": 1.610867749416033e-06, + "loss": 0.6356, + "step": 1117 + }, + { + "epoch": 0.9322493224932249, + "grad_norm": 0.10509736090898514, + "learning_rate": 1.6101562357822491e-06, + "loss": 0.6337, + "step": 1118 + }, + { + "epoch": 0.933083176985616, + "grad_norm": 0.10801702737808228, + "learning_rate": 1.6094442297127935e-06, + "loss": 0.615, + "step": 1119 + }, + { + "epoch": 0.9339170314780071, + "grad_norm": 0.11627451330423355, + "learning_rate": 1.6087317317823007e-06, + "loss": 0.6059, + "step": 1120 + }, + { + "epoch": 0.9347508859703981, + "grad_norm": 0.11316727101802826, + "learning_rate": 1.6080187425658033e-06, + "loss": 0.6092, + "step": 1121 + }, + { + "epoch": 0.9355847404627893, + "grad_norm": 0.10899489372968674, + "learning_rate": 1.6073052626387296e-06, + "loss": 0.5787, + "step": 1122 + }, + { + "epoch": 0.9364185949551803, + "grad_norm": 0.10977214574813843, + "learning_rate": 1.606591292576904e-06, + "loss": 0.6022, + "step": 1123 + }, + { + "epoch": 0.9372524494475714, + "grad_norm": 0.10926050692796707, + "learning_rate": 1.6058768329565469e-06, + "loss": 0.6626, + "step": 1124 + }, + { + "epoch": 0.9380863039399625, + "grad_norm": 0.10779014229774475, + "learning_rate": 1.6051618843542736e-06, + "loss": 0.6381, + "step": 1125 + }, + { + "epoch": 0.9389201584323535, + "grad_norm": 0.11509402096271515, + "learning_rate": 1.6044464473470939e-06, + "loss": 0.5817, + "step": 1126 + }, + { + "epoch": 0.9397540129247446, + "grad_norm": 0.10561665147542953, + "learning_rate": 1.6037305225124121e-06, + "loss": 0.5634, + "step": 1127 + }, + { + "epoch": 0.9405878674171357, + "grad_norm": 0.1081278994679451, + "learning_rate": 1.6030141104280253e-06, + "loss": 0.6437, + "step": 1128 + }, + { + "epoch": 0.9414217219095268, + "grad_norm": 0.10470031946897507, + "learning_rate": 1.6022972116721256e-06, + "loss": 0.5853, + "step": 1129 + }, + { + "epoch": 0.9422555764019178, + "grad_norm": 0.1069476306438446, + "learning_rate": 1.601579826823296e-06, + "loss": 0.5653, + "step": 1130 + }, + { + "epoch": 0.943089430894309, + "grad_norm": 0.1095786914229393, + "learning_rate": 1.600861956460513e-06, + "loss": 0.6191, + "step": 1131 + }, + { + "epoch": 0.9439232853867, + "grad_norm": 0.10863108187913895, + "learning_rate": 1.600143601163144e-06, + "loss": 0.6166, + "step": 1132 + }, + { + "epoch": 0.944757139879091, + "grad_norm": 0.110601507127285, + "learning_rate": 1.5994247615109498e-06, + "loss": 0.5716, + "step": 1133 + }, + { + "epoch": 0.9455909943714822, + "grad_norm": 0.108455128967762, + "learning_rate": 1.598705438084079e-06, + "loss": 0.5888, + "step": 1134 + }, + { + "epoch": 0.9464248488638732, + "grad_norm": 0.11156295984983444, + "learning_rate": 1.5979856314630728e-06, + "loss": 0.6723, + "step": 1135 + }, + { + "epoch": 0.9472587033562644, + "grad_norm": 0.11098117381334305, + "learning_rate": 1.5972653422288622e-06, + "loss": 0.5671, + "step": 1136 + }, + { + "epoch": 0.9480925578486554, + "grad_norm": 0.11285511404275894, + "learning_rate": 1.5965445709627669e-06, + "loss": 0.6437, + "step": 1137 + }, + { + "epoch": 0.9489264123410465, + "grad_norm": 0.10554268956184387, + "learning_rate": 1.5958233182464965e-06, + "loss": 0.6289, + "step": 1138 + }, + { + "epoch": 0.9497602668334376, + "grad_norm": 0.10124699026346207, + "learning_rate": 1.5951015846621483e-06, + "loss": 0.6068, + "step": 1139 + }, + { + "epoch": 0.9505941213258287, + "grad_norm": 0.11128734052181244, + "learning_rate": 1.5943793707922084e-06, + "loss": 0.5907, + "step": 1140 + }, + { + "epoch": 0.9514279758182197, + "grad_norm": 0.11676076054573059, + "learning_rate": 1.5936566772195503e-06, + "loss": 0.602, + "step": 1141 + }, + { + "epoch": 0.9522618303106108, + "grad_norm": 0.11664719134569168, + "learning_rate": 1.5929335045274343e-06, + "loss": 0.6197, + "step": 1142 + }, + { + "epoch": 0.9530956848030019, + "grad_norm": 0.1060841903090477, + "learning_rate": 1.5922098532995083e-06, + "loss": 0.553, + "step": 1143 + }, + { + "epoch": 0.9539295392953929, + "grad_norm": 0.11080587655305862, + "learning_rate": 1.591485724119805e-06, + "loss": 0.6357, + "step": 1144 + }, + { + "epoch": 0.9547633937877841, + "grad_norm": 0.10428017377853394, + "learning_rate": 1.5907611175727442e-06, + "loss": 0.5693, + "step": 1145 + }, + { + "epoch": 0.9555972482801751, + "grad_norm": 0.10886859893798828, + "learning_rate": 1.59003603424313e-06, + "loss": 0.6142, + "step": 1146 + }, + { + "epoch": 0.9564311027725662, + "grad_norm": 0.1091897264122963, + "learning_rate": 1.5893104747161522e-06, + "loss": 0.6234, + "step": 1147 + }, + { + "epoch": 0.9572649572649573, + "grad_norm": 0.11705927550792694, + "learning_rate": 1.5885844395773841e-06, + "loss": 0.6322, + "step": 1148 + }, + { + "epoch": 0.9580988117573483, + "grad_norm": 0.11114629358053207, + "learning_rate": 1.5878579294127831e-06, + "loss": 0.6233, + "step": 1149 + }, + { + "epoch": 0.9589326662497394, + "grad_norm": 0.11006072908639908, + "learning_rate": 1.5871309448086903e-06, + "loss": 0.6271, + "step": 1150 + }, + { + "epoch": 0.9597665207421305, + "grad_norm": 0.10859903693199158, + "learning_rate": 1.5864034863518292e-06, + "loss": 0.6101, + "step": 1151 + }, + { + "epoch": 0.9606003752345216, + "grad_norm": 0.11398789286613464, + "learning_rate": 1.585675554629306e-06, + "loss": 0.6183, + "step": 1152 + }, + { + "epoch": 0.9614342297269126, + "grad_norm": 0.10914972424507141, + "learning_rate": 1.5849471502286087e-06, + "loss": 0.6687, + "step": 1153 + }, + { + "epoch": 0.9622680842193038, + "grad_norm": 0.11407602578401566, + "learning_rate": 1.5842182737376074e-06, + "loss": 0.6513, + "step": 1154 + }, + { + "epoch": 0.9631019387116948, + "grad_norm": 0.11537064611911774, + "learning_rate": 1.5834889257445525e-06, + "loss": 0.6284, + "step": 1155 + }, + { + "epoch": 0.9639357932040858, + "grad_norm": 0.10912200808525085, + "learning_rate": 1.582759106838075e-06, + "loss": 0.634, + "step": 1156 + }, + { + "epoch": 0.964769647696477, + "grad_norm": 0.11783526092767715, + "learning_rate": 1.5820288176071861e-06, + "loss": 0.6435, + "step": 1157 + }, + { + "epoch": 0.965603502188868, + "grad_norm": 0.11490530520677567, + "learning_rate": 1.5812980586412765e-06, + "loss": 0.5872, + "step": 1158 + }, + { + "epoch": 0.9664373566812591, + "grad_norm": 0.10283337533473969, + "learning_rate": 1.5805668305301167e-06, + "loss": 0.565, + "step": 1159 + }, + { + "epoch": 0.9672712111736502, + "grad_norm": 0.10747389495372772, + "learning_rate": 1.5798351338638548e-06, + "loss": 0.5878, + "step": 1160 + }, + { + "epoch": 0.9681050656660413, + "grad_norm": 0.11634895950555801, + "learning_rate": 1.5791029692330172e-06, + "loss": 0.6547, + "step": 1161 + }, + { + "epoch": 0.9689389201584323, + "grad_norm": 0.11201413720846176, + "learning_rate": 1.5783703372285086e-06, + "loss": 0.6039, + "step": 1162 + }, + { + "epoch": 0.9697727746508235, + "grad_norm": 0.11117483675479889, + "learning_rate": 1.5776372384416105e-06, + "loss": 0.6447, + "step": 1163 + }, + { + "epoch": 0.9706066291432145, + "grad_norm": 0.11274691671133041, + "learning_rate": 1.5769036734639815e-06, + "loss": 0.5847, + "step": 1164 + }, + { + "epoch": 0.9714404836356055, + "grad_norm": 0.11471108347177505, + "learning_rate": 1.5761696428876556e-06, + "loss": 0.6523, + "step": 1165 + }, + { + "epoch": 0.9722743381279967, + "grad_norm": 0.11089842766523361, + "learning_rate": 1.5754351473050434e-06, + "loss": 0.5949, + "step": 1166 + }, + { + "epoch": 0.9731081926203877, + "grad_norm": 0.10404518991708755, + "learning_rate": 1.5747001873089306e-06, + "loss": 0.5496, + "step": 1167 + }, + { + "epoch": 0.9739420471127789, + "grad_norm": 0.116578109562397, + "learning_rate": 1.5739647634924773e-06, + "loss": 0.6103, + "step": 1168 + }, + { + "epoch": 0.9747759016051699, + "grad_norm": 0.1054786667227745, + "learning_rate": 1.5732288764492184e-06, + "loss": 0.623, + "step": 1169 + }, + { + "epoch": 0.975609756097561, + "grad_norm": 0.11369701474905014, + "learning_rate": 1.5724925267730624e-06, + "loss": 0.567, + "step": 1170 + }, + { + "epoch": 0.9764436105899521, + "grad_norm": 0.1198234111070633, + "learning_rate": 1.5717557150582917e-06, + "loss": 0.6188, + "step": 1171 + }, + { + "epoch": 0.9772774650823431, + "grad_norm": 0.12100456655025482, + "learning_rate": 1.5710184418995604e-06, + "loss": 0.641, + "step": 1172 + }, + { + "epoch": 0.9781113195747342, + "grad_norm": 0.10783829540014267, + "learning_rate": 1.5702807078918965e-06, + "loss": 0.5889, + "step": 1173 + }, + { + "epoch": 0.9789451740671253, + "grad_norm": 0.11204763501882553, + "learning_rate": 1.5695425136306987e-06, + "loss": 0.6271, + "step": 1174 + }, + { + "epoch": 0.9797790285595164, + "grad_norm": 0.11363273113965988, + "learning_rate": 1.568803859711738e-06, + "loss": 0.6099, + "step": 1175 + }, + { + "epoch": 0.9806128830519074, + "grad_norm": 0.10527540743350983, + "learning_rate": 1.5680647467311555e-06, + "loss": 0.5782, + "step": 1176 + }, + { + "epoch": 0.9814467375442986, + "grad_norm": 0.11106487363576889, + "learning_rate": 1.5673251752854645e-06, + "loss": 0.645, + "step": 1177 + }, + { + "epoch": 0.9822805920366896, + "grad_norm": 0.11001728475093842, + "learning_rate": 1.5665851459715457e-06, + "loss": 0.611, + "step": 1178 + }, + { + "epoch": 0.9831144465290806, + "grad_norm": 0.10988467186689377, + "learning_rate": 1.5658446593866517e-06, + "loss": 0.6229, + "step": 1179 + }, + { + "epoch": 0.9839483010214718, + "grad_norm": 0.10910095274448395, + "learning_rate": 1.565103716128403e-06, + "loss": 0.5964, + "step": 1180 + }, + { + "epoch": 0.9847821555138628, + "grad_norm": 0.11653497070074081, + "learning_rate": 1.5643623167947891e-06, + "loss": 0.6451, + "step": 1181 + }, + { + "epoch": 0.9856160100062539, + "grad_norm": 0.11314839869737625, + "learning_rate": 1.5636204619841667e-06, + "loss": 0.6419, + "step": 1182 + }, + { + "epoch": 0.986449864498645, + "grad_norm": 0.10827460885047913, + "learning_rate": 1.5628781522952611e-06, + "loss": 0.6129, + "step": 1183 + }, + { + "epoch": 0.9872837189910361, + "grad_norm": 0.1053231805562973, + "learning_rate": 1.5621353883271649e-06, + "loss": 0.6455, + "step": 1184 + }, + { + "epoch": 0.9881175734834271, + "grad_norm": 0.11166463792324066, + "learning_rate": 1.561392170679336e-06, + "loss": 0.594, + "step": 1185 + }, + { + "epoch": 0.9889514279758183, + "grad_norm": 0.11850026249885559, + "learning_rate": 1.5606484999516e-06, + "loss": 0.575, + "step": 1186 + }, + { + "epoch": 0.9897852824682093, + "grad_norm": 0.11536535620689392, + "learning_rate": 1.5599043767441471e-06, + "loss": 0.645, + "step": 1187 + }, + { + "epoch": 0.9906191369606003, + "grad_norm": 0.1191059872508049, + "learning_rate": 1.5591598016575328e-06, + "loss": 0.6007, + "step": 1188 + }, + { + "epoch": 0.9914529914529915, + "grad_norm": 0.1131712794303894, + "learning_rate": 1.558414775292678e-06, + "loss": 0.5926, + "step": 1189 + }, + { + "epoch": 0.9922868459453825, + "grad_norm": 0.11502361297607422, + "learning_rate": 1.5576692982508663e-06, + "loss": 0.6223, + "step": 1190 + }, + { + "epoch": 0.9931207004377736, + "grad_norm": 0.10861647129058838, + "learning_rate": 1.5569233711337474e-06, + "loss": 0.5948, + "step": 1191 + }, + { + "epoch": 0.9939545549301647, + "grad_norm": 0.11109884828329086, + "learning_rate": 1.5561769945433323e-06, + "loss": 0.5778, + "step": 1192 + }, + { + "epoch": 0.9947884094225558, + "grad_norm": 0.1124456375837326, + "learning_rate": 1.5554301690819951e-06, + "loss": 0.6732, + "step": 1193 + }, + { + "epoch": 0.9956222639149468, + "grad_norm": 0.10982295870780945, + "learning_rate": 1.5546828953524727e-06, + "loss": 0.6221, + "step": 1194 + }, + { + "epoch": 0.9964561184073379, + "grad_norm": 0.10954145342111588, + "learning_rate": 1.553935173957863e-06, + "loss": 0.6045, + "step": 1195 + }, + { + "epoch": 0.997289972899729, + "grad_norm": 0.10329638421535492, + "learning_rate": 1.5531870055016265e-06, + "loss": 0.5827, + "step": 1196 + }, + { + "epoch": 0.99812382739212, + "grad_norm": 0.1147105023264885, + "learning_rate": 1.5524383905875828e-06, + "loss": 0.6676, + "step": 1197 + }, + { + "epoch": 0.9989576818845112, + "grad_norm": 0.10970206558704376, + "learning_rate": 1.5516893298199133e-06, + "loss": 0.6159, + "step": 1198 + }, + { + "epoch": 0.9997915363769022, + "grad_norm": 0.10848627239465714, + "learning_rate": 1.5509398238031585e-06, + "loss": 0.6106, + "step": 1199 + }, + { + "epoch": 1.0, + "grad_norm": 0.10848627239465714, + "learning_rate": 1.5501898731422187e-06, + "loss": 0.1335, + "step": 1200 + }, + { + "epoch": 1.0, + "eval_loss": 0.6266602873802185, + "eval_runtime": 321.4177, + "eval_samples_per_second": 16.748, + "eval_steps_per_second": 2.794, + "step": 1200 + }, + { + "epoch": 1.0006253908692933, + "grad_norm": 0.10542084276676178, + "learning_rate": 1.5494394784423524e-06, + "loss": 0.4569, + "step": 1201 + }, + { + "epoch": 1.0008338544923912, + "grad_norm": 0.11667848378419876, + "learning_rate": 1.5486886403091765e-06, + "loss": 0.6476, + "step": 1202 + }, + { + "epoch": 1.001667708984782, + "grad_norm": 0.12027855962514877, + "learning_rate": 1.5479373593486666e-06, + "loss": 0.6156, + "step": 1203 + }, + { + "epoch": 1.0025015634771732, + "grad_norm": 0.115195631980896, + "learning_rate": 1.547185636167155e-06, + "loss": 0.603, + "step": 1204 + }, + { + "epoch": 1.0033354179695644, + "grad_norm": 0.11124297231435776, + "learning_rate": 1.5464334713713309e-06, + "loss": 0.6328, + "step": 1205 + }, + { + "epoch": 1.0041692724619553, + "grad_norm": 0.11163129657506943, + "learning_rate": 1.5456808655682405e-06, + "loss": 0.658, + "step": 1206 + }, + { + "epoch": 1.0050031269543465, + "grad_norm": 0.10940606147050858, + "learning_rate": 1.5449278193652853e-06, + "loss": 0.6146, + "step": 1207 + }, + { + "epoch": 1.0058369814467376, + "grad_norm": 0.10748612880706787, + "learning_rate": 1.5441743333702224e-06, + "loss": 0.6022, + "step": 1208 + }, + { + "epoch": 1.0066708359391285, + "grad_norm": 0.11005198210477829, + "learning_rate": 1.5434204081911641e-06, + "loss": 0.595, + "step": 1209 + }, + { + "epoch": 1.0075046904315197, + "grad_norm": 0.11207201331853867, + "learning_rate": 1.5426660444365766e-06, + "loss": 0.6294, + "step": 1210 + }, + { + "epoch": 1.0083385449239108, + "grad_norm": 0.12536509335041046, + "learning_rate": 1.5419112427152807e-06, + "loss": 0.6097, + "step": 1211 + }, + { + "epoch": 1.0091723994163018, + "grad_norm": 0.11734244227409363, + "learning_rate": 1.54115600363645e-06, + "loss": 0.6301, + "step": 1212 + }, + { + "epoch": 1.010006253908693, + "grad_norm": 0.10745465755462646, + "learning_rate": 1.5404003278096118e-06, + "loss": 0.5472, + "step": 1213 + }, + { + "epoch": 1.010840108401084, + "grad_norm": 0.10949809849262238, + "learning_rate": 1.5396442158446456e-06, + "loss": 0.5806, + "step": 1214 + }, + { + "epoch": 1.011673962893475, + "grad_norm": 0.10704012960195541, + "learning_rate": 1.5388876683517825e-06, + "loss": 0.6133, + "step": 1215 + }, + { + "epoch": 1.0125078173858661, + "grad_norm": 0.11208271235227585, + "learning_rate": 1.5381306859416054e-06, + "loss": 0.6766, + "step": 1216 + }, + { + "epoch": 1.0133416718782573, + "grad_norm": 0.11088071763515472, + "learning_rate": 1.5373732692250484e-06, + "loss": 0.6471, + "step": 1217 + }, + { + "epoch": 1.0141755263706482, + "grad_norm": 0.11159610748291016, + "learning_rate": 1.536615418813396e-06, + "loss": 0.5996, + "step": 1218 + }, + { + "epoch": 1.0150093808630394, + "grad_norm": 0.11161819100379944, + "learning_rate": 1.5358571353182824e-06, + "loss": 0.6552, + "step": 1219 + }, + { + "epoch": 1.0158432353554305, + "grad_norm": 0.1059797927737236, + "learning_rate": 1.5350984193516917e-06, + "loss": 0.663, + "step": 1220 + }, + { + "epoch": 1.0166770898478215, + "grad_norm": 0.11861960589885712, + "learning_rate": 1.5343392715259572e-06, + "loss": 0.6603, + "step": 1221 + }, + { + "epoch": 1.0175109443402126, + "grad_norm": 0.1215897724032402, + "learning_rate": 1.5335796924537598e-06, + "loss": 0.6286, + "step": 1222 + }, + { + "epoch": 1.0183447988326038, + "grad_norm": 0.10159210860729218, + "learning_rate": 1.53281968274813e-06, + "loss": 0.571, + "step": 1223 + }, + { + "epoch": 1.0191786533249947, + "grad_norm": 0.11154468357563019, + "learning_rate": 1.5320592430224442e-06, + "loss": 0.5887, + "step": 1224 + }, + { + "epoch": 1.0200125078173858, + "grad_norm": 0.11331478506326675, + "learning_rate": 1.5312983738904266e-06, + "loss": 0.5954, + "step": 1225 + }, + { + "epoch": 1.020846362309777, + "grad_norm": 0.11255330592393875, + "learning_rate": 1.5305370759661481e-06, + "loss": 0.5458, + "step": 1226 + }, + { + "epoch": 1.021680216802168, + "grad_norm": 0.10835165530443192, + "learning_rate": 1.529775349864026e-06, + "loss": 0.6382, + "step": 1227 + }, + { + "epoch": 1.022514071294559, + "grad_norm": 0.11464633047580719, + "learning_rate": 1.5290131961988225e-06, + "loss": 0.6012, + "step": 1228 + }, + { + "epoch": 1.0233479257869502, + "grad_norm": 0.11286839097738266, + "learning_rate": 1.528250615585644e-06, + "loss": 0.6138, + "step": 1229 + }, + { + "epoch": 1.0241817802793411, + "grad_norm": 0.1153862252831459, + "learning_rate": 1.5274876086399441e-06, + "loss": 0.6682, + "step": 1230 + }, + { + "epoch": 1.0250156347717323, + "grad_norm": 0.11663507670164108, + "learning_rate": 1.526724175977518e-06, + "loss": 0.6318, + "step": 1231 + }, + { + "epoch": 1.0258494892641234, + "grad_norm": 0.1101766973733902, + "learning_rate": 1.5259603182145057e-06, + "loss": 0.6107, + "step": 1232 + }, + { + "epoch": 1.0266833437565146, + "grad_norm": 0.11496511846780777, + "learning_rate": 1.5251960359673897e-06, + "loss": 0.6275, + "step": 1233 + }, + { + "epoch": 1.0275171982489055, + "grad_norm": 0.10905829817056656, + "learning_rate": 1.5244313298529959e-06, + "loss": 0.623, + "step": 1234 + }, + { + "epoch": 1.0283510527412967, + "grad_norm": 0.10732358694076538, + "learning_rate": 1.523666200488491e-06, + "loss": 0.5313, + "step": 1235 + }, + { + "epoch": 1.0291849072336878, + "grad_norm": 0.11587117612361908, + "learning_rate": 1.5229006484913847e-06, + "loss": 0.6362, + "step": 1236 + }, + { + "epoch": 1.0300187617260788, + "grad_norm": 0.10961401462554932, + "learning_rate": 1.5221346744795267e-06, + "loss": 0.6235, + "step": 1237 + }, + { + "epoch": 1.03085261621847, + "grad_norm": 0.11670937389135361, + "learning_rate": 1.521368279071108e-06, + "loss": 0.5856, + "step": 1238 + }, + { + "epoch": 1.031686470710861, + "grad_norm": 0.11099621653556824, + "learning_rate": 1.5206014628846593e-06, + "loss": 0.5848, + "step": 1239 + }, + { + "epoch": 1.032520325203252, + "grad_norm": 0.11106451600790024, + "learning_rate": 1.5198342265390508e-06, + "loss": 0.6128, + "step": 1240 + }, + { + "epoch": 1.0333541796956431, + "grad_norm": 0.11923278123140335, + "learning_rate": 1.5190665706534923e-06, + "loss": 0.5776, + "step": 1241 + }, + { + "epoch": 1.0341880341880343, + "grad_norm": 0.11535177379846573, + "learning_rate": 1.5182984958475321e-06, + "loss": 0.6155, + "step": 1242 + }, + { + "epoch": 1.0350218886804252, + "grad_norm": 0.10793700069189072, + "learning_rate": 1.5175300027410565e-06, + "loss": 0.5962, + "step": 1243 + }, + { + "epoch": 1.0358557431728164, + "grad_norm": 0.11252836883068085, + "learning_rate": 1.5167610919542883e-06, + "loss": 0.5944, + "step": 1244 + }, + { + "epoch": 1.0366895976652075, + "grad_norm": 0.11895468831062317, + "learning_rate": 1.5159917641077893e-06, + "loss": 0.6314, + "step": 1245 + }, + { + "epoch": 1.0375234521575984, + "grad_norm": 0.1106170117855072, + "learning_rate": 1.515222019822457e-06, + "loss": 0.6092, + "step": 1246 + }, + { + "epoch": 1.0383573066499896, + "grad_norm": 0.11284390091896057, + "learning_rate": 1.5144518597195242e-06, + "loss": 0.5649, + "step": 1247 + }, + { + "epoch": 1.0391911611423807, + "grad_norm": 0.1112569123506546, + "learning_rate": 1.5136812844205608e-06, + "loss": 0.6286, + "step": 1248 + }, + { + "epoch": 1.0400250156347717, + "grad_norm": 0.12193046510219574, + "learning_rate": 1.512910294547471e-06, + "loss": 0.6077, + "step": 1249 + }, + { + "epoch": 1.0408588701271628, + "grad_norm": 0.11813755333423615, + "learning_rate": 1.5121388907224936e-06, + "loss": 0.6649, + "step": 1250 + }, + { + "epoch": 1.041692724619554, + "grad_norm": 0.11095897108316422, + "learning_rate": 1.511367073568201e-06, + "loss": 0.5702, + "step": 1251 + }, + { + "epoch": 1.042526579111945, + "grad_norm": 0.10681331902742386, + "learning_rate": 1.5105948437075008e-06, + "loss": 0.5554, + "step": 1252 + }, + { + "epoch": 1.043360433604336, + "grad_norm": 0.1169658675789833, + "learning_rate": 1.5098222017636316e-06, + "loss": 0.614, + "step": 1253 + }, + { + "epoch": 1.0441942880967272, + "grad_norm": 0.11909131705760956, + "learning_rate": 1.509049148360166e-06, + "loss": 0.6348, + "step": 1254 + }, + { + "epoch": 1.0450281425891181, + "grad_norm": 0.11616687476634979, + "learning_rate": 1.5082756841210085e-06, + "loss": 0.6457, + "step": 1255 + }, + { + "epoch": 1.0458619970815093, + "grad_norm": 0.10854938626289368, + "learning_rate": 1.5075018096703945e-06, + "loss": 0.6224, + "step": 1256 + }, + { + "epoch": 1.0466958515739004, + "grad_norm": 0.1108667403459549, + "learning_rate": 1.506727525632891e-06, + "loss": 0.6122, + "step": 1257 + }, + { + "epoch": 1.0475297060662914, + "grad_norm": 0.11282949894666672, + "learning_rate": 1.5059528326333952e-06, + "loss": 0.6331, + "step": 1258 + }, + { + "epoch": 1.0483635605586825, + "grad_norm": 0.10732194781303406, + "learning_rate": 1.5051777312971354e-06, + "loss": 0.6162, + "step": 1259 + }, + { + "epoch": 1.0491974150510737, + "grad_norm": 0.11677113175392151, + "learning_rate": 1.5044022222496682e-06, + "loss": 0.6325, + "step": 1260 + }, + { + "epoch": 1.0500312695434646, + "grad_norm": 0.12134989351034164, + "learning_rate": 1.5036263061168796e-06, + "loss": 0.6091, + "step": 1261 + }, + { + "epoch": 1.0508651240358557, + "grad_norm": 0.11596612632274628, + "learning_rate": 1.5028499835249842e-06, + "loss": 0.6314, + "step": 1262 + }, + { + "epoch": 1.051698978528247, + "grad_norm": 0.11390900611877441, + "learning_rate": 1.5020732551005249e-06, + "loss": 0.627, + "step": 1263 + }, + { + "epoch": 1.0525328330206378, + "grad_norm": 0.11059928685426712, + "learning_rate": 1.501296121470372e-06, + "loss": 0.6124, + "step": 1264 + }, + { + "epoch": 1.053366687513029, + "grad_norm": 0.1167077049612999, + "learning_rate": 1.5005185832617228e-06, + "loss": 0.6523, + "step": 1265 + }, + { + "epoch": 1.0542005420054201, + "grad_norm": 0.11124616861343384, + "learning_rate": 1.4997406411021007e-06, + "loss": 0.5474, + "step": 1266 + }, + { + "epoch": 1.055034396497811, + "grad_norm": 0.11309702694416046, + "learning_rate": 1.4989622956193559e-06, + "loss": 0.5877, + "step": 1267 + }, + { + "epoch": 1.0558682509902022, + "grad_norm": 0.11402348428964615, + "learning_rate": 1.4981835474416638e-06, + "loss": 0.6443, + "step": 1268 + }, + { + "epoch": 1.0567021054825934, + "grad_norm": 0.12587358057498932, + "learning_rate": 1.497404397197524e-06, + "loss": 0.6602, + "step": 1269 + }, + { + "epoch": 1.0575359599749843, + "grad_norm": 0.10854615271091461, + "learning_rate": 1.496624845515762e-06, + "loss": 0.6295, + "step": 1270 + }, + { + "epoch": 1.0583698144673754, + "grad_norm": 0.11264639347791672, + "learning_rate": 1.4958448930255262e-06, + "loss": 0.616, + "step": 1271 + }, + { + "epoch": 1.0592036689597666, + "grad_norm": 0.11427608877420425, + "learning_rate": 1.4950645403562896e-06, + "loss": 0.5909, + "step": 1272 + }, + { + "epoch": 1.0600375234521575, + "grad_norm": 0.11705349385738373, + "learning_rate": 1.4942837881378462e-06, + "loss": 0.6587, + "step": 1273 + }, + { + "epoch": 1.0608713779445487, + "grad_norm": 0.11993349343538284, + "learning_rate": 1.4935026370003148e-06, + "loss": 0.607, + "step": 1274 + }, + { + "epoch": 1.0617052324369398, + "grad_norm": 0.13291329145431519, + "learning_rate": 1.4927210875741345e-06, + "loss": 0.716, + "step": 1275 + }, + { + "epoch": 1.0625390869293307, + "grad_norm": 0.11170666664838791, + "learning_rate": 1.4919391404900665e-06, + "loss": 0.6195, + "step": 1276 + }, + { + "epoch": 1.063372941421722, + "grad_norm": 0.10838773101568222, + "learning_rate": 1.4911567963791927e-06, + "loss": 0.6319, + "step": 1277 + }, + { + "epoch": 1.064206795914113, + "grad_norm": 0.10962139070034027, + "learning_rate": 1.490374055872916e-06, + "loss": 0.6101, + "step": 1278 + }, + { + "epoch": 1.065040650406504, + "grad_norm": 0.11577238142490387, + "learning_rate": 1.4895909196029584e-06, + "loss": 0.6285, + "step": 1279 + }, + { + "epoch": 1.0658745048988951, + "grad_norm": 0.11450095474720001, + "learning_rate": 1.4888073882013615e-06, + "loss": 0.63, + "step": 1280 + }, + { + "epoch": 1.0667083593912863, + "grad_norm": 0.11657418310642242, + "learning_rate": 1.4880234623004866e-06, + "loss": 0.638, + "step": 1281 + }, + { + "epoch": 1.0675422138836772, + "grad_norm": 0.11416659504175186, + "learning_rate": 1.4872391425330122e-06, + "loss": 0.6171, + "step": 1282 + }, + { + "epoch": 1.0683760683760684, + "grad_norm": 0.11568942666053772, + "learning_rate": 1.4864544295319355e-06, + "loss": 0.617, + "step": 1283 + }, + { + "epoch": 1.0692099228684595, + "grad_norm": 0.10842934995889664, + "learning_rate": 1.485669323930571e-06, + "loss": 0.5741, + "step": 1284 + }, + { + "epoch": 1.0700437773608504, + "grad_norm": 0.11735663563013077, + "learning_rate": 1.4848838263625493e-06, + "loss": 0.6143, + "step": 1285 + }, + { + "epoch": 1.0708776318532416, + "grad_norm": 0.11187799274921417, + "learning_rate": 1.4840979374618186e-06, + "loss": 0.5643, + "step": 1286 + }, + { + "epoch": 1.0717114863456327, + "grad_norm": 0.11164343357086182, + "learning_rate": 1.4833116578626415e-06, + "loss": 0.596, + "step": 1287 + }, + { + "epoch": 1.0725453408380237, + "grad_norm": 0.11518281698226929, + "learning_rate": 1.4825249881995972e-06, + "loss": 0.5958, + "step": 1288 + }, + { + "epoch": 1.0733791953304148, + "grad_norm": 0.12009905278682709, + "learning_rate": 1.481737929107579e-06, + "loss": 0.6025, + "step": 1289 + }, + { + "epoch": 1.074213049822806, + "grad_norm": 0.11446750909090042, + "learning_rate": 1.4809504812217949e-06, + "loss": 0.6771, + "step": 1290 + }, + { + "epoch": 1.075046904315197, + "grad_norm": 0.11930940300226212, + "learning_rate": 1.4801626451777655e-06, + "loss": 0.5431, + "step": 1291 + }, + { + "epoch": 1.075880758807588, + "grad_norm": 0.11522958427667618, + "learning_rate": 1.4793744216113266e-06, + "loss": 0.5958, + "step": 1292 + }, + { + "epoch": 1.0767146132999792, + "grad_norm": 0.10718630254268646, + "learning_rate": 1.4785858111586256e-06, + "loss": 0.6057, + "step": 1293 + }, + { + "epoch": 1.0775484677923703, + "grad_norm": 0.10954626649618149, + "learning_rate": 1.477796814456122e-06, + "loss": 0.5653, + "step": 1294 + }, + { + "epoch": 1.0783823222847613, + "grad_norm": 0.11433891206979752, + "learning_rate": 1.4770074321405876e-06, + "loss": 0.6031, + "step": 1295 + }, + { + "epoch": 1.0792161767771524, + "grad_norm": 0.11025510728359222, + "learning_rate": 1.4762176648491049e-06, + "loss": 0.6165, + "step": 1296 + }, + { + "epoch": 1.0800500312695434, + "grad_norm": 0.11881809681653976, + "learning_rate": 1.4754275132190678e-06, + "loss": 0.6047, + "step": 1297 + }, + { + "epoch": 1.0808838857619345, + "grad_norm": 0.12111616879701614, + "learning_rate": 1.474636977888179e-06, + "loss": 0.585, + "step": 1298 + }, + { + "epoch": 1.0817177402543257, + "grad_norm": 0.1101752445101738, + "learning_rate": 1.4738460594944528e-06, + "loss": 0.6288, + "step": 1299 + }, + { + "epoch": 1.0825515947467168, + "grad_norm": 0.11202886700630188, + "learning_rate": 1.4730547586762113e-06, + "loss": 0.6521, + "step": 1300 + }, + { + "epoch": 1.0833854492391077, + "grad_norm": 0.1188650131225586, + "learning_rate": 1.4722630760720856e-06, + "loss": 0.5981, + "step": 1301 + }, + { + "epoch": 1.0842193037314989, + "grad_norm": 0.11928634345531464, + "learning_rate": 1.4714710123210146e-06, + "loss": 0.6131, + "step": 1302 + }, + { + "epoch": 1.08505315822389, + "grad_norm": 0.11217010766267776, + "learning_rate": 1.470678568062246e-06, + "loss": 0.6369, + "step": 1303 + }, + { + "epoch": 1.085887012716281, + "grad_norm": 0.11031745374202728, + "learning_rate": 1.469885743935333e-06, + "loss": 0.5831, + "step": 1304 + }, + { + "epoch": 1.0867208672086721, + "grad_norm": 0.11733289062976837, + "learning_rate": 1.469092540580136e-06, + "loss": 0.6509, + "step": 1305 + }, + { + "epoch": 1.0875547217010633, + "grad_norm": 0.1268366575241089, + "learning_rate": 1.4682989586368218e-06, + "loss": 0.6563, + "step": 1306 + }, + { + "epoch": 1.0883885761934542, + "grad_norm": 0.11494706571102142, + "learning_rate": 1.4675049987458628e-06, + "loss": 0.6659, + "step": 1307 + }, + { + "epoch": 1.0892224306858453, + "grad_norm": 0.11928484588861465, + "learning_rate": 1.4667106615480361e-06, + "loss": 0.63, + "step": 1308 + }, + { + "epoch": 1.0900562851782365, + "grad_norm": 0.11468034982681274, + "learning_rate": 1.465915947684423e-06, + "loss": 0.6123, + "step": 1309 + }, + { + "epoch": 1.0908901396706274, + "grad_norm": 0.118593730032444, + "learning_rate": 1.4651208577964095e-06, + "loss": 0.6211, + "step": 1310 + }, + { + "epoch": 1.0917239941630186, + "grad_norm": 0.11647184193134308, + "learning_rate": 1.4643253925256844e-06, + "loss": 0.5772, + "step": 1311 + }, + { + "epoch": 1.0925578486554097, + "grad_norm": 0.11724021285772324, + "learning_rate": 1.4635295525142404e-06, + "loss": 0.6618, + "step": 1312 + }, + { + "epoch": 1.0933917031478007, + "grad_norm": 0.10851619392633438, + "learning_rate": 1.4627333384043712e-06, + "loss": 0.6384, + "step": 1313 + }, + { + "epoch": 1.0942255576401918, + "grad_norm": 0.11257562786340714, + "learning_rate": 1.4619367508386734e-06, + "loss": 0.6153, + "step": 1314 + }, + { + "epoch": 1.095059412132583, + "grad_norm": 0.10816484689712524, + "learning_rate": 1.4611397904600456e-06, + "loss": 0.5628, + "step": 1315 + }, + { + "epoch": 1.0958932666249739, + "grad_norm": 0.11537015438079834, + "learning_rate": 1.4603424579116859e-06, + "loss": 0.5909, + "step": 1316 + }, + { + "epoch": 1.096727121117365, + "grad_norm": 0.11379721760749817, + "learning_rate": 1.4595447538370932e-06, + "loss": 0.6503, + "step": 1317 + }, + { + "epoch": 1.0975609756097562, + "grad_norm": 0.1122441440820694, + "learning_rate": 1.4587466788800673e-06, + "loss": 0.6081, + "step": 1318 + }, + { + "epoch": 1.0983948301021471, + "grad_norm": 0.10928847640752792, + "learning_rate": 1.4579482336847059e-06, + "loss": 0.6015, + "step": 1319 + }, + { + "epoch": 1.0992286845945383, + "grad_norm": 0.1118772029876709, + "learning_rate": 1.4571494188954056e-06, + "loss": 0.6146, + "step": 1320 + }, + { + "epoch": 1.1000625390869294, + "grad_norm": 0.12636762857437134, + "learning_rate": 1.4563502351568624e-06, + "loss": 0.6321, + "step": 1321 + }, + { + "epoch": 1.1008963935793203, + "grad_norm": 0.11796654760837555, + "learning_rate": 1.4555506831140695e-06, + "loss": 0.6043, + "step": 1322 + }, + { + "epoch": 1.1017302480717115, + "grad_norm": 0.10846695303916931, + "learning_rate": 1.4547507634123175e-06, + "loss": 0.5693, + "step": 1323 + }, + { + "epoch": 1.1025641025641026, + "grad_norm": 0.11494018882513046, + "learning_rate": 1.4539504766971927e-06, + "loss": 0.6297, + "step": 1324 + }, + { + "epoch": 1.1033979570564936, + "grad_norm": 0.11545376479625702, + "learning_rate": 1.4531498236145786e-06, + "loss": 0.6224, + "step": 1325 + }, + { + "epoch": 1.1042318115488847, + "grad_norm": 0.118451327085495, + "learning_rate": 1.4523488048106548e-06, + "loss": 0.6033, + "step": 1326 + }, + { + "epoch": 1.1050656660412759, + "grad_norm": 0.12059949338436127, + "learning_rate": 1.4515474209318947e-06, + "loss": 0.621, + "step": 1327 + }, + { + "epoch": 1.1058995205336668, + "grad_norm": 0.11343257129192352, + "learning_rate": 1.4507456726250678e-06, + "loss": 0.6262, + "step": 1328 + }, + { + "epoch": 1.106733375026058, + "grad_norm": 0.11061260104179382, + "learning_rate": 1.4499435605372365e-06, + "loss": 0.5978, + "step": 1329 + }, + { + "epoch": 1.107567229518449, + "grad_norm": 0.1068677082657814, + "learning_rate": 1.4491410853157575e-06, + "loss": 0.5494, + "step": 1330 + }, + { + "epoch": 1.10840108401084, + "grad_norm": 0.12015094608068466, + "learning_rate": 1.44833824760828e-06, + "loss": 0.6331, + "step": 1331 + }, + { + "epoch": 1.1092349385032312, + "grad_norm": 0.12817353010177612, + "learning_rate": 1.447535048062747e-06, + "loss": 0.6544, + "step": 1332 + }, + { + "epoch": 1.1100687929956223, + "grad_norm": 0.10849379003047943, + "learning_rate": 1.4467314873273916e-06, + "loss": 0.5527, + "step": 1333 + }, + { + "epoch": 1.1109026474880133, + "grad_norm": 0.11941130459308624, + "learning_rate": 1.4459275660507403e-06, + "loss": 0.595, + "step": 1334 + }, + { + "epoch": 1.1117365019804044, + "grad_norm": 0.11218126118183136, + "learning_rate": 1.4451232848816087e-06, + "loss": 0.6288, + "step": 1335 + }, + { + "epoch": 1.1125703564727956, + "grad_norm": 0.11410936713218689, + "learning_rate": 1.444318644469105e-06, + "loss": 0.607, + "step": 1336 + }, + { + "epoch": 1.1134042109651865, + "grad_norm": 0.11492687463760376, + "learning_rate": 1.4435136454626262e-06, + "loss": 0.6503, + "step": 1337 + }, + { + "epoch": 1.1142380654575776, + "grad_norm": 0.11537691950798035, + "learning_rate": 1.4427082885118575e-06, + "loss": 0.6566, + "step": 1338 + }, + { + "epoch": 1.1150719199499688, + "grad_norm": 0.1288568675518036, + "learning_rate": 1.4419025742667757e-06, + "loss": 0.6144, + "step": 1339 + }, + { + "epoch": 1.1159057744423597, + "grad_norm": 0.12686489522457123, + "learning_rate": 1.4410965033776437e-06, + "loss": 0.6014, + "step": 1340 + }, + { + "epoch": 1.1167396289347509, + "grad_norm": 0.11200912296772003, + "learning_rate": 1.440290076495013e-06, + "loss": 0.6042, + "step": 1341 + }, + { + "epoch": 1.117573483427142, + "grad_norm": 0.11744599044322968, + "learning_rate": 1.439483294269722e-06, + "loss": 0.579, + "step": 1342 + }, + { + "epoch": 1.118407337919533, + "grad_norm": 0.11925812810659409, + "learning_rate": 1.4386761573528973e-06, + "loss": 0.6255, + "step": 1343 + }, + { + "epoch": 1.119241192411924, + "grad_norm": 0.12317737191915512, + "learning_rate": 1.4378686663959505e-06, + "loss": 0.5868, + "step": 1344 + }, + { + "epoch": 1.1200750469043153, + "grad_norm": 0.11927246302366257, + "learning_rate": 1.4370608220505788e-06, + "loss": 0.5806, + "step": 1345 + }, + { + "epoch": 1.1209089013967062, + "grad_norm": 0.10501258075237274, + "learning_rate": 1.436252624968765e-06, + "loss": 0.4934, + "step": 1346 + }, + { + "epoch": 1.1217427558890973, + "grad_norm": 0.11004379391670227, + "learning_rate": 1.4354440758027771e-06, + "loss": 0.6332, + "step": 1347 + }, + { + "epoch": 1.1225766103814885, + "grad_norm": 0.11404111236333847, + "learning_rate": 1.4346351752051662e-06, + "loss": 0.6288, + "step": 1348 + }, + { + "epoch": 1.1234104648738794, + "grad_norm": 0.1158967912197113, + "learning_rate": 1.4338259238287678e-06, + "loss": 0.648, + "step": 1349 + }, + { + "epoch": 1.1242443193662706, + "grad_norm": 0.11354929208755493, + "learning_rate": 1.4330163223267002e-06, + "loss": 0.6073, + "step": 1350 + }, + { + "epoch": 1.1250781738586617, + "grad_norm": 0.11661235243082047, + "learning_rate": 1.4322063713523644e-06, + "loss": 0.6168, + "step": 1351 + }, + { + "epoch": 1.1259120283510526, + "grad_norm": 0.11652939021587372, + "learning_rate": 1.4313960715594435e-06, + "loss": 0.6285, + "step": 1352 + }, + { + "epoch": 1.1267458828434438, + "grad_norm": 0.10853532701730728, + "learning_rate": 1.4305854236019017e-06, + "loss": 0.5987, + "step": 1353 + }, + { + "epoch": 1.127579737335835, + "grad_norm": 0.11252254992723465, + "learning_rate": 1.429774428133985e-06, + "loss": 0.5752, + "step": 1354 + }, + { + "epoch": 1.128413591828226, + "grad_norm": 0.12270165979862213, + "learning_rate": 1.4289630858102188e-06, + "loss": 0.6592, + "step": 1355 + }, + { + "epoch": 1.129247446320617, + "grad_norm": 0.11131463944911957, + "learning_rate": 1.4281513972854094e-06, + "loss": 0.623, + "step": 1356 + }, + { + "epoch": 1.1300813008130082, + "grad_norm": 0.11877856403589249, + "learning_rate": 1.4273393632146418e-06, + "loss": 0.64, + "step": 1357 + }, + { + "epoch": 1.130915155305399, + "grad_norm": 0.11009548604488373, + "learning_rate": 1.4265269842532805e-06, + "loss": 0.5545, + "step": 1358 + }, + { + "epoch": 1.1317490097977903, + "grad_norm": 0.11712039262056351, + "learning_rate": 1.425714261056968e-06, + "loss": 0.5737, + "step": 1359 + }, + { + "epoch": 1.1325828642901814, + "grad_norm": 0.12244511395692825, + "learning_rate": 1.4249011942816243e-06, + "loss": 0.6165, + "step": 1360 + }, + { + "epoch": 1.1334167187825726, + "grad_norm": 0.1117769330739975, + "learning_rate": 1.424087784583447e-06, + "loss": 0.6471, + "step": 1361 + }, + { + "epoch": 1.1342505732749635, + "grad_norm": 0.11851704865694046, + "learning_rate": 1.4232740326189113e-06, + "loss": 0.5935, + "step": 1362 + }, + { + "epoch": 1.1350844277673546, + "grad_norm": 0.11291871964931488, + "learning_rate": 1.4224599390447672e-06, + "loss": 0.6288, + "step": 1363 + }, + { + "epoch": 1.1359182822597456, + "grad_norm": 0.12204591929912567, + "learning_rate": 1.421645504518041e-06, + "loss": 0.5833, + "step": 1364 + }, + { + "epoch": 1.1367521367521367, + "grad_norm": 0.1169772669672966, + "learning_rate": 1.4208307296960343e-06, + "loss": 0.5812, + "step": 1365 + }, + { + "epoch": 1.1375859912445279, + "grad_norm": 0.11904870718717575, + "learning_rate": 1.4200156152363236e-06, + "loss": 0.6731, + "step": 1366 + }, + { + "epoch": 1.138419845736919, + "grad_norm": 0.11016523838043213, + "learning_rate": 1.4192001617967585e-06, + "loss": 0.6453, + "step": 1367 + }, + { + "epoch": 1.13925370022931, + "grad_norm": 0.12760315835475922, + "learning_rate": 1.4183843700354635e-06, + "loss": 0.6603, + "step": 1368 + }, + { + "epoch": 1.140087554721701, + "grad_norm": 0.12259379029273987, + "learning_rate": 1.417568240610835e-06, + "loss": 0.667, + "step": 1369 + }, + { + "epoch": 1.140921409214092, + "grad_norm": 0.116600401699543, + "learning_rate": 1.416751774181543e-06, + "loss": 0.5901, + "step": 1370 + }, + { + "epoch": 1.1417552637064832, + "grad_norm": 0.11861781775951385, + "learning_rate": 1.415934971406528e-06, + "loss": 0.5813, + "step": 1371 + }, + { + "epoch": 1.1425891181988743, + "grad_norm": 0.1124439537525177, + "learning_rate": 1.415117832945003e-06, + "loss": 0.5673, + "step": 1372 + }, + { + "epoch": 1.1434229726912655, + "grad_norm": 0.11245383322238922, + "learning_rate": 1.4143003594564526e-06, + "loss": 0.5528, + "step": 1373 + }, + { + "epoch": 1.1442568271836564, + "grad_norm": 0.12306279689073563, + "learning_rate": 1.4134825516006307e-06, + "loss": 0.6384, + "step": 1374 + }, + { + "epoch": 1.1450906816760476, + "grad_norm": 0.11250312626361847, + "learning_rate": 1.4126644100375603e-06, + "loss": 0.6299, + "step": 1375 + }, + { + "epoch": 1.1459245361684387, + "grad_norm": 0.12479723989963531, + "learning_rate": 1.411845935427536e-06, + "loss": 0.6302, + "step": 1376 + }, + { + "epoch": 1.1467583906608296, + "grad_norm": 0.11576981842517853, + "learning_rate": 1.4110271284311189e-06, + "loss": 0.6142, + "step": 1377 + }, + { + "epoch": 1.1475922451532208, + "grad_norm": 0.1186005100607872, + "learning_rate": 1.41020798970914e-06, + "loss": 0.6573, + "step": 1378 + }, + { + "epoch": 1.148426099645612, + "grad_norm": 0.12041036039590836, + "learning_rate": 1.4093885199226969e-06, + "loss": 0.5843, + "step": 1379 + }, + { + "epoch": 1.1492599541380029, + "grad_norm": 0.11400541663169861, + "learning_rate": 1.4085687197331558e-06, + "loss": 0.6108, + "step": 1380 + }, + { + "epoch": 1.150093808630394, + "grad_norm": 0.11786925792694092, + "learning_rate": 1.4077485898021477e-06, + "loss": 0.5824, + "step": 1381 + }, + { + "epoch": 1.1509276631227852, + "grad_norm": 0.12048111855983734, + "learning_rate": 1.406928130791571e-06, + "loss": 0.6496, + "step": 1382 + }, + { + "epoch": 1.151761517615176, + "grad_norm": 0.11885000020265579, + "learning_rate": 1.4061073433635898e-06, + "loss": 0.6486, + "step": 1383 + }, + { + "epoch": 1.1525953721075672, + "grad_norm": 0.11726827174425125, + "learning_rate": 1.4052862281806318e-06, + "loss": 0.6299, + "step": 1384 + }, + { + "epoch": 1.1534292265999584, + "grad_norm": 0.11078763753175735, + "learning_rate": 1.4044647859053913e-06, + "loss": 0.608, + "step": 1385 + }, + { + "epoch": 1.1542630810923493, + "grad_norm": 0.11806345731019974, + "learning_rate": 1.4036430172008246e-06, + "loss": 0.5859, + "step": 1386 + }, + { + "epoch": 1.1550969355847405, + "grad_norm": 0.11999361962080002, + "learning_rate": 1.4028209227301532e-06, + "loss": 0.6445, + "step": 1387 + }, + { + "epoch": 1.1559307900771316, + "grad_norm": 0.1165865957736969, + "learning_rate": 1.4019985031568604e-06, + "loss": 0.5671, + "step": 1388 + }, + { + "epoch": 1.1567646445695225, + "grad_norm": 0.12415032833814621, + "learning_rate": 1.4011757591446915e-06, + "loss": 0.6278, + "step": 1389 + }, + { + "epoch": 1.1575984990619137, + "grad_norm": 0.11192337423563004, + "learning_rate": 1.4003526913576554e-06, + "loss": 0.5964, + "step": 1390 + }, + { + "epoch": 1.1584323535543049, + "grad_norm": 0.11573266237974167, + "learning_rate": 1.3995293004600203e-06, + "loss": 0.6284, + "step": 1391 + }, + { + "epoch": 1.1592662080466958, + "grad_norm": 0.10940641909837723, + "learning_rate": 1.3987055871163168e-06, + "loss": 0.6426, + "step": 1392 + }, + { + "epoch": 1.160100062539087, + "grad_norm": 0.11915182322263718, + "learning_rate": 1.3978815519913342e-06, + "loss": 0.57, + "step": 1393 + }, + { + "epoch": 1.160933917031478, + "grad_norm": 0.12485363334417343, + "learning_rate": 1.3970571957501226e-06, + "loss": 0.5625, + "step": 1394 + }, + { + "epoch": 1.161767771523869, + "grad_norm": 0.11141739040613174, + "learning_rate": 1.3962325190579919e-06, + "loss": 0.5876, + "step": 1395 + }, + { + "epoch": 1.1626016260162602, + "grad_norm": 0.11316776275634766, + "learning_rate": 1.3954075225805085e-06, + "loss": 0.5888, + "step": 1396 + }, + { + "epoch": 1.1634354805086513, + "grad_norm": 0.11194372922182083, + "learning_rate": 1.3945822069834982e-06, + "loss": 0.5907, + "step": 1397 + }, + { + "epoch": 1.1642693350010422, + "grad_norm": 0.11734558641910553, + "learning_rate": 1.3937565729330452e-06, + "loss": 0.6379, + "step": 1398 + }, + { + "epoch": 1.1651031894934334, + "grad_norm": 0.11923021823167801, + "learning_rate": 1.3929306210954888e-06, + "loss": 0.624, + "step": 1399 + }, + { + "epoch": 1.1659370439858245, + "grad_norm": 0.11734186112880707, + "learning_rate": 1.392104352137426e-06, + "loss": 0.5796, + "step": 1400 + }, + { + "epoch": 1.1667708984782155, + "grad_norm": 0.11634799838066101, + "learning_rate": 1.3912777667257092e-06, + "loss": 0.5915, + "step": 1401 + }, + { + "epoch": 1.1676047529706066, + "grad_norm": 0.11459887027740479, + "learning_rate": 1.3904508655274472e-06, + "loss": 0.523, + "step": 1402 + }, + { + "epoch": 1.1684386074629978, + "grad_norm": 0.12342048436403275, + "learning_rate": 1.3896236492100025e-06, + "loss": 0.5816, + "step": 1403 + }, + { + "epoch": 1.1692724619553887, + "grad_norm": 0.12243616580963135, + "learning_rate": 1.388796118440992e-06, + "loss": 0.6229, + "step": 1404 + }, + { + "epoch": 1.1701063164477798, + "grad_norm": 0.11432825773954391, + "learning_rate": 1.3879682738882871e-06, + "loss": 0.5549, + "step": 1405 + }, + { + "epoch": 1.170940170940171, + "grad_norm": 0.12253974378108978, + "learning_rate": 1.3871401162200122e-06, + "loss": 0.6121, + "step": 1406 + }, + { + "epoch": 1.171774025432562, + "grad_norm": 0.12205984443426132, + "learning_rate": 1.3863116461045438e-06, + "loss": 0.6503, + "step": 1407 + }, + { + "epoch": 1.172607879924953, + "grad_norm": 0.12476064264774323, + "learning_rate": 1.3854828642105116e-06, + "loss": 0.6309, + "step": 1408 + }, + { + "epoch": 1.1734417344173442, + "grad_norm": 0.12862101197242737, + "learning_rate": 1.384653771206796e-06, + "loss": 0.6598, + "step": 1409 + }, + { + "epoch": 1.1742755889097352, + "grad_norm": 0.11716258525848389, + "learning_rate": 1.3838243677625289e-06, + "loss": 0.619, + "step": 1410 + }, + { + "epoch": 1.1751094434021263, + "grad_norm": 0.12143337726593018, + "learning_rate": 1.3829946545470927e-06, + "loss": 0.6529, + "step": 1411 + }, + { + "epoch": 1.1759432978945175, + "grad_norm": 0.1213587149977684, + "learning_rate": 1.38216463223012e-06, + "loss": 0.642, + "step": 1412 + }, + { + "epoch": 1.1767771523869084, + "grad_norm": 0.11838190257549286, + "learning_rate": 1.3813343014814925e-06, + "loss": 0.6738, + "step": 1413 + }, + { + "epoch": 1.1776110068792995, + "grad_norm": 0.12712059915065765, + "learning_rate": 1.3805036629713406e-06, + "loss": 0.6289, + "step": 1414 + }, + { + "epoch": 1.1784448613716907, + "grad_norm": 0.12272265553474426, + "learning_rate": 1.3796727173700443e-06, + "loss": 0.6046, + "step": 1415 + }, + { + "epoch": 1.1792787158640818, + "grad_norm": 0.11514297872781754, + "learning_rate": 1.3788414653482296e-06, + "loss": 0.5965, + "step": 1416 + }, + { + "epoch": 1.1801125703564728, + "grad_norm": 0.12112387269735336, + "learning_rate": 1.3780099075767719e-06, + "loss": 0.6412, + "step": 1417 + }, + { + "epoch": 1.180946424848864, + "grad_norm": 0.12124162167310715, + "learning_rate": 1.3771780447267916e-06, + "loss": 0.5654, + "step": 1418 + }, + { + "epoch": 1.1817802793412548, + "grad_norm": 0.12093062698841095, + "learning_rate": 1.3763458774696562e-06, + "loss": 0.6209, + "step": 1419 + }, + { + "epoch": 1.182614133833646, + "grad_norm": 0.12662725150585175, + "learning_rate": 1.3755134064769786e-06, + "loss": 0.6165, + "step": 1420 + }, + { + "epoch": 1.1834479883260371, + "grad_norm": 0.12158691138029099, + "learning_rate": 1.3746806324206173e-06, + "loss": 0.5725, + "step": 1421 + }, + { + "epoch": 1.1842818428184283, + "grad_norm": 0.11839036643505096, + "learning_rate": 1.373847555972674e-06, + "loss": 0.627, + "step": 1422 + }, + { + "epoch": 1.1851156973108192, + "grad_norm": 0.1114383339881897, + "learning_rate": 1.3730141778054962e-06, + "loss": 0.6054, + "step": 1423 + }, + { + "epoch": 1.1859495518032104, + "grad_norm": 0.11537707597017288, + "learning_rate": 1.3721804985916743e-06, + "loss": 0.625, + "step": 1424 + }, + { + "epoch": 1.1867834062956013, + "grad_norm": 0.12441827356815338, + "learning_rate": 1.3713465190040414e-06, + "loss": 0.6006, + "step": 1425 + }, + { + "epoch": 1.1876172607879925, + "grad_norm": 0.11192117631435394, + "learning_rate": 1.3705122397156726e-06, + "loss": 0.6492, + "step": 1426 + }, + { + "epoch": 1.1884511152803836, + "grad_norm": 0.12034382671117783, + "learning_rate": 1.369677661399886e-06, + "loss": 0.6147, + "step": 1427 + }, + { + "epoch": 1.1892849697727748, + "grad_norm": 0.12263434380292892, + "learning_rate": 1.3688427847302403e-06, + "loss": 0.6354, + "step": 1428 + }, + { + "epoch": 1.1901188242651657, + "grad_norm": 0.12424664944410324, + "learning_rate": 1.368007610380535e-06, + "loss": 0.593, + "step": 1429 + }, + { + "epoch": 1.1909526787575568, + "grad_norm": 0.11351408064365387, + "learning_rate": 1.3671721390248099e-06, + "loss": 0.5836, + "step": 1430 + }, + { + "epoch": 1.1917865332499478, + "grad_norm": 0.11502416431903839, + "learning_rate": 1.3663363713373453e-06, + "loss": 0.5834, + "step": 1431 + }, + { + "epoch": 1.192620387742339, + "grad_norm": 0.1204412430524826, + "learning_rate": 1.3655003079926593e-06, + "loss": 0.5815, + "step": 1432 + }, + { + "epoch": 1.19345424223473, + "grad_norm": 0.12347093969583511, + "learning_rate": 1.364663949665509e-06, + "loss": 0.6212, + "step": 1433 + }, + { + "epoch": 1.1942880967271212, + "grad_norm": 0.12269889563322067, + "learning_rate": 1.3638272970308906e-06, + "loss": 0.6115, + "step": 1434 + }, + { + "epoch": 1.1951219512195121, + "grad_norm": 0.12062874436378479, + "learning_rate": 1.3629903507640367e-06, + "loss": 0.6147, + "step": 1435 + }, + { + "epoch": 1.1959558057119033, + "grad_norm": 0.12827728688716888, + "learning_rate": 1.362153111540417e-06, + "loss": 0.6462, + "step": 1436 + }, + { + "epoch": 1.1967896602042944, + "grad_norm": 0.12133876234292984, + "learning_rate": 1.3613155800357384e-06, + "loss": 0.6544, + "step": 1437 + }, + { + "epoch": 1.1976235146966854, + "grad_norm": 0.12726116180419922, + "learning_rate": 1.3604777569259427e-06, + "loss": 0.6589, + "step": 1438 + }, + { + "epoch": 1.1984573691890765, + "grad_norm": 0.12063872814178467, + "learning_rate": 1.3596396428872078e-06, + "loss": 0.6368, + "step": 1439 + }, + { + "epoch": 1.1992912236814677, + "grad_norm": 0.11616161465644836, + "learning_rate": 1.3588012385959462e-06, + "loss": 0.5165, + "step": 1440 + }, + { + "epoch": 1.2001250781738586, + "grad_norm": 0.1210310310125351, + "learning_rate": 1.3579625447288043e-06, + "loss": 0.5979, + "step": 1441 + }, + { + "epoch": 1.2009589326662498, + "grad_norm": 0.11908716708421707, + "learning_rate": 1.3571235619626624e-06, + "loss": 0.647, + "step": 1442 + }, + { + "epoch": 1.201792787158641, + "grad_norm": 0.11108589917421341, + "learning_rate": 1.3562842909746342e-06, + "loss": 0.6023, + "step": 1443 + }, + { + "epoch": 1.2026266416510318, + "grad_norm": 0.12645037472248077, + "learning_rate": 1.3554447324420659e-06, + "loss": 0.6239, + "step": 1444 + }, + { + "epoch": 1.203460496143423, + "grad_norm": 0.11542391031980515, + "learning_rate": 1.3546048870425354e-06, + "loss": 0.601, + "step": 1445 + }, + { + "epoch": 1.2042943506358141, + "grad_norm": 0.12430214136838913, + "learning_rate": 1.3537647554538532e-06, + "loss": 0.6483, + "step": 1446 + }, + { + "epoch": 1.205128205128205, + "grad_norm": 0.11422111093997955, + "learning_rate": 1.352924338354059e-06, + "loss": 0.5866, + "step": 1447 + }, + { + "epoch": 1.2059620596205962, + "grad_norm": 0.11492970585823059, + "learning_rate": 1.3520836364214246e-06, + "loss": 0.6163, + "step": 1448 + }, + { + "epoch": 1.2067959141129874, + "grad_norm": 0.11202309280633926, + "learning_rate": 1.3512426503344507e-06, + "loss": 0.6274, + "step": 1449 + }, + { + "epoch": 1.2076297686053783, + "grad_norm": 0.11905910074710846, + "learning_rate": 1.350401380771868e-06, + "loss": 0.6036, + "step": 1450 + }, + { + "epoch": 1.2084636230977694, + "grad_norm": 0.11703864485025406, + "learning_rate": 1.3495598284126348e-06, + "loss": 0.6599, + "step": 1451 + }, + { + "epoch": 1.2092974775901606, + "grad_norm": 0.12923474609851837, + "learning_rate": 1.3487179939359393e-06, + "loss": 0.6587, + "step": 1452 + }, + { + "epoch": 1.2101313320825515, + "grad_norm": 0.12412197142839432, + "learning_rate": 1.3478758780211964e-06, + "loss": 0.6702, + "step": 1453 + }, + { + "epoch": 1.2109651865749427, + "grad_norm": 0.12546297907829285, + "learning_rate": 1.3470334813480482e-06, + "loss": 0.6502, + "step": 1454 + }, + { + "epoch": 1.2117990410673338, + "grad_norm": 0.11630617082118988, + "learning_rate": 1.3461908045963633e-06, + "loss": 0.6105, + "step": 1455 + }, + { + "epoch": 1.2126328955597248, + "grad_norm": 0.11860398203134537, + "learning_rate": 1.3453478484462366e-06, + "loss": 0.6031, + "step": 1456 + }, + { + "epoch": 1.213466750052116, + "grad_norm": 0.12004498392343521, + "learning_rate": 1.3445046135779885e-06, + "loss": 0.5755, + "step": 1457 + }, + { + "epoch": 1.214300604544507, + "grad_norm": 0.11858487129211426, + "learning_rate": 1.3436611006721641e-06, + "loss": 0.6452, + "step": 1458 + }, + { + "epoch": 1.215134459036898, + "grad_norm": 0.12253355234861374, + "learning_rate": 1.342817310409533e-06, + "loss": 0.5958, + "step": 1459 + }, + { + "epoch": 1.2159683135292891, + "grad_norm": 0.1240239366889, + "learning_rate": 1.3419732434710893e-06, + "loss": 0.6035, + "step": 1460 + }, + { + "epoch": 1.2168021680216803, + "grad_norm": 0.10970788449048996, + "learning_rate": 1.3411289005380493e-06, + "loss": 0.5539, + "step": 1461 + }, + { + "epoch": 1.2176360225140712, + "grad_norm": 0.12056974321603775, + "learning_rate": 1.340284282291852e-06, + "loss": 0.6337, + "step": 1462 + }, + { + "epoch": 1.2184698770064624, + "grad_norm": 0.12024622410535812, + "learning_rate": 1.3394393894141602e-06, + "loss": 0.6247, + "step": 1463 + }, + { + "epoch": 1.2193037314988535, + "grad_norm": 0.11444680392742157, + "learning_rate": 1.3385942225868568e-06, + "loss": 0.6319, + "step": 1464 + }, + { + "epoch": 1.2201375859912444, + "grad_norm": 0.1220422014594078, + "learning_rate": 1.3377487824920457e-06, + "loss": 0.5785, + "step": 1465 + }, + { + "epoch": 1.2209714404836356, + "grad_norm": 0.11991899460554123, + "learning_rate": 1.336903069812053e-06, + "loss": 0.5802, + "step": 1466 + }, + { + "epoch": 1.2218052949760267, + "grad_norm": 0.12127123773097992, + "learning_rate": 1.3360570852294226e-06, + "loss": 0.6411, + "step": 1467 + }, + { + "epoch": 1.2226391494684177, + "grad_norm": 0.11670736223459244, + "learning_rate": 1.33521082942692e-06, + "loss": 0.6309, + "step": 1468 + }, + { + "epoch": 1.2234730039608088, + "grad_norm": 0.12089689821004868, + "learning_rate": 1.3343643030875276e-06, + "loss": 0.5576, + "step": 1469 + }, + { + "epoch": 1.2243068584532, + "grad_norm": 0.10762432962656021, + "learning_rate": 1.3335175068944477e-06, + "loss": 0.6016, + "step": 1470 + }, + { + "epoch": 1.225140712945591, + "grad_norm": 0.12298929691314697, + "learning_rate": 1.3326704415310998e-06, + "loss": 0.5854, + "step": 1471 + }, + { + "epoch": 1.225974567437982, + "grad_norm": 0.12037690728902817, + "learning_rate": 1.3318231076811203e-06, + "loss": 0.59, + "step": 1472 + }, + { + "epoch": 1.2268084219303732, + "grad_norm": 0.12309654802083969, + "learning_rate": 1.3309755060283626e-06, + "loss": 0.5617, + "step": 1473 + }, + { + "epoch": 1.2276422764227641, + "grad_norm": 0.11756734549999237, + "learning_rate": 1.3301276372568962e-06, + "loss": 0.6191, + "step": 1474 + }, + { + "epoch": 1.2284761309151553, + "grad_norm": 0.12193117290735245, + "learning_rate": 1.3292795020510065e-06, + "loss": 0.5838, + "step": 1475 + }, + { + "epoch": 1.2293099854075464, + "grad_norm": 0.11809369921684265, + "learning_rate": 1.3284311010951933e-06, + "loss": 0.6189, + "step": 1476 + }, + { + "epoch": 1.2301438398999376, + "grad_norm": 0.11490161716938019, + "learning_rate": 1.3275824350741716e-06, + "loss": 0.5843, + "step": 1477 + }, + { + "epoch": 1.2309776943923285, + "grad_norm": 0.12500688433647156, + "learning_rate": 1.3267335046728698e-06, + "loss": 0.6587, + "step": 1478 + }, + { + "epoch": 1.2318115488847197, + "grad_norm": 0.12077947705984116, + "learning_rate": 1.3258843105764298e-06, + "loss": 0.5794, + "step": 1479 + }, + { + "epoch": 1.2326454033771106, + "grad_norm": 0.1252153217792511, + "learning_rate": 1.3250348534702061e-06, + "loss": 0.6054, + "step": 1480 + }, + { + "epoch": 1.2334792578695017, + "grad_norm": 0.11973953247070312, + "learning_rate": 1.3241851340397657e-06, + "loss": 0.6504, + "step": 1481 + }, + { + "epoch": 1.234313112361893, + "grad_norm": 0.12270364910364151, + "learning_rate": 1.3233351529708875e-06, + "loss": 0.5973, + "step": 1482 + }, + { + "epoch": 1.235146966854284, + "grad_norm": 0.11802547425031662, + "learning_rate": 1.322484910949562e-06, + "loss": 0.5782, + "step": 1483 + }, + { + "epoch": 1.235980821346675, + "grad_norm": 0.11984699219465256, + "learning_rate": 1.3216344086619887e-06, + "loss": 0.6318, + "step": 1484 + }, + { + "epoch": 1.2368146758390661, + "grad_norm": 0.11720497161149979, + "learning_rate": 1.3207836467945784e-06, + "loss": 0.5948, + "step": 1485 + }, + { + "epoch": 1.237648530331457, + "grad_norm": 0.11490039527416229, + "learning_rate": 1.3199326260339514e-06, + "loss": 0.6022, + "step": 1486 + }, + { + "epoch": 1.2384823848238482, + "grad_norm": 0.11624352633953094, + "learning_rate": 1.3190813470669361e-06, + "loss": 0.5964, + "step": 1487 + }, + { + "epoch": 1.2393162393162394, + "grad_norm": 0.11498794704675674, + "learning_rate": 1.3182298105805704e-06, + "loss": 0.614, + "step": 1488 + }, + { + "epoch": 1.2401500938086305, + "grad_norm": 0.11788622289896011, + "learning_rate": 1.3173780172620996e-06, + "loss": 0.56, + "step": 1489 + }, + { + "epoch": 1.2409839483010214, + "grad_norm": 0.11602766811847687, + "learning_rate": 1.316525967798976e-06, + "loss": 0.5926, + "step": 1490 + }, + { + "epoch": 1.2418178027934126, + "grad_norm": 0.11411117762327194, + "learning_rate": 1.3156736628788582e-06, + "loss": 0.6047, + "step": 1491 + }, + { + "epoch": 1.2426516572858035, + "grad_norm": 0.12349656969308853, + "learning_rate": 1.3148211031896127e-06, + "loss": 0.6254, + "step": 1492 + }, + { + "epoch": 1.2434855117781947, + "grad_norm": 0.11581690609455109, + "learning_rate": 1.3139682894193098e-06, + "loss": 0.5767, + "step": 1493 + }, + { + "epoch": 1.2443193662705858, + "grad_norm": 0.12168504297733307, + "learning_rate": 1.3131152222562253e-06, + "loss": 0.6076, + "step": 1494 + }, + { + "epoch": 1.245153220762977, + "grad_norm": 0.12642395496368408, + "learning_rate": 1.3122619023888402e-06, + "loss": 0.5493, + "step": 1495 + }, + { + "epoch": 1.245987075255368, + "grad_norm": 0.11577428877353668, + "learning_rate": 1.3114083305058388e-06, + "loss": 0.5811, + "step": 1496 + }, + { + "epoch": 1.246820929747759, + "grad_norm": 0.12375449389219284, + "learning_rate": 1.3105545072961092e-06, + "loss": 0.6018, + "step": 1497 + }, + { + "epoch": 1.24765478424015, + "grad_norm": 0.12210660427808762, + "learning_rate": 1.309700433448742e-06, + "loss": 0.5596, + "step": 1498 + }, + { + "epoch": 1.2484886387325411, + "grad_norm": 0.11797598749399185, + "learning_rate": 1.3088461096530304e-06, + "loss": 0.6141, + "step": 1499 + }, + { + "epoch": 1.2493224932249323, + "grad_norm": 0.11823847144842148, + "learning_rate": 1.3079915365984688e-06, + "loss": 0.6102, + "step": 1500 + }, + { + "epoch": 1.2501563477173234, + "grad_norm": 0.12179305404424667, + "learning_rate": 1.3071367149747534e-06, + "loss": 0.6268, + "step": 1501 + }, + { + "epoch": 1.2509902022097144, + "grad_norm": 0.12396464496850967, + "learning_rate": 1.3062816454717803e-06, + "loss": 0.5904, + "step": 1502 + }, + { + "epoch": 1.2518240567021055, + "grad_norm": 0.12284725159406662, + "learning_rate": 1.3054263287796462e-06, + "loss": 0.63, + "step": 1503 + }, + { + "epoch": 1.2526579111944964, + "grad_norm": 0.11863099783658981, + "learning_rate": 1.3045707655886478e-06, + "loss": 0.6018, + "step": 1504 + }, + { + "epoch": 1.2534917656868876, + "grad_norm": 0.11341888457536697, + "learning_rate": 1.3037149565892793e-06, + "loss": 0.5945, + "step": 1505 + }, + { + "epoch": 1.2543256201792787, + "grad_norm": 0.11802545189857483, + "learning_rate": 1.3028589024722345e-06, + "loss": 0.6707, + "step": 1506 + }, + { + "epoch": 1.2551594746716699, + "grad_norm": 0.12044145166873932, + "learning_rate": 1.3020026039284044e-06, + "loss": 0.6184, + "step": 1507 + }, + { + "epoch": 1.2559933291640608, + "grad_norm": 0.1108793094754219, + "learning_rate": 1.301146061648878e-06, + "loss": 0.607, + "step": 1508 + }, + { + "epoch": 1.256827183656452, + "grad_norm": 0.11588302254676819, + "learning_rate": 1.3002892763249395e-06, + "loss": 0.6171, + "step": 1509 + }, + { + "epoch": 1.257661038148843, + "grad_norm": 0.1190834641456604, + "learning_rate": 1.2994322486480714e-06, + "loss": 0.7073, + "step": 1510 + }, + { + "epoch": 1.258494892641234, + "grad_norm": 0.11875880509614944, + "learning_rate": 1.2985749793099498e-06, + "loss": 0.5856, + "step": 1511 + }, + { + "epoch": 1.2593287471336252, + "grad_norm": 0.1210557222366333, + "learning_rate": 1.2977174690024473e-06, + "loss": 0.599, + "step": 1512 + }, + { + "epoch": 1.2601626016260163, + "grad_norm": 0.11437001079320908, + "learning_rate": 1.2968597184176298e-06, + "loss": 0.5626, + "step": 1513 + }, + { + "epoch": 1.2609964561184073, + "grad_norm": 0.11603326350450516, + "learning_rate": 1.296001728247758e-06, + "loss": 0.6017, + "step": 1514 + }, + { + "epoch": 1.2618303106107984, + "grad_norm": 0.10780469328165054, + "learning_rate": 1.2951434991852855e-06, + "loss": 0.524, + "step": 1515 + }, + { + "epoch": 1.2626641651031896, + "grad_norm": 0.11808061599731445, + "learning_rate": 1.2942850319228585e-06, + "loss": 0.5907, + "step": 1516 + }, + { + "epoch": 1.2634980195955805, + "grad_norm": 0.12312963604927063, + "learning_rate": 1.2934263271533169e-06, + "loss": 0.6118, + "step": 1517 + }, + { + "epoch": 1.2643318740879717, + "grad_norm": 0.12425033748149872, + "learning_rate": 1.2925673855696899e-06, + "loss": 0.5697, + "step": 1518 + }, + { + "epoch": 1.2651657285803628, + "grad_norm": 0.122237429022789, + "learning_rate": 1.2917082078652e-06, + "loss": 0.6109, + "step": 1519 + }, + { + "epoch": 1.2659995830727537, + "grad_norm": 0.1070125550031662, + "learning_rate": 1.2908487947332586e-06, + "loss": 0.6049, + "step": 1520 + }, + { + "epoch": 1.2668334375651449, + "grad_norm": 0.12275154143571854, + "learning_rate": 1.2899891468674686e-06, + "loss": 0.6722, + "step": 1521 + }, + { + "epoch": 1.267667292057536, + "grad_norm": 0.11539646238088608, + "learning_rate": 1.289129264961621e-06, + "loss": 0.6129, + "step": 1522 + }, + { + "epoch": 1.268501146549927, + "grad_norm": 0.12436021119356155, + "learning_rate": 1.2882691497096967e-06, + "loss": 0.571, + "step": 1523 + }, + { + "epoch": 1.2693350010423181, + "grad_norm": 0.1118353083729744, + "learning_rate": 1.2874088018058644e-06, + "loss": 0.5885, + "step": 1524 + }, + { + "epoch": 1.2701688555347093, + "grad_norm": 0.1265314221382141, + "learning_rate": 1.2865482219444803e-06, + "loss": 0.6319, + "step": 1525 + }, + { + "epoch": 1.2710027100271002, + "grad_norm": 0.11715365201234818, + "learning_rate": 1.2856874108200887e-06, + "loss": 0.592, + "step": 1526 + }, + { + "epoch": 1.2718365645194913, + "grad_norm": 0.12056981027126312, + "learning_rate": 1.28482636912742e-06, + "loss": 0.6248, + "step": 1527 + }, + { + "epoch": 1.2726704190118825, + "grad_norm": 0.1140827089548111, + "learning_rate": 1.2839650975613904e-06, + "loss": 0.6356, + "step": 1528 + }, + { + "epoch": 1.2735042735042734, + "grad_norm": 0.12372983992099762, + "learning_rate": 1.2831035968171024e-06, + "loss": 0.6428, + "step": 1529 + }, + { + "epoch": 1.2743381279966646, + "grad_norm": 0.11513406038284302, + "learning_rate": 1.2822418675898425e-06, + "loss": 0.6471, + "step": 1530 + }, + { + "epoch": 1.2751719824890557, + "grad_norm": 0.12804672122001648, + "learning_rate": 1.2813799105750821e-06, + "loss": 0.627, + "step": 1531 + }, + { + "epoch": 1.2760058369814469, + "grad_norm": 0.12233211100101471, + "learning_rate": 1.280517726468477e-06, + "loss": 0.6289, + "step": 1532 + }, + { + "epoch": 1.2768396914738378, + "grad_norm": 0.11950721591711044, + "learning_rate": 1.2796553159658651e-06, + "loss": 0.6586, + "step": 1533 + }, + { + "epoch": 1.277673545966229, + "grad_norm": 0.11652793735265732, + "learning_rate": 1.2787926797632681e-06, + "loss": 0.5945, + "step": 1534 + }, + { + "epoch": 1.2785074004586199, + "grad_norm": 0.11893709748983383, + "learning_rate": 1.277929818556889e-06, + "loss": 0.5765, + "step": 1535 + }, + { + "epoch": 1.279341254951011, + "grad_norm": 0.12274635583162308, + "learning_rate": 1.277066733043113e-06, + "loss": 0.6151, + "step": 1536 + }, + { + "epoch": 1.2801751094434022, + "grad_norm": 0.12275136262178421, + "learning_rate": 1.2762034239185062e-06, + "loss": 0.6484, + "step": 1537 + }, + { + "epoch": 1.2810089639357933, + "grad_norm": 0.12041904777288437, + "learning_rate": 1.2753398918798149e-06, + "loss": 0.5766, + "step": 1538 + }, + { + "epoch": 1.2818428184281843, + "grad_norm": 0.12134743481874466, + "learning_rate": 1.2744761376239655e-06, + "loss": 0.5816, + "step": 1539 + }, + { + "epoch": 1.2826766729205754, + "grad_norm": 0.1321423202753067, + "learning_rate": 1.2736121618480637e-06, + "loss": 0.6209, + "step": 1540 + }, + { + "epoch": 1.2835105274129663, + "grad_norm": 0.11608442664146423, + "learning_rate": 1.2727479652493942e-06, + "loss": 0.5762, + "step": 1541 + }, + { + "epoch": 1.2843443819053575, + "grad_norm": 0.11591994017362595, + "learning_rate": 1.2718835485254195e-06, + "loss": 0.6242, + "step": 1542 + }, + { + "epoch": 1.2851782363977486, + "grad_norm": 0.12969833612442017, + "learning_rate": 1.2710189123737802e-06, + "loss": 0.6901, + "step": 1543 + }, + { + "epoch": 1.2860120908901398, + "grad_norm": 0.12043966352939606, + "learning_rate": 1.2701540574922937e-06, + "loss": 0.6434, + "step": 1544 + }, + { + "epoch": 1.2868459453825307, + "grad_norm": 0.12212436646223068, + "learning_rate": 1.2692889845789537e-06, + "loss": 0.5581, + "step": 1545 + }, + { + "epoch": 1.2876797998749219, + "grad_norm": 0.11876165866851807, + "learning_rate": 1.2684236943319306e-06, + "loss": 0.5852, + "step": 1546 + }, + { + "epoch": 1.2885136543673128, + "grad_norm": 0.11841654032468796, + "learning_rate": 1.2675581874495695e-06, + "loss": 0.5714, + "step": 1547 + }, + { + "epoch": 1.289347508859704, + "grad_norm": 0.11584402620792389, + "learning_rate": 1.266692464630391e-06, + "loss": 0.6423, + "step": 1548 + }, + { + "epoch": 1.290181363352095, + "grad_norm": 0.1268841028213501, + "learning_rate": 1.265826526573089e-06, + "loss": 0.6126, + "step": 1549 + }, + { + "epoch": 1.2910152178444863, + "grad_norm": 0.1187991052865982, + "learning_rate": 1.2649603739765323e-06, + "loss": 0.6014, + "step": 1550 + }, + { + "epoch": 1.2918490723368772, + "grad_norm": 0.12219411134719849, + "learning_rate": 1.2640940075397616e-06, + "loss": 0.6377, + "step": 1551 + }, + { + "epoch": 1.2926829268292683, + "grad_norm": 0.11915986984968185, + "learning_rate": 1.2632274279619917e-06, + "loss": 0.589, + "step": 1552 + }, + { + "epoch": 1.2935167813216593, + "grad_norm": 0.1265590786933899, + "learning_rate": 1.2623606359426075e-06, + "loss": 0.6224, + "step": 1553 + }, + { + "epoch": 1.2943506358140504, + "grad_norm": 0.12014889717102051, + "learning_rate": 1.2614936321811671e-06, + "loss": 0.6103, + "step": 1554 + }, + { + "epoch": 1.2951844903064416, + "grad_norm": 0.12002978473901749, + "learning_rate": 1.2606264173773986e-06, + "loss": 0.6196, + "step": 1555 + }, + { + "epoch": 1.2960183447988327, + "grad_norm": 0.12422852218151093, + "learning_rate": 1.2597589922312007e-06, + "loss": 0.6181, + "step": 1556 + }, + { + "epoch": 1.2968521992912236, + "grad_norm": 0.12736350297927856, + "learning_rate": 1.2588913574426419e-06, + "loss": 0.5203, + "step": 1557 + }, + { + "epoch": 1.2976860537836148, + "grad_norm": 0.1263066977262497, + "learning_rate": 1.2580235137119593e-06, + "loss": 0.6256, + "step": 1558 + }, + { + "epoch": 1.2985199082760057, + "grad_norm": 0.11893312633037567, + "learning_rate": 1.2571554617395596e-06, + "loss": 0.6223, + "step": 1559 + }, + { + "epoch": 1.2993537627683969, + "grad_norm": 0.1329938918352127, + "learning_rate": 1.2562872022260167e-06, + "loss": 0.6485, + "step": 1560 + }, + { + "epoch": 1.300187617260788, + "grad_norm": 0.1189078688621521, + "learning_rate": 1.2554187358720723e-06, + "loss": 0.5941, + "step": 1561 + }, + { + "epoch": 1.3010214717531792, + "grad_norm": 0.12884029746055603, + "learning_rate": 1.2545500633786356e-06, + "loss": 0.6785, + "step": 1562 + }, + { + "epoch": 1.30185532624557, + "grad_norm": 0.12023568153381348, + "learning_rate": 1.2536811854467815e-06, + "loss": 0.62, + "step": 1563 + }, + { + "epoch": 1.3026891807379612, + "grad_norm": 0.1196177676320076, + "learning_rate": 1.2528121027777505e-06, + "loss": 0.5589, + "step": 1564 + }, + { + "epoch": 1.3035230352303522, + "grad_norm": 0.11578637361526489, + "learning_rate": 1.2519428160729491e-06, + "loss": 0.5884, + "step": 1565 + }, + { + "epoch": 1.3043568897227433, + "grad_norm": 0.12561044096946716, + "learning_rate": 1.2510733260339478e-06, + "loss": 0.5634, + "step": 1566 + }, + { + "epoch": 1.3051907442151345, + "grad_norm": 0.11461557447910309, + "learning_rate": 1.2502036333624815e-06, + "loss": 0.5752, + "step": 1567 + }, + { + "epoch": 1.3060245987075256, + "grad_norm": 0.12718208134174347, + "learning_rate": 1.249333738760449e-06, + "loss": 0.6245, + "step": 1568 + }, + { + "epoch": 1.3068584531999166, + "grad_norm": 0.12742236256599426, + "learning_rate": 1.2484636429299111e-06, + "loss": 0.6428, + "step": 1569 + }, + { + "epoch": 1.3076923076923077, + "grad_norm": 0.11149119585752487, + "learning_rate": 1.247593346573092e-06, + "loss": 0.6066, + "step": 1570 + }, + { + "epoch": 1.3085261621846986, + "grad_norm": 0.1338125318288803, + "learning_rate": 1.2467228503923772e-06, + "loss": 0.5512, + "step": 1571 + }, + { + "epoch": 1.3093600166770898, + "grad_norm": 0.1266402006149292, + "learning_rate": 1.2458521550903138e-06, + "loss": 0.6163, + "step": 1572 + }, + { + "epoch": 1.310193871169481, + "grad_norm": 0.12411841750144958, + "learning_rate": 1.2449812613696094e-06, + "loss": 0.6586, + "step": 1573 + }, + { + "epoch": 1.311027725661872, + "grad_norm": 0.12587663531303406, + "learning_rate": 1.2441101699331311e-06, + "loss": 0.6382, + "step": 1574 + }, + { + "epoch": 1.311861580154263, + "grad_norm": 0.1223389133810997, + "learning_rate": 1.243238881483907e-06, + "loss": 0.6087, + "step": 1575 + }, + { + "epoch": 1.3126954346466542, + "grad_norm": 0.12871743738651276, + "learning_rate": 1.2423673967251227e-06, + "loss": 0.6235, + "step": 1576 + }, + { + "epoch": 1.3135292891390453, + "grad_norm": 0.12110795080661774, + "learning_rate": 1.2414957163601237e-06, + "loss": 0.5741, + "step": 1577 + }, + { + "epoch": 1.3143631436314362, + "grad_norm": 0.11542119085788727, + "learning_rate": 1.240623841092412e-06, + "loss": 0.5549, + "step": 1578 + }, + { + "epoch": 1.3151969981238274, + "grad_norm": 0.11896467208862305, + "learning_rate": 1.2397517716256478e-06, + "loss": 0.6275, + "step": 1579 + }, + { + "epoch": 1.3160308526162185, + "grad_norm": 0.12287244945764542, + "learning_rate": 1.2388795086636479e-06, + "loss": 0.6034, + "step": 1580 + }, + { + "epoch": 1.3168647071086095, + "grad_norm": 0.13043585419654846, + "learning_rate": 1.238007052910385e-06, + "loss": 0.6218, + "step": 1581 + }, + { + "epoch": 1.3176985616010006, + "grad_norm": 0.12255072593688965, + "learning_rate": 1.2371344050699871e-06, + "loss": 0.556, + "step": 1582 + }, + { + "epoch": 1.3185324160933918, + "grad_norm": 0.12159581482410431, + "learning_rate": 1.2362615658467377e-06, + "loss": 0.6142, + "step": 1583 + }, + { + "epoch": 1.3193662705857827, + "grad_norm": 0.11910532414913177, + "learning_rate": 1.2353885359450756e-06, + "loss": 0.6148, + "step": 1584 + }, + { + "epoch": 1.3202001250781739, + "grad_norm": 0.11686382442712784, + "learning_rate": 1.2345153160695917e-06, + "loss": 0.5947, + "step": 1585 + }, + { + "epoch": 1.321033979570565, + "grad_norm": 0.12290305644273758, + "learning_rate": 1.2336419069250308e-06, + "loss": 0.5934, + "step": 1586 + }, + { + "epoch": 1.321867834062956, + "grad_norm": 0.12619248032569885, + "learning_rate": 1.2327683092162918e-06, + "loss": 0.6704, + "step": 1587 + }, + { + "epoch": 1.322701688555347, + "grad_norm": 0.13024446368217468, + "learning_rate": 1.231894523648424e-06, + "loss": 0.5824, + "step": 1588 + }, + { + "epoch": 1.3235355430477382, + "grad_norm": 0.1289645880460739, + "learning_rate": 1.2310205509266292e-06, + "loss": 0.656, + "step": 1589 + }, + { + "epoch": 1.3243693975401292, + "grad_norm": 0.1269690990447998, + "learning_rate": 1.23014639175626e-06, + "loss": 0.6433, + "step": 1590 + }, + { + "epoch": 1.3252032520325203, + "grad_norm": 0.12446628510951996, + "learning_rate": 1.2292720468428199e-06, + "loss": 0.57, + "step": 1591 + }, + { + "epoch": 1.3260371065249115, + "grad_norm": 0.12336724996566772, + "learning_rate": 1.2283975168919617e-06, + "loss": 0.6324, + "step": 1592 + }, + { + "epoch": 1.3268709610173026, + "grad_norm": 0.11296447366476059, + "learning_rate": 1.2275228026094879e-06, + "loss": 0.6097, + "step": 1593 + }, + { + "epoch": 1.3277048155096935, + "grad_norm": 0.12164086103439331, + "learning_rate": 1.2266479047013498e-06, + "loss": 0.5969, + "step": 1594 + }, + { + "epoch": 1.3285386700020847, + "grad_norm": 0.11765522509813309, + "learning_rate": 1.2257728238736467e-06, + "loss": 0.5404, + "step": 1595 + }, + { + "epoch": 1.3293725244944756, + "grad_norm": 0.12349137663841248, + "learning_rate": 1.224897560832625e-06, + "loss": 0.6574, + "step": 1596 + }, + { + "epoch": 1.3302063789868668, + "grad_norm": 0.1171487495303154, + "learning_rate": 1.22402211628468e-06, + "loss": 0.5834, + "step": 1597 + }, + { + "epoch": 1.331040233479258, + "grad_norm": 0.11913489550352097, + "learning_rate": 1.2231464909363512e-06, + "loss": 0.6022, + "step": 1598 + }, + { + "epoch": 1.331874087971649, + "grad_norm": 0.11552874743938446, + "learning_rate": 1.2222706854943255e-06, + "loss": 0.6164, + "step": 1599 + }, + { + "epoch": 1.33270794246404, + "grad_norm": 0.11986060440540314, + "learning_rate": 1.2213947006654345e-06, + "loss": 0.6062, + "step": 1600 + }, + { + "epoch": 1.33270794246404, + "eval_loss": 0.6176180839538574, + "eval_runtime": 321.4896, + "eval_samples_per_second": 16.744, + "eval_steps_per_second": 2.793, + "step": 1600 + }, + { + "epoch": 1.3335417969564312, + "grad_norm": 0.12321524322032928, + "learning_rate": 1.2205185371566554e-06, + "loss": 0.6164, + "step": 1601 + }, + { + "epoch": 1.334375651448822, + "grad_norm": 0.12141884118318558, + "learning_rate": 1.2196421956751083e-06, + "loss": 0.601, + "step": 1602 + }, + { + "epoch": 1.3352095059412132, + "grad_norm": 0.12699675559997559, + "learning_rate": 1.2187656769280577e-06, + "loss": 0.5379, + "step": 1603 + }, + { + "epoch": 1.3360433604336044, + "grad_norm": 0.12791088223457336, + "learning_rate": 1.2178889816229116e-06, + "loss": 0.6292, + "step": 1604 + }, + { + "epoch": 1.3368772149259955, + "grad_norm": 0.12027748674154282, + "learning_rate": 1.2170121104672196e-06, + "loss": 0.564, + "step": 1605 + }, + { + "epoch": 1.3377110694183865, + "grad_norm": 0.11741901189088821, + "learning_rate": 1.2161350641686737e-06, + "loss": 0.5772, + "step": 1606 + }, + { + "epoch": 1.3385449239107776, + "grad_norm": 0.11572231352329254, + "learning_rate": 1.2152578434351072e-06, + "loss": 0.5819, + "step": 1607 + }, + { + "epoch": 1.3393787784031685, + "grad_norm": 0.1255660504102707, + "learning_rate": 1.214380448974494e-06, + "loss": 0.6341, + "step": 1608 + }, + { + "epoch": 1.3402126328955597, + "grad_norm": 0.11974442005157471, + "learning_rate": 1.2135028814949485e-06, + "loss": 0.6233, + "step": 1609 + }, + { + "epoch": 1.3410464873879508, + "grad_norm": 0.12625208497047424, + "learning_rate": 1.212625141704725e-06, + "loss": 0.6231, + "step": 1610 + }, + { + "epoch": 1.341880341880342, + "grad_norm": 0.12386664003133774, + "learning_rate": 1.2117472303122157e-06, + "loss": 0.617, + "step": 1611 + }, + { + "epoch": 1.342714196372733, + "grad_norm": 0.13161849975585938, + "learning_rate": 1.210869148025952e-06, + "loss": 0.654, + "step": 1612 + }, + { + "epoch": 1.343548050865124, + "grad_norm": 0.11914031207561493, + "learning_rate": 1.2099908955546043e-06, + "loss": 0.6282, + "step": 1613 + }, + { + "epoch": 1.344381905357515, + "grad_norm": 0.12541799247264862, + "learning_rate": 1.2091124736069788e-06, + "loss": 0.6133, + "step": 1614 + }, + { + "epoch": 1.3452157598499062, + "grad_norm": 0.12099876254796982, + "learning_rate": 1.2082338828920185e-06, + "loss": 0.5661, + "step": 1615 + }, + { + "epoch": 1.3460496143422973, + "grad_norm": 0.12484125792980194, + "learning_rate": 1.2073551241188037e-06, + "loss": 0.6238, + "step": 1616 + }, + { + "epoch": 1.3468834688346885, + "grad_norm": 0.12073414772748947, + "learning_rate": 1.2064761979965497e-06, + "loss": 0.5491, + "step": 1617 + }, + { + "epoch": 1.3477173233270794, + "grad_norm": 0.12662388384342194, + "learning_rate": 1.2055971052346067e-06, + "loss": 0.6235, + "step": 1618 + }, + { + "epoch": 1.3485511778194705, + "grad_norm": 0.12112808972597122, + "learning_rate": 1.2047178465424594e-06, + "loss": 0.5493, + "step": 1619 + }, + { + "epoch": 1.3493850323118615, + "grad_norm": 0.13022390007972717, + "learning_rate": 1.2038384226297271e-06, + "loss": 0.6531, + "step": 1620 + }, + { + "epoch": 1.3502188868042526, + "grad_norm": 0.12017896771430969, + "learning_rate": 1.202958834206162e-06, + "loss": 0.5819, + "step": 1621 + }, + { + "epoch": 1.3510527412966438, + "grad_norm": 0.12317414581775665, + "learning_rate": 1.2020790819816484e-06, + "loss": 0.5736, + "step": 1622 + }, + { + "epoch": 1.351886595789035, + "grad_norm": 0.12808437645435333, + "learning_rate": 1.2011991666662043e-06, + "loss": 0.6196, + "step": 1623 + }, + { + "epoch": 1.3527204502814258, + "grad_norm": 0.11925622075796127, + "learning_rate": 1.2003190889699776e-06, + "loss": 0.5703, + "step": 1624 + }, + { + "epoch": 1.353554304773817, + "grad_norm": 0.1228049024939537, + "learning_rate": 1.1994388496032485e-06, + "loss": 0.6357, + "step": 1625 + }, + { + "epoch": 1.354388159266208, + "grad_norm": 0.12009157985448837, + "learning_rate": 1.1985584492764273e-06, + "loss": 0.5559, + "step": 1626 + }, + { + "epoch": 1.355222013758599, + "grad_norm": 0.1287962794303894, + "learning_rate": 1.1976778887000543e-06, + "loss": 0.6269, + "step": 1627 + }, + { + "epoch": 1.3560558682509902, + "grad_norm": 0.11851408332586288, + "learning_rate": 1.1967971685847988e-06, + "loss": 0.5966, + "step": 1628 + }, + { + "epoch": 1.3568897227433814, + "grad_norm": 0.12402857095003128, + "learning_rate": 1.1959162896414589e-06, + "loss": 0.5611, + "step": 1629 + }, + { + "epoch": 1.3577235772357723, + "grad_norm": 0.12092249095439911, + "learning_rate": 1.1950352525809617e-06, + "loss": 0.5631, + "step": 1630 + }, + { + "epoch": 1.3585574317281635, + "grad_norm": 0.12502413988113403, + "learning_rate": 1.1941540581143608e-06, + "loss": 0.584, + "step": 1631 + }, + { + "epoch": 1.3593912862205544, + "grad_norm": 0.12327523529529572, + "learning_rate": 1.1932727069528373e-06, + "loss": 0.6027, + "step": 1632 + }, + { + "epoch": 1.3602251407129455, + "grad_norm": 0.11953040957450867, + "learning_rate": 1.1923911998076988e-06, + "loss": 0.6104, + "step": 1633 + }, + { + "epoch": 1.3610589952053367, + "grad_norm": 0.11636977642774582, + "learning_rate": 1.1915095373903788e-06, + "loss": 0.6116, + "step": 1634 + }, + { + "epoch": 1.3618928496977278, + "grad_norm": 0.12088527530431747, + "learning_rate": 1.1906277204124363e-06, + "loss": 0.6055, + "step": 1635 + }, + { + "epoch": 1.3627267041901188, + "grad_norm": 0.119356170296669, + "learning_rate": 1.1897457495855544e-06, + "loss": 0.606, + "step": 1636 + }, + { + "epoch": 1.36356055868251, + "grad_norm": 0.12436091899871826, + "learning_rate": 1.1888636256215412e-06, + "loss": 0.6475, + "step": 1637 + }, + { + "epoch": 1.3643944131749008, + "grad_norm": 0.12362188845872879, + "learning_rate": 1.187981349232328e-06, + "loss": 0.5592, + "step": 1638 + }, + { + "epoch": 1.365228267667292, + "grad_norm": 0.12188667058944702, + "learning_rate": 1.1870989211299684e-06, + "loss": 0.6392, + "step": 1639 + }, + { + "epoch": 1.3660621221596831, + "grad_norm": 0.12498065829277039, + "learning_rate": 1.1862163420266398e-06, + "loss": 0.5989, + "step": 1640 + }, + { + "epoch": 1.3668959766520743, + "grad_norm": 0.11544735729694366, + "learning_rate": 1.1853336126346406e-06, + "loss": 0.6508, + "step": 1641 + }, + { + "epoch": 1.3677298311444652, + "grad_norm": 0.1186600849032402, + "learning_rate": 1.1844507336663906e-06, + "loss": 0.5887, + "step": 1642 + }, + { + "epoch": 1.3685636856368564, + "grad_norm": 0.12236911803483963, + "learning_rate": 1.1835677058344308e-06, + "loss": 0.5881, + "step": 1643 + }, + { + "epoch": 1.3693975401292475, + "grad_norm": 0.11984533816576004, + "learning_rate": 1.1826845298514214e-06, + "loss": 0.5811, + "step": 1644 + }, + { + "epoch": 1.3702313946216385, + "grad_norm": 0.12120435386896133, + "learning_rate": 1.1818012064301432e-06, + "loss": 0.6307, + "step": 1645 + }, + { + "epoch": 1.3710652491140296, + "grad_norm": 0.11709383875131607, + "learning_rate": 1.1809177362834955e-06, + "loss": 0.6239, + "step": 1646 + }, + { + "epoch": 1.3718991036064208, + "grad_norm": 0.13146185874938965, + "learning_rate": 1.1800341201244952e-06, + "loss": 0.5956, + "step": 1647 + }, + { + "epoch": 1.3727329580988117, + "grad_norm": 0.11478230357170105, + "learning_rate": 1.1791503586662793e-06, + "loss": 0.5437, + "step": 1648 + }, + { + "epoch": 1.3735668125912028, + "grad_norm": 0.11662175506353378, + "learning_rate": 1.1782664526220992e-06, + "loss": 0.6086, + "step": 1649 + }, + { + "epoch": 1.374400667083594, + "grad_norm": 0.11829902231693268, + "learning_rate": 1.1773824027053255e-06, + "loss": 0.5965, + "step": 1650 + }, + { + "epoch": 1.375234521575985, + "grad_norm": 0.12727516889572144, + "learning_rate": 1.1764982096294431e-06, + "loss": 0.573, + "step": 1651 + }, + { + "epoch": 1.376068376068376, + "grad_norm": 0.12310794740915298, + "learning_rate": 1.1756138741080536e-06, + "loss": 0.6462, + "step": 1652 + }, + { + "epoch": 1.3769022305607672, + "grad_norm": 0.11653224378824234, + "learning_rate": 1.1747293968548734e-06, + "loss": 0.5782, + "step": 1653 + }, + { + "epoch": 1.3777360850531581, + "grad_norm": 0.12128982692956924, + "learning_rate": 1.1738447785837322e-06, + "loss": 0.6108, + "step": 1654 + }, + { + "epoch": 1.3785699395455493, + "grad_norm": 0.12801648676395416, + "learning_rate": 1.1729600200085752e-06, + "loss": 0.6026, + "step": 1655 + }, + { + "epoch": 1.3794037940379404, + "grad_norm": 0.11884930729866028, + "learning_rate": 1.1720751218434596e-06, + "loss": 0.5554, + "step": 1656 + }, + { + "epoch": 1.3802376485303314, + "grad_norm": 0.12543028593063354, + "learning_rate": 1.1711900848025554e-06, + "loss": 0.6485, + "step": 1657 + }, + { + "epoch": 1.3810715030227225, + "grad_norm": 0.12205296754837036, + "learning_rate": 1.1703049096001452e-06, + "loss": 0.585, + "step": 1658 + }, + { + "epoch": 1.3819053575151137, + "grad_norm": 0.12094337493181229, + "learning_rate": 1.1694195969506229e-06, + "loss": 0.588, + "step": 1659 + }, + { + "epoch": 1.3827392120075048, + "grad_norm": 0.12128522247076035, + "learning_rate": 1.1685341475684934e-06, + "loss": 0.6043, + "step": 1660 + }, + { + "epoch": 1.3835730664998958, + "grad_norm": 0.12067275494337082, + "learning_rate": 1.1676485621683711e-06, + "loss": 0.5896, + "step": 1661 + }, + { + "epoch": 1.384406920992287, + "grad_norm": 0.11607006937265396, + "learning_rate": 1.1667628414649815e-06, + "loss": 0.5454, + "step": 1662 + }, + { + "epoch": 1.3852407754846778, + "grad_norm": 0.11472892016172409, + "learning_rate": 1.1658769861731584e-06, + "loss": 0.5768, + "step": 1663 + }, + { + "epoch": 1.386074629977069, + "grad_norm": 0.12670141458511353, + "learning_rate": 1.1649909970078447e-06, + "loss": 0.6426, + "step": 1664 + }, + { + "epoch": 1.3869084844694601, + "grad_norm": 0.12123548239469528, + "learning_rate": 1.1641048746840912e-06, + "loss": 0.6508, + "step": 1665 + }, + { + "epoch": 1.3877423389618513, + "grad_norm": 0.12310612946748734, + "learning_rate": 1.1632186199170555e-06, + "loss": 0.6414, + "step": 1666 + }, + { + "epoch": 1.3885761934542422, + "grad_norm": 0.11864950507879257, + "learning_rate": 1.1623322334220037e-06, + "loss": 0.5058, + "step": 1667 + }, + { + "epoch": 1.3894100479466334, + "grad_norm": 0.12240031361579895, + "learning_rate": 1.1614457159143064e-06, + "loss": 0.5974, + "step": 1668 + }, + { + "epoch": 1.3902439024390243, + "grad_norm": 0.13276803493499756, + "learning_rate": 1.1605590681094408e-06, + "loss": 0.5648, + "step": 1669 + }, + { + "epoch": 1.3910777569314154, + "grad_norm": 0.12129524350166321, + "learning_rate": 1.1596722907229898e-06, + "loss": 0.6506, + "step": 1670 + }, + { + "epoch": 1.3919116114238066, + "grad_norm": 0.12178556621074677, + "learning_rate": 1.1587853844706396e-06, + "loss": 0.5844, + "step": 1671 + }, + { + "epoch": 1.3927454659161977, + "grad_norm": 0.12161284685134888, + "learning_rate": 1.1578983500681816e-06, + "loss": 0.5732, + "step": 1672 + }, + { + "epoch": 1.3935793204085887, + "grad_norm": 0.12326084077358246, + "learning_rate": 1.1570111882315098e-06, + "loss": 0.5727, + "step": 1673 + }, + { + "epoch": 1.3944131749009798, + "grad_norm": 0.12302473187446594, + "learning_rate": 1.1561238996766216e-06, + "loss": 0.5913, + "step": 1674 + }, + { + "epoch": 1.3952470293933708, + "grad_norm": 0.12436919659376144, + "learning_rate": 1.1552364851196166e-06, + "loss": 0.5877, + "step": 1675 + }, + { + "epoch": 1.396080883885762, + "grad_norm": 0.12807847559452057, + "learning_rate": 1.154348945276695e-06, + "loss": 0.6274, + "step": 1676 + }, + { + "epoch": 1.396914738378153, + "grad_norm": 0.12182601541280746, + "learning_rate": 1.1534612808641602e-06, + "loss": 0.6319, + "step": 1677 + }, + { + "epoch": 1.3977485928705442, + "grad_norm": 0.13187600672245026, + "learning_rate": 1.1525734925984143e-06, + "loss": 0.6148, + "step": 1678 + }, + { + "epoch": 1.3985824473629351, + "grad_norm": 0.11592843383550644, + "learning_rate": 1.1516855811959602e-06, + "loss": 0.5472, + "step": 1679 + }, + { + "epoch": 1.3994163018553263, + "grad_norm": 0.11742573231458664, + "learning_rate": 1.1507975473734e-06, + "loss": 0.5666, + "step": 1680 + }, + { + "epoch": 1.4002501563477172, + "grad_norm": 0.12708871066570282, + "learning_rate": 1.1499093918474347e-06, + "loss": 0.6051, + "step": 1681 + }, + { + "epoch": 1.4010840108401084, + "grad_norm": 0.11694274097681046, + "learning_rate": 1.1490211153348634e-06, + "loss": 0.5253, + "step": 1682 + }, + { + "epoch": 1.4019178653324995, + "grad_norm": 0.12378162145614624, + "learning_rate": 1.1481327185525827e-06, + "loss": 0.6387, + "step": 1683 + }, + { + "epoch": 1.4027517198248907, + "grad_norm": 0.1409406065940857, + "learning_rate": 1.1472442022175867e-06, + "loss": 0.5597, + "step": 1684 + }, + { + "epoch": 1.4035855743172816, + "grad_norm": 0.12476393580436707, + "learning_rate": 1.1463555670469657e-06, + "loss": 0.5394, + "step": 1685 + }, + { + "epoch": 1.4044194288096727, + "grad_norm": 0.13176341354846954, + "learning_rate": 1.1454668137579057e-06, + "loss": 0.6395, + "step": 1686 + }, + { + "epoch": 1.4052532833020637, + "grad_norm": 0.13114789128303528, + "learning_rate": 1.1445779430676883e-06, + "loss": 0.6215, + "step": 1687 + }, + { + "epoch": 1.4060871377944548, + "grad_norm": 0.12079852819442749, + "learning_rate": 1.1436889556936904e-06, + "loss": 0.5878, + "step": 1688 + }, + { + "epoch": 1.406920992286846, + "grad_norm": 0.1296648234128952, + "learning_rate": 1.142799852353382e-06, + "loss": 0.5442, + "step": 1689 + }, + { + "epoch": 1.4077548467792371, + "grad_norm": 0.1254526525735855, + "learning_rate": 1.141910633764327e-06, + "loss": 0.6245, + "step": 1690 + }, + { + "epoch": 1.408588701271628, + "grad_norm": 0.12021210789680481, + "learning_rate": 1.1410213006441827e-06, + "loss": 0.6083, + "step": 1691 + }, + { + "epoch": 1.4094225557640192, + "grad_norm": 0.12678121030330658, + "learning_rate": 1.1401318537106987e-06, + "loss": 0.6064, + "step": 1692 + }, + { + "epoch": 1.4102564102564101, + "grad_norm": 0.12526234984397888, + "learning_rate": 1.1392422936817165e-06, + "loss": 0.6203, + "step": 1693 + }, + { + "epoch": 1.4110902647488013, + "grad_norm": 0.11436313390731812, + "learning_rate": 1.1383526212751686e-06, + "loss": 0.5449, + "step": 1694 + }, + { + "epoch": 1.4119241192411924, + "grad_norm": 0.11527229845523834, + "learning_rate": 1.1374628372090781e-06, + "loss": 0.5821, + "step": 1695 + }, + { + "epoch": 1.4127579737335836, + "grad_norm": 0.12278933078050613, + "learning_rate": 1.1365729422015593e-06, + "loss": 0.6249, + "step": 1696 + }, + { + "epoch": 1.4135918282259745, + "grad_norm": 0.12223772704601288, + "learning_rate": 1.1356829369708144e-06, + "loss": 0.5978, + "step": 1697 + }, + { + "epoch": 1.4144256827183657, + "grad_norm": 0.12515763938426971, + "learning_rate": 1.1347928222351354e-06, + "loss": 0.5633, + "step": 1698 + }, + { + "epoch": 1.4152595372107566, + "grad_norm": 0.12488391250371933, + "learning_rate": 1.1339025987129031e-06, + "loss": 0.6073, + "step": 1699 + }, + { + "epoch": 1.4160933917031477, + "grad_norm": 0.13267357647418976, + "learning_rate": 1.1330122671225855e-06, + "loss": 0.5895, + "step": 1700 + }, + { + "epoch": 1.416927246195539, + "grad_norm": 0.121218241751194, + "learning_rate": 1.1321218281827377e-06, + "loss": 0.5815, + "step": 1701 + }, + { + "epoch": 1.41776110068793, + "grad_norm": 0.11679316312074661, + "learning_rate": 1.1312312826120019e-06, + "loss": 0.6053, + "step": 1702 + }, + { + "epoch": 1.418594955180321, + "grad_norm": 0.12714731693267822, + "learning_rate": 1.1303406311291064e-06, + "loss": 0.5996, + "step": 1703 + }, + { + "epoch": 1.4194288096727121, + "grad_norm": 0.1288938969373703, + "learning_rate": 1.1294498744528645e-06, + "loss": 0.622, + "step": 1704 + }, + { + "epoch": 1.4202626641651033, + "grad_norm": 0.11887186765670776, + "learning_rate": 1.128559013302174e-06, + "loss": 0.5892, + "step": 1705 + }, + { + "epoch": 1.4210965186574942, + "grad_norm": 0.12163923680782318, + "learning_rate": 1.1276680483960188e-06, + "loss": 0.5992, + "step": 1706 + }, + { + "epoch": 1.4219303731498854, + "grad_norm": 0.1262393295764923, + "learning_rate": 1.1267769804534647e-06, + "loss": 0.6356, + "step": 1707 + }, + { + "epoch": 1.4227642276422765, + "grad_norm": 0.12177680432796478, + "learning_rate": 1.1258858101936615e-06, + "loss": 0.5822, + "step": 1708 + }, + { + "epoch": 1.4235980821346674, + "grad_norm": 0.1219385415315628, + "learning_rate": 1.1249945383358413e-06, + "loss": 0.5946, + "step": 1709 + }, + { + "epoch": 1.4244319366270586, + "grad_norm": 0.12739478051662445, + "learning_rate": 1.1241031655993187e-06, + "loss": 0.5985, + "step": 1710 + }, + { + "epoch": 1.4252657911194497, + "grad_norm": 0.12883968651294708, + "learning_rate": 1.1232116927034892e-06, + "loss": 0.6438, + "step": 1711 + }, + { + "epoch": 1.4260996456118407, + "grad_norm": 0.1276615411043167, + "learning_rate": 1.1223201203678287e-06, + "loss": 0.6173, + "step": 1712 + }, + { + "epoch": 1.4269335001042318, + "grad_norm": 0.11872417479753494, + "learning_rate": 1.1214284493118946e-06, + "loss": 0.5831, + "step": 1713 + }, + { + "epoch": 1.427767354596623, + "grad_norm": 0.12630686163902283, + "learning_rate": 1.1205366802553228e-06, + "loss": 0.6089, + "step": 1714 + }, + { + "epoch": 1.428601209089014, + "grad_norm": 0.13489751517772675, + "learning_rate": 1.1196448139178295e-06, + "loss": 0.6001, + "step": 1715 + }, + { + "epoch": 1.429435063581405, + "grad_norm": 0.12865175306797028, + "learning_rate": 1.118752851019208e-06, + "loss": 0.6606, + "step": 1716 + }, + { + "epoch": 1.4302689180737962, + "grad_norm": 0.12072816491127014, + "learning_rate": 1.1178607922793307e-06, + "loss": 0.6222, + "step": 1717 + }, + { + "epoch": 1.4311027725661871, + "grad_norm": 0.11538699269294739, + "learning_rate": 1.1169686384181467e-06, + "loss": 0.5956, + "step": 1718 + }, + { + "epoch": 1.4319366270585783, + "grad_norm": 0.1280529797077179, + "learning_rate": 1.116076390155682e-06, + "loss": 0.5582, + "step": 1719 + }, + { + "epoch": 1.4327704815509694, + "grad_norm": 0.11776122450828552, + "learning_rate": 1.1151840482120385e-06, + "loss": 0.5765, + "step": 1720 + }, + { + "epoch": 1.4336043360433606, + "grad_norm": 0.12921258807182312, + "learning_rate": 1.1142916133073946e-06, + "loss": 0.6283, + "step": 1721 + }, + { + "epoch": 1.4344381905357515, + "grad_norm": 0.11978008598089218, + "learning_rate": 1.113399086162003e-06, + "loss": 0.5641, + "step": 1722 + }, + { + "epoch": 1.4352720450281427, + "grad_norm": 0.12989699840545654, + "learning_rate": 1.1125064674961913e-06, + "loss": 0.5957, + "step": 1723 + }, + { + "epoch": 1.4361058995205336, + "grad_norm": 0.12442397326231003, + "learning_rate": 1.1116137580303602e-06, + "loss": 0.5626, + "step": 1724 + }, + { + "epoch": 1.4369397540129247, + "grad_norm": 0.12246950715780258, + "learning_rate": 1.1107209584849844e-06, + "loss": 0.6279, + "step": 1725 + }, + { + "epoch": 1.4377736085053159, + "grad_norm": 0.12757597863674164, + "learning_rate": 1.109828069580611e-06, + "loss": 0.6326, + "step": 1726 + }, + { + "epoch": 1.438607462997707, + "grad_norm": 0.12838071584701538, + "learning_rate": 1.108935092037859e-06, + "loss": 0.6689, + "step": 1727 + }, + { + "epoch": 1.439441317490098, + "grad_norm": 0.13266290724277496, + "learning_rate": 1.10804202657742e-06, + "loss": 0.6087, + "step": 1728 + }, + { + "epoch": 1.440275171982489, + "grad_norm": 0.1249866783618927, + "learning_rate": 1.107148873920055e-06, + "loss": 0.5671, + "step": 1729 + }, + { + "epoch": 1.44110902647488, + "grad_norm": 0.12549594044685364, + "learning_rate": 1.1062556347865966e-06, + "loss": 0.6026, + "step": 1730 + }, + { + "epoch": 1.4419428809672712, + "grad_norm": 0.11809445917606354, + "learning_rate": 1.1053623098979464e-06, + "loss": 0.6609, + "step": 1731 + }, + { + "epoch": 1.4427767354596623, + "grad_norm": 0.1284630000591278, + "learning_rate": 1.1044688999750757e-06, + "loss": 0.6976, + "step": 1732 + }, + { + "epoch": 1.4436105899520535, + "grad_norm": 0.12497500330209732, + "learning_rate": 1.1035754057390245e-06, + "loss": 0.6037, + "step": 1733 + }, + { + "epoch": 1.4444444444444444, + "grad_norm": 0.12468233704566956, + "learning_rate": 1.1026818279109002e-06, + "loss": 0.5953, + "step": 1734 + }, + { + "epoch": 1.4452782989368356, + "grad_norm": 0.1244860291481018, + "learning_rate": 1.1017881672118785e-06, + "loss": 0.5816, + "step": 1735 + }, + { + "epoch": 1.4461121534292265, + "grad_norm": 0.12283075600862503, + "learning_rate": 1.1008944243632012e-06, + "loss": 0.5896, + "step": 1736 + }, + { + "epoch": 1.4469460079216176, + "grad_norm": 0.12096507102251053, + "learning_rate": 1.100000600086177e-06, + "loss": 0.5848, + "step": 1737 + }, + { + "epoch": 1.4477798624140088, + "grad_norm": 0.13141663372516632, + "learning_rate": 1.0991066951021802e-06, + "loss": 0.5944, + "step": 1738 + }, + { + "epoch": 1.4486137169064, + "grad_norm": 0.11670444160699844, + "learning_rate": 1.0982127101326497e-06, + "loss": 0.5609, + "step": 1739 + }, + { + "epoch": 1.4494475713987909, + "grad_norm": 0.1190413385629654, + "learning_rate": 1.09731864589909e-06, + "loss": 0.639, + "step": 1740 + }, + { + "epoch": 1.450281425891182, + "grad_norm": 0.12622325122356415, + "learning_rate": 1.0964245031230684e-06, + "loss": 0.6146, + "step": 1741 + }, + { + "epoch": 1.451115280383573, + "grad_norm": 0.12580569088459015, + "learning_rate": 1.0955302825262161e-06, + "loss": 0.5947, + "step": 1742 + }, + { + "epoch": 1.451949134875964, + "grad_norm": 0.1237453743815422, + "learning_rate": 1.0946359848302273e-06, + "loss": 0.6203, + "step": 1743 + }, + { + "epoch": 1.4527829893683553, + "grad_norm": 0.11954555660486221, + "learning_rate": 1.0937416107568586e-06, + "loss": 0.5816, + "step": 1744 + }, + { + "epoch": 1.4536168438607464, + "grad_norm": 0.125918909907341, + "learning_rate": 1.0928471610279278e-06, + "loss": 0.5414, + "step": 1745 + }, + { + "epoch": 1.4544506983531373, + "grad_norm": 0.1135488972067833, + "learning_rate": 1.0919526363653133e-06, + "loss": 0.5737, + "step": 1746 + }, + { + "epoch": 1.4552845528455285, + "grad_norm": 0.1270725578069687, + "learning_rate": 1.091058037490955e-06, + "loss": 0.7054, + "step": 1747 + }, + { + "epoch": 1.4561184073379194, + "grad_norm": 0.11943162232637405, + "learning_rate": 1.0901633651268524e-06, + "loss": 0.601, + "step": 1748 + }, + { + "epoch": 1.4569522618303106, + "grad_norm": 0.12362517416477203, + "learning_rate": 1.089268619995064e-06, + "loss": 0.6438, + "step": 1749 + }, + { + "epoch": 1.4577861163227017, + "grad_norm": 0.11327839642763138, + "learning_rate": 1.0883738028177068e-06, + "loss": 0.5466, + "step": 1750 + }, + { + "epoch": 1.4586199708150929, + "grad_norm": 0.11802560091018677, + "learning_rate": 1.0874789143169567e-06, + "loss": 0.5889, + "step": 1751 + }, + { + "epoch": 1.4594538253074838, + "grad_norm": 0.1234131008386612, + "learning_rate": 1.0865839552150469e-06, + "loss": 0.6463, + "step": 1752 + }, + { + "epoch": 1.460287679799875, + "grad_norm": 0.11657127737998962, + "learning_rate": 1.085688926234267e-06, + "loss": 0.5814, + "step": 1753 + }, + { + "epoch": 1.4611215342922659, + "grad_norm": 0.12287198007106781, + "learning_rate": 1.084793828096964e-06, + "loss": 0.6292, + "step": 1754 + }, + { + "epoch": 1.461955388784657, + "grad_norm": 0.11891543120145798, + "learning_rate": 1.0838986615255397e-06, + "loss": 0.6176, + "step": 1755 + }, + { + "epoch": 1.4627892432770482, + "grad_norm": 0.12395941466093063, + "learning_rate": 1.0830034272424518e-06, + "loss": 0.5912, + "step": 1756 + }, + { + "epoch": 1.4636230977694393, + "grad_norm": 0.12994913756847382, + "learning_rate": 1.0821081259702126e-06, + "loss": 0.6032, + "step": 1757 + }, + { + "epoch": 1.4644569522618303, + "grad_norm": 0.12760955095291138, + "learning_rate": 1.0812127584313882e-06, + "loss": 0.619, + "step": 1758 + }, + { + "epoch": 1.4652908067542214, + "grad_norm": 0.13421405851840973, + "learning_rate": 1.0803173253485982e-06, + "loss": 0.5927, + "step": 1759 + }, + { + "epoch": 1.4661246612466123, + "grad_norm": 0.1249236911535263, + "learning_rate": 1.0794218274445153e-06, + "loss": 0.5979, + "step": 1760 + }, + { + "epoch": 1.4669585157390035, + "grad_norm": 0.12936027348041534, + "learning_rate": 1.0785262654418645e-06, + "loss": 0.5907, + "step": 1761 + }, + { + "epoch": 1.4677923702313946, + "grad_norm": 0.1282528042793274, + "learning_rate": 1.0776306400634225e-06, + "loss": 0.6585, + "step": 1762 + }, + { + "epoch": 1.4686262247237858, + "grad_norm": 0.12033521384000778, + "learning_rate": 1.0767349520320165e-06, + "loss": 0.5572, + "step": 1763 + }, + { + "epoch": 1.4694600792161767, + "grad_norm": 0.12083300203084946, + "learning_rate": 1.0758392020705255e-06, + "loss": 0.6449, + "step": 1764 + }, + { + "epoch": 1.4702939337085679, + "grad_norm": 0.1208571195602417, + "learning_rate": 1.0749433909018776e-06, + "loss": 0.6389, + "step": 1765 + }, + { + "epoch": 1.4711277882009588, + "grad_norm": 0.11647844314575195, + "learning_rate": 1.0740475192490509e-06, + "loss": 0.6427, + "step": 1766 + }, + { + "epoch": 1.47196164269335, + "grad_norm": 0.12908311188220978, + "learning_rate": 1.073151587835071e-06, + "loss": 0.6074, + "step": 1767 + }, + { + "epoch": 1.472795497185741, + "grad_norm": 0.12080005556344986, + "learning_rate": 1.0722555973830138e-06, + "loss": 0.6037, + "step": 1768 + }, + { + "epoch": 1.4736293516781322, + "grad_norm": 0.12868741154670715, + "learning_rate": 1.0713595486160011e-06, + "loss": 0.6149, + "step": 1769 + }, + { + "epoch": 1.4744632061705232, + "grad_norm": 0.12102212756872177, + "learning_rate": 1.0704634422572027e-06, + "loss": 0.5885, + "step": 1770 + }, + { + "epoch": 1.4752970606629143, + "grad_norm": 0.11719702929258347, + "learning_rate": 1.069567279029834e-06, + "loss": 0.5335, + "step": 1771 + }, + { + "epoch": 1.4761309151553055, + "grad_norm": 0.12366301566362381, + "learning_rate": 1.0686710596571578e-06, + "loss": 0.5753, + "step": 1772 + }, + { + "epoch": 1.4769647696476964, + "grad_norm": 0.1275297850370407, + "learning_rate": 1.0677747848624803e-06, + "loss": 0.6315, + "step": 1773 + }, + { + "epoch": 1.4777986241400876, + "grad_norm": 0.11969911307096481, + "learning_rate": 1.066878455369154e-06, + "loss": 0.5857, + "step": 1774 + }, + { + "epoch": 1.4786324786324787, + "grad_norm": 0.13167330622673035, + "learning_rate": 1.0659820719005749e-06, + "loss": 0.6526, + "step": 1775 + }, + { + "epoch": 1.4794663331248696, + "grad_norm": 0.12339498847723007, + "learning_rate": 1.0650856351801824e-06, + "loss": 0.6355, + "step": 1776 + }, + { + "epoch": 1.4803001876172608, + "grad_norm": 0.12983542680740356, + "learning_rate": 1.0641891459314595e-06, + "loss": 0.6069, + "step": 1777 + }, + { + "epoch": 1.481134042109652, + "grad_norm": 0.1269686222076416, + "learning_rate": 1.063292604877931e-06, + "loss": 0.6478, + "step": 1778 + }, + { + "epoch": 1.4819678966020429, + "grad_norm": 0.12114999443292618, + "learning_rate": 1.0623960127431635e-06, + "loss": 0.5406, + "step": 1779 + }, + { + "epoch": 1.482801751094434, + "grad_norm": 0.11573608964681625, + "learning_rate": 1.061499370250765e-06, + "loss": 0.5789, + "step": 1780 + }, + { + "epoch": 1.4836356055868252, + "grad_norm": 0.1279306858778, + "learning_rate": 1.0606026781243847e-06, + "loss": 0.6475, + "step": 1781 + }, + { + "epoch": 1.484469460079216, + "grad_norm": 0.12379437685012817, + "learning_rate": 1.0597059370877103e-06, + "loss": 0.5484, + "step": 1782 + }, + { + "epoch": 1.4853033145716072, + "grad_norm": 0.12157581001520157, + "learning_rate": 1.0588091478644714e-06, + "loss": 0.5866, + "step": 1783 + }, + { + "epoch": 1.4861371690639984, + "grad_norm": 0.13414230942726135, + "learning_rate": 1.0579123111784342e-06, + "loss": 0.6632, + "step": 1784 + }, + { + "epoch": 1.4869710235563893, + "grad_norm": 0.12635232508182526, + "learning_rate": 1.057015427753404e-06, + "loss": 0.6577, + "step": 1785 + }, + { + "epoch": 1.4878048780487805, + "grad_norm": 0.13163447380065918, + "learning_rate": 1.0561184983132244e-06, + "loss": 0.6335, + "step": 1786 + }, + { + "epoch": 1.4886387325411716, + "grad_norm": 0.12453741580247879, + "learning_rate": 1.0552215235817754e-06, + "loss": 0.6312, + "step": 1787 + }, + { + "epoch": 1.4894725870335628, + "grad_norm": 0.12222108989953995, + "learning_rate": 1.054324504282974e-06, + "loss": 0.6371, + "step": 1788 + }, + { + "epoch": 1.4903064415259537, + "grad_norm": 0.1175316870212555, + "learning_rate": 1.0534274411407725e-06, + "loss": 0.5479, + "step": 1789 + }, + { + "epoch": 1.4911402960183449, + "grad_norm": 0.1285734921693802, + "learning_rate": 1.0525303348791596e-06, + "loss": 0.6339, + "step": 1790 + }, + { + "epoch": 1.4919741505107358, + "grad_norm": 0.1263425052165985, + "learning_rate": 1.0516331862221582e-06, + "loss": 0.6376, + "step": 1791 + }, + { + "epoch": 1.492808005003127, + "grad_norm": 0.13151004910469055, + "learning_rate": 1.050735995893825e-06, + "loss": 0.5966, + "step": 1792 + }, + { + "epoch": 1.493641859495518, + "grad_norm": 0.12273474782705307, + "learning_rate": 1.0498387646182511e-06, + "loss": 0.6469, + "step": 1793 + }, + { + "epoch": 1.4944757139879092, + "grad_norm": 0.1246555745601654, + "learning_rate": 1.0489414931195603e-06, + "loss": 0.5489, + "step": 1794 + }, + { + "epoch": 1.4953095684803002, + "grad_norm": 0.12344285845756531, + "learning_rate": 1.0480441821219096e-06, + "loss": 0.5895, + "step": 1795 + }, + { + "epoch": 1.4961434229726913, + "grad_norm": 0.12281202524900436, + "learning_rate": 1.047146832349486e-06, + "loss": 0.595, + "step": 1796 + }, + { + "epoch": 1.4969772774650822, + "grad_norm": 0.13203196227550507, + "learning_rate": 1.0462494445265098e-06, + "loss": 0.6427, + "step": 1797 + }, + { + "epoch": 1.4978111319574734, + "grad_norm": 0.12045428156852722, + "learning_rate": 1.045352019377231e-06, + "loss": 0.5412, + "step": 1798 + }, + { + "epoch": 1.4986449864498645, + "grad_norm": 0.1290232390165329, + "learning_rate": 1.0444545576259302e-06, + "loss": 0.6301, + "step": 1799 + }, + { + "epoch": 1.4994788409422557, + "grad_norm": 0.12524542212486267, + "learning_rate": 1.0435570599969165e-06, + "loss": 0.6474, + "step": 1800 + }, + { + "epoch": 1.5003126954346466, + "grad_norm": 0.11619977653026581, + "learning_rate": 1.0426595272145295e-06, + "loss": 0.5225, + "step": 1801 + }, + { + "epoch": 1.5011465499270378, + "grad_norm": 0.12204474955797195, + "learning_rate": 1.041761960003136e-06, + "loss": 0.5938, + "step": 1802 + }, + { + "epoch": 1.5019804044194287, + "grad_norm": 0.13166320323944092, + "learning_rate": 1.0408643590871312e-06, + "loss": 0.6535, + "step": 1803 + }, + { + "epoch": 1.5028142589118199, + "grad_norm": 0.12408415228128433, + "learning_rate": 1.0399667251909368e-06, + "loss": 0.5745, + "step": 1804 + }, + { + "epoch": 1.503648113404211, + "grad_norm": 0.1306605190038681, + "learning_rate": 1.039069059039002e-06, + "loss": 0.6438, + "step": 1805 + }, + { + "epoch": 1.5044819678966022, + "grad_norm": 0.11977241933345795, + "learning_rate": 1.0381713613558022e-06, + "loss": 0.6051, + "step": 1806 + }, + { + "epoch": 1.505315822388993, + "grad_norm": 0.11851506680250168, + "learning_rate": 1.0372736328658363e-06, + "loss": 0.6246, + "step": 1807 + }, + { + "epoch": 1.5061496768813842, + "grad_norm": 0.12838518619537354, + "learning_rate": 1.0363758742936306e-06, + "loss": 0.6446, + "step": 1808 + }, + { + "epoch": 1.5069835313737752, + "grad_norm": 0.11811474710702896, + "learning_rate": 1.0354780863637337e-06, + "loss": 0.5982, + "step": 1809 + }, + { + "epoch": 1.5078173858661663, + "grad_norm": 0.12509089708328247, + "learning_rate": 1.0345802698007196e-06, + "loss": 0.6321, + "step": 1810 + }, + { + "epoch": 1.5086512403585575, + "grad_norm": 0.1284075677394867, + "learning_rate": 1.0336824253291835e-06, + "loss": 0.641, + "step": 1811 + }, + { + "epoch": 1.5094850948509486, + "grad_norm": 0.12310284376144409, + "learning_rate": 1.0327845536737453e-06, + "loss": 0.6306, + "step": 1812 + }, + { + "epoch": 1.5103189493433395, + "grad_norm": 0.12668739259243011, + "learning_rate": 1.031886655559045e-06, + "loss": 0.6587, + "step": 1813 + }, + { + "epoch": 1.5111528038357307, + "grad_norm": 0.12098442763090134, + "learning_rate": 1.0309887317097444e-06, + "loss": 0.6573, + "step": 1814 + }, + { + "epoch": 1.5119866583281216, + "grad_norm": 0.1259874552488327, + "learning_rate": 1.0300907828505267e-06, + "loss": 0.641, + "step": 1815 + }, + { + "epoch": 1.5128205128205128, + "grad_norm": 0.13358241319656372, + "learning_rate": 1.0291928097060948e-06, + "loss": 0.6342, + "step": 1816 + }, + { + "epoch": 1.513654367312904, + "grad_norm": 0.12032781541347504, + "learning_rate": 1.0282948130011715e-06, + "loss": 0.5766, + "step": 1817 + }, + { + "epoch": 1.514488221805295, + "grad_norm": 0.11778460443019867, + "learning_rate": 1.027396793460498e-06, + "loss": 0.5576, + "step": 1818 + }, + { + "epoch": 1.515322076297686, + "grad_norm": 0.12484540045261383, + "learning_rate": 1.0264987518088346e-06, + "loss": 0.6544, + "step": 1819 + }, + { + "epoch": 1.5161559307900772, + "grad_norm": 0.12196074426174164, + "learning_rate": 1.0256006887709592e-06, + "loss": 0.5936, + "step": 1820 + }, + { + "epoch": 1.516989785282468, + "grad_norm": 0.13073596358299255, + "learning_rate": 1.024702605071667e-06, + "loss": 0.6153, + "step": 1821 + }, + { + "epoch": 1.5178236397748592, + "grad_norm": 0.124057836830616, + "learning_rate": 1.0238045014357692e-06, + "loss": 0.6284, + "step": 1822 + }, + { + "epoch": 1.5186574942672504, + "grad_norm": 0.12679198384284973, + "learning_rate": 1.0229063785880948e-06, + "loss": 0.5893, + "step": 1823 + }, + { + "epoch": 1.5194913487596415, + "grad_norm": 0.128417506814003, + "learning_rate": 1.022008237253486e-06, + "loss": 0.5923, + "step": 1824 + }, + { + "epoch": 1.5203252032520327, + "grad_norm": 0.12480554729700089, + "learning_rate": 1.0211100781568023e-06, + "loss": 0.606, + "step": 1825 + }, + { + "epoch": 1.5211590577444236, + "grad_norm": 0.1268758475780487, + "learning_rate": 1.0202119020229154e-06, + "loss": 0.6155, + "step": 1826 + }, + { + "epoch": 1.5219929122368145, + "grad_norm": 0.12820713222026825, + "learning_rate": 1.0193137095767123e-06, + "loss": 0.5935, + "step": 1827 + }, + { + "epoch": 1.5228267667292057, + "grad_norm": 0.11923402547836304, + "learning_rate": 1.0184155015430926e-06, + "loss": 0.6327, + "step": 1828 + }, + { + "epoch": 1.5236606212215968, + "grad_norm": 0.11409208923578262, + "learning_rate": 1.017517278646968e-06, + "loss": 0.5759, + "step": 1829 + }, + { + "epoch": 1.524494475713988, + "grad_norm": 0.12284267693758011, + "learning_rate": 1.016619041613263e-06, + "loss": 0.5775, + "step": 1830 + }, + { + "epoch": 1.5253283302063791, + "grad_norm": 0.12808358669281006, + "learning_rate": 1.015720791166913e-06, + "loss": 0.5413, + "step": 1831 + }, + { + "epoch": 1.52616218469877, + "grad_norm": 0.11841126531362534, + "learning_rate": 1.0148225280328647e-06, + "loss": 0.573, + "step": 1832 + }, + { + "epoch": 1.526996039191161, + "grad_norm": 0.12383410334587097, + "learning_rate": 1.0139242529360744e-06, + "loss": 0.594, + "step": 1833 + }, + { + "epoch": 1.5278298936835522, + "grad_norm": 0.12949687242507935, + "learning_rate": 1.0130259666015084e-06, + "loss": 0.5782, + "step": 1834 + }, + { + "epoch": 1.5286637481759433, + "grad_norm": 0.12045048922300339, + "learning_rate": 1.0121276697541426e-06, + "loss": 0.6006, + "step": 1835 + }, + { + "epoch": 1.5294976026683345, + "grad_norm": 0.1283283829689026, + "learning_rate": 1.01122936311896e-06, + "loss": 0.638, + "step": 1836 + }, + { + "epoch": 1.5303314571607256, + "grad_norm": 0.12735942006111145, + "learning_rate": 1.0103310474209527e-06, + "loss": 0.6341, + "step": 1837 + }, + { + "epoch": 1.5311653116531165, + "grad_norm": 0.12286734580993652, + "learning_rate": 1.0094327233851199e-06, + "loss": 0.5176, + "step": 1838 + }, + { + "epoch": 1.5319991661455075, + "grad_norm": 0.12337525188922882, + "learning_rate": 1.0085343917364675e-06, + "loss": 0.5797, + "step": 1839 + }, + { + "epoch": 1.5328330206378986, + "grad_norm": 0.12428612262010574, + "learning_rate": 1.0076360532000068e-06, + "loss": 0.5969, + "step": 1840 + }, + { + "epoch": 1.5336668751302898, + "grad_norm": 0.12380371987819672, + "learning_rate": 1.006737708500756e-06, + "loss": 0.6301, + "step": 1841 + }, + { + "epoch": 1.534500729622681, + "grad_norm": 0.12506607174873352, + "learning_rate": 1.0058393583637376e-06, + "loss": 0.6016, + "step": 1842 + }, + { + "epoch": 1.535334584115072, + "grad_norm": 0.12334273755550385, + "learning_rate": 1.0049410035139774e-06, + "loss": 0.5499, + "step": 1843 + }, + { + "epoch": 1.536168438607463, + "grad_norm": 0.12097372114658356, + "learning_rate": 1.0040426446765074e-06, + "loss": 0.6183, + "step": 1844 + }, + { + "epoch": 1.537002293099854, + "grad_norm": 0.13287068903446198, + "learning_rate": 1.0031442825763603e-06, + "loss": 0.66, + "step": 1845 + }, + { + "epoch": 1.537836147592245, + "grad_norm": 0.12525850534439087, + "learning_rate": 1.0022459179385736e-06, + "loss": 0.6031, + "step": 1846 + }, + { + "epoch": 1.5386700020846362, + "grad_norm": 0.11843365430831909, + "learning_rate": 1.001347551488185e-06, + "loss": 0.6228, + "step": 1847 + }, + { + "epoch": 1.5395038565770274, + "grad_norm": 0.12727557122707367, + "learning_rate": 1.0004491839502352e-06, + "loss": 0.5877, + "step": 1848 + }, + { + "epoch": 1.5403377110694185, + "grad_norm": 0.12447907030582428, + "learning_rate": 9.99550816049765e-07, + "loss": 0.6149, + "step": 1849 + }, + { + "epoch": 1.5411715655618095, + "grad_norm": 0.12982895970344543, + "learning_rate": 9.986524485118152e-07, + "loss": 0.6134, + "step": 1850 + }, + { + "epoch": 1.5420054200542004, + "grad_norm": 0.13101699948310852, + "learning_rate": 9.977540820614265e-07, + "loss": 0.5873, + "step": 1851 + }, + { + "epoch": 1.5428392745465915, + "grad_norm": 0.126919686794281, + "learning_rate": 9.968557174236396e-07, + "loss": 0.5707, + "step": 1852 + }, + { + "epoch": 1.5436731290389827, + "grad_norm": 0.12582388520240784, + "learning_rate": 9.95957355323493e-07, + "loss": 0.5985, + "step": 1853 + }, + { + "epoch": 1.5445069835313738, + "grad_norm": 0.13102799654006958, + "learning_rate": 9.950589964860226e-07, + "loss": 0.5939, + "step": 1854 + }, + { + "epoch": 1.545340838023765, + "grad_norm": 0.13151919841766357, + "learning_rate": 9.941606416362628e-07, + "loss": 0.6505, + "step": 1855 + }, + { + "epoch": 1.546174692516156, + "grad_norm": 0.12860901653766632, + "learning_rate": 9.932622914992437e-07, + "loss": 0.6129, + "step": 1856 + }, + { + "epoch": 1.547008547008547, + "grad_norm": 0.1313966065645218, + "learning_rate": 9.923639467999929e-07, + "loss": 0.6383, + "step": 1857 + }, + { + "epoch": 1.547842401500938, + "grad_norm": 0.12255886942148209, + "learning_rate": 9.914656082635326e-07, + "loss": 0.5915, + "step": 1858 + }, + { + "epoch": 1.5486762559933291, + "grad_norm": 0.12750515341758728, + "learning_rate": 9.9056727661488e-07, + "loss": 0.5972, + "step": 1859 + }, + { + "epoch": 1.5495101104857203, + "grad_norm": 0.12609492242336273, + "learning_rate": 9.896689525790472e-07, + "loss": 0.6454, + "step": 1860 + }, + { + "epoch": 1.5503439649781114, + "grad_norm": 0.1195141151547432, + "learning_rate": 9.887706368810402e-07, + "loss": 0.5384, + "step": 1861 + }, + { + "epoch": 1.5511778194705024, + "grad_norm": 0.12656457722187042, + "learning_rate": 9.878723302458576e-07, + "loss": 0.6791, + "step": 1862 + }, + { + "epoch": 1.5520116739628935, + "grad_norm": 0.1253191977739334, + "learning_rate": 9.869740333984917e-07, + "loss": 0.5846, + "step": 1863 + }, + { + "epoch": 1.5528455284552845, + "grad_norm": 0.12685538828372955, + "learning_rate": 9.860757470639258e-07, + "loss": 0.6074, + "step": 1864 + }, + { + "epoch": 1.5536793829476756, + "grad_norm": 0.12786641716957092, + "learning_rate": 9.851774719671355e-07, + "loss": 0.6289, + "step": 1865 + }, + { + "epoch": 1.5545132374400668, + "grad_norm": 0.12862898409366608, + "learning_rate": 9.84279208833087e-07, + "loss": 0.6372, + "step": 1866 + }, + { + "epoch": 1.555347091932458, + "grad_norm": 0.1340702474117279, + "learning_rate": 9.833809583867374e-07, + "loss": 0.6325, + "step": 1867 + }, + { + "epoch": 1.5561809464248488, + "grad_norm": 0.12695550918579102, + "learning_rate": 9.824827213530321e-07, + "loss": 0.631, + "step": 1868 + }, + { + "epoch": 1.55701480091724, + "grad_norm": 0.12067730724811554, + "learning_rate": 9.815844984569077e-07, + "loss": 0.51, + "step": 1869 + }, + { + "epoch": 1.557848655409631, + "grad_norm": 0.1230134516954422, + "learning_rate": 9.806862904232878e-07, + "loss": 0.5481, + "step": 1870 + }, + { + "epoch": 1.558682509902022, + "grad_norm": 0.11921508610248566, + "learning_rate": 9.797880979770843e-07, + "loss": 0.5222, + "step": 1871 + }, + { + "epoch": 1.5595163643944132, + "grad_norm": 0.12758903205394745, + "learning_rate": 9.788899218431978e-07, + "loss": 0.5841, + "step": 1872 + }, + { + "epoch": 1.5603502188868044, + "grad_norm": 0.12362825870513916, + "learning_rate": 9.779917627465138e-07, + "loss": 0.5896, + "step": 1873 + }, + { + "epoch": 1.5611840733791953, + "grad_norm": 0.11825423687696457, + "learning_rate": 9.770936214119053e-07, + "loss": 0.5742, + "step": 1874 + }, + { + "epoch": 1.5620179278715864, + "grad_norm": 0.12133418768644333, + "learning_rate": 9.761954985642307e-07, + "loss": 0.5397, + "step": 1875 + }, + { + "epoch": 1.5628517823639774, + "grad_norm": 0.1267905831336975, + "learning_rate": 9.752973949283332e-07, + "loss": 0.5854, + "step": 1876 + }, + { + "epoch": 1.5636856368563685, + "grad_norm": 0.12750284373760223, + "learning_rate": 9.743993112290408e-07, + "loss": 0.6361, + "step": 1877 + }, + { + "epoch": 1.5645194913487597, + "grad_norm": 0.12900686264038086, + "learning_rate": 9.735012481911653e-07, + "loss": 0.5673, + "step": 1878 + }, + { + "epoch": 1.5653533458411508, + "grad_norm": 0.12558184564113617, + "learning_rate": 9.726032065395022e-07, + "loss": 0.6087, + "step": 1879 + }, + { + "epoch": 1.5661872003335418, + "grad_norm": 0.12879346311092377, + "learning_rate": 9.717051869988286e-07, + "loss": 0.6159, + "step": 1880 + }, + { + "epoch": 1.567021054825933, + "grad_norm": 0.12538638710975647, + "learning_rate": 9.708071902939053e-07, + "loss": 0.5894, + "step": 1881 + }, + { + "epoch": 1.5678549093183238, + "grad_norm": 0.12777774035930634, + "learning_rate": 9.699092171494735e-07, + "loss": 0.6216, + "step": 1882 + }, + { + "epoch": 1.568688763810715, + "grad_norm": 0.12464749068021774, + "learning_rate": 9.690112682902557e-07, + "loss": 0.6147, + "step": 1883 + }, + { + "epoch": 1.5695226183031061, + "grad_norm": 0.12391242384910583, + "learning_rate": 9.681133444409552e-07, + "loss": 0.6487, + "step": 1884 + }, + { + "epoch": 1.5703564727954973, + "grad_norm": 0.12629380822181702, + "learning_rate": 9.672154463262546e-07, + "loss": 0.5764, + "step": 1885 + }, + { + "epoch": 1.5711903272878882, + "grad_norm": 0.13301175832748413, + "learning_rate": 9.663175746708162e-07, + "loss": 0.5439, + "step": 1886 + }, + { + "epoch": 1.5720241817802794, + "grad_norm": 0.1252022087574005, + "learning_rate": 9.654197301992805e-07, + "loss": 0.6084, + "step": 1887 + }, + { + "epoch": 1.5728580362726703, + "grad_norm": 0.12484809011220932, + "learning_rate": 9.64521913636266e-07, + "loss": 0.6254, + "step": 1888 + }, + { + "epoch": 1.5736918907650614, + "grad_norm": 0.13050304353237152, + "learning_rate": 9.636241257063695e-07, + "loss": 0.5849, + "step": 1889 + }, + { + "epoch": 1.5745257452574526, + "grad_norm": 0.1395888775587082, + "learning_rate": 9.627263671341638e-07, + "loss": 0.6367, + "step": 1890 + }, + { + "epoch": 1.5753595997498437, + "grad_norm": 0.12380406260490417, + "learning_rate": 9.61828638644198e-07, + "loss": 0.6102, + "step": 1891 + }, + { + "epoch": 1.576193454242235, + "grad_norm": 0.1276310682296753, + "learning_rate": 9.609309409609978e-07, + "loss": 0.6069, + "step": 1892 + }, + { + "epoch": 1.5770273087346258, + "grad_norm": 0.12831759452819824, + "learning_rate": 9.600332748090633e-07, + "loss": 0.6047, + "step": 1893 + }, + { + "epoch": 1.5778611632270167, + "grad_norm": 0.12802143394947052, + "learning_rate": 9.59135640912869e-07, + "loss": 0.5987, + "step": 1894 + }, + { + "epoch": 1.578695017719408, + "grad_norm": 0.12463508546352386, + "learning_rate": 9.582380399968642e-07, + "loss": 0.5831, + "step": 1895 + }, + { + "epoch": 1.579528872211799, + "grad_norm": 0.13273175060749054, + "learning_rate": 9.573404727854708e-07, + "loss": 0.5763, + "step": 1896 + }, + { + "epoch": 1.5803627267041902, + "grad_norm": 0.1212095320224762, + "learning_rate": 9.564429400030836e-07, + "loss": 0.6321, + "step": 1897 + }, + { + "epoch": 1.5811965811965814, + "grad_norm": 0.13839462399482727, + "learning_rate": 9.555454423740702e-07, + "loss": 0.565, + "step": 1898 + }, + { + "epoch": 1.5820304356889723, + "grad_norm": 0.12162493169307709, + "learning_rate": 9.54647980622769e-07, + "loss": 0.633, + "step": 1899 + }, + { + "epoch": 1.5828642901813632, + "grad_norm": 0.13095051050186157, + "learning_rate": 9.5375055547349e-07, + "loss": 0.6058, + "step": 1900 + }, + { + "epoch": 1.5836981446737544, + "grad_norm": 0.12818747758865356, + "learning_rate": 9.528531676505139e-07, + "loss": 0.63, + "step": 1901 + }, + { + "epoch": 1.5845319991661455, + "grad_norm": 0.12812934815883636, + "learning_rate": 9.519558178780903e-07, + "loss": 0.6276, + "step": 1902 + }, + { + "epoch": 1.5853658536585367, + "grad_norm": 0.12593896687030792, + "learning_rate": 9.510585068804393e-07, + "loss": 0.6017, + "step": 1903 + }, + { + "epoch": 1.5861997081509278, + "grad_norm": 0.12588874995708466, + "learning_rate": 9.501612353817489e-07, + "loss": 0.5889, + "step": 1904 + }, + { + "epoch": 1.5870335626433187, + "grad_norm": 0.12574082612991333, + "learning_rate": 9.492640041061751e-07, + "loss": 0.642, + "step": 1905 + }, + { + "epoch": 1.5878674171357097, + "grad_norm": 0.125298872590065, + "learning_rate": 9.48366813777842e-07, + "loss": 0.5936, + "step": 1906 + }, + { + "epoch": 1.5887012716281008, + "grad_norm": 0.13028943538665771, + "learning_rate": 9.474696651208405e-07, + "loss": 0.5672, + "step": 1907 + }, + { + "epoch": 1.589535126120492, + "grad_norm": 0.12832380831241608, + "learning_rate": 9.465725588592276e-07, + "loss": 0.5677, + "step": 1908 + }, + { + "epoch": 1.5903689806128831, + "grad_norm": 0.1250237077474594, + "learning_rate": 9.456754957170261e-07, + "loss": 0.5445, + "step": 1909 + }, + { + "epoch": 1.5912028351052743, + "grad_norm": 0.12492944300174713, + "learning_rate": 9.447784764182246e-07, + "loss": 0.503, + "step": 1910 + }, + { + "epoch": 1.5920366895976652, + "grad_norm": 0.13353389501571655, + "learning_rate": 9.438815016867757e-07, + "loss": 0.5951, + "step": 1911 + }, + { + "epoch": 1.5928705440900561, + "grad_norm": 0.12580597400665283, + "learning_rate": 9.429845722465959e-07, + "loss": 0.6326, + "step": 1912 + }, + { + "epoch": 1.5937043985824473, + "grad_norm": 0.13248313963413239, + "learning_rate": 9.42087688821566e-07, + "loss": 0.6321, + "step": 1913 + }, + { + "epoch": 1.5945382530748384, + "grad_norm": 0.12739625573158264, + "learning_rate": 9.411908521355288e-07, + "loss": 0.5681, + "step": 1914 + }, + { + "epoch": 1.5953721075672296, + "grad_norm": 0.12767314910888672, + "learning_rate": 9.402940629122893e-07, + "loss": 0.6109, + "step": 1915 + }, + { + "epoch": 1.5962059620596207, + "grad_norm": 0.12399329990148544, + "learning_rate": 9.393973218756155e-07, + "loss": 0.6677, + "step": 1916 + }, + { + "epoch": 1.5970398165520117, + "grad_norm": 0.12876498699188232, + "learning_rate": 9.385006297492348e-07, + "loss": 0.6214, + "step": 1917 + }, + { + "epoch": 1.5978736710444026, + "grad_norm": 0.13340961933135986, + "learning_rate": 9.376039872568367e-07, + "loss": 0.5851, + "step": 1918 + }, + { + "epoch": 1.5987075255367937, + "grad_norm": 0.12147720158100128, + "learning_rate": 9.367073951220692e-07, + "loss": 0.6, + "step": 1919 + }, + { + "epoch": 1.599541380029185, + "grad_norm": 0.12896296381950378, + "learning_rate": 9.358108540685404e-07, + "loss": 0.6626, + "step": 1920 + }, + { + "epoch": 1.600375234521576, + "grad_norm": 0.12212686985731125, + "learning_rate": 9.349143648198174e-07, + "loss": 0.5925, + "step": 1921 + }, + { + "epoch": 1.6012090890139672, + "grad_norm": 0.125599205493927, + "learning_rate": 9.340179280994251e-07, + "loss": 0.6106, + "step": 1922 + }, + { + "epoch": 1.6020429435063581, + "grad_norm": 0.1354406625032425, + "learning_rate": 9.331215446308459e-07, + "loss": 0.6041, + "step": 1923 + }, + { + "epoch": 1.6028767979987493, + "grad_norm": 0.12907730042934418, + "learning_rate": 9.322252151375199e-07, + "loss": 0.6636, + "step": 1924 + }, + { + "epoch": 1.6037106524911402, + "grad_norm": 0.13119882345199585, + "learning_rate": 9.313289403428425e-07, + "loss": 0.581, + "step": 1925 + }, + { + "epoch": 1.6045445069835313, + "grad_norm": 0.13200044631958008, + "learning_rate": 9.304327209701661e-07, + "loss": 0.6078, + "step": 1926 + }, + { + "epoch": 1.6053783614759225, + "grad_norm": 0.12081551551818848, + "learning_rate": 9.295365577427975e-07, + "loss": 0.6248, + "step": 1927 + }, + { + "epoch": 1.6062122159683137, + "grad_norm": 0.12003374099731445, + "learning_rate": 9.286404513839991e-07, + "loss": 0.5611, + "step": 1928 + }, + { + "epoch": 1.6070460704607046, + "grad_norm": 0.12585462629795074, + "learning_rate": 9.27744402616986e-07, + "loss": 0.612, + "step": 1929 + }, + { + "epoch": 1.6078799249530957, + "grad_norm": 0.12522676587104797, + "learning_rate": 9.268484121649287e-07, + "loss": 0.5522, + "step": 1930 + }, + { + "epoch": 1.6087137794454867, + "grad_norm": 0.13690777122974396, + "learning_rate": 9.25952480750949e-07, + "loss": 0.6234, + "step": 1931 + }, + { + "epoch": 1.6095476339378778, + "grad_norm": 0.12669777870178223, + "learning_rate": 9.250566090981221e-07, + "loss": 0.5642, + "step": 1932 + }, + { + "epoch": 1.610381488430269, + "grad_norm": 0.12308058142662048, + "learning_rate": 9.241607979294744e-07, + "loss": 0.5617, + "step": 1933 + }, + { + "epoch": 1.61121534292266, + "grad_norm": 0.12228291481733322, + "learning_rate": 9.232650479679835e-07, + "loss": 0.6025, + "step": 1934 + }, + { + "epoch": 1.612049197415051, + "grad_norm": 0.12602399289608002, + "learning_rate": 9.223693599365776e-07, + "loss": 0.5623, + "step": 1935 + }, + { + "epoch": 1.6128830519074422, + "grad_norm": 0.12143569439649582, + "learning_rate": 9.214737345581355e-07, + "loss": 0.603, + "step": 1936 + }, + { + "epoch": 1.6137169063998331, + "grad_norm": 0.13084720075130463, + "learning_rate": 9.205781725554848e-07, + "loss": 0.6486, + "step": 1937 + }, + { + "epoch": 1.6145507608922243, + "grad_norm": 0.1243109330534935, + "learning_rate": 9.196826746514017e-07, + "loss": 0.5865, + "step": 1938 + }, + { + "epoch": 1.6153846153846154, + "grad_norm": 0.12693606317043304, + "learning_rate": 9.187872415686118e-07, + "loss": 0.5937, + "step": 1939 + }, + { + "epoch": 1.6162184698770066, + "grad_norm": 0.13238000869750977, + "learning_rate": 9.178918740297875e-07, + "loss": 0.5982, + "step": 1940 + }, + { + "epoch": 1.6170523243693975, + "grad_norm": 0.12275899946689606, + "learning_rate": 9.169965727575482e-07, + "loss": 0.6161, + "step": 1941 + }, + { + "epoch": 1.6178861788617886, + "grad_norm": 0.13157354295253754, + "learning_rate": 9.161013384744605e-07, + "loss": 0.5869, + "step": 1942 + }, + { + "epoch": 1.6187200333541796, + "grad_norm": 0.13102547824382782, + "learning_rate": 9.152061719030364e-07, + "loss": 0.613, + "step": 1943 + }, + { + "epoch": 1.6195538878465707, + "grad_norm": 0.12562698125839233, + "learning_rate": 9.143110737657329e-07, + "loss": 0.5896, + "step": 1944 + }, + { + "epoch": 1.6203877423389619, + "grad_norm": 0.13320913910865784, + "learning_rate": 9.134160447849532e-07, + "loss": 0.6378, + "step": 1945 + }, + { + "epoch": 1.621221596831353, + "grad_norm": 0.12788859009742737, + "learning_rate": 9.125210856830432e-07, + "loss": 0.6269, + "step": 1946 + }, + { + "epoch": 1.622055451323744, + "grad_norm": 0.11895091086626053, + "learning_rate": 9.116261971822932e-07, + "loss": 0.6095, + "step": 1947 + }, + { + "epoch": 1.622889305816135, + "grad_norm": 0.12486100941896439, + "learning_rate": 9.107313800049362e-07, + "loss": 0.6176, + "step": 1948 + }, + { + "epoch": 1.623723160308526, + "grad_norm": 0.12875628471374512, + "learning_rate": 9.098366348731476e-07, + "loss": 0.6406, + "step": 1949 + }, + { + "epoch": 1.6245570148009172, + "grad_norm": 0.131727933883667, + "learning_rate": 9.089419625090448e-07, + "loss": 0.6074, + "step": 1950 + }, + { + "epoch": 1.6253908692933083, + "grad_norm": 0.1188148781657219, + "learning_rate": 9.080473636346868e-07, + "loss": 0.4721, + "step": 1951 + }, + { + "epoch": 1.6262247237856995, + "grad_norm": 0.13043268024921417, + "learning_rate": 9.071528389720723e-07, + "loss": 0.5997, + "step": 1952 + }, + { + "epoch": 1.6270585782780906, + "grad_norm": 0.12843738496303558, + "learning_rate": 9.062583892431414e-07, + "loss": 0.5812, + "step": 1953 + }, + { + "epoch": 1.6278924327704816, + "grad_norm": 0.13008764386177063, + "learning_rate": 9.053640151697727e-07, + "loss": 0.624, + "step": 1954 + }, + { + "epoch": 1.6287262872628725, + "grad_norm": 0.12412194907665253, + "learning_rate": 9.044697174737842e-07, + "loss": 0.5941, + "step": 1955 + }, + { + "epoch": 1.6295601417552636, + "grad_norm": 0.12464737147092819, + "learning_rate": 9.035754968769319e-07, + "loss": 0.5998, + "step": 1956 + }, + { + "epoch": 1.6303939962476548, + "grad_norm": 0.1221119835972786, + "learning_rate": 9.026813541009104e-07, + "loss": 0.5577, + "step": 1957 + }, + { + "epoch": 1.631227850740046, + "grad_norm": 0.1313888281583786, + "learning_rate": 9.017872898673501e-07, + "loss": 0.6428, + "step": 1958 + }, + { + "epoch": 1.632061705232437, + "grad_norm": 0.13170261681079865, + "learning_rate": 9.008933048978198e-07, + "loss": 0.627, + "step": 1959 + }, + { + "epoch": 1.632895559724828, + "grad_norm": 0.13136637210845947, + "learning_rate": 8.999993999138227e-07, + "loss": 0.5789, + "step": 1960 + }, + { + "epoch": 1.633729414217219, + "grad_norm": 0.12489289045333862, + "learning_rate": 8.991055756367986e-07, + "loss": 0.6105, + "step": 1961 + }, + { + "epoch": 1.63456326870961, + "grad_norm": 0.12347642332315445, + "learning_rate": 8.982118327881216e-07, + "loss": 0.679, + "step": 1962 + }, + { + "epoch": 1.6353971232020013, + "grad_norm": 0.12213364243507385, + "learning_rate": 8.973181720890998e-07, + "loss": 0.6041, + "step": 1963 + }, + { + "epoch": 1.6362309776943924, + "grad_norm": 0.1292726844549179, + "learning_rate": 8.964245942609755e-07, + "loss": 0.6276, + "step": 1964 + }, + { + "epoch": 1.6370648321867836, + "grad_norm": 0.1329735517501831, + "learning_rate": 8.955311000249243e-07, + "loss": 0.6325, + "step": 1965 + }, + { + "epoch": 1.6378986866791745, + "grad_norm": 0.12430353462696075, + "learning_rate": 8.946376901020538e-07, + "loss": 0.5724, + "step": 1966 + }, + { + "epoch": 1.6387325411715654, + "grad_norm": 0.12788113951683044, + "learning_rate": 8.937443652134036e-07, + "loss": 0.6015, + "step": 1967 + }, + { + "epoch": 1.6395663956639566, + "grad_norm": 0.12255391478538513, + "learning_rate": 8.928511260799452e-07, + "loss": 0.5728, + "step": 1968 + }, + { + "epoch": 1.6404002501563477, + "grad_norm": 0.12283147871494293, + "learning_rate": 8.919579734225803e-07, + "loss": 0.5962, + "step": 1969 + }, + { + "epoch": 1.6412341046487389, + "grad_norm": 0.12731313705444336, + "learning_rate": 8.910649079621409e-07, + "loss": 0.6335, + "step": 1970 + }, + { + "epoch": 1.64206795914113, + "grad_norm": 0.1330873817205429, + "learning_rate": 8.901719304193893e-07, + "loss": 0.5865, + "step": 1971 + }, + { + "epoch": 1.642901813633521, + "grad_norm": 0.12057091295719147, + "learning_rate": 8.892790415150159e-07, + "loss": 0.5224, + "step": 1972 + }, + { + "epoch": 1.6437356681259119, + "grad_norm": 0.12532652914524078, + "learning_rate": 8.883862419696397e-07, + "loss": 0.657, + "step": 1973 + }, + { + "epoch": 1.644569522618303, + "grad_norm": 0.12516506016254425, + "learning_rate": 8.874935325038086e-07, + "loss": 0.5994, + "step": 1974 + }, + { + "epoch": 1.6454033771106942, + "grad_norm": 0.12904782593250275, + "learning_rate": 8.866009138379966e-07, + "loss": 0.6043, + "step": 1975 + }, + { + "epoch": 1.6462372316030853, + "grad_norm": 0.12918344140052795, + "learning_rate": 8.857083866926053e-07, + "loss": 0.5468, + "step": 1976 + }, + { + "epoch": 1.6470710860954765, + "grad_norm": 0.13387003540992737, + "learning_rate": 8.848159517879615e-07, + "loss": 0.6415, + "step": 1977 + }, + { + "epoch": 1.6479049405878674, + "grad_norm": 0.12973016500473022, + "learning_rate": 8.839236098443181e-07, + "loss": 0.6151, + "step": 1978 + }, + { + "epoch": 1.6487387950802583, + "grad_norm": 0.13009996712207794, + "learning_rate": 8.830313615818534e-07, + "loss": 0.5864, + "step": 1979 + }, + { + "epoch": 1.6495726495726495, + "grad_norm": 0.1275186836719513, + "learning_rate": 8.821392077206694e-07, + "loss": 0.5637, + "step": 1980 + }, + { + "epoch": 1.6504065040650406, + "grad_norm": 0.13509085774421692, + "learning_rate": 8.81247148980792e-07, + "loss": 0.568, + "step": 1981 + }, + { + "epoch": 1.6512403585574318, + "grad_norm": 0.12902608513832092, + "learning_rate": 8.803551860821705e-07, + "loss": 0.6641, + "step": 1982 + }, + { + "epoch": 1.652074213049823, + "grad_norm": 0.12584422528743744, + "learning_rate": 8.79463319744677e-07, + "loss": 0.6211, + "step": 1983 + }, + { + "epoch": 1.6529080675422139, + "grad_norm": 0.12634308636188507, + "learning_rate": 8.785715506881056e-07, + "loss": 0.5961, + "step": 1984 + }, + { + "epoch": 1.653741922034605, + "grad_norm": 0.12757942080497742, + "learning_rate": 8.776798796321713e-07, + "loss": 0.6216, + "step": 1985 + }, + { + "epoch": 1.654575776526996, + "grad_norm": 0.1245681568980217, + "learning_rate": 8.767883072965112e-07, + "loss": 0.557, + "step": 1986 + }, + { + "epoch": 1.655409631019387, + "grad_norm": 0.13303357362747192, + "learning_rate": 8.758968344006811e-07, + "loss": 0.6282, + "step": 1987 + }, + { + "epoch": 1.6562434855117782, + "grad_norm": 0.135984405875206, + "learning_rate": 8.750054616641584e-07, + "loss": 0.6216, + "step": 1988 + }, + { + "epoch": 1.6570773400041694, + "grad_norm": 0.13211168348789215, + "learning_rate": 8.741141898063385e-07, + "loss": 0.5596, + "step": 1989 + }, + { + "epoch": 1.6579111944965603, + "grad_norm": 0.13526758551597595, + "learning_rate": 8.732230195465352e-07, + "loss": 0.6068, + "step": 1990 + }, + { + "epoch": 1.6587450489889515, + "grad_norm": 0.1245320737361908, + "learning_rate": 8.723319516039813e-07, + "loss": 0.5431, + "step": 1991 + }, + { + "epoch": 1.6595789034813424, + "grad_norm": 0.13527055084705353, + "learning_rate": 8.714409866978259e-07, + "loss": 0.6634, + "step": 1992 + }, + { + "epoch": 1.6604127579737336, + "grad_norm": 0.1319219022989273, + "learning_rate": 8.705501255471357e-07, + "loss": 0.63, + "step": 1993 + }, + { + "epoch": 1.6612466124661247, + "grad_norm": 0.13556063175201416, + "learning_rate": 8.696593688708938e-07, + "loss": 0.6061, + "step": 1994 + }, + { + "epoch": 1.6620804669585159, + "grad_norm": 0.13379541039466858, + "learning_rate": 8.68768717387998e-07, + "loss": 0.5935, + "step": 1995 + }, + { + "epoch": 1.6629143214509068, + "grad_norm": 0.13060715794563293, + "learning_rate": 8.678781718172622e-07, + "loss": 0.6509, + "step": 1996 + }, + { + "epoch": 1.663748175943298, + "grad_norm": 0.13264665007591248, + "learning_rate": 8.669877328774145e-07, + "loss": 0.6026, + "step": 1997 + }, + { + "epoch": 1.6645820304356889, + "grad_norm": 0.12658947706222534, + "learning_rate": 8.66097401287097e-07, + "loss": 0.5882, + "step": 1998 + }, + { + "epoch": 1.66541588492808, + "grad_norm": 0.12481638789176941, + "learning_rate": 8.652071777648645e-07, + "loss": 0.5412, + "step": 1999 + }, + { + "epoch": 1.6662497394204712, + "grad_norm": 0.13407550752162933, + "learning_rate": 8.643170630291859e-07, + "loss": 0.5861, + "step": 2000 + }, + { + "epoch": 1.6662497394204712, + "eval_loss": 0.6118956804275513, + "eval_runtime": 321.3552, + "eval_samples_per_second": 16.751, + "eval_steps_per_second": 2.794, + "step": 2000 + }, + { + "epoch": 1.6670835939128623, + "grad_norm": 0.13629695773124695, + "learning_rate": 8.63427057798441e-07, + "loss": 0.6098, + "step": 2001 + }, + { + "epoch": 1.6679174484052532, + "grad_norm": 0.12663327157497406, + "learning_rate": 8.625371627909215e-07, + "loss": 0.6582, + "step": 2002 + }, + { + "epoch": 1.6687513028976444, + "grad_norm": 0.12094560265541077, + "learning_rate": 8.616473787248312e-07, + "loss": 0.6911, + "step": 2003 + }, + { + "epoch": 1.6695851573900353, + "grad_norm": 0.13008569180965424, + "learning_rate": 8.607577063182833e-07, + "loss": 0.6286, + "step": 2004 + }, + { + "epoch": 1.6704190118824265, + "grad_norm": 0.12234228849411011, + "learning_rate": 8.598681462893011e-07, + "loss": 0.5862, + "step": 2005 + }, + { + "epoch": 1.6712528663748176, + "grad_norm": 0.12451769411563873, + "learning_rate": 8.589786993558172e-07, + "loss": 0.5655, + "step": 2006 + }, + { + "epoch": 1.6720867208672088, + "grad_norm": 0.12555864453315735, + "learning_rate": 8.58089366235673e-07, + "loss": 0.537, + "step": 2007 + }, + { + "epoch": 1.6729205753595997, + "grad_norm": 0.12302037328481674, + "learning_rate": 8.572001476466182e-07, + "loss": 0.6041, + "step": 2008 + }, + { + "epoch": 1.6737544298519909, + "grad_norm": 0.13534630835056305, + "learning_rate": 8.563110443063098e-07, + "loss": 0.6523, + "step": 2009 + }, + { + "epoch": 1.6745882843443818, + "grad_norm": 0.13176940381526947, + "learning_rate": 8.554220569323117e-07, + "loss": 0.6085, + "step": 2010 + }, + { + "epoch": 1.675422138836773, + "grad_norm": 0.12548188865184784, + "learning_rate": 8.545331862420943e-07, + "loss": 0.6374, + "step": 2011 + }, + { + "epoch": 1.676255993329164, + "grad_norm": 0.12622122466564178, + "learning_rate": 8.536444329530346e-07, + "loss": 0.6144, + "step": 2012 + }, + { + "epoch": 1.6770898478215552, + "grad_norm": 0.12633056938648224, + "learning_rate": 8.527557977824136e-07, + "loss": 0.5992, + "step": 2013 + }, + { + "epoch": 1.6779237023139462, + "grad_norm": 0.1307809054851532, + "learning_rate": 8.518672814474174e-07, + "loss": 0.6072, + "step": 2014 + }, + { + "epoch": 1.6787575568063373, + "grad_norm": 0.1285247951745987, + "learning_rate": 8.509788846651369e-07, + "loss": 0.5737, + "step": 2015 + }, + { + "epoch": 1.6795914112987282, + "grad_norm": 0.12834513187408447, + "learning_rate": 8.500906081525656e-07, + "loss": 0.6045, + "step": 2016 + }, + { + "epoch": 1.6804252657911194, + "grad_norm": 0.12550753355026245, + "learning_rate": 8.492024526265999e-07, + "loss": 0.576, + "step": 2017 + }, + { + "epoch": 1.6812591202835105, + "grad_norm": 0.1294783502817154, + "learning_rate": 8.483144188040399e-07, + "loss": 0.6527, + "step": 2018 + }, + { + "epoch": 1.6820929747759017, + "grad_norm": 0.12203014642000198, + "learning_rate": 8.474265074015856e-07, + "loss": 0.5592, + "step": 2019 + }, + { + "epoch": 1.6829268292682928, + "grad_norm": 0.12280543148517609, + "learning_rate": 8.465387191358398e-07, + "loss": 0.5638, + "step": 2020 + }, + { + "epoch": 1.6837606837606838, + "grad_norm": 0.13221971690654755, + "learning_rate": 8.456510547233048e-07, + "loss": 0.636, + "step": 2021 + }, + { + "epoch": 1.6845945382530747, + "grad_norm": 0.13103988766670227, + "learning_rate": 8.447635148803835e-07, + "loss": 0.5729, + "step": 2022 + }, + { + "epoch": 1.6854283927454659, + "grad_norm": 0.13310641050338745, + "learning_rate": 8.438761003233783e-07, + "loss": 0.6808, + "step": 2023 + }, + { + "epoch": 1.686262247237857, + "grad_norm": 0.13218644261360168, + "learning_rate": 8.429888117684902e-07, + "loss": 0.5781, + "step": 2024 + }, + { + "epoch": 1.6870961017302482, + "grad_norm": 0.1348038613796234, + "learning_rate": 8.421016499318183e-07, + "loss": 0.6018, + "step": 2025 + }, + { + "epoch": 1.6879299562226393, + "grad_norm": 0.12556737661361694, + "learning_rate": 8.412146155293605e-07, + "loss": 0.6126, + "step": 2026 + }, + { + "epoch": 1.6887638107150302, + "grad_norm": 0.12600624561309814, + "learning_rate": 8.403277092770104e-07, + "loss": 0.6138, + "step": 2027 + }, + { + "epoch": 1.6895976652074212, + "grad_norm": 0.1401616632938385, + "learning_rate": 8.394409318905592e-07, + "loss": 0.5923, + "step": 2028 + }, + { + "epoch": 1.6904315196998123, + "grad_norm": 0.12791836261749268, + "learning_rate": 8.385542840856939e-07, + "loss": 0.5626, + "step": 2029 + }, + { + "epoch": 1.6912653741922035, + "grad_norm": 0.12238767743110657, + "learning_rate": 8.376677665779967e-07, + "loss": 0.5742, + "step": 2030 + }, + { + "epoch": 1.6920992286845946, + "grad_norm": 0.12230707705020905, + "learning_rate": 8.367813800829442e-07, + "loss": 0.6064, + "step": 2031 + }, + { + "epoch": 1.6929330831769858, + "grad_norm": 0.12692691385746002, + "learning_rate": 8.358951253159089e-07, + "loss": 0.6256, + "step": 2032 + }, + { + "epoch": 1.6937669376693767, + "grad_norm": 0.12948474287986755, + "learning_rate": 8.350090029921551e-07, + "loss": 0.6033, + "step": 2033 + }, + { + "epoch": 1.6946007921617676, + "grad_norm": 0.12658190727233887, + "learning_rate": 8.341230138268415e-07, + "loss": 0.5697, + "step": 2034 + }, + { + "epoch": 1.6954346466541588, + "grad_norm": 0.12975440919399261, + "learning_rate": 8.332371585350185e-07, + "loss": 0.6117, + "step": 2035 + }, + { + "epoch": 1.69626850114655, + "grad_norm": 0.11741238087415695, + "learning_rate": 8.323514378316288e-07, + "loss": 0.5446, + "step": 2036 + }, + { + "epoch": 1.697102355638941, + "grad_norm": 0.12554077804088593, + "learning_rate": 8.314658524315068e-07, + "loss": 0.5718, + "step": 2037 + }, + { + "epoch": 1.6979362101313322, + "grad_norm": 0.12683430314064026, + "learning_rate": 8.30580403049377e-07, + "loss": 0.6252, + "step": 2038 + }, + { + "epoch": 1.6987700646237232, + "grad_norm": 0.12512685358524323, + "learning_rate": 8.296950903998549e-07, + "loss": 0.6054, + "step": 2039 + }, + { + "epoch": 1.699603919116114, + "grad_norm": 0.13208059966564178, + "learning_rate": 8.288099151974446e-07, + "loss": 0.6566, + "step": 2040 + }, + { + "epoch": 1.7004377736085052, + "grad_norm": 0.1311858892440796, + "learning_rate": 8.279248781565407e-07, + "loss": 0.6073, + "step": 2041 + }, + { + "epoch": 1.7012716281008964, + "grad_norm": 0.12754926085472107, + "learning_rate": 8.27039979991425e-07, + "loss": 0.5843, + "step": 2042 + }, + { + "epoch": 1.7021054825932875, + "grad_norm": 0.12932710349559784, + "learning_rate": 8.261552214162678e-07, + "loss": 0.5742, + "step": 2043 + }, + { + "epoch": 1.7029393370856787, + "grad_norm": 0.12164090573787689, + "learning_rate": 8.252706031451269e-07, + "loss": 0.6054, + "step": 2044 + }, + { + "epoch": 1.7037731915780696, + "grad_norm": 0.1326964795589447, + "learning_rate": 8.243861258919465e-07, + "loss": 0.5884, + "step": 2045 + }, + { + "epoch": 1.7046070460704605, + "grad_norm": 0.12942658364772797, + "learning_rate": 8.235017903705568e-07, + "loss": 0.5781, + "step": 2046 + }, + { + "epoch": 1.7054409005628517, + "grad_norm": 0.12789803743362427, + "learning_rate": 8.226175972946746e-07, + "loss": 0.6674, + "step": 2047 + }, + { + "epoch": 1.7062747550552428, + "grad_norm": 0.13167151808738708, + "learning_rate": 8.217335473779007e-07, + "loss": 0.6454, + "step": 2048 + }, + { + "epoch": 1.707108609547634, + "grad_norm": 0.13229158520698547, + "learning_rate": 8.208496413337209e-07, + "loss": 0.6279, + "step": 2049 + }, + { + "epoch": 1.7079424640400251, + "grad_norm": 0.12817689776420593, + "learning_rate": 8.199658798755047e-07, + "loss": 0.6782, + "step": 2050 + }, + { + "epoch": 1.708776318532416, + "grad_norm": 0.13753139972686768, + "learning_rate": 8.190822637165047e-07, + "loss": 0.6016, + "step": 2051 + }, + { + "epoch": 1.7096101730248072, + "grad_norm": 0.12560203671455383, + "learning_rate": 8.181987935698568e-07, + "loss": 0.6165, + "step": 2052 + }, + { + "epoch": 1.7104440275171982, + "grad_norm": 0.13002927601337433, + "learning_rate": 8.173154701485786e-07, + "loss": 0.6121, + "step": 2053 + }, + { + "epoch": 1.7112778820095893, + "grad_norm": 0.13231289386749268, + "learning_rate": 8.164322941655692e-07, + "loss": 0.6832, + "step": 2054 + }, + { + "epoch": 1.7121117365019805, + "grad_norm": 0.12360356748104095, + "learning_rate": 8.155492663336093e-07, + "loss": 0.5825, + "step": 2055 + }, + { + "epoch": 1.7129455909943716, + "grad_norm": 0.12095890194177628, + "learning_rate": 8.146663873653596e-07, + "loss": 0.6056, + "step": 2056 + }, + { + "epoch": 1.7137794454867625, + "grad_norm": 0.1325632631778717, + "learning_rate": 8.137836579733604e-07, + "loss": 0.6321, + "step": 2057 + }, + { + "epoch": 1.7146132999791537, + "grad_norm": 0.1254180669784546, + "learning_rate": 8.129010788700317e-07, + "loss": 0.5731, + "step": 2058 + }, + { + "epoch": 1.7154471544715446, + "grad_norm": 0.13764861226081848, + "learning_rate": 8.120186507676724e-07, + "loss": 0.5629, + "step": 2059 + }, + { + "epoch": 1.7162810089639358, + "grad_norm": 0.13155192136764526, + "learning_rate": 8.111363743784586e-07, + "loss": 0.6458, + "step": 2060 + }, + { + "epoch": 1.717114863456327, + "grad_norm": 0.13540032505989075, + "learning_rate": 8.102542504144454e-07, + "loss": 0.585, + "step": 2061 + }, + { + "epoch": 1.717948717948718, + "grad_norm": 0.1272411197423935, + "learning_rate": 8.093722795875635e-07, + "loss": 0.5677, + "step": 2062 + }, + { + "epoch": 1.718782572441109, + "grad_norm": 0.13317130506038666, + "learning_rate": 8.084904626096209e-07, + "loss": 0.6681, + "step": 2063 + }, + { + "epoch": 1.7196164269335001, + "grad_norm": 0.12464749813079834, + "learning_rate": 8.076088001923012e-07, + "loss": 0.6021, + "step": 2064 + }, + { + "epoch": 1.720450281425891, + "grad_norm": 0.12794774770736694, + "learning_rate": 8.06727293047163e-07, + "loss": 0.6102, + "step": 2065 + }, + { + "epoch": 1.7212841359182822, + "grad_norm": 0.11895739287137985, + "learning_rate": 8.058459418856393e-07, + "loss": 0.538, + "step": 2066 + }, + { + "epoch": 1.7221179904106734, + "grad_norm": 0.12413477897644043, + "learning_rate": 8.049647474190383e-07, + "loss": 0.5926, + "step": 2067 + }, + { + "epoch": 1.7229518449030645, + "grad_norm": 0.1245928704738617, + "learning_rate": 8.04083710358541e-07, + "loss": 0.5464, + "step": 2068 + }, + { + "epoch": 1.7237856993954555, + "grad_norm": 0.12484681606292725, + "learning_rate": 8.032028314152012e-07, + "loss": 0.5695, + "step": 2069 + }, + { + "epoch": 1.7246195538878466, + "grad_norm": 0.12726366519927979, + "learning_rate": 8.023221112999458e-07, + "loss": 0.5966, + "step": 2070 + }, + { + "epoch": 1.7254534083802375, + "grad_norm": 0.12166636437177658, + "learning_rate": 8.014415507235727e-07, + "loss": 0.6285, + "step": 2071 + }, + { + "epoch": 1.7262872628726287, + "grad_norm": 0.1254647970199585, + "learning_rate": 8.005611503967515e-07, + "loss": 0.6124, + "step": 2072 + }, + { + "epoch": 1.7271211173650198, + "grad_norm": 0.12607796490192413, + "learning_rate": 7.996809110300226e-07, + "loss": 0.5675, + "step": 2073 + }, + { + "epoch": 1.727954971857411, + "grad_norm": 0.13107630610466003, + "learning_rate": 7.988008333337961e-07, + "loss": 0.6774, + "step": 2074 + }, + { + "epoch": 1.728788826349802, + "grad_norm": 0.12005145847797394, + "learning_rate": 7.979209180183514e-07, + "loss": 0.6335, + "step": 2075 + }, + { + "epoch": 1.729622680842193, + "grad_norm": 0.11902257800102234, + "learning_rate": 7.970411657938381e-07, + "loss": 0.5516, + "step": 2076 + }, + { + "epoch": 1.730456535334584, + "grad_norm": 0.11752620339393616, + "learning_rate": 7.961615773702727e-07, + "loss": 0.5586, + "step": 2077 + }, + { + "epoch": 1.7312903898269751, + "grad_norm": 0.12146768718957901, + "learning_rate": 7.952821534575405e-07, + "loss": 0.6039, + "step": 2078 + }, + { + "epoch": 1.7321242443193663, + "grad_norm": 0.1350381076335907, + "learning_rate": 7.944028947653936e-07, + "loss": 0.5445, + "step": 2079 + }, + { + "epoch": 1.7329580988117574, + "grad_norm": 0.11824685335159302, + "learning_rate": 7.935238020034503e-07, + "loss": 0.5534, + "step": 2080 + }, + { + "epoch": 1.7337919533041486, + "grad_norm": 0.12622307240962982, + "learning_rate": 7.926448758811963e-07, + "loss": 0.5945, + "step": 2081 + }, + { + "epoch": 1.7346258077965395, + "grad_norm": 0.12936830520629883, + "learning_rate": 7.917661171079816e-07, + "loss": 0.5754, + "step": 2082 + }, + { + "epoch": 1.7354596622889304, + "grad_norm": 0.12486603856086731, + "learning_rate": 7.908875263930214e-07, + "loss": 0.6099, + "step": 2083 + }, + { + "epoch": 1.7362935167813216, + "grad_norm": 0.13117629289627075, + "learning_rate": 7.900091044453956e-07, + "loss": 0.6207, + "step": 2084 + }, + { + "epoch": 1.7371273712737128, + "grad_norm": 0.1298520714044571, + "learning_rate": 7.891308519740479e-07, + "loss": 0.5873, + "step": 2085 + }, + { + "epoch": 1.737961225766104, + "grad_norm": 0.12560497224330902, + "learning_rate": 7.882527696877846e-07, + "loss": 0.5623, + "step": 2086 + }, + { + "epoch": 1.738795080258495, + "grad_norm": 0.13026194274425507, + "learning_rate": 7.873748582952753e-07, + "loss": 0.6185, + "step": 2087 + }, + { + "epoch": 1.739628934750886, + "grad_norm": 0.13659796118736267, + "learning_rate": 7.864971185050515e-07, + "loss": 0.6017, + "step": 2088 + }, + { + "epoch": 1.740462789243277, + "grad_norm": 0.13491465151309967, + "learning_rate": 7.856195510255058e-07, + "loss": 0.5749, + "step": 2089 + }, + { + "epoch": 1.741296643735668, + "grad_norm": 0.12798424065113068, + "learning_rate": 7.847421565648928e-07, + "loss": 0.6356, + "step": 2090 + }, + { + "epoch": 1.7421304982280592, + "grad_norm": 0.12282812595367432, + "learning_rate": 7.838649358313262e-07, + "loss": 0.5922, + "step": 2091 + }, + { + "epoch": 1.7429643527204504, + "grad_norm": 0.1255982518196106, + "learning_rate": 7.829878895327804e-07, + "loss": 0.645, + "step": 2092 + }, + { + "epoch": 1.7437982072128415, + "grad_norm": 0.1324257254600525, + "learning_rate": 7.821110183770884e-07, + "loss": 0.5499, + "step": 2093 + }, + { + "epoch": 1.7446320617052324, + "grad_norm": 0.12864653766155243, + "learning_rate": 7.812343230719422e-07, + "loss": 0.5602, + "step": 2094 + }, + { + "epoch": 1.7454659161976234, + "grad_norm": 0.13943764567375183, + "learning_rate": 7.803578043248918e-07, + "loss": 0.6665, + "step": 2095 + }, + { + "epoch": 1.7462997706900145, + "grad_norm": 0.12609517574310303, + "learning_rate": 7.794814628433448e-07, + "loss": 0.6295, + "step": 2096 + }, + { + "epoch": 1.7471336251824057, + "grad_norm": 0.12663273513317108, + "learning_rate": 7.786052993345655e-07, + "loss": 0.5585, + "step": 2097 + }, + { + "epoch": 1.7479674796747968, + "grad_norm": 0.13500264286994934, + "learning_rate": 7.777293145056745e-07, + "loss": 0.6277, + "step": 2098 + }, + { + "epoch": 1.748801334167188, + "grad_norm": 0.12323655933141708, + "learning_rate": 7.76853509063649e-07, + "loss": 0.5386, + "step": 2099 + }, + { + "epoch": 1.749635188659579, + "grad_norm": 0.13133323192596436, + "learning_rate": 7.759778837153204e-07, + "loss": 0.5523, + "step": 2100 + }, + { + "epoch": 1.7504690431519698, + "grad_norm": 0.12879957258701324, + "learning_rate": 7.751024391673749e-07, + "loss": 0.6167, + "step": 2101 + }, + { + "epoch": 1.751302897644361, + "grad_norm": 0.13135939836502075, + "learning_rate": 7.742271761263536e-07, + "loss": 0.6001, + "step": 2102 + }, + { + "epoch": 1.7521367521367521, + "grad_norm": 0.13290925323963165, + "learning_rate": 7.733520952986505e-07, + "loss": 0.5734, + "step": 2103 + }, + { + "epoch": 1.7529706066291433, + "grad_norm": 0.1278628557920456, + "learning_rate": 7.724771973905119e-07, + "loss": 0.6017, + "step": 2104 + }, + { + "epoch": 1.7538044611215344, + "grad_norm": 0.12137078493833542, + "learning_rate": 7.716024831080383e-07, + "loss": 0.5904, + "step": 2105 + }, + { + "epoch": 1.7546383156139254, + "grad_norm": 0.12834006547927856, + "learning_rate": 7.707279531571799e-07, + "loss": 0.5976, + "step": 2106 + }, + { + "epoch": 1.7554721701063163, + "grad_norm": 0.12812873721122742, + "learning_rate": 7.698536082437399e-07, + "loss": 0.577, + "step": 2107 + }, + { + "epoch": 1.7563060245987074, + "grad_norm": 0.12574169039726257, + "learning_rate": 7.68979449073371e-07, + "loss": 0.6299, + "step": 2108 + }, + { + "epoch": 1.7571398790910986, + "grad_norm": 0.14230380952358246, + "learning_rate": 7.681054763515759e-07, + "loss": 0.5561, + "step": 2109 + }, + { + "epoch": 1.7579737335834897, + "grad_norm": 0.1346198320388794, + "learning_rate": 7.672316907837084e-07, + "loss": 0.6105, + "step": 2110 + }, + { + "epoch": 1.758807588075881, + "grad_norm": 0.12239620089530945, + "learning_rate": 7.663580930749692e-07, + "loss": 0.5807, + "step": 2111 + }, + { + "epoch": 1.7596414425682718, + "grad_norm": 0.13581721484661102, + "learning_rate": 7.654846839304086e-07, + "loss": 0.6454, + "step": 2112 + }, + { + "epoch": 1.760475297060663, + "grad_norm": 0.13057507574558258, + "learning_rate": 7.646114640549246e-07, + "loss": 0.5808, + "step": 2113 + }, + { + "epoch": 1.761309151553054, + "grad_norm": 0.1373625248670578, + "learning_rate": 7.637384341532622e-07, + "loss": 0.6481, + "step": 2114 + }, + { + "epoch": 1.762143006045445, + "grad_norm": 0.14396755397319794, + "learning_rate": 7.628655949300132e-07, + "loss": 0.5838, + "step": 2115 + }, + { + "epoch": 1.7629768605378362, + "grad_norm": 0.13005192577838898, + "learning_rate": 7.619929470896154e-07, + "loss": 0.5892, + "step": 2116 + }, + { + "epoch": 1.7638107150302273, + "grad_norm": 0.1325499266386032, + "learning_rate": 7.611204913363522e-07, + "loss": 0.6032, + "step": 2117 + }, + { + "epoch": 1.7646445695226183, + "grad_norm": 0.12955497205257416, + "learning_rate": 7.602482283743523e-07, + "loss": 0.5703, + "step": 2118 + }, + { + "epoch": 1.7654784240150094, + "grad_norm": 0.11970066279172897, + "learning_rate": 7.593761589075879e-07, + "loss": 0.5528, + "step": 2119 + }, + { + "epoch": 1.7663122785074004, + "grad_norm": 0.12704381346702576, + "learning_rate": 7.585042836398762e-07, + "loss": 0.5799, + "step": 2120 + }, + { + "epoch": 1.7671461329997915, + "grad_norm": 0.1244860365986824, + "learning_rate": 7.576326032748771e-07, + "loss": 0.5843, + "step": 2121 + }, + { + "epoch": 1.7679799874921827, + "grad_norm": 0.13251663744449615, + "learning_rate": 7.567611185160931e-07, + "loss": 0.5768, + "step": 2122 + }, + { + "epoch": 1.7688138419845738, + "grad_norm": 0.12903398275375366, + "learning_rate": 7.55889830066869e-07, + "loss": 0.56, + "step": 2123 + }, + { + "epoch": 1.7696476964769647, + "grad_norm": 0.13495059311389923, + "learning_rate": 7.550187386303908e-07, + "loss": 0.6376, + "step": 2124 + }, + { + "epoch": 1.7704815509693559, + "grad_norm": 0.1355675309896469, + "learning_rate": 7.541478449096861e-07, + "loss": 0.6375, + "step": 2125 + }, + { + "epoch": 1.7713154054617468, + "grad_norm": 0.12867653369903564, + "learning_rate": 7.532771496076226e-07, + "loss": 0.5449, + "step": 2126 + }, + { + "epoch": 1.772149259954138, + "grad_norm": 0.13269716501235962, + "learning_rate": 7.524066534269078e-07, + "loss": 0.6504, + "step": 2127 + }, + { + "epoch": 1.7729831144465291, + "grad_norm": 0.12922774255275726, + "learning_rate": 7.515363570700889e-07, + "loss": 0.5892, + "step": 2128 + }, + { + "epoch": 1.7738169689389203, + "grad_norm": 0.1222226545214653, + "learning_rate": 7.506662612395513e-07, + "loss": 0.5652, + "step": 2129 + }, + { + "epoch": 1.7746508234313112, + "grad_norm": 0.1284438967704773, + "learning_rate": 7.497963666375185e-07, + "loss": 0.5996, + "step": 2130 + }, + { + "epoch": 1.7754846779237023, + "grad_norm": 0.1273675560951233, + "learning_rate": 7.489266739660523e-07, + "loss": 0.6324, + "step": 2131 + }, + { + "epoch": 1.7763185324160933, + "grad_norm": 0.13269834220409393, + "learning_rate": 7.480571839270512e-07, + "loss": 0.6446, + "step": 2132 + }, + { + "epoch": 1.7771523869084844, + "grad_norm": 0.12966395914554596, + "learning_rate": 7.471878972222494e-07, + "loss": 0.5247, + "step": 2133 + }, + { + "epoch": 1.7779862414008756, + "grad_norm": 0.12394700199365616, + "learning_rate": 7.463188145532185e-07, + "loss": 0.6179, + "step": 2134 + }, + { + "epoch": 1.7788200958932667, + "grad_norm": 0.13124272227287292, + "learning_rate": 7.454499366213642e-07, + "loss": 0.5627, + "step": 2135 + }, + { + "epoch": 1.7796539503856577, + "grad_norm": 0.13032935559749603, + "learning_rate": 7.445812641279274e-07, + "loss": 0.5894, + "step": 2136 + }, + { + "epoch": 1.7804878048780488, + "grad_norm": 0.13481009006500244, + "learning_rate": 7.437127977739835e-07, + "loss": 0.6706, + "step": 2137 + }, + { + "epoch": 1.7813216593704397, + "grad_norm": 0.1252714842557907, + "learning_rate": 7.428445382604405e-07, + "loss": 0.6167, + "step": 2138 + }, + { + "epoch": 1.7821555138628309, + "grad_norm": 0.13593217730522156, + "learning_rate": 7.419764862880406e-07, + "loss": 0.6212, + "step": 2139 + }, + { + "epoch": 1.782989368355222, + "grad_norm": 0.128921777009964, + "learning_rate": 7.411086425573583e-07, + "loss": 0.5644, + "step": 2140 + }, + { + "epoch": 1.7838232228476132, + "grad_norm": 0.1279333531856537, + "learning_rate": 7.402410077687993e-07, + "loss": 0.6109, + "step": 2141 + }, + { + "epoch": 1.7846570773400041, + "grad_norm": 0.1240469217300415, + "learning_rate": 7.393735826226013e-07, + "loss": 0.6266, + "step": 2142 + }, + { + "epoch": 1.7854909318323953, + "grad_norm": 0.12809887528419495, + "learning_rate": 7.38506367818833e-07, + "loss": 0.6296, + "step": 2143 + }, + { + "epoch": 1.7863247863247862, + "grad_norm": 0.12371251732110977, + "learning_rate": 7.376393640573927e-07, + "loss": 0.5797, + "step": 2144 + }, + { + "epoch": 1.7871586408171773, + "grad_norm": 0.13013000786304474, + "learning_rate": 7.367725720380086e-07, + "loss": 0.5641, + "step": 2145 + }, + { + "epoch": 1.7879924953095685, + "grad_norm": 0.12911874055862427, + "learning_rate": 7.359059924602385e-07, + "loss": 0.6069, + "step": 2146 + }, + { + "epoch": 1.7888263498019596, + "grad_norm": 0.12248188257217407, + "learning_rate": 7.35039626023468e-07, + "loss": 0.5962, + "step": 2147 + }, + { + "epoch": 1.7896602042943508, + "grad_norm": 0.1306978166103363, + "learning_rate": 7.34173473426911e-07, + "loss": 0.6135, + "step": 2148 + }, + { + "epoch": 1.7904940587867417, + "grad_norm": 0.11989414691925049, + "learning_rate": 7.333075353696089e-07, + "loss": 0.5457, + "step": 2149 + }, + { + "epoch": 1.7913279132791327, + "grad_norm": 0.1320810168981552, + "learning_rate": 7.324418125504304e-07, + "loss": 0.616, + "step": 2150 + }, + { + "epoch": 1.7921617677715238, + "grad_norm": 0.1229398101568222, + "learning_rate": 7.315763056680694e-07, + "loss": 0.5835, + "step": 2151 + }, + { + "epoch": 1.792995622263915, + "grad_norm": 0.12493004649877548, + "learning_rate": 7.307110154210463e-07, + "loss": 0.5619, + "step": 2152 + }, + { + "epoch": 1.793829476756306, + "grad_norm": 0.12893696129322052, + "learning_rate": 7.298459425077063e-07, + "loss": 0.603, + "step": 2153 + }, + { + "epoch": 1.7946633312486973, + "grad_norm": 0.1281341016292572, + "learning_rate": 7.289810876262199e-07, + "loss": 0.56, + "step": 2154 + }, + { + "epoch": 1.7954971857410882, + "grad_norm": 0.1304667741060257, + "learning_rate": 7.281164514745805e-07, + "loss": 0.539, + "step": 2155 + }, + { + "epoch": 1.7963310402334791, + "grad_norm": 0.12886284291744232, + "learning_rate": 7.272520347506058e-07, + "loss": 0.6319, + "step": 2156 + }, + { + "epoch": 1.7971648947258703, + "grad_norm": 0.12648077309131622, + "learning_rate": 7.263878381519364e-07, + "loss": 0.5886, + "step": 2157 + }, + { + "epoch": 1.7979987492182614, + "grad_norm": 0.12707141041755676, + "learning_rate": 7.255238623760347e-07, + "loss": 0.5575, + "step": 2158 + }, + { + "epoch": 1.7988326037106526, + "grad_norm": 0.13646213710308075, + "learning_rate": 7.24660108120185e-07, + "loss": 0.5984, + "step": 2159 + }, + { + "epoch": 1.7996664582030437, + "grad_norm": 0.1271483451128006, + "learning_rate": 7.237965760814939e-07, + "loss": 0.5451, + "step": 2160 + }, + { + "epoch": 1.8005003126954346, + "grad_norm": 0.1413080096244812, + "learning_rate": 7.22933266956887e-07, + "loss": 0.609, + "step": 2161 + }, + { + "epoch": 1.8013341671878256, + "grad_norm": 0.1268233358860016, + "learning_rate": 7.220701814431109e-07, + "loss": 0.5813, + "step": 2162 + }, + { + "epoch": 1.8021680216802167, + "grad_norm": 0.13267411291599274, + "learning_rate": 7.212073202367319e-07, + "loss": 0.6314, + "step": 2163 + }, + { + "epoch": 1.8030018761726079, + "grad_norm": 0.12609542906284332, + "learning_rate": 7.203446840341347e-07, + "loss": 0.5809, + "step": 2164 + }, + { + "epoch": 1.803835730664999, + "grad_norm": 0.12579114735126495, + "learning_rate": 7.194822735315229e-07, + "loss": 0.6272, + "step": 2165 + }, + { + "epoch": 1.8046695851573902, + "grad_norm": 0.1351184844970703, + "learning_rate": 7.186200894249177e-07, + "loss": 0.5507, + "step": 2166 + }, + { + "epoch": 1.805503439649781, + "grad_norm": 0.13242338597774506, + "learning_rate": 7.177581324101575e-07, + "loss": 0.604, + "step": 2167 + }, + { + "epoch": 1.806337294142172, + "grad_norm": 0.13314031064510345, + "learning_rate": 7.168964031828976e-07, + "loss": 0.6186, + "step": 2168 + }, + { + "epoch": 1.8071711486345632, + "grad_norm": 0.12832415103912354, + "learning_rate": 7.160349024386095e-07, + "loss": 0.5621, + "step": 2169 + }, + { + "epoch": 1.8080050031269543, + "grad_norm": 0.13182200491428375, + "learning_rate": 7.151736308725801e-07, + "loss": 0.6296, + "step": 2170 + }, + { + "epoch": 1.8088388576193455, + "grad_norm": 0.13076068460941315, + "learning_rate": 7.143125891799111e-07, + "loss": 0.5908, + "step": 2171 + }, + { + "epoch": 1.8096727121117366, + "grad_norm": 0.1297273486852646, + "learning_rate": 7.134517780555197e-07, + "loss": 0.5424, + "step": 2172 + }, + { + "epoch": 1.8105065666041276, + "grad_norm": 0.13636283576488495, + "learning_rate": 7.125911981941359e-07, + "loss": 0.6602, + "step": 2173 + }, + { + "epoch": 1.8113404210965185, + "grad_norm": 0.13110437989234924, + "learning_rate": 7.117308502903035e-07, + "loss": 0.5969, + "step": 2174 + }, + { + "epoch": 1.8121742755889096, + "grad_norm": 0.1326896697282791, + "learning_rate": 7.108707350383791e-07, + "loss": 0.6131, + "step": 2175 + }, + { + "epoch": 1.8130081300813008, + "grad_norm": 0.13135559856891632, + "learning_rate": 7.100108531325317e-07, + "loss": 0.5867, + "step": 2176 + }, + { + "epoch": 1.813841984573692, + "grad_norm": 0.12924109399318695, + "learning_rate": 7.091512052667412e-07, + "loss": 0.6186, + "step": 2177 + }, + { + "epoch": 1.814675839066083, + "grad_norm": 0.12946565449237823, + "learning_rate": 7.082917921348001e-07, + "loss": 0.6381, + "step": 2178 + }, + { + "epoch": 1.815509693558474, + "grad_norm": 0.12695160508155823, + "learning_rate": 7.0743261443031e-07, + "loss": 0.5888, + "step": 2179 + }, + { + "epoch": 1.8163435480508652, + "grad_norm": 0.12655752897262573, + "learning_rate": 7.065736728466832e-07, + "loss": 0.6633, + "step": 2180 + }, + { + "epoch": 1.817177402543256, + "grad_norm": 0.13403771817684174, + "learning_rate": 7.057149680771413e-07, + "loss": 0.608, + "step": 2181 + }, + { + "epoch": 1.8180112570356473, + "grad_norm": 0.12965700030326843, + "learning_rate": 7.048565008147146e-07, + "loss": 0.6786, + "step": 2182 + }, + { + "epoch": 1.8188451115280384, + "grad_norm": 0.1377706527709961, + "learning_rate": 7.039982717522422e-07, + "loss": 0.6328, + "step": 2183 + }, + { + "epoch": 1.8196789660204296, + "grad_norm": 0.14265958964824677, + "learning_rate": 7.031402815823704e-07, + "loss": 0.6493, + "step": 2184 + }, + { + "epoch": 1.8205128205128205, + "grad_norm": 0.13361769914627075, + "learning_rate": 7.022825309975529e-07, + "loss": 0.618, + "step": 2185 + }, + { + "epoch": 1.8213466750052116, + "grad_norm": 0.12783940136432648, + "learning_rate": 7.014250206900502e-07, + "loss": 0.6005, + "step": 2186 + }, + { + "epoch": 1.8221805294976026, + "grad_norm": 0.12289536744356155, + "learning_rate": 7.005677513519288e-07, + "loss": 0.577, + "step": 2187 + }, + { + "epoch": 1.8230143839899937, + "grad_norm": 0.12831023335456848, + "learning_rate": 6.997107236750604e-07, + "loss": 0.5951, + "step": 2188 + }, + { + "epoch": 1.8238482384823849, + "grad_norm": 0.12774531543254852, + "learning_rate": 6.988539383511222e-07, + "loss": 0.6544, + "step": 2189 + }, + { + "epoch": 1.824682092974776, + "grad_norm": 0.13053590059280396, + "learning_rate": 6.979973960715957e-07, + "loss": 0.5647, + "step": 2190 + }, + { + "epoch": 1.825515947467167, + "grad_norm": 0.12190835922956467, + "learning_rate": 6.971410975277654e-07, + "loss": 0.6, + "step": 2191 + }, + { + "epoch": 1.826349801959558, + "grad_norm": 0.13813957571983337, + "learning_rate": 6.962850434107206e-07, + "loss": 0.6417, + "step": 2192 + }, + { + "epoch": 1.827183656451949, + "grad_norm": 0.13274197280406952, + "learning_rate": 6.954292344113521e-07, + "loss": 0.6578, + "step": 2193 + }, + { + "epoch": 1.8280175109443402, + "grad_norm": 0.12455936521291733, + "learning_rate": 6.945736712203536e-07, + "loss": 0.5857, + "step": 2194 + }, + { + "epoch": 1.8288513654367313, + "grad_norm": 0.13513398170471191, + "learning_rate": 6.937183545282198e-07, + "loss": 0.5971, + "step": 2195 + }, + { + "epoch": 1.8296852199291225, + "grad_norm": 0.1251787394285202, + "learning_rate": 6.928632850252467e-07, + "loss": 0.6098, + "step": 2196 + }, + { + "epoch": 1.8305190744215134, + "grad_norm": 0.12823201715946198, + "learning_rate": 6.920084634015313e-07, + "loss": 0.6184, + "step": 2197 + }, + { + "epoch": 1.8313529289139046, + "grad_norm": 0.12458322197198868, + "learning_rate": 6.911538903469697e-07, + "loss": 0.5409, + "step": 2198 + }, + { + "epoch": 1.8321867834062955, + "grad_norm": 0.13494035601615906, + "learning_rate": 6.902995665512581e-07, + "loss": 0.5969, + "step": 2199 + }, + { + "epoch": 1.8330206378986866, + "grad_norm": 0.1326414942741394, + "learning_rate": 6.894454927038906e-07, + "loss": 0.5327, + "step": 2200 + }, + { + "epoch": 1.8338544923910778, + "grad_norm": 0.12999312579631805, + "learning_rate": 6.885916694941612e-07, + "loss": 0.6224, + "step": 2201 + }, + { + "epoch": 1.834688346883469, + "grad_norm": 0.12325473874807358, + "learning_rate": 6.8773809761116e-07, + "loss": 0.5234, + "step": 2202 + }, + { + "epoch": 1.8355222013758599, + "grad_norm": 0.12891866266727448, + "learning_rate": 6.868847777437747e-07, + "loss": 0.5782, + "step": 2203 + }, + { + "epoch": 1.836356055868251, + "grad_norm": 0.1282792091369629, + "learning_rate": 6.860317105806905e-07, + "loss": 0.5817, + "step": 2204 + }, + { + "epoch": 1.837189910360642, + "grad_norm": 0.12752637267112732, + "learning_rate": 6.851788968103876e-07, + "loss": 0.6053, + "step": 2205 + }, + { + "epoch": 1.838023764853033, + "grad_norm": 0.1264016330242157, + "learning_rate": 6.843263371211414e-07, + "loss": 0.5788, + "step": 2206 + }, + { + "epoch": 1.8388576193454242, + "grad_norm": 0.12685662508010864, + "learning_rate": 6.83474032201024e-07, + "loss": 0.5849, + "step": 2207 + }, + { + "epoch": 1.8396914738378154, + "grad_norm": 0.1329505294561386, + "learning_rate": 6.826219827379002e-07, + "loss": 0.5886, + "step": 2208 + }, + { + "epoch": 1.8405253283302065, + "grad_norm": 0.13913948833942413, + "learning_rate": 6.817701894194293e-07, + "loss": 0.6189, + "step": 2209 + }, + { + "epoch": 1.8413591828225975, + "grad_norm": 0.12883232533931732, + "learning_rate": 6.809186529330639e-07, + "loss": 0.5634, + "step": 2210 + }, + { + "epoch": 1.8421930373149884, + "grad_norm": 0.12113146483898163, + "learning_rate": 6.800673739660488e-07, + "loss": 0.597, + "step": 2211 + }, + { + "epoch": 1.8430268918073796, + "grad_norm": 0.1325560361146927, + "learning_rate": 6.792163532054217e-07, + "loss": 0.551, + "step": 2212 + }, + { + "epoch": 1.8438607462997707, + "grad_norm": 0.14157071709632874, + "learning_rate": 6.783655913380115e-07, + "loss": 0.563, + "step": 2213 + }, + { + "epoch": 1.8446946007921619, + "grad_norm": 0.12442418932914734, + "learning_rate": 6.775150890504381e-07, + "loss": 0.5486, + "step": 2214 + }, + { + "epoch": 1.845528455284553, + "grad_norm": 0.1348833441734314, + "learning_rate": 6.766648470291123e-07, + "loss": 0.5977, + "step": 2215 + }, + { + "epoch": 1.846362309776944, + "grad_norm": 0.12434475868940353, + "learning_rate": 6.758148659602344e-07, + "loss": 0.5013, + "step": 2216 + }, + { + "epoch": 1.8471961642693349, + "grad_norm": 0.12582679092884064, + "learning_rate": 6.749651465297943e-07, + "loss": 0.5939, + "step": 2217 + }, + { + "epoch": 1.848030018761726, + "grad_norm": 0.12978383898735046, + "learning_rate": 6.741156894235705e-07, + "loss": 0.6158, + "step": 2218 + }, + { + "epoch": 1.8488638732541172, + "grad_norm": 0.1360258311033249, + "learning_rate": 6.732664953271305e-07, + "loss": 0.6059, + "step": 2219 + }, + { + "epoch": 1.8496977277465083, + "grad_norm": 0.13842739164829254, + "learning_rate": 6.724175649258286e-07, + "loss": 0.5801, + "step": 2220 + }, + { + "epoch": 1.8505315822388995, + "grad_norm": 0.12137693911790848, + "learning_rate": 6.715688989048065e-07, + "loss": 0.5945, + "step": 2221 + }, + { + "epoch": 1.8513654367312904, + "grad_norm": 0.1248222142457962, + "learning_rate": 6.707204979489933e-07, + "loss": 0.5963, + "step": 2222 + }, + { + "epoch": 1.8521992912236813, + "grad_norm": 0.12999458611011505, + "learning_rate": 6.698723627431037e-07, + "loss": 0.6183, + "step": 2223 + }, + { + "epoch": 1.8530331457160725, + "grad_norm": 0.13486354053020477, + "learning_rate": 6.690244939716375e-07, + "loss": 0.6077, + "step": 2224 + }, + { + "epoch": 1.8538670002084636, + "grad_norm": 0.12799599766731262, + "learning_rate": 6.681768923188798e-07, + "loss": 0.6049, + "step": 2225 + }, + { + "epoch": 1.8547008547008548, + "grad_norm": 0.13147741556167603, + "learning_rate": 6.673295584689003e-07, + "loss": 0.6467, + "step": 2226 + }, + { + "epoch": 1.855534709193246, + "grad_norm": 0.13251982629299164, + "learning_rate": 6.664824931055522e-07, + "loss": 0.6342, + "step": 2227 + }, + { + "epoch": 1.8563685636856369, + "grad_norm": 0.13429106771945953, + "learning_rate": 6.656356969124723e-07, + "loss": 0.6237, + "step": 2228 + }, + { + "epoch": 1.8572024181780278, + "grad_norm": 0.1291757971048355, + "learning_rate": 6.647891705730801e-07, + "loss": 0.5665, + "step": 2229 + }, + { + "epoch": 1.858036272670419, + "grad_norm": 0.13031582534313202, + "learning_rate": 6.639429147705772e-07, + "loss": 0.5835, + "step": 2230 + }, + { + "epoch": 1.85887012716281, + "grad_norm": 0.13219709694385529, + "learning_rate": 6.630969301879473e-07, + "loss": 0.5901, + "step": 2231 + }, + { + "epoch": 1.8597039816552012, + "grad_norm": 0.13087095320224762, + "learning_rate": 6.622512175079542e-07, + "loss": 0.6978, + "step": 2232 + }, + { + "epoch": 1.8605378361475924, + "grad_norm": 0.1454305499792099, + "learning_rate": 6.614057774131436e-07, + "loss": 0.5978, + "step": 2233 + }, + { + "epoch": 1.8613716906399833, + "grad_norm": 0.12761370837688446, + "learning_rate": 6.605606105858401e-07, + "loss": 0.5659, + "step": 2234 + }, + { + "epoch": 1.8622055451323742, + "grad_norm": 0.13098569214344025, + "learning_rate": 6.597157177081477e-07, + "loss": 0.5773, + "step": 2235 + }, + { + "epoch": 1.8630393996247654, + "grad_norm": 0.1315004825592041, + "learning_rate": 6.588710994619508e-07, + "loss": 0.5531, + "step": 2236 + }, + { + "epoch": 1.8638732541171565, + "grad_norm": 0.1302976906299591, + "learning_rate": 6.580267565289106e-07, + "loss": 0.5603, + "step": 2237 + }, + { + "epoch": 1.8647071086095477, + "grad_norm": 0.12400949001312256, + "learning_rate": 6.571826895904666e-07, + "loss": 0.5228, + "step": 2238 + }, + { + "epoch": 1.8655409631019388, + "grad_norm": 0.13697467744350433, + "learning_rate": 6.563388993278359e-07, + "loss": 0.6525, + "step": 2239 + }, + { + "epoch": 1.8663748175943298, + "grad_norm": 0.12461115419864655, + "learning_rate": 6.554953864220114e-07, + "loss": 0.6154, + "step": 2240 + }, + { + "epoch": 1.867208672086721, + "grad_norm": 0.13096342980861664, + "learning_rate": 6.546521515537635e-07, + "loss": 0.612, + "step": 2241 + }, + { + "epoch": 1.8680425265791119, + "grad_norm": 0.12971916794776917, + "learning_rate": 6.538091954036369e-07, + "loss": 0.6228, + "step": 2242 + }, + { + "epoch": 1.868876381071503, + "grad_norm": 0.13440938293933868, + "learning_rate": 6.529665186519519e-07, + "loss": 0.6226, + "step": 2243 + }, + { + "epoch": 1.8697102355638942, + "grad_norm": 0.13049529492855072, + "learning_rate": 6.521241219788037e-07, + "loss": 0.6174, + "step": 2244 + }, + { + "epoch": 1.8705440900562853, + "grad_norm": 0.12830443680286407, + "learning_rate": 6.512820060640607e-07, + "loss": 0.5727, + "step": 2245 + }, + { + "epoch": 1.8713779445486762, + "grad_norm": 0.1377989947795868, + "learning_rate": 6.504401715873653e-07, + "loss": 0.6148, + "step": 2246 + }, + { + "epoch": 1.8722117990410674, + "grad_norm": 0.12611956894397736, + "learning_rate": 6.495986192281324e-07, + "loss": 0.5785, + "step": 2247 + }, + { + "epoch": 1.8730456535334583, + "grad_norm": 0.13040220737457275, + "learning_rate": 6.487573496655495e-07, + "loss": 0.6567, + "step": 2248 + }, + { + "epoch": 1.8738795080258495, + "grad_norm": 0.13693012297153473, + "learning_rate": 6.479163635785758e-07, + "loss": 0.5877, + "step": 2249 + }, + { + "epoch": 1.8747133625182406, + "grad_norm": 0.13410882651805878, + "learning_rate": 6.47075661645941e-07, + "loss": 0.6076, + "step": 2250 + }, + { + "epoch": 1.8755472170106318, + "grad_norm": 0.12973110377788544, + "learning_rate": 6.462352445461467e-07, + "loss": 0.6246, + "step": 2251 + }, + { + "epoch": 1.8763810715030227, + "grad_norm": 0.12795211374759674, + "learning_rate": 6.453951129574643e-07, + "loss": 0.6205, + "step": 2252 + }, + { + "epoch": 1.8772149259954138, + "grad_norm": 0.1331990361213684, + "learning_rate": 6.44555267557934e-07, + "loss": 0.6227, + "step": 2253 + }, + { + "epoch": 1.8780487804878048, + "grad_norm": 0.13448093831539154, + "learning_rate": 6.437157090253658e-07, + "loss": 0.5703, + "step": 2254 + }, + { + "epoch": 1.878882634980196, + "grad_norm": 0.12563006579875946, + "learning_rate": 6.428764380373375e-07, + "loss": 0.5943, + "step": 2255 + }, + { + "epoch": 1.879716489472587, + "grad_norm": 0.13857470452785492, + "learning_rate": 6.420374552711958e-07, + "loss": 0.6263, + "step": 2256 + }, + { + "epoch": 1.8805503439649782, + "grad_norm": 0.13378775119781494, + "learning_rate": 6.411987614040539e-07, + "loss": 0.6188, + "step": 2257 + }, + { + "epoch": 1.8813841984573692, + "grad_norm": 0.13999195396900177, + "learning_rate": 6.40360357112792e-07, + "loss": 0.6461, + "step": 2258 + }, + { + "epoch": 1.8822180529497603, + "grad_norm": 0.13252532482147217, + "learning_rate": 6.395222430740572e-07, + "loss": 0.6247, + "step": 2259 + }, + { + "epoch": 1.8830519074421512, + "grad_norm": 0.13208715617656708, + "learning_rate": 6.386844199642618e-07, + "loss": 0.5912, + "step": 2260 + }, + { + "epoch": 1.8838857619345424, + "grad_norm": 0.12673957645893097, + "learning_rate": 6.378468884595829e-07, + "loss": 0.529, + "step": 2261 + }, + { + "epoch": 1.8847196164269335, + "grad_norm": 0.1289004683494568, + "learning_rate": 6.370096492359635e-07, + "loss": 0.5575, + "step": 2262 + }, + { + "epoch": 1.8855534709193247, + "grad_norm": 0.1387024074792862, + "learning_rate": 6.361727029691097e-07, + "loss": 0.6183, + "step": 2263 + }, + { + "epoch": 1.8863873254117156, + "grad_norm": 0.13052384555339813, + "learning_rate": 6.353360503344909e-07, + "loss": 0.6069, + "step": 2264 + }, + { + "epoch": 1.8872211799041068, + "grad_norm": 0.13136428594589233, + "learning_rate": 6.344996920073409e-07, + "loss": 0.6139, + "step": 2265 + }, + { + "epoch": 1.8880550343964977, + "grad_norm": 0.12757977843284607, + "learning_rate": 6.336636286626547e-07, + "loss": 0.6091, + "step": 2266 + }, + { + "epoch": 1.8888888888888888, + "grad_norm": 0.13187870383262634, + "learning_rate": 6.328278609751898e-07, + "loss": 0.6339, + "step": 2267 + }, + { + "epoch": 1.88972274338128, + "grad_norm": 0.13330577313899994, + "learning_rate": 6.319923896194651e-07, + "loss": 0.5872, + "step": 2268 + }, + { + "epoch": 1.8905565978736711, + "grad_norm": 0.13062001764774323, + "learning_rate": 6.311572152697598e-07, + "loss": 0.618, + "step": 2269 + }, + { + "epoch": 1.891390452366062, + "grad_norm": 0.1248871386051178, + "learning_rate": 6.30322338600114e-07, + "loss": 0.576, + "step": 2270 + }, + { + "epoch": 1.8922243068584532, + "grad_norm": 0.1286626160144806, + "learning_rate": 6.294877602843275e-07, + "loss": 0.5739, + "step": 2271 + }, + { + "epoch": 1.8930581613508441, + "grad_norm": 0.137839674949646, + "learning_rate": 6.286534809959587e-07, + "loss": 0.5684, + "step": 2272 + }, + { + "epoch": 1.8938920158432353, + "grad_norm": 0.13219650089740753, + "learning_rate": 6.278195014083256e-07, + "loss": 0.6107, + "step": 2273 + }, + { + "epoch": 1.8947258703356264, + "grad_norm": 0.13113045692443848, + "learning_rate": 6.269858221945037e-07, + "loss": 0.5889, + "step": 2274 + }, + { + "epoch": 1.8955597248280176, + "grad_norm": 0.13249342143535614, + "learning_rate": 6.261524440273262e-07, + "loss": 0.5744, + "step": 2275 + }, + { + "epoch": 1.8963935793204088, + "grad_norm": 0.12959535419940948, + "learning_rate": 6.253193675793831e-07, + "loss": 0.597, + "step": 2276 + }, + { + "epoch": 1.8972274338127997, + "grad_norm": 0.12774960696697235, + "learning_rate": 6.244865935230214e-07, + "loss": 0.5293, + "step": 2277 + }, + { + "epoch": 1.8980612883051906, + "grad_norm": 0.13584597408771515, + "learning_rate": 6.23654122530344e-07, + "loss": 0.6379, + "step": 2278 + }, + { + "epoch": 1.8988951427975818, + "grad_norm": 0.13189005851745605, + "learning_rate": 6.228219552732083e-07, + "loss": 0.5781, + "step": 2279 + }, + { + "epoch": 1.899728997289973, + "grad_norm": 0.12410996854305267, + "learning_rate": 6.219900924232278e-07, + "loss": 0.5808, + "step": 2280 + }, + { + "epoch": 1.900562851782364, + "grad_norm": 0.12360142171382904, + "learning_rate": 6.211585346517701e-07, + "loss": 0.6029, + "step": 2281 + }, + { + "epoch": 1.9013967062747552, + "grad_norm": 0.12962640821933746, + "learning_rate": 6.203272826299559e-07, + "loss": 0.6324, + "step": 2282 + }, + { + "epoch": 1.9022305607671461, + "grad_norm": 0.1311754435300827, + "learning_rate": 6.194963370286595e-07, + "loss": 0.5782, + "step": 2283 + }, + { + "epoch": 1.903064415259537, + "grad_norm": 0.13473135232925415, + "learning_rate": 6.186656985185077e-07, + "loss": 0.6378, + "step": 2284 + }, + { + "epoch": 1.9038982697519282, + "grad_norm": 0.1313280314207077, + "learning_rate": 6.178353677698801e-07, + "loss": 0.6647, + "step": 2285 + }, + { + "epoch": 1.9047321242443194, + "grad_norm": 0.13595938682556152, + "learning_rate": 6.170053454529073e-07, + "loss": 0.5646, + "step": 2286 + }, + { + "epoch": 1.9055659787367105, + "grad_norm": 0.13668006658554077, + "learning_rate": 6.16175632237471e-07, + "loss": 0.6181, + "step": 2287 + }, + { + "epoch": 1.9063998332291017, + "grad_norm": 0.13875062763690948, + "learning_rate": 6.15346228793204e-07, + "loss": 0.5855, + "step": 2288 + }, + { + "epoch": 1.9072336877214926, + "grad_norm": 0.12911437451839447, + "learning_rate": 6.145171357894885e-07, + "loss": 0.6139, + "step": 2289 + }, + { + "epoch": 1.9080675422138835, + "grad_norm": 0.13118845224380493, + "learning_rate": 6.136883538954561e-07, + "loss": 0.5864, + "step": 2290 + }, + { + "epoch": 1.9089013967062747, + "grad_norm": 0.12767499685287476, + "learning_rate": 6.128598837799879e-07, + "loss": 0.6385, + "step": 2291 + }, + { + "epoch": 1.9097352511986658, + "grad_norm": 0.1284511238336563, + "learning_rate": 6.12031726111713e-07, + "loss": 0.5868, + "step": 2292 + }, + { + "epoch": 1.910569105691057, + "grad_norm": 0.13237819075584412, + "learning_rate": 6.112038815590079e-07, + "loss": 0.5604, + "step": 2293 + }, + { + "epoch": 1.9114029601834481, + "grad_norm": 0.13073354959487915, + "learning_rate": 6.103763507899977e-07, + "loss": 0.621, + "step": 2294 + }, + { + "epoch": 1.912236814675839, + "grad_norm": 0.1346769481897354, + "learning_rate": 6.095491344725527e-07, + "loss": 0.5636, + "step": 2295 + }, + { + "epoch": 1.91307066916823, + "grad_norm": 0.13018903136253357, + "learning_rate": 6.087222332742905e-07, + "loss": 0.582, + "step": 2296 + }, + { + "epoch": 1.9139045236606211, + "grad_norm": 0.13262416422367096, + "learning_rate": 6.078956478625742e-07, + "loss": 0.6244, + "step": 2297 + }, + { + "epoch": 1.9147383781530123, + "grad_norm": 0.1358201801776886, + "learning_rate": 6.070693789045113e-07, + "loss": 0.6072, + "step": 2298 + }, + { + "epoch": 1.9155722326454034, + "grad_norm": 0.12717005610466003, + "learning_rate": 6.06243427066955e-07, + "loss": 0.6003, + "step": 2299 + }, + { + "epoch": 1.9164060871377946, + "grad_norm": 0.13791069388389587, + "learning_rate": 6.054177930165016e-07, + "loss": 0.6324, + "step": 2300 + }, + { + "epoch": 1.9172399416301855, + "grad_norm": 0.12626376748085022, + "learning_rate": 6.045924774194916e-07, + "loss": 0.5243, + "step": 2301 + }, + { + "epoch": 1.9180737961225764, + "grad_norm": 0.12443426251411438, + "learning_rate": 6.037674809420083e-07, + "loss": 0.5733, + "step": 2302 + }, + { + "epoch": 1.9189076506149676, + "grad_norm": 0.13789716362953186, + "learning_rate": 6.029428042498772e-07, + "loss": 0.6085, + "step": 2303 + }, + { + "epoch": 1.9197415051073587, + "grad_norm": 0.12876053154468536, + "learning_rate": 6.02118448008666e-07, + "loss": 0.5843, + "step": 2304 + }, + { + "epoch": 1.92057535959975, + "grad_norm": 0.1320531666278839, + "learning_rate": 6.012944128836835e-07, + "loss": 0.6165, + "step": 2305 + }, + { + "epoch": 1.921409214092141, + "grad_norm": 0.12669098377227783, + "learning_rate": 6.004706995399797e-07, + "loss": 0.5911, + "step": 2306 + }, + { + "epoch": 1.922243068584532, + "grad_norm": 0.12740486860275269, + "learning_rate": 5.996473086423449e-07, + "loss": 0.5759, + "step": 2307 + }, + { + "epoch": 1.9230769230769231, + "grad_norm": 0.1266666054725647, + "learning_rate": 5.988242408553083e-07, + "loss": 0.6185, + "step": 2308 + }, + { + "epoch": 1.923910777569314, + "grad_norm": 0.14473924040794373, + "learning_rate": 5.980014968431395e-07, + "loss": 0.5639, + "step": 2309 + }, + { + "epoch": 1.9247446320617052, + "grad_norm": 0.13156810402870178, + "learning_rate": 5.971790772698466e-07, + "loss": 0.5898, + "step": 2310 + }, + { + "epoch": 1.9255784865540964, + "grad_norm": 0.13030414283275604, + "learning_rate": 5.963569827991751e-07, + "loss": 0.5725, + "step": 2311 + }, + { + "epoch": 1.9264123410464875, + "grad_norm": 0.1348404586315155, + "learning_rate": 5.955352140946088e-07, + "loss": 0.5914, + "step": 2312 + }, + { + "epoch": 1.9272461955388784, + "grad_norm": 0.12505322694778442, + "learning_rate": 5.94713771819368e-07, + "loss": 0.5284, + "step": 2313 + }, + { + "epoch": 1.9280800500312696, + "grad_norm": 0.1333184838294983, + "learning_rate": 5.938926566364105e-07, + "loss": 0.6533, + "step": 2314 + }, + { + "epoch": 1.9289139045236605, + "grad_norm": 0.14823047816753387, + "learning_rate": 5.930718692084289e-07, + "loss": 0.6039, + "step": 2315 + }, + { + "epoch": 1.9297477590160517, + "grad_norm": 0.1254442185163498, + "learning_rate": 5.922514101978522e-07, + "loss": 0.5922, + "step": 2316 + }, + { + "epoch": 1.9305816135084428, + "grad_norm": 0.13495124876499176, + "learning_rate": 5.914312802668444e-07, + "loss": 0.6377, + "step": 2317 + }, + { + "epoch": 1.931415468000834, + "grad_norm": 0.12776747345924377, + "learning_rate": 5.90611480077303e-07, + "loss": 0.5803, + "step": 2318 + }, + { + "epoch": 1.932249322493225, + "grad_norm": 0.12187840789556503, + "learning_rate": 5.897920102908602e-07, + "loss": 0.6001, + "step": 2319 + }, + { + "epoch": 1.933083176985616, + "grad_norm": 0.13361410796642303, + "learning_rate": 5.889728715688814e-07, + "loss": 0.6101, + "step": 2320 + }, + { + "epoch": 1.933917031478007, + "grad_norm": 0.13355158269405365, + "learning_rate": 5.881540645724645e-07, + "loss": 0.5752, + "step": 2321 + }, + { + "epoch": 1.9347508859703981, + "grad_norm": 0.12796255946159363, + "learning_rate": 5.8733558996244e-07, + "loss": 0.5554, + "step": 2322 + }, + { + "epoch": 1.9355847404627893, + "grad_norm": 0.132767915725708, + "learning_rate": 5.865174483993696e-07, + "loss": 0.5639, + "step": 2323 + }, + { + "epoch": 1.9364185949551804, + "grad_norm": 0.1333564668893814, + "learning_rate": 5.856996405435471e-07, + "loss": 0.6437, + "step": 2324 + }, + { + "epoch": 1.9372524494475714, + "grad_norm": 0.13327006995677948, + "learning_rate": 5.848821670549967e-07, + "loss": 0.6275, + "step": 2325 + }, + { + "epoch": 1.9380863039399625, + "grad_norm": 0.13549090921878815, + "learning_rate": 5.840650285934722e-07, + "loss": 0.5768, + "step": 2326 + }, + { + "epoch": 1.9389201584323534, + "grad_norm": 0.12455825507640839, + "learning_rate": 5.832482258184574e-07, + "loss": 0.5617, + "step": 2327 + }, + { + "epoch": 1.9397540129247446, + "grad_norm": 0.1259452998638153, + "learning_rate": 5.824317593891651e-07, + "loss": 0.5639, + "step": 2328 + }, + { + "epoch": 1.9405878674171357, + "grad_norm": 0.12992453575134277, + "learning_rate": 5.816156299645362e-07, + "loss": 0.6418, + "step": 2329 + }, + { + "epoch": 1.9414217219095269, + "grad_norm": 0.12744390964508057, + "learning_rate": 5.807998382032413e-07, + "loss": 0.5801, + "step": 2330 + }, + { + "epoch": 1.9422555764019178, + "grad_norm": 0.13386395573616028, + "learning_rate": 5.799843847636765e-07, + "loss": 0.6192, + "step": 2331 + }, + { + "epoch": 1.943089430894309, + "grad_norm": 0.13299784064292908, + "learning_rate": 5.791692703039656e-07, + "loss": 0.5963, + "step": 2332 + }, + { + "epoch": 1.9439232853867, + "grad_norm": 0.12975671887397766, + "learning_rate": 5.783544954819591e-07, + "loss": 0.6515, + "step": 2333 + }, + { + "epoch": 1.944757139879091, + "grad_norm": 0.13418138027191162, + "learning_rate": 5.775400609552331e-07, + "loss": 0.5566, + "step": 2334 + }, + { + "epoch": 1.9455909943714822, + "grad_norm": 0.13558447360992432, + "learning_rate": 5.767259673810889e-07, + "loss": 0.6067, + "step": 2335 + }, + { + "epoch": 1.9464248488638733, + "grad_norm": 0.12238030880689621, + "learning_rate": 5.759122154165528e-07, + "loss": 0.5724, + "step": 2336 + }, + { + "epoch": 1.9472587033562645, + "grad_norm": 0.14128273725509644, + "learning_rate": 5.750988057183754e-07, + "loss": 0.6078, + "step": 2337 + }, + { + "epoch": 1.9480925578486554, + "grad_norm": 0.14713209867477417, + "learning_rate": 5.742857389430319e-07, + "loss": 0.6125, + "step": 2338 + }, + { + "epoch": 1.9489264123410464, + "grad_norm": 0.1362263411283493, + "learning_rate": 5.734730157467194e-07, + "loss": 0.5711, + "step": 2339 + }, + { + "epoch": 1.9497602668334375, + "grad_norm": 0.13019274175167084, + "learning_rate": 5.72660636785358e-07, + "loss": 0.5702, + "step": 2340 + }, + { + "epoch": 1.9505941213258287, + "grad_norm": 0.14090077579021454, + "learning_rate": 5.718486027145905e-07, + "loss": 0.6286, + "step": 2341 + }, + { + "epoch": 1.9514279758182198, + "grad_norm": 0.13111229240894318, + "learning_rate": 5.710369141897813e-07, + "loss": 0.6162, + "step": 2342 + }, + { + "epoch": 1.952261830310611, + "grad_norm": 0.14520929753780365, + "learning_rate": 5.702255718660148e-07, + "loss": 0.6235, + "step": 2343 + }, + { + "epoch": 1.9530956848030019, + "grad_norm": 0.12398312240839005, + "learning_rate": 5.69414576398098e-07, + "loss": 0.5896, + "step": 2344 + }, + { + "epoch": 1.9539295392953928, + "grad_norm": 0.14264680445194244, + "learning_rate": 5.686039284405563e-07, + "loss": 0.5586, + "step": 2345 + }, + { + "epoch": 1.954763393787784, + "grad_norm": 0.13142624497413635, + "learning_rate": 5.677936286476356e-07, + "loss": 0.5624, + "step": 2346 + }, + { + "epoch": 1.9555972482801751, + "grad_norm": 0.13193339109420776, + "learning_rate": 5.669836776732999e-07, + "loss": 0.6235, + "step": 2347 + }, + { + "epoch": 1.9564311027725663, + "grad_norm": 0.12483777850866318, + "learning_rate": 5.661740761712324e-07, + "loss": 0.6564, + "step": 2348 + }, + { + "epoch": 1.9572649572649574, + "grad_norm": 0.13298740983009338, + "learning_rate": 5.653648247948341e-07, + "loss": 0.6364, + "step": 2349 + }, + { + "epoch": 1.9580988117573483, + "grad_norm": 0.12461144477128983, + "learning_rate": 5.64555924197223e-07, + "loss": 0.5612, + "step": 2350 + }, + { + "epoch": 1.9589326662497393, + "grad_norm": 0.12801668047904968, + "learning_rate": 5.637473750312349e-07, + "loss": 0.5265, + "step": 2351 + }, + { + "epoch": 1.9597665207421304, + "grad_norm": 0.13269507884979248, + "learning_rate": 5.62939177949421e-07, + "loss": 0.5956, + "step": 2352 + }, + { + "epoch": 1.9606003752345216, + "grad_norm": 0.14033639430999756, + "learning_rate": 5.621313336040494e-07, + "loss": 0.6257, + "step": 2353 + }, + { + "epoch": 1.9614342297269127, + "grad_norm": 0.12311799824237823, + "learning_rate": 5.613238426471024e-07, + "loss": 0.5558, + "step": 2354 + }, + { + "epoch": 1.9622680842193039, + "grad_norm": 0.12262886762619019, + "learning_rate": 5.605167057302777e-07, + "loss": 0.5878, + "step": 2355 + }, + { + "epoch": 1.9631019387116948, + "grad_norm": 0.12075518816709518, + "learning_rate": 5.597099235049874e-07, + "loss": 0.5562, + "step": 2356 + }, + { + "epoch": 1.9639357932040857, + "grad_norm": 0.12985512614250183, + "learning_rate": 5.589034966223566e-07, + "loss": 0.6132, + "step": 2357 + }, + { + "epoch": 1.9647696476964769, + "grad_norm": 0.1268652081489563, + "learning_rate": 5.580974257332242e-07, + "loss": 0.6154, + "step": 2358 + }, + { + "epoch": 1.965603502188868, + "grad_norm": 0.13775032758712769, + "learning_rate": 5.572917114881422e-07, + "loss": 0.6034, + "step": 2359 + }, + { + "epoch": 1.9664373566812592, + "grad_norm": 0.13750946521759033, + "learning_rate": 5.564863545373739e-07, + "loss": 0.5833, + "step": 2360 + }, + { + "epoch": 1.9672712111736503, + "grad_norm": 0.12842507660388947, + "learning_rate": 5.556813555308948e-07, + "loss": 0.5999, + "step": 2361 + }, + { + "epoch": 1.9681050656660413, + "grad_norm": 0.13427001237869263, + "learning_rate": 5.548767151183912e-07, + "loss": 0.6044, + "step": 2362 + }, + { + "epoch": 1.9689389201584322, + "grad_norm": 0.13650521636009216, + "learning_rate": 5.540724339492602e-07, + "loss": 0.6156, + "step": 2363 + }, + { + "epoch": 1.9697727746508233, + "grad_norm": 0.13917742669582367, + "learning_rate": 5.532685126726086e-07, + "loss": 0.5875, + "step": 2364 + }, + { + "epoch": 1.9706066291432145, + "grad_norm": 0.1281631588935852, + "learning_rate": 5.524649519372532e-07, + "loss": 0.5927, + "step": 2365 + }, + { + "epoch": 1.9714404836356056, + "grad_norm": 0.13355673849582672, + "learning_rate": 5.516617523917195e-07, + "loss": 0.5991, + "step": 2366 + }, + { + "epoch": 1.9722743381279968, + "grad_norm": 0.12803775072097778, + "learning_rate": 5.508589146842423e-07, + "loss": 0.6001, + "step": 2367 + }, + { + "epoch": 1.9731081926203877, + "grad_norm": 0.1371046006679535, + "learning_rate": 5.500564394627634e-07, + "loss": 0.5983, + "step": 2368 + }, + { + "epoch": 1.9739420471127789, + "grad_norm": 0.1317107230424881, + "learning_rate": 5.492543273749321e-07, + "loss": 0.6025, + "step": 2369 + }, + { + "epoch": 1.9747759016051698, + "grad_norm": 0.13639812171459198, + "learning_rate": 5.484525790681052e-07, + "loss": 0.598, + "step": 2370 + }, + { + "epoch": 1.975609756097561, + "grad_norm": 0.13002654910087585, + "learning_rate": 5.476511951893454e-07, + "loss": 0.602, + "step": 2371 + }, + { + "epoch": 1.976443610589952, + "grad_norm": 0.12995946407318115, + "learning_rate": 5.468501763854214e-07, + "loss": 0.575, + "step": 2372 + }, + { + "epoch": 1.9772774650823433, + "grad_norm": 0.13288018107414246, + "learning_rate": 5.460495233028073e-07, + "loss": 0.6301, + "step": 2373 + }, + { + "epoch": 1.9781113195747342, + "grad_norm": 0.13422086834907532, + "learning_rate": 5.452492365876827e-07, + "loss": 0.6567, + "step": 2374 + }, + { + "epoch": 1.9789451740671253, + "grad_norm": 0.13234974443912506, + "learning_rate": 5.444493168859303e-07, + "loss": 0.5892, + "step": 2375 + }, + { + "epoch": 1.9797790285595163, + "grad_norm": 0.13048966228961945, + "learning_rate": 5.436497648431375e-07, + "loss": 0.6009, + "step": 2376 + }, + { + "epoch": 1.9806128830519074, + "grad_norm": 0.124395951628685, + "learning_rate": 5.428505811045947e-07, + "loss": 0.5759, + "step": 2377 + }, + { + "epoch": 1.9814467375442986, + "grad_norm": 0.12869031727313995, + "learning_rate": 5.420517663152948e-07, + "loss": 0.6303, + "step": 2378 + }, + { + "epoch": 1.9822805920366897, + "grad_norm": 0.13104695081710815, + "learning_rate": 5.412533211199329e-07, + "loss": 0.6003, + "step": 2379 + }, + { + "epoch": 1.9831144465290806, + "grad_norm": 0.13080964982509613, + "learning_rate": 5.404552461629068e-07, + "loss": 0.6617, + "step": 2380 + }, + { + "epoch": 1.9839483010214718, + "grad_norm": 0.12720125913619995, + "learning_rate": 5.39657542088314e-07, + "loss": 0.6138, + "step": 2381 + }, + { + "epoch": 1.9847821555138627, + "grad_norm": 0.12719741463661194, + "learning_rate": 5.388602095399544e-07, + "loss": 0.6034, + "step": 2382 + }, + { + "epoch": 1.9856160100062539, + "grad_norm": 0.13190889358520508, + "learning_rate": 5.380632491613264e-07, + "loss": 0.6329, + "step": 2383 + }, + { + "epoch": 1.986449864498645, + "grad_norm": 0.13070254027843475, + "learning_rate": 5.372666615956289e-07, + "loss": 0.5807, + "step": 2384 + }, + { + "epoch": 1.9872837189910362, + "grad_norm": 0.12874957919120789, + "learning_rate": 5.3647044748576e-07, + "loss": 0.5942, + "step": 2385 + }, + { + "epoch": 1.988117573483427, + "grad_norm": 0.13363546133041382, + "learning_rate": 5.356746074743157e-07, + "loss": 0.5778, + "step": 2386 + }, + { + "epoch": 1.9889514279758183, + "grad_norm": 0.1323724091053009, + "learning_rate": 5.348791422035903e-07, + "loss": 0.6128, + "step": 2387 + }, + { + "epoch": 1.9897852824682092, + "grad_norm": 0.1363963931798935, + "learning_rate": 5.340840523155768e-07, + "loss": 0.6101, + "step": 2388 + }, + { + "epoch": 1.9906191369606003, + "grad_norm": 0.13670435547828674, + "learning_rate": 5.332893384519639e-07, + "loss": 0.6146, + "step": 2389 + }, + { + "epoch": 1.9914529914529915, + "grad_norm": 0.12916621565818787, + "learning_rate": 5.324950012541372e-07, + "loss": 0.6372, + "step": 2390 + }, + { + "epoch": 1.9922868459453826, + "grad_norm": 0.12862586975097656, + "learning_rate": 5.317010413631782e-07, + "loss": 0.5652, + "step": 2391 + }, + { + "epoch": 1.9931207004377736, + "grad_norm": 0.1284669190645218, + "learning_rate": 5.309074594198644e-07, + "loss": 0.6468, + "step": 2392 + }, + { + "epoch": 1.9939545549301647, + "grad_norm": 0.12397977709770203, + "learning_rate": 5.301142560646676e-07, + "loss": 0.603, + "step": 2393 + }, + { + "epoch": 1.9947884094225556, + "grad_norm": 0.13170576095581055, + "learning_rate": 5.293214319377542e-07, + "loss": 0.63, + "step": 2394 + }, + { + "epoch": 1.9956222639149468, + "grad_norm": 0.11900755017995834, + "learning_rate": 5.285289876789849e-07, + "loss": 0.5428, + "step": 2395 + }, + { + "epoch": 1.996456118407338, + "grad_norm": 0.13693659007549286, + "learning_rate": 5.277369239279143e-07, + "loss": 0.6014, + "step": 2396 + }, + { + "epoch": 1.997289972899729, + "grad_norm": 0.14242641627788544, + "learning_rate": 5.269452413237885e-07, + "loss": 0.6007, + "step": 2397 + }, + { + "epoch": 1.99812382739212, + "grad_norm": 0.12357503920793533, + "learning_rate": 5.261539405055471e-07, + "loss": 0.5221, + "step": 2398 + } + ], + "logging_steps": 1, + "max_steps": 3597, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 1199, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.0907687802820887e+19, + "train_batch_size": 6, + "trial_name": null, + "trial_params": null +}