v0.7b-adapter / trainer_state.json
gotzmann's picture
..
001bc04
raw
history blame
No virus
153 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9994810586403736,
"eval_steps": 500,
"global_step": 963,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 0.1219773218035698,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.3759,
"step": 1
},
{
"epoch": 0.0,
"grad_norm": 0.1272401511669159,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.3538,
"step": 2
},
{
"epoch": 0.0,
"grad_norm": 0.13286344707012177,
"learning_rate": 3e-06,
"loss": 1.3886,
"step": 3
},
{
"epoch": 0.0,
"grad_norm": 0.1291491985321045,
"learning_rate": 4.000000000000001e-06,
"loss": 1.3143,
"step": 4
},
{
"epoch": 0.01,
"grad_norm": 0.12544086575508118,
"learning_rate": 5e-06,
"loss": 1.317,
"step": 5
},
{
"epoch": 0.01,
"grad_norm": 0.15034854412078857,
"learning_rate": 6e-06,
"loss": 1.3537,
"step": 6
},
{
"epoch": 0.01,
"grad_norm": 0.14342203736305237,
"learning_rate": 7.000000000000001e-06,
"loss": 1.3083,
"step": 7
},
{
"epoch": 0.01,
"grad_norm": 0.1392456442117691,
"learning_rate": 8.000000000000001e-06,
"loss": 1.3501,
"step": 8
},
{
"epoch": 0.01,
"grad_norm": 0.13647185266017914,
"learning_rate": 9e-06,
"loss": 1.1859,
"step": 9
},
{
"epoch": 0.01,
"grad_norm": 0.12714740633964539,
"learning_rate": 1e-05,
"loss": 1.3746,
"step": 10
},
{
"epoch": 0.01,
"grad_norm": 0.12918396294116974,
"learning_rate": 1.1000000000000001e-05,
"loss": 1.3672,
"step": 11
},
{
"epoch": 0.01,
"grad_norm": 0.2055857628583908,
"learning_rate": 1.2e-05,
"loss": 1.3692,
"step": 12
},
{
"epoch": 0.01,
"grad_norm": 0.1577647179365158,
"learning_rate": 1.3000000000000001e-05,
"loss": 1.2316,
"step": 13
},
{
"epoch": 0.01,
"grad_norm": 0.1306232064962387,
"learning_rate": 1.4000000000000001e-05,
"loss": 1.2453,
"step": 14
},
{
"epoch": 0.02,
"grad_norm": 0.1246161013841629,
"learning_rate": 1.5e-05,
"loss": 1.3613,
"step": 15
},
{
"epoch": 0.02,
"grad_norm": 0.1408628672361374,
"learning_rate": 1.6000000000000003e-05,
"loss": 1.3905,
"step": 16
},
{
"epoch": 0.02,
"grad_norm": 0.1561778485774994,
"learning_rate": 1.7000000000000003e-05,
"loss": 1.2617,
"step": 17
},
{
"epoch": 0.02,
"grad_norm": 0.13680391013622284,
"learning_rate": 1.8e-05,
"loss": 1.3604,
"step": 18
},
{
"epoch": 0.02,
"grad_norm": 0.12442106008529663,
"learning_rate": 1.9e-05,
"loss": 1.3812,
"step": 19
},
{
"epoch": 0.02,
"grad_norm": 0.13730080425739288,
"learning_rate": 2e-05,
"loss": 1.2252,
"step": 20
},
{
"epoch": 0.02,
"grad_norm": 0.11723917722702026,
"learning_rate": 2.1e-05,
"loss": 1.1512,
"step": 21
},
{
"epoch": 0.02,
"grad_norm": 0.13034303486347198,
"learning_rate": 2.2000000000000003e-05,
"loss": 1.2698,
"step": 22
},
{
"epoch": 0.02,
"grad_norm": 0.15671290457248688,
"learning_rate": 2.3000000000000003e-05,
"loss": 1.2823,
"step": 23
},
{
"epoch": 0.02,
"grad_norm": 0.11263088881969452,
"learning_rate": 2.4e-05,
"loss": 1.375,
"step": 24
},
{
"epoch": 0.03,
"grad_norm": 0.13272984325885773,
"learning_rate": 2.5e-05,
"loss": 1.2721,
"step": 25
},
{
"epoch": 0.03,
"grad_norm": 0.13560321927070618,
"learning_rate": 2.6000000000000002e-05,
"loss": 1.3223,
"step": 26
},
{
"epoch": 0.03,
"grad_norm": 0.13235026597976685,
"learning_rate": 2.7000000000000002e-05,
"loss": 1.2181,
"step": 27
},
{
"epoch": 0.03,
"grad_norm": 0.12842555344104767,
"learning_rate": 2.8000000000000003e-05,
"loss": 1.3783,
"step": 28
},
{
"epoch": 0.03,
"grad_norm": 0.13284699618816376,
"learning_rate": 2.9e-05,
"loss": 1.3668,
"step": 29
},
{
"epoch": 0.03,
"grad_norm": 0.12673886120319366,
"learning_rate": 3e-05,
"loss": 1.3657,
"step": 30
},
{
"epoch": 0.03,
"grad_norm": 0.1453860104084015,
"learning_rate": 3.1e-05,
"loss": 1.2646,
"step": 31
},
{
"epoch": 0.03,
"grad_norm": 0.17890802025794983,
"learning_rate": 3.2000000000000005e-05,
"loss": 1.1986,
"step": 32
},
{
"epoch": 0.03,
"grad_norm": 0.1185929998755455,
"learning_rate": 3.3e-05,
"loss": 1.1962,
"step": 33
},
{
"epoch": 0.04,
"grad_norm": 0.12994253635406494,
"learning_rate": 3.4000000000000007e-05,
"loss": 1.2893,
"step": 34
},
{
"epoch": 0.04,
"grad_norm": 0.1479964405298233,
"learning_rate": 3.5e-05,
"loss": 1.1221,
"step": 35
},
{
"epoch": 0.04,
"grad_norm": 0.13566361367702484,
"learning_rate": 3.6e-05,
"loss": 1.2425,
"step": 36
},
{
"epoch": 0.04,
"grad_norm": 0.1559300720691681,
"learning_rate": 3.7e-05,
"loss": 1.2995,
"step": 37
},
{
"epoch": 0.04,
"grad_norm": 0.14187201857566833,
"learning_rate": 3.8e-05,
"loss": 1.3666,
"step": 38
},
{
"epoch": 0.04,
"grad_norm": 0.13634175062179565,
"learning_rate": 3.9000000000000006e-05,
"loss": 1.3111,
"step": 39
},
{
"epoch": 0.04,
"grad_norm": 0.1550116240978241,
"learning_rate": 4e-05,
"loss": 1.3729,
"step": 40
},
{
"epoch": 0.04,
"grad_norm": 0.14702185988426208,
"learning_rate": 4.1e-05,
"loss": 1.2447,
"step": 41
},
{
"epoch": 0.04,
"grad_norm": 0.1325961798429489,
"learning_rate": 4.2e-05,
"loss": 1.174,
"step": 42
},
{
"epoch": 0.04,
"grad_norm": 0.1384323686361313,
"learning_rate": 4.3e-05,
"loss": 1.2333,
"step": 43
},
{
"epoch": 0.05,
"grad_norm": 0.13662421703338623,
"learning_rate": 4.4000000000000006e-05,
"loss": 1.1815,
"step": 44
},
{
"epoch": 0.05,
"grad_norm": 0.13229218125343323,
"learning_rate": 4.5e-05,
"loss": 1.2125,
"step": 45
},
{
"epoch": 0.05,
"grad_norm": 0.14347675442695618,
"learning_rate": 4.600000000000001e-05,
"loss": 1.1498,
"step": 46
},
{
"epoch": 0.05,
"grad_norm": 0.18536897003650665,
"learning_rate": 4.7e-05,
"loss": 1.2404,
"step": 47
},
{
"epoch": 0.05,
"grad_norm": 0.13089048862457275,
"learning_rate": 4.8e-05,
"loss": 1.2904,
"step": 48
},
{
"epoch": 0.05,
"grad_norm": 0.12752604484558105,
"learning_rate": 4.9e-05,
"loss": 1.2702,
"step": 49
},
{
"epoch": 0.05,
"grad_norm": 0.15005581080913544,
"learning_rate": 5e-05,
"loss": 1.2866,
"step": 50
},
{
"epoch": 0.05,
"grad_norm": 0.16374410688877106,
"learning_rate": 5.1000000000000006e-05,
"loss": 1.333,
"step": 51
},
{
"epoch": 0.05,
"grad_norm": 0.1684991866350174,
"learning_rate": 5.2000000000000004e-05,
"loss": 1.4278,
"step": 52
},
{
"epoch": 0.06,
"grad_norm": 0.1903846263885498,
"learning_rate": 5.300000000000001e-05,
"loss": 1.3035,
"step": 53
},
{
"epoch": 0.06,
"grad_norm": 0.16056384146213531,
"learning_rate": 5.4000000000000005e-05,
"loss": 1.2323,
"step": 54
},
{
"epoch": 0.06,
"grad_norm": 0.14987166225910187,
"learning_rate": 5.500000000000001e-05,
"loss": 1.0263,
"step": 55
},
{
"epoch": 0.06,
"grad_norm": 0.13468973338603973,
"learning_rate": 5.6000000000000006e-05,
"loss": 1.2534,
"step": 56
},
{
"epoch": 0.06,
"grad_norm": 0.1467704474925995,
"learning_rate": 5.6999999999999996e-05,
"loss": 1.3025,
"step": 57
},
{
"epoch": 0.06,
"grad_norm": 0.17560411989688873,
"learning_rate": 5.8e-05,
"loss": 1.2307,
"step": 58
},
{
"epoch": 0.06,
"grad_norm": 0.1423458606004715,
"learning_rate": 5.9e-05,
"loss": 1.1628,
"step": 59
},
{
"epoch": 0.06,
"grad_norm": 0.12863807380199432,
"learning_rate": 6e-05,
"loss": 1.1174,
"step": 60
},
{
"epoch": 0.06,
"grad_norm": 0.13459883630275726,
"learning_rate": 6.1e-05,
"loss": 1.2223,
"step": 61
},
{
"epoch": 0.06,
"grad_norm": 0.13762715458869934,
"learning_rate": 6.2e-05,
"loss": 1.2393,
"step": 62
},
{
"epoch": 0.07,
"grad_norm": 0.1310354322195053,
"learning_rate": 6.3e-05,
"loss": 1.2487,
"step": 63
},
{
"epoch": 0.07,
"grad_norm": 0.12477163225412369,
"learning_rate": 6.400000000000001e-05,
"loss": 1.2632,
"step": 64
},
{
"epoch": 0.07,
"grad_norm": 0.15380260348320007,
"learning_rate": 6.500000000000001e-05,
"loss": 1.3762,
"step": 65
},
{
"epoch": 0.07,
"grad_norm": 0.1248747706413269,
"learning_rate": 6.6e-05,
"loss": 1.1336,
"step": 66
},
{
"epoch": 0.07,
"grad_norm": 0.12641173601150513,
"learning_rate": 6.7e-05,
"loss": 1.2674,
"step": 67
},
{
"epoch": 0.07,
"grad_norm": 0.13872478902339935,
"learning_rate": 6.800000000000001e-05,
"loss": 1.3014,
"step": 68
},
{
"epoch": 0.07,
"grad_norm": 0.1377810686826706,
"learning_rate": 6.9e-05,
"loss": 1.2816,
"step": 69
},
{
"epoch": 0.07,
"grad_norm": 0.13459351658821106,
"learning_rate": 7e-05,
"loss": 1.2872,
"step": 70
},
{
"epoch": 0.07,
"grad_norm": 0.13774238526821136,
"learning_rate": 7.1e-05,
"loss": 1.1481,
"step": 71
},
{
"epoch": 0.07,
"grad_norm": 0.1276445835828781,
"learning_rate": 7.2e-05,
"loss": 1.1924,
"step": 72
},
{
"epoch": 0.08,
"grad_norm": 0.12415210157632828,
"learning_rate": 7.3e-05,
"loss": 1.3093,
"step": 73
},
{
"epoch": 0.08,
"grad_norm": 0.11668894439935684,
"learning_rate": 7.4e-05,
"loss": 1.1941,
"step": 74
},
{
"epoch": 0.08,
"grad_norm": 0.16712161898612976,
"learning_rate": 7.500000000000001e-05,
"loss": 1.2581,
"step": 75
},
{
"epoch": 0.08,
"grad_norm": 0.1783616989850998,
"learning_rate": 7.6e-05,
"loss": 1.0885,
"step": 76
},
{
"epoch": 0.08,
"grad_norm": 0.1388709843158722,
"learning_rate": 7.7e-05,
"loss": 1.1999,
"step": 77
},
{
"epoch": 0.08,
"grad_norm": 0.13248670101165771,
"learning_rate": 7.800000000000001e-05,
"loss": 1.1312,
"step": 78
},
{
"epoch": 0.08,
"grad_norm": 0.8402003645896912,
"learning_rate": 7.900000000000001e-05,
"loss": 1.281,
"step": 79
},
{
"epoch": 0.08,
"grad_norm": 0.1560419201850891,
"learning_rate": 8e-05,
"loss": 1.3314,
"step": 80
},
{
"epoch": 0.08,
"grad_norm": 0.13449999690055847,
"learning_rate": 8.1e-05,
"loss": 1.1742,
"step": 81
},
{
"epoch": 0.09,
"grad_norm": 0.14383117854595184,
"learning_rate": 8.2e-05,
"loss": 1.1958,
"step": 82
},
{
"epoch": 0.09,
"grad_norm": 0.15058472752571106,
"learning_rate": 8.3e-05,
"loss": 1.2992,
"step": 83
},
{
"epoch": 0.09,
"grad_norm": 0.14776362478733063,
"learning_rate": 8.4e-05,
"loss": 1.3611,
"step": 84
},
{
"epoch": 0.09,
"grad_norm": 0.14592403173446655,
"learning_rate": 8.5e-05,
"loss": 1.1165,
"step": 85
},
{
"epoch": 0.09,
"grad_norm": 0.1357845813035965,
"learning_rate": 8.6e-05,
"loss": 1.2509,
"step": 86
},
{
"epoch": 0.09,
"grad_norm": 0.15681371092796326,
"learning_rate": 8.7e-05,
"loss": 1.0857,
"step": 87
},
{
"epoch": 0.09,
"grad_norm": 0.14496199786663055,
"learning_rate": 8.800000000000001e-05,
"loss": 1.1111,
"step": 88
},
{
"epoch": 0.09,
"grad_norm": 0.1501457244157791,
"learning_rate": 8.900000000000001e-05,
"loss": 1.2556,
"step": 89
},
{
"epoch": 0.09,
"grad_norm": 0.14578907191753387,
"learning_rate": 9e-05,
"loss": 1.1509,
"step": 90
},
{
"epoch": 0.09,
"grad_norm": 0.13890138268470764,
"learning_rate": 9.1e-05,
"loss": 1.1687,
"step": 91
},
{
"epoch": 0.1,
"grad_norm": 0.14948885142803192,
"learning_rate": 9.200000000000001e-05,
"loss": 1.0968,
"step": 92
},
{
"epoch": 0.1,
"grad_norm": 0.13117392361164093,
"learning_rate": 9.300000000000001e-05,
"loss": 1.3073,
"step": 93
},
{
"epoch": 0.1,
"grad_norm": 0.13922806084156036,
"learning_rate": 9.4e-05,
"loss": 1.2502,
"step": 94
},
{
"epoch": 0.1,
"grad_norm": 0.16392332315444946,
"learning_rate": 9.5e-05,
"loss": 1.3148,
"step": 95
},
{
"epoch": 0.1,
"grad_norm": 0.11756494641304016,
"learning_rate": 9.6e-05,
"loss": 1.3215,
"step": 96
},
{
"epoch": 0.1,
"grad_norm": 0.11909925937652588,
"learning_rate": 9.7e-05,
"loss": 1.2593,
"step": 97
},
{
"epoch": 0.1,
"grad_norm": 0.1371874213218689,
"learning_rate": 9.8e-05,
"loss": 1.1537,
"step": 98
},
{
"epoch": 0.1,
"grad_norm": 0.14141380786895752,
"learning_rate": 9.900000000000001e-05,
"loss": 1.221,
"step": 99
},
{
"epoch": 0.1,
"grad_norm": 0.13502179086208344,
"learning_rate": 0.0001,
"loss": 1.3492,
"step": 100
},
{
"epoch": 0.1,
"grad_norm": 0.15468691289424896,
"learning_rate": 0.000101,
"loss": 1.2592,
"step": 101
},
{
"epoch": 0.11,
"grad_norm": 0.13041648268699646,
"learning_rate": 0.00010200000000000001,
"loss": 1.3792,
"step": 102
},
{
"epoch": 0.11,
"grad_norm": 0.11658702045679092,
"learning_rate": 0.00010300000000000001,
"loss": 1.1845,
"step": 103
},
{
"epoch": 0.11,
"grad_norm": 0.13084186613559723,
"learning_rate": 0.00010400000000000001,
"loss": 1.1808,
"step": 104
},
{
"epoch": 0.11,
"grad_norm": 0.13180844485759735,
"learning_rate": 0.000105,
"loss": 1.1721,
"step": 105
},
{
"epoch": 0.11,
"grad_norm": 0.1375647485256195,
"learning_rate": 0.00010600000000000002,
"loss": 1.3622,
"step": 106
},
{
"epoch": 0.11,
"grad_norm": 0.11864122748374939,
"learning_rate": 0.00010700000000000001,
"loss": 1.238,
"step": 107
},
{
"epoch": 0.11,
"grad_norm": 0.11780542880296707,
"learning_rate": 0.00010800000000000001,
"loss": 1.1581,
"step": 108
},
{
"epoch": 0.11,
"grad_norm": 0.11449900269508362,
"learning_rate": 0.000109,
"loss": 1.3076,
"step": 109
},
{
"epoch": 0.11,
"grad_norm": 0.11585865169763565,
"learning_rate": 0.00011000000000000002,
"loss": 1.3553,
"step": 110
},
{
"epoch": 0.12,
"grad_norm": 0.12288644164800644,
"learning_rate": 0.00011100000000000001,
"loss": 1.1448,
"step": 111
},
{
"epoch": 0.12,
"grad_norm": 0.12131894379854202,
"learning_rate": 0.00011200000000000001,
"loss": 1.3456,
"step": 112
},
{
"epoch": 0.12,
"grad_norm": 0.13827624917030334,
"learning_rate": 0.000113,
"loss": 1.3183,
"step": 113
},
{
"epoch": 0.12,
"grad_norm": 0.13708753883838654,
"learning_rate": 0.00011399999999999999,
"loss": 1.3293,
"step": 114
},
{
"epoch": 0.12,
"grad_norm": 0.11893680691719055,
"learning_rate": 0.00011499999999999999,
"loss": 1.1954,
"step": 115
},
{
"epoch": 0.12,
"grad_norm": 0.12022116035223007,
"learning_rate": 0.000116,
"loss": 1.1822,
"step": 116
},
{
"epoch": 0.12,
"grad_norm": 0.11215274035930634,
"learning_rate": 0.000117,
"loss": 1.2453,
"step": 117
},
{
"epoch": 0.12,
"grad_norm": 0.11778578907251358,
"learning_rate": 0.000118,
"loss": 1.1939,
"step": 118
},
{
"epoch": 0.12,
"grad_norm": 0.12714551389217377,
"learning_rate": 0.000119,
"loss": 1.1917,
"step": 119
},
{
"epoch": 0.12,
"grad_norm": 0.11972431093454361,
"learning_rate": 0.00012,
"loss": 1.2769,
"step": 120
},
{
"epoch": 0.13,
"grad_norm": 0.12945646047592163,
"learning_rate": 0.000121,
"loss": 1.2929,
"step": 121
},
{
"epoch": 0.13,
"grad_norm": 0.11721435189247131,
"learning_rate": 0.000122,
"loss": 1.3606,
"step": 122
},
{
"epoch": 0.13,
"grad_norm": 0.117381751537323,
"learning_rate": 0.000123,
"loss": 1.3187,
"step": 123
},
{
"epoch": 0.13,
"grad_norm": 0.13110613822937012,
"learning_rate": 0.000124,
"loss": 1.3786,
"step": 124
},
{
"epoch": 0.13,
"grad_norm": 0.12571324408054352,
"learning_rate": 0.000125,
"loss": 1.0165,
"step": 125
},
{
"epoch": 0.13,
"grad_norm": 0.12401648610830307,
"learning_rate": 0.000126,
"loss": 1.1858,
"step": 126
},
{
"epoch": 0.13,
"grad_norm": 0.11960897594690323,
"learning_rate": 0.000127,
"loss": 1.2638,
"step": 127
},
{
"epoch": 0.13,
"grad_norm": 0.11195061355829239,
"learning_rate": 0.00012800000000000002,
"loss": 1.2469,
"step": 128
},
{
"epoch": 0.13,
"grad_norm": 0.1292635053396225,
"learning_rate": 0.00012900000000000002,
"loss": 1.3642,
"step": 129
},
{
"epoch": 0.13,
"grad_norm": 0.11140226572751999,
"learning_rate": 0.00013000000000000002,
"loss": 1.2238,
"step": 130
},
{
"epoch": 0.14,
"grad_norm": 0.10454066097736359,
"learning_rate": 0.000131,
"loss": 1.2418,
"step": 131
},
{
"epoch": 0.14,
"grad_norm": 0.11138935387134552,
"learning_rate": 0.000132,
"loss": 1.1964,
"step": 132
},
{
"epoch": 0.14,
"grad_norm": 0.12519143521785736,
"learning_rate": 0.000133,
"loss": 1.1889,
"step": 133
},
{
"epoch": 0.14,
"grad_norm": 0.1129814088344574,
"learning_rate": 0.000134,
"loss": 1.3425,
"step": 134
},
{
"epoch": 0.14,
"grad_norm": 0.11608937382698059,
"learning_rate": 0.00013500000000000003,
"loss": 1.0762,
"step": 135
},
{
"epoch": 0.14,
"grad_norm": 0.1410205215215683,
"learning_rate": 0.00013600000000000003,
"loss": 1.2502,
"step": 136
},
{
"epoch": 0.14,
"grad_norm": 0.11246364563703537,
"learning_rate": 0.00013700000000000002,
"loss": 1.3605,
"step": 137
},
{
"epoch": 0.14,
"grad_norm": 0.1213245838880539,
"learning_rate": 0.000138,
"loss": 1.2632,
"step": 138
},
{
"epoch": 0.14,
"grad_norm": 0.11144573241472244,
"learning_rate": 0.000139,
"loss": 1.1771,
"step": 139
},
{
"epoch": 0.15,
"grad_norm": 0.1700059324502945,
"learning_rate": 0.00014,
"loss": 1.3396,
"step": 140
},
{
"epoch": 0.15,
"grad_norm": 0.11950355768203735,
"learning_rate": 0.000141,
"loss": 1.1329,
"step": 141
},
{
"epoch": 0.15,
"grad_norm": 0.11367864161729813,
"learning_rate": 0.000142,
"loss": 1.3323,
"step": 142
},
{
"epoch": 0.15,
"grad_norm": 0.22367842495441437,
"learning_rate": 0.000143,
"loss": 1.201,
"step": 143
},
{
"epoch": 0.15,
"grad_norm": 0.11639752984046936,
"learning_rate": 0.000144,
"loss": 1.2507,
"step": 144
},
{
"epoch": 0.15,
"grad_norm": 0.12356352061033249,
"learning_rate": 0.000145,
"loss": 1.3437,
"step": 145
},
{
"epoch": 0.15,
"grad_norm": 0.10859037935733795,
"learning_rate": 0.000146,
"loss": 1.1677,
"step": 146
},
{
"epoch": 0.15,
"grad_norm": 0.11901260167360306,
"learning_rate": 0.000147,
"loss": 1.1125,
"step": 147
},
{
"epoch": 0.15,
"grad_norm": 0.1092861220240593,
"learning_rate": 0.000148,
"loss": 1.1939,
"step": 148
},
{
"epoch": 0.15,
"grad_norm": 0.10728005319833755,
"learning_rate": 0.00014900000000000002,
"loss": 1.1974,
"step": 149
},
{
"epoch": 0.16,
"grad_norm": 0.11113261431455612,
"learning_rate": 0.00015000000000000001,
"loss": 1.1107,
"step": 150
},
{
"epoch": 0.16,
"grad_norm": 0.14674580097198486,
"learning_rate": 0.000151,
"loss": 1.2339,
"step": 151
},
{
"epoch": 0.16,
"grad_norm": 0.10521555691957474,
"learning_rate": 0.000152,
"loss": 1.2917,
"step": 152
},
{
"epoch": 0.16,
"grad_norm": 0.12526080012321472,
"learning_rate": 0.000153,
"loss": 1.3247,
"step": 153
},
{
"epoch": 0.16,
"grad_norm": 0.2366744726896286,
"learning_rate": 0.000154,
"loss": 1.3637,
"step": 154
},
{
"epoch": 0.16,
"grad_norm": 0.11178407818078995,
"learning_rate": 0.000155,
"loss": 1.2649,
"step": 155
},
{
"epoch": 0.16,
"grad_norm": 0.10156022757291794,
"learning_rate": 0.00015600000000000002,
"loss": 1.1637,
"step": 156
},
{
"epoch": 0.16,
"grad_norm": 0.09755547344684601,
"learning_rate": 0.00015700000000000002,
"loss": 1.241,
"step": 157
},
{
"epoch": 0.16,
"grad_norm": 0.11344686150550842,
"learning_rate": 0.00015800000000000002,
"loss": 1.2208,
"step": 158
},
{
"epoch": 0.17,
"grad_norm": 0.1021842285990715,
"learning_rate": 0.00015900000000000002,
"loss": 1.1936,
"step": 159
},
{
"epoch": 0.17,
"grad_norm": 0.11096473783254623,
"learning_rate": 0.00016,
"loss": 1.2908,
"step": 160
},
{
"epoch": 0.17,
"grad_norm": 0.11365038901567459,
"learning_rate": 0.000161,
"loss": 1.1754,
"step": 161
},
{
"epoch": 0.17,
"grad_norm": 0.1159452423453331,
"learning_rate": 0.000162,
"loss": 1.2411,
"step": 162
},
{
"epoch": 0.17,
"grad_norm": 0.1240570917725563,
"learning_rate": 0.000163,
"loss": 1.2655,
"step": 163
},
{
"epoch": 0.17,
"grad_norm": 0.11343025416135788,
"learning_rate": 0.000164,
"loss": 1.3486,
"step": 164
},
{
"epoch": 0.17,
"grad_norm": 0.10576006770133972,
"learning_rate": 0.000165,
"loss": 1.2787,
"step": 165
},
{
"epoch": 0.17,
"grad_norm": 0.182003453373909,
"learning_rate": 0.000166,
"loss": 1.2045,
"step": 166
},
{
"epoch": 0.17,
"grad_norm": 0.10536637157201767,
"learning_rate": 0.000167,
"loss": 1.1388,
"step": 167
},
{
"epoch": 0.17,
"grad_norm": 0.10617152601480484,
"learning_rate": 0.000168,
"loss": 1.2686,
"step": 168
},
{
"epoch": 0.18,
"grad_norm": 0.11899720877408981,
"learning_rate": 0.00016900000000000002,
"loss": 1.1732,
"step": 169
},
{
"epoch": 0.18,
"grad_norm": 0.11736088991165161,
"learning_rate": 0.00017,
"loss": 1.2772,
"step": 170
},
{
"epoch": 0.18,
"grad_norm": 0.10500625520944595,
"learning_rate": 0.000171,
"loss": 1.2193,
"step": 171
},
{
"epoch": 0.18,
"grad_norm": 0.10299772769212723,
"learning_rate": 0.000172,
"loss": 1.2021,
"step": 172
},
{
"epoch": 0.18,
"grad_norm": 0.10550544410943985,
"learning_rate": 0.000173,
"loss": 1.2393,
"step": 173
},
{
"epoch": 0.18,
"grad_norm": 0.1132335364818573,
"learning_rate": 0.000174,
"loss": 1.2713,
"step": 174
},
{
"epoch": 0.18,
"grad_norm": 0.11581739038228989,
"learning_rate": 0.000175,
"loss": 1.3282,
"step": 175
},
{
"epoch": 0.18,
"grad_norm": 0.11211756616830826,
"learning_rate": 0.00017600000000000002,
"loss": 1.2374,
"step": 176
},
{
"epoch": 0.18,
"grad_norm": 0.10848426073789597,
"learning_rate": 0.00017700000000000002,
"loss": 1.2916,
"step": 177
},
{
"epoch": 0.18,
"grad_norm": 0.11354674398899078,
"learning_rate": 0.00017800000000000002,
"loss": 1.1914,
"step": 178
},
{
"epoch": 0.19,
"grad_norm": 0.12108771502971649,
"learning_rate": 0.00017900000000000001,
"loss": 1.3693,
"step": 179
},
{
"epoch": 0.19,
"grad_norm": 0.11068333685398102,
"learning_rate": 0.00018,
"loss": 1.086,
"step": 180
},
{
"epoch": 0.19,
"grad_norm": 0.11752796918153763,
"learning_rate": 0.000181,
"loss": 1.1508,
"step": 181
},
{
"epoch": 0.19,
"grad_norm": 0.10650038719177246,
"learning_rate": 0.000182,
"loss": 1.204,
"step": 182
},
{
"epoch": 0.19,
"grad_norm": 0.12248033285140991,
"learning_rate": 0.000183,
"loss": 1.2288,
"step": 183
},
{
"epoch": 0.19,
"grad_norm": 0.10320232063531876,
"learning_rate": 0.00018400000000000003,
"loss": 1.1309,
"step": 184
},
{
"epoch": 0.19,
"grad_norm": 0.1450732797384262,
"learning_rate": 0.00018500000000000002,
"loss": 1.2696,
"step": 185
},
{
"epoch": 0.19,
"grad_norm": 0.10568010061979294,
"learning_rate": 0.00018600000000000002,
"loss": 1.2643,
"step": 186
},
{
"epoch": 0.19,
"grad_norm": 0.12027282267808914,
"learning_rate": 0.00018700000000000002,
"loss": 1.2455,
"step": 187
},
{
"epoch": 0.2,
"grad_norm": 0.09627476334571838,
"learning_rate": 0.000188,
"loss": 1.1999,
"step": 188
},
{
"epoch": 0.2,
"grad_norm": 0.1297677904367447,
"learning_rate": 0.00018899999999999999,
"loss": 1.0843,
"step": 189
},
{
"epoch": 0.2,
"grad_norm": 0.1251845359802246,
"learning_rate": 0.00019,
"loss": 1.2957,
"step": 190
},
{
"epoch": 0.2,
"grad_norm": 0.11355076730251312,
"learning_rate": 0.000191,
"loss": 1.2705,
"step": 191
},
{
"epoch": 0.2,
"grad_norm": 0.11948814988136292,
"learning_rate": 0.000192,
"loss": 1.2352,
"step": 192
},
{
"epoch": 0.2,
"grad_norm": 0.12006771564483643,
"learning_rate": 0.000193,
"loss": 1.2489,
"step": 193
},
{
"epoch": 0.2,
"grad_norm": 0.11943700909614563,
"learning_rate": 0.000194,
"loss": 1.0768,
"step": 194
},
{
"epoch": 0.2,
"grad_norm": 0.12115556001663208,
"learning_rate": 0.000195,
"loss": 1.2614,
"step": 195
},
{
"epoch": 0.2,
"grad_norm": 0.1361999660730362,
"learning_rate": 0.000196,
"loss": 1.1418,
"step": 196
},
{
"epoch": 0.2,
"grad_norm": 0.10292702168226242,
"learning_rate": 0.00019700000000000002,
"loss": 1.1649,
"step": 197
},
{
"epoch": 0.21,
"grad_norm": 0.12569741904735565,
"learning_rate": 0.00019800000000000002,
"loss": 1.2054,
"step": 198
},
{
"epoch": 0.21,
"grad_norm": 0.09684847295284271,
"learning_rate": 0.000199,
"loss": 1.0854,
"step": 199
},
{
"epoch": 0.21,
"grad_norm": 0.10697442293167114,
"learning_rate": 0.0002,
"loss": 1.1895,
"step": 200
},
{
"epoch": 0.21,
"grad_norm": 0.10017522424459457,
"learning_rate": 0.00019999915234318063,
"loss": 1.2775,
"step": 201
},
{
"epoch": 0.21,
"grad_norm": 0.0991068184375763,
"learning_rate": 0.0001999966093870929,
"loss": 1.3151,
"step": 202
},
{
"epoch": 0.21,
"grad_norm": 0.10074901580810547,
"learning_rate": 0.00019999237117484796,
"loss": 1.2666,
"step": 203
},
{
"epoch": 0.21,
"grad_norm": 0.10549337416887283,
"learning_rate": 0.00019998643777829672,
"loss": 1.2483,
"step": 204
},
{
"epoch": 0.21,
"grad_norm": 0.10364524275064468,
"learning_rate": 0.00019997880929802894,
"loss": 1.0562,
"step": 205
},
{
"epoch": 0.21,
"grad_norm": 0.21587559580802917,
"learning_rate": 0.00019996948586337125,
"loss": 1.3881,
"step": 206
},
{
"epoch": 0.21,
"grad_norm": 0.10044362396001816,
"learning_rate": 0.00019995846763238512,
"loss": 1.2114,
"step": 207
},
{
"epoch": 0.22,
"grad_norm": 0.1037297397851944,
"learning_rate": 0.0001999457547918641,
"loss": 1.2611,
"step": 208
},
{
"epoch": 0.22,
"grad_norm": 0.11380340903997421,
"learning_rate": 0.00019993134755733074,
"loss": 1.3762,
"step": 209
},
{
"epoch": 0.22,
"grad_norm": 0.11318037658929825,
"learning_rate": 0.0001999152461730328,
"loss": 1.2906,
"step": 210
},
{
"epoch": 0.22,
"grad_norm": 0.1097273975610733,
"learning_rate": 0.0001998974509119393,
"loss": 1.0855,
"step": 211
},
{
"epoch": 0.22,
"grad_norm": 0.11358966678380966,
"learning_rate": 0.00019987796207573573,
"loss": 1.2061,
"step": 212
},
{
"epoch": 0.22,
"grad_norm": 0.11139972507953644,
"learning_rate": 0.00019985677999481894,
"loss": 1.2453,
"step": 213
},
{
"epoch": 0.22,
"grad_norm": 0.1275399923324585,
"learning_rate": 0.00019983390502829166,
"loss": 1.1917,
"step": 214
},
{
"epoch": 0.22,
"grad_norm": 0.1333334743976593,
"learning_rate": 0.00019980933756395634,
"loss": 1.2102,
"step": 215
},
{
"epoch": 0.22,
"grad_norm": 0.10892786085605621,
"learning_rate": 0.00019978307801830854,
"loss": 1.1177,
"step": 216
},
{
"epoch": 0.23,
"grad_norm": 0.10664794594049454,
"learning_rate": 0.00019975512683652986,
"loss": 1.1484,
"step": 217
},
{
"epoch": 0.23,
"grad_norm": 0.20084381103515625,
"learning_rate": 0.0001997254844924806,
"loss": 1.0399,
"step": 218
},
{
"epoch": 0.23,
"grad_norm": 0.11178454756736755,
"learning_rate": 0.0001996941514886914,
"loss": 1.3401,
"step": 219
},
{
"epoch": 0.23,
"grad_norm": 0.10734831541776657,
"learning_rate": 0.00019966112835635493,
"loss": 1.1964,
"step": 220
},
{
"epoch": 0.23,
"grad_norm": 0.12020622938871384,
"learning_rate": 0.00019962641565531692,
"loss": 1.1535,
"step": 221
},
{
"epoch": 0.23,
"grad_norm": 0.10945051908493042,
"learning_rate": 0.00019959001397406646,
"loss": 1.2086,
"step": 222
},
{
"epoch": 0.23,
"grad_norm": 0.11263303458690643,
"learning_rate": 0.00019955192392972628,
"loss": 1.1488,
"step": 223
},
{
"epoch": 0.23,
"grad_norm": 0.119475819170475,
"learning_rate": 0.00019951214616804202,
"loss": 1.3071,
"step": 224
},
{
"epoch": 0.23,
"grad_norm": 0.10934551060199738,
"learning_rate": 0.00019947068136337158,
"loss": 1.2639,
"step": 225
},
{
"epoch": 0.23,
"grad_norm": 0.11675988882780075,
"learning_rate": 0.0001994275302186734,
"loss": 1.142,
"step": 226
},
{
"epoch": 0.24,
"grad_norm": 0.28856417536735535,
"learning_rate": 0.00019938269346549475,
"loss": 1.2437,
"step": 227
},
{
"epoch": 0.24,
"grad_norm": 0.11922767013311386,
"learning_rate": 0.00019933617186395917,
"loss": 1.1668,
"step": 228
},
{
"epoch": 0.24,
"grad_norm": 0.12038363516330719,
"learning_rate": 0.00019928796620275377,
"loss": 1.0465,
"step": 229
},
{
"epoch": 0.24,
"grad_norm": 0.10866128653287888,
"learning_rate": 0.00019923807729911567,
"loss": 1.1417,
"step": 230
},
{
"epoch": 0.24,
"grad_norm": 0.134980708360672,
"learning_rate": 0.00019918650599881827,
"loss": 1.3574,
"step": 231
},
{
"epoch": 0.24,
"grad_norm": 0.11622115224599838,
"learning_rate": 0.00019913325317615684,
"loss": 1.2457,
"step": 232
},
{
"epoch": 0.24,
"grad_norm": 0.11134631186723709,
"learning_rate": 0.0001990783197339338,
"loss": 1.2194,
"step": 233
},
{
"epoch": 0.24,
"grad_norm": 0.11312945932149887,
"learning_rate": 0.0001990217066034432,
"loss": 1.179,
"step": 234
},
{
"epoch": 0.24,
"grad_norm": 0.10070719569921494,
"learning_rate": 0.00019896341474445525,
"loss": 1.299,
"step": 235
},
{
"epoch": 0.24,
"grad_norm": 0.11468905210494995,
"learning_rate": 0.00019890344514519973,
"loss": 1.2141,
"step": 236
},
{
"epoch": 0.25,
"grad_norm": 0.12275680154561996,
"learning_rate": 0.00019884179882234944,
"loss": 1.2581,
"step": 237
},
{
"epoch": 0.25,
"grad_norm": 0.1063394546508789,
"learning_rate": 0.00019877847682100294,
"loss": 1.3027,
"step": 238
},
{
"epoch": 0.25,
"grad_norm": 0.10849407315254211,
"learning_rate": 0.00019871348021466672,
"loss": 1.1914,
"step": 239
},
{
"epoch": 0.25,
"grad_norm": 0.1102900579571724,
"learning_rate": 0.0001986468101052371,
"loss": 1.2589,
"step": 240
},
{
"epoch": 0.25,
"grad_norm": 0.11001666635274887,
"learning_rate": 0.00019857846762298155,
"loss": 1.141,
"step": 241
},
{
"epoch": 0.25,
"grad_norm": 0.10907820612192154,
"learning_rate": 0.0001985084539265195,
"loss": 1.1139,
"step": 242
},
{
"epoch": 0.25,
"grad_norm": 0.1067519262433052,
"learning_rate": 0.0001984367702028027,
"loss": 1.2389,
"step": 243
},
{
"epoch": 0.25,
"grad_norm": 0.11710456013679504,
"learning_rate": 0.0001983634176670951,
"loss": 1.2421,
"step": 244
},
{
"epoch": 0.25,
"grad_norm": 0.10886801034212112,
"learning_rate": 0.00019828839756295223,
"loss": 1.1696,
"step": 245
},
{
"epoch": 0.26,
"grad_norm": 0.1404796689748764,
"learning_rate": 0.00019821171116220012,
"loss": 1.29,
"step": 246
},
{
"epoch": 0.26,
"grad_norm": 0.1096196323633194,
"learning_rate": 0.00019813335976491385,
"loss": 1.3536,
"step": 247
},
{
"epoch": 0.26,
"grad_norm": 0.11487864702939987,
"learning_rate": 0.00019805334469939528,
"loss": 1.1371,
"step": 248
},
{
"epoch": 0.26,
"grad_norm": 0.1555541455745697,
"learning_rate": 0.00019797166732215076,
"loss": 1.1812,
"step": 249
},
{
"epoch": 0.26,
"grad_norm": 0.10709488391876221,
"learning_rate": 0.000197888329017868,
"loss": 1.101,
"step": 250
},
{
"epoch": 0.26,
"grad_norm": 0.10238959640264511,
"learning_rate": 0.00019780333119939262,
"loss": 1.2363,
"step": 251
},
{
"epoch": 0.26,
"grad_norm": 0.12807561457157135,
"learning_rate": 0.00019771667530770425,
"loss": 1.217,
"step": 252
},
{
"epoch": 0.26,
"grad_norm": 0.10418405383825302,
"learning_rate": 0.00019762836281189206,
"loss": 1.36,
"step": 253
},
{
"epoch": 0.26,
"grad_norm": 0.10400766879320145,
"learning_rate": 0.00019753839520912983,
"loss": 1.2368,
"step": 254
},
{
"epoch": 0.26,
"grad_norm": 0.1629195362329483,
"learning_rate": 0.00019744677402465052,
"loss": 1.2236,
"step": 255
},
{
"epoch": 0.27,
"grad_norm": 0.11335594952106476,
"learning_rate": 0.00019735350081172067,
"loss": 1.2062,
"step": 256
},
{
"epoch": 0.27,
"grad_norm": 0.18272048234939575,
"learning_rate": 0.00019725857715161373,
"loss": 1.3138,
"step": 257
},
{
"epoch": 0.27,
"grad_norm": 0.10922593623399734,
"learning_rate": 0.0001971620046535835,
"loss": 1.1697,
"step": 258
},
{
"epoch": 0.27,
"grad_norm": 0.10762567073106766,
"learning_rate": 0.00019706378495483665,
"loss": 1.1303,
"step": 259
},
{
"epoch": 0.27,
"grad_norm": 0.11802787333726883,
"learning_rate": 0.00019696391972050514,
"loss": 1.2909,
"step": 260
},
{
"epoch": 0.27,
"grad_norm": 0.11715908348560333,
"learning_rate": 0.00019686241064361792,
"loss": 1.1456,
"step": 261
},
{
"epoch": 0.27,
"grad_norm": 0.11491627991199493,
"learning_rate": 0.00019675925944507224,
"loss": 1.2083,
"step": 262
},
{
"epoch": 0.27,
"grad_norm": 0.11131855100393295,
"learning_rate": 0.0001966544678736044,
"loss": 1.2579,
"step": 263
},
{
"epoch": 0.27,
"grad_norm": 0.12063883244991302,
"learning_rate": 0.00019654803770576026,
"loss": 1.3887,
"step": 264
},
{
"epoch": 0.28,
"grad_norm": 0.14915111660957336,
"learning_rate": 0.0001964399707458649,
"loss": 1.2351,
"step": 265
},
{
"epoch": 0.28,
"grad_norm": 0.11454611271619797,
"learning_rate": 0.00019633026882599227,
"loss": 1.3325,
"step": 266
},
{
"epoch": 0.28,
"grad_norm": 0.10960078984498978,
"learning_rate": 0.00019621893380593396,
"loss": 1.3,
"step": 267
},
{
"epoch": 0.28,
"grad_norm": 0.11405951529741287,
"learning_rate": 0.0001961059675731678,
"loss": 1.0641,
"step": 268
},
{
"epoch": 0.28,
"grad_norm": 0.10570972412824631,
"learning_rate": 0.00019599137204282565,
"loss": 1.1543,
"step": 269
},
{
"epoch": 0.28,
"grad_norm": 0.11385346204042435,
"learning_rate": 0.00019587514915766124,
"loss": 1.1546,
"step": 270
},
{
"epoch": 0.28,
"grad_norm": 0.10122133791446686,
"learning_rate": 0.00019575730088801697,
"loss": 1.1946,
"step": 271
},
{
"epoch": 0.28,
"grad_norm": 0.14278005063533783,
"learning_rate": 0.00019563782923179063,
"loss": 1.1853,
"step": 272
},
{
"epoch": 0.28,
"grad_norm": 0.11730815470218658,
"learning_rate": 0.00019551673621440147,
"loss": 1.4478,
"step": 273
},
{
"epoch": 0.28,
"grad_norm": 0.12385374307632446,
"learning_rate": 0.00019539402388875598,
"loss": 1.4392,
"step": 274
},
{
"epoch": 0.29,
"grad_norm": 0.10802686959505081,
"learning_rate": 0.00019526969433521297,
"loss": 1.3126,
"step": 275
},
{
"epoch": 0.29,
"grad_norm": 0.12256834656000137,
"learning_rate": 0.00019514374966154825,
"loss": 1.1855,
"step": 276
},
{
"epoch": 0.29,
"grad_norm": 0.10675772279500961,
"learning_rate": 0.00019501619200291908,
"loss": 1.2578,
"step": 277
},
{
"epoch": 0.29,
"grad_norm": 0.1070014238357544,
"learning_rate": 0.00019488702352182786,
"loss": 1.1426,
"step": 278
},
{
"epoch": 0.29,
"grad_norm": 0.11141553521156311,
"learning_rate": 0.00019475624640808543,
"loss": 1.1046,
"step": 279
},
{
"epoch": 0.29,
"grad_norm": 0.10741408169269562,
"learning_rate": 0.00019462386287877407,
"loss": 1.1991,
"step": 280
},
{
"epoch": 0.29,
"grad_norm": 0.11526710540056229,
"learning_rate": 0.00019448987517820982,
"loss": 1.3507,
"step": 281
},
{
"epoch": 0.29,
"grad_norm": 0.1025758609175682,
"learning_rate": 0.0001943542855779044,
"loss": 1.3,
"step": 282
},
{
"epoch": 0.29,
"grad_norm": 0.11231940984725952,
"learning_rate": 0.0001942170963765268,
"loss": 1.2408,
"step": 283
},
{
"epoch": 0.29,
"grad_norm": 0.1058710590004921,
"learning_rate": 0.0001940783098998643,
"loss": 1.1668,
"step": 284
},
{
"epoch": 0.3,
"grad_norm": 0.1024676039814949,
"learning_rate": 0.00019393792850078292,
"loss": 1.2192,
"step": 285
},
{
"epoch": 0.3,
"grad_norm": 0.10104867815971375,
"learning_rate": 0.00019379595455918774,
"loss": 1.2881,
"step": 286
},
{
"epoch": 0.3,
"grad_norm": 0.1211661621928215,
"learning_rate": 0.00019365239048198225,
"loss": 1.3614,
"step": 287
},
{
"epoch": 0.3,
"grad_norm": 0.1058136522769928,
"learning_rate": 0.0001935072387030279,
"loss": 1.2043,
"step": 288
},
{
"epoch": 0.3,
"grad_norm": 0.3414977490901947,
"learning_rate": 0.0001933605016831026,
"loss": 1.1638,
"step": 289
},
{
"epoch": 0.3,
"grad_norm": 0.09908904135227203,
"learning_rate": 0.00019321218190985904,
"loss": 1.0565,
"step": 290
},
{
"epoch": 0.3,
"grad_norm": 0.11039788275957108,
"learning_rate": 0.00019306228189778254,
"loss": 1.1324,
"step": 291
},
{
"epoch": 0.3,
"grad_norm": 0.1092277467250824,
"learning_rate": 0.0001929108041881485,
"loss": 1.0857,
"step": 292
},
{
"epoch": 0.3,
"grad_norm": 0.11304416507482529,
"learning_rate": 0.00019275775134897918,
"loss": 1.3365,
"step": 293
},
{
"epoch": 0.31,
"grad_norm": 0.1239667758345604,
"learning_rate": 0.00019260312597500022,
"loss": 1.2251,
"step": 294
},
{
"epoch": 0.31,
"grad_norm": 0.1069445088505745,
"learning_rate": 0.00019244693068759666,
"loss": 1.1466,
"step": 295
},
{
"epoch": 0.31,
"grad_norm": 0.11692642420530319,
"learning_rate": 0.00019228916813476853,
"loss": 0.947,
"step": 296
},
{
"epoch": 0.31,
"grad_norm": 0.11571913212537766,
"learning_rate": 0.00019212984099108594,
"loss": 1.1865,
"step": 297
},
{
"epoch": 0.31,
"grad_norm": 0.14679546654224396,
"learning_rate": 0.00019196895195764362,
"loss": 1.2747,
"step": 298
},
{
"epoch": 0.31,
"grad_norm": 0.11377564072608948,
"learning_rate": 0.00019180650376201535,
"loss": 1.2149,
"step": 299
},
{
"epoch": 0.31,
"grad_norm": 0.14246807992458344,
"learning_rate": 0.0001916424991582075,
"loss": 1.3234,
"step": 300
},
{
"epoch": 0.31,
"grad_norm": 0.11343812942504883,
"learning_rate": 0.00019147694092661255,
"loss": 1.2125,
"step": 301
},
{
"epoch": 0.31,
"grad_norm": 0.49394461512565613,
"learning_rate": 0.00019130983187396172,
"loss": 1.2545,
"step": 302
},
{
"epoch": 0.31,
"grad_norm": 0.12038710713386536,
"learning_rate": 0.0001911411748332776,
"loss": 1.2441,
"step": 303
},
{
"epoch": 0.32,
"grad_norm": 0.112415611743927,
"learning_rate": 0.00019097097266382597,
"loss": 1.1986,
"step": 304
},
{
"epoch": 0.32,
"grad_norm": 0.1328456848859787,
"learning_rate": 0.0001907992282510675,
"loss": 1.0427,
"step": 305
},
{
"epoch": 0.32,
"grad_norm": 0.12180749326944351,
"learning_rate": 0.00019062594450660857,
"loss": 1.2805,
"step": 306
},
{
"epoch": 0.32,
"grad_norm": 0.16928140819072723,
"learning_rate": 0.00019045112436815215,
"loss": 1.1842,
"step": 307
},
{
"epoch": 0.32,
"grad_norm": 0.15169909596443176,
"learning_rate": 0.00019027477079944788,
"loss": 1.1867,
"step": 308
},
{
"epoch": 0.32,
"grad_norm": 0.11884856224060059,
"learning_rate": 0.0001900968867902419,
"loss": 1.2883,
"step": 309
},
{
"epoch": 0.32,
"grad_norm": 0.10564276576042175,
"learning_rate": 0.0001899174753562261,
"loss": 1.264,
"step": 310
},
{
"epoch": 0.32,
"grad_norm": 0.12126502394676208,
"learning_rate": 0.0001897365395389869,
"loss": 1.2866,
"step": 311
},
{
"epoch": 0.32,
"grad_norm": 0.11135692894458771,
"learning_rate": 0.00018955408240595395,
"loss": 1.3224,
"step": 312
},
{
"epoch": 0.32,
"grad_norm": 0.11942031234502792,
"learning_rate": 0.00018937010705034787,
"loss": 1.2606,
"step": 313
},
{
"epoch": 0.33,
"grad_norm": 0.11544723808765411,
"learning_rate": 0.00018918461659112802,
"loss": 1.2025,
"step": 314
},
{
"epoch": 0.33,
"grad_norm": 0.12025148421525955,
"learning_rate": 0.00018899761417293941,
"loss": 1.4092,
"step": 315
},
{
"epoch": 0.33,
"grad_norm": 0.126939594745636,
"learning_rate": 0.00018880910296605955,
"loss": 1.1811,
"step": 316
},
{
"epoch": 0.33,
"grad_norm": 0.29961127042770386,
"learning_rate": 0.00018861908616634465,
"loss": 1.1998,
"step": 317
},
{
"epoch": 0.33,
"grad_norm": 0.12669076025485992,
"learning_rate": 0.00018842756699517538,
"loss": 1.1652,
"step": 318
},
{
"epoch": 0.33,
"grad_norm": 0.12721942365169525,
"learning_rate": 0.00018823454869940242,
"loss": 1.2145,
"step": 319
},
{
"epoch": 0.33,
"grad_norm": 0.1199335977435112,
"learning_rate": 0.0001880400345512913,
"loss": 1.2953,
"step": 320
},
{
"epoch": 0.33,
"grad_norm": 0.13686604797840118,
"learning_rate": 0.00018784402784846683,
"loss": 1.2455,
"step": 321
},
{
"epoch": 0.33,
"grad_norm": 0.1259993016719818,
"learning_rate": 0.00018764653191385736,
"loss": 0.9901,
"step": 322
},
{
"epoch": 0.34,
"grad_norm": 0.13265901803970337,
"learning_rate": 0.00018744755009563848,
"loss": 1.2294,
"step": 323
},
{
"epoch": 0.34,
"grad_norm": 0.11366987973451614,
"learning_rate": 0.00018724708576717605,
"loss": 1.3416,
"step": 324
},
{
"epoch": 0.34,
"grad_norm": 0.11126185208559036,
"learning_rate": 0.0001870451423269692,
"loss": 1.1254,
"step": 325
},
{
"epoch": 0.34,
"grad_norm": 0.11551456153392792,
"learning_rate": 0.0001868417231985926,
"loss": 1.2113,
"step": 326
},
{
"epoch": 0.34,
"grad_norm": 0.11155180633068085,
"learning_rate": 0.00018663683183063845,
"loss": 1.1587,
"step": 327
},
{
"epoch": 0.34,
"grad_norm": 0.1066342443227768,
"learning_rate": 0.0001864304716966581,
"loss": 1.3596,
"step": 328
},
{
"epoch": 0.34,
"grad_norm": 0.11637596786022186,
"learning_rate": 0.00018622264629510299,
"loss": 1.2117,
"step": 329
},
{
"epoch": 0.34,
"grad_norm": 0.10947410017251968,
"learning_rate": 0.00018601335914926557,
"loss": 0.9963,
"step": 330
},
{
"epoch": 0.34,
"grad_norm": 0.1826743334531784,
"learning_rate": 0.0001858026138072193,
"loss": 1.1548,
"step": 331
},
{
"epoch": 0.34,
"grad_norm": 0.12238079309463501,
"learning_rate": 0.00018559041384175876,
"loss": 1.0507,
"step": 332
},
{
"epoch": 0.35,
"grad_norm": 0.1196766272187233,
"learning_rate": 0.00018537676285033887,
"loss": 1.1443,
"step": 333
},
{
"epoch": 0.35,
"grad_norm": 0.1195516511797905,
"learning_rate": 0.00018516166445501405,
"loss": 1.083,
"step": 334
},
{
"epoch": 0.35,
"grad_norm": 0.111576147377491,
"learning_rate": 0.0001849451223023767,
"loss": 1.192,
"step": 335
},
{
"epoch": 0.35,
"grad_norm": 0.12735474109649658,
"learning_rate": 0.0001847271400634955,
"loss": 1.2692,
"step": 336
},
{
"epoch": 0.35,
"grad_norm": 0.10551782697439194,
"learning_rate": 0.00018450772143385307,
"loss": 1.2476,
"step": 337
},
{
"epoch": 0.35,
"grad_norm": 0.11405996978282928,
"learning_rate": 0.00018428687013328336,
"loss": 1.1888,
"step": 338
},
{
"epoch": 0.35,
"grad_norm": 0.13341975212097168,
"learning_rate": 0.00018406458990590858,
"loss": 1.2129,
"step": 339
},
{
"epoch": 0.35,
"grad_norm": 0.13987426459789276,
"learning_rate": 0.00018384088452007578,
"loss": 1.2114,
"step": 340
},
{
"epoch": 0.35,
"grad_norm": 0.1929798275232315,
"learning_rate": 0.0001836157577682928,
"loss": 1.1411,
"step": 341
},
{
"epoch": 0.35,
"grad_norm": 0.11936288326978683,
"learning_rate": 0.00018338921346716425,
"loss": 1.2696,
"step": 342
},
{
"epoch": 0.36,
"grad_norm": 0.10896290838718414,
"learning_rate": 0.0001831612554573265,
"loss": 1.2438,
"step": 343
},
{
"epoch": 0.36,
"grad_norm": 0.11335362493991852,
"learning_rate": 0.00018293188760338286,
"loss": 1.0288,
"step": 344
},
{
"epoch": 0.36,
"grad_norm": 0.11729809641838074,
"learning_rate": 0.00018270111379383775,
"loss": 1.1118,
"step": 345
},
{
"epoch": 0.36,
"grad_norm": 0.11622367799282074,
"learning_rate": 0.0001824689379410311,
"loss": 1.1515,
"step": 346
},
{
"epoch": 0.36,
"grad_norm": 0.11727786809206009,
"learning_rate": 0.00018223536398107176,
"loss": 1.2468,
"step": 347
},
{
"epoch": 0.36,
"grad_norm": 0.11702138185501099,
"learning_rate": 0.00018200039587377097,
"loss": 1.2647,
"step": 348
},
{
"epoch": 0.36,
"grad_norm": 0.12109760195016861,
"learning_rate": 0.0001817640376025751,
"loss": 1.2282,
"step": 349
},
{
"epoch": 0.36,
"grad_norm": 0.11537209153175354,
"learning_rate": 0.00018152629317449813,
"loss": 1.178,
"step": 350
},
{
"epoch": 0.36,
"grad_norm": 0.1228351965546608,
"learning_rate": 0.00018128716662005383,
"loss": 1.2099,
"step": 351
},
{
"epoch": 0.37,
"grad_norm": 0.15511314570903778,
"learning_rate": 0.0001810466619931872,
"loss": 1.2881,
"step": 352
},
{
"epoch": 0.37,
"grad_norm": 0.1076040267944336,
"learning_rate": 0.00018080478337120605,
"loss": 1.264,
"step": 353
},
{
"epoch": 0.37,
"grad_norm": 0.14106139540672302,
"learning_rate": 0.00018056153485471166,
"loss": 1.2821,
"step": 354
},
{
"epoch": 0.37,
"grad_norm": 0.11988777667284012,
"learning_rate": 0.00018031692056752924,
"loss": 1.2771,
"step": 355
},
{
"epoch": 0.37,
"grad_norm": 0.10703491419553757,
"learning_rate": 0.0001800709446566382,
"loss": 1.2525,
"step": 356
},
{
"epoch": 0.37,
"grad_norm": 0.11741764843463898,
"learning_rate": 0.00017982361129210172,
"loss": 1.2151,
"step": 357
},
{
"epoch": 0.37,
"grad_norm": 0.19867832958698273,
"learning_rate": 0.00017957492466699605,
"loss": 1.2559,
"step": 358
},
{
"epoch": 0.37,
"grad_norm": 0.12436340004205704,
"learning_rate": 0.00017932488899733947,
"loss": 1.3145,
"step": 359
},
{
"epoch": 0.37,
"grad_norm": 0.12323563545942307,
"learning_rate": 0.00017907350852202077,
"loss": 1.2145,
"step": 360
},
{
"epoch": 0.37,
"grad_norm": 0.126682311296463,
"learning_rate": 0.00017882078750272742,
"loss": 1.2805,
"step": 361
},
{
"epoch": 0.38,
"grad_norm": 0.10707955062389374,
"learning_rate": 0.00017856673022387338,
"loss": 1.1415,
"step": 362
},
{
"epoch": 0.38,
"grad_norm": 0.1281125843524933,
"learning_rate": 0.0001783113409925263,
"loss": 1.3117,
"step": 363
},
{
"epoch": 0.38,
"grad_norm": 0.11295267939567566,
"learning_rate": 0.00017805462413833468,
"loss": 1.1453,
"step": 364
},
{
"epoch": 0.38,
"grad_norm": 0.12421952188014984,
"learning_rate": 0.00017779658401345438,
"loss": 1.2293,
"step": 365
},
{
"epoch": 0.38,
"grad_norm": 0.11468549817800522,
"learning_rate": 0.00017753722499247478,
"loss": 1.1391,
"step": 366
},
{
"epoch": 0.38,
"grad_norm": 0.12964022159576416,
"learning_rate": 0.00017727655147234476,
"loss": 1.3975,
"step": 367
},
{
"epoch": 0.38,
"grad_norm": 0.11601359397172928,
"learning_rate": 0.00017701456787229804,
"loss": 1.2622,
"step": 368
},
{
"epoch": 0.38,
"grad_norm": 0.12835174798965454,
"learning_rate": 0.00017675127863377838,
"loss": 1.1703,
"step": 369
},
{
"epoch": 0.38,
"grad_norm": 0.11938933283090591,
"learning_rate": 0.00017648668822036408,
"loss": 1.2702,
"step": 370
},
{
"epoch": 0.39,
"grad_norm": 0.12092038989067078,
"learning_rate": 0.00017622080111769257,
"loss": 1.3653,
"step": 371
},
{
"epoch": 0.39,
"grad_norm": 0.1410731077194214,
"learning_rate": 0.00017595362183338412,
"loss": 1.2465,
"step": 372
},
{
"epoch": 0.39,
"grad_norm": 0.11769446730613708,
"learning_rate": 0.00017568515489696558,
"loss": 1.235,
"step": 373
},
{
"epoch": 0.39,
"grad_norm": 0.10854797065258026,
"learning_rate": 0.00017541540485979359,
"loss": 1.1758,
"step": 374
},
{
"epoch": 0.39,
"grad_norm": 0.10094323009252548,
"learning_rate": 0.0001751443762949772,
"loss": 1.0939,
"step": 375
},
{
"epoch": 0.39,
"grad_norm": 0.10836099833250046,
"learning_rate": 0.00017487207379730078,
"loss": 1.1618,
"step": 376
},
{
"epoch": 0.39,
"grad_norm": 0.1138240173459053,
"learning_rate": 0.00017459850198314563,
"loss": 1.1423,
"step": 377
},
{
"epoch": 0.39,
"grad_norm": 0.11237498372793198,
"learning_rate": 0.000174323665490412,
"loss": 1.0702,
"step": 378
},
{
"epoch": 0.39,
"grad_norm": 0.1078115925192833,
"learning_rate": 0.00017404756897844052,
"loss": 1.2523,
"step": 379
},
{
"epoch": 0.39,
"grad_norm": 0.11401957273483276,
"learning_rate": 0.00017377021712793297,
"loss": 1.2443,
"step": 380
},
{
"epoch": 0.4,
"grad_norm": 0.1109926849603653,
"learning_rate": 0.00017349161464087312,
"loss": 1.264,
"step": 381
},
{
"epoch": 0.4,
"grad_norm": 0.10931116342544556,
"learning_rate": 0.00017321176624044687,
"loss": 1.294,
"step": 382
},
{
"epoch": 0.4,
"grad_norm": 0.10655161738395691,
"learning_rate": 0.00017293067667096239,
"loss": 1.2629,
"step": 383
},
{
"epoch": 0.4,
"grad_norm": 0.11112368851900101,
"learning_rate": 0.00017264835069776945,
"loss": 1.25,
"step": 384
},
{
"epoch": 0.4,
"grad_norm": 0.1312362551689148,
"learning_rate": 0.0001723647931071788,
"loss": 1.1807,
"step": 385
},
{
"epoch": 0.4,
"grad_norm": 0.1174694374203682,
"learning_rate": 0.00017208000870638093,
"loss": 1.1719,
"step": 386
},
{
"epoch": 0.4,
"grad_norm": 0.11656120419502258,
"learning_rate": 0.00017179400232336463,
"loss": 1.4266,
"step": 387
},
{
"epoch": 0.4,
"grad_norm": 0.11490406095981598,
"learning_rate": 0.00017150677880683514,
"loss": 1.3041,
"step": 388
},
{
"epoch": 0.4,
"grad_norm": 0.1138778105378151,
"learning_rate": 0.00017121834302613186,
"loss": 1.3364,
"step": 389
},
{
"epoch": 0.4,
"grad_norm": 0.11219751089811325,
"learning_rate": 0.000170928699871146,
"loss": 1.3409,
"step": 390
},
{
"epoch": 0.41,
"grad_norm": 0.10369167476892471,
"learning_rate": 0.0001706378542522374,
"loss": 1.1775,
"step": 391
},
{
"epoch": 0.41,
"grad_norm": 0.11245618760585785,
"learning_rate": 0.00017034581110015156,
"loss": 1.1841,
"step": 392
},
{
"epoch": 0.41,
"grad_norm": 0.10895467549562454,
"learning_rate": 0.00017005257536593577,
"loss": 1.1855,
"step": 393
},
{
"epoch": 0.41,
"grad_norm": 0.10433463007211685,
"learning_rate": 0.00016975815202085554,
"loss": 1.2853,
"step": 394
},
{
"epoch": 0.41,
"grad_norm": 0.10486618429422379,
"learning_rate": 0.00016946254605630992,
"loss": 1.2029,
"step": 395
},
{
"epoch": 0.41,
"grad_norm": 0.12566353380680084,
"learning_rate": 0.00016916576248374719,
"loss": 1.187,
"step": 396
},
{
"epoch": 0.41,
"grad_norm": 0.11196866631507874,
"learning_rate": 0.00016886780633457973,
"loss": 1.1956,
"step": 397
},
{
"epoch": 0.41,
"grad_norm": 0.12952035665512085,
"learning_rate": 0.00016856868266009874,
"loss": 1.24,
"step": 398
},
{
"epoch": 0.41,
"grad_norm": 0.1032567098736763,
"learning_rate": 0.0001682683965313887,
"loss": 1.207,
"step": 399
},
{
"epoch": 0.42,
"grad_norm": 0.10673535615205765,
"learning_rate": 0.00016796695303924127,
"loss": 1.23,
"step": 400
},
{
"epoch": 0.42,
"grad_norm": 0.1103532612323761,
"learning_rate": 0.00016766435729406913,
"loss": 1.2994,
"step": 401
},
{
"epoch": 0.42,
"grad_norm": 0.11205995827913284,
"learning_rate": 0.0001673606144258192,
"loss": 1.1168,
"step": 402
},
{
"epoch": 0.42,
"grad_norm": 0.11950290203094482,
"learning_rate": 0.00016705572958388576,
"loss": 1.2152,
"step": 403
},
{
"epoch": 0.42,
"grad_norm": 0.17785818874835968,
"learning_rate": 0.0001667497079370231,
"loss": 1.2523,
"step": 404
},
{
"epoch": 0.42,
"grad_norm": 0.10789604485034943,
"learning_rate": 0.00016644255467325794,
"loss": 1.289,
"step": 405
},
{
"epoch": 0.42,
"grad_norm": 0.12129693478345871,
"learning_rate": 0.00016613427499980143,
"loss": 1.257,
"step": 406
},
{
"epoch": 0.42,
"grad_norm": 0.11083897203207016,
"learning_rate": 0.00016582487414296097,
"loss": 1.3349,
"step": 407
},
{
"epoch": 0.42,
"grad_norm": 0.09888826310634613,
"learning_rate": 0.0001655143573480515,
"loss": 1.1737,
"step": 408
},
{
"epoch": 0.42,
"grad_norm": 0.11880184710025787,
"learning_rate": 0.00016520272987930652,
"loss": 1.0338,
"step": 409
},
{
"epoch": 0.43,
"grad_norm": 0.11628952622413635,
"learning_rate": 0.00016488999701978903,
"loss": 1.2898,
"step": 410
},
{
"epoch": 0.43,
"grad_norm": 0.113735631108284,
"learning_rate": 0.0001645761640713019,
"loss": 1.1826,
"step": 411
},
{
"epoch": 0.43,
"grad_norm": 0.10678713023662567,
"learning_rate": 0.00016426123635429786,
"loss": 1.2231,
"step": 412
},
{
"epoch": 0.43,
"grad_norm": 0.11631446331739426,
"learning_rate": 0.00016394521920778948,
"loss": 1.0241,
"step": 413
},
{
"epoch": 0.43,
"grad_norm": 0.11213452368974686,
"learning_rate": 0.00016362811798925853,
"loss": 1.4935,
"step": 414
},
{
"epoch": 0.43,
"grad_norm": 0.1304715871810913,
"learning_rate": 0.0001633099380745652,
"loss": 1.1686,
"step": 415
},
{
"epoch": 0.43,
"grad_norm": 0.10429395735263824,
"learning_rate": 0.000162990684857857,
"loss": 1.0118,
"step": 416
},
{
"epoch": 0.43,
"grad_norm": 0.10091353952884674,
"learning_rate": 0.00016267036375147725,
"loss": 1.1401,
"step": 417
},
{
"epoch": 0.43,
"grad_norm": 0.10777878761291504,
"learning_rate": 0.00016234898018587337,
"loss": 1.33,
"step": 418
},
{
"epoch": 0.43,
"grad_norm": 0.11149658262729645,
"learning_rate": 0.00016202653960950474,
"loss": 1.2201,
"step": 419
},
{
"epoch": 0.44,
"grad_norm": 0.10549131035804749,
"learning_rate": 0.0001617030474887505,
"loss": 1.0758,
"step": 420
},
{
"epoch": 0.44,
"grad_norm": 0.12649282813072205,
"learning_rate": 0.0001613785093078166,
"loss": 1.2241,
"step": 421
},
{
"epoch": 0.44,
"grad_norm": 0.11720848083496094,
"learning_rate": 0.00016105293056864315,
"loss": 1.3001,
"step": 422
},
{
"epoch": 0.44,
"grad_norm": 0.12630730867385864,
"learning_rate": 0.0001607263167908109,
"loss": 0.986,
"step": 423
},
{
"epoch": 0.44,
"grad_norm": 0.11931440979242325,
"learning_rate": 0.00016039867351144778,
"loss": 1.2037,
"step": 424
},
{
"epoch": 0.44,
"grad_norm": 0.1198689192533493,
"learning_rate": 0.000160070006285135,
"loss": 1.2556,
"step": 425
},
{
"epoch": 0.44,
"grad_norm": 0.11019983142614365,
"learning_rate": 0.0001597403206838128,
"loss": 1.004,
"step": 426
},
{
"epoch": 0.44,
"grad_norm": 0.11280027031898499,
"learning_rate": 0.00015940962229668625,
"loss": 1.3164,
"step": 427
},
{
"epoch": 0.44,
"grad_norm": 0.12924431264400482,
"learning_rate": 0.00015907791673013016,
"loss": 1.1612,
"step": 428
},
{
"epoch": 0.45,
"grad_norm": 0.1206149235367775,
"learning_rate": 0.00015874520960759423,
"loss": 1.2171,
"step": 429
},
{
"epoch": 0.45,
"grad_norm": 0.12173935025930405,
"learning_rate": 0.0001584115065695077,
"loss": 1.2215,
"step": 430
},
{
"epoch": 0.45,
"grad_norm": 0.126984640955925,
"learning_rate": 0.0001580768132731837,
"loss": 1.2379,
"step": 431
},
{
"epoch": 0.45,
"grad_norm": 0.12142551690340042,
"learning_rate": 0.0001577411353927233,
"loss": 1.1171,
"step": 432
},
{
"epoch": 0.45,
"grad_norm": 0.11669143289327621,
"learning_rate": 0.00015740447861891944,
"loss": 1.1148,
"step": 433
},
{
"epoch": 0.45,
"grad_norm": 0.12235064059495926,
"learning_rate": 0.00015706684865916026,
"loss": 1.2705,
"step": 434
},
{
"epoch": 0.45,
"grad_norm": 0.11695750802755356,
"learning_rate": 0.00015672825123733258,
"loss": 1.1602,
"step": 435
},
{
"epoch": 0.45,
"grad_norm": 0.10883978754281998,
"learning_rate": 0.00015638869209372457,
"loss": 1.3629,
"step": 436
},
{
"epoch": 0.45,
"grad_norm": 0.09961367398500443,
"learning_rate": 0.00015604817698492885,
"loss": 1.1844,
"step": 437
},
{
"epoch": 0.45,
"grad_norm": 0.10849742591381073,
"learning_rate": 0.00015570671168374438,
"loss": 1.2453,
"step": 438
},
{
"epoch": 0.46,
"grad_norm": 0.13158155977725983,
"learning_rate": 0.00015536430197907904,
"loss": 1.0024,
"step": 439
},
{
"epoch": 0.46,
"grad_norm": 0.11819962412118912,
"learning_rate": 0.00015502095367585122,
"loss": 1.2432,
"step": 440
},
{
"epoch": 0.46,
"grad_norm": 0.10652721673250198,
"learning_rate": 0.0001546766725948916,
"loss": 1.2042,
"step": 441
},
{
"epoch": 0.46,
"grad_norm": 0.10861877351999283,
"learning_rate": 0.0001543314645728442,
"loss": 1.4347,
"step": 442
},
{
"epoch": 0.46,
"grad_norm": 0.11992961913347244,
"learning_rate": 0.00015398533546206778,
"loss": 1.2897,
"step": 443
},
{
"epoch": 0.46,
"grad_norm": 0.11120472103357315,
"learning_rate": 0.00015363829113053634,
"loss": 1.1729,
"step": 444
},
{
"epoch": 0.46,
"grad_norm": 0.11962945014238358,
"learning_rate": 0.00015329033746173975,
"loss": 1.0863,
"step": 445
},
{
"epoch": 0.46,
"grad_norm": 0.10950727015733719,
"learning_rate": 0.00015294148035458404,
"loss": 1.2127,
"step": 446
},
{
"epoch": 0.46,
"grad_norm": 0.11483786255121231,
"learning_rate": 0.0001525917257232913,
"loss": 1.1465,
"step": 447
},
{
"epoch": 0.46,
"grad_norm": 0.11853177845478058,
"learning_rate": 0.0001522410794972995,
"loss": 1.2527,
"step": 448
},
{
"epoch": 0.47,
"grad_norm": 0.10667990148067474,
"learning_rate": 0.00015188954762116195,
"loss": 1.2347,
"step": 449
},
{
"epoch": 0.47,
"grad_norm": 0.1277850717306137,
"learning_rate": 0.0001515371360544465,
"loss": 1.2875,
"step": 450
},
{
"epoch": 0.47,
"grad_norm": 0.10839621722698212,
"learning_rate": 0.00015118385077163445,
"loss": 1.3415,
"step": 451
},
{
"epoch": 0.47,
"grad_norm": 0.11405637860298157,
"learning_rate": 0.00015082969776201947,
"loss": 1.1112,
"step": 452
},
{
"epoch": 0.47,
"grad_norm": 0.1116739809513092,
"learning_rate": 0.00015047468302960577,
"loss": 1.2876,
"step": 453
},
{
"epoch": 0.47,
"grad_norm": 0.10504814237356186,
"learning_rate": 0.00015011881259300653,
"loss": 1.2027,
"step": 454
},
{
"epoch": 0.47,
"grad_norm": 0.10286784917116165,
"learning_rate": 0.00014976209248534182,
"loss": 1.1372,
"step": 455
},
{
"epoch": 0.47,
"grad_norm": 0.1141451969742775,
"learning_rate": 0.00014940452875413627,
"loss": 1.0566,
"step": 456
},
{
"epoch": 0.47,
"grad_norm": 0.12613342702388763,
"learning_rate": 0.00014904612746121659,
"loss": 0.9762,
"step": 457
},
{
"epoch": 0.48,
"grad_norm": 0.11521808058023453,
"learning_rate": 0.00014868689468260874,
"loss": 1.144,
"step": 458
},
{
"epoch": 0.48,
"grad_norm": 0.10090262442827225,
"learning_rate": 0.0001483268365084351,
"loss": 1.0197,
"step": 459
},
{
"epoch": 0.48,
"grad_norm": 0.11367885768413544,
"learning_rate": 0.0001479659590428109,
"loss": 1.3636,
"step": 460
},
{
"epoch": 0.48,
"grad_norm": 0.11813089996576309,
"learning_rate": 0.0001476042684037411,
"loss": 1.314,
"step": 461
},
{
"epoch": 0.48,
"grad_norm": 0.10709206759929657,
"learning_rate": 0.0001472417707230164,
"loss": 1.2899,
"step": 462
},
{
"epoch": 0.48,
"grad_norm": 0.11429055780172348,
"learning_rate": 0.00014687847214610943,
"loss": 1.2237,
"step": 463
},
{
"epoch": 0.48,
"grad_norm": 0.10949011892080307,
"learning_rate": 0.00014651437883207054,
"loss": 1.3596,
"step": 464
},
{
"epoch": 0.48,
"grad_norm": 0.1138731986284256,
"learning_rate": 0.00014614949695342336,
"loss": 1.0538,
"step": 465
},
{
"epoch": 0.48,
"grad_norm": 0.10718726366758347,
"learning_rate": 0.00014578383269606005,
"loss": 1.2639,
"step": 466
},
{
"epoch": 0.48,
"grad_norm": 0.10555477440357208,
"learning_rate": 0.0001454173922591367,
"loss": 1.2917,
"step": 467
},
{
"epoch": 0.49,
"grad_norm": 0.10175728797912598,
"learning_rate": 0.00014505018185496802,
"loss": 1.111,
"step": 468
},
{
"epoch": 0.49,
"grad_norm": 0.113253653049469,
"learning_rate": 0.00014468220770892206,
"loss": 1.0426,
"step": 469
},
{
"epoch": 0.49,
"grad_norm": 0.11308235675096512,
"learning_rate": 0.0001443134760593147,
"loss": 1.2584,
"step": 470
},
{
"epoch": 0.49,
"grad_norm": 0.10383269935846329,
"learning_rate": 0.0001439439931573039,
"loss": 1.0795,
"step": 471
},
{
"epoch": 0.49,
"grad_norm": 0.10998781770467758,
"learning_rate": 0.00014357376526678367,
"loss": 1.2224,
"step": 472
},
{
"epoch": 0.49,
"grad_norm": 0.12233158946037292,
"learning_rate": 0.00014320279866427796,
"loss": 1.3125,
"step": 473
},
{
"epoch": 0.49,
"grad_norm": 0.11626295000314713,
"learning_rate": 0.00014283109963883419,
"loss": 1.3173,
"step": 474
},
{
"epoch": 0.49,
"grad_norm": 0.12676700949668884,
"learning_rate": 0.0001424586744919166,
"loss": 1.3502,
"step": 475
},
{
"epoch": 0.49,
"grad_norm": 0.11780429631471634,
"learning_rate": 0.00014208552953729947,
"loss": 1.3161,
"step": 476
},
{
"epoch": 0.5,
"grad_norm": 0.11618795245885849,
"learning_rate": 0.00014171167110096016,
"loss": 1.2574,
"step": 477
},
{
"epoch": 0.5,
"grad_norm": 0.10930506885051727,
"learning_rate": 0.00014133710552097174,
"loss": 1.243,
"step": 478
},
{
"epoch": 0.5,
"grad_norm": 0.11596094816923141,
"learning_rate": 0.00014096183914739553,
"loss": 1.2908,
"step": 479
},
{
"epoch": 0.5,
"grad_norm": 0.13881249725818634,
"learning_rate": 0.00014058587834217355,
"loss": 1.1379,
"step": 480
},
{
"epoch": 0.5,
"grad_norm": 0.09807305783033371,
"learning_rate": 0.00014020922947902066,
"loss": 1.1799,
"step": 481
},
{
"epoch": 0.5,
"grad_norm": 0.10549982637166977,
"learning_rate": 0.00013983189894331636,
"loss": 1.2467,
"step": 482
},
{
"epoch": 0.5,
"grad_norm": 0.7992545366287231,
"learning_rate": 0.00013945389313199669,
"loss": 1.2711,
"step": 483
},
{
"epoch": 0.5,
"grad_norm": 0.18441331386566162,
"learning_rate": 0.00013907521845344571,
"loss": 1.0915,
"step": 484
},
{
"epoch": 0.5,
"grad_norm": 0.12779419124126434,
"learning_rate": 0.00013869588132738695,
"loss": 1.2633,
"step": 485
},
{
"epoch": 0.5,
"grad_norm": 0.13042038679122925,
"learning_rate": 0.00013831588818477436,
"loss": 1.2233,
"step": 486
},
{
"epoch": 0.51,
"grad_norm": 0.14113464951515198,
"learning_rate": 0.00013793524546768356,
"loss": 1.0653,
"step": 487
},
{
"epoch": 0.51,
"grad_norm": 0.12155605107545853,
"learning_rate": 0.0001375539596292025,
"loss": 1.2181,
"step": 488
},
{
"epoch": 0.51,
"grad_norm": 0.12651774287223816,
"learning_rate": 0.00013717203713332188,
"loss": 1.3338,
"step": 489
},
{
"epoch": 0.51,
"grad_norm": 0.2244076430797577,
"learning_rate": 0.00013678948445482598,
"loss": 1.3459,
"step": 490
},
{
"epoch": 0.51,
"grad_norm": 0.1257290095090866,
"learning_rate": 0.00013640630807918247,
"loss": 1.0061,
"step": 491
},
{
"epoch": 0.51,
"grad_norm": 0.11858661472797394,
"learning_rate": 0.00013602251450243273,
"loss": 1.1775,
"step": 492
},
{
"epoch": 0.51,
"grad_norm": 0.1279936581850052,
"learning_rate": 0.00013563811023108158,
"loss": 1.3702,
"step": 493
},
{
"epoch": 0.51,
"grad_norm": 0.13254624605178833,
"learning_rate": 0.00013525310178198705,
"loss": 1.3612,
"step": 494
},
{
"epoch": 0.51,
"grad_norm": 0.1243981122970581,
"learning_rate": 0.00013486749568225001,
"loss": 1.1652,
"step": 495
},
{
"epoch": 0.51,
"grad_norm": 0.1259343922138214,
"learning_rate": 0.0001344812984691031,
"loss": 1.2998,
"step": 496
},
{
"epoch": 0.52,
"grad_norm": 0.12421772629022598,
"learning_rate": 0.00013409451668980048,
"loss": 1.2688,
"step": 497
},
{
"epoch": 0.52,
"grad_norm": 0.11533673852682114,
"learning_rate": 0.0001337071569015063,
"loss": 1.2123,
"step": 498
},
{
"epoch": 0.52,
"grad_norm": 0.2135774940252304,
"learning_rate": 0.00013331922567118393,
"loss": 1.173,
"step": 499
},
{
"epoch": 0.52,
"grad_norm": 0.1305667608976364,
"learning_rate": 0.00013293072957548443,
"loss": 1.2795,
"step": 500
},
{
"epoch": 0.52,
"grad_norm": 0.11869475990533829,
"learning_rate": 0.00013254167520063509,
"loss": 1.1984,
"step": 501
},
{
"epoch": 0.52,
"grad_norm": 0.13584646582603455,
"learning_rate": 0.0001321520691423278,
"loss": 1.2018,
"step": 502
},
{
"epoch": 0.52,
"grad_norm": 0.12778714299201965,
"learning_rate": 0.0001317619180056072,
"loss": 1.2817,
"step": 503
},
{
"epoch": 0.52,
"grad_norm": 0.12326094508171082,
"learning_rate": 0.0001313712284047587,
"loss": 1.3434,
"step": 504
},
{
"epoch": 0.52,
"grad_norm": 0.11255846172571182,
"learning_rate": 0.0001309800069631964,
"loss": 1.1986,
"step": 505
},
{
"epoch": 0.53,
"grad_norm": 0.11870656907558441,
"learning_rate": 0.0001305882603133508,
"loss": 1.2409,
"step": 506
},
{
"epoch": 0.53,
"grad_norm": 0.11277451366186142,
"learning_rate": 0.0001301959950965562,
"loss": 1.4011,
"step": 507
},
{
"epoch": 0.53,
"grad_norm": 0.11900844424962997,
"learning_rate": 0.00012980321796293836,
"loss": 1.2379,
"step": 508
},
{
"epoch": 0.53,
"grad_norm": 0.14404135942459106,
"learning_rate": 0.00012940993557130164,
"loss": 1.3902,
"step": 509
},
{
"epoch": 0.53,
"grad_norm": 0.12229623645544052,
"learning_rate": 0.000129016154589016,
"loss": 1.2842,
"step": 510
},
{
"epoch": 0.53,
"grad_norm": 0.1197381243109703,
"learning_rate": 0.00012862188169190418,
"loss": 1.3836,
"step": 511
},
{
"epoch": 0.53,
"grad_norm": 0.12395986169576645,
"learning_rate": 0.00012822712356412838,
"loss": 1.3508,
"step": 512
},
{
"epoch": 0.53,
"grad_norm": 0.12418900430202484,
"learning_rate": 0.00012783188689807697,
"loss": 1.1969,
"step": 513
},
{
"epoch": 0.53,
"grad_norm": 0.1138986125588417,
"learning_rate": 0.00012743617839425098,
"loss": 1.3654,
"step": 514
},
{
"epoch": 0.53,
"grad_norm": 0.11219038814306259,
"learning_rate": 0.0001270400047611508,
"loss": 1.2621,
"step": 515
},
{
"epoch": 0.54,
"grad_norm": 0.11220891773700714,
"learning_rate": 0.00012664337271516194,
"loss": 1.2384,
"step": 516
},
{
"epoch": 0.54,
"grad_norm": 0.11203695833683014,
"learning_rate": 0.0001262462889804416,
"loss": 1.1551,
"step": 517
},
{
"epoch": 0.54,
"grad_norm": 0.10459273308515549,
"learning_rate": 0.00012584876028880453,
"loss": 1.2276,
"step": 518
},
{
"epoch": 0.54,
"grad_norm": 0.10843029618263245,
"learning_rate": 0.00012545079337960882,
"loss": 1.2414,
"step": 519
},
{
"epoch": 0.54,
"grad_norm": 0.10928460955619812,
"learning_rate": 0.00012505239499964178,
"loss": 1.0391,
"step": 520
},
{
"epoch": 0.54,
"grad_norm": 0.1170789897441864,
"learning_rate": 0.0001246535719030055,
"loss": 1.2558,
"step": 521
},
{
"epoch": 0.54,
"grad_norm": 0.118638776242733,
"learning_rate": 0.00012425433085100224,
"loss": 1.2906,
"step": 522
},
{
"epoch": 0.54,
"grad_norm": 0.12077479064464569,
"learning_rate": 0.0001238546786120201,
"loss": 1.3222,
"step": 523
},
{
"epoch": 0.54,
"grad_norm": 0.12042129039764404,
"learning_rate": 0.00012345462196141797,
"loss": 1.2399,
"step": 524
},
{
"epoch": 0.54,
"grad_norm": 0.11980915814638138,
"learning_rate": 0.00012305416768141082,
"loss": 1.2426,
"step": 525
},
{
"epoch": 0.55,
"grad_norm": 0.10789747536182404,
"learning_rate": 0.00012265332256095464,
"loss": 1.3437,
"step": 526
},
{
"epoch": 0.55,
"grad_norm": 0.10967132449150085,
"learning_rate": 0.00012225209339563145,
"loss": 1.1329,
"step": 527
},
{
"epoch": 0.55,
"grad_norm": 0.11903247982263565,
"learning_rate": 0.00012185048698753403,
"loss": 1.2107,
"step": 528
},
{
"epoch": 0.55,
"grad_norm": 0.10917498171329498,
"learning_rate": 0.00012144851014515055,
"loss": 1.1689,
"step": 529
},
{
"epoch": 0.55,
"grad_norm": 0.11308782547712326,
"learning_rate": 0.00012104616968324927,
"loss": 1.2487,
"step": 530
},
{
"epoch": 0.55,
"grad_norm": 0.10030585527420044,
"learning_rate": 0.00012064347242276293,
"loss": 1.1451,
"step": 531
},
{
"epoch": 0.55,
"grad_norm": 0.1195712685585022,
"learning_rate": 0.00012024042519067309,
"loss": 1.2525,
"step": 532
},
{
"epoch": 0.55,
"grad_norm": 0.12115343660116196,
"learning_rate": 0.00011983703481989443,
"loss": 1.2348,
"step": 533
},
{
"epoch": 0.55,
"grad_norm": 0.1060243546962738,
"learning_rate": 0.00011943330814915896,
"loss": 1.0223,
"step": 534
},
{
"epoch": 0.56,
"grad_norm": 0.2178022414445877,
"learning_rate": 0.00011902925202289996,
"loss": 1.1158,
"step": 535
},
{
"epoch": 0.56,
"grad_norm": 0.13022154569625854,
"learning_rate": 0.00011862487329113606,
"loss": 1.2742,
"step": 536
},
{
"epoch": 0.56,
"grad_norm": 0.11495082825422287,
"learning_rate": 0.00011822017880935505,
"loss": 1.2562,
"step": 537
},
{
"epoch": 0.56,
"grad_norm": 0.11310116201639175,
"learning_rate": 0.00011781517543839768,
"loss": 1.2719,
"step": 538
},
{
"epoch": 0.56,
"grad_norm": 0.13280071318149567,
"learning_rate": 0.00011740987004434137,
"loss": 1.1481,
"step": 539
},
{
"epoch": 0.56,
"grad_norm": 0.12752439081668854,
"learning_rate": 0.00011700426949838363,
"loss": 1.3024,
"step": 540
},
{
"epoch": 0.56,
"grad_norm": 0.11650800704956055,
"learning_rate": 0.00011659838067672599,
"loss": 1.2154,
"step": 541
},
{
"epoch": 0.56,
"grad_norm": 0.11279530823230743,
"learning_rate": 0.00011619221046045687,
"loss": 1.1627,
"step": 542
},
{
"epoch": 0.56,
"grad_norm": 0.12641020119190216,
"learning_rate": 0.0001157857657354354,
"loss": 1.0148,
"step": 543
},
{
"epoch": 0.56,
"grad_norm": 0.10685007274150848,
"learning_rate": 0.00011537905339217448,
"loss": 1.288,
"step": 544
},
{
"epoch": 0.57,
"grad_norm": 0.10929278284311295,
"learning_rate": 0.00011497208032572384,
"loss": 1.123,
"step": 545
},
{
"epoch": 0.57,
"grad_norm": 0.11334774643182755,
"learning_rate": 0.00011456485343555344,
"loss": 1.2894,
"step": 546
},
{
"epoch": 0.57,
"grad_norm": 0.10330837219953537,
"learning_rate": 0.0001141573796254363,
"loss": 1.1484,
"step": 547
},
{
"epoch": 0.57,
"grad_norm": 0.09870249778032303,
"learning_rate": 0.00011374966580333147,
"loss": 1.0223,
"step": 548
},
{
"epoch": 0.57,
"grad_norm": 0.12955228984355927,
"learning_rate": 0.00011334171888126697,
"loss": 1.2058,
"step": 549
},
{
"epoch": 0.57,
"grad_norm": 0.11018647998571396,
"learning_rate": 0.00011293354577522263,
"loss": 1.1589,
"step": 550
},
{
"epoch": 0.57,
"grad_norm": 0.12055356800556183,
"learning_rate": 0.00011252515340501281,
"loss": 1.091,
"step": 551
},
{
"epoch": 0.57,
"grad_norm": 0.10831710696220398,
"learning_rate": 0.00011211654869416901,
"loss": 1.2076,
"step": 552
},
{
"epoch": 0.57,
"grad_norm": 0.11912648379802704,
"learning_rate": 0.00011170773856982268,
"loss": 1.2953,
"step": 553
},
{
"epoch": 0.57,
"grad_norm": 0.11210989207029343,
"learning_rate": 0.00011129872996258757,
"loss": 1.0549,
"step": 554
},
{
"epoch": 0.58,
"grad_norm": 0.10992615669965744,
"learning_rate": 0.0001108895298064424,
"loss": 1.1818,
"step": 555
},
{
"epoch": 0.58,
"grad_norm": 0.1124982088804245,
"learning_rate": 0.00011048014503861321,
"loss": 1.3218,
"step": 556
},
{
"epoch": 0.58,
"grad_norm": 0.12512408196926117,
"learning_rate": 0.00011007058259945584,
"loss": 1.2262,
"step": 557
},
{
"epoch": 0.58,
"grad_norm": 0.1207667887210846,
"learning_rate": 0.00010966084943233817,
"loss": 1.1636,
"step": 558
},
{
"epoch": 0.58,
"grad_norm": 0.10861886292695999,
"learning_rate": 0.00010925095248352239,
"loss": 1.1363,
"step": 559
},
{
"epoch": 0.58,
"grad_norm": 0.11505457758903503,
"learning_rate": 0.0001088408987020475,
"loss": 1.1823,
"step": 560
},
{
"epoch": 0.58,
"grad_norm": 0.10509267449378967,
"learning_rate": 0.00010843069503961111,
"loss": 1.2269,
"step": 561
},
{
"epoch": 0.58,
"grad_norm": 0.10376826673746109,
"learning_rate": 0.00010802034845045188,
"loss": 1.0244,
"step": 562
},
{
"epoch": 0.58,
"grad_norm": 0.10977223515510559,
"learning_rate": 0.00010760986589123146,
"loss": 1.2713,
"step": 563
},
{
"epoch": 0.59,
"grad_norm": 0.11492714285850525,
"learning_rate": 0.0001071992543209167,
"loss": 1.2242,
"step": 564
},
{
"epoch": 0.59,
"grad_norm": 0.10544589161872864,
"learning_rate": 0.00010678852070066159,
"loss": 1.2959,
"step": 565
},
{
"epoch": 0.59,
"grad_norm": 0.11260811984539032,
"learning_rate": 0.00010637767199368912,
"loss": 1.3398,
"step": 566
},
{
"epoch": 0.59,
"grad_norm": 0.10535457730293274,
"learning_rate": 0.00010596671516517356,
"loss": 1.3178,
"step": 567
},
{
"epoch": 0.59,
"grad_norm": 0.10959797352552414,
"learning_rate": 0.00010555565718212197,
"loss": 1.0378,
"step": 568
},
{
"epoch": 0.59,
"grad_norm": 0.10014626383781433,
"learning_rate": 0.00010514450501325645,
"loss": 1.2384,
"step": 569
},
{
"epoch": 0.59,
"grad_norm": 0.1097472757101059,
"learning_rate": 0.00010473326562889582,
"loss": 1.1366,
"step": 570
},
{
"epoch": 0.59,
"grad_norm": 0.10998505353927612,
"learning_rate": 0.0001043219460008374,
"loss": 1.1083,
"step": 571
},
{
"epoch": 0.59,
"grad_norm": 0.11129113286733627,
"learning_rate": 0.00010391055310223899,
"loss": 1.1951,
"step": 572
},
{
"epoch": 0.59,
"grad_norm": 0.10348080098628998,
"learning_rate": 0.00010349909390750046,
"loss": 1.1871,
"step": 573
},
{
"epoch": 0.6,
"grad_norm": 0.10865020006895065,
"learning_rate": 0.00010308757539214572,
"loss": 1.2269,
"step": 574
},
{
"epoch": 0.6,
"grad_norm": 0.1128375232219696,
"learning_rate": 0.00010267600453270421,
"loss": 1.2358,
"step": 575
},
{
"epoch": 0.6,
"grad_norm": 0.13256721198558807,
"learning_rate": 0.0001022643883065929,
"loss": 1.3445,
"step": 576
},
{
"epoch": 0.6,
"grad_norm": 0.10738494992256165,
"learning_rate": 0.00010185273369199781,
"loss": 1.1042,
"step": 577
},
{
"epoch": 0.6,
"grad_norm": 0.10427848994731903,
"learning_rate": 0.00010144104766775572,
"loss": 1.0901,
"step": 578
},
{
"epoch": 0.6,
"grad_norm": 0.11618473380804062,
"learning_rate": 0.000101029337213236,
"loss": 0.781,
"step": 579
},
{
"epoch": 0.6,
"grad_norm": 0.20147624611854553,
"learning_rate": 0.00010061760930822211,
"loss": 1.2707,
"step": 580
},
{
"epoch": 0.6,
"grad_norm": 0.11436042189598083,
"learning_rate": 0.0001002058709327934,
"loss": 1.193,
"step": 581
},
{
"epoch": 0.6,
"grad_norm": 0.12396758794784546,
"learning_rate": 9.979412906720663e-05,
"loss": 1.0342,
"step": 582
},
{
"epoch": 0.61,
"grad_norm": 0.11723118275403976,
"learning_rate": 9.938239069177791e-05,
"loss": 1.1791,
"step": 583
},
{
"epoch": 0.61,
"grad_norm": 0.12710683047771454,
"learning_rate": 9.897066278676404e-05,
"loss": 1.023,
"step": 584
},
{
"epoch": 0.61,
"grad_norm": 0.10067608207464218,
"learning_rate": 9.85589523322443e-05,
"loss": 1.1753,
"step": 585
},
{
"epoch": 0.61,
"grad_norm": 0.11074398458003998,
"learning_rate": 9.81472663080022e-05,
"loss": 1.2567,
"step": 586
},
{
"epoch": 0.61,
"grad_norm": 0.16948086023330688,
"learning_rate": 9.773561169340712e-05,
"loss": 1.3789,
"step": 587
},
{
"epoch": 0.61,
"grad_norm": 0.10852295160293579,
"learning_rate": 9.732399546729578e-05,
"loss": 1.3642,
"step": 588
},
{
"epoch": 0.61,
"grad_norm": 0.10131103545427322,
"learning_rate": 9.691242460785433e-05,
"loss": 1.129,
"step": 589
},
{
"epoch": 0.61,
"grad_norm": 0.11037011444568634,
"learning_rate": 9.650090609249956e-05,
"loss": 1.1974,
"step": 590
},
{
"epoch": 0.61,
"grad_norm": 0.12829488515853882,
"learning_rate": 9.608944689776103e-05,
"loss": 1.2126,
"step": 591
},
{
"epoch": 0.61,
"grad_norm": 0.16550996899604797,
"learning_rate": 9.567805399916259e-05,
"loss": 1.228,
"step": 592
},
{
"epoch": 0.62,
"grad_norm": 0.1353956162929535,
"learning_rate": 9.526673437110419e-05,
"loss": 1.2877,
"step": 593
},
{
"epoch": 0.62,
"grad_norm": 0.17490524053573608,
"learning_rate": 9.485549498674358e-05,
"loss": 1.212,
"step": 594
},
{
"epoch": 0.62,
"grad_norm": 0.10544750839471817,
"learning_rate": 9.444434281787807e-05,
"loss": 1.3147,
"step": 595
},
{
"epoch": 0.62,
"grad_norm": 0.10780143737792969,
"learning_rate": 9.403328483482646e-05,
"loss": 1.1851,
"step": 596
},
{
"epoch": 0.62,
"grad_norm": 0.12775209546089172,
"learning_rate": 9.362232800631088e-05,
"loss": 1.1219,
"step": 597
},
{
"epoch": 0.62,
"grad_norm": 0.11476773023605347,
"learning_rate": 9.321147929933847e-05,
"loss": 1.2289,
"step": 598
},
{
"epoch": 0.62,
"grad_norm": 0.11137551069259644,
"learning_rate": 9.28007456790833e-05,
"loss": 1.2016,
"step": 599
},
{
"epoch": 0.62,
"grad_norm": 0.13261277973651886,
"learning_rate": 9.239013410876857e-05,
"loss": 1.389,
"step": 600
},
{
"epoch": 0.62,
"grad_norm": 0.10396253317594528,
"learning_rate": 9.197965154954815e-05,
"loss": 1.3316,
"step": 601
},
{
"epoch": 0.62,
"grad_norm": 0.11388034373521805,
"learning_rate": 9.15693049603889e-05,
"loss": 1.2501,
"step": 602
},
{
"epoch": 0.63,
"grad_norm": 0.1430211067199707,
"learning_rate": 9.115910129795251e-05,
"loss": 1.2259,
"step": 603
},
{
"epoch": 0.63,
"grad_norm": 0.12369144707918167,
"learning_rate": 9.074904751647762e-05,
"loss": 1.138,
"step": 604
},
{
"epoch": 0.63,
"grad_norm": 0.11555325239896774,
"learning_rate": 9.033915056766187e-05,
"loss": 1.1845,
"step": 605
},
{
"epoch": 0.63,
"grad_norm": 0.15159161388874054,
"learning_rate": 8.992941740054418e-05,
"loss": 1.3823,
"step": 606
},
{
"epoch": 0.63,
"grad_norm": 0.12028995901346207,
"learning_rate": 8.951985496138678e-05,
"loss": 1.0985,
"step": 607
},
{
"epoch": 0.63,
"grad_norm": 0.12002217024564743,
"learning_rate": 8.911047019355764e-05,
"loss": 1.2431,
"step": 608
},
{
"epoch": 0.63,
"grad_norm": 0.11859162151813507,
"learning_rate": 8.870127003741244e-05,
"loss": 1.1669,
"step": 609
},
{
"epoch": 0.63,
"grad_norm": 0.10495835542678833,
"learning_rate": 8.829226143017735e-05,
"loss": 1.1424,
"step": 610
},
{
"epoch": 0.63,
"grad_norm": 0.11981993168592453,
"learning_rate": 8.788345130583099e-05,
"loss": 1.0444,
"step": 611
},
{
"epoch": 0.64,
"grad_norm": 0.11924871802330017,
"learning_rate": 8.747484659498724e-05,
"loss": 1.1574,
"step": 612
},
{
"epoch": 0.64,
"grad_norm": 0.15531030297279358,
"learning_rate": 8.706645422477739e-05,
"loss": 1.1786,
"step": 613
},
{
"epoch": 0.64,
"grad_norm": 0.11026681959629059,
"learning_rate": 8.665828111873307e-05,
"loss": 1.2112,
"step": 614
},
{
"epoch": 0.64,
"grad_norm": 0.12397878617048264,
"learning_rate": 8.625033419666857e-05,
"loss": 1.2931,
"step": 615
},
{
"epoch": 0.64,
"grad_norm": 0.10594267398118973,
"learning_rate": 8.584262037456373e-05,
"loss": 1.1422,
"step": 616
},
{
"epoch": 0.64,
"grad_norm": 0.11834708601236343,
"learning_rate": 8.54351465644466e-05,
"loss": 1.16,
"step": 617
},
{
"epoch": 0.64,
"grad_norm": 0.12605226039886475,
"learning_rate": 8.502791967427618e-05,
"loss": 1.3575,
"step": 618
},
{
"epoch": 0.64,
"grad_norm": 0.12767426669597626,
"learning_rate": 8.462094660782556e-05,
"loss": 1.3343,
"step": 619
},
{
"epoch": 0.64,
"grad_norm": 0.12377568334341049,
"learning_rate": 8.42142342645646e-05,
"loss": 1.2027,
"step": 620
},
{
"epoch": 0.64,
"grad_norm": 0.11611881852149963,
"learning_rate": 8.380778953954313e-05,
"loss": 1.1748,
"step": 621
},
{
"epoch": 0.65,
"grad_norm": 0.10847889631986618,
"learning_rate": 8.340161932327404e-05,
"loss": 1.138,
"step": 622
},
{
"epoch": 0.65,
"grad_norm": 0.11104707419872284,
"learning_rate": 8.299573050161638e-05,
"loss": 1.2403,
"step": 623
},
{
"epoch": 0.65,
"grad_norm": 0.10397801548242569,
"learning_rate": 8.259012995565867e-05,
"loss": 1.2159,
"step": 624
},
{
"epoch": 0.65,
"grad_norm": 0.10560380667448044,
"learning_rate": 8.21848245616023e-05,
"loss": 1.0932,
"step": 625
},
{
"epoch": 0.65,
"grad_norm": 0.10818043351173401,
"learning_rate": 8.177982119064496e-05,
"loss": 1.2667,
"step": 626
},
{
"epoch": 0.65,
"grad_norm": 0.1053985133767128,
"learning_rate": 8.137512670886397e-05,
"loss": 1.3267,
"step": 627
},
{
"epoch": 0.65,
"grad_norm": 0.10604721307754517,
"learning_rate": 8.097074797710006e-05,
"loss": 1.2254,
"step": 628
},
{
"epoch": 0.65,
"grad_norm": 0.1083521693944931,
"learning_rate": 8.056669185084108e-05,
"loss": 1.389,
"step": 629
},
{
"epoch": 0.65,
"grad_norm": 0.12045227736234665,
"learning_rate": 8.016296518010558e-05,
"loss": 1.245,
"step": 630
},
{
"epoch": 0.65,
"grad_norm": 0.11613083630800247,
"learning_rate": 7.975957480932695e-05,
"loss": 1.2618,
"step": 631
},
{
"epoch": 0.66,
"grad_norm": 0.11401872336864471,
"learning_rate": 7.935652757723708e-05,
"loss": 1.2073,
"step": 632
},
{
"epoch": 0.66,
"grad_norm": 0.1841176450252533,
"learning_rate": 7.895383031675074e-05,
"loss": 1.2132,
"step": 633
},
{
"epoch": 0.66,
"grad_norm": 0.10668601095676422,
"learning_rate": 7.855148985484946e-05,
"loss": 1.2177,
"step": 634
},
{
"epoch": 0.66,
"grad_norm": 0.10970567911863327,
"learning_rate": 7.814951301246598e-05,
"loss": 1.2306,
"step": 635
},
{
"epoch": 0.66,
"grad_norm": 0.10744346678256989,
"learning_rate": 7.774790660436858e-05,
"loss": 1.2155,
"step": 636
},
{
"epoch": 0.66,
"grad_norm": 0.09847363084554672,
"learning_rate": 7.734667743904539e-05,
"loss": 1.1383,
"step": 637
},
{
"epoch": 0.66,
"grad_norm": 0.12585873901844025,
"learning_rate": 7.69458323185892e-05,
"loss": 1.2411,
"step": 638
},
{
"epoch": 0.66,
"grad_norm": 0.13609972596168518,
"learning_rate": 7.654537803858204e-05,
"loss": 1.1972,
"step": 639
},
{
"epoch": 0.66,
"grad_norm": 0.10885506868362427,
"learning_rate": 7.614532138797993e-05,
"loss": 1.2197,
"step": 640
},
{
"epoch": 0.67,
"grad_norm": 0.10706409066915512,
"learning_rate": 7.57456691489978e-05,
"loss": 1.0704,
"step": 641
},
{
"epoch": 0.67,
"grad_norm": 0.11232727020978928,
"learning_rate": 7.534642809699455e-05,
"loss": 1.2443,
"step": 642
},
{
"epoch": 0.67,
"grad_norm": 0.1036246195435524,
"learning_rate": 7.494760500035823e-05,
"loss": 1.1623,
"step": 643
},
{
"epoch": 0.67,
"grad_norm": 0.11857923120260239,
"learning_rate": 7.454920662039117e-05,
"loss": 1.3161,
"step": 644
},
{
"epoch": 0.67,
"grad_norm": 0.1063128188252449,
"learning_rate": 7.415123971119549e-05,
"loss": 1.37,
"step": 645
},
{
"epoch": 0.67,
"grad_norm": 0.11207219213247299,
"learning_rate": 7.375371101955841e-05,
"loss": 1.1433,
"step": 646
},
{
"epoch": 0.67,
"grad_norm": 0.1198432594537735,
"learning_rate": 7.335662728483808e-05,
"loss": 1.2341,
"step": 647
},
{
"epoch": 0.67,
"grad_norm": 0.14110444486141205,
"learning_rate": 7.295999523884921e-05,
"loss": 1.2201,
"step": 648
},
{
"epoch": 0.67,
"grad_norm": 0.10053621232509613,
"learning_rate": 7.256382160574901e-05,
"loss": 1.1276,
"step": 649
},
{
"epoch": 0.67,
"grad_norm": 0.1087542399764061,
"learning_rate": 7.216811310192308e-05,
"loss": 1.3124,
"step": 650
},
{
"epoch": 0.68,
"grad_norm": 0.10470487177371979,
"learning_rate": 7.177287643587163e-05,
"loss": 1.0677,
"step": 651
},
{
"epoch": 0.68,
"grad_norm": 0.12348546087741852,
"learning_rate": 7.137811830809583e-05,
"loss": 1.0808,
"step": 652
},
{
"epoch": 0.68,
"grad_norm": 0.12709330022335052,
"learning_rate": 7.0983845410984e-05,
"loss": 1.1746,
"step": 653
},
{
"epoch": 0.68,
"grad_norm": 0.0996909812092781,
"learning_rate": 7.059006442869839e-05,
"loss": 1.2121,
"step": 654
},
{
"epoch": 0.68,
"grad_norm": 0.12972918152809143,
"learning_rate": 7.019678203706163e-05,
"loss": 1.149,
"step": 655
},
{
"epoch": 0.68,
"grad_norm": 0.10252442955970764,
"learning_rate": 6.980400490344383e-05,
"loss": 1.2643,
"step": 656
},
{
"epoch": 0.68,
"grad_norm": 0.11937206983566284,
"learning_rate": 6.941173968664923e-05,
"loss": 1.2484,
"step": 657
},
{
"epoch": 0.68,
"grad_norm": 0.11440838873386383,
"learning_rate": 6.901999303680358e-05,
"loss": 1.3256,
"step": 658
},
{
"epoch": 0.68,
"grad_norm": 0.10103975981473923,
"learning_rate": 6.862877159524132e-05,
"loss": 1.3196,
"step": 659
},
{
"epoch": 0.69,
"grad_norm": 0.10081090033054352,
"learning_rate": 6.823808199439284e-05,
"loss": 1.1371,
"step": 660
},
{
"epoch": 0.69,
"grad_norm": 0.16644109785556793,
"learning_rate": 6.784793085767221e-05,
"loss": 1.1239,
"step": 661
},
{
"epoch": 0.69,
"grad_norm": 0.1086033284664154,
"learning_rate": 6.745832479936491e-05,
"loss": 1.3244,
"step": 662
},
{
"epoch": 0.69,
"grad_norm": 0.11194080859422684,
"learning_rate": 6.70692704245156e-05,
"loss": 1.3153,
"step": 663
},
{
"epoch": 0.69,
"grad_norm": 0.1124996691942215,
"learning_rate": 6.668077432881609e-05,
"loss": 1.2942,
"step": 664
},
{
"epoch": 0.69,
"grad_norm": 0.3689382076263428,
"learning_rate": 6.629284309849372e-05,
"loss": 1.2606,
"step": 665
},
{
"epoch": 0.69,
"grad_norm": 0.11704660952091217,
"learning_rate": 6.590548331019957e-05,
"loss": 1.3214,
"step": 666
},
{
"epoch": 0.69,
"grad_norm": 0.11636264622211456,
"learning_rate": 6.55187015308969e-05,
"loss": 1.3366,
"step": 667
},
{
"epoch": 0.69,
"grad_norm": 0.10623035579919815,
"learning_rate": 6.513250431775004e-05,
"loss": 1.1485,
"step": 668
},
{
"epoch": 0.69,
"grad_norm": 0.10182057321071625,
"learning_rate": 6.474689821801295e-05,
"loss": 1.1521,
"step": 669
},
{
"epoch": 0.7,
"grad_norm": 0.11016492545604706,
"learning_rate": 6.436188976891846e-05,
"loss": 1.2067,
"step": 670
},
{
"epoch": 0.7,
"grad_norm": 0.11019518971443176,
"learning_rate": 6.39774854975673e-05,
"loss": 1.2559,
"step": 671
},
{
"epoch": 0.7,
"grad_norm": 0.11215123534202576,
"learning_rate": 6.359369192081756e-05,
"loss": 1.2557,
"step": 672
},
{
"epoch": 0.7,
"grad_norm": 0.1162608414888382,
"learning_rate": 6.321051554517407e-05,
"loss": 1.1967,
"step": 673
},
{
"epoch": 0.7,
"grad_norm": 0.1348588615655899,
"learning_rate": 6.282796286667814e-05,
"loss": 1.1944,
"step": 674
},
{
"epoch": 0.7,
"grad_norm": 0.12534910440444946,
"learning_rate": 6.244604037079753e-05,
"loss": 1.1354,
"step": 675
},
{
"epoch": 0.7,
"grad_norm": 0.11349747329950333,
"learning_rate": 6.206475453231644e-05,
"loss": 1.3001,
"step": 676
},
{
"epoch": 0.7,
"grad_norm": 0.11473750323057175,
"learning_rate": 6.168411181522569e-05,
"loss": 1.3084,
"step": 677
},
{
"epoch": 0.7,
"grad_norm": 0.11263560503721237,
"learning_rate": 6.13041186726131e-05,
"loss": 1.0361,
"step": 678
},
{
"epoch": 0.7,
"grad_norm": 0.11515972018241882,
"learning_rate": 6.092478154655431e-05,
"loss": 1.1342,
"step": 679
},
{
"epoch": 0.71,
"grad_norm": 0.10681987553834915,
"learning_rate": 6.0546106868003324e-05,
"loss": 1.2221,
"step": 680
},
{
"epoch": 0.71,
"grad_norm": 0.10119231045246124,
"learning_rate": 6.016810105668365e-05,
"loss": 1.1237,
"step": 681
},
{
"epoch": 0.71,
"grad_norm": 0.10480265319347382,
"learning_rate": 5.979077052097936e-05,
"loss": 1.2506,
"step": 682
},
{
"epoch": 0.71,
"grad_norm": 0.11073455959558487,
"learning_rate": 5.941412165782645e-05,
"loss": 1.0603,
"step": 683
},
{
"epoch": 0.71,
"grad_norm": 0.11075324565172195,
"learning_rate": 5.9038160852604476e-05,
"loss": 1.1351,
"step": 684
},
{
"epoch": 0.71,
"grad_norm": 0.10219030827283859,
"learning_rate": 5.866289447902828e-05,
"loss": 1.2163,
"step": 685
},
{
"epoch": 0.71,
"grad_norm": 0.10446413606405258,
"learning_rate": 5.8288328899039834e-05,
"loss": 1.2578,
"step": 686
},
{
"epoch": 0.71,
"grad_norm": 0.10746556520462036,
"learning_rate": 5.791447046270056e-05,
"loss": 1.2526,
"step": 687
},
{
"epoch": 0.71,
"grad_norm": 0.12443575263023376,
"learning_rate": 5.754132550808344e-05,
"loss": 1.3376,
"step": 688
},
{
"epoch": 0.72,
"grad_norm": 0.12434745579957962,
"learning_rate": 5.7168900361165824e-05,
"loss": 1.369,
"step": 689
},
{
"epoch": 0.72,
"grad_norm": 0.11743592470884323,
"learning_rate": 5.679720133572206e-05,
"loss": 1.2044,
"step": 690
},
{
"epoch": 0.72,
"grad_norm": 0.12590321898460388,
"learning_rate": 5.6426234733216376e-05,
"loss": 1.0729,
"step": 691
},
{
"epoch": 0.72,
"grad_norm": 0.10635906457901001,
"learning_rate": 5.605600684269614e-05,
"loss": 1.2448,
"step": 692
},
{
"epoch": 0.72,
"grad_norm": 0.10504303872585297,
"learning_rate": 5.568652394068532e-05,
"loss": 1.0663,
"step": 693
},
{
"epoch": 0.72,
"grad_norm": 0.10200861841440201,
"learning_rate": 5.5317792291077965e-05,
"loss": 1.0939,
"step": 694
},
{
"epoch": 0.72,
"grad_norm": 0.11620642244815826,
"learning_rate": 5.4949818145031986e-05,
"loss": 1.1002,
"step": 695
},
{
"epoch": 0.72,
"grad_norm": 0.10879629105329514,
"learning_rate": 5.4582607740863324e-05,
"loss": 1.1695,
"step": 696
},
{
"epoch": 0.72,
"grad_norm": 0.11083561182022095,
"learning_rate": 5.421616730394e-05,
"loss": 1.2967,
"step": 697
},
{
"epoch": 0.72,
"grad_norm": 0.11363320797681808,
"learning_rate": 5.385050304657669e-05,
"loss": 1.0099,
"step": 698
},
{
"epoch": 0.73,
"grad_norm": 0.10837159305810928,
"learning_rate": 5.348562116792946e-05,
"loss": 1.1575,
"step": 699
},
{
"epoch": 0.73,
"grad_norm": 0.7101715207099915,
"learning_rate": 5.312152785389055e-05,
"loss": 1.3229,
"step": 700
},
{
"epoch": 0.73,
"grad_norm": 0.11040350794792175,
"learning_rate": 5.2758229276983614e-05,
"loss": 1.183,
"step": 701
},
{
"epoch": 0.73,
"grad_norm": 0.10885446518659592,
"learning_rate": 5.2395731596258925e-05,
"loss": 1.139,
"step": 702
},
{
"epoch": 0.73,
"grad_norm": 0.10934137552976608,
"learning_rate": 5.20340409571891e-05,
"loss": 1.291,
"step": 703
},
{
"epoch": 0.73,
"grad_norm": 0.11229011416435242,
"learning_rate": 5.167316349156495e-05,
"loss": 1.1406,
"step": 704
},
{
"epoch": 0.73,
"grad_norm": 0.12293226271867752,
"learning_rate": 5.1313105317391284e-05,
"loss": 1.1102,
"step": 705
},
{
"epoch": 0.73,
"grad_norm": 0.10751207172870636,
"learning_rate": 5.095387253878345e-05,
"loss": 1.1296,
"step": 706
},
{
"epoch": 0.73,
"grad_norm": 0.10970565676689148,
"learning_rate": 5.059547124586375e-05,
"loss": 1.3482,
"step": 707
},
{
"epoch": 0.73,
"grad_norm": 0.10699622333049774,
"learning_rate": 5.0237907514658176e-05,
"loss": 1.3642,
"step": 708
},
{
"epoch": 0.74,
"grad_norm": 0.1132645532488823,
"learning_rate": 4.988118740699346e-05,
"loss": 1.1504,
"step": 709
},
{
"epoch": 0.74,
"grad_norm": 0.12392119318246841,
"learning_rate": 4.9525316970394245e-05,
"loss": 1.3906,
"step": 710
},
{
"epoch": 0.74,
"grad_norm": 0.10416343808174133,
"learning_rate": 4.917030223798057e-05,
"loss": 1.3419,
"step": 711
},
{
"epoch": 0.74,
"grad_norm": 0.11010722070932388,
"learning_rate": 4.881614922836555e-05,
"loss": 1.1175,
"step": 712
},
{
"epoch": 0.74,
"grad_norm": 0.11078814417123795,
"learning_rate": 4.8462863945553516e-05,
"loss": 1.3513,
"step": 713
},
{
"epoch": 0.74,
"grad_norm": 0.12069610506296158,
"learning_rate": 4.8110452378838035e-05,
"loss": 1.3913,
"step": 714
},
{
"epoch": 0.74,
"grad_norm": 0.11017792671918869,
"learning_rate": 4.775892050270051e-05,
"loss": 1.2217,
"step": 715
},
{
"epoch": 0.74,
"grad_norm": 0.10235986858606339,
"learning_rate": 4.740827427670871e-05,
"loss": 1.1375,
"step": 716
},
{
"epoch": 0.74,
"grad_norm": 0.11740509420633316,
"learning_rate": 4.7058519645416e-05,
"loss": 1.1649,
"step": 717
},
{
"epoch": 0.75,
"grad_norm": 0.16506601870059967,
"learning_rate": 4.6709662538260267e-05,
"loss": 1.0392,
"step": 718
},
{
"epoch": 0.75,
"grad_norm": 0.22638647258281708,
"learning_rate": 4.63617088694637e-05,
"loss": 1.21,
"step": 719
},
{
"epoch": 0.75,
"grad_norm": 0.11359802633523941,
"learning_rate": 4.601466453793224e-05,
"loss": 1.2481,
"step": 720
},
{
"epoch": 0.75,
"grad_norm": 0.10808828473091125,
"learning_rate": 4.566853542715581e-05,
"loss": 1.2008,
"step": 721
},
{
"epoch": 0.75,
"grad_norm": 0.09867287427186966,
"learning_rate": 4.5323327405108416e-05,
"loss": 1.0815,
"step": 722
},
{
"epoch": 0.75,
"grad_norm": 0.11430750787258148,
"learning_rate": 4.4979046324148785e-05,
"loss": 1.3082,
"step": 723
},
{
"epoch": 0.75,
"grad_norm": 0.1082565113902092,
"learning_rate": 4.4635698020921014e-05,
"loss": 1.0386,
"step": 724
},
{
"epoch": 0.75,
"grad_norm": 0.11467121541500092,
"learning_rate": 4.4293288316255653e-05,
"loss": 1.299,
"step": 725
},
{
"epoch": 0.75,
"grad_norm": 0.10791204124689102,
"learning_rate": 4.395182301507118e-05,
"loss": 1.2306,
"step": 726
},
{
"epoch": 0.75,
"grad_norm": 0.11149678379297256,
"learning_rate": 4.3611307906275414e-05,
"loss": 1.3074,
"step": 727
},
{
"epoch": 0.76,
"grad_norm": 0.11212314665317535,
"learning_rate": 4.327174876266743e-05,
"loss": 1.1512,
"step": 728
},
{
"epoch": 0.76,
"grad_norm": 0.1070462167263031,
"learning_rate": 4.293315134083975e-05,
"loss": 1.2378,
"step": 729
},
{
"epoch": 0.76,
"grad_norm": 0.11427627503871918,
"learning_rate": 4.2595521381080604e-05,
"loss": 1.2729,
"step": 730
},
{
"epoch": 0.76,
"grad_norm": 0.1135973185300827,
"learning_rate": 4.2258864607276705e-05,
"loss": 1.1399,
"step": 731
},
{
"epoch": 0.76,
"grad_norm": 0.11445244401693344,
"learning_rate": 4.192318672681631e-05,
"loss": 1.1549,
"step": 732
},
{
"epoch": 0.76,
"grad_norm": 0.10163676738739014,
"learning_rate": 4.158849343049232e-05,
"loss": 1.2751,
"step": 733
},
{
"epoch": 0.76,
"grad_norm": 0.11175056546926498,
"learning_rate": 4.1254790392405797e-05,
"loss": 1.2371,
"step": 734
},
{
"epoch": 0.76,
"grad_norm": 0.11040692776441574,
"learning_rate": 4.092208326986986e-05,
"loss": 1.3218,
"step": 735
},
{
"epoch": 0.76,
"grad_norm": 0.10620775073766708,
"learning_rate": 4.0590377703313785e-05,
"loss": 1.1667,
"step": 736
},
{
"epoch": 0.76,
"grad_norm": 0.10938204824924469,
"learning_rate": 4.0259679316187214e-05,
"loss": 1.0897,
"step": 737
},
{
"epoch": 0.77,
"grad_norm": 0.10027039796113968,
"learning_rate": 3.992999371486508e-05,
"loss": 1.1646,
"step": 738
},
{
"epoch": 0.77,
"grad_norm": 0.13585035502910614,
"learning_rate": 3.960132648855226e-05,
"loss": 1.1759,
"step": 739
},
{
"epoch": 0.77,
"grad_norm": 0.12253659218549728,
"learning_rate": 3.9273683209189115e-05,
"loss": 1.3585,
"step": 740
},
{
"epoch": 0.77,
"grad_norm": 0.14041317999362946,
"learning_rate": 3.8947069431356855e-05,
"loss": 1.2736,
"step": 741
},
{
"epoch": 0.77,
"grad_norm": 0.10640061646699905,
"learning_rate": 3.862149069218343e-05,
"loss": 1.2859,
"step": 742
},
{
"epoch": 0.77,
"grad_norm": 0.10962343961000443,
"learning_rate": 3.829695251124953e-05,
"loss": 1.2581,
"step": 743
},
{
"epoch": 0.77,
"grad_norm": 0.12311571091413498,
"learning_rate": 3.797346039049528e-05,
"loss": 1.3115,
"step": 744
},
{
"epoch": 0.77,
"grad_norm": 0.12078052759170532,
"learning_rate": 3.7651019814126654e-05,
"loss": 1.03,
"step": 745
},
{
"epoch": 0.77,
"grad_norm": 0.10591412335634232,
"learning_rate": 3.732963624852275e-05,
"loss": 1.0906,
"step": 746
},
{
"epoch": 0.78,
"grad_norm": 0.1085178554058075,
"learning_rate": 3.700931514214303e-05,
"loss": 1.1478,
"step": 747
},
{
"epoch": 0.78,
"grad_norm": 0.10265811532735825,
"learning_rate": 3.6690061925434814e-05,
"loss": 1.1375,
"step": 748
},
{
"epoch": 0.78,
"grad_norm": 0.10623796284198761,
"learning_rate": 3.637188201074149e-05,
"loss": 1.2655,
"step": 749
},
{
"epoch": 0.78,
"grad_norm": 0.1077788770198822,
"learning_rate": 3.605478079221054e-05,
"loss": 1.2805,
"step": 750
},
{
"epoch": 0.78,
"grad_norm": 0.1078350618481636,
"learning_rate": 3.573876364570214e-05,
"loss": 1.356,
"step": 751
},
{
"epoch": 0.78,
"grad_norm": 0.10767360776662827,
"learning_rate": 3.5423835928698125e-05,
"loss": 1.2346,
"step": 752
},
{
"epoch": 0.78,
"grad_norm": 0.10546786338090897,
"learning_rate": 3.5110002980210975e-05,
"loss": 1.2541,
"step": 753
},
{
"epoch": 0.78,
"grad_norm": 0.09910236299037933,
"learning_rate": 3.479727012069349e-05,
"loss": 1.2015,
"step": 754
},
{
"epoch": 0.78,
"grad_norm": 0.10072822123765945,
"learning_rate": 3.4485642651948516e-05,
"loss": 1.3005,
"step": 755
},
{
"epoch": 0.78,
"grad_norm": 0.11063572019338608,
"learning_rate": 3.417512585703903e-05,
"loss": 1.0739,
"step": 756
},
{
"epoch": 0.79,
"grad_norm": 0.09937085956335068,
"learning_rate": 3.38657250001986e-05,
"loss": 1.3106,
"step": 757
},
{
"epoch": 0.79,
"grad_norm": 0.10743856430053711,
"learning_rate": 3.355744532674211e-05,
"loss": 1.1492,
"step": 758
},
{
"epoch": 0.79,
"grad_norm": 0.11000777781009674,
"learning_rate": 3.325029206297694e-05,
"loss": 1.1876,
"step": 759
},
{
"epoch": 0.79,
"grad_norm": 0.11820725351572037,
"learning_rate": 3.294427041611425e-05,
"loss": 1.2457,
"step": 760
},
{
"epoch": 0.79,
"grad_norm": 0.10308333486318588,
"learning_rate": 3.263938557418082e-05,
"loss": 1.2801,
"step": 761
},
{
"epoch": 0.79,
"grad_norm": 0.1042330265045166,
"learning_rate": 3.233564270593088e-05,
"loss": 1.2237,
"step": 762
},
{
"epoch": 0.79,
"grad_norm": 0.10803329199552536,
"learning_rate": 3.2033046960758764e-05,
"loss": 1.2841,
"step": 763
},
{
"epoch": 0.79,
"grad_norm": 0.11426287144422531,
"learning_rate": 3.1731603468611337e-05,
"loss": 1.355,
"step": 764
},
{
"epoch": 0.79,
"grad_norm": 0.10273044556379318,
"learning_rate": 3.143131733990127e-05,
"loss": 1.2975,
"step": 765
},
{
"epoch": 0.8,
"grad_norm": 0.11743076890707016,
"learning_rate": 3.113219366542031e-05,
"loss": 1.2532,
"step": 766
},
{
"epoch": 0.8,
"grad_norm": 0.1157737672328949,
"learning_rate": 3.083423751625282e-05,
"loss": 1.4476,
"step": 767
},
{
"epoch": 0.8,
"grad_norm": 0.10297975689172745,
"learning_rate": 3.0537453943690074e-05,
"loss": 1.1543,
"step": 768
},
{
"epoch": 0.8,
"grad_norm": 0.10176219791173935,
"learning_rate": 3.0241847979144487e-05,
"loss": 1.0901,
"step": 769
},
{
"epoch": 0.8,
"grad_norm": 0.11567193269729614,
"learning_rate": 2.994742463406427e-05,
"loss": 1.25,
"step": 770
},
{
"epoch": 0.8,
"grad_norm": 0.12137153744697571,
"learning_rate": 2.96541888998485e-05,
"loss": 1.3956,
"step": 771
},
{
"epoch": 0.8,
"grad_norm": 0.10772846639156342,
"learning_rate": 2.9362145747762625e-05,
"loss": 1.4177,
"step": 772
},
{
"epoch": 0.8,
"grad_norm": 0.10832470655441284,
"learning_rate": 2.9071300128854007e-05,
"loss": 1.1614,
"step": 773
},
{
"epoch": 0.8,
"grad_norm": 0.09898839145898819,
"learning_rate": 2.878165697386812e-05,
"loss": 1.2994,
"step": 774
},
{
"epoch": 0.8,
"grad_norm": 0.1113181784749031,
"learning_rate": 2.849322119316489e-05,
"loss": 1.1914,
"step": 775
},
{
"epoch": 0.81,
"grad_norm": 0.10062463581562042,
"learning_rate": 2.82059976766354e-05,
"loss": 1.1165,
"step": 776
},
{
"epoch": 0.81,
"grad_norm": 0.10819394886493683,
"learning_rate": 2.7919991293619098e-05,
"loss": 1.283,
"step": 777
},
{
"epoch": 0.81,
"grad_norm": 0.1365766078233719,
"learning_rate": 2.763520689282122e-05,
"loss": 1.1963,
"step": 778
},
{
"epoch": 0.81,
"grad_norm": 0.11200159043073654,
"learning_rate": 2.7351649302230553e-05,
"loss": 1.1885,
"step": 779
},
{
"epoch": 0.81,
"grad_norm": 0.11218269914388657,
"learning_rate": 2.7069323329037634e-05,
"loss": 1.2944,
"step": 780
},
{
"epoch": 0.81,
"grad_norm": 0.12175281345844269,
"learning_rate": 2.678823375955314e-05,
"loss": 1.2379,
"step": 781
},
{
"epoch": 0.81,
"grad_norm": 0.125010147690773,
"learning_rate": 2.650838535912692e-05,
"loss": 1.2309,
"step": 782
},
{
"epoch": 0.81,
"grad_norm": 0.12338108569383621,
"learning_rate": 2.6229782872067045e-05,
"loss": 1.1489,
"step": 783
},
{
"epoch": 0.81,
"grad_norm": 0.11083382368087769,
"learning_rate": 2.595243102155951e-05,
"loss": 1.1595,
"step": 784
},
{
"epoch": 0.81,
"grad_norm": 0.11741047352552414,
"learning_rate": 2.5676334509588008e-05,
"loss": 1.2438,
"step": 785
},
{
"epoch": 0.82,
"grad_norm": 0.10495740175247192,
"learning_rate": 2.540149801685441e-05,
"loss": 1.299,
"step": 786
},
{
"epoch": 0.82,
"grad_norm": 0.11125294119119644,
"learning_rate": 2.5127926202699238e-05,
"loss": 1.1244,
"step": 787
},
{
"epoch": 0.82,
"grad_norm": 0.10036681592464447,
"learning_rate": 2.485562370502279e-05,
"loss": 1.2892,
"step": 788
},
{
"epoch": 0.82,
"grad_norm": 0.11835762113332748,
"learning_rate": 2.4584595140206458e-05,
"loss": 1.1257,
"step": 789
},
{
"epoch": 0.82,
"grad_norm": 0.11104938387870789,
"learning_rate": 2.4314845103034456e-05,
"loss": 1.1601,
"step": 790
},
{
"epoch": 0.82,
"grad_norm": 0.10659179091453552,
"learning_rate": 2.40463781666159e-05,
"loss": 1.2066,
"step": 791
},
{
"epoch": 0.82,
"grad_norm": 0.10941512882709503,
"learning_rate": 2.377919888230744e-05,
"loss": 1.3887,
"step": 792
},
{
"epoch": 0.82,
"grad_norm": 0.18816524744033813,
"learning_rate": 2.3513311779635905e-05,
"loss": 1.2191,
"step": 793
},
{
"epoch": 0.82,
"grad_norm": 0.1119375005364418,
"learning_rate": 2.3248721366221638e-05,
"loss": 1.2773,
"step": 794
},
{
"epoch": 0.83,
"grad_norm": 0.10966973006725311,
"learning_rate": 2.2985432127701946e-05,
"loss": 1.2283,
"step": 795
},
{
"epoch": 0.83,
"grad_norm": 0.10751962661743164,
"learning_rate": 2.2723448527655266e-05,
"loss": 1.375,
"step": 796
},
{
"epoch": 0.83,
"grad_norm": 0.19195051491260529,
"learning_rate": 2.246277500752524e-05,
"loss": 1.3717,
"step": 797
},
{
"epoch": 0.83,
"grad_norm": 0.4861750900745392,
"learning_rate": 2.220341598654565e-05,
"loss": 1.1786,
"step": 798
},
{
"epoch": 0.83,
"grad_norm": 0.10594101250171661,
"learning_rate": 2.194537586166532e-05,
"loss": 1.2528,
"step": 799
},
{
"epoch": 0.83,
"grad_norm": 0.11542479693889618,
"learning_rate": 2.16886590074737e-05,
"loss": 1.2277,
"step": 800
},
{
"epoch": 0.83,
"grad_norm": 0.10653986781835556,
"learning_rate": 2.143326977612662e-05,
"loss": 1.1785,
"step": 801
},
{
"epoch": 0.83,
"grad_norm": 0.11760350316762924,
"learning_rate": 2.1179212497272584e-05,
"loss": 1.2573,
"step": 802
},
{
"epoch": 0.83,
"grad_norm": 0.1091405376791954,
"learning_rate": 2.092649147797927e-05,
"loss": 1.0923,
"step": 803
},
{
"epoch": 0.83,
"grad_norm": 0.10867162048816681,
"learning_rate": 2.0675111002660562e-05,
"loss": 1.2358,
"step": 804
},
{
"epoch": 0.84,
"grad_norm": 0.10017253458499908,
"learning_rate": 2.0425075333003953e-05,
"loss": 1.2741,
"step": 805
},
{
"epoch": 0.84,
"grad_norm": 0.11259835958480835,
"learning_rate": 2.0176388707898274e-05,
"loss": 1.0486,
"step": 806
},
{
"epoch": 0.84,
"grad_norm": 0.1280423402786255,
"learning_rate": 1.992905534336178e-05,
"loss": 1.3091,
"step": 807
},
{
"epoch": 0.84,
"grad_norm": 0.10603731125593185,
"learning_rate": 1.9683079432470775e-05,
"loss": 1.2541,
"step": 808
},
{
"epoch": 0.84,
"grad_norm": 0.10933735966682434,
"learning_rate": 1.9438465145288374e-05,
"loss": 1.2186,
"step": 809
},
{
"epoch": 0.84,
"grad_norm": 0.11166153848171234,
"learning_rate": 1.9195216628793956e-05,
"loss": 1.2563,
"step": 810
},
{
"epoch": 0.84,
"grad_norm": 0.11184462904930115,
"learning_rate": 1.8953338006812805e-05,
"loss": 1.3052,
"step": 811
},
{
"epoch": 0.84,
"grad_norm": 0.11681969463825226,
"learning_rate": 1.8712833379946215e-05,
"loss": 1.1702,
"step": 812
},
{
"epoch": 0.84,
"grad_norm": 0.11982112377882004,
"learning_rate": 1.847370682550187e-05,
"loss": 1.3025,
"step": 813
},
{
"epoch": 0.84,
"grad_norm": 0.11721136420965195,
"learning_rate": 1.82359623974249e-05,
"loss": 1.1182,
"step": 814
},
{
"epoch": 0.85,
"grad_norm": 0.11441758275032043,
"learning_rate": 1.799960412622904e-05,
"loss": 1.1366,
"step": 815
},
{
"epoch": 0.85,
"grad_norm": 0.11492711305618286,
"learning_rate": 1.776463601892825e-05,
"loss": 1.1784,
"step": 816
},
{
"epoch": 0.85,
"grad_norm": 0.11135811358690262,
"learning_rate": 1.753106205896895e-05,
"loss": 1.2411,
"step": 817
},
{
"epoch": 0.85,
"grad_norm": 0.0985221266746521,
"learning_rate": 1.729888620616228e-05,
"loss": 1.283,
"step": 818
},
{
"epoch": 0.85,
"grad_norm": 0.10769841074943542,
"learning_rate": 1.7068112396617163e-05,
"loss": 1.2442,
"step": 819
},
{
"epoch": 0.85,
"grad_norm": 0.09943833947181702,
"learning_rate": 1.683874454267349e-05,
"loss": 1.1121,
"step": 820
},
{
"epoch": 0.85,
"grad_norm": 0.11267413944005966,
"learning_rate": 1.6610786532835774e-05,
"loss": 1.2912,
"step": 821
},
{
"epoch": 0.85,
"grad_norm": 0.10905832797288895,
"learning_rate": 1.6384242231707204e-05,
"loss": 1.1264,
"step": 822
},
{
"epoch": 0.85,
"grad_norm": 0.11152827739715576,
"learning_rate": 1.6159115479924257e-05,
"loss": 1.0625,
"step": 823
},
{
"epoch": 0.86,
"grad_norm": 0.10507339239120483,
"learning_rate": 1.593541009409143e-05,
"loss": 1.282,
"step": 824
},
{
"epoch": 0.86,
"grad_norm": 0.11833300441503525,
"learning_rate": 1.5713129866716648e-05,
"loss": 1.0683,
"step": 825
},
{
"epoch": 0.86,
"grad_norm": 0.10734719038009644,
"learning_rate": 1.5492278566146946e-05,
"loss": 1.2312,
"step": 826
},
{
"epoch": 0.86,
"grad_norm": 0.09920140355825424,
"learning_rate": 1.5272859936504514e-05,
"loss": 1.3253,
"step": 827
},
{
"epoch": 0.86,
"grad_norm": 0.10309410840272903,
"learning_rate": 1.5054877697623304e-05,
"loss": 1.3389,
"step": 828
},
{
"epoch": 0.86,
"grad_norm": 0.09966228157281876,
"learning_rate": 1.4838335544985981e-05,
"loss": 1.1592,
"step": 829
},
{
"epoch": 0.86,
"grad_norm": 0.10500723123550415,
"learning_rate": 1.462323714966114e-05,
"loss": 1.2318,
"step": 830
},
{
"epoch": 0.86,
"grad_norm": 0.10912671685218811,
"learning_rate": 1.4409586158241272e-05,
"loss": 1.2195,
"step": 831
},
{
"epoch": 0.86,
"grad_norm": 0.1132998913526535,
"learning_rate": 1.4197386192780715e-05,
"loss": 1.2476,
"step": 832
},
{
"epoch": 0.86,
"grad_norm": 0.11414579302072525,
"learning_rate": 1.3986640850734444e-05,
"loss": 1.229,
"step": 833
},
{
"epoch": 0.87,
"grad_norm": 0.10963235795497894,
"learning_rate": 1.3777353704897e-05,
"loss": 1.1468,
"step": 834
},
{
"epoch": 0.87,
"grad_norm": 0.11412128806114197,
"learning_rate": 1.3569528303341927e-05,
"loss": 1.141,
"step": 835
},
{
"epoch": 0.87,
"grad_norm": 0.11492784321308136,
"learning_rate": 1.3363168169361573e-05,
"loss": 1.0967,
"step": 836
},
{
"epoch": 0.87,
"grad_norm": 0.11600396037101746,
"learning_rate": 1.3158276801407431e-05,
"loss": 1.2414,
"step": 837
},
{
"epoch": 0.87,
"grad_norm": 0.11931241303682327,
"learning_rate": 1.2954857673030807e-05,
"loss": 1.2318,
"step": 838
},
{
"epoch": 0.87,
"grad_norm": 0.10490193963050842,
"learning_rate": 1.2752914232823942e-05,
"loss": 1.2481,
"step": 839
},
{
"epoch": 0.87,
"grad_norm": 0.11982220411300659,
"learning_rate": 1.2552449904361529e-05,
"loss": 1.0291,
"step": 840
},
{
"epoch": 0.87,
"grad_norm": 0.0949595496058464,
"learning_rate": 1.235346808614264e-05,
"loss": 1.0835,
"step": 841
},
{
"epoch": 0.87,
"grad_norm": 0.11206669360399246,
"learning_rate": 1.2155972151533224e-05,
"loss": 1.2174,
"step": 842
},
{
"epoch": 0.87,
"grad_norm": 0.10925093293190002,
"learning_rate": 1.1959965448708733e-05,
"loss": 1.359,
"step": 843
},
{
"epoch": 0.88,
"grad_norm": 0.10597091913223267,
"learning_rate": 1.1765451300597573e-05,
"loss": 1.2567,
"step": 844
},
{
"epoch": 0.88,
"grad_norm": 0.12670493125915527,
"learning_rate": 1.1572433004824635e-05,
"loss": 1.1798,
"step": 845
},
{
"epoch": 0.88,
"grad_norm": 0.10978806763887405,
"learning_rate": 1.1380913833655382e-05,
"loss": 1.2303,
"step": 846
},
{
"epoch": 0.88,
"grad_norm": 0.1855391412973404,
"learning_rate": 1.1190897033940461e-05,
"loss": 1.2306,
"step": 847
},
{
"epoch": 0.88,
"grad_norm": 0.11130934208631516,
"learning_rate": 1.1002385827060602e-05,
"loss": 1.373,
"step": 848
},
{
"epoch": 0.88,
"grad_norm": 0.10244546085596085,
"learning_rate": 1.0815383408871982e-05,
"loss": 1.3044,
"step": 849
},
{
"epoch": 0.88,
"grad_norm": 0.11533805727958679,
"learning_rate": 1.0629892949652132e-05,
"loss": 1.0704,
"step": 850
},
{
"epoch": 0.88,
"grad_norm": 0.12121516466140747,
"learning_rate": 1.0445917594046072e-05,
"loss": 1.265,
"step": 851
},
{
"epoch": 0.88,
"grad_norm": 0.1047421246767044,
"learning_rate": 1.026346046101312e-05,
"loss": 1.1294,
"step": 852
},
{
"epoch": 0.89,
"grad_norm": 0.10979347676038742,
"learning_rate": 1.0082524643773916e-05,
"loss": 0.9608,
"step": 853
},
{
"epoch": 0.89,
"grad_norm": 0.12287425249814987,
"learning_rate": 9.903113209758096e-06,
"loss": 0.8795,
"step": 854
},
{
"epoch": 0.89,
"grad_norm": 0.11486369371414185,
"learning_rate": 9.725229200552122e-06,
"loss": 1.3881,
"step": 855
},
{
"epoch": 0.89,
"grad_norm": 0.11395483464002609,
"learning_rate": 9.548875631847875e-06,
"loss": 1.1552,
"step": 856
},
{
"epoch": 0.89,
"grad_norm": 0.10814516991376877,
"learning_rate": 9.374055493391455e-06,
"loss": 1.2155,
"step": 857
},
{
"epoch": 0.89,
"grad_norm": 0.10679764300584793,
"learning_rate": 9.200771748932513e-06,
"loss": 1.1449,
"step": 858
},
{
"epoch": 0.89,
"grad_norm": 0.1089312732219696,
"learning_rate": 9.029027336174022e-06,
"loss": 1.2221,
"step": 859
},
{
"epoch": 0.89,
"grad_norm": 0.11471953988075256,
"learning_rate": 8.858825166722417e-06,
"loss": 0.9955,
"step": 860
},
{
"epoch": 0.89,
"grad_norm": 0.11357836425304413,
"learning_rate": 8.6901681260383e-06,
"loss": 1.0855,
"step": 861
},
{
"epoch": 0.89,
"grad_norm": 0.11274585127830505,
"learning_rate": 8.523059073387474e-06,
"loss": 1.1812,
"step": 862
},
{
"epoch": 0.9,
"grad_norm": 0.10156439989805222,
"learning_rate": 8.35750084179251e-06,
"loss": 1.1702,
"step": 863
},
{
"epoch": 0.9,
"grad_norm": 0.10595791786909103,
"learning_rate": 8.193496237984677e-06,
"loss": 1.1719,
"step": 864
},
{
"epoch": 0.9,
"grad_norm": 0.10934492200613022,
"learning_rate": 8.031048042356392e-06,
"loss": 1.1735,
"step": 865
},
{
"epoch": 0.9,
"grad_norm": 0.10104786604642868,
"learning_rate": 7.870159008914069e-06,
"loss": 1.1967,
"step": 866
},
{
"epoch": 0.9,
"grad_norm": 0.10521814227104187,
"learning_rate": 7.710831865231461e-06,
"loss": 1.3094,
"step": 867
},
{
"epoch": 0.9,
"grad_norm": 0.1166355162858963,
"learning_rate": 7.553069312403349e-06,
"loss": 1.2912,
"step": 868
},
{
"epoch": 0.9,
"grad_norm": 0.09746871888637543,
"learning_rate": 7.3968740249998115e-06,
"loss": 1.1133,
"step": 869
},
{
"epoch": 0.9,
"grad_norm": 0.10336558520793915,
"learning_rate": 7.242248651020844e-06,
"loss": 1.1893,
"step": 870
},
{
"epoch": 0.9,
"grad_norm": 0.11591080576181412,
"learning_rate": 7.089195811851501e-06,
"loss": 1.1395,
"step": 871
},
{
"epoch": 0.91,
"grad_norm": 0.22089900076389313,
"learning_rate": 6.937718102217461e-06,
"loss": 1.1993,
"step": 872
},
{
"epoch": 0.91,
"grad_norm": 0.108499675989151,
"learning_rate": 6.787818090140985e-06,
"loss": 1.1195,
"step": 873
},
{
"epoch": 0.91,
"grad_norm": 0.10180932283401489,
"learning_rate": 6.639498316897419e-06,
"loss": 1.3188,
"step": 874
},
{
"epoch": 0.91,
"grad_norm": 0.10950213670730591,
"learning_rate": 6.492761296972116e-06,
"loss": 1.1373,
"step": 875
},
{
"epoch": 0.91,
"grad_norm": 0.09740964323282242,
"learning_rate": 6.347609518017761e-06,
"loss": 1.1551,
"step": 876
},
{
"epoch": 0.91,
"grad_norm": 0.10599476099014282,
"learning_rate": 6.2040454408123e-06,
"loss": 1.2161,
"step": 877
},
{
"epoch": 0.91,
"grad_norm": 0.11540497094392776,
"learning_rate": 6.062071499217081e-06,
"loss": 1.256,
"step": 878
},
{
"epoch": 0.91,
"grad_norm": 0.10363679379224777,
"learning_rate": 5.921690100135713e-06,
"loss": 1.2462,
"step": 879
},
{
"epoch": 0.91,
"grad_norm": 0.10643389075994492,
"learning_rate": 5.782903623473201e-06,
"loss": 1.34,
"step": 880
},
{
"epoch": 0.91,
"grad_norm": 0.14184726774692535,
"learning_rate": 5.645714422095627e-06,
"loss": 1.0978,
"step": 881
},
{
"epoch": 0.92,
"grad_norm": 0.11375614255666733,
"learning_rate": 5.510124821790208e-06,
"loss": 1.2334,
"step": 882
},
{
"epoch": 0.92,
"grad_norm": 0.11842206120491028,
"learning_rate": 5.376137121225933e-06,
"loss": 1.2043,
"step": 883
},
{
"epoch": 0.92,
"grad_norm": 0.1027233749628067,
"learning_rate": 5.24375359191458e-06,
"loss": 1.2038,
"step": 884
},
{
"epoch": 0.92,
"grad_norm": 0.09779934585094452,
"learning_rate": 5.11297647817216e-06,
"loss": 1.2799,
"step": 885
},
{
"epoch": 0.92,
"grad_norm": 0.28459811210632324,
"learning_rate": 4.983807997080925e-06,
"loss": 1.2972,
"step": 886
},
{
"epoch": 0.92,
"grad_norm": 0.12350057065486908,
"learning_rate": 4.856250338451762e-06,
"loss": 1.1445,
"step": 887
},
{
"epoch": 0.92,
"grad_norm": 0.10094188153743744,
"learning_rate": 4.73030566478706e-06,
"loss": 1.0144,
"step": 888
},
{
"epoch": 0.92,
"grad_norm": 0.10270640254020691,
"learning_rate": 4.605976111244015e-06,
"loss": 1.1673,
"step": 889
},
{
"epoch": 0.92,
"grad_norm": 0.12842753529548645,
"learning_rate": 4.483263785598524e-06,
"loss": 1.1681,
"step": 890
},
{
"epoch": 0.92,
"grad_norm": 0.10717935115098953,
"learning_rate": 4.362170768209406e-06,
"loss": 1.2661,
"step": 891
},
{
"epoch": 0.93,
"grad_norm": 0.20119836926460266,
"learning_rate": 4.24269911198305e-06,
"loss": 1.1492,
"step": 892
},
{
"epoch": 0.93,
"grad_norm": 0.11531922966241837,
"learning_rate": 4.124850842338779e-06,
"loss": 1.2629,
"step": 893
},
{
"epoch": 0.93,
"grad_norm": 0.10086618363857269,
"learning_rate": 4.008627957174372e-06,
"loss": 1.1223,
"step": 894
},
{
"epoch": 0.93,
"grad_norm": 0.10228842496871948,
"learning_rate": 3.8940324268322285e-06,
"loss": 1.2285,
"step": 895
},
{
"epoch": 0.93,
"grad_norm": 0.1152658760547638,
"learning_rate": 3.7810661940660517e-06,
"loss": 1.3102,
"step": 896
},
{
"epoch": 0.93,
"grad_norm": 0.09905364364385605,
"learning_rate": 3.66973117400774e-06,
"loss": 1.2029,
"step": 897
},
{
"epoch": 0.93,
"grad_norm": 0.11741360276937485,
"learning_rate": 3.5600292541351e-06,
"loss": 1.0504,
"step": 898
},
{
"epoch": 0.93,
"grad_norm": 0.10663048177957535,
"learning_rate": 3.451962294239741e-06,
"loss": 1.0434,
"step": 899
},
{
"epoch": 0.93,
"grad_norm": 0.13997888565063477,
"learning_rate": 3.3455321263955786e-06,
"loss": 1.1909,
"step": 900
},
{
"epoch": 0.94,
"grad_norm": 0.10722199827432632,
"learning_rate": 3.240740554927768e-06,
"loss": 1.1066,
"step": 901
},
{
"epoch": 0.94,
"grad_norm": 0.10431613773107529,
"learning_rate": 3.1375893563820756e-06,
"loss": 1.1097,
"step": 902
},
{
"epoch": 0.94,
"grad_norm": 0.10055112838745117,
"learning_rate": 3.0360802794948686e-06,
"loss": 1.119,
"step": 903
},
{
"epoch": 0.94,
"grad_norm": 0.14287303388118744,
"learning_rate": 2.936215045163371e-06,
"loss": 1.1977,
"step": 904
},
{
"epoch": 0.94,
"grad_norm": 0.11945986747741699,
"learning_rate": 2.837995346416533e-06,
"loss": 1.0618,
"step": 905
},
{
"epoch": 0.94,
"grad_norm": 0.11288944631814957,
"learning_rate": 2.7414228483862657e-06,
"loss": 1.1562,
"step": 906
},
{
"epoch": 0.94,
"grad_norm": 0.11274345964193344,
"learning_rate": 2.6464991882793278e-06,
"loss": 1.2794,
"step": 907
},
{
"epoch": 0.94,
"grad_norm": 0.10257246345281601,
"learning_rate": 2.5532259753494827e-06,
"loss": 1.1981,
"step": 908
},
{
"epoch": 0.94,
"grad_norm": 0.11306274682283401,
"learning_rate": 2.4616047908702087e-06,
"loss": 1.2817,
"step": 909
},
{
"epoch": 0.94,
"grad_norm": 0.1063089668750763,
"learning_rate": 2.3716371881079557e-06,
"loss": 1.2429,
"step": 910
},
{
"epoch": 0.95,
"grad_norm": 0.11371003836393356,
"learning_rate": 2.2833246922957407e-06,
"loss": 1.2017,
"step": 911
},
{
"epoch": 0.95,
"grad_norm": 0.10964678227901459,
"learning_rate": 2.1966688006073823e-06,
"loss": 1.0438,
"step": 912
},
{
"epoch": 0.95,
"grad_norm": 0.10942777991294861,
"learning_rate": 2.111670982132008e-06,
"loss": 1.1411,
"step": 913
},
{
"epoch": 0.95,
"grad_norm": 0.20406045019626617,
"learning_rate": 2.028332677849254e-06,
"loss": 1.26,
"step": 914
},
{
"epoch": 0.95,
"grad_norm": 0.1080511286854744,
"learning_rate": 1.9466553006047383e-06,
"loss": 1.2078,
"step": 915
},
{
"epoch": 0.95,
"grad_norm": 0.10461273789405823,
"learning_rate": 1.86664023508617e-06,
"loss": 1.0991,
"step": 916
},
{
"epoch": 0.95,
"grad_norm": 0.14060625433921814,
"learning_rate": 1.7882888377998786e-06,
"loss": 1.1999,
"step": 917
},
{
"epoch": 0.95,
"grad_norm": 0.10702642798423767,
"learning_rate": 1.7116024370477878e-06,
"loss": 1.1649,
"step": 918
},
{
"epoch": 0.95,
"grad_norm": 0.12004446983337402,
"learning_rate": 1.6365823329049124e-06,
"loss": 1.1923,
"step": 919
},
{
"epoch": 0.95,
"grad_norm": 0.12491834163665771,
"learning_rate": 1.5632297971972965e-06,
"loss": 1.1329,
"step": 920
},
{
"epoch": 0.96,
"grad_norm": 0.11271212995052338,
"learning_rate": 1.4915460734805097e-06,
"loss": 1.1198,
"step": 921
},
{
"epoch": 0.96,
"grad_norm": 0.11066664755344391,
"learning_rate": 1.4215323770184641e-06,
"loss": 1.2727,
"step": 922
},
{
"epoch": 0.96,
"grad_norm": 0.12216173112392426,
"learning_rate": 1.3531898947629296e-06,
"loss": 1.1179,
"step": 923
},
{
"epoch": 0.96,
"grad_norm": 0.10309845954179764,
"learning_rate": 1.2865197853333178e-06,
"loss": 1.328,
"step": 924
},
{
"epoch": 0.96,
"grad_norm": 0.1231243684887886,
"learning_rate": 1.221523178997075e-06,
"loss": 1.1911,
"step": 925
},
{
"epoch": 0.96,
"grad_norm": 0.12335013598203659,
"learning_rate": 1.158201177650553e-06,
"loss": 1.2257,
"step": 926
},
{
"epoch": 0.96,
"grad_norm": 0.11128874868154526,
"learning_rate": 1.0965548548002803e-06,
"loss": 1.2306,
"step": 927
},
{
"epoch": 0.96,
"grad_norm": 0.10876341909170151,
"learning_rate": 1.036585255544764e-06,
"loss": 1.2044,
"step": 928
},
{
"epoch": 0.96,
"grad_norm": 0.12024974822998047,
"learning_rate": 9.782933965567954e-07,
"loss": 1.1953,
"step": 929
},
{
"epoch": 0.97,
"grad_norm": 0.11236795783042908,
"learning_rate": 9.21680266066216e-07,
"loss": 1.3033,
"step": 930
},
{
"epoch": 0.97,
"grad_norm": 0.11030973494052887,
"learning_rate": 8.667468238431453e-07,
"loss": 1.2,
"step": 931
},
{
"epoch": 0.97,
"grad_norm": 0.10306066274642944,
"learning_rate": 8.13494001181725e-07,
"loss": 1.3141,
"step": 932
},
{
"epoch": 0.97,
"grad_norm": 0.10356069356203079,
"learning_rate": 7.619227008843322e-07,
"loss": 1.2244,
"step": 933
},
{
"epoch": 0.97,
"grad_norm": 0.10682324320077896,
"learning_rate": 7.120337972462365e-07,
"loss": 1.1282,
"step": 934
},
{
"epoch": 0.97,
"grad_norm": 0.11171166598796844,
"learning_rate": 6.638281360408339e-07,
"loss": 1.1534,
"step": 935
},
{
"epoch": 0.97,
"grad_norm": 0.09517676383256912,
"learning_rate": 6.173065345052687e-07,
"loss": 1.1292,
"step": 936
},
{
"epoch": 0.97,
"grad_norm": 0.11539721488952637,
"learning_rate": 5.724697813265901e-07,
"loss": 1.1716,
"step": 937
},
{
"epoch": 0.97,
"grad_norm": 0.11124927550554276,
"learning_rate": 5.293186366284176e-07,
"loss": 1.1894,
"step": 938
},
{
"epoch": 0.97,
"grad_norm": 0.261902779340744,
"learning_rate": 4.878538319579629e-07,
"loss": 1.2252,
"step": 939
},
{
"epoch": 0.98,
"grad_norm": 0.11507264524698257,
"learning_rate": 4.4807607027372855e-07,
"loss": 1.2918,
"step": 940
},
{
"epoch": 0.98,
"grad_norm": 0.1210726797580719,
"learning_rate": 4.0998602593352863e-07,
"loss": 1.1893,
"step": 941
},
{
"epoch": 0.98,
"grad_norm": 0.11546792089939117,
"learning_rate": 3.7358434468308667e-07,
"loss": 1.1486,
"step": 942
},
{
"epoch": 0.98,
"grad_norm": 0.12196337431669235,
"learning_rate": 3.388716436450667e-07,
"loss": 1.3495,
"step": 943
},
{
"epoch": 0.98,
"grad_norm": 0.10631462186574936,
"learning_rate": 3.0584851130861513e-07,
"loss": 1.1866,
"step": 944
},
{
"epoch": 0.98,
"grad_norm": 0.14002542197704315,
"learning_rate": 2.745155075194128e-07,
"loss": 0.9632,
"step": 945
},
{
"epoch": 0.98,
"grad_norm": 0.15344469249248505,
"learning_rate": 2.448731634701384e-07,
"loss": 1.3407,
"step": 946
},
{
"epoch": 0.98,
"grad_norm": 0.11114881932735443,
"learning_rate": 2.169219816914869e-07,
"loss": 1.1723,
"step": 947
},
{
"epoch": 0.98,
"grad_norm": 0.09970024973154068,
"learning_rate": 1.9066243604367594e-07,
"loss": 1.3012,
"step": 948
},
{
"epoch": 0.98,
"grad_norm": 0.10021547228097916,
"learning_rate": 1.6609497170834155e-07,
"loss": 1.1621,
"step": 949
},
{
"epoch": 0.99,
"grad_norm": 0.11028821766376495,
"learning_rate": 1.4322000518106616e-07,
"loss": 1.092,
"step": 950
},
{
"epoch": 0.99,
"grad_norm": 0.10497645288705826,
"learning_rate": 1.220379242642844e-07,
"loss": 1.1194,
"step": 951
},
{
"epoch": 0.99,
"grad_norm": 0.10754618048667908,
"learning_rate": 1.025490880606883e-07,
"loss": 1.0586,
"step": 952
},
{
"epoch": 0.99,
"grad_norm": 0.1118793711066246,
"learning_rate": 8.475382696719881e-08,
"loss": 1.1977,
"step": 953
},
{
"epoch": 0.99,
"grad_norm": 0.11205057799816132,
"learning_rate": 6.865244266928139e-08,
"loss": 0.944,
"step": 954
},
{
"epoch": 0.99,
"grad_norm": 0.10550331324338913,
"learning_rate": 5.424520813590572e-08,
"loss": 1.1463,
"step": 955
},
{
"epoch": 0.99,
"grad_norm": 0.10324594378471375,
"learning_rate": 4.153236761488266e-08,
"loss": 1.1325,
"step": 956
},
{
"epoch": 0.99,
"grad_norm": 0.10522052645683289,
"learning_rate": 3.051413662874536e-08,
"loss": 1.1922,
"step": 957
},
{
"epoch": 0.99,
"grad_norm": 0.12036368250846863,
"learning_rate": 2.1190701971052218e-08,
"loss": 1.123,
"step": 958
},
{
"epoch": 1.0,
"grad_norm": 0.11026199162006378,
"learning_rate": 1.3562221703267152e-08,
"loss": 1.261,
"step": 959
},
{
"epoch": 1.0,
"grad_norm": 0.28202083706855774,
"learning_rate": 7.628825152050657e-09,
"loss": 1.2547,
"step": 960
},
{
"epoch": 1.0,
"grad_norm": 0.10787337273359299,
"learning_rate": 3.390612907094859e-09,
"loss": 1.2832,
"step": 961
},
{
"epoch": 1.0,
"grad_norm": 0.10484518110752106,
"learning_rate": 8.476568193804824e-10,
"loss": 1.2884,
"step": 962
},
{
"epoch": 1.0,
"grad_norm": 0.11853493005037308,
"learning_rate": 0.0,
"loss": 1.2526,
"step": 963
},
{
"epoch": 1.0,
"step": 963,
"total_flos": 1.3034774175330337e+19,
"train_loss": 1.2195818047285822,
"train_runtime": 50845.1793,
"train_samples_per_second": 0.152,
"train_steps_per_second": 0.019
}
],
"logging_steps": 1.0,
"max_steps": 963,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"total_flos": 1.3034774175330337e+19,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}