{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9942418426103647, "eval_steps": 130, "global_step": 520, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003838771593090211, "grad_norm": 41.942832946777344, "learning_rate": 5e-06, "loss": 29.5342, "step": 1 }, { "epoch": 0.003838771593090211, "eval_loss": 1.869257926940918, "eval_runtime": 31.5851, "eval_samples_per_second": 9.372, "eval_steps_per_second": 4.686, "step": 1 }, { "epoch": 0.007677543186180422, "grad_norm": 37.62504577636719, "learning_rate": 1e-05, "loss": 29.602, "step": 2 }, { "epoch": 0.011516314779270634, "grad_norm": 27.579742431640625, "learning_rate": 1.5e-05, "loss": 28.6884, "step": 3 }, { "epoch": 0.015355086372360844, "grad_norm": 36.99763870239258, "learning_rate": 2e-05, "loss": 29.9575, "step": 4 }, { "epoch": 0.019193857965451054, "grad_norm": 37.934635162353516, "learning_rate": 2.5e-05, "loss": 28.0644, "step": 5 }, { "epoch": 0.023032629558541268, "grad_norm": 45.45661163330078, "learning_rate": 3e-05, "loss": 29.0135, "step": 6 }, { "epoch": 0.026871401151631478, "grad_norm": 31.74431800842285, "learning_rate": 3.5e-05, "loss": 29.0581, "step": 7 }, { "epoch": 0.030710172744721688, "grad_norm": 32.59209060668945, "learning_rate": 4e-05, "loss": 28.5253, "step": 8 }, { "epoch": 0.0345489443378119, "grad_norm": 26.19696617126465, "learning_rate": 4.5e-05, "loss": 27.7504, "step": 9 }, { "epoch": 0.03838771593090211, "grad_norm": 30.916200637817383, "learning_rate": 5e-05, "loss": 29.3754, "step": 10 }, { "epoch": 0.04222648752399232, "grad_norm": 25.58551597595215, "learning_rate": 4.9999525683718174e-05, "loss": 29.0332, "step": 11 }, { "epoch": 0.046065259117082535, "grad_norm": 30.068437576293945, "learning_rate": 4.9998102752870765e-05, "loss": 29.2469, "step": 12 }, { "epoch": 0.04990403071017274, "grad_norm": 25.17665672302246, "learning_rate": 4.999573126145132e-05, "loss": 28.2097, "step": 13 }, { "epoch": 0.053742802303262956, "grad_norm": 27.615007400512695, "learning_rate": 4.99924112994468e-05, "loss": 29.7609, "step": 14 }, { "epoch": 0.05758157389635317, "grad_norm": 26.473073959350586, "learning_rate": 4.998814299283415e-05, "loss": 29.3846, "step": 15 }, { "epoch": 0.061420345489443376, "grad_norm": 22.732168197631836, "learning_rate": 4.998292650357558e-05, "loss": 28.2741, "step": 16 }, { "epoch": 0.06525911708253358, "grad_norm": 26.874069213867188, "learning_rate": 4.9976762029612335e-05, "loss": 30.2697, "step": 17 }, { "epoch": 0.0690978886756238, "grad_norm": 22.940404891967773, "learning_rate": 4.9969649804857257e-05, "loss": 29.0416, "step": 18 }, { "epoch": 0.07293666026871401, "grad_norm": 22.93003273010254, "learning_rate": 4.996159009918585e-05, "loss": 30.0236, "step": 19 }, { "epoch": 0.07677543186180422, "grad_norm": 21.159198760986328, "learning_rate": 4.995258321842611e-05, "loss": 29.995, "step": 20 }, { "epoch": 0.08061420345489444, "grad_norm": 23.158321380615234, "learning_rate": 4.9942629504346827e-05, "loss": 30.1427, "step": 21 }, { "epoch": 0.08445297504798464, "grad_norm": 22.745277404785156, "learning_rate": 4.993172933464471e-05, "loss": 29.0171, "step": 22 }, { "epoch": 0.08829174664107485, "grad_norm": 21.473155975341797, "learning_rate": 4.9919883122929976e-05, "loss": 29.1784, "step": 23 }, { "epoch": 0.09213051823416507, "grad_norm": 21.882991790771484, "learning_rate": 4.990709131871074e-05, "loss": 27.7198, "step": 24 }, { "epoch": 0.09596928982725528, "grad_norm": 20.5384464263916, "learning_rate": 4.989335440737586e-05, "loss": 29.3608, "step": 25 }, { "epoch": 0.09980806142034548, "grad_norm": 21.028078079223633, "learning_rate": 4.9878672910176615e-05, "loss": 30.7017, "step": 26 }, { "epoch": 0.1036468330134357, "grad_norm": 20.912294387817383, "learning_rate": 4.9863047384206835e-05, "loss": 29.2454, "step": 27 }, { "epoch": 0.10748560460652591, "grad_norm": 22.990535736083984, "learning_rate": 4.984647842238185e-05, "loss": 30.5442, "step": 28 }, { "epoch": 0.11132437619961612, "grad_norm": 20.818967819213867, "learning_rate": 4.982896665341591e-05, "loss": 28.0971, "step": 29 }, { "epoch": 0.11516314779270634, "grad_norm": 22.41204071044922, "learning_rate": 4.98105127417984e-05, "loss": 30.0213, "step": 30 }, { "epoch": 0.11900191938579655, "grad_norm": 20.42460823059082, "learning_rate": 4.979111738776857e-05, "loss": 29.0625, "step": 31 }, { "epoch": 0.12284069097888675, "grad_norm": 23.12067413330078, "learning_rate": 4.977078132728901e-05, "loss": 30.4828, "step": 32 }, { "epoch": 0.12667946257197696, "grad_norm": 19.101734161376953, "learning_rate": 4.974950533201767e-05, "loss": 27.8432, "step": 33 }, { "epoch": 0.13051823416506717, "grad_norm": 18.871788024902344, "learning_rate": 4.972729020927865e-05, "loss": 28.2289, "step": 34 }, { "epoch": 0.1343570057581574, "grad_norm": 19.23995590209961, "learning_rate": 4.9704136802031485e-05, "loss": 28.7979, "step": 35 }, { "epoch": 0.1381957773512476, "grad_norm": 19.988222122192383, "learning_rate": 4.968004598883923e-05, "loss": 29.0405, "step": 36 }, { "epoch": 0.1420345489443378, "grad_norm": 17.621200561523438, "learning_rate": 4.965501868383506e-05, "loss": 28.8017, "step": 37 }, { "epoch": 0.14587332053742802, "grad_norm": 19.57996368408203, "learning_rate": 4.9629055836687665e-05, "loss": 28.5596, "step": 38 }, { "epoch": 0.14971209213051823, "grad_norm": 19.62120819091797, "learning_rate": 4.960215843256511e-05, "loss": 28.6677, "step": 39 }, { "epoch": 0.15355086372360843, "grad_norm": 18.79466438293457, "learning_rate": 4.957432749209755e-05, "loss": 30.2908, "step": 40 }, { "epoch": 0.15738963531669867, "grad_norm": 21.75792694091797, "learning_rate": 4.9545564071338426e-05, "loss": 28.7169, "step": 41 }, { "epoch": 0.16122840690978887, "grad_norm": 21.684478759765625, "learning_rate": 4.9515869261724446e-05, "loss": 28.8571, "step": 42 }, { "epoch": 0.16506717850287908, "grad_norm": 20.90673065185547, "learning_rate": 4.948524419003415e-05, "loss": 28.5535, "step": 43 }, { "epoch": 0.1689059500959693, "grad_norm": 20.5564022064209, "learning_rate": 4.9453690018345144e-05, "loss": 29.5709, "step": 44 }, { "epoch": 0.1727447216890595, "grad_norm": 21.395124435424805, "learning_rate": 4.942120794399002e-05, "loss": 28.016, "step": 45 }, { "epoch": 0.1765834932821497, "grad_norm": 18.09257698059082, "learning_rate": 4.938779919951092e-05, "loss": 29.6793, "step": 46 }, { "epoch": 0.18042226487523993, "grad_norm": 19.043846130371094, "learning_rate": 4.935346505261276e-05, "loss": 30.1972, "step": 47 }, { "epoch": 0.18426103646833014, "grad_norm": 20.1807804107666, "learning_rate": 4.931820680611512e-05, "loss": 28.7825, "step": 48 }, { "epoch": 0.18809980806142035, "grad_norm": 18.534324645996094, "learning_rate": 4.928202579790285e-05, "loss": 27.895, "step": 49 }, { "epoch": 0.19193857965451055, "grad_norm": 19.662391662597656, "learning_rate": 4.9244923400875245e-05, "loss": 30.0118, "step": 50 }, { "epoch": 0.19577735124760076, "grad_norm": 19.72050666809082, "learning_rate": 4.920690102289397e-05, "loss": 28.6979, "step": 51 }, { "epoch": 0.19961612284069097, "grad_norm": 20.052833557128906, "learning_rate": 4.916796010672969e-05, "loss": 29.5174, "step": 52 }, { "epoch": 0.2034548944337812, "grad_norm": 18.85672378540039, "learning_rate": 4.9128102130007225e-05, "loss": 28.3002, "step": 53 }, { "epoch": 0.2072936660268714, "grad_norm": 18.159709930419922, "learning_rate": 4.908732860514958e-05, "loss": 28.7818, "step": 54 }, { "epoch": 0.21113243761996162, "grad_norm": 20.378021240234375, "learning_rate": 4.9045641079320484e-05, "loss": 29.458, "step": 55 }, { "epoch": 0.21497120921305182, "grad_norm": 21.459836959838867, "learning_rate": 4.900304113436571e-05, "loss": 27.3894, "step": 56 }, { "epoch": 0.21880998080614203, "grad_norm": 19.784387588500977, "learning_rate": 4.895953038675306e-05, "loss": 29.3604, "step": 57 }, { "epoch": 0.22264875239923224, "grad_norm": 19.705406188964844, "learning_rate": 4.891511048751102e-05, "loss": 27.4372, "step": 58 }, { "epoch": 0.22648752399232247, "grad_norm": 19.9382381439209, "learning_rate": 4.886978312216611e-05, "loss": 28.2482, "step": 59 }, { "epoch": 0.23032629558541268, "grad_norm": 19.08539581298828, "learning_rate": 4.882355001067892e-05, "loss": 28.4456, "step": 60 }, { "epoch": 0.23416506717850288, "grad_norm": 18.285186767578125, "learning_rate": 4.877641290737884e-05, "loss": 28.745, "step": 61 }, { "epoch": 0.2380038387715931, "grad_norm": 19.981107711791992, "learning_rate": 4.872837360089754e-05, "loss": 27.6915, "step": 62 }, { "epoch": 0.2418426103646833, "grad_norm": 18.624874114990234, "learning_rate": 4.8679433914101006e-05, "loss": 28.0039, "step": 63 }, { "epoch": 0.2456813819577735, "grad_norm": 22.298728942871094, "learning_rate": 4.862959570402049e-05, "loss": 28.3383, "step": 64 }, { "epoch": 0.2495201535508637, "grad_norm": 20.383586883544922, "learning_rate": 4.857886086178194e-05, "loss": 28.918, "step": 65 }, { "epoch": 0.2533589251439539, "grad_norm": 18.7249755859375, "learning_rate": 4.85272313125343e-05, "loss": 29.0783, "step": 66 }, { "epoch": 0.2571976967370441, "grad_norm": 18.3210391998291, "learning_rate": 4.8474709015376416e-05, "loss": 27.3239, "step": 67 }, { "epoch": 0.26103646833013433, "grad_norm": 19.459135055541992, "learning_rate": 4.842129596328277e-05, "loss": 29.2664, "step": 68 }, { "epoch": 0.2648752399232246, "grad_norm": 18.80237579345703, "learning_rate": 4.836699418302777e-05, "loss": 28.8449, "step": 69 }, { "epoch": 0.2687140115163148, "grad_norm": 18.05824851989746, "learning_rate": 4.8311805735108894e-05, "loss": 27.5182, "step": 70 }, { "epoch": 0.272552783109405, "grad_norm": 18.784578323364258, "learning_rate": 4.825573271366851e-05, "loss": 28.4831, "step": 71 }, { "epoch": 0.2763915547024952, "grad_norm": 18.89769172668457, "learning_rate": 4.8198777246414373e-05, "loss": 27.7881, "step": 72 }, { "epoch": 0.2802303262955854, "grad_norm": 17.64800453186035, "learning_rate": 4.814094149453891e-05, "loss": 27.7358, "step": 73 }, { "epoch": 0.2840690978886756, "grad_norm": 19.012086868286133, "learning_rate": 4.808222765263724e-05, "loss": 29.8711, "step": 74 }, { "epoch": 0.28790786948176583, "grad_norm": 18.49270248413086, "learning_rate": 4.802263794862385e-05, "loss": 28.3761, "step": 75 }, { "epoch": 0.29174664107485604, "grad_norm": 21.752199172973633, "learning_rate": 4.796217464364808e-05, "loss": 28.3796, "step": 76 }, { "epoch": 0.29558541266794625, "grad_norm": 19.340524673461914, "learning_rate": 4.790084003200834e-05, "loss": 29.4554, "step": 77 }, { "epoch": 0.29942418426103645, "grad_norm": 21.60015869140625, "learning_rate": 4.783863644106502e-05, "loss": 28.0435, "step": 78 }, { "epoch": 0.30326295585412666, "grad_norm": 20.29178237915039, "learning_rate": 4.777556623115221e-05, "loss": 27.563, "step": 79 }, { "epoch": 0.30710172744721687, "grad_norm": 18.13334846496582, "learning_rate": 4.7711631795488096e-05, "loss": 27.4931, "step": 80 }, { "epoch": 0.31094049904030713, "grad_norm": 18.14622688293457, "learning_rate": 4.764683556008418e-05, "loss": 28.0506, "step": 81 }, { "epoch": 0.31477927063339733, "grad_norm": 19.75128936767578, "learning_rate": 4.758117998365322e-05, "loss": 27.2963, "step": 82 }, { "epoch": 0.31861804222648754, "grad_norm": 17.486196517944336, "learning_rate": 4.751466755751594e-05, "loss": 28.7996, "step": 83 }, { "epoch": 0.32245681381957775, "grad_norm": 19.39826011657715, "learning_rate": 4.7447300805506455e-05, "loss": 27.6073, "step": 84 }, { "epoch": 0.32629558541266795, "grad_norm": 17.352880477905273, "learning_rate": 4.7379082283876566e-05, "loss": 28.5087, "step": 85 }, { "epoch": 0.33013435700575816, "grad_norm": 20.071483612060547, "learning_rate": 4.7310014581198695e-05, "loss": 27.1697, "step": 86 }, { "epoch": 0.33397312859884837, "grad_norm": 19.04390525817871, "learning_rate": 4.7240100318267746e-05, "loss": 27.9563, "step": 87 }, { "epoch": 0.3378119001919386, "grad_norm": 20.634225845336914, "learning_rate": 4.716934214800155e-05, "loss": 27.6816, "step": 88 }, { "epoch": 0.3416506717850288, "grad_norm": 19.746788024902344, "learning_rate": 4.70977427553403e-05, "loss": 27.3921, "step": 89 }, { "epoch": 0.345489443378119, "grad_norm": 19.04279899597168, "learning_rate": 4.702530485714461e-05, "loss": 27.2845, "step": 90 }, { "epoch": 0.3493282149712092, "grad_norm": 18.288244247436523, "learning_rate": 4.695203120209245e-05, "loss": 28.8947, "step": 91 }, { "epoch": 0.3531669865642994, "grad_norm": 18.553823471069336, "learning_rate": 4.6877924570574817e-05, "loss": 27.7966, "step": 92 }, { "epoch": 0.3570057581573896, "grad_norm": 18.615053176879883, "learning_rate": 4.6802987774590275e-05, "loss": 27.8787, "step": 93 }, { "epoch": 0.36084452975047987, "grad_norm": 18.30453109741211, "learning_rate": 4.672722365763821e-05, "loss": 29.0792, "step": 94 }, { "epoch": 0.3646833013435701, "grad_norm": 18.134958267211914, "learning_rate": 4.665063509461097e-05, "loss": 29.07, "step": 95 }, { "epoch": 0.3685220729366603, "grad_norm": 19.019006729125977, "learning_rate": 4.6573224991684744e-05, "loss": 27.708, "step": 96 }, { "epoch": 0.3723608445297505, "grad_norm": 18.863723754882812, "learning_rate": 4.649499628620931e-05, "loss": 27.4551, "step": 97 }, { "epoch": 0.3761996161228407, "grad_norm": 18.62997055053711, "learning_rate": 4.641595194659657e-05, "loss": 26.7616, "step": 98 }, { "epoch": 0.3800383877159309, "grad_norm": 18.19691276550293, "learning_rate": 4.63360949722079e-05, "loss": 27.6041, "step": 99 }, { "epoch": 0.3838771593090211, "grad_norm": 19.37999153137207, "learning_rate": 4.625542839324036e-05, "loss": 26.9613, "step": 100 }, { "epoch": 0.3877159309021113, "grad_norm": 18.335630416870117, "learning_rate": 4.617395527061168e-05, "loss": 27.4261, "step": 101 }, { "epoch": 0.3915547024952015, "grad_norm": 19.16864585876465, "learning_rate": 4.6091678695844164e-05, "loss": 28.8591, "step": 102 }, { "epoch": 0.39539347408829173, "grad_norm": 20.256633758544922, "learning_rate": 4.600860179094732e-05, "loss": 26.8054, "step": 103 }, { "epoch": 0.39923224568138194, "grad_norm": 18.64192008972168, "learning_rate": 4.5924727708299445e-05, "loss": 27.3249, "step": 104 }, { "epoch": 0.40307101727447214, "grad_norm": 17.976713180541992, "learning_rate": 4.584005963052799e-05, "loss": 29.1977, "step": 105 }, { "epoch": 0.4069097888675624, "grad_norm": 17.99713706970215, "learning_rate": 4.575460077038877e-05, "loss": 29.0127, "step": 106 }, { "epoch": 0.4107485604606526, "grad_norm": 18.080854415893555, "learning_rate": 4.5668354370644086e-05, "loss": 29.716, "step": 107 }, { "epoch": 0.4145873320537428, "grad_norm": 17.458669662475586, "learning_rate": 4.558132370393968e-05, "loss": 29.7465, "step": 108 }, { "epoch": 0.418426103646833, "grad_norm": 18.444486618041992, "learning_rate": 4.5493512072680536e-05, "loss": 30.3556, "step": 109 }, { "epoch": 0.42226487523992323, "grad_norm": 17.931936264038086, "learning_rate": 4.540492280890555e-05, "loss": 27.0296, "step": 110 }, { "epoch": 0.42610364683301344, "grad_norm": 17.657825469970703, "learning_rate": 4.5315559274161144e-05, "loss": 28.3838, "step": 111 }, { "epoch": 0.42994241842610365, "grad_norm": 17.2442626953125, "learning_rate": 4.522542485937369e-05, "loss": 26.963, "step": 112 }, { "epoch": 0.43378119001919385, "grad_norm": 16.4385929107666, "learning_rate": 4.5134522984720816e-05, "loss": 28.0953, "step": 113 }, { "epoch": 0.43761996161228406, "grad_norm": 16.927474975585938, "learning_rate": 4.504285709950167e-05, "loss": 29.0116, "step": 114 }, { "epoch": 0.44145873320537427, "grad_norm": 18.508386611938477, "learning_rate": 4.4950430682006e-05, "loss": 27.5212, "step": 115 }, { "epoch": 0.44529750479846447, "grad_norm": 17.00088882446289, "learning_rate": 4.485724723938215e-05, "loss": 27.2209, "step": 116 }, { "epoch": 0.4491362763915547, "grad_norm": 17.35956573486328, "learning_rate": 4.476331030750408e-05, "loss": 28.9031, "step": 117 }, { "epoch": 0.45297504798464494, "grad_norm": 17.835237503051758, "learning_rate": 4.4668623450837085e-05, "loss": 25.8466, "step": 118 }, { "epoch": 0.45681381957773515, "grad_norm": 18.052587509155273, "learning_rate": 4.457319026230257e-05, "loss": 27.7288, "step": 119 }, { "epoch": 0.46065259117082535, "grad_norm": 21.071945190429688, "learning_rate": 4.447701436314176e-05, "loss": 27.2769, "step": 120 }, { "epoch": 0.46449136276391556, "grad_norm": 17.07927703857422, "learning_rate": 4.4380099402778244e-05, "loss": 27.3407, "step": 121 }, { "epoch": 0.46833013435700577, "grad_norm": 17.430988311767578, "learning_rate": 4.428244905867952e-05, "loss": 27.7631, "step": 122 }, { "epoch": 0.472168905950096, "grad_norm": 18.140281677246094, "learning_rate": 4.418406703621743e-05, "loss": 27.4634, "step": 123 }, { "epoch": 0.4760076775431862, "grad_norm": 18.464765548706055, "learning_rate": 4.408495706852758e-05, "loss": 27.1883, "step": 124 }, { "epoch": 0.4798464491362764, "grad_norm": 17.243370056152344, "learning_rate": 4.398512291636768e-05, "loss": 29.4247, "step": 125 }, { "epoch": 0.4836852207293666, "grad_norm": 18.10504722595215, "learning_rate": 4.3884568367974845e-05, "loss": 27.3026, "step": 126 }, { "epoch": 0.4875239923224568, "grad_norm": 18.39180564880371, "learning_rate": 4.378329723892184e-05, "loss": 28.0279, "step": 127 }, { "epoch": 0.491362763915547, "grad_norm": 16.66558265686035, "learning_rate": 4.3681313371972276e-05, "loss": 27.5308, "step": 128 }, { "epoch": 0.4952015355086372, "grad_norm": 16.756423950195312, "learning_rate": 4.357862063693486e-05, "loss": 28.4946, "step": 129 }, { "epoch": 0.4990403071017274, "grad_norm": 17.056476593017578, "learning_rate": 4.347522293051648e-05, "loss": 27.8562, "step": 130 }, { "epoch": 0.4990403071017274, "eval_loss": 1.7600558996200562, "eval_runtime": 31.0188, "eval_samples_per_second": 9.543, "eval_steps_per_second": 4.771, "step": 130 }, { "epoch": 0.5028790786948176, "grad_norm": 17.216785430908203, "learning_rate": 4.337112417617439e-05, "loss": 27.1427, "step": 131 }, { "epoch": 0.5067178502879078, "grad_norm": 16.08450698852539, "learning_rate": 4.3266328323967333e-05, "loss": 27.4919, "step": 132 }, { "epoch": 0.510556621880998, "grad_norm": 17.361270904541016, "learning_rate": 4.3160839350405606e-05, "loss": 26.1541, "step": 133 }, { "epoch": 0.5143953934740882, "grad_norm": 17.303585052490234, "learning_rate": 4.305466125830023e-05, "loss": 26.0347, "step": 134 }, { "epoch": 0.5182341650671785, "grad_norm": 17.440980911254883, "learning_rate": 4.294779807661105e-05, "loss": 27.2826, "step": 135 }, { "epoch": 0.5220729366602687, "grad_norm": 19.435768127441406, "learning_rate": 4.284025386029381e-05, "loss": 27.5317, "step": 136 }, { "epoch": 0.525911708253359, "grad_norm": 17.796663284301758, "learning_rate": 4.273203269014634e-05, "loss": 27.8947, "step": 137 }, { "epoch": 0.5297504798464492, "grad_norm": 17.049524307250977, "learning_rate": 4.2623138672653684e-05, "loss": 29.3983, "step": 138 }, { "epoch": 0.5335892514395394, "grad_norm": 17.124839782714844, "learning_rate": 4.2513575939832275e-05, "loss": 28.4112, "step": 139 }, { "epoch": 0.5374280230326296, "grad_norm": 17.724750518798828, "learning_rate": 4.2403348649073174e-05, "loss": 29.1517, "step": 140 }, { "epoch": 0.5412667946257198, "grad_norm": 17.679073333740234, "learning_rate": 4.2292460982984255e-05, "loss": 27.1786, "step": 141 }, { "epoch": 0.54510556621881, "grad_norm": 18.939489364624023, "learning_rate": 4.218091714923157e-05, "loss": 28.6373, "step": 142 }, { "epoch": 0.5489443378119002, "grad_norm": 17.122732162475586, "learning_rate": 4.206872138037964e-05, "loss": 27.252, "step": 143 }, { "epoch": 0.5527831094049904, "grad_norm": 18.2526912689209, "learning_rate": 4.1955877933730855e-05, "loss": 29.9347, "step": 144 }, { "epoch": 0.5566218809980806, "grad_norm": 18.54884147644043, "learning_rate": 4.184239109116393e-05, "loss": 27.988, "step": 145 }, { "epoch": 0.5604606525911708, "grad_norm": 18.026559829711914, "learning_rate": 4.172826515897146e-05, "loss": 28.1473, "step": 146 }, { "epoch": 0.564299424184261, "grad_norm": 17.47412109375, "learning_rate": 4.161350446769645e-05, "loss": 26.7597, "step": 147 }, { "epoch": 0.5681381957773513, "grad_norm": 19.019777297973633, "learning_rate": 4.149811337196807e-05, "loss": 27.5097, "step": 148 }, { "epoch": 0.5719769673704415, "grad_norm": 18.873889923095703, "learning_rate": 4.1382096250336346e-05, "loss": 26.3739, "step": 149 }, { "epoch": 0.5758157389635317, "grad_norm": 17.524677276611328, "learning_rate": 4.126545750510605e-05, "loss": 26.7282, "step": 150 }, { "epoch": 0.5796545105566219, "grad_norm": 17.283926010131836, "learning_rate": 4.1148201562169685e-05, "loss": 26.9749, "step": 151 }, { "epoch": 0.5834932821497121, "grad_norm": 17.261751174926758, "learning_rate": 4.103033287083946e-05, "loss": 27.2926, "step": 152 }, { "epoch": 0.5873320537428023, "grad_norm": 16.952129364013672, "learning_rate": 4.0911855903678534e-05, "loss": 29.1231, "step": 153 }, { "epoch": 0.5911708253358925, "grad_norm": 16.523609161376953, "learning_rate": 4.079277515633127e-05, "loss": 27.9168, "step": 154 }, { "epoch": 0.5950095969289827, "grad_norm": 17.648927688598633, "learning_rate": 4.067309514735267e-05, "loss": 26.9956, "step": 155 }, { "epoch": 0.5988483685220729, "grad_norm": 18.750659942626953, "learning_rate": 4.055282041803685e-05, "loss": 27.5584, "step": 156 }, { "epoch": 0.6026871401151631, "grad_norm": 17.183300018310547, "learning_rate": 4.0431955532244827e-05, "loss": 27.9265, "step": 157 }, { "epoch": 0.6065259117082533, "grad_norm": 17.270681381225586, "learning_rate": 4.031050507623125e-05, "loss": 28.9735, "step": 158 }, { "epoch": 0.6103646833013435, "grad_norm": 19.801315307617188, "learning_rate": 4.0188473658470426e-05, "loss": 26.7811, "step": 159 }, { "epoch": 0.6142034548944337, "grad_norm": 16.823190689086914, "learning_rate": 4.0065865909481417e-05, "loss": 26.9071, "step": 160 }, { "epoch": 0.6180422264875239, "grad_norm": 17.074491500854492, "learning_rate": 3.9942686481652344e-05, "loss": 27.9241, "step": 161 }, { "epoch": 0.6218809980806143, "grad_norm": 18.242040634155273, "learning_rate": 3.981894004906388e-05, "loss": 28.2823, "step": 162 }, { "epoch": 0.6257197696737045, "grad_norm": 16.200469970703125, "learning_rate": 3.969463130731183e-05, "loss": 27.2651, "step": 163 }, { "epoch": 0.6295585412667947, "grad_norm": 17.807695388793945, "learning_rate": 3.9569764973329026e-05, "loss": 28.572, "step": 164 }, { "epoch": 0.6333973128598849, "grad_norm": 16.356901168823242, "learning_rate": 3.9444345785206285e-05, "loss": 27.4942, "step": 165 }, { "epoch": 0.6372360844529751, "grad_norm": 18.459415435791016, "learning_rate": 3.931837850201263e-05, "loss": 27.4934, "step": 166 }, { "epoch": 0.6410748560460653, "grad_norm": 18.079626083374023, "learning_rate": 3.919186790361475e-05, "loss": 28.0261, "step": 167 }, { "epoch": 0.6449136276391555, "grad_norm": 18.058738708496094, "learning_rate": 3.906481879049558e-05, "loss": 27.2611, "step": 168 }, { "epoch": 0.6487523992322457, "grad_norm": 17.553970336914062, "learning_rate": 3.893723598357214e-05, "loss": 27.5768, "step": 169 }, { "epoch": 0.6525911708253359, "grad_norm": 18.69593620300293, "learning_rate": 3.880912432401265e-05, "loss": 27.5784, "step": 170 }, { "epoch": 0.6564299424184261, "grad_norm": 17.491533279418945, "learning_rate": 3.8680488673052784e-05, "loss": 29.1586, "step": 171 }, { "epoch": 0.6602687140115163, "grad_norm": 18.12783432006836, "learning_rate": 3.855133391181124e-05, "loss": 27.3099, "step": 172 }, { "epoch": 0.6641074856046065, "grad_norm": 18.322479248046875, "learning_rate": 3.842166494110451e-05, "loss": 29.5131, "step": 173 }, { "epoch": 0.6679462571976967, "grad_norm": 17.041065216064453, "learning_rate": 3.82914866812609e-05, "loss": 28.5525, "step": 174 }, { "epoch": 0.6717850287907869, "grad_norm": 17.609098434448242, "learning_rate": 3.81608040719339e-05, "loss": 28.8543, "step": 175 }, { "epoch": 0.6756238003838771, "grad_norm": 17.111724853515625, "learning_rate": 3.802962207191463e-05, "loss": 25.6639, "step": 176 }, { "epoch": 0.6794625719769674, "grad_norm": 17.328916549682617, "learning_rate": 3.789794565894378e-05, "loss": 26.0536, "step": 177 }, { "epoch": 0.6833013435700576, "grad_norm": 17.837627410888672, "learning_rate": 3.7765779829522675e-05, "loss": 27.5278, "step": 178 }, { "epoch": 0.6871401151631478, "grad_norm": 15.394762992858887, "learning_rate": 3.7633129598723704e-05, "loss": 27.7244, "step": 179 }, { "epoch": 0.690978886756238, "grad_norm": 19.647441864013672, "learning_rate": 3.7500000000000003e-05, "loss": 25.729, "step": 180 }, { "epoch": 0.6948176583493282, "grad_norm": 17.077699661254883, "learning_rate": 3.7366396084994475e-05, "loss": 26.2709, "step": 181 }, { "epoch": 0.6986564299424184, "grad_norm": 16.858604431152344, "learning_rate": 3.723232292334809e-05, "loss": 27.8621, "step": 182 }, { "epoch": 0.7024952015355086, "grad_norm": 17.455453872680664, "learning_rate": 3.709778560250754e-05, "loss": 27.0243, "step": 183 }, { "epoch": 0.7063339731285988, "grad_norm": 16.5456600189209, "learning_rate": 3.696278922753216e-05, "loss": 27.6667, "step": 184 }, { "epoch": 0.710172744721689, "grad_norm": 17.52065086364746, "learning_rate": 3.6827338920900254e-05, "loss": 27.4649, "step": 185 }, { "epoch": 0.7140115163147792, "grad_norm": 15.55224323272705, "learning_rate": 3.669143982231467e-05, "loss": 26.4226, "step": 186 }, { "epoch": 0.7178502879078695, "grad_norm": 16.189666748046875, "learning_rate": 3.655509708850783e-05, "loss": 26.3024, "step": 187 }, { "epoch": 0.7216890595009597, "grad_norm": 17.70796775817871, "learning_rate": 3.641831589304602e-05, "loss": 27.1759, "step": 188 }, { "epoch": 0.72552783109405, "grad_norm": 16.10495376586914, "learning_rate": 3.628110142613308e-05, "loss": 26.5754, "step": 189 }, { "epoch": 0.7293666026871402, "grad_norm": 17.34556007385254, "learning_rate": 3.6143458894413465e-05, "loss": 27.8653, "step": 190 }, { "epoch": 0.7332053742802304, "grad_norm": 18.321557998657227, "learning_rate": 3.600539352077469e-05, "loss": 27.7326, "step": 191 }, { "epoch": 0.7370441458733206, "grad_norm": 15.874064445495605, "learning_rate": 3.586691054414913e-05, "loss": 26.683, "step": 192 }, { "epoch": 0.7408829174664108, "grad_norm": 16.701417922973633, "learning_rate": 3.572801521931522e-05, "loss": 25.815, "step": 193 }, { "epoch": 0.744721689059501, "grad_norm": 17.957345962524414, "learning_rate": 3.558871281669811e-05, "loss": 26.726, "step": 194 }, { "epoch": 0.7485604606525912, "grad_norm": 17.00701141357422, "learning_rate": 3.544900862216959e-05, "loss": 26.8739, "step": 195 }, { "epoch": 0.7523992322456814, "grad_norm": 19.22977638244629, "learning_rate": 3.5308907936847594e-05, "loss": 26.1091, "step": 196 }, { "epoch": 0.7562380038387716, "grad_norm": 18.498205184936523, "learning_rate": 3.516841607689501e-05, "loss": 25.4755, "step": 197 }, { "epoch": 0.7600767754318618, "grad_norm": 18.955081939697266, "learning_rate": 3.502753837331797e-05, "loss": 26.3684, "step": 198 }, { "epoch": 0.763915547024952, "grad_norm": 18.34401512145996, "learning_rate": 3.488628017176356e-05, "loss": 28.767, "step": 199 }, { "epoch": 0.7677543186180422, "grad_norm": 18.423044204711914, "learning_rate": 3.474464683231698e-05, "loss": 27.2117, "step": 200 }, { "epoch": 0.7715930902111324, "grad_norm": 18.909847259521484, "learning_rate": 3.460264372929815e-05, "loss": 26.408, "step": 201 }, { "epoch": 0.7754318618042226, "grad_norm": 16.31786346435547, "learning_rate": 3.446027625105776e-05, "loss": 26.7065, "step": 202 }, { "epoch": 0.7792706333973128, "grad_norm": 18.030500411987305, "learning_rate": 3.431754979977285e-05, "loss": 27.028, "step": 203 }, { "epoch": 0.783109404990403, "grad_norm": 16.70123863220215, "learning_rate": 3.4174469791241806e-05, "loss": 27.5544, "step": 204 }, { "epoch": 0.7869481765834933, "grad_norm": 16.9814453125, "learning_rate": 3.403104165467883e-05, "loss": 27.5372, "step": 205 }, { "epoch": 0.7907869481765835, "grad_norm": 17.28445053100586, "learning_rate": 3.388727083250795e-05, "loss": 27.4625, "step": 206 }, { "epoch": 0.7946257197696737, "grad_norm": 17.412431716918945, "learning_rate": 3.374316278015653e-05, "loss": 27.897, "step": 207 }, { "epoch": 0.7984644913627639, "grad_norm": 16.263408660888672, "learning_rate": 3.3598722965848204e-05, "loss": 27.2817, "step": 208 }, { "epoch": 0.8023032629558541, "grad_norm": 16.9997615814209, "learning_rate": 3.345395687039543e-05, "loss": 26.4631, "step": 209 }, { "epoch": 0.8061420345489443, "grad_norm": 17.816619873046875, "learning_rate": 3.330886998699149e-05, "loss": 25.6059, "step": 210 }, { "epoch": 0.8099808061420346, "grad_norm": 17.235605239868164, "learning_rate": 3.316346782100208e-05, "loss": 27.4531, "step": 211 }, { "epoch": 0.8138195777351248, "grad_norm": 15.711729049682617, "learning_rate": 3.301775588975638e-05, "loss": 27.0936, "step": 212 }, { "epoch": 0.817658349328215, "grad_norm": 16.274784088134766, "learning_rate": 3.28717397223377e-05, "loss": 25.6661, "step": 213 }, { "epoch": 0.8214971209213052, "grad_norm": 15.97848129272461, "learning_rate": 3.272542485937369e-05, "loss": 26.772, "step": 214 }, { "epoch": 0.8253358925143954, "grad_norm": 16.191444396972656, "learning_rate": 3.257881685282609e-05, "loss": 25.8711, "step": 215 }, { "epoch": 0.8291746641074856, "grad_norm": 16.814664840698242, "learning_rate": 3.243192126578007e-05, "loss": 25.5683, "step": 216 }, { "epoch": 0.8330134357005758, "grad_norm": 15.929566383361816, "learning_rate": 3.228474367223312e-05, "loss": 27.5315, "step": 217 }, { "epoch": 0.836852207293666, "grad_norm": 16.453189849853516, "learning_rate": 3.2137289656883556e-05, "loss": 27.1406, "step": 218 }, { "epoch": 0.8406909788867563, "grad_norm": 15.205329895019531, "learning_rate": 3.19895648149186e-05, "loss": 26.2816, "step": 219 }, { "epoch": 0.8445297504798465, "grad_norm": 16.378446578979492, "learning_rate": 3.1841574751802076e-05, "loss": 26.6438, "step": 220 }, { "epoch": 0.8483685220729367, "grad_norm": 16.06751251220703, "learning_rate": 3.16933250830617e-05, "loss": 27.0177, "step": 221 }, { "epoch": 0.8522072936660269, "grad_norm": 15.032289505004883, "learning_rate": 3.1544821434076014e-05, "loss": 27.5912, "step": 222 }, { "epoch": 0.8560460652591171, "grad_norm": 16.624408721923828, "learning_rate": 3.1396069439860894e-05, "loss": 28.211, "step": 223 }, { "epoch": 0.8598848368522073, "grad_norm": 17.585723876953125, "learning_rate": 3.124707474485577e-05, "loss": 26.8424, "step": 224 }, { "epoch": 0.8637236084452975, "grad_norm": 15.791353225708008, "learning_rate": 3.109784300270943e-05, "loss": 26.7544, "step": 225 }, { "epoch": 0.8675623800383877, "grad_norm": 16.265228271484375, "learning_rate": 3.094837987606547e-05, "loss": 26.9071, "step": 226 }, { "epoch": 0.8714011516314779, "grad_norm": 15.219544410705566, "learning_rate": 3.0798691036347453e-05, "loss": 25.127, "step": 227 }, { "epoch": 0.8752399232245681, "grad_norm": 16.0206241607666, "learning_rate": 3.064878216354369e-05, "loss": 27.3355, "step": 228 }, { "epoch": 0.8790786948176583, "grad_norm": 15.717583656311035, "learning_rate": 3.049865894599172e-05, "loss": 27.7224, "step": 229 }, { "epoch": 0.8829174664107485, "grad_norm": 16.433361053466797, "learning_rate": 3.0348327080162435e-05, "loss": 27.4405, "step": 230 }, { "epoch": 0.8867562380038387, "grad_norm": 16.457277297973633, "learning_rate": 3.0197792270443982e-05, "loss": 26.5396, "step": 231 }, { "epoch": 0.8905950095969289, "grad_norm": 15.61976146697998, "learning_rate": 3.0047060228925256e-05, "loss": 27.3457, "step": 232 }, { "epoch": 0.8944337811900192, "grad_norm": 16.303308486938477, "learning_rate": 2.9896136675179176e-05, "loss": 26.2405, "step": 233 }, { "epoch": 0.8982725527831094, "grad_norm": 15.62144660949707, "learning_rate": 2.974502733604565e-05, "loss": 26.1536, "step": 234 }, { "epoch": 0.9021113243761996, "grad_norm": 16.30591583251953, "learning_rate": 2.9593737945414264e-05, "loss": 26.1949, "step": 235 }, { "epoch": 0.9059500959692899, "grad_norm": 16.510095596313477, "learning_rate": 2.9442274244006722e-05, "loss": 26.6893, "step": 236 }, { "epoch": 0.9097888675623801, "grad_norm": 16.378929138183594, "learning_rate": 2.9290641979158993e-05, "loss": 28.0321, "step": 237 }, { "epoch": 0.9136276391554703, "grad_norm": 15.976604461669922, "learning_rate": 2.913884690460325e-05, "loss": 27.17, "step": 238 }, { "epoch": 0.9174664107485605, "grad_norm": 16.334959030151367, "learning_rate": 2.8986894780249524e-05, "loss": 26.3522, "step": 239 }, { "epoch": 0.9213051823416507, "grad_norm": 16.02887725830078, "learning_rate": 2.8834791371967142e-05, "loss": 28.4084, "step": 240 }, { "epoch": 0.9251439539347409, "grad_norm": 17.058687210083008, "learning_rate": 2.868254245136594e-05, "loss": 26.801, "step": 241 }, { "epoch": 0.9289827255278311, "grad_norm": 15.894112586975098, "learning_rate": 2.8530153795577286e-05, "loss": 25.7278, "step": 242 }, { "epoch": 0.9328214971209213, "grad_norm": 16.269180297851562, "learning_rate": 2.8377631187034825e-05, "loss": 26.3214, "step": 243 }, { "epoch": 0.9366602687140115, "grad_norm": 15.76980209350586, "learning_rate": 2.8224980413255086e-05, "loss": 27.0605, "step": 244 }, { "epoch": 0.9404990403071017, "grad_norm": 15.396162986755371, "learning_rate": 2.8072207266617855e-05, "loss": 26.77, "step": 245 }, { "epoch": 0.944337811900192, "grad_norm": 16.089946746826172, "learning_rate": 2.7919317544146405e-05, "loss": 26.3307, "step": 246 }, { "epoch": 0.9481765834932822, "grad_norm": 15.087519645690918, "learning_rate": 2.776631704728752e-05, "loss": 25.4705, "step": 247 }, { "epoch": 0.9520153550863724, "grad_norm": 16.0484619140625, "learning_rate": 2.761321158169134e-05, "loss": 25.8905, "step": 248 }, { "epoch": 0.9558541266794626, "grad_norm": 16.0047607421875, "learning_rate": 2.746000695699107e-05, "loss": 26.0246, "step": 249 }, { "epoch": 0.9596928982725528, "grad_norm": 16.781606674194336, "learning_rate": 2.7306708986582553e-05, "loss": 25.6385, "step": 250 }, { "epoch": 0.963531669865643, "grad_norm": 15.778640747070312, "learning_rate": 2.7153323487403655e-05, "loss": 26.7195, "step": 251 }, { "epoch": 0.9673704414587332, "grad_norm": 17.09006118774414, "learning_rate": 2.6999856279713544e-05, "loss": 25.475, "step": 252 }, { "epoch": 0.9712092130518234, "grad_norm": 16.174684524536133, "learning_rate": 2.6846313186871853e-05, "loss": 24.9302, "step": 253 }, { "epoch": 0.9750479846449136, "grad_norm": 15.83430290222168, "learning_rate": 2.669270003511769e-05, "loss": 25.4569, "step": 254 }, { "epoch": 0.9788867562380038, "grad_norm": 16.559917449951172, "learning_rate": 2.653902265334858e-05, "loss": 26.9698, "step": 255 }, { "epoch": 0.982725527831094, "grad_norm": 15.691744804382324, "learning_rate": 2.638528687289925e-05, "loss": 25.7697, "step": 256 }, { "epoch": 0.9865642994241842, "grad_norm": 15.908937454223633, "learning_rate": 2.6231498527320425e-05, "loss": 26.0816, "step": 257 }, { "epoch": 0.9904030710172744, "grad_norm": 15.256813049316406, "learning_rate": 2.6077663452157397e-05, "loss": 27.2372, "step": 258 }, { "epoch": 0.9942418426103646, "grad_norm": 16.329204559326172, "learning_rate": 2.592378748472863e-05, "loss": 26.6932, "step": 259 }, { "epoch": 0.9980806142034548, "grad_norm": 15.60101318359375, "learning_rate": 2.5769876463904265e-05, "loss": 26.632, "step": 260 }, { "epoch": 0.9980806142034548, "eval_loss": 1.698987364768982, "eval_runtime": 31.0703, "eval_samples_per_second": 9.527, "eval_steps_per_second": 4.763, "step": 260 }, { "epoch": 1.0, "grad_norm": 10.983232498168945, "learning_rate": 2.5615936229884557e-05, "loss": 12.2146, "step": 261 }, { "epoch": 1.0038387715930903, "grad_norm": 31.90011978149414, "learning_rate": 2.5461972623978247e-05, "loss": 20.352, "step": 262 }, { "epoch": 1.0076775431861804, "grad_norm": 28.391979217529297, "learning_rate": 2.530799148838096e-05, "loss": 21.9385, "step": 263 }, { "epoch": 1.0115163147792707, "grad_norm": 23.981868743896484, "learning_rate": 2.5153998665953473e-05, "loss": 21.5485, "step": 264 }, { "epoch": 1.0153550863723608, "grad_norm": 35.79429244995117, "learning_rate": 2.5e-05, "loss": 21.8402, "step": 265 }, { "epoch": 1.0191938579654511, "grad_norm": 38.60832214355469, "learning_rate": 2.4846001334046537e-05, "loss": 22.1631, "step": 266 }, { "epoch": 1.0230326295585412, "grad_norm": 28.950275421142578, "learning_rate": 2.4692008511619042e-05, "loss": 22.083, "step": 267 }, { "epoch": 1.0268714011516316, "grad_norm": 24.666053771972656, "learning_rate": 2.453802737602176e-05, "loss": 22.934, "step": 268 }, { "epoch": 1.0307101727447217, "grad_norm": 27.369165420532227, "learning_rate": 2.4384063770115455e-05, "loss": 20.742, "step": 269 }, { "epoch": 1.034548944337812, "grad_norm": 22.31073570251465, "learning_rate": 2.4230123536095748e-05, "loss": 20.5333, "step": 270 }, { "epoch": 1.038387715930902, "grad_norm": 21.734153747558594, "learning_rate": 2.4076212515271383e-05, "loss": 20.3299, "step": 271 }, { "epoch": 1.0422264875239924, "grad_norm": 22.82332992553711, "learning_rate": 2.3922336547842616e-05, "loss": 20.2011, "step": 272 }, { "epoch": 1.0460652591170825, "grad_norm": 21.255210876464844, "learning_rate": 2.376850147267958e-05, "loss": 21.1856, "step": 273 }, { "epoch": 1.0499040307101728, "grad_norm": 19.702167510986328, "learning_rate": 2.361471312710075e-05, "loss": 21.5977, "step": 274 }, { "epoch": 1.053742802303263, "grad_norm": 19.20767593383789, "learning_rate": 2.346097734665143e-05, "loss": 20.1011, "step": 275 }, { "epoch": 1.0575815738963532, "grad_norm": 21.382646560668945, "learning_rate": 2.3307299964882315e-05, "loss": 20.8318, "step": 276 }, { "epoch": 1.0614203454894433, "grad_norm": 20.11638641357422, "learning_rate": 2.3153686813128153e-05, "loss": 23.0071, "step": 277 }, { "epoch": 1.0652591170825336, "grad_norm": 17.689075469970703, "learning_rate": 2.3000143720286462e-05, "loss": 21.3526, "step": 278 }, { "epoch": 1.0690978886756237, "grad_norm": 19.41102409362793, "learning_rate": 2.2846676512596354e-05, "loss": 20.1244, "step": 279 }, { "epoch": 1.072936660268714, "grad_norm": 19.403839111328125, "learning_rate": 2.2693291013417453e-05, "loss": 21.4532, "step": 280 }, { "epoch": 1.0767754318618041, "grad_norm": 17.783912658691406, "learning_rate": 2.2539993043008937e-05, "loss": 20.0096, "step": 281 }, { "epoch": 1.0806142034548945, "grad_norm": 18.40188980102539, "learning_rate": 2.238678841830867e-05, "loss": 21.923, "step": 282 }, { "epoch": 1.0844529750479845, "grad_norm": 16.70290756225586, "learning_rate": 2.2233682952712485e-05, "loss": 22.0604, "step": 283 }, { "epoch": 1.0882917466410749, "grad_norm": 17.57340431213379, "learning_rate": 2.2080682455853597e-05, "loss": 20.0495, "step": 284 }, { "epoch": 1.092130518234165, "grad_norm": 17.793813705444336, "learning_rate": 2.192779273338215e-05, "loss": 19.6591, "step": 285 }, { "epoch": 1.0959692898272553, "grad_norm": 17.183975219726562, "learning_rate": 2.1775019586744923e-05, "loss": 21.4261, "step": 286 }, { "epoch": 1.0998080614203456, "grad_norm": 17.502321243286133, "learning_rate": 2.162236881296518e-05, "loss": 20.4338, "step": 287 }, { "epoch": 1.1036468330134357, "grad_norm": 16.18671226501465, "learning_rate": 2.1469846204422723e-05, "loss": 20.8599, "step": 288 }, { "epoch": 1.107485604606526, "grad_norm": 16.251554489135742, "learning_rate": 2.1317457548634058e-05, "loss": 21.8091, "step": 289 }, { "epoch": 1.111324376199616, "grad_norm": 16.855628967285156, "learning_rate": 2.116520862803286e-05, "loss": 22.1352, "step": 290 }, { "epoch": 1.1151631477927064, "grad_norm": 17.795156478881836, "learning_rate": 2.1013105219750478e-05, "loss": 21.5766, "step": 291 }, { "epoch": 1.1190019193857965, "grad_norm": 17.897003173828125, "learning_rate": 2.0861153095396748e-05, "loss": 22.034, "step": 292 }, { "epoch": 1.1228406909788868, "grad_norm": 17.152738571166992, "learning_rate": 2.070935802084101e-05, "loss": 21.4751, "step": 293 }, { "epoch": 1.126679462571977, "grad_norm": 18.294397354125977, "learning_rate": 2.0557725755993283e-05, "loss": 21.8068, "step": 294 }, { "epoch": 1.1305182341650672, "grad_norm": 16.99103355407715, "learning_rate": 2.0406262054585738e-05, "loss": 20.2328, "step": 295 }, { "epoch": 1.1343570057581573, "grad_norm": 17.327777862548828, "learning_rate": 2.0254972663954357e-05, "loss": 21.6014, "step": 296 }, { "epoch": 1.1381957773512477, "grad_norm": 16.88957405090332, "learning_rate": 2.0103863324820833e-05, "loss": 20.027, "step": 297 }, { "epoch": 1.1420345489443378, "grad_norm": 17.30256462097168, "learning_rate": 1.995293977107475e-05, "loss": 21.1921, "step": 298 }, { "epoch": 1.145873320537428, "grad_norm": 16.07200050354004, "learning_rate": 1.980220772955602e-05, "loss": 21.384, "step": 299 }, { "epoch": 1.1497120921305182, "grad_norm": 16.85907745361328, "learning_rate": 1.965167291983757e-05, "loss": 21.8715, "step": 300 }, { "epoch": 1.1535508637236085, "grad_norm": 17.825361251831055, "learning_rate": 1.9501341054008292e-05, "loss": 21.3469, "step": 301 }, { "epoch": 1.1573896353166986, "grad_norm": 17.06654930114746, "learning_rate": 1.9351217836456313e-05, "loss": 21.4779, "step": 302 }, { "epoch": 1.161228406909789, "grad_norm": 17.776653289794922, "learning_rate": 1.9201308963652552e-05, "loss": 19.489, "step": 303 }, { "epoch": 1.165067178502879, "grad_norm": 16.5079402923584, "learning_rate": 1.9051620123934537e-05, "loss": 20.2179, "step": 304 }, { "epoch": 1.1689059500959693, "grad_norm": 18.33095932006836, "learning_rate": 1.890215699729057e-05, "loss": 20.6006, "step": 305 }, { "epoch": 1.1727447216890594, "grad_norm": 16.471492767333984, "learning_rate": 1.8752925255144226e-05, "loss": 20.8791, "step": 306 }, { "epoch": 1.1765834932821497, "grad_norm": 16.422468185424805, "learning_rate": 1.860393056013911e-05, "loss": 20.9214, "step": 307 }, { "epoch": 1.18042226487524, "grad_norm": 16.32746696472168, "learning_rate": 1.8455178565923992e-05, "loss": 18.8008, "step": 308 }, { "epoch": 1.1842610364683301, "grad_norm": 15.671878814697266, "learning_rate": 1.8306674916938303e-05, "loss": 21.7142, "step": 309 }, { "epoch": 1.1880998080614202, "grad_norm": 17.514930725097656, "learning_rate": 1.815842524819793e-05, "loss": 20.0245, "step": 310 }, { "epoch": 1.1919385796545106, "grad_norm": 17.478960037231445, "learning_rate": 1.8010435185081404e-05, "loss": 20.6899, "step": 311 }, { "epoch": 1.1957773512476009, "grad_norm": 17.39609718322754, "learning_rate": 1.786271034311645e-05, "loss": 22.0616, "step": 312 }, { "epoch": 1.199616122840691, "grad_norm": 17.008588790893555, "learning_rate": 1.7715256327766886e-05, "loss": 20.1968, "step": 313 }, { "epoch": 1.2034548944337813, "grad_norm": 18.373451232910156, "learning_rate": 1.7568078734219934e-05, "loss": 21.7361, "step": 314 }, { "epoch": 1.2072936660268714, "grad_norm": 17.162620544433594, "learning_rate": 1.7421183147173915e-05, "loss": 20.7115, "step": 315 }, { "epoch": 1.2111324376199617, "grad_norm": 17.564990997314453, "learning_rate": 1.7274575140626318e-05, "loss": 20.3664, "step": 316 }, { "epoch": 1.2149712092130518, "grad_norm": 16.083980560302734, "learning_rate": 1.7128260277662307e-05, "loss": 21.2208, "step": 317 }, { "epoch": 1.2188099808061421, "grad_norm": 16.320255279541016, "learning_rate": 1.6982244110243627e-05, "loss": 19.6076, "step": 318 }, { "epoch": 1.2226487523992322, "grad_norm": 16.825092315673828, "learning_rate": 1.683653217899792e-05, "loss": 22.2103, "step": 319 }, { "epoch": 1.2264875239923225, "grad_norm": 16.72588539123535, "learning_rate": 1.6691130013008514e-05, "loss": 20.7865, "step": 320 }, { "epoch": 1.2303262955854126, "grad_norm": 16.987138748168945, "learning_rate": 1.654604312960457e-05, "loss": 20.3091, "step": 321 }, { "epoch": 1.234165067178503, "grad_norm": 17.015310287475586, "learning_rate": 1.6401277034151798e-05, "loss": 22.1052, "step": 322 }, { "epoch": 1.238003838771593, "grad_norm": 18.03606605529785, "learning_rate": 1.6256837219843472e-05, "loss": 21.0887, "step": 323 }, { "epoch": 1.2418426103646834, "grad_norm": 18.330631256103516, "learning_rate": 1.611272916749205e-05, "loss": 19.8342, "step": 324 }, { "epoch": 1.2456813819577734, "grad_norm": 17.45516586303711, "learning_rate": 1.5968958345321178e-05, "loss": 21.2693, "step": 325 }, { "epoch": 1.2495201535508638, "grad_norm": 16.1998348236084, "learning_rate": 1.58255302087582e-05, "loss": 20.9778, "step": 326 }, { "epoch": 1.2533589251439539, "grad_norm": 17.573469161987305, "learning_rate": 1.568245020022715e-05, "loss": 21.1408, "step": 327 }, { "epoch": 1.2571976967370442, "grad_norm": 17.92475700378418, "learning_rate": 1.5539723748942245e-05, "loss": 20.7829, "step": 328 }, { "epoch": 1.2610364683301343, "grad_norm": 17.939800262451172, "learning_rate": 1.5397356270701856e-05, "loss": 20.7192, "step": 329 }, { "epoch": 1.2648752399232246, "grad_norm": 16.325746536254883, "learning_rate": 1.5255353167683017e-05, "loss": 20.8419, "step": 330 }, { "epoch": 1.2687140115163147, "grad_norm": 16.557632446289062, "learning_rate": 1.5113719828236438e-05, "loss": 22.6103, "step": 331 }, { "epoch": 1.272552783109405, "grad_norm": 22.712125778198242, "learning_rate": 1.4972461626682033e-05, "loss": 20.6929, "step": 332 }, { "epoch": 1.2763915547024953, "grad_norm": 16.070537567138672, "learning_rate": 1.4831583923104999e-05, "loss": 20.8418, "step": 333 }, { "epoch": 1.2802303262955854, "grad_norm": 16.767885208129883, "learning_rate": 1.4691092063152417e-05, "loss": 22.7305, "step": 334 }, { "epoch": 1.2840690978886755, "grad_norm": 16.39026641845703, "learning_rate": 1.4550991377830426e-05, "loss": 22.0039, "step": 335 }, { "epoch": 1.2879078694817658, "grad_norm": 16.63512420654297, "learning_rate": 1.4411287183301902e-05, "loss": 21.4299, "step": 336 }, { "epoch": 1.2917466410748562, "grad_norm": 17.900527954101562, "learning_rate": 1.4271984780684778e-05, "loss": 22.4775, "step": 337 }, { "epoch": 1.2955854126679462, "grad_norm": 17.05221939086914, "learning_rate": 1.4133089455850879e-05, "loss": 21.9468, "step": 338 }, { "epoch": 1.2994241842610363, "grad_norm": 17.210445404052734, "learning_rate": 1.399460647922532e-05, "loss": 20.1815, "step": 339 }, { "epoch": 1.3032629558541267, "grad_norm": 16.748929977416992, "learning_rate": 1.3856541105586545e-05, "loss": 22.235, "step": 340 }, { "epoch": 1.307101727447217, "grad_norm": 15.610675811767578, "learning_rate": 1.371889857386693e-05, "loss": 20.9521, "step": 341 }, { "epoch": 1.310940499040307, "grad_norm": 17.04247283935547, "learning_rate": 1.3581684106953987e-05, "loss": 20.1059, "step": 342 }, { "epoch": 1.3147792706333974, "grad_norm": 16.33750343322754, "learning_rate": 1.3444902911492174e-05, "loss": 22.5783, "step": 343 }, { "epoch": 1.3186180422264875, "grad_norm": 16.650222778320312, "learning_rate": 1.3308560177685334e-05, "loss": 20.9683, "step": 344 }, { "epoch": 1.3224568138195778, "grad_norm": 16.820764541625977, "learning_rate": 1.3172661079099752e-05, "loss": 19.93, "step": 345 }, { "epoch": 1.326295585412668, "grad_norm": 17.346710205078125, "learning_rate": 1.303721077246784e-05, "loss": 20.52, "step": 346 }, { "epoch": 1.3301343570057582, "grad_norm": 17.263507843017578, "learning_rate": 1.2902214397492462e-05, "loss": 21.0038, "step": 347 }, { "epoch": 1.3339731285988483, "grad_norm": 17.55599594116211, "learning_rate": 1.2767677076651913e-05, "loss": 20.0553, "step": 348 }, { "epoch": 1.3378119001919386, "grad_norm": 16.44485092163086, "learning_rate": 1.2633603915005534e-05, "loss": 19.7822, "step": 349 }, { "epoch": 1.3416506717850287, "grad_norm": 16.858556747436523, "learning_rate": 1.2500000000000006e-05, "loss": 19.2506, "step": 350 }, { "epoch": 1.345489443378119, "grad_norm": 16.211315155029297, "learning_rate": 1.2366870401276304e-05, "loss": 21.2656, "step": 351 }, { "epoch": 1.3493282149712091, "grad_norm": 16.584392547607422, "learning_rate": 1.223422017047733e-05, "loss": 20.7353, "step": 352 }, { "epoch": 1.3531669865642995, "grad_norm": 16.187578201293945, "learning_rate": 1.210205434105622e-05, "loss": 21.1905, "step": 353 }, { "epoch": 1.3570057581573896, "grad_norm": 16.611042022705078, "learning_rate": 1.1970377928085371e-05, "loss": 21.2483, "step": 354 }, { "epoch": 1.3608445297504799, "grad_norm": 16.303922653198242, "learning_rate": 1.1839195928066102e-05, "loss": 20.2743, "step": 355 }, { "epoch": 1.36468330134357, "grad_norm": 16.44252586364746, "learning_rate": 1.1708513318739095e-05, "loss": 21.34, "step": 356 }, { "epoch": 1.3685220729366603, "grad_norm": 16.870229721069336, "learning_rate": 1.1578335058895498e-05, "loss": 20.5273, "step": 357 }, { "epoch": 1.3723608445297506, "grad_norm": 17.659400939941406, "learning_rate": 1.1448666088188764e-05, "loss": 19.7263, "step": 358 }, { "epoch": 1.3761996161228407, "grad_norm": 16.4289493560791, "learning_rate": 1.1319511326947222e-05, "loss": 21.7212, "step": 359 }, { "epoch": 1.3800383877159308, "grad_norm": 15.762267112731934, "learning_rate": 1.1190875675987356e-05, "loss": 20.5204, "step": 360 }, { "epoch": 1.383877159309021, "grad_norm": 16.620811462402344, "learning_rate": 1.1062764016427862e-05, "loss": 20.9438, "step": 361 }, { "epoch": 1.3877159309021114, "grad_norm": 15.502689361572266, "learning_rate": 1.0935181209504422e-05, "loss": 20.7568, "step": 362 }, { "epoch": 1.3915547024952015, "grad_norm": 16.41027069091797, "learning_rate": 1.0808132096385248e-05, "loss": 20.323, "step": 363 }, { "epoch": 1.3953934740882916, "grad_norm": 17.651681900024414, "learning_rate": 1.068162149798737e-05, "loss": 20.1112, "step": 364 }, { "epoch": 1.399232245681382, "grad_norm": 17.555715560913086, "learning_rate": 1.0555654214793722e-05, "loss": 22.2747, "step": 365 }, { "epoch": 1.4030710172744723, "grad_norm": 17.43577766418457, "learning_rate": 1.0430235026670978e-05, "loss": 21.3942, "step": 366 }, { "epoch": 1.4069097888675623, "grad_norm": 16.61440658569336, "learning_rate": 1.0305368692688174e-05, "loss": 22.5395, "step": 367 }, { "epoch": 1.4107485604606527, "grad_norm": 15.602707862854004, "learning_rate": 1.018105995093613e-05, "loss": 21.2575, "step": 368 }, { "epoch": 1.4145873320537428, "grad_norm": 17.061851501464844, "learning_rate": 1.005731351834766e-05, "loss": 21.1615, "step": 369 }, { "epoch": 1.418426103646833, "grad_norm": 16.832231521606445, "learning_rate": 9.934134090518593e-06, "loss": 21.3402, "step": 370 }, { "epoch": 1.4222648752399232, "grad_norm": 17.012224197387695, "learning_rate": 9.811526341529582e-06, "loss": 19.6323, "step": 371 }, { "epoch": 1.4261036468330135, "grad_norm": 16.545629501342773, "learning_rate": 9.689494923768755e-06, "loss": 19.4946, "step": 372 }, { "epoch": 1.4299424184261036, "grad_norm": 16.054431915283203, "learning_rate": 9.568044467755181e-06, "loss": 21.2497, "step": 373 }, { "epoch": 1.433781190019194, "grad_norm": 16.18979835510254, "learning_rate": 9.447179581963156e-06, "loss": 22.5627, "step": 374 }, { "epoch": 1.437619961612284, "grad_norm": 16.570592880249023, "learning_rate": 9.326904852647344e-06, "loss": 19.4798, "step": 375 }, { "epoch": 1.4414587332053743, "grad_norm": 16.715749740600586, "learning_rate": 9.207224843668732e-06, "loss": 22.351, "step": 376 }, { "epoch": 1.4452975047984644, "grad_norm": 16.769977569580078, "learning_rate": 9.088144096321472e-06, "loss": 21.9362, "step": 377 }, { "epoch": 1.4491362763915547, "grad_norm": 17.325634002685547, "learning_rate": 8.969667129160547e-06, "loss": 21.8168, "step": 378 }, { "epoch": 1.452975047984645, "grad_norm": 16.272539138793945, "learning_rate": 8.851798437830323e-06, "loss": 20.2932, "step": 379 }, { "epoch": 1.4568138195777351, "grad_norm": 16.569541931152344, "learning_rate": 8.734542494893955e-06, "loss": 22.7741, "step": 380 }, { "epoch": 1.4606525911708252, "grad_norm": 16.735231399536133, "learning_rate": 8.617903749663666e-06, "loss": 20.8389, "step": 381 }, { "epoch": 1.4644913627639156, "grad_norm": 17.50290298461914, "learning_rate": 8.50188662803194e-06, "loss": 21.3456, "step": 382 }, { "epoch": 1.4683301343570059, "grad_norm": 16.45136260986328, "learning_rate": 8.386495532303556e-06, "loss": 21.3918, "step": 383 }, { "epoch": 1.472168905950096, "grad_norm": 16.945947647094727, "learning_rate": 8.271734841028553e-06, "loss": 21.3806, "step": 384 }, { "epoch": 1.476007677543186, "grad_norm": 17.403200149536133, "learning_rate": 8.15760890883607e-06, "loss": 21.5741, "step": 385 }, { "epoch": 1.4798464491362764, "grad_norm": 15.925960540771484, "learning_rate": 8.044122066269149e-06, "loss": 20.3568, "step": 386 }, { "epoch": 1.4836852207293667, "grad_norm": 15.36002254486084, "learning_rate": 7.931278619620364e-06, "loss": 21.4072, "step": 387 }, { "epoch": 1.4875239923224568, "grad_norm": 18.0460205078125, "learning_rate": 7.819082850768434e-06, "loss": 21.0651, "step": 388 }, { "epoch": 1.491362763915547, "grad_norm": 16.68878746032715, "learning_rate": 7.70753901701575e-06, "loss": 19.6786, "step": 389 }, { "epoch": 1.4952015355086372, "grad_norm": 16.785842895507812, "learning_rate": 7.5966513509268365e-06, "loss": 21.9675, "step": 390 }, { "epoch": 1.4952015355086372, "eval_loss": 1.7116776704788208, "eval_runtime": 31.0248, "eval_samples_per_second": 9.541, "eval_steps_per_second": 4.77, "step": 390 }, { "epoch": 1.4990403071017275, "grad_norm": 16.459524154663086, "learning_rate": 7.4864240601677265e-06, "loss": 19.1034, "step": 391 }, { "epoch": 1.5028790786948176, "grad_norm": 16.578235626220703, "learning_rate": 7.3768613273463234e-06, "loss": 20.6083, "step": 392 }, { "epoch": 1.5067178502879077, "grad_norm": 15.791862487792969, "learning_rate": 7.267967309853666e-06, "loss": 20.8469, "step": 393 }, { "epoch": 1.510556621880998, "grad_norm": 16.680150985717773, "learning_rate": 7.159746139706194e-06, "loss": 21.1103, "step": 394 }, { "epoch": 1.5143953934740884, "grad_norm": 15.956221580505371, "learning_rate": 7.0522019233889545e-06, "loss": 21.1805, "step": 395 }, { "epoch": 1.5182341650671785, "grad_norm": 16.348682403564453, "learning_rate": 6.9453387416997695e-06, "loss": 20.605, "step": 396 }, { "epoch": 1.5220729366602685, "grad_norm": 16.562957763671875, "learning_rate": 6.839160649594401e-06, "loss": 21.1429, "step": 397 }, { "epoch": 1.5259117082533589, "grad_norm": 15.684383392333984, "learning_rate": 6.733671676032674e-06, "loss": 23.0287, "step": 398 }, { "epoch": 1.5297504798464492, "grad_norm": 16.632112503051758, "learning_rate": 6.628875823825612e-06, "loss": 20.5922, "step": 399 }, { "epoch": 1.5335892514395395, "grad_norm": 15.957228660583496, "learning_rate": 6.524777069483526e-06, "loss": 20.5221, "step": 400 }, { "epoch": 1.5374280230326296, "grad_norm": 16.59423828125, "learning_rate": 6.421379363065142e-06, "loss": 21.1484, "step": 401 }, { "epoch": 1.5412667946257197, "grad_norm": 16.450971603393555, "learning_rate": 6.318686628027723e-06, "loss": 19.4515, "step": 402 }, { "epoch": 1.54510556621881, "grad_norm": 15.815667152404785, "learning_rate": 6.216702761078166e-06, "loss": 20.9243, "step": 403 }, { "epoch": 1.5489443378119003, "grad_norm": 16.43648338317871, "learning_rate": 6.115431632025154e-06, "loss": 21.8014, "step": 404 }, { "epoch": 1.5527831094049904, "grad_norm": 15.677999496459961, "learning_rate": 6.01487708363232e-06, "loss": 20.509, "step": 405 }, { "epoch": 1.5566218809980805, "grad_norm": 15.919227600097656, "learning_rate": 5.915042931472425e-06, "loss": 21.2447, "step": 406 }, { "epoch": 1.5604606525911708, "grad_norm": 15.848505973815918, "learning_rate": 5.8159329637825745e-06, "loss": 21.4095, "step": 407 }, { "epoch": 1.5642994241842612, "grad_norm": 15.660775184631348, "learning_rate": 5.717550941320482e-06, "loss": 21.0583, "step": 408 }, { "epoch": 1.5681381957773513, "grad_norm": 16.026033401489258, "learning_rate": 5.619900597221753e-06, "loss": 21.2968, "step": 409 }, { "epoch": 1.5719769673704413, "grad_norm": 16.016826629638672, "learning_rate": 5.522985636858239e-06, "loss": 20.5098, "step": 410 }, { "epoch": 1.5758157389635317, "grad_norm": 16.44532012939453, "learning_rate": 5.42680973769743e-06, "loss": 20.4221, "step": 411 }, { "epoch": 1.579654510556622, "grad_norm": 16.122182846069336, "learning_rate": 5.33137654916292e-06, "loss": 20.8718, "step": 412 }, { "epoch": 1.583493282149712, "grad_norm": 16.150543212890625, "learning_rate": 5.236689692495916e-06, "loss": 20.9968, "step": 413 }, { "epoch": 1.5873320537428022, "grad_norm": 17.00517463684082, "learning_rate": 5.14275276061785e-06, "loss": 21.4716, "step": 414 }, { "epoch": 1.5911708253358925, "grad_norm": 16.143770217895508, "learning_rate": 5.049569317994013e-06, "loss": 19.5914, "step": 415 }, { "epoch": 1.5950095969289828, "grad_norm": 15.659564018249512, "learning_rate": 4.957142900498335e-06, "loss": 22.1265, "step": 416 }, { "epoch": 1.598848368522073, "grad_norm": 15.749252319335938, "learning_rate": 4.865477015279185e-06, "loss": 20.5613, "step": 417 }, { "epoch": 1.602687140115163, "grad_norm": 16.31047248840332, "learning_rate": 4.7745751406263165e-06, "loss": 23.077, "step": 418 }, { "epoch": 1.6065259117082533, "grad_norm": 15.302318572998047, "learning_rate": 4.684440725838862e-06, "loss": 21.2777, "step": 419 }, { "epoch": 1.6103646833013436, "grad_norm": 15.675249099731445, "learning_rate": 4.5950771910944605e-06, "loss": 21.9063, "step": 420 }, { "epoch": 1.6142034548944337, "grad_norm": 16.395877838134766, "learning_rate": 4.5064879273194745e-06, "loss": 22.1439, "step": 421 }, { "epoch": 1.6180422264875238, "grad_norm": 16.09247398376465, "learning_rate": 4.4186762960603224e-06, "loss": 23.4098, "step": 422 }, { "epoch": 1.6218809980806141, "grad_norm": 16.875818252563477, "learning_rate": 4.331645629355915e-06, "loss": 19.9787, "step": 423 }, { "epoch": 1.6257197696737045, "grad_norm": 15.509851455688477, "learning_rate": 4.245399229611238e-06, "loss": 20.8531, "step": 424 }, { "epoch": 1.6295585412667948, "grad_norm": 15.853813171386719, "learning_rate": 4.159940369472015e-06, "loss": 21.1502, "step": 425 }, { "epoch": 1.6333973128598849, "grad_norm": 16.189851760864258, "learning_rate": 4.075272291700557e-06, "loss": 21.4833, "step": 426 }, { "epoch": 1.637236084452975, "grad_norm": 16.603588104248047, "learning_rate": 3.991398209052686e-06, "loss": 20.5711, "step": 427 }, { "epoch": 1.6410748560460653, "grad_norm": 16.684690475463867, "learning_rate": 3.908321304155846e-06, "loss": 20.9092, "step": 428 }, { "epoch": 1.6449136276391556, "grad_norm": 16.88503646850586, "learning_rate": 3.826044729388326e-06, "loss": 19.7428, "step": 429 }, { "epoch": 1.6487523992322457, "grad_norm": 17.10723304748535, "learning_rate": 3.7445716067596503e-06, "loss": 20.6075, "step": 430 }, { "epoch": 1.6525911708253358, "grad_norm": 15.53963565826416, "learning_rate": 3.6639050277921057e-06, "loss": 22.9, "step": 431 }, { "epoch": 1.6564299424184261, "grad_norm": 15.606730461120605, "learning_rate": 3.5840480534034355e-06, "loss": 21.504, "step": 432 }, { "epoch": 1.6602687140115164, "grad_norm": 16.915599822998047, "learning_rate": 3.505003713790689e-06, "loss": 21.3113, "step": 433 }, { "epoch": 1.6641074856046065, "grad_norm": 16.729263305664062, "learning_rate": 3.4267750083152584e-06, "loss": 21.0463, "step": 434 }, { "epoch": 1.6679462571976966, "grad_norm": 16.304168701171875, "learning_rate": 3.3493649053890326e-06, "loss": 20.9805, "step": 435 }, { "epoch": 1.671785028790787, "grad_norm": 16.270374298095703, "learning_rate": 3.2727763423617913e-06, "loss": 21.1003, "step": 436 }, { "epoch": 1.6756238003838773, "grad_norm": 15.20581340789795, "learning_rate": 3.197012225409729e-06, "loss": 20.0619, "step": 437 }, { "epoch": 1.6794625719769674, "grad_norm": 17.438676834106445, "learning_rate": 3.1220754294251837e-06, "loss": 19.7497, "step": 438 }, { "epoch": 1.6833013435700575, "grad_norm": 17.613048553466797, "learning_rate": 3.0479687979075523e-06, "loss": 21.0865, "step": 439 }, { "epoch": 1.6871401151631478, "grad_norm": 16.843353271484375, "learning_rate": 2.974695142855388e-06, "loss": 21.8551, "step": 440 }, { "epoch": 1.690978886756238, "grad_norm": 16.841970443725586, "learning_rate": 2.9022572446597007e-06, "loss": 20.6149, "step": 441 }, { "epoch": 1.6948176583493282, "grad_norm": 15.736058235168457, "learning_rate": 2.8306578519984527e-06, "loss": 19.9789, "step": 442 }, { "epoch": 1.6986564299424183, "grad_norm": 17.377676010131836, "learning_rate": 2.7598996817322613e-06, "loss": 21.6122, "step": 443 }, { "epoch": 1.7024952015355086, "grad_norm": 15.351333618164062, "learning_rate": 2.6899854188013054e-06, "loss": 21.8045, "step": 444 }, { "epoch": 1.706333973128599, "grad_norm": 15.400979042053223, "learning_rate": 2.6209177161234445e-06, "loss": 21.6742, "step": 445 }, { "epoch": 1.710172744721689, "grad_norm": 15.909041404724121, "learning_rate": 2.552699194493549e-06, "loss": 21.2103, "step": 446 }, { "epoch": 1.714011516314779, "grad_norm": 16.288795471191406, "learning_rate": 2.4853324424840674e-06, "loss": 19.6584, "step": 447 }, { "epoch": 1.7178502879078694, "grad_norm": 16.88001251220703, "learning_rate": 2.418820016346779e-06, "loss": 20.9527, "step": 448 }, { "epoch": 1.7216890595009597, "grad_norm": 15.426772117614746, "learning_rate": 2.3531644399158197e-06, "loss": 22.1142, "step": 449 }, { "epoch": 1.72552783109405, "grad_norm": 16.2529354095459, "learning_rate": 2.2883682045119063e-06, "loss": 20.6667, "step": 450 }, { "epoch": 1.7293666026871402, "grad_norm": 22.660812377929688, "learning_rate": 2.224433768847789e-06, "loss": 21.2103, "step": 451 }, { "epoch": 1.7332053742802302, "grad_norm": 16.202503204345703, "learning_rate": 2.1613635589349756e-06, "loss": 20.0658, "step": 452 }, { "epoch": 1.7370441458733206, "grad_norm": 15.147746086120605, "learning_rate": 2.0991599679916624e-06, "loss": 20.9152, "step": 453 }, { "epoch": 1.7408829174664109, "grad_norm": 15.011345863342285, "learning_rate": 2.0378253563519247e-06, "loss": 20.4116, "step": 454 }, { "epoch": 1.744721689059501, "grad_norm": 15.8588228225708, "learning_rate": 1.977362051376158e-06, "loss": 18.8413, "step": 455 }, { "epoch": 1.748560460652591, "grad_norm": 16.016305923461914, "learning_rate": 1.9177723473627646e-06, "loss": 19.489, "step": 456 }, { "epoch": 1.7523992322456814, "grad_norm": 15.62695026397705, "learning_rate": 1.859058505461095e-06, "loss": 20.4723, "step": 457 }, { "epoch": 1.7562380038387717, "grad_norm": 15.519407272338867, "learning_rate": 1.801222753585638e-06, "loss": 22.4733, "step": 458 }, { "epoch": 1.7600767754318618, "grad_norm": 15.624913215637207, "learning_rate": 1.7442672863314968e-06, "loss": 22.5163, "step": 459 }, { "epoch": 1.763915547024952, "grad_norm": 16.19761085510254, "learning_rate": 1.6881942648911076e-06, "loss": 21.959, "step": 460 }, { "epoch": 1.7677543186180422, "grad_norm": 15.693192481994629, "learning_rate": 1.6330058169722384e-06, "loss": 20.8804, "step": 461 }, { "epoch": 1.7715930902111325, "grad_norm": 15.414134979248047, "learning_rate": 1.578704036717238e-06, "loss": 19.8119, "step": 462 }, { "epoch": 1.7754318618042226, "grad_norm": 16.06474494934082, "learning_rate": 1.5252909846235896e-06, "loss": 21.2434, "step": 463 }, { "epoch": 1.7792706333973127, "grad_norm": 16.06039810180664, "learning_rate": 1.4727686874657143e-06, "loss": 21.5852, "step": 464 }, { "epoch": 1.783109404990403, "grad_norm": 17.113985061645508, "learning_rate": 1.421139138218064e-06, "loss": 21.6185, "step": 465 }, { "epoch": 1.7869481765834934, "grad_norm": 15.716484069824219, "learning_rate": 1.3704042959795132e-06, "loss": 21.6163, "step": 466 }, { "epoch": 1.7907869481765835, "grad_norm": 16.430994033813477, "learning_rate": 1.3205660858989944e-06, "loss": 22.8004, "step": 467 }, { "epoch": 1.7946257197696736, "grad_norm": 16.834121704101562, "learning_rate": 1.2716263991024713e-06, "loss": 20.1899, "step": 468 }, { "epoch": 1.7984644913627639, "grad_norm": 15.677217483520508, "learning_rate": 1.2235870926211619e-06, "loss": 18.9967, "step": 469 }, { "epoch": 1.8023032629558542, "grad_norm": 16.316518783569336, "learning_rate": 1.1764499893210878e-06, "loss": 21.6224, "step": 470 }, { "epoch": 1.8061420345489443, "grad_norm": 15.550005912780762, "learning_rate": 1.1302168778338922e-06, "loss": 20.9617, "step": 471 }, { "epoch": 1.8099808061420346, "grad_norm": 15.957982063293457, "learning_rate": 1.0848895124889818e-06, "loss": 20.4512, "step": 472 }, { "epoch": 1.8138195777351247, "grad_norm": 16.459383010864258, "learning_rate": 1.0404696132469444e-06, "loss": 21.528, "step": 473 }, { "epoch": 1.817658349328215, "grad_norm": 16.093690872192383, "learning_rate": 9.969588656342982e-07, "loss": 20.7164, "step": 474 }, { "epoch": 1.8214971209213053, "grad_norm": 16.058801651000977, "learning_rate": 9.54358920679524e-07, "loss": 19.863, "step": 475 }, { "epoch": 1.8253358925143954, "grad_norm": 16.003698348999023, "learning_rate": 9.126713948504228e-07, "loss": 20.8559, "step": 476 }, { "epoch": 1.8291746641074855, "grad_norm": 16.547693252563477, "learning_rate": 8.718978699927744e-07, "loss": 22.0172, "step": 477 }, { "epoch": 1.8330134357005758, "grad_norm": 15.8489408493042, "learning_rate": 8.320398932703144e-07, "loss": 21.21, "step": 478 }, { "epoch": 1.8368522072936662, "grad_norm": 16.281145095825195, "learning_rate": 7.930989771060299e-07, "loss": 20.4811, "step": 479 }, { "epoch": 1.8406909788867563, "grad_norm": 15.579625129699707, "learning_rate": 7.550765991247654e-07, "loss": 19.9756, "step": 480 }, { "epoch": 1.8445297504798464, "grad_norm": 15.547847747802734, "learning_rate": 7.179742020971519e-07, "loss": 20.5781, "step": 481 }, { "epoch": 1.8483685220729367, "grad_norm": 16.1280517578125, "learning_rate": 6.817931938848803e-07, "loss": 20.1387, "step": 482 }, { "epoch": 1.852207293666027, "grad_norm": 16.432701110839844, "learning_rate": 6.465349473872478e-07, "loss": 23.4471, "step": 483 }, { "epoch": 1.856046065259117, "grad_norm": 16.626161575317383, "learning_rate": 6.122008004890851e-07, "loss": 20.4011, "step": 484 }, { "epoch": 1.8598848368522072, "grad_norm": 16.168102264404297, "learning_rate": 5.78792056009983e-07, "loss": 21.4412, "step": 485 }, { "epoch": 1.8637236084452975, "grad_norm": 16.56131935119629, "learning_rate": 5.463099816548579e-07, "loss": 20.2194, "step": 486 }, { "epoch": 1.8675623800383878, "grad_norm": 17.292823791503906, "learning_rate": 5.147558099658529e-07, "loss": 21.643, "step": 487 }, { "epoch": 1.871401151631478, "grad_norm": 16.576915740966797, "learning_rate": 4.841307382755566e-07, "loss": 22.5503, "step": 488 }, { "epoch": 1.875239923224568, "grad_norm": 15.804405212402344, "learning_rate": 4.5443592866157846e-07, "loss": 20.4958, "step": 489 }, { "epoch": 1.8790786948176583, "grad_norm": 15.931843757629395, "learning_rate": 4.256725079024554e-07, "loss": 19.8143, "step": 490 }, { "epoch": 1.8829174664107486, "grad_norm": 17.453289031982422, "learning_rate": 3.978415674348895e-07, "loss": 22.0656, "step": 491 }, { "epoch": 1.8867562380038387, "grad_norm": 15.733434677124023, "learning_rate": 3.709441633123367e-07, "loss": 19.9285, "step": 492 }, { "epoch": 1.8905950095969288, "grad_norm": 15.915714263916016, "learning_rate": 3.4498131616493565e-07, "loss": 20.6875, "step": 493 }, { "epoch": 1.8944337811900192, "grad_norm": 15.583646774291992, "learning_rate": 3.1995401116077516e-07, "loss": 21.8348, "step": 494 }, { "epoch": 1.8982725527831095, "grad_norm": 17.053115844726562, "learning_rate": 2.958631979685156e-07, "loss": 20.8289, "step": 495 }, { "epoch": 1.9021113243761996, "grad_norm": 17.572505950927734, "learning_rate": 2.7270979072135104e-07, "loss": 20.5572, "step": 496 }, { "epoch": 1.9059500959692899, "grad_norm": 15.979856491088867, "learning_rate": 2.5049466798232614e-07, "loss": 20.125, "step": 497 }, { "epoch": 1.90978886756238, "grad_norm": 15.965109825134277, "learning_rate": 2.2921867271099295e-07, "loss": 21.0769, "step": 498 }, { "epoch": 1.9136276391554703, "grad_norm": 16.205610275268555, "learning_rate": 2.0888261223143135e-07, "loss": 21.3394, "step": 499 }, { "epoch": 1.9174664107485606, "grad_norm": 16.218839645385742, "learning_rate": 1.8948725820160662e-07, "loss": 21.4822, "step": 500 }, { "epoch": 1.9213051823416507, "grad_norm": 15.393239974975586, "learning_rate": 1.7103334658409577e-07, "loss": 20.9145, "step": 501 }, { "epoch": 1.9251439539347408, "grad_norm": 16.45865249633789, "learning_rate": 1.5352157761815977e-07, "loss": 20.0352, "step": 502 }, { "epoch": 1.9289827255278311, "grad_norm": 16.02602767944336, "learning_rate": 1.3695261579316777e-07, "loss": 19.5627, "step": 503 }, { "epoch": 1.9328214971209214, "grad_norm": 16.707592010498047, "learning_rate": 1.2132708982338926e-07, "loss": 21.6733, "step": 504 }, { "epoch": 1.9366602687140115, "grad_norm": 15.346840858459473, "learning_rate": 1.0664559262413831e-07, "loss": 20.6038, "step": 505 }, { "epoch": 1.9404990403071016, "grad_norm": 15.660294532775879, "learning_rate": 9.290868128926377e-08, "loss": 20.4834, "step": 506 }, { "epoch": 1.944337811900192, "grad_norm": 16.100744247436523, "learning_rate": 8.011687707002447e-08, "loss": 21.3543, "step": 507 }, { "epoch": 1.9481765834932823, "grad_norm": 16.333961486816406, "learning_rate": 6.827066535529946e-08, "loss": 19.4984, "step": 508 }, { "epoch": 1.9520153550863724, "grad_norm": 16.721385955810547, "learning_rate": 5.737049565317787e-08, "loss": 20.7062, "step": 509 }, { "epoch": 1.9558541266794625, "grad_norm": 15.953263282775879, "learning_rate": 4.741678157389739e-08, "loss": 21.8326, "step": 510 }, { "epoch": 1.9596928982725528, "grad_norm": 16.059139251708984, "learning_rate": 3.840990081415141e-08, "loss": 20.5162, "step": 511 }, { "epoch": 1.963531669865643, "grad_norm": 16.434776306152344, "learning_rate": 3.0350195142753166e-08, "loss": 21.5626, "step": 512 }, { "epoch": 1.9673704414587332, "grad_norm": 16.480600357055664, "learning_rate": 2.323797038767117e-08, "loss": 21.3467, "step": 513 }, { "epoch": 1.9712092130518233, "grad_norm": 15.328332901000977, "learning_rate": 1.7073496424427348e-08, "loss": 21.8149, "step": 514 }, { "epoch": 1.9750479846449136, "grad_norm": 16.419599533081055, "learning_rate": 1.1857007165852474e-08, "loss": 21.2042, "step": 515 }, { "epoch": 1.978886756238004, "grad_norm": 17.01874351501465, "learning_rate": 7.588700553209926e-09, "loss": 20.3651, "step": 516 }, { "epoch": 1.982725527831094, "grad_norm": 16.0975284576416, "learning_rate": 4.268738548682261e-09, "loss": 20.8112, "step": 517 }, { "epoch": 1.986564299424184, "grad_norm": 15.759478569030762, "learning_rate": 1.8972471292344473e-09, "loss": 21.6873, "step": 518 }, { "epoch": 1.9904030710172744, "grad_norm": 15.517256736755371, "learning_rate": 4.743162818288127e-10, "loss": 22.6064, "step": 519 }, { "epoch": 1.9942418426103647, "grad_norm": 16.582988739013672, "learning_rate": 0.0, "loss": 21.648, "step": 520 }, { "epoch": 1.9942418426103647, "eval_loss": 1.710574746131897, "eval_runtime": 31.6354, "eval_samples_per_second": 9.357, "eval_steps_per_second": 4.678, "step": 520 } ], "logging_steps": 1, "max_steps": 520, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 260, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.987149504299532e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }