{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.986078886310905, "eval_steps": 500, "global_step": 107, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02, "grad_norm": 28.83467674255371, "learning_rate": 2e-05, "loss": 2.7878, "step": 1 }, { "epoch": 0.04, "grad_norm": 28.955718994140625, "learning_rate": 4e-05, "loss": 2.7826, "step": 2 }, { "epoch": 0.06, "grad_norm": 17.856950759887695, "learning_rate": 6e-05, "loss": 2.2388, "step": 3 }, { "epoch": 0.07, "grad_norm": 16.615158081054688, "learning_rate": 8e-05, "loss": 1.7809, "step": 4 }, { "epoch": 0.09, "grad_norm": 34.79571533203125, "learning_rate": 0.0001, "loss": 1.9714, "step": 5 }, { "epoch": 0.11, "grad_norm": 6.3113694190979, "learning_rate": 9.935064935064936e-05, "loss": 1.4032, "step": 6 }, { "epoch": 0.13, "grad_norm": 8.230330467224121, "learning_rate": 9.870129870129871e-05, "loss": 1.241, "step": 7 }, { "epoch": 0.15, "grad_norm": 17.19231605529785, "learning_rate": 9.805194805194806e-05, "loss": 1.2388, "step": 8 }, { "epoch": 0.17, "grad_norm": 6.900284767150879, "learning_rate": 9.74025974025974e-05, "loss": 1.0881, "step": 9 }, { "epoch": 0.19, "grad_norm": 2.7498185634613037, "learning_rate": 9.675324675324677e-05, "loss": 1.0071, "step": 10 }, { "epoch": 0.2, "grad_norm": 4.016067028045654, "learning_rate": 9.610389610389611e-05, "loss": 0.9962, "step": 11 }, { "epoch": 0.22, "grad_norm": 2.365187883377075, "learning_rate": 9.545454545454546e-05, "loss": 0.9928, "step": 12 }, { "epoch": 0.24, "grad_norm": 1.4348371028900146, "learning_rate": 9.480519480519481e-05, "loss": 0.9107, "step": 13 }, { "epoch": 0.26, "grad_norm": 1.8148932456970215, "learning_rate": 9.415584415584417e-05, "loss": 0.853, "step": 14 }, { "epoch": 0.28, "grad_norm": 1.9855575561523438, "learning_rate": 9.35064935064935e-05, "loss": 0.8598, "step": 15 }, { "epoch": 0.3, "grad_norm": 1.9960970878601074, "learning_rate": 9.285714285714286e-05, "loss": 0.8384, "step": 16 }, { "epoch": 0.32, "grad_norm": 1.4565762281417847, "learning_rate": 9.220779220779221e-05, "loss": 0.7895, "step": 17 }, { "epoch": 0.33, "grad_norm": 1.419858694076538, "learning_rate": 9.155844155844156e-05, "loss": 0.8162, "step": 18 }, { "epoch": 0.35, "grad_norm": 1.3651608228683472, "learning_rate": 9.090909090909092e-05, "loss": 0.7924, "step": 19 }, { "epoch": 0.37, "grad_norm": 1.1524626016616821, "learning_rate": 9.025974025974027e-05, "loss": 0.7484, "step": 20 }, { "epoch": 0.39, "grad_norm": 0.9174069166183472, "learning_rate": 8.961038961038961e-05, "loss": 0.7276, "step": 21 }, { "epoch": 0.41, "grad_norm": 1.1832919120788574, "learning_rate": 8.896103896103896e-05, "loss": 0.6658, "step": 22 }, { "epoch": 0.43, "grad_norm": 1.3024309873580933, "learning_rate": 8.831168831168831e-05, "loss": 0.7379, "step": 23 }, { "epoch": 0.45, "grad_norm": 1.1416062116622925, "learning_rate": 8.766233766233767e-05, "loss": 0.7202, "step": 24 }, { "epoch": 0.46, "grad_norm": 1.2058277130126953, "learning_rate": 8.701298701298701e-05, "loss": 0.6777, "step": 25 }, { "epoch": 0.48, "grad_norm": 1.0915583372116089, "learning_rate": 8.636363636363637e-05, "loss": 0.7063, "step": 26 }, { "epoch": 0.5, "grad_norm": 1.0384303331375122, "learning_rate": 8.571428571428571e-05, "loss": 0.6447, "step": 27 }, { "epoch": 0.52, "grad_norm": 1.131259560585022, "learning_rate": 8.506493506493507e-05, "loss": 0.6911, "step": 28 }, { "epoch": 0.54, "grad_norm": 1.1505099534988403, "learning_rate": 8.441558441558442e-05, "loss": 0.6142, "step": 29 }, { "epoch": 0.56, "grad_norm": 2.6675026416778564, "learning_rate": 8.376623376623377e-05, "loss": 0.6181, "step": 30 }, { "epoch": 0.58, "grad_norm": 1.2021816968917847, "learning_rate": 8.311688311688312e-05, "loss": 0.6542, "step": 31 }, { "epoch": 0.59, "grad_norm": 3.363269090652466, "learning_rate": 8.246753246753248e-05, "loss": 0.5968, "step": 32 }, { "epoch": 0.61, "grad_norm": 3.8007972240448, "learning_rate": 8.181818181818183e-05, "loss": 0.5983, "step": 33 }, { "epoch": 0.63, "grad_norm": 3.570556163787842, "learning_rate": 8.116883116883117e-05, "loss": 0.6375, "step": 34 }, { "epoch": 0.65, "grad_norm": 4.428549766540527, "learning_rate": 8.051948051948052e-05, "loss": 0.6519, "step": 35 }, { "epoch": 0.67, "grad_norm": 3.7240254878997803, "learning_rate": 7.987012987012987e-05, "loss": 0.5935, "step": 36 }, { "epoch": 0.69, "grad_norm": 2.453350782394409, "learning_rate": 7.922077922077923e-05, "loss": 0.604, "step": 37 }, { "epoch": 0.71, "grad_norm": 3.341071844100952, "learning_rate": 7.857142857142858e-05, "loss": 0.5994, "step": 38 }, { "epoch": 0.72, "grad_norm": 3.4222161769866943, "learning_rate": 7.792207792207793e-05, "loss": 0.5731, "step": 39 }, { "epoch": 0.74, "grad_norm": 2.9063804149627686, "learning_rate": 7.727272727272727e-05, "loss": 0.5566, "step": 40 }, { "epoch": 0.76, "grad_norm": 3.170330762863159, "learning_rate": 7.662337662337662e-05, "loss": 0.5684, "step": 41 }, { "epoch": 0.78, "grad_norm": 2.8363170623779297, "learning_rate": 7.597402597402598e-05, "loss": 0.5936, "step": 42 }, { "epoch": 0.8, "grad_norm": 2.3944592475891113, "learning_rate": 7.532467532467533e-05, "loss": 0.5571, "step": 43 }, { "epoch": 0.82, "grad_norm": 2.816237688064575, "learning_rate": 7.467532467532467e-05, "loss": 0.5598, "step": 44 }, { "epoch": 0.84, "grad_norm": 2.57438325881958, "learning_rate": 7.402597402597404e-05, "loss": 0.5615, "step": 45 }, { "epoch": 0.85, "grad_norm": 2.3164243698120117, "learning_rate": 7.337662337662338e-05, "loss": 0.5381, "step": 46 }, { "epoch": 0.87, "grad_norm": 2.4226622581481934, "learning_rate": 7.272727272727273e-05, "loss": 0.5592, "step": 47 }, { "epoch": 0.89, "grad_norm": 1.5568475723266602, "learning_rate": 7.207792207792208e-05, "loss": 0.5729, "step": 48 }, { "epoch": 0.91, "grad_norm": 1.1320440769195557, "learning_rate": 7.142857142857143e-05, "loss": 0.5546, "step": 49 }, { "epoch": 0.93, "grad_norm": 2.220273733139038, "learning_rate": 7.077922077922077e-05, "loss": 0.5587, "step": 50 }, { "epoch": 0.95, "grad_norm": 1.8099664449691772, "learning_rate": 7.012987012987014e-05, "loss": 0.5403, "step": 51 }, { "epoch": 0.97, "grad_norm": 1.7992054224014282, "learning_rate": 6.948051948051948e-05, "loss": 0.5374, "step": 52 }, { "epoch": 0.98, "grad_norm": 1.5891600847244263, "learning_rate": 6.883116883116883e-05, "loss": 0.5188, "step": 53 }, { "epoch": 1.0, "grad_norm": 1.713174819946289, "learning_rate": 6.818181818181818e-05, "loss": 0.5632, "step": 54 }, { "epoch": 1.02, "grad_norm": 1.529587984085083, "learning_rate": 6.753246753246754e-05, "loss": 0.4607, "step": 55 }, { "epoch": 1.04, "grad_norm": 1.243905782699585, "learning_rate": 6.688311688311688e-05, "loss": 0.4807, "step": 56 }, { "epoch": 1.06, "grad_norm": 0.92779141664505, "learning_rate": 6.623376623376624e-05, "loss": 0.4515, "step": 57 }, { "epoch": 1.08, "grad_norm": 1.8945612907409668, "learning_rate": 6.55844155844156e-05, "loss": 0.4788, "step": 58 }, { "epoch": 1.1, "grad_norm": 1.6997075080871582, "learning_rate": 6.493506493506494e-05, "loss": 0.4411, "step": 59 }, { "epoch": 1.11, "grad_norm": 1.4320181608200073, "learning_rate": 6.428571428571429e-05, "loss": 0.4474, "step": 60 }, { "epoch": 1.13, "grad_norm": 1.5225821733474731, "learning_rate": 6.363636363636364e-05, "loss": 0.4578, "step": 61 }, { "epoch": 1.15, "grad_norm": 1.4867883920669556, "learning_rate": 6.2987012987013e-05, "loss": 0.4838, "step": 62 }, { "epoch": 1.17, "grad_norm": 1.3935158252716064, "learning_rate": 6.233766233766233e-05, "loss": 0.4383, "step": 63 }, { "epoch": 1.19, "grad_norm": 1.3193836212158203, "learning_rate": 6.16883116883117e-05, "loss": 0.4369, "step": 64 }, { "epoch": 1.21, "grad_norm": 1.0565460920333862, "learning_rate": 6.103896103896104e-05, "loss": 0.4618, "step": 65 }, { "epoch": 1.23, "grad_norm": 1.9069595336914062, "learning_rate": 6.03896103896104e-05, "loss": 0.4636, "step": 66 }, { "epoch": 1.24, "grad_norm": 1.592588186264038, "learning_rate": 5.9740259740259744e-05, "loss": 0.4363, "step": 67 }, { "epoch": 1.26, "grad_norm": 1.392088770866394, "learning_rate": 5.90909090909091e-05, "loss": 0.4573, "step": 68 }, { "epoch": 1.28, "grad_norm": 1.4312165975570679, "learning_rate": 5.844155844155844e-05, "loss": 0.4404, "step": 69 }, { "epoch": 1.3, "grad_norm": 1.4288945198059082, "learning_rate": 5.7792207792207796e-05, "loss": 0.4517, "step": 70 }, { "epoch": 1.32, "grad_norm": 1.247237205505371, "learning_rate": 5.714285714285714e-05, "loss": 0.4731, "step": 71 }, { "epoch": 1.34, "grad_norm": 1.4292420148849487, "learning_rate": 5.64935064935065e-05, "loss": 0.4561, "step": 72 }, { "epoch": 1.35, "grad_norm": 1.1647123098373413, "learning_rate": 5.584415584415584e-05, "loss": 0.4702, "step": 73 }, { "epoch": 1.37, "grad_norm": 1.6123448610305786, "learning_rate": 5.51948051948052e-05, "loss": 0.4681, "step": 74 }, { "epoch": 1.39, "grad_norm": 1.3807860612869263, "learning_rate": 5.4545454545454546e-05, "loss": 0.4217, "step": 75 }, { "epoch": 1.41, "grad_norm": 1.4928926229476929, "learning_rate": 5.38961038961039e-05, "loss": 0.4344, "step": 76 }, { "epoch": 1.43, "grad_norm": 1.3118643760681152, "learning_rate": 5.3246753246753245e-05, "loss": 0.4509, "step": 77 }, { "epoch": 1.45, "grad_norm": 1.1260207891464233, "learning_rate": 5.25974025974026e-05, "loss": 0.4423, "step": 78 }, { "epoch": 1.47, "grad_norm": 1.1223907470703125, "learning_rate": 5.1948051948051944e-05, "loss": 0.4238, "step": 79 }, { "epoch": 1.48, "grad_norm": 1.123289704322815, "learning_rate": 5.1298701298701304e-05, "loss": 0.4415, "step": 80 }, { "epoch": 1.5, "grad_norm": 0.8493527770042419, "learning_rate": 5.064935064935065e-05, "loss": 0.4278, "step": 81 }, { "epoch": 1.52, "grad_norm": 1.3558621406555176, "learning_rate": 5e-05, "loss": 0.4326, "step": 82 }, { "epoch": 1.54, "grad_norm": 1.198364496231079, "learning_rate": 4.9350649350649355e-05, "loss": 0.4504, "step": 83 }, { "epoch": 1.56, "grad_norm": 1.318117380142212, "learning_rate": 4.87012987012987e-05, "loss": 0.4389, "step": 84 }, { "epoch": 1.58, "grad_norm": 1.323292851448059, "learning_rate": 4.8051948051948054e-05, "loss": 0.4559, "step": 85 }, { "epoch": 1.6, "grad_norm": 1.0558339357376099, "learning_rate": 4.740259740259741e-05, "loss": 0.4613, "step": 86 }, { "epoch": 1.61, "grad_norm": 1.0301963090896606, "learning_rate": 4.675324675324675e-05, "loss": 0.4194, "step": 87 }, { "epoch": 1.63, "grad_norm": 0.9772946834564209, "learning_rate": 4.6103896103896106e-05, "loss": 0.4278, "step": 88 }, { "epoch": 1.65, "grad_norm": 0.93486487865448, "learning_rate": 4.545454545454546e-05, "loss": 0.454, "step": 89 }, { "epoch": 1.67, "grad_norm": 1.1458832025527954, "learning_rate": 4.4805194805194805e-05, "loss": 0.4363, "step": 90 }, { "epoch": 1.69, "grad_norm": 0.9339586496353149, "learning_rate": 4.415584415584416e-05, "loss": 0.434, "step": 91 }, { "epoch": 1.71, "grad_norm": 1.3206239938735962, "learning_rate": 4.3506493506493503e-05, "loss": 0.4431, "step": 92 }, { "epoch": 1.73, "grad_norm": 1.1199604272842407, "learning_rate": 4.2857142857142856e-05, "loss": 0.4472, "step": 93 }, { "epoch": 1.74, "grad_norm": 1.0672180652618408, "learning_rate": 4.220779220779221e-05, "loss": 0.4091, "step": 94 }, { "epoch": 1.76, "grad_norm": 1.0920594930648804, "learning_rate": 4.155844155844156e-05, "loss": 0.417, "step": 95 }, { "epoch": 1.78, "grad_norm": 0.9002637267112732, "learning_rate": 4.0909090909090915e-05, "loss": 0.4103, "step": 96 }, { "epoch": 1.8, "grad_norm": 0.8468071222305298, "learning_rate": 4.025974025974026e-05, "loss": 0.4167, "step": 97 }, { "epoch": 1.82, "grad_norm": 0.9398940205574036, "learning_rate": 3.9610389610389614e-05, "loss": 0.4201, "step": 98 }, { "epoch": 1.84, "grad_norm": 0.7842636108398438, "learning_rate": 3.8961038961038966e-05, "loss": 0.4064, "step": 99 }, { "epoch": 1.86, "grad_norm": 1.0805550813674927, "learning_rate": 3.831168831168831e-05, "loss": 0.4271, "step": 100 }, { "epoch": 1.87, "grad_norm": 0.8424351215362549, "learning_rate": 3.7662337662337665e-05, "loss": 0.4087, "step": 101 }, { "epoch": 1.89, "grad_norm": 1.0850481986999512, "learning_rate": 3.701298701298702e-05, "loss": 0.409, "step": 102 }, { "epoch": 1.91, "grad_norm": 0.9894604086875916, "learning_rate": 3.6363636363636364e-05, "loss": 0.4139, "step": 103 }, { "epoch": 1.93, "grad_norm": 0.97883540391922, "learning_rate": 3.571428571428572e-05, "loss": 0.4281, "step": 104 }, { "epoch": 1.95, "grad_norm": 0.9291779398918152, "learning_rate": 3.506493506493507e-05, "loss": 0.4376, "step": 105 }, { "epoch": 1.97, "grad_norm": 0.8272209167480469, "learning_rate": 3.4415584415584416e-05, "loss": 0.42, "step": 106 }, { "epoch": 1.99, "grad_norm": 0.8428053855895996, "learning_rate": 3.376623376623377e-05, "loss": 0.4131, "step": 107 } ], "logging_steps": 1, "max_steps": 159, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 5, "total_flos": 1.6171847189151744e+17, "train_batch_size": 16, "trial_name": null, "trial_params": null }