| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.9790481891649208, |
| "eval_steps": 26, |
| "global_step": 208, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.009577970667464832, |
| "grad_norm": 10.084560608592307, |
| "learning_rate": 1.7241379310344828e-07, |
| "loss": 1.579, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.009577970667464832, |
| "eval_loss": 2.5250306129455566, |
| "eval_runtime": 107.4458, |
| "eval_samples_per_second": 13.16, |
| "eval_steps_per_second": 3.295, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.019155941334929663, |
| "grad_norm": 8.306669565661105, |
| "learning_rate": 3.4482758620689656e-07, |
| "loss": 1.5724, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.02873391200239449, |
| "grad_norm": 13.938049536284893, |
| "learning_rate": 5.172413793103449e-07, |
| "loss": 1.5871, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.038311882669859326, |
| "grad_norm": 12.43456292626288, |
| "learning_rate": 6.896551724137931e-07, |
| "loss": 1.5681, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.04788985333732416, |
| "grad_norm": 13.870879646573128, |
| "learning_rate": 8.620689655172415e-07, |
| "loss": 1.5744, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.05746782400478898, |
| "grad_norm": 15.247654309196745, |
| "learning_rate": 1.0344827586206898e-06, |
| "loss": 1.5925, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.06704579467225381, |
| "grad_norm": 15.680512101057806, |
| "learning_rate": 1.2068965517241381e-06, |
| "loss": 1.5704, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.07662376533971865, |
| "grad_norm": 14.30414461091009, |
| "learning_rate": 1.3793103448275862e-06, |
| "loss": 1.5732, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.08620173600718348, |
| "grad_norm": 11.033868746409794, |
| "learning_rate": 1.5517241379310346e-06, |
| "loss": 1.5325, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.09577970667464832, |
| "grad_norm": 9.293155363204939, |
| "learning_rate": 1.724137931034483e-06, |
| "loss": 1.5525, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.10535767734211314, |
| "grad_norm": 10.55909566144827, |
| "learning_rate": 1.896551724137931e-06, |
| "loss": 1.5283, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.11493564800957796, |
| "grad_norm": 7.362528707126726, |
| "learning_rate": 2.0689655172413796e-06, |
| "loss": 1.5246, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.1245136186770428, |
| "grad_norm": 7.368215078656617, |
| "learning_rate": 2.241379310344828e-06, |
| "loss": 1.5313, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.13409158934450763, |
| "grad_norm": 6.065170717786516, |
| "learning_rate": 2.4137931034482762e-06, |
| "loss": 1.5027, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.14366956001197245, |
| "grad_norm": 5.328528823161362, |
| "learning_rate": 2.5862068965517246e-06, |
| "loss": 1.481, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.1532475306794373, |
| "grad_norm": 4.425999183762783, |
| "learning_rate": 2.7586206896551725e-06, |
| "loss": 1.4494, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.16282550134690213, |
| "grad_norm": 2.3104583142533675, |
| "learning_rate": 2.931034482758621e-06, |
| "loss": 1.4645, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.17240347201436695, |
| "grad_norm": 1.595394748941364, |
| "learning_rate": 3.103448275862069e-06, |
| "loss": 1.4619, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.18198144268183178, |
| "grad_norm": 1.2488731383034972, |
| "learning_rate": 3.2758620689655175e-06, |
| "loss": 1.4641, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.19155941334929663, |
| "grad_norm": 1.5772662843657, |
| "learning_rate": 3.448275862068966e-06, |
| "loss": 1.4029, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.20113738401676146, |
| "grad_norm": 2.556424014112241, |
| "learning_rate": 3.620689655172414e-06, |
| "loss": 1.4453, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.21071535468422628, |
| "grad_norm": 2.0581192872654483, |
| "learning_rate": 3.793103448275862e-06, |
| "loss": 1.4135, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.2202933253516911, |
| "grad_norm": 1.6613052346475512, |
| "learning_rate": 3.96551724137931e-06, |
| "loss": 1.4336, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.22987129601915593, |
| "grad_norm": 1.2670811596205898, |
| "learning_rate": 4.137931034482759e-06, |
| "loss": 1.3898, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.23944926668662078, |
| "grad_norm": 1.4594637064715403, |
| "learning_rate": 4.310344827586207e-06, |
| "loss": 1.392, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.2490272373540856, |
| "grad_norm": 1.6947460151500366, |
| "learning_rate": 4.482758620689656e-06, |
| "loss": 1.3967, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.2490272373540856, |
| "eval_loss": 2.319483518600464, |
| "eval_runtime": 107.1009, |
| "eval_samples_per_second": 13.202, |
| "eval_steps_per_second": 3.305, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.25860520802155046, |
| "grad_norm": 1.4794556532045955, |
| "learning_rate": 4.655172413793104e-06, |
| "loss": 1.3882, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.26818317868901526, |
| "grad_norm": 1.275878657564078, |
| "learning_rate": 4.8275862068965525e-06, |
| "loss": 1.4152, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.2777611493564801, |
| "grad_norm": 1.0273810925450593, |
| "learning_rate": 5e-06, |
| "loss": 1.3897, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.2873391200239449, |
| "grad_norm": 1.3658855156304837, |
| "learning_rate": 4.9998459603839726e-06, |
| "loss": 1.3539, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.29691709069140976, |
| "grad_norm": 1.160650318212732, |
| "learning_rate": 4.9993838605184505e-06, |
| "loss": 1.3461, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.3064950613588746, |
| "grad_norm": 0.9334705830010439, |
| "learning_rate": 4.998613757348784e-06, |
| "loss": 1.3575, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.3160730320263394, |
| "grad_norm": 1.0269221075865582, |
| "learning_rate": 4.99753574577609e-06, |
| "loss": 1.3503, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.32565100269380426, |
| "grad_norm": 0.9951200682896573, |
| "learning_rate": 4.996149958645559e-06, |
| "loss": 1.3718, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.33522897336126906, |
| "grad_norm": 0.8568405246328175, |
| "learning_rate": 4.994456566730085e-06, |
| "loss": 1.3515, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.3448069440287339, |
| "grad_norm": 0.8752926728569858, |
| "learning_rate": 4.992455778709222e-06, |
| "loss": 1.3571, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.35438491469619876, |
| "grad_norm": 0.9195979878575848, |
| "learning_rate": 4.990147841143462e-06, |
| "loss": 1.3335, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.36396288536366356, |
| "grad_norm": 0.8848215909446233, |
| "learning_rate": 4.98753303844386e-06, |
| "loss": 1.3093, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.3735408560311284, |
| "grad_norm": 0.8261733197817335, |
| "learning_rate": 4.984611692836979e-06, |
| "loss": 1.3376, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.38311882669859326, |
| "grad_norm": 0.7643849735934586, |
| "learning_rate": 4.981384164325184e-06, |
| "loss": 1.3172, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.39269679736605806, |
| "grad_norm": 0.8302859072234411, |
| "learning_rate": 4.977850850642275e-06, |
| "loss": 1.352, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.4022747680335229, |
| "grad_norm": 0.8019795318623388, |
| "learning_rate": 4.97401218720448e-06, |
| "loss": 1.3271, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.4118527387009877, |
| "grad_norm": 0.7856123291749388, |
| "learning_rate": 4.969868647056793e-06, |
| "loss": 1.3302, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.42143070936845256, |
| "grad_norm": 0.7212471859830762, |
| "learning_rate": 4.965420740814679e-06, |
| "loss": 1.3215, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.4310086800359174, |
| "grad_norm": 0.7660292329930958, |
| "learning_rate": 4.960669016601155e-06, |
| "loss": 1.3435, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.4405866507033822, |
| "grad_norm": 0.7247198414191649, |
| "learning_rate": 4.95561405997924e-06, |
| "loss": 1.3163, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.45016462137084706, |
| "grad_norm": 0.7419070442778594, |
| "learning_rate": 4.950256493879795e-06, |
| "loss": 1.3209, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.45974259203831186, |
| "grad_norm": 0.7024643859790418, |
| "learning_rate": 4.94459697852476e-06, |
| "loss": 1.2684, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.4693205627057767, |
| "grad_norm": 0.7208397492740805, |
| "learning_rate": 4.938636211345792e-06, |
| "loss": 1.2818, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.47889853337324156, |
| "grad_norm": 0.7159719760236076, |
| "learning_rate": 4.932374926898321e-06, |
| "loss": 1.3094, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.48847650404070636, |
| "grad_norm": 0.7100286359014379, |
| "learning_rate": 4.92581389677103e-06, |
| "loss": 1.3177, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.4980544747081712, |
| "grad_norm": 0.664062518173294, |
| "learning_rate": 4.918953929490768e-06, |
| "loss": 1.2868, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.4980544747081712, |
| "eval_loss": 2.239407777786255, |
| "eval_runtime": 107.263, |
| "eval_samples_per_second": 13.183, |
| "eval_steps_per_second": 3.3, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.507632445375636, |
| "grad_norm": 0.8658636506450442, |
| "learning_rate": 4.911795870422916e-06, |
| "loss": 1.2904, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.5172104160431009, |
| "grad_norm": 0.6715121564275828, |
| "learning_rate": 4.904340601667208e-06, |
| "loss": 1.326, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.5267883867105657, |
| "grad_norm": 0.8518222183690225, |
| "learning_rate": 4.896589041949036e-06, |
| "loss": 1.2757, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.5363663573780305, |
| "grad_norm": 0.6780934729098863, |
| "learning_rate": 4.888542146506224e-06, |
| "loss": 1.3027, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.5459443280454953, |
| "grad_norm": 0.8407110074770763, |
| "learning_rate": 4.880200906971321e-06, |
| "loss": 1.2965, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.5555222987129602, |
| "grad_norm": 0.654501814705368, |
| "learning_rate": 4.8715663512493924e-06, |
| "loss": 1.2764, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.565100269380425, |
| "grad_norm": 0.7722805216190872, |
| "learning_rate": 4.8626395433913595e-06, |
| "loss": 1.2799, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.5746782400478898, |
| "grad_norm": 0.6575468000608066, |
| "learning_rate": 4.853421583462866e-06, |
| "loss": 1.3009, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.5842562107153547, |
| "grad_norm": 0.6919845481307941, |
| "learning_rate": 4.8439136074087165e-06, |
| "loss": 1.2885, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.5938341813828195, |
| "grad_norm": 0.652693683934317, |
| "learning_rate": 4.834116786912897e-06, |
| "loss": 1.2564, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.6034121520502843, |
| "grad_norm": 0.6684643483116979, |
| "learning_rate": 4.82403232925418e-06, |
| "loss": 1.278, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.6129901227177492, |
| "grad_norm": 0.6735443956477082, |
| "learning_rate": 4.813661477157355e-06, |
| "loss": 1.2895, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.622568093385214, |
| "grad_norm": 0.6574494336528988, |
| "learning_rate": 4.803005508640083e-06, |
| "loss": 1.2481, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.6321460640526788, |
| "grad_norm": 0.7061153031772025, |
| "learning_rate": 4.7920657368554e-06, |
| "loss": 1.3023, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.6417240347201437, |
| "grad_norm": 0.6609850544647713, |
| "learning_rate": 4.780843509929905e-06, |
| "loss": 1.2619, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.6513020053876085, |
| "grad_norm": 0.6958172104041147, |
| "learning_rate": 4.769340210797618e-06, |
| "loss": 1.2633, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.6608799760550733, |
| "grad_norm": 0.6532872905224688, |
| "learning_rate": 4.757557257029563e-06, |
| "loss": 1.2581, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.6704579467225381, |
| "grad_norm": 0.693714390508834, |
| "learning_rate": 4.745496100659083e-06, |
| "loss": 1.2499, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.680035917390003, |
| "grad_norm": 0.6749996898449282, |
| "learning_rate": 4.733158228002891e-06, |
| "loss": 1.2536, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.6896138880574678, |
| "grad_norm": 0.6753612400656019, |
| "learning_rate": 4.720545159477921e-06, |
| "loss": 1.2605, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.6991918587249326, |
| "grad_norm": 0.6950386791904168, |
| "learning_rate": 4.707658449413961e-06, |
| "loss": 1.2489, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.7087698293923975, |
| "grad_norm": 0.6396387112266337, |
| "learning_rate": 4.694499685862106e-06, |
| "loss": 1.264, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.7183478000598623, |
| "grad_norm": 0.6809655013846588, |
| "learning_rate": 4.681070490399064e-06, |
| "loss": 1.2477, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.7279257707273271, |
| "grad_norm": 0.6814836664342683, |
| "learning_rate": 4.667372517927323e-06, |
| "loss": 1.2349, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.737503741394792, |
| "grad_norm": 0.6502075268222723, |
| "learning_rate": 4.653407456471222e-06, |
| "loss": 1.243, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.7470817120622568, |
| "grad_norm": 0.6579341200451629, |
| "learning_rate": 4.639177026968924e-06, |
| "loss": 1.2549, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.7470817120622568, |
| "eval_loss": 2.2078425884246826, |
| "eval_runtime": 107.0636, |
| "eval_samples_per_second": 13.207, |
| "eval_steps_per_second": 3.306, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.7566596827297216, |
| "grad_norm": 0.6264741505964025, |
| "learning_rate": 4.624682983060346e-06, |
| "loss": 1.2903, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.7662376533971865, |
| "grad_norm": 0.6533395420906253, |
| "learning_rate": 4.609927110871053e-06, |
| "loss": 1.2371, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.7758156240646513, |
| "grad_norm": 0.6366166912748572, |
| "learning_rate": 4.594911228792156e-06, |
| "loss": 1.2554, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.7853935947321161, |
| "grad_norm": 0.6435835690637465, |
| "learning_rate": 4.579637187256222e-06, |
| "loss": 1.2855, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.7949715653995809, |
| "grad_norm": 0.6410872090826751, |
| "learning_rate": 4.564106868509246e-06, |
| "loss": 1.232, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.8045495360670458, |
| "grad_norm": 0.6260242741257913, |
| "learning_rate": 4.5483221863786965e-06, |
| "loss": 1.2458, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.8141275067345106, |
| "grad_norm": 0.6588265965096135, |
| "learning_rate": 4.5322850860376744e-06, |
| "loss": 1.2474, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.8237054774019754, |
| "grad_norm": 0.6372013969893753, |
| "learning_rate": 4.515997543765202e-06, |
| "loss": 1.2563, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.8332834480694403, |
| "grad_norm": 0.683356686747451, |
| "learning_rate": 4.499461566702685e-06, |
| "loss": 1.2447, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.8428614187369051, |
| "grad_norm": 0.6520958114219059, |
| "learning_rate": 4.48267919260657e-06, |
| "loss": 1.2243, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.8524393894043699, |
| "grad_norm": 0.6468861797594448, |
| "learning_rate": 4.465652489597226e-06, |
| "loss": 1.2254, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.8620173600718348, |
| "grad_norm": 0.6675355862176291, |
| "learning_rate": 4.4483835559040885e-06, |
| "loss": 1.2116, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.8715953307392996, |
| "grad_norm": 0.6318507194646, |
| "learning_rate": 4.430874519607089e-06, |
| "loss": 1.2634, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.8811733014067644, |
| "grad_norm": 0.6496099541936005, |
| "learning_rate": 4.413127538374411e-06, |
| "loss": 1.2129, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.8907512720742293, |
| "grad_norm": 0.6026396711785842, |
| "learning_rate": 4.395144799196593e-06, |
| "loss": 1.2483, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.9003292427416941, |
| "grad_norm": 0.6709684350468395, |
| "learning_rate": 4.376928518117028e-06, |
| "loss": 1.2193, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.9099072134091589, |
| "grad_norm": 0.6237262552476821, |
| "learning_rate": 4.358480939958867e-06, |
| "loss": 1.218, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.9194851840766237, |
| "grad_norm": 0.6582242790059232, |
| "learning_rate": 4.339804338048397e-06, |
| "loss": 1.229, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.9290631547440886, |
| "grad_norm": 0.6235719312223321, |
| "learning_rate": 4.320901013934887e-06, |
| "loss": 1.2098, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.9386411254115534, |
| "grad_norm": 0.6295163336318428, |
| "learning_rate": 4.301773297106968e-06, |
| "loss": 1.205, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.9482190960790182, |
| "grad_norm": 0.6250959313071772, |
| "learning_rate": 4.282423544705564e-06, |
| "loss": 1.2054, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.9577970667464831, |
| "grad_norm": 0.6086898991547662, |
| "learning_rate": 4.262854141233419e-06, |
| "loss": 1.2118, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.9673750374139479, |
| "grad_norm": 0.5764067645719498, |
| "learning_rate": 4.243067498261251e-06, |
| "loss": 1.2372, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.9769530080814127, |
| "grad_norm": 0.6406315852737573, |
| "learning_rate": 4.223066054130568e-06, |
| "loss": 1.2251, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.9865309787488776, |
| "grad_norm": 0.5834984455673559, |
| "learning_rate": 4.2028522736531895e-06, |
| "loss": 1.2258, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.9961089494163424, |
| "grad_norm": 0.5911139350878512, |
| "learning_rate": 4.182428647807503e-06, |
| "loss": 1.2286, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.9961089494163424, |
| "eval_loss": 2.184576988220215, |
| "eval_runtime": 107.5576, |
| "eval_samples_per_second": 13.146, |
| "eval_steps_per_second": 3.291, |
| "step": 104 |
| }, |
| { |
| "epoch": 1.0056869200838072, |
| "grad_norm": 0.6299976497698655, |
| "learning_rate": 4.161797693431493e-06, |
| "loss": 1.2383, |
| "step": 105 |
| }, |
| { |
| "epoch": 1.002095181083508, |
| "grad_norm": 0.5986176560633782, |
| "learning_rate": 4.140961952912594e-06, |
| "loss": 1.2182, |
| "step": 106 |
| }, |
| { |
| "epoch": 1.0116731517509727, |
| "grad_norm": 0.7138997909374802, |
| "learning_rate": 4.11992399387438e-06, |
| "loss": 1.1894, |
| "step": 107 |
| }, |
| { |
| "epoch": 1.0212511224184375, |
| "grad_norm": 0.6431525283411005, |
| "learning_rate": 4.098686408860157e-06, |
| "loss": 1.1741, |
| "step": 108 |
| }, |
| { |
| "epoch": 1.0308290930859023, |
| "grad_norm": 0.7490910983392529, |
| "learning_rate": 4.077251815013477e-06, |
| "loss": 1.1849, |
| "step": 109 |
| }, |
| { |
| "epoch": 1.0404070637533673, |
| "grad_norm": 0.6667698353299697, |
| "learning_rate": 4.055622853755627e-06, |
| "loss": 1.1833, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.0499850344208321, |
| "grad_norm": 0.7240102351414811, |
| "learning_rate": 4.033802190460114e-06, |
| "loss": 1.1915, |
| "step": 111 |
| }, |
| { |
| "epoch": 1.059563005088297, |
| "grad_norm": 0.6281393232743739, |
| "learning_rate": 4.011792514124217e-06, |
| "loss": 1.1557, |
| "step": 112 |
| }, |
| { |
| "epoch": 1.0691409757557617, |
| "grad_norm": 0.6735415717178005, |
| "learning_rate": 3.989596537037608e-06, |
| "loss": 1.1878, |
| "step": 113 |
| }, |
| { |
| "epoch": 1.0787189464232265, |
| "grad_norm": 0.5939146155666697, |
| "learning_rate": 3.967216994448116e-06, |
| "loss": 1.1639, |
| "step": 114 |
| }, |
| { |
| "epoch": 1.0882969170906913, |
| "grad_norm": 0.6932505538102671, |
| "learning_rate": 3.9446566442246615e-06, |
| "loss": 1.1759, |
| "step": 115 |
| }, |
| { |
| "epoch": 1.0978748877581563, |
| "grad_norm": 0.5763908483496408, |
| "learning_rate": 3.921918266517392e-06, |
| "loss": 1.1781, |
| "step": 116 |
| }, |
| { |
| "epoch": 1.1074528584256211, |
| "grad_norm": 0.6818836608860367, |
| "learning_rate": 3.899004663415083e-06, |
| "loss": 1.1869, |
| "step": 117 |
| }, |
| { |
| "epoch": 1.117030829093086, |
| "grad_norm": 0.5998154432302447, |
| "learning_rate": 3.875918658599837e-06, |
| "loss": 1.1692, |
| "step": 118 |
| }, |
| { |
| "epoch": 1.1266087997605507, |
| "grad_norm": 0.6596200288243683, |
| "learning_rate": 3.852663096999104e-06, |
| "loss": 1.2059, |
| "step": 119 |
| }, |
| { |
| "epoch": 1.1361867704280155, |
| "grad_norm": 0.5918812335768482, |
| "learning_rate": 3.829240844435109e-06, |
| "loss": 1.1798, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.1457647410954803, |
| "grad_norm": 0.6232580849345692, |
| "learning_rate": 3.8056547872716865e-06, |
| "loss": 1.1517, |
| "step": 121 |
| }, |
| { |
| "epoch": 1.1553427117629451, |
| "grad_norm": 0.5903843042319051, |
| "learning_rate": 3.7819078320585865e-06, |
| "loss": 1.1906, |
| "step": 122 |
| }, |
| { |
| "epoch": 1.1649206824304101, |
| "grad_norm": 0.5896678764206408, |
| "learning_rate": 3.7580029051732992e-06, |
| "loss": 1.1832, |
| "step": 123 |
| }, |
| { |
| "epoch": 1.174498653097875, |
| "grad_norm": 0.5666656027289849, |
| "learning_rate": 3.733942952460432e-06, |
| "loss": 1.1911, |
| "step": 124 |
| }, |
| { |
| "epoch": 1.1840766237653397, |
| "grad_norm": 0.565358825737842, |
| "learning_rate": 3.7097309388686865e-06, |
| "loss": 1.1945, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.1936545944328045, |
| "grad_norm": 0.645159266559964, |
| "learning_rate": 3.6853698480854853e-06, |
| "loss": 1.1988, |
| "step": 126 |
| }, |
| { |
| "epoch": 1.2032325651002693, |
| "grad_norm": 0.5661828443152349, |
| "learning_rate": 3.660862682169283e-06, |
| "loss": 1.1683, |
| "step": 127 |
| }, |
| { |
| "epoch": 1.2128105357677341, |
| "grad_norm": 0.5590652900384634, |
| "learning_rate": 3.636212461179623e-06, |
| "loss": 1.1401, |
| "step": 128 |
| }, |
| { |
| "epoch": 1.2223885064351991, |
| "grad_norm": 0.5772830186331369, |
| "learning_rate": 3.6114222228049657e-06, |
| "loss": 1.1457, |
| "step": 129 |
| }, |
| { |
| "epoch": 1.231966477102664, |
| "grad_norm": 0.5638028162671672, |
| "learning_rate": 3.5864950219883514e-06, |
| "loss": 1.1599, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.231966477102664, |
| "eval_loss": 2.181441068649292, |
| "eval_runtime": 107.4543, |
| "eval_samples_per_second": 13.159, |
| "eval_steps_per_second": 3.294, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.2415444477701287, |
| "grad_norm": 0.5674264221381613, |
| "learning_rate": 3.561433930550934e-06, |
| "loss": 1.1439, |
| "step": 131 |
| }, |
| { |
| "epoch": 1.2511224184375935, |
| "grad_norm": 0.5548457286136358, |
| "learning_rate": 3.536242036813436e-06, |
| "loss": 1.1455, |
| "step": 132 |
| }, |
| { |
| "epoch": 1.2607003891050583, |
| "grad_norm": 0.5681860545302818, |
| "learning_rate": 3.510922445215568e-06, |
| "loss": 1.1619, |
| "step": 133 |
| }, |
| { |
| "epoch": 1.2702783597725231, |
| "grad_norm": 0.5189655726956113, |
| "learning_rate": 3.4854782759334625e-06, |
| "loss": 1.1647, |
| "step": 134 |
| }, |
| { |
| "epoch": 1.279856330439988, |
| "grad_norm": 0.5482759127528988, |
| "learning_rate": 3.4599126644951758e-06, |
| "loss": 1.1963, |
| "step": 135 |
| }, |
| { |
| "epoch": 1.289434301107453, |
| "grad_norm": 0.5545710145438582, |
| "learning_rate": 3.4342287613942804e-06, |
| "loss": 1.1673, |
| "step": 136 |
| }, |
| { |
| "epoch": 1.2990122717749177, |
| "grad_norm": 0.5616911560631516, |
| "learning_rate": 3.4084297317016353e-06, |
| "loss": 1.1482, |
| "step": 137 |
| }, |
| { |
| "epoch": 1.3085902424423825, |
| "grad_norm": 0.5429625311889626, |
| "learning_rate": 3.3825187546753426e-06, |
| "loss": 1.1459, |
| "step": 138 |
| }, |
| { |
| "epoch": 1.3181682131098473, |
| "grad_norm": 0.5775738090552808, |
| "learning_rate": 3.3564990233689632e-06, |
| "loss": 1.1744, |
| "step": 139 |
| }, |
| { |
| "epoch": 1.3277461837773121, |
| "grad_norm": 0.5422962267277087, |
| "learning_rate": 3.330373744238033e-06, |
| "loss": 1.1796, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.3373241544447771, |
| "grad_norm": 0.5383626495155892, |
| "learning_rate": 3.3041461367449256e-06, |
| "loss": 1.1646, |
| "step": 141 |
| }, |
| { |
| "epoch": 1.346902125112242, |
| "grad_norm": 0.5588657340470299, |
| "learning_rate": 3.2778194329621104e-06, |
| "loss": 1.1842, |
| "step": 142 |
| }, |
| { |
| "epoch": 1.3564800957797067, |
| "grad_norm": 0.5198196148369068, |
| "learning_rate": 3.2513968771738606e-06, |
| "loss": 1.1708, |
| "step": 143 |
| }, |
| { |
| "epoch": 1.3660580664471715, |
| "grad_norm": 0.5453371169769571, |
| "learning_rate": 3.224881725476456e-06, |
| "loss": 1.1636, |
| "step": 144 |
| }, |
| { |
| "epoch": 1.3756360371146363, |
| "grad_norm": 0.5692897944097868, |
| "learning_rate": 3.198277245376924e-06, |
| "loss": 1.1273, |
| "step": 145 |
| }, |
| { |
| "epoch": 1.3852140077821011, |
| "grad_norm": 0.5423704486470122, |
| "learning_rate": 3.1715867153903844e-06, |
| "loss": 1.1405, |
| "step": 146 |
| }, |
| { |
| "epoch": 1.394791978449566, |
| "grad_norm": 0.5819177408649716, |
| "learning_rate": 3.144813424636031e-06, |
| "loss": 1.1665, |
| "step": 147 |
| }, |
| { |
| "epoch": 1.4043699491170307, |
| "grad_norm": 0.554870749454361, |
| "learning_rate": 3.1179606724318052e-06, |
| "loss": 1.1872, |
| "step": 148 |
| }, |
| { |
| "epoch": 1.4139479197844955, |
| "grad_norm": 0.5493659769746441, |
| "learning_rate": 3.091031767887817e-06, |
| "loss": 1.1906, |
| "step": 149 |
| }, |
| { |
| "epoch": 1.4235258904519605, |
| "grad_norm": 0.6008378552179591, |
| "learning_rate": 3.0640300294985613e-06, |
| "loss": 1.1635, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.4331038611194253, |
| "grad_norm": 0.5078261653762177, |
| "learning_rate": 3.036958784733967e-06, |
| "loss": 1.1438, |
| "step": 151 |
| }, |
| { |
| "epoch": 1.4426818317868901, |
| "grad_norm": 0.5559592542323409, |
| "learning_rate": 3.0098213696293542e-06, |
| "loss": 1.1642, |
| "step": 152 |
| }, |
| { |
| "epoch": 1.452259802454355, |
| "grad_norm": 0.5461821050739424, |
| "learning_rate": 2.982621128374325e-06, |
| "loss": 1.1725, |
| "step": 153 |
| }, |
| { |
| "epoch": 1.46183777312182, |
| "grad_norm": 0.5412862095154186, |
| "learning_rate": 2.9553614129006543e-06, |
| "loss": 1.1654, |
| "step": 154 |
| }, |
| { |
| "epoch": 1.4714157437892847, |
| "grad_norm": 0.5658659296771973, |
| "learning_rate": 2.9280455824692255e-06, |
| "loss": 1.1655, |
| "step": 155 |
| }, |
| { |
| "epoch": 1.4809937144567495, |
| "grad_norm": 0.5525850336445564, |
| "learning_rate": 2.9006770032560637e-06, |
| "loss": 1.1577, |
| "step": 156 |
| }, |
| { |
| "epoch": 1.4809937144567495, |
| "eval_loss": 2.1755869388580322, |
| "eval_runtime": 107.2159, |
| "eval_samples_per_second": 13.188, |
| "eval_steps_per_second": 3.302, |
| "step": 156 |
| }, |
| { |
| "epoch": 1.4905716851242143, |
| "grad_norm": 0.5710362202768258, |
| "learning_rate": 2.8732590479375167e-06, |
| "loss": 1.1595, |
| "step": 157 |
| }, |
| { |
| "epoch": 1.5001496557916791, |
| "grad_norm": 0.5369626897696785, |
| "learning_rate": 2.8457950952746293e-06, |
| "loss": 1.1622, |
| "step": 158 |
| }, |
| { |
| "epoch": 1.509727626459144, |
| "grad_norm": 0.5194143574454793, |
| "learning_rate": 2.8182885296967833e-06, |
| "loss": 1.1313, |
| "step": 159 |
| }, |
| { |
| "epoch": 1.5193055971266087, |
| "grad_norm": 0.5220817246963333, |
| "learning_rate": 2.7907427408846156e-06, |
| "loss": 1.1493, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.5288835677940735, |
| "grad_norm": 0.5307538609855902, |
| "learning_rate": 2.763161123352314e-06, |
| "loss": 1.1571, |
| "step": 161 |
| }, |
| { |
| "epoch": 1.5384615384615383, |
| "grad_norm": 0.5133921578064818, |
| "learning_rate": 2.735547076029296e-06, |
| "loss": 1.1398, |
| "step": 162 |
| }, |
| { |
| "epoch": 1.5480395091290033, |
| "grad_norm": 0.528392253063443, |
| "learning_rate": 2.7079040018413586e-06, |
| "loss": 1.169, |
| "step": 163 |
| }, |
| { |
| "epoch": 1.5576174797964681, |
| "grad_norm": 0.5033775123091357, |
| "learning_rate": 2.6802353072913307e-06, |
| "loss": 1.1396, |
| "step": 164 |
| }, |
| { |
| "epoch": 1.567195450463933, |
| "grad_norm": 0.5429413779707357, |
| "learning_rate": 2.6525444020392794e-06, |
| "loss": 1.1558, |
| "step": 165 |
| }, |
| { |
| "epoch": 1.5767734211313977, |
| "grad_norm": 0.5391198899526514, |
| "learning_rate": 2.6248346984823325e-06, |
| "loss": 1.1584, |
| "step": 166 |
| }, |
| { |
| "epoch": 1.5863513917988628, |
| "grad_norm": 0.5237711725405991, |
| "learning_rate": 2.5971096113341692e-06, |
| "loss": 1.1399, |
| "step": 167 |
| }, |
| { |
| "epoch": 1.5959293624663276, |
| "grad_norm": 0.522431379990406, |
| "learning_rate": 2.5693725572042135e-06, |
| "loss": 1.146, |
| "step": 168 |
| }, |
| { |
| "epoch": 1.6055073331337923, |
| "grad_norm": 0.540451111257001, |
| "learning_rate": 2.5416269541765963e-06, |
| "loss": 1.1347, |
| "step": 169 |
| }, |
| { |
| "epoch": 1.6150853038012571, |
| "grad_norm": 0.542474309771266, |
| "learning_rate": 2.5138762213889493e-06, |
| "loss": 1.1507, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.624663274468722, |
| "grad_norm": 0.5339716680549861, |
| "learning_rate": 2.486123778611051e-06, |
| "loss": 1.1428, |
| "step": 171 |
| }, |
| { |
| "epoch": 1.6342412451361867, |
| "grad_norm": 0.5194346219437855, |
| "learning_rate": 2.458373045823404e-06, |
| "loss": 1.1717, |
| "step": 172 |
| }, |
| { |
| "epoch": 1.6438192158036515, |
| "grad_norm": 0.5486922902738444, |
| "learning_rate": 2.4306274427957878e-06, |
| "loss": 1.1405, |
| "step": 173 |
| }, |
| { |
| "epoch": 1.6533971864711163, |
| "grad_norm": 0.5364703724723029, |
| "learning_rate": 2.402890388665831e-06, |
| "loss": 1.1397, |
| "step": 174 |
| }, |
| { |
| "epoch": 1.6629751571385811, |
| "grad_norm": 0.5151838009534813, |
| "learning_rate": 2.375165301517668e-06, |
| "loss": 1.1625, |
| "step": 175 |
| }, |
| { |
| "epoch": 1.6725531278060461, |
| "grad_norm": 0.5387178228054901, |
| "learning_rate": 2.3474555979607214e-06, |
| "loss": 1.1586, |
| "step": 176 |
| }, |
| { |
| "epoch": 1.682131098473511, |
| "grad_norm": 0.5264984610535657, |
| "learning_rate": 2.3197646927086697e-06, |
| "loss": 1.1654, |
| "step": 177 |
| }, |
| { |
| "epoch": 1.6917090691409757, |
| "grad_norm": 0.5272357155280125, |
| "learning_rate": 2.2920959981586426e-06, |
| "loss": 1.1934, |
| "step": 178 |
| }, |
| { |
| "epoch": 1.7012870398084405, |
| "grad_norm": 0.5252339989768573, |
| "learning_rate": 2.2644529239707054e-06, |
| "loss": 1.1426, |
| "step": 179 |
| }, |
| { |
| "epoch": 1.7108650104759056, |
| "grad_norm": 0.4974185735061034, |
| "learning_rate": 2.2368388766476875e-06, |
| "loss": 1.1597, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.7204429811433704, |
| "grad_norm": 0.5361098970394095, |
| "learning_rate": 2.2092572591153843e-06, |
| "loss": 1.1637, |
| "step": 181 |
| }, |
| { |
| "epoch": 1.7300209518108352, |
| "grad_norm": 0.5305009042993176, |
| "learning_rate": 2.1817114703032176e-06, |
| "loss": 1.1637, |
| "step": 182 |
| }, |
| { |
| "epoch": 1.7300209518108352, |
| "eval_loss": 2.1710658073425293, |
| "eval_runtime": 107.1212, |
| "eval_samples_per_second": 13.2, |
| "eval_steps_per_second": 3.305, |
| "step": 182 |
| }, |
| { |
| "epoch": 1.7395989224783, |
| "grad_norm": 0.5012187069773779, |
| "learning_rate": 2.154204904725371e-06, |
| "loss": 1.1447, |
| "step": 183 |
| }, |
| { |
| "epoch": 1.7491768931457647, |
| "grad_norm": 0.5763812037469009, |
| "learning_rate": 2.126740952062484e-06, |
| "loss": 1.1565, |
| "step": 184 |
| }, |
| { |
| "epoch": 1.7587548638132295, |
| "grad_norm": 0.5129804478325861, |
| "learning_rate": 2.099322996743936e-06, |
| "loss": 1.1798, |
| "step": 185 |
| }, |
| { |
| "epoch": 1.7683328344806943, |
| "grad_norm": 0.5107704085635135, |
| "learning_rate": 2.0719544175307754e-06, |
| "loss": 1.1486, |
| "step": 186 |
| }, |
| { |
| "epoch": 1.7779108051481591, |
| "grad_norm": 0.5225266432128085, |
| "learning_rate": 2.044638587099347e-06, |
| "loss": 1.1457, |
| "step": 187 |
| }, |
| { |
| "epoch": 1.787488775815624, |
| "grad_norm": 0.48553711881118367, |
| "learning_rate": 2.0173788716256758e-06, |
| "loss": 1.1557, |
| "step": 188 |
| }, |
| { |
| "epoch": 1.797066746483089, |
| "grad_norm": 0.5155245524580911, |
| "learning_rate": 1.9901786303706466e-06, |
| "loss": 1.1667, |
| "step": 189 |
| }, |
| { |
| "epoch": 1.8066447171505537, |
| "grad_norm": 0.5394238331941211, |
| "learning_rate": 1.9630412152660333e-06, |
| "loss": 1.1639, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.8162226878180185, |
| "grad_norm": 0.5208012650775928, |
| "learning_rate": 1.93596997050144e-06, |
| "loss": 1.167, |
| "step": 191 |
| }, |
| { |
| "epoch": 1.8258006584854833, |
| "grad_norm": 0.5084683728452081, |
| "learning_rate": 1.9089682321121834e-06, |
| "loss": 1.146, |
| "step": 192 |
| }, |
| { |
| "epoch": 1.8353786291529484, |
| "grad_norm": 0.5107216674575125, |
| "learning_rate": 1.8820393275681954e-06, |
| "loss": 1.1299, |
| "step": 193 |
| }, |
| { |
| "epoch": 1.8449565998204132, |
| "grad_norm": 0.5037220655522233, |
| "learning_rate": 1.8551865753639692e-06, |
| "loss": 1.1705, |
| "step": 194 |
| }, |
| { |
| "epoch": 1.854534570487878, |
| "grad_norm": 0.5081083073272432, |
| "learning_rate": 1.8284132846096164e-06, |
| "loss": 1.1232, |
| "step": 195 |
| }, |
| { |
| "epoch": 1.8641125411553428, |
| "grad_norm": 0.4960779996118519, |
| "learning_rate": 1.801722754623077e-06, |
| "loss": 1.1356, |
| "step": 196 |
| }, |
| { |
| "epoch": 1.8736905118228075, |
| "grad_norm": 0.5194537399766056, |
| "learning_rate": 1.775118274523545e-06, |
| "loss": 1.1321, |
| "step": 197 |
| }, |
| { |
| "epoch": 1.8832684824902723, |
| "grad_norm": 0.5149057994299137, |
| "learning_rate": 1.74860312282614e-06, |
| "loss": 1.1306, |
| "step": 198 |
| }, |
| { |
| "epoch": 1.8928464531577371, |
| "grad_norm": 0.5061127962699723, |
| "learning_rate": 1.72218056703789e-06, |
| "loss": 1.1302, |
| "step": 199 |
| }, |
| { |
| "epoch": 1.902424423825202, |
| "grad_norm": 0.49704736224795454, |
| "learning_rate": 1.6958538632550753e-06, |
| "loss": 1.1479, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.9120023944926667, |
| "grad_norm": 0.4976492539596855, |
| "learning_rate": 1.6696262557619677e-06, |
| "loss": 1.135, |
| "step": 201 |
| }, |
| { |
| "epoch": 1.9215803651601315, |
| "grad_norm": 0.5438558597014863, |
| "learning_rate": 1.6435009766310372e-06, |
| "loss": 1.1677, |
| "step": 202 |
| }, |
| { |
| "epoch": 1.9311583358275966, |
| "grad_norm": 0.49386649254244525, |
| "learning_rate": 1.6174812453246582e-06, |
| "loss": 1.1396, |
| "step": 203 |
| }, |
| { |
| "epoch": 1.9407363064950613, |
| "grad_norm": 0.5039832884638089, |
| "learning_rate": 1.5915702682983657e-06, |
| "loss": 1.1857, |
| "step": 204 |
| }, |
| { |
| "epoch": 1.9503142771625261, |
| "grad_norm": 0.4892382263387271, |
| "learning_rate": 1.5657712386057202e-06, |
| "loss": 1.15, |
| "step": 205 |
| }, |
| { |
| "epoch": 1.9598922478299912, |
| "grad_norm": 0.5084631284159544, |
| "learning_rate": 1.5400873355048248e-06, |
| "loss": 1.1572, |
| "step": 206 |
| }, |
| { |
| "epoch": 1.969470218497456, |
| "grad_norm": 0.5008750617477549, |
| "learning_rate": 1.5145217240665373e-06, |
| "loss": 1.1326, |
| "step": 207 |
| }, |
| { |
| "epoch": 1.9790481891649208, |
| "grad_norm": 0.4980386882470781, |
| "learning_rate": 1.489077554784432e-06, |
| "loss": 1.143, |
| "step": 208 |
| }, |
| { |
| "epoch": 1.9790481891649208, |
| "eval_loss": 2.1687815189361572, |
| "eval_runtime": 107.1708, |
| "eval_samples_per_second": 13.194, |
| "eval_steps_per_second": 3.303, |
| "step": 208 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 312, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 52, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 4.9974933303014195e+17, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|