| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 6.0, |
| "eval_steps": 500, |
| "global_step": 3426, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.035056967572304996, |
| "grad_norm": 12.451324462890625, |
| "learning_rate": 1.8446601941747574e-06, |
| "loss": 3.237, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.07011393514460999, |
| "grad_norm": 4.131791114807129, |
| "learning_rate": 3.7864077669902915e-06, |
| "loss": 2.5266, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.10517090271691498, |
| "grad_norm": 2.2451958656311035, |
| "learning_rate": 5.728155339805825e-06, |
| "loss": 1.8755, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.14022787028921999, |
| "grad_norm": 1.4421552419662476, |
| "learning_rate": 7.66990291262136e-06, |
| "loss": 1.3649, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.175284837861525, |
| "grad_norm": 0.9530179500579834, |
| "learning_rate": 9.611650485436894e-06, |
| "loss": 1.0674, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.21034180543382996, |
| "grad_norm": 0.7194477319717407, |
| "learning_rate": 9.99942798060303e-06, |
| "loss": 0.9241, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.24539877300613497, |
| "grad_norm": 0.6556061506271362, |
| "learning_rate": 9.997104376116195e-06, |
| "loss": 0.8575, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.28045574057843997, |
| "grad_norm": 0.5718048810958862, |
| "learning_rate": 9.992994265395959e-06, |
| "loss": 0.829, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.31551270815074495, |
| "grad_norm": 0.4922148287296295, |
| "learning_rate": 9.987099117840969e-06, |
| "loss": 0.8034, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.35056967572305, |
| "grad_norm": 0.47302234172821045, |
| "learning_rate": 9.979421041015336e-06, |
| "loss": 0.7839, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.38562664329535495, |
| "grad_norm": 0.49009189009666443, |
| "learning_rate": 9.969962779895172e-06, |
| "loss": 0.768, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.42068361086765993, |
| "grad_norm": 0.4963654577732086, |
| "learning_rate": 9.958727715887218e-06, |
| "loss": 0.7628, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.45574057843996496, |
| "grad_norm": 0.5206854343414307, |
| "learning_rate": 9.94571986561998e-06, |
| "loss": 0.7488, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.49079754601226994, |
| "grad_norm": 0.48924869298934937, |
| "learning_rate": 9.930943879507748e-06, |
| "loss": 0.7436, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.5258545135845749, |
| "grad_norm": 0.43540337681770325, |
| "learning_rate": 9.914405040088026e-06, |
| "loss": 0.7375, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.5609114811568799, |
| "grad_norm": 0.44258421659469604, |
| "learning_rate": 9.896109260132993e-06, |
| "loss": 0.7277, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.595968448729185, |
| "grad_norm": 0.4955386519432068, |
| "learning_rate": 9.876063080535627e-06, |
| "loss": 0.7284, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.6310254163014899, |
| "grad_norm": 0.5027541518211365, |
| "learning_rate": 9.85427366797129e-06, |
| "loss": 0.7231, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.6660823838737949, |
| "grad_norm": 0.4675957262516022, |
| "learning_rate": 9.830748812335576e-06, |
| "loss": 0.7212, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.7011393514461, |
| "grad_norm": 0.4283595383167267, |
| "learning_rate": 9.805496923959363e-06, |
| "loss": 0.7164, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.7361963190184049, |
| "grad_norm": 0.452084481716156, |
| "learning_rate": 9.778527030602049e-06, |
| "loss": 0.711, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.7712532865907099, |
| "grad_norm": 0.4737929105758667, |
| "learning_rate": 9.74984877422405e-06, |
| "loss": 0.7084, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.8063102541630149, |
| "grad_norm": 0.4964485466480255, |
| "learning_rate": 9.719472407539725e-06, |
| "loss": 0.7028, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.8413672217353199, |
| "grad_norm": 0.44363030791282654, |
| "learning_rate": 9.68740879035194e-06, |
| "loss": 0.7045, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.8764241893076249, |
| "grad_norm": 0.5004998445510864, |
| "learning_rate": 9.6536693856696e-06, |
| "loss": 0.6937, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.9114811568799299, |
| "grad_norm": 0.4564264118671417, |
| "learning_rate": 9.618266255609533e-06, |
| "loss": 0.699, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.9465381244522348, |
| "grad_norm": 0.4558616280555725, |
| "learning_rate": 9.58121205708418e-06, |
| "loss": 0.691, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.9815950920245399, |
| "grad_norm": 0.413114458322525, |
| "learning_rate": 9.542520037276636e-06, |
| "loss": 0.6891, |
| "step": 560 |
| }, |
| { |
| "epoch": 1.0157756354075373, |
| "grad_norm": 0.403679758310318, |
| "learning_rate": 9.502204028904687e-06, |
| "loss": 0.6812, |
| "step": 580 |
| }, |
| { |
| "epoch": 1.0508326029798423, |
| "grad_norm": 0.40308722853660583, |
| "learning_rate": 9.46027844527549e-06, |
| "loss": 0.6791, |
| "step": 600 |
| }, |
| { |
| "epoch": 1.0858895705521472, |
| "grad_norm": 0.4085083603858948, |
| "learning_rate": 9.416758275132693e-06, |
| "loss": 0.6803, |
| "step": 620 |
| }, |
| { |
| "epoch": 1.1209465381244523, |
| "grad_norm": 0.4475920796394348, |
| "learning_rate": 9.371659077297843e-06, |
| "loss": 0.6789, |
| "step": 640 |
| }, |
| { |
| "epoch": 1.1560035056967572, |
| "grad_norm": 0.4604188799858093, |
| "learning_rate": 9.324996975107978e-06, |
| "loss": 0.674, |
| "step": 660 |
| }, |
| { |
| "epoch": 1.1910604732690622, |
| "grad_norm": 0.4190482795238495, |
| "learning_rate": 9.276788650651392e-06, |
| "loss": 0.6746, |
| "step": 680 |
| }, |
| { |
| "epoch": 1.2261174408413673, |
| "grad_norm": 0.420953631401062, |
| "learning_rate": 9.227051338803656e-06, |
| "loss": 0.6692, |
| "step": 700 |
| }, |
| { |
| "epoch": 1.2611744084136722, |
| "grad_norm": 0.4463854432106018, |
| "learning_rate": 9.175802821066009e-06, |
| "loss": 0.6737, |
| "step": 720 |
| }, |
| { |
| "epoch": 1.2962313759859772, |
| "grad_norm": 0.44004735350608826, |
| "learning_rate": 9.12306141920832e-06, |
| "loss": 0.6673, |
| "step": 740 |
| }, |
| { |
| "epoch": 1.331288343558282, |
| "grad_norm": 0.42015475034713745, |
| "learning_rate": 9.068845988718906e-06, |
| "loss": 0.6676, |
| "step": 760 |
| }, |
| { |
| "epoch": 1.3663453111305872, |
| "grad_norm": 0.43683475255966187, |
| "learning_rate": 9.013175912063534e-06, |
| "loss": 0.6649, |
| "step": 780 |
| }, |
| { |
| "epoch": 1.4014022787028921, |
| "grad_norm": 0.4281805753707886, |
| "learning_rate": 8.956071091756036e-06, |
| "loss": 0.6658, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.4364592462751973, |
| "grad_norm": 0.4270734190940857, |
| "learning_rate": 8.89755194324299e-06, |
| "loss": 0.6646, |
| "step": 820 |
| }, |
| { |
| "epoch": 1.4715162138475022, |
| "grad_norm": 0.4163481593132019, |
| "learning_rate": 8.837639387605031e-06, |
| "loss": 0.6658, |
| "step": 840 |
| }, |
| { |
| "epoch": 1.5065731814198071, |
| "grad_norm": 0.45280900597572327, |
| "learning_rate": 8.776354844077389e-06, |
| "loss": 0.6592, |
| "step": 860 |
| }, |
| { |
| "epoch": 1.541630148992112, |
| "grad_norm": 0.40485361218452454, |
| "learning_rate": 8.713720222392338e-06, |
| "loss": 0.6579, |
| "step": 880 |
| }, |
| { |
| "epoch": 1.5766871165644172, |
| "grad_norm": 0.42039763927459717, |
| "learning_rate": 8.649757914946284e-06, |
| "loss": 0.6616, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.6117440841367223, |
| "grad_norm": 0.4760454595088959, |
| "learning_rate": 8.584490788794296e-06, |
| "loss": 0.6572, |
| "step": 920 |
| }, |
| { |
| "epoch": 1.6468010517090272, |
| "grad_norm": 0.43802690505981445, |
| "learning_rate": 8.517942177474943e-06, |
| "loss": 0.6548, |
| "step": 940 |
| }, |
| { |
| "epoch": 1.6818580192813322, |
| "grad_norm": 0.5002708435058594, |
| "learning_rate": 8.450135872668369e-06, |
| "loss": 0.6557, |
| "step": 960 |
| }, |
| { |
| "epoch": 1.716914986853637, |
| "grad_norm": 0.4160609543323517, |
| "learning_rate": 8.38109611569056e-06, |
| "loss": 0.6529, |
| "step": 980 |
| }, |
| { |
| "epoch": 1.751971954425942, |
| "grad_norm": 0.43179649114608765, |
| "learning_rate": 8.310847588826876e-06, |
| "loss": 0.6529, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.7870289219982471, |
| "grad_norm": 0.4322780668735504, |
| "learning_rate": 8.239415406507934e-06, |
| "loss": 0.6535, |
| "step": 1020 |
| }, |
| { |
| "epoch": 1.8220858895705523, |
| "grad_norm": 0.4642186462879181, |
| "learning_rate": 8.166825106330985e-06, |
| "loss": 0.649, |
| "step": 1040 |
| }, |
| { |
| "epoch": 1.8571428571428572, |
| "grad_norm": 0.42697349190711975, |
| "learning_rate": 8.093102639930013e-06, |
| "loss": 0.65, |
| "step": 1060 |
| }, |
| { |
| "epoch": 1.8921998247151621, |
| "grad_norm": 0.4486387372016907, |
| "learning_rate": 8.01827436369781e-06, |
| "loss": 0.6492, |
| "step": 1080 |
| }, |
| { |
| "epoch": 1.927256792287467, |
| "grad_norm": 0.42962825298309326, |
| "learning_rate": 7.942367029363351e-06, |
| "loss": 0.6518, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.962313759859772, |
| "grad_norm": 0.43645408749580383, |
| "learning_rate": 7.865407774427828e-06, |
| "loss": 0.6475, |
| "step": 1120 |
| }, |
| { |
| "epoch": 1.997370727432077, |
| "grad_norm": 0.4662039875984192, |
| "learning_rate": 7.787424112462758e-06, |
| "loss": 0.649, |
| "step": 1140 |
| }, |
| { |
| "epoch": 2.0315512708150747, |
| "grad_norm": 0.44152551889419556, |
| "learning_rate": 7.708443923273671e-06, |
| "loss": 0.6401, |
| "step": 1160 |
| }, |
| { |
| "epoch": 2.0666082383873796, |
| "grad_norm": 0.39342495799064636, |
| "learning_rate": 7.628495442932838e-06, |
| "loss": 0.6386, |
| "step": 1180 |
| }, |
| { |
| "epoch": 2.1016652059596845, |
| "grad_norm": 0.4356766641139984, |
| "learning_rate": 7.54760725368464e-06, |
| "loss": 0.6391, |
| "step": 1200 |
| }, |
| { |
| "epoch": 2.1367221735319895, |
| "grad_norm": 0.3897708058357239, |
| "learning_rate": 7.465808273727182e-06, |
| "loss": 0.6383, |
| "step": 1220 |
| }, |
| { |
| "epoch": 2.1717791411042944, |
| "grad_norm": 0.4168529510498047, |
| "learning_rate": 7.383127746873796e-06, |
| "loss": 0.6361, |
| "step": 1240 |
| }, |
| { |
| "epoch": 2.2068361086765993, |
| "grad_norm": 0.39462465047836304, |
| "learning_rate": 7.2995952320981356e-06, |
| "loss": 0.6371, |
| "step": 1260 |
| }, |
| { |
| "epoch": 2.2418930762489047, |
| "grad_norm": 0.42870041728019714, |
| "learning_rate": 7.215240592966603e-06, |
| "loss": 0.6318, |
| "step": 1280 |
| }, |
| { |
| "epoch": 2.2769500438212096, |
| "grad_norm": 0.46848800778388977, |
| "learning_rate": 7.130093986961868e-06, |
| "loss": 0.633, |
| "step": 1300 |
| }, |
| { |
| "epoch": 2.3120070113935145, |
| "grad_norm": 0.415912002325058, |
| "learning_rate": 7.044185854701321e-06, |
| "loss": 0.6367, |
| "step": 1320 |
| }, |
| { |
| "epoch": 2.3470639789658194, |
| "grad_norm": 0.4347931444644928, |
| "learning_rate": 6.957546909054304e-06, |
| "loss": 0.6374, |
| "step": 1340 |
| }, |
| { |
| "epoch": 2.3821209465381243, |
| "grad_norm": 0.4282444417476654, |
| "learning_rate": 6.870208124161998e-06, |
| "loss": 0.6353, |
| "step": 1360 |
| }, |
| { |
| "epoch": 2.4171779141104293, |
| "grad_norm": 0.43224233388900757, |
| "learning_rate": 6.78220072436392e-06, |
| "loss": 0.6348, |
| "step": 1380 |
| }, |
| { |
| "epoch": 2.4522348816827346, |
| "grad_norm": 0.4176190495491028, |
| "learning_rate": 6.693556173034953e-06, |
| "loss": 0.633, |
| "step": 1400 |
| }, |
| { |
| "epoch": 2.4872918492550395, |
| "grad_norm": 0.4181615710258484, |
| "learning_rate": 6.6043061613369356e-06, |
| "loss": 0.6324, |
| "step": 1420 |
| }, |
| { |
| "epoch": 2.5223488168273445, |
| "grad_norm": 0.38148173689842224, |
| "learning_rate": 6.514482596888807e-06, |
| "loss": 0.6301, |
| "step": 1440 |
| }, |
| { |
| "epoch": 2.5574057843996494, |
| "grad_norm": 0.41031816601753235, |
| "learning_rate": 6.424117592359367e-06, |
| "loss": 0.6332, |
| "step": 1460 |
| }, |
| { |
| "epoch": 2.5924627519719543, |
| "grad_norm": 0.4005562961101532, |
| "learning_rate": 6.333243453986734e-06, |
| "loss": 0.6265, |
| "step": 1480 |
| }, |
| { |
| "epoch": 2.6275197195442592, |
| "grad_norm": 0.3940238058567047, |
| "learning_rate": 6.241892670028595e-06, |
| "loss": 0.6315, |
| "step": 1500 |
| }, |
| { |
| "epoch": 2.662576687116564, |
| "grad_norm": 0.4001730680465698, |
| "learning_rate": 6.150097899147384e-06, |
| "loss": 0.6299, |
| "step": 1520 |
| }, |
| { |
| "epoch": 2.6976336546888695, |
| "grad_norm": 0.3857872188091278, |
| "learning_rate": 6.057891958734538e-06, |
| "loss": 0.6304, |
| "step": 1540 |
| }, |
| { |
| "epoch": 2.7326906222611744, |
| "grad_norm": 0.4169263243675232, |
| "learning_rate": 5.965307813178015e-06, |
| "loss": 0.6315, |
| "step": 1560 |
| }, |
| { |
| "epoch": 2.7677475898334793, |
| "grad_norm": 0.4010975658893585, |
| "learning_rate": 5.872378562077241e-06, |
| "loss": 0.6297, |
| "step": 1580 |
| }, |
| { |
| "epoch": 2.8028045574057843, |
| "grad_norm": 0.4302142262458801, |
| "learning_rate": 5.779137428409738e-06, |
| "loss": 0.6302, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.8378615249780896, |
| "grad_norm": 0.39000585675239563, |
| "learning_rate": 5.685617746653629e-06, |
| "loss": 0.6312, |
| "step": 1620 |
| }, |
| { |
| "epoch": 2.8729184925503946, |
| "grad_norm": 0.4292212724685669, |
| "learning_rate": 5.591852950870287e-06, |
| "loss": 0.6312, |
| "step": 1640 |
| }, |
| { |
| "epoch": 2.9079754601226995, |
| "grad_norm": 0.3838886320590973, |
| "learning_rate": 5.497876562751384e-06, |
| "loss": 0.6302, |
| "step": 1660 |
| }, |
| { |
| "epoch": 2.9430324276950044, |
| "grad_norm": 0.36835259199142456, |
| "learning_rate": 5.403722179634602e-06, |
| "loss": 0.6292, |
| "step": 1680 |
| }, |
| { |
| "epoch": 2.9780893952673093, |
| "grad_norm": 0.3884848654270172, |
| "learning_rate": 5.309423462492314e-06, |
| "loss": 0.6261, |
| "step": 1700 |
| }, |
| { |
| "epoch": 3.0122699386503067, |
| "grad_norm": 0.3762246072292328, |
| "learning_rate": 5.215014123897504e-06, |
| "loss": 0.6202, |
| "step": 1720 |
| }, |
| { |
| "epoch": 3.0473269062226116, |
| "grad_norm": 0.38138872385025024, |
| "learning_rate": 5.120527915971235e-06, |
| "loss": 0.6205, |
| "step": 1740 |
| }, |
| { |
| "epoch": 3.0823838737949165, |
| "grad_norm": 0.38698920607566833, |
| "learning_rate": 5.0259986183160006e-06, |
| "loss": 0.6186, |
| "step": 1760 |
| }, |
| { |
| "epoch": 3.117440841367222, |
| "grad_norm": 0.378830224275589, |
| "learning_rate": 4.931460025939226e-06, |
| "loss": 0.6214, |
| "step": 1780 |
| }, |
| { |
| "epoch": 3.152497808939527, |
| "grad_norm": 0.3751004934310913, |
| "learning_rate": 4.836945937171279e-06, |
| "loss": 0.6209, |
| "step": 1800 |
| }, |
| { |
| "epoch": 3.1875547765118317, |
| "grad_norm": 0.3829745054244995, |
| "learning_rate": 4.742490141582279e-06, |
| "loss": 0.6213, |
| "step": 1820 |
| }, |
| { |
| "epoch": 3.2226117440841366, |
| "grad_norm": 0.4228389859199524, |
| "learning_rate": 4.648126407902058e-06, |
| "loss": 0.6193, |
| "step": 1840 |
| }, |
| { |
| "epoch": 3.2576687116564416, |
| "grad_norm": 0.367960125207901, |
| "learning_rate": 4.553888471947546e-06, |
| "loss": 0.6198, |
| "step": 1860 |
| }, |
| { |
| "epoch": 3.292725679228747, |
| "grad_norm": 0.39815646409988403, |
| "learning_rate": 4.4598100245619505e-06, |
| "loss": 0.6187, |
| "step": 1880 |
| }, |
| { |
| "epoch": 3.327782646801052, |
| "grad_norm": 0.3625248074531555, |
| "learning_rate": 4.3659246995699845e-06, |
| "loss": 0.6176, |
| "step": 1900 |
| }, |
| { |
| "epoch": 3.3628396143733568, |
| "grad_norm": 0.37671083211898804, |
| "learning_rate": 4.2722660617535105e-06, |
| "loss": 0.6182, |
| "step": 1920 |
| }, |
| { |
| "epoch": 3.3978965819456617, |
| "grad_norm": 0.3727245032787323, |
| "learning_rate": 4.178867594851849e-06, |
| "loss": 0.616, |
| "step": 1940 |
| }, |
| { |
| "epoch": 3.4329535495179666, |
| "grad_norm": 0.361914724111557, |
| "learning_rate": 4.085762689591054e-06, |
| "loss": 0.6157, |
| "step": 1960 |
| }, |
| { |
| "epoch": 3.4680105170902715, |
| "grad_norm": 0.3587988018989563, |
| "learning_rate": 3.992984631746469e-06, |
| "loss": 0.6188, |
| "step": 1980 |
| }, |
| { |
| "epoch": 3.5030674846625764, |
| "grad_norm": 0.3729381561279297, |
| "learning_rate": 3.9005665902427695e-06, |
| "loss": 0.6208, |
| "step": 2000 |
| }, |
| { |
| "epoch": 3.538124452234882, |
| "grad_norm": 0.35941800475120544, |
| "learning_rate": 3.8085416052958107e-06, |
| "loss": 0.6183, |
| "step": 2020 |
| }, |
| { |
| "epoch": 3.5731814198071867, |
| "grad_norm": 0.35061511397361755, |
| "learning_rate": 3.7169425766004653e-06, |
| "loss": 0.6148, |
| "step": 2040 |
| }, |
| { |
| "epoch": 3.6082383873794917, |
| "grad_norm": 0.36660343408584595, |
| "learning_rate": 3.6258022515687215e-06, |
| "loss": 0.6186, |
| "step": 2060 |
| }, |
| { |
| "epoch": 3.6432953549517966, |
| "grad_norm": 0.4035237729549408, |
| "learning_rate": 3.5351532136222012e-06, |
| "loss": 0.616, |
| "step": 2080 |
| }, |
| { |
| "epoch": 3.678352322524102, |
| "grad_norm": 0.3646794855594635, |
| "learning_rate": 3.445027870543323e-06, |
| "loss": 0.6145, |
| "step": 2100 |
| }, |
| { |
| "epoch": 3.713409290096407, |
| "grad_norm": 0.38012927770614624, |
| "learning_rate": 3.3554584428892488e-06, |
| "loss": 0.6181, |
| "step": 2120 |
| }, |
| { |
| "epoch": 3.7484662576687118, |
| "grad_norm": 0.36509743332862854, |
| "learning_rate": 3.2664769524727712e-06, |
| "loss": 0.616, |
| "step": 2140 |
| }, |
| { |
| "epoch": 3.7835232252410167, |
| "grad_norm": 0.38520926237106323, |
| "learning_rate": 3.178115210914242e-06, |
| "loss": 0.6169, |
| "step": 2160 |
| }, |
| { |
| "epoch": 3.8185801928133216, |
| "grad_norm": 0.3496517837047577, |
| "learning_rate": 3.0904048082686655e-06, |
| "loss": 0.6158, |
| "step": 2180 |
| }, |
| { |
| "epoch": 3.8536371603856265, |
| "grad_norm": 0.37065935134887695, |
| "learning_rate": 3.00337710173198e-06, |
| "loss": 0.6162, |
| "step": 2200 |
| }, |
| { |
| "epoch": 3.8886941279579315, |
| "grad_norm": 0.339855819940567, |
| "learning_rate": 2.9170632044306137e-06, |
| "loss": 0.6156, |
| "step": 2220 |
| }, |
| { |
| "epoch": 3.9237510955302364, |
| "grad_norm": 0.3591175675392151, |
| "learning_rate": 2.8314939742982673e-06, |
| "loss": 0.6144, |
| "step": 2240 |
| }, |
| { |
| "epoch": 3.9588080631025417, |
| "grad_norm": 0.3600142300128937, |
| "learning_rate": 2.746700003043964e-06, |
| "loss": 0.6195, |
| "step": 2260 |
| }, |
| { |
| "epoch": 3.9938650306748467, |
| "grad_norm": 0.3738831579685211, |
| "learning_rate": 2.662711605215248e-06, |
| "loss": 0.6135, |
| "step": 2280 |
| }, |
| { |
| "epoch": 4.028045574057844, |
| "grad_norm": 0.3281383514404297, |
| "learning_rate": 2.579558807360489e-06, |
| "loss": 0.6104, |
| "step": 2300 |
| }, |
| { |
| "epoch": 4.063102541630149, |
| "grad_norm": 0.3486866354942322, |
| "learning_rate": 2.4972713372941406e-06, |
| "loss": 0.6095, |
| "step": 2320 |
| }, |
| { |
| "epoch": 4.098159509202454, |
| "grad_norm": 0.3392680287361145, |
| "learning_rate": 2.4158786134687966e-06, |
| "loss": 0.6108, |
| "step": 2340 |
| }, |
| { |
| "epoch": 4.133216476774759, |
| "grad_norm": 0.3901905417442322, |
| "learning_rate": 2.3354097344578565e-06, |
| "loss": 0.6123, |
| "step": 2360 |
| }, |
| { |
| "epoch": 4.168273444347064, |
| "grad_norm": 0.33477798104286194, |
| "learning_rate": 2.25589346855254e-06, |
| "loss": 0.6079, |
| "step": 2380 |
| }, |
| { |
| "epoch": 4.203330411919369, |
| "grad_norm": 0.35764065384864807, |
| "learning_rate": 2.1773582434769854e-06, |
| "loss": 0.6066, |
| "step": 2400 |
| }, |
| { |
| "epoch": 4.238387379491674, |
| "grad_norm": 0.34379830956459045, |
| "learning_rate": 2.0998321362251036e-06, |
| "loss": 0.608, |
| "step": 2420 |
| }, |
| { |
| "epoch": 4.273444347063979, |
| "grad_norm": 0.3394622206687927, |
| "learning_rate": 2.023342863022819e-06, |
| "loss": 0.6142, |
| "step": 2440 |
| }, |
| { |
| "epoch": 4.308501314636284, |
| "grad_norm": 0.35136643052101135, |
| "learning_rate": 1.94791776941929e-06, |
| "loss": 0.6115, |
| "step": 2460 |
| }, |
| { |
| "epoch": 4.343558282208589, |
| "grad_norm": 0.3298383355140686, |
| "learning_rate": 1.873583820510647e-06, |
| "loss": 0.6089, |
| "step": 2480 |
| }, |
| { |
| "epoch": 4.378615249780894, |
| "grad_norm": 0.3390386402606964, |
| "learning_rate": 1.8003675912997487e-06, |
| "loss": 0.612, |
| "step": 2500 |
| }, |
| { |
| "epoch": 4.413672217353199, |
| "grad_norm": 0.3399540185928345, |
| "learning_rate": 1.7282952571953987e-06, |
| "loss": 0.6123, |
| "step": 2520 |
| }, |
| { |
| "epoch": 4.448729184925504, |
| "grad_norm": 0.3504091203212738, |
| "learning_rate": 1.657392584654412e-06, |
| "loss": 0.6078, |
| "step": 2540 |
| }, |
| { |
| "epoch": 4.483786152497809, |
| "grad_norm": 0.3221462070941925, |
| "learning_rate": 1.587684921969912e-06, |
| "loss": 0.6142, |
| "step": 2560 |
| }, |
| { |
| "epoch": 4.518843120070114, |
| "grad_norm": 0.34956008195877075, |
| "learning_rate": 1.5191971902090797e-06, |
| "loss": 0.6085, |
| "step": 2580 |
| }, |
| { |
| "epoch": 4.553900087642419, |
| "grad_norm": 0.34091153740882874, |
| "learning_rate": 1.4519538743036927e-06, |
| "loss": 0.6117, |
| "step": 2600 |
| }, |
| { |
| "epoch": 4.588957055214724, |
| "grad_norm": 0.34645262360572815, |
| "learning_rate": 1.385979014296533e-06, |
| "loss": 0.6076, |
| "step": 2620 |
| }, |
| { |
| "epoch": 4.624014022787029, |
| "grad_norm": 0.3348851799964905, |
| "learning_rate": 1.3212961967468985e-06, |
| "loss": 0.6116, |
| "step": 2640 |
| }, |
| { |
| "epoch": 4.659070990359334, |
| "grad_norm": 0.32889384031295776, |
| "learning_rate": 1.2579285462981855e-06, |
| "loss": 0.6129, |
| "step": 2660 |
| }, |
| { |
| "epoch": 4.694127957931639, |
| "grad_norm": 0.3461220860481262, |
| "learning_rate": 1.195898717410664e-06, |
| "loss": 0.6106, |
| "step": 2680 |
| }, |
| { |
| "epoch": 4.729184925503944, |
| "grad_norm": 0.319545716047287, |
| "learning_rate": 1.1352288862622968e-06, |
| "loss": 0.6068, |
| "step": 2700 |
| }, |
| { |
| "epoch": 4.764241893076249, |
| "grad_norm": 0.3272876739501953, |
| "learning_rate": 1.075940742820588e-06, |
| "loss": 0.6082, |
| "step": 2720 |
| }, |
| { |
| "epoch": 4.799298860648554, |
| "grad_norm": 0.3398887515068054, |
| "learning_rate": 1.0180554830882333e-06, |
| "loss": 0.6069, |
| "step": 2740 |
| }, |
| { |
| "epoch": 4.8343558282208585, |
| "grad_norm": 0.32064250111579895, |
| "learning_rate": 9.615938015253723e-07, |
| "loss": 0.6086, |
| "step": 2760 |
| }, |
| { |
| "epoch": 4.869412795793163, |
| "grad_norm": 0.32411012053489685, |
| "learning_rate": 9.065758836511556e-07, |
| "loss": 0.6083, |
| "step": 2780 |
| }, |
| { |
| "epoch": 4.904469763365469, |
| "grad_norm": 0.31640660762786865, |
| "learning_rate": 8.53021398827239e-07, |
| "loss": 0.6094, |
| "step": 2800 |
| }, |
| { |
| "epoch": 4.939526730937774, |
| "grad_norm": 0.3243345320224762, |
| "learning_rate": 8.009494932258427e-07, |
| "loss": 0.6104, |
| "step": 2820 |
| }, |
| { |
| "epoch": 4.974583698510079, |
| "grad_norm": 0.31760430335998535, |
| "learning_rate": 7.503787829848191e-07, |
| "loss": 0.6078, |
| "step": 2840 |
| }, |
| { |
| "epoch": 5.0087642418930765, |
| "grad_norm": 0.32336053252220154, |
| "learning_rate": 7.013273475522392e-07, |
| "loss": 0.609, |
| "step": 2860 |
| }, |
| { |
| "epoch": 5.043821209465381, |
| "grad_norm": 0.31958338618278503, |
| "learning_rate": 6.53812723222838e-07, |
| "loss": 0.6073, |
| "step": 2880 |
| }, |
| { |
| "epoch": 5.078878177037686, |
| "grad_norm": 0.319396436214447, |
| "learning_rate": 6.078518968686426e-07, |
| "loss": 0.6096, |
| "step": 2900 |
| }, |
| { |
| "epoch": 5.113935144609991, |
| "grad_norm": 0.33287620544433594, |
| "learning_rate": 5.634612998660249e-07, |
| "loss": 0.6011, |
| "step": 2920 |
| }, |
| { |
| "epoch": 5.148992112182296, |
| "grad_norm": 0.32154789566993713, |
| "learning_rate": 5.206568022213482e-07, |
| "loss": 0.6081, |
| "step": 2940 |
| }, |
| { |
| "epoch": 5.184049079754601, |
| "grad_norm": 0.3217925727367401, |
| "learning_rate": 4.794537068973093e-07, |
| "loss": 0.607, |
| "step": 2960 |
| }, |
| { |
| "epoch": 5.219106047326906, |
| "grad_norm": 0.3143027722835541, |
| "learning_rate": 4.398667443420029e-07, |
| "loss": 0.6065, |
| "step": 2980 |
| }, |
| { |
| "epoch": 5.254163014899211, |
| "grad_norm": 0.3175284266471863, |
| "learning_rate": 4.019100672226617e-07, |
| "loss": 0.6056, |
| "step": 3000 |
| }, |
| { |
| "epoch": 5.289219982471516, |
| "grad_norm": 0.32859277725219727, |
| "learning_rate": 3.65597245365964e-07, |
| "loss": 0.6063, |
| "step": 3020 |
| }, |
| { |
| "epoch": 5.324276950043822, |
| "grad_norm": 0.3200486898422241, |
| "learning_rate": 3.3094126090670477e-07, |
| "loss": 0.6074, |
| "step": 3040 |
| }, |
| { |
| "epoch": 5.3593339176161265, |
| "grad_norm": 0.3193705976009369, |
| "learning_rate": 2.9795450364657865e-07, |
| "loss": 0.607, |
| "step": 3060 |
| }, |
| { |
| "epoch": 5.3943908851884315, |
| "grad_norm": 0.32681772112846375, |
| "learning_rate": 2.6664876662471697e-07, |
| "loss": 0.6053, |
| "step": 3080 |
| }, |
| { |
| "epoch": 5.429447852760736, |
| "grad_norm": 0.31692323088645935, |
| "learning_rate": 2.370352419015892e-07, |
| "loss": 0.608, |
| "step": 3100 |
| }, |
| { |
| "epoch": 5.464504820333041, |
| "grad_norm": 0.32039353251457214, |
| "learning_rate": 2.091245165577349e-07, |
| "loss": 0.6079, |
| "step": 3120 |
| }, |
| { |
| "epoch": 5.499561787905346, |
| "grad_norm": 0.31619492173194885, |
| "learning_rate": 1.8292656890880722e-07, |
| "loss": 0.6045, |
| "step": 3140 |
| }, |
| { |
| "epoch": 5.534618755477651, |
| "grad_norm": 0.31279516220092773, |
| "learning_rate": 1.5845076493823331e-07, |
| "loss": 0.6066, |
| "step": 3160 |
| }, |
| { |
| "epoch": 5.569675723049956, |
| "grad_norm": 0.3182794451713562, |
| "learning_rate": 1.3570585494880328e-07, |
| "loss": 0.6085, |
| "step": 3180 |
| }, |
| { |
| "epoch": 5.604732690622261, |
| "grad_norm": 0.31176891922950745, |
| "learning_rate": 1.1469997043436154e-07, |
| "loss": 0.6048, |
| "step": 3200 |
| }, |
| { |
| "epoch": 5.639789658194566, |
| "grad_norm": 0.31051337718963623, |
| "learning_rate": 9.544062117273045e-08, |
| "loss": 0.607, |
| "step": 3220 |
| }, |
| { |
| "epoch": 5.674846625766871, |
| "grad_norm": 0.31119874119758606, |
| "learning_rate": 7.793469254090524e-08, |
| "loss": 0.6054, |
| "step": 3240 |
| }, |
| { |
| "epoch": 5.709903593339176, |
| "grad_norm": 0.3019949793815613, |
| "learning_rate": 6.218844305346916e-08, |
| "loss": 0.6074, |
| "step": 3260 |
| }, |
| { |
| "epoch": 5.7449605609114816, |
| "grad_norm": 0.30909818410873413, |
| "learning_rate": 4.820750212513048e-08, |
| "loss": 0.6085, |
| "step": 3280 |
| }, |
| { |
| "epoch": 5.7800175284837865, |
| "grad_norm": 0.30439695715904236, |
| "learning_rate": 3.599686805815128e-08, |
| "loss": 0.6101, |
| "step": 3300 |
| }, |
| { |
| "epoch": 5.815074496056091, |
| "grad_norm": 0.3216073215007782, |
| "learning_rate": 2.5560906255420737e-08, |
| "loss": 0.6092, |
| "step": 3320 |
| }, |
| { |
| "epoch": 5.850131463628396, |
| "grad_norm": 0.3213990330696106, |
| "learning_rate": 1.6903347659781856e-08, |
| "loss": 0.6071, |
| "step": 3340 |
| }, |
| { |
| "epoch": 5.885188431200701, |
| "grad_norm": 0.3153989613056183, |
| "learning_rate": 1.0027287420192322e-08, |
| "loss": 0.6074, |
| "step": 3360 |
| }, |
| { |
| "epoch": 5.920245398773006, |
| "grad_norm": 0.308703750371933, |
| "learning_rate": 4.935183785180209e-09, |
| "loss": 0.6058, |
| "step": 3380 |
| }, |
| { |
| "epoch": 5.955302366345311, |
| "grad_norm": 0.3108364939689636, |
| "learning_rate": 1.6288572240014123e-09, |
| "loss": 0.6038, |
| "step": 3400 |
| }, |
| { |
| "epoch": 5.990359333917616, |
| "grad_norm": 0.313579261302948, |
| "learning_rate": 1.0948977580638176e-10, |
| "loss": 0.6079, |
| "step": 3420 |
| }, |
| { |
| "epoch": 6.0, |
| "step": 3426, |
| "total_flos": 9.485754681343869e+18, |
| "train_loss": 0.6827397172519218, |
| "train_runtime": 47545.1809, |
| "train_samples_per_second": 4.607, |
| "train_steps_per_second": 0.072 |
| } |
| ], |
| "logging_steps": 20, |
| "max_steps": 3426, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 6, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 9.485754681343869e+18, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|