| { |
| "best_global_step": 7028, |
| "best_metric": 0.767158567905426, |
| "best_model_checkpoint": "./mcqa_qwen3_letter/checkpoint-7028", |
| "epoch": 0.9998932953939178, |
| "eval_steps": 500, |
| "global_step": 7028, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.007113640405477503, |
| "grad_norm": 20.374778747558594, |
| "learning_rate": 1.337126600284495e-06, |
| "loss": 2.342, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.014227280810955006, |
| "grad_norm": 23.720001220703125, |
| "learning_rate": 2.7596017069701283e-06, |
| "loss": 1.6092, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.02134092121643251, |
| "grad_norm": 14.003034591674805, |
| "learning_rate": 4.182076813655762e-06, |
| "loss": 1.4564, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.028454561621910013, |
| "grad_norm": 11.145376205444336, |
| "learning_rate": 5.604551920341395e-06, |
| "loss": 1.481, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.035568202027387516, |
| "grad_norm": 14.064179420471191, |
| "learning_rate": 7.027027027027028e-06, |
| "loss": 1.4483, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.04268184243286502, |
| "grad_norm": 14.6947021484375, |
| "learning_rate": 8.44950213371266e-06, |
| "loss": 1.4355, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.04979548283834252, |
| "grad_norm": 11.520552635192871, |
| "learning_rate": 9.871977240398294e-06, |
| "loss": 1.4314, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.056909123243820026, |
| "grad_norm": 11.234253883361816, |
| "learning_rate": 1.1294452347083926e-05, |
| "loss": 1.3494, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.06402276364929753, |
| "grad_norm": 9.99629020690918, |
| "learning_rate": 1.271692745376956e-05, |
| "loss": 1.3342, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.07113640405477503, |
| "grad_norm": 8.978431701660156, |
| "learning_rate": 1.4139402560455193e-05, |
| "loss": 1.2951, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.07825004446025254, |
| "grad_norm": 9.42021369934082, |
| "learning_rate": 1.5561877667140826e-05, |
| "loss": 1.3058, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.08536368486573004, |
| "grad_norm": 9.690632820129395, |
| "learning_rate": 1.698435277382646e-05, |
| "loss": 1.2561, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.09247732527120754, |
| "grad_norm": 10.151556015014648, |
| "learning_rate": 1.8406827880512092e-05, |
| "loss": 1.2415, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.09959096567668504, |
| "grad_norm": 8.446418762207031, |
| "learning_rate": 1.9829302987197725e-05, |
| "loss": 1.3303, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.10670460608216255, |
| "grad_norm": 9.451847076416016, |
| "learning_rate": 1.9860869565217395e-05, |
| "loss": 1.2735, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.11381824648764005, |
| "grad_norm": 9.343132019042969, |
| "learning_rate": 1.9702766798418973e-05, |
| "loss": 1.3471, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.12093188689311755, |
| "grad_norm": 9.069653511047363, |
| "learning_rate": 1.9544664031620554e-05, |
| "loss": 1.2535, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.12804552729859506, |
| "grad_norm": 5.554582595825195, |
| "learning_rate": 1.9386561264822135e-05, |
| "loss": 1.2887, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.13515916770407255, |
| "grad_norm": 8.713581085205078, |
| "learning_rate": 1.9228458498023717e-05, |
| "loss": 1.1681, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.14227280810955006, |
| "grad_norm": 8.742165565490723, |
| "learning_rate": 1.9070355731225298e-05, |
| "loss": 1.2271, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.14938644851502755, |
| "grad_norm": 8.755561828613281, |
| "learning_rate": 1.891225296442688e-05, |
| "loss": 1.2557, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.15650008892050507, |
| "grad_norm": 7.881649017333984, |
| "learning_rate": 1.875415019762846e-05, |
| "loss": 1.1999, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.16361372932598256, |
| "grad_norm": 9.955330848693848, |
| "learning_rate": 1.8596047430830042e-05, |
| "loss": 1.1375, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.17072736973146008, |
| "grad_norm": 9.453360557556152, |
| "learning_rate": 1.843794466403162e-05, |
| "loss": 1.2173, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.17784101013693757, |
| "grad_norm": 6.806760787963867, |
| "learning_rate": 1.8279841897233205e-05, |
| "loss": 1.2116, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.18495465054241508, |
| "grad_norm": 13.604923248291016, |
| "learning_rate": 1.8121739130434783e-05, |
| "loss": 1.0788, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.19206829094789257, |
| "grad_norm": 10.025015830993652, |
| "learning_rate": 1.7963636363636364e-05, |
| "loss": 1.149, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.1991819313533701, |
| "grad_norm": 9.246174812316895, |
| "learning_rate": 1.7805533596837945e-05, |
| "loss": 1.1171, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.20629557175884758, |
| "grad_norm": 9.556463241577148, |
| "learning_rate": 1.7647430830039527e-05, |
| "loss": 1.1744, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.2134092121643251, |
| "grad_norm": 13.477895736694336, |
| "learning_rate": 1.7489328063241108e-05, |
| "loss": 1.0723, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.22052285256980259, |
| "grad_norm": 7.583223342895508, |
| "learning_rate": 1.733122529644269e-05, |
| "loss": 1.1004, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.2276364929752801, |
| "grad_norm": 9.056886672973633, |
| "learning_rate": 1.717312252964427e-05, |
| "loss": 1.1413, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.2347501333807576, |
| "grad_norm": 8.818636894226074, |
| "learning_rate": 1.701501976284585e-05, |
| "loss": 1.0248, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.2418637737862351, |
| "grad_norm": 11.717618942260742, |
| "learning_rate": 1.6856916996047434e-05, |
| "loss": 1.0386, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.2489774141917126, |
| "grad_norm": 7.414971828460693, |
| "learning_rate": 1.6698814229249015e-05, |
| "loss": 1.0885, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.2560910545971901, |
| "grad_norm": 6.960204601287842, |
| "learning_rate": 1.6540711462450593e-05, |
| "loss": 1.053, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.26320469500266763, |
| "grad_norm": 6.861100673675537, |
| "learning_rate": 1.6382608695652174e-05, |
| "loss": 1.0398, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.2703183354081451, |
| "grad_norm": 7.645370960235596, |
| "learning_rate": 1.6224505928853756e-05, |
| "loss": 0.9453, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.2774319758136226, |
| "grad_norm": 7.169046878814697, |
| "learning_rate": 1.6066403162055337e-05, |
| "loss": 0.9847, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.28454561621910013, |
| "grad_norm": 6.290506362915039, |
| "learning_rate": 1.5908300395256918e-05, |
| "loss": 0.9592, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.29165925662457765, |
| "grad_norm": 8.001459121704102, |
| "learning_rate": 1.57501976284585e-05, |
| "loss": 0.9674, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.2987728970300551, |
| "grad_norm": 8.543235778808594, |
| "learning_rate": 1.559525691699605e-05, |
| "loss": 1.003, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.3058865374355326, |
| "grad_norm": 10.702069282531738, |
| "learning_rate": 1.543715415019763e-05, |
| "loss": 0.9914, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.31300017784101014, |
| "grad_norm": 5.234061241149902, |
| "learning_rate": 1.527905138339921e-05, |
| "loss": 1.0809, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.32011381824648766, |
| "grad_norm": 8.00466251373291, |
| "learning_rate": 1.5120948616600793e-05, |
| "loss": 0.9713, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.3272274586519651, |
| "grad_norm": 7.484270095825195, |
| "learning_rate": 1.4962845849802372e-05, |
| "loss": 0.906, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.33434109905744264, |
| "grad_norm": 17.150543212890625, |
| "learning_rate": 1.4804743083003954e-05, |
| "loss": 1.0206, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.34145473946292015, |
| "grad_norm": 8.58785343170166, |
| "learning_rate": 1.4646640316205533e-05, |
| "loss": 0.9868, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.34856837986839767, |
| "grad_norm": 7.040744304656982, |
| "learning_rate": 1.4488537549407116e-05, |
| "loss": 0.9225, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.35568202027387513, |
| "grad_norm": 8.559785842895508, |
| "learning_rate": 1.4330434782608698e-05, |
| "loss": 0.9345, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.36279566067935265, |
| "grad_norm": 9.140971183776855, |
| "learning_rate": 1.4172332015810277e-05, |
| "loss": 1.004, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.36990930108483017, |
| "grad_norm": 8.379377365112305, |
| "learning_rate": 1.4014229249011859e-05, |
| "loss": 1.0199, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.3770229414903077, |
| "grad_norm": 6.753004550933838, |
| "learning_rate": 1.3856126482213438e-05, |
| "loss": 0.9463, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.38413658189578515, |
| "grad_norm": 7.884097576141357, |
| "learning_rate": 1.3698023715415021e-05, |
| "loss": 0.9367, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.39125022230126266, |
| "grad_norm": 8.405110359191895, |
| "learning_rate": 1.3539920948616603e-05, |
| "loss": 0.9036, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.3983638627067402, |
| "grad_norm": 7.3707661628723145, |
| "learning_rate": 1.3381818181818182e-05, |
| "loss": 0.9868, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.4054775031122177, |
| "grad_norm": 9.472613334655762, |
| "learning_rate": 1.3223715415019764e-05, |
| "loss": 0.9809, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.41259114351769516, |
| "grad_norm": 8.191240310668945, |
| "learning_rate": 1.3065612648221343e-05, |
| "loss": 0.9249, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.4197047839231727, |
| "grad_norm": 7.100245952606201, |
| "learning_rate": 1.2907509881422926e-05, |
| "loss": 0.9303, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.4268184243286502, |
| "grad_norm": 6.30331563949585, |
| "learning_rate": 1.2749407114624508e-05, |
| "loss": 0.9459, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.4339320647341277, |
| "grad_norm": 5.755526065826416, |
| "learning_rate": 1.2591304347826087e-05, |
| "loss": 0.8818, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.44104570513960517, |
| "grad_norm": 7.514272689819336, |
| "learning_rate": 1.243320158102767e-05, |
| "loss": 0.8817, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.4481593455450827, |
| "grad_norm": 6.345920562744141, |
| "learning_rate": 1.227509881422925e-05, |
| "loss": 0.8257, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.4552729859505602, |
| "grad_norm": 5.915317058563232, |
| "learning_rate": 1.2120158102766799e-05, |
| "loss": 0.9691, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.4623866263560377, |
| "grad_norm": 4.785401821136475, |
| "learning_rate": 1.1962055335968382e-05, |
| "loss": 0.97, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.4695002667615152, |
| "grad_norm": 8.403907775878906, |
| "learning_rate": 1.1803952569169962e-05, |
| "loss": 0.8699, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.4766139071669927, |
| "grad_norm": 8.583006858825684, |
| "learning_rate": 1.1645849802371543e-05, |
| "loss": 0.9233, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.4837275475724702, |
| "grad_norm": 7.836755752563477, |
| "learning_rate": 1.1487747035573123e-05, |
| "loss": 0.8716, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.49084118797794773, |
| "grad_norm": 4.714875221252441, |
| "learning_rate": 1.1329644268774704e-05, |
| "loss": 0.8499, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.4979548283834252, |
| "grad_norm": 7.097923755645752, |
| "learning_rate": 1.1171541501976287e-05, |
| "loss": 0.912, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.5050684687889028, |
| "grad_norm": 4.853347301483154, |
| "learning_rate": 1.1013438735177867e-05, |
| "loss": 0.8411, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.5121821091943802, |
| "grad_norm": 6.642279148101807, |
| "learning_rate": 1.0855335968379448e-05, |
| "loss": 0.9325, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.5192957495998577, |
| "grad_norm": 6.702445030212402, |
| "learning_rate": 1.0697233201581028e-05, |
| "loss": 0.8354, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.5264093900053353, |
| "grad_norm": 6.634207725524902, |
| "learning_rate": 1.0539130434782609e-05, |
| "loss": 0.8813, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.5335230304108127, |
| "grad_norm": 4.122372627258301, |
| "learning_rate": 1.0381027667984192e-05, |
| "loss": 0.7811, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.5406366708162902, |
| "grad_norm": 8.899633407592773, |
| "learning_rate": 1.0222924901185772e-05, |
| "loss": 0.8707, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.5477503112217678, |
| "grad_norm": 5.374997615814209, |
| "learning_rate": 1.0064822134387353e-05, |
| "loss": 0.8854, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.5548639516272452, |
| "grad_norm": 7.9275102615356445, |
| "learning_rate": 9.906719367588934e-06, |
| "loss": 0.9255, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.5619775920327228, |
| "grad_norm": 6.925314903259277, |
| "learning_rate": 9.748616600790514e-06, |
| "loss": 0.9027, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.5690912324382003, |
| "grad_norm": 9.117793083190918, |
| "learning_rate": 9.590513833992095e-06, |
| "loss": 0.8831, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.5762048728436777, |
| "grad_norm": 5.861302852630615, |
| "learning_rate": 9.432411067193677e-06, |
| "loss": 0.8329, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.5833185132491553, |
| "grad_norm": 6.011074066162109, |
| "learning_rate": 9.274308300395256e-06, |
| "loss": 0.8766, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.5904321536546328, |
| "grad_norm": 5.763038158416748, |
| "learning_rate": 9.11620553359684e-06, |
| "loss": 0.9374, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.5975457940601102, |
| "grad_norm": 5.520111083984375, |
| "learning_rate": 8.95810276679842e-06, |
| "loss": 0.8813, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.6046594344655878, |
| "grad_norm": 7.351500511169434, |
| "learning_rate": 8.8e-06, |
| "loss": 0.8677, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.6117730748710652, |
| "grad_norm": 5.451608657836914, |
| "learning_rate": 8.641897233201582e-06, |
| "loss": 0.8606, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.6188867152765428, |
| "grad_norm": 5.458895683288574, |
| "learning_rate": 8.483794466403163e-06, |
| "loss": 0.8522, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.6260003556820203, |
| "grad_norm": 5.554731369018555, |
| "learning_rate": 8.325691699604743e-06, |
| "loss": 0.9042, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.6331139960874977, |
| "grad_norm": 3.6643424034118652, |
| "learning_rate": 8.167588932806326e-06, |
| "loss": 0.8696, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.6402276364929753, |
| "grad_norm": 5.593344688415527, |
| "learning_rate": 8.009486166007906e-06, |
| "loss": 0.7663, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.6473412768984528, |
| "grad_norm": 7.597021102905273, |
| "learning_rate": 7.851383399209487e-06, |
| "loss": 0.7752, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.6544549173039302, |
| "grad_norm": 4.095972537994385, |
| "learning_rate": 7.693280632411068e-06, |
| "loss": 0.897, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.6615685577094078, |
| "grad_norm": 7.817249298095703, |
| "learning_rate": 7.535177865612649e-06, |
| "loss": 0.7448, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.6686821981148853, |
| "grad_norm": 3.296755313873291, |
| "learning_rate": 7.37707509881423e-06, |
| "loss": 0.8065, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.6757958385203628, |
| "grad_norm": 5.186858177185059, |
| "learning_rate": 7.218972332015811e-06, |
| "loss": 0.8618, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.6829094789258403, |
| "grad_norm": 6.484861373901367, |
| "learning_rate": 7.060869565217392e-06, |
| "loss": 0.8549, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.6900231193313178, |
| "grad_norm": 4.90764856338501, |
| "learning_rate": 6.902766798418972e-06, |
| "loss": 0.9441, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.6971367597367953, |
| "grad_norm": 3.6083099842071533, |
| "learning_rate": 6.744664031620554e-06, |
| "loss": 0.7467, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.7042504001422728, |
| "grad_norm": 4.79904317855835, |
| "learning_rate": 6.586561264822134e-06, |
| "loss": 0.7974, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.7113640405477503, |
| "grad_norm": 8.30154800415039, |
| "learning_rate": 6.4284584980237165e-06, |
| "loss": 0.8927, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.7184776809532278, |
| "grad_norm": 5.582029819488525, |
| "learning_rate": 6.270355731225297e-06, |
| "loss": 0.8348, |
| "step": 5050 |
| }, |
| { |
| "epoch": 0.7255913213587053, |
| "grad_norm": 7.575035572052002, |
| "learning_rate": 6.112252964426878e-06, |
| "loss": 0.8691, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.7327049617641829, |
| "grad_norm": 6.100114345550537, |
| "learning_rate": 5.954150197628459e-06, |
| "loss": 0.788, |
| "step": 5150 |
| }, |
| { |
| "epoch": 0.7398186021696603, |
| "grad_norm": 7.05925178527832, |
| "learning_rate": 5.796047430830039e-06, |
| "loss": 0.8325, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.7469322425751378, |
| "grad_norm": 6.5520920753479, |
| "learning_rate": 5.6379446640316215e-06, |
| "loss": 0.7792, |
| "step": 5250 |
| }, |
| { |
| "epoch": 0.7540458829806154, |
| "grad_norm": 6.9081807136535645, |
| "learning_rate": 5.479841897233202e-06, |
| "loss": 0.805, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.7611595233860928, |
| "grad_norm": 7.752345085144043, |
| "learning_rate": 5.321739130434783e-06, |
| "loss": 0.7998, |
| "step": 5350 |
| }, |
| { |
| "epoch": 0.7682731637915703, |
| "grad_norm": 3.8013064861297607, |
| "learning_rate": 5.163636363636364e-06, |
| "loss": 0.7748, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.7753868041970479, |
| "grad_norm": 5.316600322723389, |
| "learning_rate": 5.005533596837945e-06, |
| "loss": 0.7644, |
| "step": 5450 |
| }, |
| { |
| "epoch": 0.7825004446025253, |
| "grad_norm": 8.250207901000977, |
| "learning_rate": 4.8474308300395266e-06, |
| "loss": 0.7766, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.7896140850080029, |
| "grad_norm": 5.134970664978027, |
| "learning_rate": 4.689328063241107e-06, |
| "loss": 0.79, |
| "step": 5550 |
| }, |
| { |
| "epoch": 0.7967277254134804, |
| "grad_norm": 5.832414627075195, |
| "learning_rate": 4.531225296442688e-06, |
| "loss": 0.9154, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.8038413658189578, |
| "grad_norm": 6.164233684539795, |
| "learning_rate": 4.373122529644269e-06, |
| "loss": 0.8288, |
| "step": 5650 |
| }, |
| { |
| "epoch": 0.8109550062244354, |
| "grad_norm": 11.60009479522705, |
| "learning_rate": 4.21501976284585e-06, |
| "loss": 0.7047, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.8180686466299129, |
| "grad_norm": 3.422287702560425, |
| "learning_rate": 4.056916996047432e-06, |
| "loss": 0.7387, |
| "step": 5750 |
| }, |
| { |
| "epoch": 0.8251822870353903, |
| "grad_norm": 3.786428689956665, |
| "learning_rate": 3.898814229249012e-06, |
| "loss": 0.806, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.8322959274408679, |
| "grad_norm": 3.7693772315979004, |
| "learning_rate": 3.740711462450593e-06, |
| "loss": 0.8136, |
| "step": 5850 |
| }, |
| { |
| "epoch": 0.8394095678463454, |
| "grad_norm": 3.2537221908569336, |
| "learning_rate": 3.5826086956521744e-06, |
| "loss": 0.7834, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.8465232082518229, |
| "grad_norm": 2.938720226287842, |
| "learning_rate": 3.4245059288537553e-06, |
| "loss": 0.7607, |
| "step": 5950 |
| }, |
| { |
| "epoch": 0.8536368486573004, |
| "grad_norm": 6.700362205505371, |
| "learning_rate": 3.266403162055336e-06, |
| "loss": 0.8713, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.8607504890627778, |
| "grad_norm": 4.768782615661621, |
| "learning_rate": 3.1083003952569176e-06, |
| "loss": 0.8086, |
| "step": 6050 |
| }, |
| { |
| "epoch": 0.8678641294682554, |
| "grad_norm": 6.537806034088135, |
| "learning_rate": 2.950197628458498e-06, |
| "loss": 0.8722, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.8749777698737329, |
| "grad_norm": 6.817727088928223, |
| "learning_rate": 2.792094861660079e-06, |
| "loss": 0.7126, |
| "step": 6150 |
| }, |
| { |
| "epoch": 0.8820914102792103, |
| "grad_norm": 5.932295799255371, |
| "learning_rate": 2.6339920948616604e-06, |
| "loss": 0.778, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.8892050506846879, |
| "grad_norm": 6.475405693054199, |
| "learning_rate": 2.4758893280632413e-06, |
| "loss": 0.7812, |
| "step": 6250 |
| }, |
| { |
| "epoch": 0.8963186910901654, |
| "grad_norm": 8.57321834564209, |
| "learning_rate": 2.3177865612648222e-06, |
| "loss": 0.7654, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.903432331495643, |
| "grad_norm": 1.9723538160324097, |
| "learning_rate": 2.1596837944664036e-06, |
| "loss": 0.7701, |
| "step": 6350 |
| }, |
| { |
| "epoch": 0.9105459719011204, |
| "grad_norm": 6.585085391998291, |
| "learning_rate": 2.001581027667984e-06, |
| "loss": 0.8419, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.9176596123065979, |
| "grad_norm": 4.754767894744873, |
| "learning_rate": 1.8434782608695654e-06, |
| "loss": 0.9006, |
| "step": 6450 |
| }, |
| { |
| "epoch": 0.9247732527120754, |
| "grad_norm": 4.661940574645996, |
| "learning_rate": 1.6853754940711464e-06, |
| "loss": 0.7784, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.9318868931175529, |
| "grad_norm": 4.972583770751953, |
| "learning_rate": 1.5272727272727275e-06, |
| "loss": 0.8112, |
| "step": 6550 |
| }, |
| { |
| "epoch": 0.9390005335230304, |
| "grad_norm": 6.641800403594971, |
| "learning_rate": 1.3691699604743084e-06, |
| "loss": 0.7466, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.9461141739285079, |
| "grad_norm": 5.098995208740234, |
| "learning_rate": 1.2110671936758893e-06, |
| "loss": 0.824, |
| "step": 6650 |
| }, |
| { |
| "epoch": 0.9532278143339854, |
| "grad_norm": 7.060574054718018, |
| "learning_rate": 1.0529644268774705e-06, |
| "loss": 0.8502, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.960341454739463, |
| "grad_norm": 4.678645610809326, |
| "learning_rate": 8.948616600790515e-07, |
| "loss": 0.8004, |
| "step": 6750 |
| }, |
| { |
| "epoch": 0.9674550951449404, |
| "grad_norm": 5.895313739776611, |
| "learning_rate": 7.367588932806324e-07, |
| "loss": 0.6811, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.9745687355504179, |
| "grad_norm": 5.316034317016602, |
| "learning_rate": 5.786561264822135e-07, |
| "loss": 0.7955, |
| "step": 6850 |
| }, |
| { |
| "epoch": 0.9816823759558955, |
| "grad_norm": 5.032801628112793, |
| "learning_rate": 4.205533596837945e-07, |
| "loss": 0.8206, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.9887960163613729, |
| "grad_norm": 5.486656188964844, |
| "learning_rate": 2.6245059288537554e-07, |
| "loss": 0.795, |
| "step": 6950 |
| }, |
| { |
| "epoch": 0.9959096567668504, |
| "grad_norm": 6.862821102142334, |
| "learning_rate": 1.0434782608695654e-07, |
| "loss": 0.676, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.9998932953939178, |
| "eval_loss": 0.767158567905426, |
| "eval_runtime": 341.0802, |
| "eval_samples_per_second": 15.454, |
| "eval_steps_per_second": 1.932, |
| "step": 7028 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 7028, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3.803879769007718e+16, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|