| { |
| "best_global_step": 14058, |
| "best_metric": 0.6273267865180969, |
| "best_model_checkpoint": "./mcqa_qwen3_letter_m3/checkpoint-14058", |
| "epoch": 2.0, |
| "eval_steps": 500, |
| "global_step": 14058, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.007113640405477503, |
| "grad_norm": 20.266469955444336, |
| "learning_rate": 4.551920341394026e-07, |
| "loss": 2.1439, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.014227280810955006, |
| "grad_norm": 21.066892623901367, |
| "learning_rate": 9.293504030346136e-07, |
| "loss": 1.4471, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.02134092121643251, |
| "grad_norm": 18.918508529663086, |
| "learning_rate": 1.4035087719298246e-06, |
| "loss": 1.3067, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.028454561621910013, |
| "grad_norm": 9.563344955444336, |
| "learning_rate": 1.8681839734471315e-06, |
| "loss": 1.2745, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.035568202027387516, |
| "grad_norm": 20.408035278320312, |
| "learning_rate": 2.3423423423423424e-06, |
| "loss": 1.2323, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.04268184243286502, |
| "grad_norm": 17.020856857299805, |
| "learning_rate": 2.8165007112375536e-06, |
| "loss": 1.2476, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.04979548283834252, |
| "grad_norm": 10.942744255065918, |
| "learning_rate": 3.2906590801327647e-06, |
| "loss": 1.2084, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.056909123243820026, |
| "grad_norm": 11.239660263061523, |
| "learning_rate": 3.7648174490279754e-06, |
| "loss": 1.1453, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.06402276364929753, |
| "grad_norm": 8.421320915222168, |
| "learning_rate": 4.2389758179231865e-06, |
| "loss": 1.1257, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.07113640405477503, |
| "grad_norm": 7.429409027099609, |
| "learning_rate": 4.713134186818398e-06, |
| "loss": 1.0896, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.07825004446025254, |
| "grad_norm": 8.3453950881958, |
| "learning_rate": 5.187292555713609e-06, |
| "loss": 1.0589, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.08536368486573004, |
| "grad_norm": 9.157875061035156, |
| "learning_rate": 5.66145092460882e-06, |
| "loss": 1.0628, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.09247732527120754, |
| "grad_norm": 9.339110374450684, |
| "learning_rate": 6.135609293504031e-06, |
| "loss": 1.0211, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.09959096567668504, |
| "grad_norm": 7.010004043579102, |
| "learning_rate": 6.609767662399243e-06, |
| "loss": 1.0684, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.10670460608216255, |
| "grad_norm": 8.618551254272461, |
| "learning_rate": 7.0839260312944525e-06, |
| "loss": 1.0467, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.11381824648764005, |
| "grad_norm": 9.206182479858398, |
| "learning_rate": 7.558084400189663e-06, |
| "loss": 1.0855, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.12093188689311755, |
| "grad_norm": 11.882657051086426, |
| "learning_rate": 8.032242769084875e-06, |
| "loss": 0.9897, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.12804552729859506, |
| "grad_norm": 6.381322860717773, |
| "learning_rate": 8.506401137980086e-06, |
| "loss": 1.0227, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.13515916770407255, |
| "grad_norm": 7.592447280883789, |
| "learning_rate": 8.980559506875298e-06, |
| "loss": 0.9508, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.14227280810955006, |
| "grad_norm": 10.622550010681152, |
| "learning_rate": 9.454717875770508e-06, |
| "loss": 0.9916, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.14938644851502755, |
| "grad_norm": 16.836315155029297, |
| "learning_rate": 9.92887624466572e-06, |
| "loss": 0.9838, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.15650008892050507, |
| "grad_norm": 8.732917785644531, |
| "learning_rate": 1.040303461356093e-05, |
| "loss": 0.9678, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.16361372932598256, |
| "grad_norm": 10.75668716430664, |
| "learning_rate": 1.0877192982456142e-05, |
| "loss": 0.9535, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.17072736973146008, |
| "grad_norm": 10.604599952697754, |
| "learning_rate": 1.1351351351351352e-05, |
| "loss": 1.0164, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.17784101013693757, |
| "grad_norm": 11.192643165588379, |
| "learning_rate": 1.1825509720246564e-05, |
| "loss": 0.9811, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.18495465054241508, |
| "grad_norm": 11.667268753051758, |
| "learning_rate": 1.2299668089141774e-05, |
| "loss": 0.9329, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.19206829094789257, |
| "grad_norm": 12.122210502624512, |
| "learning_rate": 1.2773826458036987e-05, |
| "loss": 0.9693, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.1991819313533701, |
| "grad_norm": 10.696669578552246, |
| "learning_rate": 1.3247984826932197e-05, |
| "loss": 0.9118, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.20629557175884758, |
| "grad_norm": 9.4218168258667, |
| "learning_rate": 1.3722143195827407e-05, |
| "loss": 0.9242, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.2134092121643251, |
| "grad_norm": 15.12156867980957, |
| "learning_rate": 1.4196301564722618e-05, |
| "loss": 0.9161, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.22052285256980259, |
| "grad_norm": 10.341136932373047, |
| "learning_rate": 1.4670459933617828e-05, |
| "loss": 0.9112, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.2276364929752801, |
| "grad_norm": 8.621809005737305, |
| "learning_rate": 1.5144618302513041e-05, |
| "loss": 0.923, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.2347501333807576, |
| "grad_norm": 8.037166595458984, |
| "learning_rate": 1.561877667140825e-05, |
| "loss": 0.8646, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.2418637737862351, |
| "grad_norm": 12.160765647888184, |
| "learning_rate": 1.6092935040303463e-05, |
| "loss": 0.8811, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.2489774141917126, |
| "grad_norm": 9.850648880004883, |
| "learning_rate": 1.6567093409198673e-05, |
| "loss": 0.9158, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.2560910545971901, |
| "grad_norm": 6.849952220916748, |
| "learning_rate": 1.7041251778093886e-05, |
| "loss": 0.8903, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.26320469500266763, |
| "grad_norm": 7.7248334884643555, |
| "learning_rate": 1.7515410146989096e-05, |
| "loss": 0.8853, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.2703183354081451, |
| "grad_norm": 5.612997531890869, |
| "learning_rate": 1.7989568515884306e-05, |
| "loss": 0.8407, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.2774319758136226, |
| "grad_norm": 6.731159210205078, |
| "learning_rate": 1.846372688477952e-05, |
| "loss": 0.8447, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.28454561621910013, |
| "grad_norm": 8.783773422241211, |
| "learning_rate": 1.893788525367473e-05, |
| "loss": 0.8201, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.29165925662457765, |
| "grad_norm": 8.748090744018555, |
| "learning_rate": 1.9412043622569942e-05, |
| "loss": 0.8202, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.2987728970300551, |
| "grad_norm": 6.985095977783203, |
| "learning_rate": 1.9886201991465152e-05, |
| "loss": 0.8686, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.3058865374355326, |
| "grad_norm": 9.010405540466309, |
| "learning_rate": 1.9959947299077735e-05, |
| "loss": 0.8541, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.31300017784101014, |
| "grad_norm": 10.884967803955078, |
| "learning_rate": 1.9907246376811596e-05, |
| "loss": 0.9021, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.32011381824648766, |
| "grad_norm": 6.696805953979492, |
| "learning_rate": 1.9854545454545456e-05, |
| "loss": 0.8352, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.3272274586519651, |
| "grad_norm": 6.158838272094727, |
| "learning_rate": 1.9801844532279317e-05, |
| "loss": 0.8078, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.33434109905744264, |
| "grad_norm": 10.230331420898438, |
| "learning_rate": 1.9749143610013177e-05, |
| "loss": 0.8659, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.34145473946292015, |
| "grad_norm": 7.0724945068359375, |
| "learning_rate": 1.9696442687747038e-05, |
| "loss": 0.862, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.34856837986839767, |
| "grad_norm": 9.820209503173828, |
| "learning_rate": 1.9643741765480898e-05, |
| "loss": 0.8036, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.35568202027387513, |
| "grad_norm": 6.326645374298096, |
| "learning_rate": 1.959104084321476e-05, |
| "loss": 0.8255, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.36279566067935265, |
| "grad_norm": 4.881910800933838, |
| "learning_rate": 1.953939393939394e-05, |
| "loss": 0.861, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.36990930108483017, |
| "grad_norm": 6.977218151092529, |
| "learning_rate": 1.94866930171278e-05, |
| "loss": 0.8487, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.3770229414903077, |
| "grad_norm": 6.354854106903076, |
| "learning_rate": 1.943399209486166e-05, |
| "loss": 0.8122, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.38413658189578515, |
| "grad_norm": 6.679897308349609, |
| "learning_rate": 1.9381291172595522e-05, |
| "loss": 0.8239, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.39125022230126266, |
| "grad_norm": 7.958705902099609, |
| "learning_rate": 1.9328590250329382e-05, |
| "loss": 0.7808, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.3983638627067402, |
| "grad_norm": 6.450643062591553, |
| "learning_rate": 1.9275889328063243e-05, |
| "loss": 0.8359, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.4054775031122177, |
| "grad_norm": 5.986859321594238, |
| "learning_rate": 1.9223188405797103e-05, |
| "loss": 0.8278, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.41259114351769516, |
| "grad_norm": 8.40491771697998, |
| "learning_rate": 1.9170487483530964e-05, |
| "loss": 0.7932, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.4197047839231727, |
| "grad_norm": 7.491307735443115, |
| "learning_rate": 1.9117786561264824e-05, |
| "loss": 0.7932, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.4268184243286502, |
| "grad_norm": 5.475683689117432, |
| "learning_rate": 1.9065085638998684e-05, |
| "loss": 0.8292, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.4339320647341277, |
| "grad_norm": 6.489156723022461, |
| "learning_rate": 1.9012384716732545e-05, |
| "loss": 0.7585, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.44104570513960517, |
| "grad_norm": 6.417465686798096, |
| "learning_rate": 1.8959683794466402e-05, |
| "loss": 0.7628, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.4481593455450827, |
| "grad_norm": 4.215211868286133, |
| "learning_rate": 1.8906982872200266e-05, |
| "loss": 0.7418, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.4552729859505602, |
| "grad_norm": 5.503742218017578, |
| "learning_rate": 1.8854281949934126e-05, |
| "loss": 0.8346, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.4623866263560377, |
| "grad_norm": 5.3524651527404785, |
| "learning_rate": 1.8801581027667987e-05, |
| "loss": 0.8285, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.4695002667615152, |
| "grad_norm": 5.5221781730651855, |
| "learning_rate": 1.8748880105401844e-05, |
| "loss": 0.7363, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.4766139071669927, |
| "grad_norm": 7.9843268394470215, |
| "learning_rate": 1.8696179183135704e-05, |
| "loss": 0.7964, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.4837275475724702, |
| "grad_norm": 6.0546417236328125, |
| "learning_rate": 1.8643478260869568e-05, |
| "loss": 0.7696, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.49084118797794773, |
| "grad_norm": 5.521228313446045, |
| "learning_rate": 1.859077733860343e-05, |
| "loss": 0.7309, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.4979548283834252, |
| "grad_norm": 9.225383758544922, |
| "learning_rate": 1.853807641633729e-05, |
| "loss": 0.7869, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.5050684687889028, |
| "grad_norm": 4.0471601486206055, |
| "learning_rate": 1.8485375494071146e-05, |
| "loss": 0.7323, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.5121821091943802, |
| "grad_norm": 4.831805229187012, |
| "learning_rate": 1.8432674571805006e-05, |
| "loss": 0.7795, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.5192957495998577, |
| "grad_norm": 5.822267055511475, |
| "learning_rate": 1.837997364953887e-05, |
| "loss": 0.712, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.5264093900053353, |
| "grad_norm": 4.78722620010376, |
| "learning_rate": 1.832727272727273e-05, |
| "loss": 0.7408, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.5335230304108127, |
| "grad_norm": 5.776618480682373, |
| "learning_rate": 1.8274571805006588e-05, |
| "loss": 0.6803, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.5406366708162902, |
| "grad_norm": 11.762743949890137, |
| "learning_rate": 1.8221870882740448e-05, |
| "loss": 0.7343, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.5477503112217678, |
| "grad_norm": 3.884990692138672, |
| "learning_rate": 1.816916996047431e-05, |
| "loss": 0.7551, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.5548639516272452, |
| "grad_norm": 9.219622611999512, |
| "learning_rate": 1.8116469038208173e-05, |
| "loss": 0.764, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.5619775920327228, |
| "grad_norm": 7.808120250701904, |
| "learning_rate": 1.806376811594203e-05, |
| "loss": 0.745, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.5690912324382003, |
| "grad_norm": 7.188344955444336, |
| "learning_rate": 1.801106719367589e-05, |
| "loss": 0.7642, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.5762048728436777, |
| "grad_norm": 4.781955242156982, |
| "learning_rate": 1.795836627140975e-05, |
| "loss": 0.7089, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.5833185132491553, |
| "grad_norm": 4.489506721496582, |
| "learning_rate": 1.790566534914361e-05, |
| "loss": 0.7512, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.5904321536546328, |
| "grad_norm": 5.0979905128479, |
| "learning_rate": 1.785296442687747e-05, |
| "loss": 0.7753, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.5975457940601102, |
| "grad_norm": 5.302441596984863, |
| "learning_rate": 1.7800263504611332e-05, |
| "loss": 0.7412, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.6046594344655878, |
| "grad_norm": 4.942580223083496, |
| "learning_rate": 1.7747562582345192e-05, |
| "loss": 0.7474, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.6117730748710652, |
| "grad_norm": 4.685510158538818, |
| "learning_rate": 1.7694861660079053e-05, |
| "loss": 0.7173, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.6188867152765428, |
| "grad_norm": 4.8454999923706055, |
| "learning_rate": 1.7642160737812913e-05, |
| "loss": 0.7207, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.6260003556820203, |
| "grad_norm": 7.190736293792725, |
| "learning_rate": 1.7589459815546774e-05, |
| "loss": 0.754, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.6331139960874977, |
| "grad_norm": 7.0048675537109375, |
| "learning_rate": 1.7536758893280634e-05, |
| "loss": 0.7282, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.6402276364929753, |
| "grad_norm": 5.777256011962891, |
| "learning_rate": 1.7484057971014495e-05, |
| "loss": 0.6597, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.6473412768984528, |
| "grad_norm": 8.8043851852417, |
| "learning_rate": 1.7431357048748355e-05, |
| "loss": 0.6672, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.6544549173039302, |
| "grad_norm": 7.159558296203613, |
| "learning_rate": 1.7378656126482215e-05, |
| "loss": 0.7266, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.6615685577094078, |
| "grad_norm": 5.110541820526123, |
| "learning_rate": 1.7325955204216076e-05, |
| "loss": 0.6399, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.6686821981148853, |
| "grad_norm": 3.4727582931518555, |
| "learning_rate": 1.7273254281949936e-05, |
| "loss": 0.68, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.6757958385203628, |
| "grad_norm": 5.788763999938965, |
| "learning_rate": 1.7220553359683797e-05, |
| "loss": 0.7349, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.6829094789258403, |
| "grad_norm": 4.407217979431152, |
| "learning_rate": 1.7167852437417657e-05, |
| "loss": 0.695, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.6900231193313178, |
| "grad_norm": 4.685910224914551, |
| "learning_rate": 1.7115151515151514e-05, |
| "loss": 0.7816, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.6971367597367953, |
| "grad_norm": 2.9753873348236084, |
| "learning_rate": 1.7062450592885378e-05, |
| "loss": 0.6595, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.7042504001422728, |
| "grad_norm": 3.8326473236083984, |
| "learning_rate": 1.700974967061924e-05, |
| "loss": 0.6875, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.7113640405477503, |
| "grad_norm": 5.974637985229492, |
| "learning_rate": 1.69570487483531e-05, |
| "loss": 0.7414, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.7184776809532278, |
| "grad_norm": 3.4977550506591797, |
| "learning_rate": 1.6904347826086956e-05, |
| "loss": 0.6832, |
| "step": 5050 |
| }, |
| { |
| "epoch": 0.7255913213587053, |
| "grad_norm": 5.954402446746826, |
| "learning_rate": 1.6851646903820817e-05, |
| "loss": 0.7257, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.7327049617641829, |
| "grad_norm": 4.272510528564453, |
| "learning_rate": 1.679894598155468e-05, |
| "loss": 0.6808, |
| "step": 5150 |
| }, |
| { |
| "epoch": 0.7398186021696603, |
| "grad_norm": 4.155337810516357, |
| "learning_rate": 1.674624505928854e-05, |
| "loss": 0.7062, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.7469322425751378, |
| "grad_norm": 4.898423671722412, |
| "learning_rate": 1.6693544137022398e-05, |
| "loss": 0.6549, |
| "step": 5250 |
| }, |
| { |
| "epoch": 0.7540458829806154, |
| "grad_norm": 6.021943092346191, |
| "learning_rate": 1.664084321475626e-05, |
| "loss": 0.6886, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.7611595233860928, |
| "grad_norm": 4.547012805938721, |
| "learning_rate": 1.658814229249012e-05, |
| "loss": 0.6653, |
| "step": 5350 |
| }, |
| { |
| "epoch": 0.7682731637915703, |
| "grad_norm": 3.5163490772247314, |
| "learning_rate": 1.6535441370223983e-05, |
| "loss": 0.6549, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.7753868041970479, |
| "grad_norm": 4.255113124847412, |
| "learning_rate": 1.648274044795784e-05, |
| "loss": 0.6473, |
| "step": 5450 |
| }, |
| { |
| "epoch": 0.7825004446025253, |
| "grad_norm": 8.696562767028809, |
| "learning_rate": 1.64300395256917e-05, |
| "loss": 0.67, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.7896140850080029, |
| "grad_norm": 3.098172903060913, |
| "learning_rate": 1.637733860342556e-05, |
| "loss": 0.6626, |
| "step": 5550 |
| }, |
| { |
| "epoch": 0.7967277254134804, |
| "grad_norm": 5.950908660888672, |
| "learning_rate": 1.632463768115942e-05, |
| "loss": 0.7476, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.8038413658189578, |
| "grad_norm": 5.9870686531066895, |
| "learning_rate": 1.627193675889328e-05, |
| "loss": 0.7006, |
| "step": 5650 |
| }, |
| { |
| "epoch": 0.8109550062244354, |
| "grad_norm": 8.210447311401367, |
| "learning_rate": 1.6219235836627142e-05, |
| "loss": 0.6104, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.8180686466299129, |
| "grad_norm": 2.8877291679382324, |
| "learning_rate": 1.6166534914361002e-05, |
| "loss": 0.6183, |
| "step": 5750 |
| }, |
| { |
| "epoch": 0.8251822870353903, |
| "grad_norm": 3.5387609004974365, |
| "learning_rate": 1.6113833992094863e-05, |
| "loss": 0.6816, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.8322959274408679, |
| "grad_norm": 3.3335936069488525, |
| "learning_rate": 1.6061133069828723e-05, |
| "loss": 0.6948, |
| "step": 5850 |
| }, |
| { |
| "epoch": 0.8394095678463454, |
| "grad_norm": 3.515953302383423, |
| "learning_rate": 1.6008432147562584e-05, |
| "loss": 0.6664, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.8465232082518229, |
| "grad_norm": 2.2991325855255127, |
| "learning_rate": 1.5955731225296444e-05, |
| "loss": 0.6363, |
| "step": 5950 |
| }, |
| { |
| "epoch": 0.8536368486573004, |
| "grad_norm": 5.262946128845215, |
| "learning_rate": 1.5903030303030305e-05, |
| "loss": 0.7375, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.8607504890627778, |
| "grad_norm": 4.494121074676514, |
| "learning_rate": 1.5850329380764165e-05, |
| "loss": 0.6818, |
| "step": 6050 |
| }, |
| { |
| "epoch": 0.8678641294682554, |
| "grad_norm": 4.698576927185059, |
| "learning_rate": 1.5797628458498026e-05, |
| "loss": 0.7246, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.8749777698737329, |
| "grad_norm": 4.571809768676758, |
| "learning_rate": 1.5744927536231883e-05, |
| "loss": 0.6112, |
| "step": 6150 |
| }, |
| { |
| "epoch": 0.8820914102792103, |
| "grad_norm": 7.1781110763549805, |
| "learning_rate": 1.5692226613965746e-05, |
| "loss": 0.6593, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.8892050506846879, |
| "grad_norm": 5.867689609527588, |
| "learning_rate": 1.5639525691699607e-05, |
| "loss": 0.6491, |
| "step": 6250 |
| }, |
| { |
| "epoch": 0.8963186910901654, |
| "grad_norm": 5.427711009979248, |
| "learning_rate": 1.5586824769433467e-05, |
| "loss": 0.6419, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.903432331495643, |
| "grad_norm": 5.6072869300842285, |
| "learning_rate": 1.5534123847167328e-05, |
| "loss": 0.6443, |
| "step": 6350 |
| }, |
| { |
| "epoch": 0.9105459719011204, |
| "grad_norm": 4.355414867401123, |
| "learning_rate": 1.5481422924901185e-05, |
| "loss": 0.7021, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.9176596123065979, |
| "grad_norm": 4.039976596832275, |
| "learning_rate": 1.542872200263505e-05, |
| "loss": 0.7351, |
| "step": 6450 |
| }, |
| { |
| "epoch": 0.9247732527120754, |
| "grad_norm": 3.736238956451416, |
| "learning_rate": 1.537602108036891e-05, |
| "loss": 0.6629, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.9318868931175529, |
| "grad_norm": 3.91310453414917, |
| "learning_rate": 1.532332015810277e-05, |
| "loss": 0.6856, |
| "step": 6550 |
| }, |
| { |
| "epoch": 0.9390005335230304, |
| "grad_norm": 5.676443576812744, |
| "learning_rate": 1.5270619235836627e-05, |
| "loss": 0.6344, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.9461141739285079, |
| "grad_norm": 3.6127190589904785, |
| "learning_rate": 1.5217918313570487e-05, |
| "loss": 0.6845, |
| "step": 6650 |
| }, |
| { |
| "epoch": 0.9532278143339854, |
| "grad_norm": 4.957319259643555, |
| "learning_rate": 1.516521739130435e-05, |
| "loss": 0.7074, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.960341454739463, |
| "grad_norm": 3.351602554321289, |
| "learning_rate": 1.511251646903821e-05, |
| "loss": 0.6722, |
| "step": 6750 |
| }, |
| { |
| "epoch": 0.9674550951449404, |
| "grad_norm": 5.316050052642822, |
| "learning_rate": 1.505981554677207e-05, |
| "loss": 0.5993, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.9745687355504179, |
| "grad_norm": 3.5226705074310303, |
| "learning_rate": 1.5007114624505929e-05, |
| "loss": 0.6663, |
| "step": 6850 |
| }, |
| { |
| "epoch": 0.9816823759558955, |
| "grad_norm": 4.123106479644775, |
| "learning_rate": 1.495441370223979e-05, |
| "loss": 0.6865, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.9887960163613729, |
| "grad_norm": 3.942507743835449, |
| "learning_rate": 1.4901712779973652e-05, |
| "loss": 0.6535, |
| "step": 6950 |
| }, |
| { |
| "epoch": 0.9959096567668504, |
| "grad_norm": 4.765227317810059, |
| "learning_rate": 1.4849011857707512e-05, |
| "loss": 0.5895, |
| "step": 7000 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_loss": 0.6490142345428467, |
| "eval_runtime": 321.1223, |
| "eval_samples_per_second": 16.414, |
| "eval_steps_per_second": 2.052, |
| "step": 7029 |
| }, |
| { |
| "epoch": 1.0029877289703006, |
| "grad_norm": 3.913954496383667, |
| "learning_rate": 1.479631093544137e-05, |
| "loss": 0.5676, |
| "step": 7050 |
| }, |
| { |
| "epoch": 1.010101369375778, |
| "grad_norm": 2.70705246925354, |
| "learning_rate": 1.4743610013175231e-05, |
| "loss": 0.5268, |
| "step": 7100 |
| }, |
| { |
| "epoch": 1.0172150097812556, |
| "grad_norm": 5.550088882446289, |
| "learning_rate": 1.4690909090909092e-05, |
| "loss": 0.507, |
| "step": 7150 |
| }, |
| { |
| "epoch": 1.024328650186733, |
| "grad_norm": 6.876676082611084, |
| "learning_rate": 1.4638208168642954e-05, |
| "loss": 0.5285, |
| "step": 7200 |
| }, |
| { |
| "epoch": 1.0314422905922105, |
| "grad_norm": 2.911832809448242, |
| "learning_rate": 1.4585507246376813e-05, |
| "loss": 0.5357, |
| "step": 7250 |
| }, |
| { |
| "epoch": 1.0385559309976882, |
| "grad_norm": 5.395831108093262, |
| "learning_rate": 1.4532806324110673e-05, |
| "loss": 0.5442, |
| "step": 7300 |
| }, |
| { |
| "epoch": 1.0456695714031656, |
| "grad_norm": 8.814352035522461, |
| "learning_rate": 1.4480105401844533e-05, |
| "loss": 0.5394, |
| "step": 7350 |
| }, |
| { |
| "epoch": 1.052783211808643, |
| "grad_norm": 2.947053909301758, |
| "learning_rate": 1.4427404479578392e-05, |
| "loss": 0.5549, |
| "step": 7400 |
| }, |
| { |
| "epoch": 1.0598968522141206, |
| "grad_norm": 5.564446926116943, |
| "learning_rate": 1.4374703557312254e-05, |
| "loss": 0.5328, |
| "step": 7450 |
| }, |
| { |
| "epoch": 1.067010492619598, |
| "grad_norm": 4.452253341674805, |
| "learning_rate": 1.4322002635046115e-05, |
| "loss": 0.5197, |
| "step": 7500 |
| }, |
| { |
| "epoch": 1.0741241330250757, |
| "grad_norm": 3.506099224090576, |
| "learning_rate": 1.4269301712779975e-05, |
| "loss": 0.5509, |
| "step": 7550 |
| }, |
| { |
| "epoch": 1.0812377734305532, |
| "grad_norm": 4.477612018585205, |
| "learning_rate": 1.4216600790513834e-05, |
| "loss": 0.5005, |
| "step": 7600 |
| }, |
| { |
| "epoch": 1.0883514138360306, |
| "grad_norm": 4.912115097045898, |
| "learning_rate": 1.4163899868247694e-05, |
| "loss": 0.5351, |
| "step": 7650 |
| }, |
| { |
| "epoch": 1.095465054241508, |
| "grad_norm": 4.499685764312744, |
| "learning_rate": 1.4111198945981557e-05, |
| "loss": 0.5218, |
| "step": 7700 |
| }, |
| { |
| "epoch": 1.1025786946469855, |
| "grad_norm": 4.6450934410095215, |
| "learning_rate": 1.4058498023715417e-05, |
| "loss": 0.5195, |
| "step": 7750 |
| }, |
| { |
| "epoch": 1.109692335052463, |
| "grad_norm": 4.883353233337402, |
| "learning_rate": 1.4005797101449276e-05, |
| "loss": 0.5464, |
| "step": 7800 |
| }, |
| { |
| "epoch": 1.1168059754579407, |
| "grad_norm": 3.312135934829712, |
| "learning_rate": 1.3953096179183136e-05, |
| "loss": 0.4766, |
| "step": 7850 |
| }, |
| { |
| "epoch": 1.1239196158634182, |
| "grad_norm": 2.180053234100342, |
| "learning_rate": 1.3900395256916997e-05, |
| "loss": 0.5742, |
| "step": 7900 |
| }, |
| { |
| "epoch": 1.1310332562688956, |
| "grad_norm": 4.708530426025391, |
| "learning_rate": 1.3847694334650859e-05, |
| "loss": 0.5499, |
| "step": 7950 |
| }, |
| { |
| "epoch": 1.138146896674373, |
| "grad_norm": 3.369694709777832, |
| "learning_rate": 1.379499341238472e-05, |
| "loss": 0.4839, |
| "step": 8000 |
| }, |
| { |
| "epoch": 1.1452605370798505, |
| "grad_norm": 3.1507697105407715, |
| "learning_rate": 1.3742292490118578e-05, |
| "loss": 0.5118, |
| "step": 8050 |
| }, |
| { |
| "epoch": 1.1523741774853282, |
| "grad_norm": 3.773084878921509, |
| "learning_rate": 1.3689591567852438e-05, |
| "loss": 0.5406, |
| "step": 8100 |
| }, |
| { |
| "epoch": 1.1594878178908057, |
| "grad_norm": 3.446446180343628, |
| "learning_rate": 1.3636890645586299e-05, |
| "loss": 0.5175, |
| "step": 8150 |
| }, |
| { |
| "epoch": 1.1666014582962831, |
| "grad_norm": 5.0827155113220215, |
| "learning_rate": 1.3584189723320161e-05, |
| "loss": 0.4963, |
| "step": 8200 |
| }, |
| { |
| "epoch": 1.1737150987017606, |
| "grad_norm": 5.854413032531738, |
| "learning_rate": 1.353148880105402e-05, |
| "loss": 0.6038, |
| "step": 8250 |
| }, |
| { |
| "epoch": 1.180828739107238, |
| "grad_norm": 4.135270595550537, |
| "learning_rate": 1.347878787878788e-05, |
| "loss": 0.5876, |
| "step": 8300 |
| }, |
| { |
| "epoch": 1.1879423795127155, |
| "grad_norm": 4.9584197998046875, |
| "learning_rate": 1.342608695652174e-05, |
| "loss": 0.5272, |
| "step": 8350 |
| }, |
| { |
| "epoch": 1.1950560199181932, |
| "grad_norm": 2.152341842651367, |
| "learning_rate": 1.33733860342556e-05, |
| "loss": 0.5146, |
| "step": 8400 |
| }, |
| { |
| "epoch": 1.2021696603236707, |
| "grad_norm": 3.346052885055542, |
| "learning_rate": 1.3320685111989462e-05, |
| "loss": 0.4661, |
| "step": 8450 |
| }, |
| { |
| "epoch": 1.2092833007291481, |
| "grad_norm": 4.768698692321777, |
| "learning_rate": 1.3267984189723322e-05, |
| "loss": 0.5354, |
| "step": 8500 |
| }, |
| { |
| "epoch": 1.2163969411346256, |
| "grad_norm": 5.182507514953613, |
| "learning_rate": 1.3215283267457183e-05, |
| "loss": 0.4845, |
| "step": 8550 |
| }, |
| { |
| "epoch": 1.223510581540103, |
| "grad_norm": 4.584784030914307, |
| "learning_rate": 1.3162582345191041e-05, |
| "loss": 0.5261, |
| "step": 8600 |
| }, |
| { |
| "epoch": 1.2306242219455807, |
| "grad_norm": 3.520695209503174, |
| "learning_rate": 1.3110935441370225e-05, |
| "loss": 0.5735, |
| "step": 8650 |
| }, |
| { |
| "epoch": 1.2377378623510582, |
| "grad_norm": 5.633211612701416, |
| "learning_rate": 1.3058234519104085e-05, |
| "loss": 0.516, |
| "step": 8700 |
| }, |
| { |
| "epoch": 1.2448515027565357, |
| "grad_norm": 5.559876441955566, |
| "learning_rate": 1.3005533596837944e-05, |
| "loss": 0.5597, |
| "step": 8750 |
| }, |
| { |
| "epoch": 1.2519651431620131, |
| "grad_norm": 4.474053382873535, |
| "learning_rate": 1.2952832674571806e-05, |
| "loss": 0.5171, |
| "step": 8800 |
| }, |
| { |
| "epoch": 1.2590787835674906, |
| "grad_norm": 3.866060733795166, |
| "learning_rate": 1.2900131752305667e-05, |
| "loss": 0.5057, |
| "step": 8850 |
| }, |
| { |
| "epoch": 1.266192423972968, |
| "grad_norm": 5.632756233215332, |
| "learning_rate": 1.2847430830039527e-05, |
| "loss": 0.5514, |
| "step": 8900 |
| }, |
| { |
| "epoch": 1.2733060643784457, |
| "grad_norm": 4.323223114013672, |
| "learning_rate": 1.2794729907773386e-05, |
| "loss": 0.5052, |
| "step": 8950 |
| }, |
| { |
| "epoch": 1.2804197047839232, |
| "grad_norm": 3.6960432529449463, |
| "learning_rate": 1.2742028985507246e-05, |
| "loss": 0.5295, |
| "step": 9000 |
| }, |
| { |
| "epoch": 1.2875333451894007, |
| "grad_norm": 3.777792453765869, |
| "learning_rate": 1.2689328063241108e-05, |
| "loss": 0.49, |
| "step": 9050 |
| }, |
| { |
| "epoch": 1.2946469855948781, |
| "grad_norm": 2.103578805923462, |
| "learning_rate": 1.2636627140974969e-05, |
| "loss": 0.5437, |
| "step": 9100 |
| }, |
| { |
| "epoch": 1.3017606260003558, |
| "grad_norm": 5.730409622192383, |
| "learning_rate": 1.258392621870883e-05, |
| "loss": 0.532, |
| "step": 9150 |
| }, |
| { |
| "epoch": 1.3088742664058333, |
| "grad_norm": 6.736247539520264, |
| "learning_rate": 1.2532279314888011e-05, |
| "loss": 0.554, |
| "step": 9200 |
| }, |
| { |
| "epoch": 1.3159879068113107, |
| "grad_norm": 2.0751116275787354, |
| "learning_rate": 1.2479578392621872e-05, |
| "loss": 0.5128, |
| "step": 9250 |
| }, |
| { |
| "epoch": 1.3231015472167882, |
| "grad_norm": 2.6663055419921875, |
| "learning_rate": 1.2426877470355732e-05, |
| "loss": 0.5251, |
| "step": 9300 |
| }, |
| { |
| "epoch": 1.3302151876222656, |
| "grad_norm": 4.778317928314209, |
| "learning_rate": 1.2374176548089594e-05, |
| "loss": 0.5219, |
| "step": 9350 |
| }, |
| { |
| "epoch": 1.337328828027743, |
| "grad_norm": 5.950105667114258, |
| "learning_rate": 1.2321475625823453e-05, |
| "loss": 0.5192, |
| "step": 9400 |
| }, |
| { |
| "epoch": 1.3444424684332206, |
| "grad_norm": 3.5832479000091553, |
| "learning_rate": 1.2268774703557313e-05, |
| "loss": 0.5185, |
| "step": 9450 |
| }, |
| { |
| "epoch": 1.3515561088386983, |
| "grad_norm": 2.7390801906585693, |
| "learning_rate": 1.2216073781291174e-05, |
| "loss": 0.5375, |
| "step": 9500 |
| }, |
| { |
| "epoch": 1.3586697492441757, |
| "grad_norm": 2.688469648361206, |
| "learning_rate": 1.2163372859025033e-05, |
| "loss": 0.5352, |
| "step": 9550 |
| }, |
| { |
| "epoch": 1.3657833896496532, |
| "grad_norm": 4.956066608428955, |
| "learning_rate": 1.2110671936758893e-05, |
| "loss": 0.5341, |
| "step": 9600 |
| }, |
| { |
| "epoch": 1.3728970300551306, |
| "grad_norm": 5.483800411224365, |
| "learning_rate": 1.2057971014492755e-05, |
| "loss": 0.5644, |
| "step": 9650 |
| }, |
| { |
| "epoch": 1.3800106704606083, |
| "grad_norm": 4.272674083709717, |
| "learning_rate": 1.2005270092226616e-05, |
| "loss": 0.5044, |
| "step": 9700 |
| }, |
| { |
| "epoch": 1.3871243108660858, |
| "grad_norm": 2.5622191429138184, |
| "learning_rate": 1.1952569169960474e-05, |
| "loss": 0.5171, |
| "step": 9750 |
| }, |
| { |
| "epoch": 1.3942379512715632, |
| "grad_norm": 7.005847454071045, |
| "learning_rate": 1.1899868247694335e-05, |
| "loss": 0.5156, |
| "step": 9800 |
| }, |
| { |
| "epoch": 1.4013515916770407, |
| "grad_norm": 4.284529685974121, |
| "learning_rate": 1.1847167325428195e-05, |
| "loss": 0.5833, |
| "step": 9850 |
| }, |
| { |
| "epoch": 1.4084652320825182, |
| "grad_norm": 3.930690288543701, |
| "learning_rate": 1.1794466403162057e-05, |
| "loss": 0.5219, |
| "step": 9900 |
| }, |
| { |
| "epoch": 1.4155788724879956, |
| "grad_norm": 4.899880886077881, |
| "learning_rate": 1.1741765480895916e-05, |
| "loss": 0.5605, |
| "step": 9950 |
| }, |
| { |
| "epoch": 1.4226925128934733, |
| "grad_norm": 5.612455368041992, |
| "learning_rate": 1.1689064558629777e-05, |
| "loss": 0.5408, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.4298061532989508, |
| "grad_norm": 2.502410411834717, |
| "learning_rate": 1.1636363636363637e-05, |
| "loss": 0.5052, |
| "step": 10050 |
| }, |
| { |
| "epoch": 1.4369197937044282, |
| "grad_norm": 4.965051174163818, |
| "learning_rate": 1.1583662714097496e-05, |
| "loss": 0.5036, |
| "step": 10100 |
| }, |
| { |
| "epoch": 1.4440334341099057, |
| "grad_norm": 4.5783371925354, |
| "learning_rate": 1.1530961791831358e-05, |
| "loss": 0.5474, |
| "step": 10150 |
| }, |
| { |
| "epoch": 1.4511470745153834, |
| "grad_norm": 5.443408012390137, |
| "learning_rate": 1.1478260869565218e-05, |
| "loss": 0.5018, |
| "step": 10200 |
| }, |
| { |
| "epoch": 1.4582607149208608, |
| "grad_norm": 5.504450798034668, |
| "learning_rate": 1.1425559947299079e-05, |
| "loss": 0.4877, |
| "step": 10250 |
| }, |
| { |
| "epoch": 1.4653743553263383, |
| "grad_norm": 3.819101572036743, |
| "learning_rate": 1.137285902503294e-05, |
| "loss": 0.526, |
| "step": 10300 |
| }, |
| { |
| "epoch": 1.4724879957318158, |
| "grad_norm": 5.407709121704102, |
| "learning_rate": 1.1320158102766798e-05, |
| "loss": 0.4949, |
| "step": 10350 |
| }, |
| { |
| "epoch": 1.4796016361372932, |
| "grad_norm": 4.124641418457031, |
| "learning_rate": 1.126745718050066e-05, |
| "loss": 0.5923, |
| "step": 10400 |
| }, |
| { |
| "epoch": 1.4867152765427707, |
| "grad_norm": 4.266039848327637, |
| "learning_rate": 1.121475625823452e-05, |
| "loss": 0.5545, |
| "step": 10450 |
| }, |
| { |
| "epoch": 1.4938289169482482, |
| "grad_norm": 4.402031421661377, |
| "learning_rate": 1.1162055335968381e-05, |
| "loss": 0.505, |
| "step": 10500 |
| }, |
| { |
| "epoch": 1.5009425573537256, |
| "grad_norm": 6.378383159637451, |
| "learning_rate": 1.110935441370224e-05, |
| "loss": 0.5015, |
| "step": 10550 |
| }, |
| { |
| "epoch": 1.5080561977592033, |
| "grad_norm": 4.023849964141846, |
| "learning_rate": 1.10566534914361e-05, |
| "loss": 0.5371, |
| "step": 10600 |
| }, |
| { |
| "epoch": 1.5151698381646808, |
| "grad_norm": 4.202026844024658, |
| "learning_rate": 1.1003952569169962e-05, |
| "loss": 0.5527, |
| "step": 10650 |
| }, |
| { |
| "epoch": 1.5222834785701584, |
| "grad_norm": 4.9058685302734375, |
| "learning_rate": 1.0951251646903823e-05, |
| "loss": 0.5193, |
| "step": 10700 |
| }, |
| { |
| "epoch": 1.529397118975636, |
| "grad_norm": 5.402507781982422, |
| "learning_rate": 1.0898550724637682e-05, |
| "loss": 0.542, |
| "step": 10750 |
| }, |
| { |
| "epoch": 1.5365107593811134, |
| "grad_norm": 4.229278087615967, |
| "learning_rate": 1.0845849802371542e-05, |
| "loss": 0.5284, |
| "step": 10800 |
| }, |
| { |
| "epoch": 1.5436243997865908, |
| "grad_norm": 3.6707725524902344, |
| "learning_rate": 1.0793148880105403e-05, |
| "loss": 0.561, |
| "step": 10850 |
| }, |
| { |
| "epoch": 1.5507380401920683, |
| "grad_norm": 3.6205899715423584, |
| "learning_rate": 1.0740447957839265e-05, |
| "loss": 0.4981, |
| "step": 10900 |
| }, |
| { |
| "epoch": 1.5578516805975458, |
| "grad_norm": 4.793003559112549, |
| "learning_rate": 1.0687747035573123e-05, |
| "loss": 0.5062, |
| "step": 10950 |
| }, |
| { |
| "epoch": 1.5649653210030232, |
| "grad_norm": 3.2560999393463135, |
| "learning_rate": 1.0635046113306984e-05, |
| "loss": 0.5305, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.5720789614085007, |
| "grad_norm": 4.039142608642578, |
| "learning_rate": 1.0582345191040844e-05, |
| "loss": 0.5412, |
| "step": 11050 |
| }, |
| { |
| "epoch": 1.5791926018139784, |
| "grad_norm": 4.209068298339844, |
| "learning_rate": 1.0529644268774703e-05, |
| "loss": 0.5014, |
| "step": 11100 |
| }, |
| { |
| "epoch": 1.5863062422194558, |
| "grad_norm": 3.390684127807617, |
| "learning_rate": 1.0476943346508565e-05, |
| "loss": 0.5605, |
| "step": 11150 |
| }, |
| { |
| "epoch": 1.5934198826249333, |
| "grad_norm": 3.7350914478302, |
| "learning_rate": 1.0424242424242426e-05, |
| "loss": 0.5021, |
| "step": 11200 |
| }, |
| { |
| "epoch": 1.600533523030411, |
| "grad_norm": 3.5504932403564453, |
| "learning_rate": 1.0371541501976286e-05, |
| "loss": 0.4967, |
| "step": 11250 |
| }, |
| { |
| "epoch": 1.6076471634358884, |
| "grad_norm": 3.8655166625976562, |
| "learning_rate": 1.0318840579710145e-05, |
| "loss": 0.4857, |
| "step": 11300 |
| }, |
| { |
| "epoch": 1.6147608038413659, |
| "grad_norm": 6.647157669067383, |
| "learning_rate": 1.0266139657444005e-05, |
| "loss": 0.5033, |
| "step": 11350 |
| }, |
| { |
| "epoch": 1.6218744442468433, |
| "grad_norm": 5.384164810180664, |
| "learning_rate": 1.0213438735177868e-05, |
| "loss": 0.5486, |
| "step": 11400 |
| }, |
| { |
| "epoch": 1.6289880846523208, |
| "grad_norm": 4.699065685272217, |
| "learning_rate": 1.016179183135705e-05, |
| "loss": 0.5565, |
| "step": 11450 |
| }, |
| { |
| "epoch": 1.6361017250577983, |
| "grad_norm": 4.786224365234375, |
| "learning_rate": 1.010909090909091e-05, |
| "loss": 0.5155, |
| "step": 11500 |
| }, |
| { |
| "epoch": 1.6432153654632757, |
| "grad_norm": 4.30881929397583, |
| "learning_rate": 1.005638998682477e-05, |
| "loss": 0.5303, |
| "step": 11550 |
| }, |
| { |
| "epoch": 1.6503290058687532, |
| "grad_norm": 4.938950538635254, |
| "learning_rate": 1.000368906455863e-05, |
| "loss": 0.4891, |
| "step": 11600 |
| }, |
| { |
| "epoch": 1.6574426462742309, |
| "grad_norm": 4.875250339508057, |
| "learning_rate": 9.950988142292491e-06, |
| "loss": 0.531, |
| "step": 11650 |
| }, |
| { |
| "epoch": 1.6645562866797083, |
| "grad_norm": 5.96549654006958, |
| "learning_rate": 9.898287220026352e-06, |
| "loss": 0.5092, |
| "step": 11700 |
| }, |
| { |
| "epoch": 1.6716699270851858, |
| "grad_norm": 5.401658535003662, |
| "learning_rate": 9.845586297760212e-06, |
| "loss": 0.5018, |
| "step": 11750 |
| }, |
| { |
| "epoch": 1.6787835674906635, |
| "grad_norm": 4.49137020111084, |
| "learning_rate": 9.792885375494072e-06, |
| "loss": 0.4877, |
| "step": 11800 |
| }, |
| { |
| "epoch": 1.685897207896141, |
| "grad_norm": 3.0070083141326904, |
| "learning_rate": 9.740184453227933e-06, |
| "loss": 0.4884, |
| "step": 11850 |
| }, |
| { |
| "epoch": 1.6930108483016184, |
| "grad_norm": 3.979234218597412, |
| "learning_rate": 9.687483530961793e-06, |
| "loss": 0.5013, |
| "step": 11900 |
| }, |
| { |
| "epoch": 1.7001244887070959, |
| "grad_norm": 3.876878499984741, |
| "learning_rate": 9.634782608695654e-06, |
| "loss": 0.5322, |
| "step": 11950 |
| }, |
| { |
| "epoch": 1.7072381291125733, |
| "grad_norm": 5.199528694152832, |
| "learning_rate": 9.582081686429513e-06, |
| "loss": 0.5052, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.7143517695180508, |
| "grad_norm": 3.97282075881958, |
| "learning_rate": 9.529380764163375e-06, |
| "loss": 0.4887, |
| "step": 12050 |
| }, |
| { |
| "epoch": 1.7214654099235283, |
| "grad_norm": 4.776126384735107, |
| "learning_rate": 9.476679841897233e-06, |
| "loss": 0.5384, |
| "step": 12100 |
| }, |
| { |
| "epoch": 1.7285790503290057, |
| "grad_norm": 6.539297580718994, |
| "learning_rate": 9.423978919631094e-06, |
| "loss": 0.5283, |
| "step": 12150 |
| }, |
| { |
| "epoch": 1.7356926907344834, |
| "grad_norm": 5.380597114562988, |
| "learning_rate": 9.371277997364954e-06, |
| "loss": 0.5481, |
| "step": 12200 |
| }, |
| { |
| "epoch": 1.7428063311399609, |
| "grad_norm": 4.773643493652344, |
| "learning_rate": 9.318577075098815e-06, |
| "loss": 0.4815, |
| "step": 12250 |
| }, |
| { |
| "epoch": 1.7499199715454383, |
| "grad_norm": 1.9399245977401733, |
| "learning_rate": 9.265876152832675e-06, |
| "loss": 0.515, |
| "step": 12300 |
| }, |
| { |
| "epoch": 1.757033611950916, |
| "grad_norm": 3.586733102798462, |
| "learning_rate": 9.213175230566536e-06, |
| "loss": 0.4856, |
| "step": 12350 |
| }, |
| { |
| "epoch": 1.7641472523563935, |
| "grad_norm": 3.9114813804626465, |
| "learning_rate": 9.160474308300396e-06, |
| "loss": 0.5114, |
| "step": 12400 |
| }, |
| { |
| "epoch": 1.771260892761871, |
| "grad_norm": 3.8798410892486572, |
| "learning_rate": 9.107773386034257e-06, |
| "loss": 0.4939, |
| "step": 12450 |
| }, |
| { |
| "epoch": 1.7783745331673484, |
| "grad_norm": 2.8337159156799316, |
| "learning_rate": 9.055072463768117e-06, |
| "loss": 0.4807, |
| "step": 12500 |
| }, |
| { |
| "epoch": 1.7854881735728259, |
| "grad_norm": 4.102101802825928, |
| "learning_rate": 9.002371541501978e-06, |
| "loss": 0.5632, |
| "step": 12550 |
| }, |
| { |
| "epoch": 1.7926018139783033, |
| "grad_norm": 2.918848991394043, |
| "learning_rate": 8.949670619235838e-06, |
| "loss": 0.5012, |
| "step": 12600 |
| }, |
| { |
| "epoch": 1.7997154543837808, |
| "grad_norm": 3.708796739578247, |
| "learning_rate": 8.896969696969697e-06, |
| "loss": 0.5159, |
| "step": 12650 |
| }, |
| { |
| "epoch": 1.8068290947892582, |
| "grad_norm": 4.967949867248535, |
| "learning_rate": 8.844268774703559e-06, |
| "loss": 0.5037, |
| "step": 12700 |
| }, |
| { |
| "epoch": 1.813942735194736, |
| "grad_norm": 2.7646732330322266, |
| "learning_rate": 8.791567852437418e-06, |
| "loss": 0.4842, |
| "step": 12750 |
| }, |
| { |
| "epoch": 1.8210563756002134, |
| "grad_norm": 3.667774200439453, |
| "learning_rate": 8.73886693017128e-06, |
| "loss": 0.4744, |
| "step": 12800 |
| }, |
| { |
| "epoch": 1.828170016005691, |
| "grad_norm": 5.304251670837402, |
| "learning_rate": 8.686166007905139e-06, |
| "loss": 0.521, |
| "step": 12850 |
| }, |
| { |
| "epoch": 1.8352836564111685, |
| "grad_norm": 4.663897514343262, |
| "learning_rate": 8.633465085638999e-06, |
| "loss": 0.5143, |
| "step": 12900 |
| }, |
| { |
| "epoch": 1.842397296816646, |
| "grad_norm": 6.192160129547119, |
| "learning_rate": 8.58076416337286e-06, |
| "loss": 0.5489, |
| "step": 12950 |
| }, |
| { |
| "epoch": 1.8495109372221235, |
| "grad_norm": 3.5136263370513916, |
| "learning_rate": 8.52806324110672e-06, |
| "loss": 0.5124, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.856624577627601, |
| "grad_norm": 3.0261971950531006, |
| "learning_rate": 8.47536231884058e-06, |
| "loss": 0.5295, |
| "step": 13050 |
| }, |
| { |
| "epoch": 1.8637382180330784, |
| "grad_norm": 3.9308557510375977, |
| "learning_rate": 8.42266139657444e-06, |
| "loss": 0.5209, |
| "step": 13100 |
| }, |
| { |
| "epoch": 1.8708518584385558, |
| "grad_norm": 4.964597225189209, |
| "learning_rate": 8.369960474308301e-06, |
| "loss": 0.5187, |
| "step": 13150 |
| }, |
| { |
| "epoch": 1.8779654988440333, |
| "grad_norm": 5.161786079406738, |
| "learning_rate": 8.317259552042162e-06, |
| "loss": 0.5207, |
| "step": 13200 |
| }, |
| { |
| "epoch": 1.885079139249511, |
| "grad_norm": 2.5828144550323486, |
| "learning_rate": 8.264558629776022e-06, |
| "loss": 0.4598, |
| "step": 13250 |
| }, |
| { |
| "epoch": 1.8921927796549884, |
| "grad_norm": 3.169773817062378, |
| "learning_rate": 8.211857707509883e-06, |
| "loss": 0.531, |
| "step": 13300 |
| }, |
| { |
| "epoch": 1.899306420060466, |
| "grad_norm": 3.6591272354125977, |
| "learning_rate": 8.159156785243743e-06, |
| "loss": 0.5226, |
| "step": 13350 |
| }, |
| { |
| "epoch": 1.9064200604659436, |
| "grad_norm": 2.735844850540161, |
| "learning_rate": 8.106455862977602e-06, |
| "loss": 0.5808, |
| "step": 13400 |
| }, |
| { |
| "epoch": 1.913533700871421, |
| "grad_norm": 5.649113655090332, |
| "learning_rate": 8.053754940711464e-06, |
| "loss": 0.5154, |
| "step": 13450 |
| }, |
| { |
| "epoch": 1.9206473412768985, |
| "grad_norm": 2.830268621444702, |
| "learning_rate": 8.001054018445323e-06, |
| "loss": 0.5017, |
| "step": 13500 |
| }, |
| { |
| "epoch": 1.927760981682376, |
| "grad_norm": 5.496303558349609, |
| "learning_rate": 7.948353096179183e-06, |
| "loss": 0.4824, |
| "step": 13550 |
| }, |
| { |
| "epoch": 1.9348746220878534, |
| "grad_norm": 3.1559128761291504, |
| "learning_rate": 7.895652173913044e-06, |
| "loss": 0.5465, |
| "step": 13600 |
| }, |
| { |
| "epoch": 1.941988262493331, |
| "grad_norm": 3.265587091445923, |
| "learning_rate": 7.842951251646904e-06, |
| "loss": 0.4673, |
| "step": 13650 |
| }, |
| { |
| "epoch": 1.9491019028988084, |
| "grad_norm": 6.156945705413818, |
| "learning_rate": 7.790250329380764e-06, |
| "loss": 0.5558, |
| "step": 13700 |
| }, |
| { |
| "epoch": 1.9562155433042858, |
| "grad_norm": 7.1588029861450195, |
| "learning_rate": 7.737549407114625e-06, |
| "loss": 0.5013, |
| "step": 13750 |
| }, |
| { |
| "epoch": 1.9633291837097635, |
| "grad_norm": 3.9235057830810547, |
| "learning_rate": 7.684848484848485e-06, |
| "loss": 0.5214, |
| "step": 13800 |
| }, |
| { |
| "epoch": 1.970442824115241, |
| "grad_norm": 3.9342610836029053, |
| "learning_rate": 7.632147562582346e-06, |
| "loss": 0.5147, |
| "step": 13850 |
| }, |
| { |
| "epoch": 1.9775564645207184, |
| "grad_norm": 2.9213314056396484, |
| "learning_rate": 7.579446640316206e-06, |
| "loss": 0.4756, |
| "step": 13900 |
| }, |
| { |
| "epoch": 1.984670104926196, |
| "grad_norm": 2.211418390274048, |
| "learning_rate": 7.526745718050067e-06, |
| "loss": 0.4862, |
| "step": 13950 |
| }, |
| { |
| "epoch": 1.9917837453316736, |
| "grad_norm": 3.4956717491149902, |
| "learning_rate": 7.474044795783927e-06, |
| "loss": 0.5235, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.998897385737151, |
| "grad_norm": 4.755037784576416, |
| "learning_rate": 7.421343873517787e-06, |
| "loss": 0.5176, |
| "step": 14050 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_loss": 0.6273267865180969, |
| "eval_runtime": 321.2086, |
| "eval_samples_per_second": 16.41, |
| "eval_steps_per_second": 2.052, |
| "step": 14058 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 21084, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 7.608571407630336e+16, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|