|
{ |
|
"best_metric": 0.24052156507968903, |
|
"best_model_checkpoint": "./oral_disease_mobilevit/checkpoint-4600", |
|
"epoch": 5.0, |
|
"eval_steps": 100, |
|
"global_step": 5830, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04288164665523156, |
|
"grad_norm": 2.5394277572631836, |
|
"learning_rate": 4.957118353344769e-05, |
|
"loss": 1.7703, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.08576329331046312, |
|
"grad_norm": 2.5862715244293213, |
|
"learning_rate": 4.914236706689537e-05, |
|
"loss": 1.6871, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08576329331046312, |
|
"eval_accuracy": 0.577005577005577, |
|
"eval_loss": 1.6024609804153442, |
|
"eval_runtime": 14.1663, |
|
"eval_samples_per_second": 164.545, |
|
"eval_steps_per_second": 20.612, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.12864493996569468, |
|
"grad_norm": 3.443279504776001, |
|
"learning_rate": 4.8713550600343055e-05, |
|
"loss": 1.5656, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.17152658662092624, |
|
"grad_norm": 2.7117512226104736, |
|
"learning_rate": 4.828473413379074e-05, |
|
"loss": 1.4961, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.17152658662092624, |
|
"eval_accuracy": 0.5971685971685972, |
|
"eval_loss": 1.3290797472000122, |
|
"eval_runtime": 14.4239, |
|
"eval_samples_per_second": 161.607, |
|
"eval_steps_per_second": 20.244, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.2144082332761578, |
|
"grad_norm": 3.41520619392395, |
|
"learning_rate": 4.7855917667238424e-05, |
|
"loss": 1.3941, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.25728987993138935, |
|
"grad_norm": 2.960935354232788, |
|
"learning_rate": 4.742710120068611e-05, |
|
"loss": 1.2984, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.25728987993138935, |
|
"eval_accuracy": 0.6525096525096525, |
|
"eval_loss": 1.1308239698410034, |
|
"eval_runtime": 14.2271, |
|
"eval_samples_per_second": 163.842, |
|
"eval_steps_per_second": 20.524, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.30017152658662094, |
|
"grad_norm": 3.7776384353637695, |
|
"learning_rate": 4.699828473413379e-05, |
|
"loss": 1.2457, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.34305317324185247, |
|
"grad_norm": 6.4286088943481445, |
|
"learning_rate": 4.656946826758148e-05, |
|
"loss": 1.12, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.34305317324185247, |
|
"eval_accuracy": 0.7232947232947233, |
|
"eval_loss": 0.9413465261459351, |
|
"eval_runtime": 14.2093, |
|
"eval_samples_per_second": 164.048, |
|
"eval_steps_per_second": 20.55, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.38593481989708406, |
|
"grad_norm": 3.400545120239258, |
|
"learning_rate": 4.614065180102916e-05, |
|
"loss": 1.0816, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.4288164665523156, |
|
"grad_norm": 4.862150192260742, |
|
"learning_rate": 4.5711835334476845e-05, |
|
"loss": 1.0873, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.4288164665523156, |
|
"eval_accuracy": 0.7610467610467611, |
|
"eval_loss": 0.8114073872566223, |
|
"eval_runtime": 14.2607, |
|
"eval_samples_per_second": 163.456, |
|
"eval_steps_per_second": 20.476, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.4716981132075472, |
|
"grad_norm": 3.4492220878601074, |
|
"learning_rate": 4.528301886792453e-05, |
|
"loss": 0.9814, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.5145797598627787, |
|
"grad_norm": 3.7629287242889404, |
|
"learning_rate": 4.4854202401372214e-05, |
|
"loss": 1.0012, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5145797598627787, |
|
"eval_accuracy": 0.7919347919347919, |
|
"eval_loss": 0.6951708197593689, |
|
"eval_runtime": 13.8986, |
|
"eval_samples_per_second": 167.714, |
|
"eval_steps_per_second": 21.009, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5574614065180102, |
|
"grad_norm": 4.381726264953613, |
|
"learning_rate": 4.4425385934819905e-05, |
|
"loss": 0.9516, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.6003430531732419, |
|
"grad_norm": 3.579646348953247, |
|
"learning_rate": 4.399656946826758e-05, |
|
"loss": 0.8379, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6003430531732419, |
|
"eval_accuracy": 0.8125268125268126, |
|
"eval_loss": 0.6312650442123413, |
|
"eval_runtime": 13.8871, |
|
"eval_samples_per_second": 167.853, |
|
"eval_steps_per_second": 21.027, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6432246998284734, |
|
"grad_norm": 5.985883712768555, |
|
"learning_rate": 4.356775300171527e-05, |
|
"loss": 0.8269, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.6861063464837049, |
|
"grad_norm": 5.968649864196777, |
|
"learning_rate": 4.313893653516296e-05, |
|
"loss": 0.8627, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.6861063464837049, |
|
"eval_accuracy": 0.8181038181038182, |
|
"eval_loss": 0.5455917716026306, |
|
"eval_runtime": 13.9813, |
|
"eval_samples_per_second": 166.723, |
|
"eval_steps_per_second": 20.885, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7289879931389366, |
|
"grad_norm": 5.903740882873535, |
|
"learning_rate": 4.2710120068610635e-05, |
|
"loss": 0.8055, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.7718696397941681, |
|
"grad_norm": 6.328656196594238, |
|
"learning_rate": 4.228130360205832e-05, |
|
"loss": 0.8252, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7718696397941681, |
|
"eval_accuracy": 0.8314028314028314, |
|
"eval_loss": 0.5072463750839233, |
|
"eval_runtime": 13.8768, |
|
"eval_samples_per_second": 167.979, |
|
"eval_steps_per_second": 21.042, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.8147512864493996, |
|
"grad_norm": 4.45211124420166, |
|
"learning_rate": 4.185248713550601e-05, |
|
"loss": 0.7173, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.8576329331046312, |
|
"grad_norm": 5.749237537384033, |
|
"learning_rate": 4.142367066895369e-05, |
|
"loss": 0.7223, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8576329331046312, |
|
"eval_accuracy": 0.8378378378378378, |
|
"eval_loss": 0.46284219622612, |
|
"eval_runtime": 14.2987, |
|
"eval_samples_per_second": 163.022, |
|
"eval_steps_per_second": 20.421, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.9005145797598628, |
|
"grad_norm": 16.25695037841797, |
|
"learning_rate": 4.099485420240137e-05, |
|
"loss": 0.7153, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.9433962264150944, |
|
"grad_norm": 7.836055755615234, |
|
"learning_rate": 4.0566037735849064e-05, |
|
"loss": 0.6617, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9433962264150944, |
|
"eval_accuracy": 0.8468468468468469, |
|
"eval_loss": 0.43263953924179077, |
|
"eval_runtime": 13.9551, |
|
"eval_samples_per_second": 167.035, |
|
"eval_steps_per_second": 20.924, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9862778730703259, |
|
"grad_norm": 8.199559211730957, |
|
"learning_rate": 4.013722126929674e-05, |
|
"loss": 0.7156, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.0291595197255574, |
|
"grad_norm": 8.776689529418945, |
|
"learning_rate": 3.9708404802744425e-05, |
|
"loss": 0.7784, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.0291595197255574, |
|
"eval_accuracy": 0.8511368511368511, |
|
"eval_loss": 0.42261621356010437, |
|
"eval_runtime": 14.1288, |
|
"eval_samples_per_second": 164.982, |
|
"eval_steps_per_second": 20.667, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.072041166380789, |
|
"grad_norm": 5.693437099456787, |
|
"learning_rate": 3.9279588336192116e-05, |
|
"loss": 0.6163, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.1149228130360207, |
|
"grad_norm": 16.413869857788086, |
|
"learning_rate": 3.8850771869639794e-05, |
|
"loss": 0.6086, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.1149228130360207, |
|
"eval_accuracy": 0.8592878592878593, |
|
"eval_loss": 0.3927444517612457, |
|
"eval_runtime": 13.8914, |
|
"eval_samples_per_second": 167.801, |
|
"eval_steps_per_second": 21.02, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.1578044596912522, |
|
"grad_norm": 4.260317325592041, |
|
"learning_rate": 3.842195540308748e-05, |
|
"loss": 0.6237, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.2006861063464838, |
|
"grad_norm": 8.143071174621582, |
|
"learning_rate": 3.799313893653517e-05, |
|
"loss": 0.6196, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.2006861063464838, |
|
"eval_accuracy": 0.8618618618618619, |
|
"eval_loss": 0.3923815190792084, |
|
"eval_runtime": 13.8559, |
|
"eval_samples_per_second": 168.231, |
|
"eval_steps_per_second": 21.074, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.2435677530017153, |
|
"grad_norm": 8.181282043457031, |
|
"learning_rate": 3.756432246998285e-05, |
|
"loss": 0.5606, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.2864493996569468, |
|
"grad_norm": 14.405867576599121, |
|
"learning_rate": 3.713550600343053e-05, |
|
"loss": 0.6561, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.2864493996569468, |
|
"eval_accuracy": 0.8601458601458601, |
|
"eval_loss": 0.36444056034088135, |
|
"eval_runtime": 13.8772, |
|
"eval_samples_per_second": 167.974, |
|
"eval_steps_per_second": 21.042, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.3293310463121784, |
|
"grad_norm": 14.855607986450195, |
|
"learning_rate": 3.670668953687822e-05, |
|
"loss": 0.5548, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.3722126929674099, |
|
"grad_norm": 4.9101409912109375, |
|
"learning_rate": 3.62778730703259e-05, |
|
"loss": 0.5426, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.3722126929674099, |
|
"eval_accuracy": 0.8738738738738738, |
|
"eval_loss": 0.3387666344642639, |
|
"eval_runtime": 14.6787, |
|
"eval_samples_per_second": 158.801, |
|
"eval_steps_per_second": 19.893, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.4150943396226414, |
|
"grad_norm": 3.1763756275177, |
|
"learning_rate": 3.5849056603773584e-05, |
|
"loss": 0.5917, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.457975986277873, |
|
"grad_norm": 2.9457991123199463, |
|
"learning_rate": 3.5420240137221275e-05, |
|
"loss": 0.5858, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.457975986277873, |
|
"eval_accuracy": 0.8652938652938653, |
|
"eval_loss": 0.3525221347808838, |
|
"eval_runtime": 14.1719, |
|
"eval_samples_per_second": 164.48, |
|
"eval_steps_per_second": 20.604, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.5008576329331045, |
|
"grad_norm": 7.572704792022705, |
|
"learning_rate": 3.499142367066895e-05, |
|
"loss": 0.6098, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.5437392795883362, |
|
"grad_norm": 1.232725739479065, |
|
"learning_rate": 3.456260720411664e-05, |
|
"loss": 0.574, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.5437392795883362, |
|
"eval_accuracy": 0.8764478764478765, |
|
"eval_loss": 0.32060739398002625, |
|
"eval_runtime": 14.3858, |
|
"eval_samples_per_second": 162.035, |
|
"eval_steps_per_second": 20.298, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.5866209262435678, |
|
"grad_norm": 10.918987274169922, |
|
"learning_rate": 3.413379073756433e-05, |
|
"loss": 0.5824, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.6295025728987993, |
|
"grad_norm": 10.205852508544922, |
|
"learning_rate": 3.3704974271012005e-05, |
|
"loss": 0.5015, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.6295025728987993, |
|
"eval_accuracy": 0.882024882024882, |
|
"eval_loss": 0.32507938146591187, |
|
"eval_runtime": 14.2196, |
|
"eval_samples_per_second": 163.929, |
|
"eval_steps_per_second": 20.535, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.6723842195540308, |
|
"grad_norm": 7.081895351409912, |
|
"learning_rate": 3.3276157804459696e-05, |
|
"loss": 0.4681, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.7152658662092626, |
|
"grad_norm": 9.101844787597656, |
|
"learning_rate": 3.284734133790738e-05, |
|
"loss": 0.493, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.7152658662092626, |
|
"eval_accuracy": 0.8854568854568855, |
|
"eval_loss": 0.308988094329834, |
|
"eval_runtime": 14.3569, |
|
"eval_samples_per_second": 162.36, |
|
"eval_steps_per_second": 20.339, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.758147512864494, |
|
"grad_norm": 6.105254650115967, |
|
"learning_rate": 3.241852487135506e-05, |
|
"loss": 0.533, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.8010291595197256, |
|
"grad_norm": 14.518338203430176, |
|
"learning_rate": 3.198970840480275e-05, |
|
"loss": 0.5381, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.8010291595197256, |
|
"eval_accuracy": 0.8773058773058773, |
|
"eval_loss": 0.3106726109981537, |
|
"eval_runtime": 14.2106, |
|
"eval_samples_per_second": 164.033, |
|
"eval_steps_per_second": 20.548, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.8439108061749572, |
|
"grad_norm": 3.1229450702667236, |
|
"learning_rate": 3.1560891938250434e-05, |
|
"loss": 0.5641, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.8867924528301887, |
|
"grad_norm": 14.663033485412598, |
|
"learning_rate": 3.113207547169811e-05, |
|
"loss": 0.4979, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.8867924528301887, |
|
"eval_accuracy": 0.8901758901758902, |
|
"eval_loss": 0.2851434051990509, |
|
"eval_runtime": 14.2521, |
|
"eval_samples_per_second": 163.554, |
|
"eval_steps_per_second": 20.488, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.9296740994854202, |
|
"grad_norm": 3.774958848953247, |
|
"learning_rate": 3.07032590051458e-05, |
|
"loss": 0.5117, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.9725557461406518, |
|
"grad_norm": 25.906217575073242, |
|
"learning_rate": 3.0274442538593483e-05, |
|
"loss": 0.5145, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.9725557461406518, |
|
"eval_accuracy": 0.8927498927498928, |
|
"eval_loss": 0.2909970283508301, |
|
"eval_runtime": 14.2775, |
|
"eval_samples_per_second": 163.264, |
|
"eval_steps_per_second": 20.452, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.0154373927958833, |
|
"grad_norm": 8.114184379577637, |
|
"learning_rate": 2.9845626072041167e-05, |
|
"loss": 0.4499, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.058319039451115, |
|
"grad_norm": 9.68442153930664, |
|
"learning_rate": 2.9416809605488855e-05, |
|
"loss": 0.4846, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.058319039451115, |
|
"eval_accuracy": 0.8833118833118833, |
|
"eval_loss": 0.3212469518184662, |
|
"eval_runtime": 14.6038, |
|
"eval_samples_per_second": 159.616, |
|
"eval_steps_per_second": 19.995, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.1012006861063464, |
|
"grad_norm": 3.7848308086395264, |
|
"learning_rate": 2.8987993138936536e-05, |
|
"loss": 0.4533, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 2.144082332761578, |
|
"grad_norm": 8.865078926086426, |
|
"learning_rate": 2.855917667238422e-05, |
|
"loss": 0.4545, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.144082332761578, |
|
"eval_accuracy": 0.8828828828828829, |
|
"eval_loss": 0.3031620681285858, |
|
"eval_runtime": 14.3961, |
|
"eval_samples_per_second": 161.919, |
|
"eval_steps_per_second": 20.283, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.1869639794168094, |
|
"grad_norm": 7.019817352294922, |
|
"learning_rate": 2.8130360205831908e-05, |
|
"loss": 0.4766, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.2298456260720414, |
|
"grad_norm": 4.404627799987793, |
|
"learning_rate": 2.770154373927959e-05, |
|
"loss": 0.4345, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.2298456260720414, |
|
"eval_accuracy": 0.8923208923208923, |
|
"eval_loss": 0.2739025354385376, |
|
"eval_runtime": 14.9267, |
|
"eval_samples_per_second": 156.163, |
|
"eval_steps_per_second": 19.562, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.2727272727272725, |
|
"grad_norm": 4.765737056732178, |
|
"learning_rate": 2.7272727272727273e-05, |
|
"loss": 0.424, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.3156089193825045, |
|
"grad_norm": 31.984413146972656, |
|
"learning_rate": 2.684391080617496e-05, |
|
"loss": 0.4697, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.3156089193825045, |
|
"eval_accuracy": 0.8931788931788932, |
|
"eval_loss": 0.2758677005767822, |
|
"eval_runtime": 14.2259, |
|
"eval_samples_per_second": 163.856, |
|
"eval_steps_per_second": 20.526, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.358490566037736, |
|
"grad_norm": 5.218204498291016, |
|
"learning_rate": 2.641509433962264e-05, |
|
"loss": 0.4927, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.4013722126929675, |
|
"grad_norm": 19.66315269470215, |
|
"learning_rate": 2.5986277873070326e-05, |
|
"loss": 0.4313, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.4013722126929675, |
|
"eval_accuracy": 0.9051909051909052, |
|
"eval_loss": 0.25858309864997864, |
|
"eval_runtime": 14.6303, |
|
"eval_samples_per_second": 159.327, |
|
"eval_steps_per_second": 19.959, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.444253859348199, |
|
"grad_norm": 1.5584404468536377, |
|
"learning_rate": 2.5557461406518014e-05, |
|
"loss": 0.4136, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.4871355060034306, |
|
"grad_norm": 6.387534141540527, |
|
"learning_rate": 2.5128644939965695e-05, |
|
"loss": 0.4215, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.4871355060034306, |
|
"eval_accuracy": 0.9004719004719005, |
|
"eval_loss": 0.2652466893196106, |
|
"eval_runtime": 14.1925, |
|
"eval_samples_per_second": 164.242, |
|
"eval_steps_per_second": 20.574, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.530017152658662, |
|
"grad_norm": 24.956953048706055, |
|
"learning_rate": 2.4699828473413382e-05, |
|
"loss": 0.4174, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.5728987993138936, |
|
"grad_norm": 4.991686820983887, |
|
"learning_rate": 2.4271012006861067e-05, |
|
"loss": 0.4504, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.5728987993138936, |
|
"eval_accuracy": 0.8927498927498928, |
|
"eval_loss": 0.2829649746417999, |
|
"eval_runtime": 13.721, |
|
"eval_samples_per_second": 169.885, |
|
"eval_steps_per_second": 21.281, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.615780445969125, |
|
"grad_norm": 8.409246444702148, |
|
"learning_rate": 2.3842195540308747e-05, |
|
"loss": 0.433, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.6586620926243567, |
|
"grad_norm": 20.707319259643555, |
|
"learning_rate": 2.3413379073756435e-05, |
|
"loss": 0.4218, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.6586620926243567, |
|
"eval_accuracy": 0.8936078936078936, |
|
"eval_loss": 0.2900983989238739, |
|
"eval_runtime": 13.1413, |
|
"eval_samples_per_second": 177.38, |
|
"eval_steps_per_second": 22.22, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.7015437392795882, |
|
"grad_norm": 6.371370792388916, |
|
"learning_rate": 2.298456260720412e-05, |
|
"loss": 0.454, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.7444253859348198, |
|
"grad_norm": 2.8135769367218018, |
|
"learning_rate": 2.25557461406518e-05, |
|
"loss": 0.4889, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.7444253859348198, |
|
"eval_accuracy": 0.8910338910338911, |
|
"eval_loss": 0.2830560803413391, |
|
"eval_runtime": 13.2735, |
|
"eval_samples_per_second": 175.613, |
|
"eval_steps_per_second": 21.999, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.7873070325900513, |
|
"grad_norm": 4.3516058921813965, |
|
"learning_rate": 2.2126929674099488e-05, |
|
"loss": 0.5019, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.830188679245283, |
|
"grad_norm": 10.607779502868652, |
|
"learning_rate": 2.1698113207547172e-05, |
|
"loss": 0.47, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.830188679245283, |
|
"eval_accuracy": 0.8927498927498928, |
|
"eval_loss": 0.2812536060810089, |
|
"eval_runtime": 13.5257, |
|
"eval_samples_per_second": 172.338, |
|
"eval_steps_per_second": 21.589, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.873070325900515, |
|
"grad_norm": 3.962139844894409, |
|
"learning_rate": 2.1269296740994853e-05, |
|
"loss": 0.3699, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 2.915951972555746, |
|
"grad_norm": 10.8749361038208, |
|
"learning_rate": 2.084048027444254e-05, |
|
"loss": 0.4758, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.915951972555746, |
|
"eval_accuracy": 0.8944658944658944, |
|
"eval_loss": 0.2788408696651459, |
|
"eval_runtime": 13.5439, |
|
"eval_samples_per_second": 172.107, |
|
"eval_steps_per_second": 21.56, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.958833619210978, |
|
"grad_norm": 20.483320236206055, |
|
"learning_rate": 2.0411663807890225e-05, |
|
"loss": 0.4581, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 3.0017152658662094, |
|
"grad_norm": 12.251073837280273, |
|
"learning_rate": 1.998284734133791e-05, |
|
"loss": 0.4482, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 3.0017152658662094, |
|
"eval_accuracy": 0.8953238953238953, |
|
"eval_loss": 0.26482564210891724, |
|
"eval_runtime": 13.5212, |
|
"eval_samples_per_second": 172.396, |
|
"eval_steps_per_second": 21.596, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 3.044596912521441, |
|
"grad_norm": 6.609012603759766, |
|
"learning_rate": 1.9554030874785594e-05, |
|
"loss": 0.4557, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 3.0874785591766725, |
|
"grad_norm": 31.05156898498535, |
|
"learning_rate": 1.9125214408233278e-05, |
|
"loss": 0.3735, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 3.0874785591766725, |
|
"eval_accuracy": 0.8987558987558988, |
|
"eval_loss": 0.25668570399284363, |
|
"eval_runtime": 13.5267, |
|
"eval_samples_per_second": 172.325, |
|
"eval_steps_per_second": 21.587, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 3.130360205831904, |
|
"grad_norm": 33.095829010009766, |
|
"learning_rate": 1.8696397941680962e-05, |
|
"loss": 0.3992, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 3.1732418524871355, |
|
"grad_norm": 10.586121559143066, |
|
"learning_rate": 1.8267581475128647e-05, |
|
"loss": 0.4338, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 3.1732418524871355, |
|
"eval_accuracy": 0.9086229086229086, |
|
"eval_loss": 0.2435009777545929, |
|
"eval_runtime": 13.3274, |
|
"eval_samples_per_second": 174.903, |
|
"eval_steps_per_second": 21.91, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 3.216123499142367, |
|
"grad_norm": 2.1484551429748535, |
|
"learning_rate": 1.783876500857633e-05, |
|
"loss": 0.4019, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 3.2590051457975986, |
|
"grad_norm": 1.3401719331741333, |
|
"learning_rate": 1.7409948542024015e-05, |
|
"loss": 0.3423, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 3.2590051457975986, |
|
"eval_accuracy": 0.9021879021879022, |
|
"eval_loss": 0.25972625613212585, |
|
"eval_runtime": 14.1124, |
|
"eval_samples_per_second": 165.174, |
|
"eval_steps_per_second": 20.691, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 3.30188679245283, |
|
"grad_norm": 2.6398582458496094, |
|
"learning_rate": 1.69811320754717e-05, |
|
"loss": 0.3813, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 3.3447684391080617, |
|
"grad_norm": 8.979487419128418, |
|
"learning_rate": 1.6552315608919384e-05, |
|
"loss": 0.3689, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 3.3447684391080617, |
|
"eval_accuracy": 0.9077649077649078, |
|
"eval_loss": 0.261087566614151, |
|
"eval_runtime": 13.587, |
|
"eval_samples_per_second": 171.561, |
|
"eval_steps_per_second": 21.491, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 3.387650085763293, |
|
"grad_norm": 4.436942100524902, |
|
"learning_rate": 1.6123499142367068e-05, |
|
"loss": 0.4099, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 3.4305317324185247, |
|
"grad_norm": 4.749749660491943, |
|
"learning_rate": 1.5694682675814752e-05, |
|
"loss": 0.347, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.4305317324185247, |
|
"eval_accuracy": 0.9103389103389103, |
|
"eval_loss": 0.24434834718704224, |
|
"eval_runtime": 13.493, |
|
"eval_samples_per_second": 172.757, |
|
"eval_steps_per_second": 21.641, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.4734133790737562, |
|
"grad_norm": 62.90760040283203, |
|
"learning_rate": 1.5265866209262437e-05, |
|
"loss": 0.3681, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 3.516295025728988, |
|
"grad_norm": 29.556215286254883, |
|
"learning_rate": 1.4837049742710121e-05, |
|
"loss": 0.2854, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 3.516295025728988, |
|
"eval_accuracy": 0.9051909051909052, |
|
"eval_loss": 0.25928205251693726, |
|
"eval_runtime": 13.3683, |
|
"eval_samples_per_second": 174.368, |
|
"eval_steps_per_second": 21.843, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 3.5591766723842193, |
|
"grad_norm": 7.2456207275390625, |
|
"learning_rate": 1.4408233276157807e-05, |
|
"loss": 0.4479, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 3.6020583190394513, |
|
"grad_norm": 9.010571479797363, |
|
"learning_rate": 1.397941680960549e-05, |
|
"loss": 0.3513, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.6020583190394513, |
|
"eval_accuracy": 0.9043329043329044, |
|
"eval_loss": 0.2614918053150177, |
|
"eval_runtime": 13.6533, |
|
"eval_samples_per_second": 170.728, |
|
"eval_steps_per_second": 21.387, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.644939965694683, |
|
"grad_norm": 8.282753944396973, |
|
"learning_rate": 1.3550600343053174e-05, |
|
"loss": 0.3045, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 3.6878216123499143, |
|
"grad_norm": 8.212636947631836, |
|
"learning_rate": 1.312178387650086e-05, |
|
"loss": 0.4081, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.6878216123499143, |
|
"eval_accuracy": 0.9073359073359073, |
|
"eval_loss": 0.25328412652015686, |
|
"eval_runtime": 13.7186, |
|
"eval_samples_per_second": 169.915, |
|
"eval_steps_per_second": 21.285, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.730703259005146, |
|
"grad_norm": 1.7500250339508057, |
|
"learning_rate": 1.2692967409948542e-05, |
|
"loss": 0.3552, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 3.7735849056603774, |
|
"grad_norm": 38.88422393798828, |
|
"learning_rate": 1.2264150943396227e-05, |
|
"loss": 0.3346, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.7735849056603774, |
|
"eval_accuracy": 0.9021879021879022, |
|
"eval_loss": 0.27931204438209534, |
|
"eval_runtime": 15.0366, |
|
"eval_samples_per_second": 155.022, |
|
"eval_steps_per_second": 19.419, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.816466552315609, |
|
"grad_norm": 18.03175163269043, |
|
"learning_rate": 1.1835334476843911e-05, |
|
"loss": 0.3502, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 3.8593481989708405, |
|
"grad_norm": 6.491177082061768, |
|
"learning_rate": 1.1406518010291597e-05, |
|
"loss": 0.3495, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.8593481989708405, |
|
"eval_accuracy": 0.9056199056199056, |
|
"eval_loss": 0.2461409866809845, |
|
"eval_runtime": 14.6288, |
|
"eval_samples_per_second": 159.344, |
|
"eval_steps_per_second": 19.961, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.902229845626072, |
|
"grad_norm": 22.98536491394043, |
|
"learning_rate": 1.097770154373928e-05, |
|
"loss": 0.3662, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 3.9451114922813035, |
|
"grad_norm": 24.458087921142578, |
|
"learning_rate": 1.0548885077186965e-05, |
|
"loss": 0.3979, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.9451114922813035, |
|
"eval_accuracy": 0.9103389103389103, |
|
"eval_loss": 0.24052156507968903, |
|
"eval_runtime": 14.1187, |
|
"eval_samples_per_second": 165.1, |
|
"eval_steps_per_second": 20.682, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.987993138936535, |
|
"grad_norm": 22.207815170288086, |
|
"learning_rate": 1.012006861063465e-05, |
|
"loss": 0.379, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 4.030874785591767, |
|
"grad_norm": 19.702865600585938, |
|
"learning_rate": 9.691252144082332e-06, |
|
"loss": 0.3441, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 4.030874785591767, |
|
"eval_accuracy": 0.9017589017589017, |
|
"eval_loss": 0.27815285325050354, |
|
"eval_runtime": 14.2114, |
|
"eval_samples_per_second": 164.024, |
|
"eval_steps_per_second": 20.547, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 4.073756432246999, |
|
"grad_norm": 9.863823890686035, |
|
"learning_rate": 9.262435677530018e-06, |
|
"loss": 0.4128, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 4.11663807890223, |
|
"grad_norm": 13.863909721374512, |
|
"learning_rate": 8.833619210977703e-06, |
|
"loss": 0.3357, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 4.11663807890223, |
|
"eval_accuracy": 0.9094809094809094, |
|
"eval_loss": 0.26747584342956543, |
|
"eval_runtime": 13.9108, |
|
"eval_samples_per_second": 167.568, |
|
"eval_steps_per_second": 20.991, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 4.159519725557462, |
|
"grad_norm": 45.47288131713867, |
|
"learning_rate": 8.404802744425385e-06, |
|
"loss": 0.3301, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 4.202401372212693, |
|
"grad_norm": 4.5591864585876465, |
|
"learning_rate": 7.975986277873071e-06, |
|
"loss": 0.2867, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 4.202401372212693, |
|
"eval_accuracy": 0.9060489060489061, |
|
"eval_loss": 0.27105966210365295, |
|
"eval_runtime": 15.175, |
|
"eval_samples_per_second": 153.608, |
|
"eval_steps_per_second": 19.242, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 4.245283018867925, |
|
"grad_norm": 0.21429812908172607, |
|
"learning_rate": 7.547169811320755e-06, |
|
"loss": 0.3518, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 4.288164665523156, |
|
"grad_norm": 4.873584747314453, |
|
"learning_rate": 7.11835334476844e-06, |
|
"loss": 0.3084, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.288164665523156, |
|
"eval_accuracy": 0.9116259116259117, |
|
"eval_loss": 0.24336792528629303, |
|
"eval_runtime": 13.447, |
|
"eval_samples_per_second": 173.347, |
|
"eval_steps_per_second": 21.715, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.331046312178388, |
|
"grad_norm": 3.682224988937378, |
|
"learning_rate": 6.689536878216124e-06, |
|
"loss": 0.3335, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 4.373927958833619, |
|
"grad_norm": 8.78370475769043, |
|
"learning_rate": 6.2607204116638075e-06, |
|
"loss": 0.4865, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 4.373927958833619, |
|
"eval_accuracy": 0.9154869154869155, |
|
"eval_loss": 0.2427978515625, |
|
"eval_runtime": 13.4143, |
|
"eval_samples_per_second": 173.77, |
|
"eval_steps_per_second": 21.768, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 4.416809605488851, |
|
"grad_norm": 5.311465263366699, |
|
"learning_rate": 5.831903945111493e-06, |
|
"loss": 0.3465, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 4.459691252144083, |
|
"grad_norm": 9.669562339782715, |
|
"learning_rate": 5.403087478559177e-06, |
|
"loss": 0.3557, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 4.459691252144083, |
|
"eval_accuracy": 0.9141999141999142, |
|
"eval_loss": 0.2416524589061737, |
|
"eval_runtime": 13.3034, |
|
"eval_samples_per_second": 175.218, |
|
"eval_steps_per_second": 21.949, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 4.502572898799314, |
|
"grad_norm": 8.7540283203125, |
|
"learning_rate": 4.974271012006861e-06, |
|
"loss": 0.3254, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 4.545454545454545, |
|
"grad_norm": 0.46742549538612366, |
|
"learning_rate": 4.5454545454545455e-06, |
|
"loss": 0.3384, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 4.545454545454545, |
|
"eval_accuracy": 0.9133419133419134, |
|
"eval_loss": 0.24515940248966217, |
|
"eval_runtime": 13.6233, |
|
"eval_samples_per_second": 171.103, |
|
"eval_steps_per_second": 21.434, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 4.588336192109777, |
|
"grad_norm": 23.387781143188477, |
|
"learning_rate": 4.116638078902231e-06, |
|
"loss": 0.3508, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 4.631217838765009, |
|
"grad_norm": 3.3974390029907227, |
|
"learning_rate": 3.687821612349914e-06, |
|
"loss": 0.3147, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 4.631217838765009, |
|
"eval_accuracy": 0.9141999141999142, |
|
"eval_loss": 0.24695377051830292, |
|
"eval_runtime": 13.4806, |
|
"eval_samples_per_second": 172.915, |
|
"eval_steps_per_second": 21.661, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 4.67409948542024, |
|
"grad_norm": 26.67882537841797, |
|
"learning_rate": 3.259005145797599e-06, |
|
"loss": 0.3364, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 4.716981132075472, |
|
"grad_norm": 0.8814979791641235, |
|
"learning_rate": 2.830188679245283e-06, |
|
"loss": 0.2858, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 4.716981132075472, |
|
"eval_accuracy": 0.9103389103389103, |
|
"eval_loss": 0.24419519305229187, |
|
"eval_runtime": 13.9344, |
|
"eval_samples_per_second": 167.284, |
|
"eval_steps_per_second": 20.955, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 4.759862778730703, |
|
"grad_norm": 13.250667572021484, |
|
"learning_rate": 2.4013722126929674e-06, |
|
"loss": 0.3674, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 4.802744425385935, |
|
"grad_norm": 11.3861722946167, |
|
"learning_rate": 1.972555746140652e-06, |
|
"loss": 0.3664, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 4.802744425385935, |
|
"eval_accuracy": 0.9064779064779065, |
|
"eval_loss": 0.2746657431125641, |
|
"eval_runtime": 13.6304, |
|
"eval_samples_per_second": 171.015, |
|
"eval_steps_per_second": 21.423, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 4.845626072041166, |
|
"grad_norm": 3.5501041412353516, |
|
"learning_rate": 1.5437392795883364e-06, |
|
"loss": 0.3152, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 4.888507718696398, |
|
"grad_norm": 24.59325408935547, |
|
"learning_rate": 1.1149228130360207e-06, |
|
"loss": 0.3549, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 4.888507718696398, |
|
"eval_accuracy": 0.9099099099099099, |
|
"eval_loss": 0.26241084933280945, |
|
"eval_runtime": 15.5182, |
|
"eval_samples_per_second": 150.211, |
|
"eval_steps_per_second": 18.817, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 4.931389365351629, |
|
"grad_norm": 32.23326110839844, |
|
"learning_rate": 6.86106346483705e-07, |
|
"loss": 0.3796, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 4.974271012006861, |
|
"grad_norm": 30.20063591003418, |
|
"learning_rate": 2.572898799313894e-07, |
|
"loss": 0.3656, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 4.974271012006861, |
|
"eval_accuracy": 0.9047619047619048, |
|
"eval_loss": 0.27714791893959045, |
|
"eval_runtime": 14.0154, |
|
"eval_samples_per_second": 166.317, |
|
"eval_steps_per_second": 20.834, |
|
"step": 5800 |
|
} |
|
], |
|
"logging_steps": 50, |
|
"max_steps": 5830, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.7169922825846784e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|