|
{ |
|
"best_metric": 0.31399768590927124, |
|
"best_model_checkpoint": "deberta-v3-xsmall-zyda-2-quality-rerun/checkpoint-39948", |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 39948, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.037548813457494744, |
|
"grad_norm": 4.403994560241699, |
|
"learning_rate": 4.9374186442375094e-05, |
|
"loss": 0.6773, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.07509762691498949, |
|
"grad_norm": 3.1282551288604736, |
|
"learning_rate": 4.874837288475018e-05, |
|
"loss": 0.5374, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.11264644037248422, |
|
"grad_norm": 3.248645782470703, |
|
"learning_rate": 4.812255932712526e-05, |
|
"loss": 0.5056, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.15019525382997898, |
|
"grad_norm": 3.0767080783843994, |
|
"learning_rate": 4.749674576950035e-05, |
|
"loss": 0.4772, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.18774406728747373, |
|
"grad_norm": 4.454887866973877, |
|
"learning_rate": 4.687093221187544e-05, |
|
"loss": 0.4574, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.22529288074496845, |
|
"grad_norm": 2.398064613342285, |
|
"learning_rate": 4.624511865425053e-05, |
|
"loss": 0.4508, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.2628416942024632, |
|
"grad_norm": 3.7262675762176514, |
|
"learning_rate": 4.5619305096625615e-05, |
|
"loss": 0.432, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.30039050765995795, |
|
"grad_norm": 2.6357853412628174, |
|
"learning_rate": 4.4993491539000706e-05, |
|
"loss": 0.4303, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.3379393211174527, |
|
"grad_norm": 2.8107378482818604, |
|
"learning_rate": 4.436767798137579e-05, |
|
"loss": 0.4151, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.37548813457494745, |
|
"grad_norm": 3.1814627647399902, |
|
"learning_rate": 4.3741864423750875e-05, |
|
"loss": 0.411, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.4130369480324422, |
|
"grad_norm": 3.0693929195404053, |
|
"learning_rate": 4.3116050866125966e-05, |
|
"loss": 0.4013, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.4505857614899369, |
|
"grad_norm": 3.04325008392334, |
|
"learning_rate": 4.249023730850105e-05, |
|
"loss": 0.4042, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.4881345749474317, |
|
"grad_norm": 2.3595287799835205, |
|
"learning_rate": 4.186442375087614e-05, |
|
"loss": 0.395, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.5256833884049265, |
|
"grad_norm": 3.9787838459014893, |
|
"learning_rate": 4.123861019325123e-05, |
|
"loss": 0.3969, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.5632322018624212, |
|
"grad_norm": 5.200424671173096, |
|
"learning_rate": 4.061279663562632e-05, |
|
"loss": 0.3919, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.6007810153199159, |
|
"grad_norm": 4.694857120513916, |
|
"learning_rate": 3.99869830780014e-05, |
|
"loss": 0.3926, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.6383298287774106, |
|
"grad_norm": 2.551344633102417, |
|
"learning_rate": 3.936116952037649e-05, |
|
"loss": 0.3815, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.6758786422349053, |
|
"grad_norm": 4.175814151763916, |
|
"learning_rate": 3.873535596275158e-05, |
|
"loss": 0.3805, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.7134274556924001, |
|
"grad_norm": 3.3641836643218994, |
|
"learning_rate": 3.810954240512667e-05, |
|
"loss": 0.3813, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.7509762691498949, |
|
"grad_norm": 4.312653064727783, |
|
"learning_rate": 3.7483728847501754e-05, |
|
"loss": 0.3687, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.7885250826073896, |
|
"grad_norm": 2.912576675415039, |
|
"learning_rate": 3.6857915289876846e-05, |
|
"loss": 0.3669, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.8260738960648843, |
|
"grad_norm": 3.152151346206665, |
|
"learning_rate": 3.623210173225193e-05, |
|
"loss": 0.3746, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.8636227095223791, |
|
"grad_norm": 2.863006830215454, |
|
"learning_rate": 3.5606288174627015e-05, |
|
"loss": 0.37, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.9011715229798738, |
|
"grad_norm": 2.383753776550293, |
|
"learning_rate": 3.4980474617002106e-05, |
|
"loss": 0.3675, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.9387203364373686, |
|
"grad_norm": 2.4308314323425293, |
|
"learning_rate": 3.435466105937719e-05, |
|
"loss": 0.3607, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.9762691498948634, |
|
"grad_norm": 3.8357925415039062, |
|
"learning_rate": 3.372884750175228e-05, |
|
"loss": 0.3569, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.34360167384147644, |
|
"eval_mse": 0.34360167908090056, |
|
"eval_runtime": 105.793, |
|
"eval_samples_per_second": 945.242, |
|
"eval_steps_per_second": 118.155, |
|
"step": 13316 |
|
}, |
|
{ |
|
"epoch": 1.013817963352358, |
|
"grad_norm": 3.0041775703430176, |
|
"learning_rate": 3.310303394412737e-05, |
|
"loss": 0.3414, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.051366776809853, |
|
"grad_norm": 3.3583767414093018, |
|
"learning_rate": 3.247722038650246e-05, |
|
"loss": 0.3027, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.0889155902673475, |
|
"grad_norm": 4.092809200286865, |
|
"learning_rate": 3.185140682887754e-05, |
|
"loss": 0.3098, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.1264644037248424, |
|
"grad_norm": 2.405085802078247, |
|
"learning_rate": 3.122559327125263e-05, |
|
"loss": 0.3094, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.164013217182337, |
|
"grad_norm": 2.9946043491363525, |
|
"learning_rate": 3.059977971362772e-05, |
|
"loss": 0.3054, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.2015620306398318, |
|
"grad_norm": 2.638746976852417, |
|
"learning_rate": 2.9973966156002803e-05, |
|
"loss": 0.3053, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.2391108440973264, |
|
"grad_norm": 2.7500743865966797, |
|
"learning_rate": 2.9348152598377894e-05, |
|
"loss": 0.3027, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.2766596575548212, |
|
"grad_norm": 2.1603994369506836, |
|
"learning_rate": 2.8722339040752982e-05, |
|
"loss": 0.2978, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.314208471012316, |
|
"grad_norm": 3.147937774658203, |
|
"learning_rate": 2.8096525483128067e-05, |
|
"loss": 0.3065, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.3517572844698107, |
|
"grad_norm": 2.8613197803497314, |
|
"learning_rate": 2.7470711925503158e-05, |
|
"loss": 0.3034, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.3893060979273055, |
|
"grad_norm": 2.1770453453063965, |
|
"learning_rate": 2.684489836787824e-05, |
|
"loss": 0.296, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.4268549113848001, |
|
"grad_norm": 3.3030712604522705, |
|
"learning_rate": 2.621908481025333e-05, |
|
"loss": 0.2953, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.464403724842295, |
|
"grad_norm": 2.322561025619507, |
|
"learning_rate": 2.559327125262842e-05, |
|
"loss": 0.3018, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.5019525382997898, |
|
"grad_norm": 3.038954496383667, |
|
"learning_rate": 2.4967457695003503e-05, |
|
"loss": 0.2969, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.5395013517572844, |
|
"grad_norm": 2.606443405151367, |
|
"learning_rate": 2.4341644137378594e-05, |
|
"loss": 0.2901, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.5770501652147793, |
|
"grad_norm": 2.5407485961914062, |
|
"learning_rate": 2.3715830579753682e-05, |
|
"loss": 0.2962, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.6145989786722739, |
|
"grad_norm": 2.8466265201568604, |
|
"learning_rate": 2.3090017022128767e-05, |
|
"loss": 0.2967, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.6521477921297687, |
|
"grad_norm": 3.01155161857605, |
|
"learning_rate": 2.2464203464503855e-05, |
|
"loss": 0.294, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.6896966055872635, |
|
"grad_norm": 3.2597897052764893, |
|
"learning_rate": 2.1838389906878946e-05, |
|
"loss": 0.2938, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.7272454190447581, |
|
"grad_norm": 2.202085256576538, |
|
"learning_rate": 2.121257634925403e-05, |
|
"loss": 0.2852, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.764794232502253, |
|
"grad_norm": 2.2789671421051025, |
|
"learning_rate": 2.058676279162912e-05, |
|
"loss": 0.2927, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.8023430459597476, |
|
"grad_norm": 2.1892921924591064, |
|
"learning_rate": 1.9960949234004207e-05, |
|
"loss": 0.2912, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.8398918594172424, |
|
"grad_norm": 2.7529070377349854, |
|
"learning_rate": 1.9335135676379295e-05, |
|
"loss": 0.2886, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.8774406728747373, |
|
"grad_norm": 2.7882585525512695, |
|
"learning_rate": 1.8709322118754383e-05, |
|
"loss": 0.2933, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.9149894863322319, |
|
"grad_norm": 2.5822184085845947, |
|
"learning_rate": 1.808350856112947e-05, |
|
"loss": 0.2883, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.9525382997897265, |
|
"grad_norm": 2.326474666595459, |
|
"learning_rate": 1.7457695003504555e-05, |
|
"loss": 0.2887, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.9900871132472213, |
|
"grad_norm": 3.619825601577759, |
|
"learning_rate": 1.6831881445879643e-05, |
|
"loss": 0.2857, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.3208652436733246, |
|
"eval_mse": 0.3208652417918357, |
|
"eval_runtime": 108.1625, |
|
"eval_samples_per_second": 924.535, |
|
"eval_steps_per_second": 115.567, |
|
"step": 26632 |
|
}, |
|
{ |
|
"epoch": 2.027635926704716, |
|
"grad_norm": 3.537344217300415, |
|
"learning_rate": 1.620606788825473e-05, |
|
"loss": 0.2578, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.065184740162211, |
|
"grad_norm": 2.6542530059814453, |
|
"learning_rate": 1.558025433062982e-05, |
|
"loss": 0.2438, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 2.102733553619706, |
|
"grad_norm": 3.0753347873687744, |
|
"learning_rate": 1.4954440773004907e-05, |
|
"loss": 0.2493, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.1402823670772, |
|
"grad_norm": 2.988065004348755, |
|
"learning_rate": 1.4328627215379995e-05, |
|
"loss": 0.2484, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 2.177831180534695, |
|
"grad_norm": 2.041884422302246, |
|
"learning_rate": 1.3702813657755081e-05, |
|
"loss": 0.2509, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.21537999399219, |
|
"grad_norm": 3.0710532665252686, |
|
"learning_rate": 1.307700010013017e-05, |
|
"loss": 0.2471, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 2.2529288074496847, |
|
"grad_norm": 2.620396852493286, |
|
"learning_rate": 1.2451186542505257e-05, |
|
"loss": 0.245, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.290477620907179, |
|
"grad_norm": 3.103252410888672, |
|
"learning_rate": 1.1825372984880345e-05, |
|
"loss": 0.2468, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 2.328026434364674, |
|
"grad_norm": 2.4032323360443115, |
|
"learning_rate": 1.1199559427255433e-05, |
|
"loss": 0.2441, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.3655752478221688, |
|
"grad_norm": 3.1325552463531494, |
|
"learning_rate": 1.057374586963052e-05, |
|
"loss": 0.2431, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 2.4031240612796636, |
|
"grad_norm": 3.0277721881866455, |
|
"learning_rate": 9.947932312005607e-06, |
|
"loss": 0.2414, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.4406728747371584, |
|
"grad_norm": 2.5161995887756348, |
|
"learning_rate": 9.322118754380697e-06, |
|
"loss": 0.2412, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 2.478221688194653, |
|
"grad_norm": 1.7039533853530884, |
|
"learning_rate": 8.696305196755783e-06, |
|
"loss": 0.2434, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.5157705016521477, |
|
"grad_norm": 2.513314723968506, |
|
"learning_rate": 8.07049163913087e-06, |
|
"loss": 0.2438, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 2.5533193151096425, |
|
"grad_norm": 2.314403772354126, |
|
"learning_rate": 7.444678081505959e-06, |
|
"loss": 0.2441, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.5908681285671373, |
|
"grad_norm": 2.855802536010742, |
|
"learning_rate": 6.818864523881045e-06, |
|
"loss": 0.2408, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 2.628416942024632, |
|
"grad_norm": 2.636925458908081, |
|
"learning_rate": 6.193050966256133e-06, |
|
"loss": 0.2433, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 2.6659657554821266, |
|
"grad_norm": 2.4678878784179688, |
|
"learning_rate": 5.567237408631221e-06, |
|
"loss": 0.2367, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 2.7035145689396214, |
|
"grad_norm": 2.2376275062561035, |
|
"learning_rate": 4.941423851006309e-06, |
|
"loss": 0.2353, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 2.741063382397116, |
|
"grad_norm": 2.5792624950408936, |
|
"learning_rate": 4.315610293381396e-06, |
|
"loss": 0.2402, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 2.778612195854611, |
|
"grad_norm": 3.4142560958862305, |
|
"learning_rate": 3.6897967357564838e-06, |
|
"loss": 0.2345, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 2.816161009312106, |
|
"grad_norm": 2.2159926891326904, |
|
"learning_rate": 3.063983178131571e-06, |
|
"loss": 0.2378, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 2.8537098227696003, |
|
"grad_norm": 2.577420711517334, |
|
"learning_rate": 2.438169620506659e-06, |
|
"loss": 0.233, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 2.891258636227095, |
|
"grad_norm": 2.7511062622070312, |
|
"learning_rate": 1.8123560628817465e-06, |
|
"loss": 0.2374, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 2.92880744968459, |
|
"grad_norm": 2.820404529571533, |
|
"learning_rate": 1.186542505256834e-06, |
|
"loss": 0.2413, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 2.966356263142085, |
|
"grad_norm": 2.4552481174468994, |
|
"learning_rate": 5.607289476319215e-07, |
|
"loss": 0.2395, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.31399768590927124, |
|
"eval_mse": 0.31399767168875364, |
|
"eval_runtime": 104.865, |
|
"eval_samples_per_second": 953.607, |
|
"eval_steps_per_second": 119.201, |
|
"step": 39948 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 39948, |
|
"total_flos": 4.210143005329459e+16, |
|
"train_loss": 0.3187301256067702, |
|
"train_runtime": 5274.6963, |
|
"train_samples_per_second": 484.67, |
|
"train_steps_per_second": 7.574 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 39948, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.210143005329459e+16, |
|
"train_batch_size": 64, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|