| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 3.0, |
| "eval_steps": 500, |
| "global_step": 351, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.042735042735042736, |
| "grad_norm": 3.711402177810669, |
| "learning_rate": 3.6363636363636366e-06, |
| "loss": 0.2434, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.08547008547008547, |
| "grad_norm": 3.794971227645874, |
| "learning_rate": 8.181818181818183e-06, |
| "loss": 0.1924, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.1282051282051282, |
| "grad_norm": 2.4459402561187744, |
| "learning_rate": 9.998079135987437e-06, |
| "loss": 0.174, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.17094017094017094, |
| "grad_norm": 2.0912649631500244, |
| "learning_rate": 9.98634586692894e-06, |
| "loss": 0.1488, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.21367521367521367, |
| "grad_norm": 1.9830565452575684, |
| "learning_rate": 9.963971484502247e-06, |
| "loss": 0.1332, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.2564102564102564, |
| "grad_norm": 1.4833670854568481, |
| "learning_rate": 9.931003736767013e-06, |
| "loss": 0.1146, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.29914529914529914, |
| "grad_norm": 1.437637448310852, |
| "learning_rate": 9.887512978558329e-06, |
| "loss": 0.1175, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.3418803418803419, |
| "grad_norm": 1.591968297958374, |
| "learning_rate": 9.833592021345938e-06, |
| "loss": 0.1114, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.38461538461538464, |
| "grad_norm": 1.383669376373291, |
| "learning_rate": 9.76935593516989e-06, |
| "loss": 0.1092, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.42735042735042733, |
| "grad_norm": 1.2138135433197021, |
| "learning_rate": 9.694941803075285e-06, |
| "loss": 0.1005, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.4700854700854701, |
| "grad_norm": 1.317816138267517, |
| "learning_rate": 9.610508428570122e-06, |
| "loss": 0.1079, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.5128205128205128, |
| "grad_norm": 1.1774816513061523, |
| "learning_rate": 9.516235996730645e-06, |
| "loss": 0.1003, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.5555555555555556, |
| "grad_norm": 1.3384287357330322, |
| "learning_rate": 9.41232568967728e-06, |
| "loss": 0.1005, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.5982905982905983, |
| "grad_norm": 1.0372443199157715, |
| "learning_rate": 9.298999257241862e-06, |
| "loss": 0.1029, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.6410256410256411, |
| "grad_norm": 1.1703890562057495, |
| "learning_rate": 9.176498543742328e-06, |
| "loss": 0.0988, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.6837606837606838, |
| "grad_norm": 1.1787683963775635, |
| "learning_rate": 9.045084971874738e-06, |
| "loss": 0.0936, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.7264957264957265, |
| "grad_norm": 1.180239200592041, |
| "learning_rate": 8.905038984824079e-06, |
| "loss": 0.1036, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.7692307692307693, |
| "grad_norm": 1.1398251056671143, |
| "learning_rate": 8.756659447784367e-06, |
| "loss": 0.1, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.811965811965812, |
| "grad_norm": 1.2935456037521362, |
| "learning_rate": 8.600263010165275e-06, |
| "loss": 0.0956, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.8547008547008547, |
| "grad_norm": 1.1012020111083984, |
| "learning_rate": 8.436183429846314e-06, |
| "loss": 0.1028, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.8974358974358975, |
| "grad_norm": 1.0556975603103638, |
| "learning_rate": 8.264770860920722e-06, |
| "loss": 0.1077, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.9401709401709402, |
| "grad_norm": 1.069830298423767, |
| "learning_rate": 8.086391106448965e-06, |
| "loss": 0.0976, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.9829059829059829, |
| "grad_norm": 1.019631028175354, |
| "learning_rate": 7.90142483781658e-06, |
| "loss": 0.1008, |
| "step": 115 |
| }, |
| { |
| "epoch": 1.0256410256410255, |
| "grad_norm": 0.7582312226295471, |
| "learning_rate": 7.710266782362248e-06, |
| "loss": 0.0635, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.0683760683760684, |
| "grad_norm": 0.7042792439460754, |
| "learning_rate": 7.513324881009769e-06, |
| "loss": 0.0403, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.1111111111111112, |
| "grad_norm": 0.9356122016906738, |
| "learning_rate": 7.311019417701567e-06, |
| "loss": 0.036, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.1538461538461537, |
| "grad_norm": 0.9737276434898376, |
| "learning_rate": 7.103782122491577e-06, |
| "loss": 0.0361, |
| "step": 135 |
| }, |
| { |
| "epoch": 1.1965811965811965, |
| "grad_norm": 0.9380191564559937, |
| "learning_rate": 6.892055250211552e-06, |
| "loss": 0.0343, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.2393162393162394, |
| "grad_norm": 0.7957591414451599, |
| "learning_rate": 6.67629063667697e-06, |
| "loss": 0.0353, |
| "step": 145 |
| }, |
| { |
| "epoch": 1.282051282051282, |
| "grad_norm": 0.8701767921447754, |
| "learning_rate": 6.456948734446624e-06, |
| "loss": 0.0356, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.3247863247863247, |
| "grad_norm": 0.7799380421638489, |
| "learning_rate": 6.234497630193666e-06, |
| "loss": 0.0373, |
| "step": 155 |
| }, |
| { |
| "epoch": 1.3675213675213675, |
| "grad_norm": 0.8129424452781677, |
| "learning_rate": 6.009412045785051e-06, |
| "loss": 0.0356, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.4102564102564101, |
| "grad_norm": 0.8334594964981079, |
| "learning_rate": 5.782172325201155e-06, |
| "loss": 0.0348, |
| "step": 165 |
| }, |
| { |
| "epoch": 1.452991452991453, |
| "grad_norm": 0.8230361938476562, |
| "learning_rate": 5.553263409457504e-06, |
| "loss": 0.0322, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.4957264957264957, |
| "grad_norm": 1.0665380954742432, |
| "learning_rate": 5.323173801716222e-06, |
| "loss": 0.036, |
| "step": 175 |
| }, |
| { |
| "epoch": 1.5384615384615383, |
| "grad_norm": 1.0307265520095825, |
| "learning_rate": 5.09239452479565e-06, |
| "loss": 0.0316, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.5811965811965814, |
| "grad_norm": 0.8220146894454956, |
| "learning_rate": 4.861418073302919e-06, |
| "loss": 0.0326, |
| "step": 185 |
| }, |
| { |
| "epoch": 1.623931623931624, |
| "grad_norm": 1.0208584070205688, |
| "learning_rate": 4.630737362625631e-06, |
| "loss": 0.0333, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.6666666666666665, |
| "grad_norm": 0.9469585418701172, |
| "learning_rate": 4.400844677025585e-06, |
| "loss": 0.0361, |
| "step": 195 |
| }, |
| { |
| "epoch": 1.7094017094017095, |
| "grad_norm": 0.8382295370101929, |
| "learning_rate": 4.17223061907935e-06, |
| "loss": 0.0356, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.7521367521367521, |
| "grad_norm": 0.7920061945915222, |
| "learning_rate": 3.945383062707652e-06, |
| "loss": 0.033, |
| "step": 205 |
| }, |
| { |
| "epoch": 1.7948717948717947, |
| "grad_norm": 0.7468324899673462, |
| "learning_rate": 3.720786112027822e-06, |
| "loss": 0.0306, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.8376068376068377, |
| "grad_norm": 0.969772219657898, |
| "learning_rate": 3.498919068251237e-06, |
| "loss": 0.0332, |
| "step": 215 |
| }, |
| { |
| "epoch": 1.8803418803418803, |
| "grad_norm": 0.9195359349250793, |
| "learning_rate": 3.2802554068303595e-06, |
| "loss": 0.0321, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.9230769230769231, |
| "grad_norm": 1.082493782043457, |
| "learning_rate": 3.0652617670382745e-06, |
| "loss": 0.0321, |
| "step": 225 |
| }, |
| { |
| "epoch": 1.965811965811966, |
| "grad_norm": 0.8420332074165344, |
| "learning_rate": 2.8543969561369556e-06, |
| "loss": 0.0299, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.0085470085470085, |
| "grad_norm": 0.44002777338027954, |
| "learning_rate": 2.648110970259454e-06, |
| "loss": 0.0271, |
| "step": 235 |
| }, |
| { |
| "epoch": 2.051282051282051, |
| "grad_norm": 0.449357271194458, |
| "learning_rate": 2.4468440340954664e-06, |
| "loss": 0.01, |
| "step": 240 |
| }, |
| { |
| "epoch": 2.094017094017094, |
| "grad_norm": 0.47545549273490906, |
| "learning_rate": 2.2510256614296638e-06, |
| "loss": 0.0091, |
| "step": 245 |
| }, |
| { |
| "epoch": 2.1367521367521367, |
| "grad_norm": 0.623047411441803, |
| "learning_rate": 2.061073738537635e-06, |
| "loss": 0.0084, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.1794871794871793, |
| "grad_norm": 0.376277893781662, |
| "learning_rate": 1.8773936323955055e-06, |
| "loss": 0.0093, |
| "step": 255 |
| }, |
| { |
| "epoch": 2.2222222222222223, |
| "grad_norm": 0.676085889339447, |
| "learning_rate": 1.7003773256063882e-06, |
| "loss": 0.008, |
| "step": 260 |
| }, |
| { |
| "epoch": 2.264957264957265, |
| "grad_norm": 0.5928175449371338, |
| "learning_rate": 1.5304025798897521e-06, |
| "loss": 0.0075, |
| "step": 265 |
| }, |
| { |
| "epoch": 2.3076923076923075, |
| "grad_norm": 0.47083908319473267, |
| "learning_rate": 1.3678321299188802e-06, |
| "loss": 0.0069, |
| "step": 270 |
| }, |
| { |
| "epoch": 2.3504273504273505, |
| "grad_norm": 0.6990512609481812, |
| "learning_rate": 1.213012909226786e-06, |
| "loss": 0.0076, |
| "step": 275 |
| }, |
| { |
| "epoch": 2.393162393162393, |
| "grad_norm": 0.46621406078338623, |
| "learning_rate": 1.066275309832584e-06, |
| "loss": 0.0066, |
| "step": 280 |
| }, |
| { |
| "epoch": 2.435897435897436, |
| "grad_norm": 0.5243452787399292, |
| "learning_rate": 9.279324771682586e-07, |
| "loss": 0.007, |
| "step": 285 |
| }, |
| { |
| "epoch": 2.4786324786324787, |
| "grad_norm": 0.5895669460296631, |
| "learning_rate": 7.98279641810537e-07, |
| "loss": 0.0069, |
| "step": 290 |
| }, |
| { |
| "epoch": 2.5213675213675213, |
| "grad_norm": 0.6112285852432251, |
| "learning_rate": 6.775934894439606e-07, |
| "loss": 0.0083, |
| "step": 295 |
| }, |
| { |
| "epoch": 2.564102564102564, |
| "grad_norm": 0.5791566371917725, |
| "learning_rate": 5.661315703996905e-07, |
| "loss": 0.0087, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.606837606837607, |
| "grad_norm": 0.574300229549408, |
| "learning_rate": 4.641317500301173e-07, |
| "loss": 0.0079, |
| "step": 305 |
| }, |
| { |
| "epoch": 2.6495726495726495, |
| "grad_norm": 0.4943869113922119, |
| "learning_rate": 3.71811701092219e-07, |
| "loss": 0.0074, |
| "step": 310 |
| }, |
| { |
| "epoch": 2.6923076923076925, |
| "grad_norm": 0.4588395059108734, |
| "learning_rate": 2.893684392229185e-07, |
| "loss": 0.0065, |
| "step": 315 |
| }, |
| { |
| "epoch": 2.735042735042735, |
| "grad_norm": 0.48180243372917175, |
| "learning_rate": 2.1697790249779638e-07, |
| "loss": 0.0063, |
| "step": 320 |
| }, |
| { |
| "epoch": 2.7777777777777777, |
| "grad_norm": 0.45809242129325867, |
| "learning_rate": 1.547945759703623e-07, |
| "loss": 0.0068, |
| "step": 325 |
| }, |
| { |
| "epoch": 2.8205128205128203, |
| "grad_norm": 0.5285813212394714, |
| "learning_rate": 1.0295116199317057e-07, |
| "loss": 0.0055, |
| "step": 330 |
| }, |
| { |
| "epoch": 2.8632478632478633, |
| "grad_norm": 0.44899120926856995, |
| "learning_rate": 6.15582970243117e-08, |
| "loss": 0.0088, |
| "step": 335 |
| }, |
| { |
| "epoch": 2.905982905982906, |
| "grad_norm": 0.7157308459281921, |
| "learning_rate": 3.0704315523631956e-08, |
| "loss": 0.0073, |
| "step": 340 |
| }, |
| { |
| "epoch": 2.948717948717949, |
| "grad_norm": 0.7962710857391357, |
| "learning_rate": 1.0455061442548597e-08, |
| "loss": 0.0091, |
| "step": 345 |
| }, |
| { |
| "epoch": 2.9914529914529915, |
| "grad_norm": 0.7418946623802185, |
| "learning_rate": 8.537477097364522e-10, |
| "loss": 0.0082, |
| "step": 350 |
| }, |
| { |
| "epoch": 3.0, |
| "step": 351, |
| "total_flos": 2.73366675523371e+17, |
| "train_loss": 0.053853037605258475, |
| "train_runtime": 459.8341, |
| "train_samples_per_second": 48.755, |
| "train_steps_per_second": 0.763 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 351, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2.73366675523371e+17, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|