{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.1998400213304894, "eval_steps": 500, "global_step": 4500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03, "grad_norm": 0.5563874321368991, "learning_rate": 4.000000000000001e-06, "loss": 1.58, "step": 100 }, { "epoch": 0.05, "grad_norm": 0.4957053999852371, "learning_rate": 8.000000000000001e-06, "loss": 1.4463, "step": 200 }, { "epoch": 0.08, "grad_norm": 0.589756030703148, "learning_rate": 1.2e-05, "loss": 1.3542, "step": 300 }, { "epoch": 0.11, "grad_norm": 0.6420302678887735, "learning_rate": 1.6000000000000003e-05, "loss": 1.3331, "step": 400 }, { "epoch": 0.13, "grad_norm": 0.7649283840590947, "learning_rate": 2e-05, "loss": 1.3253, "step": 500 }, { "epoch": 0.13, "eval_loss": 1.3249897956848145, "eval_runtime": 2000.6322, "eval_samples_per_second": 6.665, "eval_steps_per_second": 0.833, "step": 500 }, { "epoch": 0.16, "grad_norm": 0.7507878377667085, "learning_rate": 1.9714285714285718e-05, "loss": 1.3221, "step": 600 }, { "epoch": 0.19, "grad_norm": 0.6176603895840452, "learning_rate": 1.942857142857143e-05, "loss": 1.3126, "step": 700 }, { "epoch": 0.21, "grad_norm": 0.7327398805078954, "learning_rate": 1.9142857142857146e-05, "loss": 1.3055, "step": 800 }, { "epoch": 0.24, "grad_norm": 0.6839292302496646, "learning_rate": 1.885714285714286e-05, "loss": 1.3093, "step": 900 }, { "epoch": 0.27, "grad_norm": 0.607074650571682, "learning_rate": 1.8571428571428575e-05, "loss": 1.2985, "step": 1000 }, { "epoch": 0.27, "eval_loss": 1.2995541095733643, "eval_runtime": 1951.5754, "eval_samples_per_second": 6.833, "eval_steps_per_second": 0.854, "step": 1000 }, { "epoch": 0.29, "grad_norm": 0.74727636749315, "learning_rate": 1.8285714285714288e-05, "loss": 1.3017, "step": 1100 }, { "epoch": 0.32, "grad_norm": 0.6311499573020434, "learning_rate": 1.8e-05, "loss": 1.2896, "step": 1200 }, { "epoch": 0.35, "grad_norm": 0.5915205156004358, "learning_rate": 1.7714285714285717e-05, "loss": 1.2834, "step": 1300 }, { "epoch": 0.37, "grad_norm": 0.5946814358633083, "learning_rate": 1.742857142857143e-05, "loss": 1.3073, "step": 1400 }, { "epoch": 0.4, "grad_norm": 0.6097257778288048, "learning_rate": 1.7142857142857142e-05, "loss": 1.2843, "step": 1500 }, { "epoch": 0.4, "eval_loss": 1.2924432754516602, "eval_runtime": 1987.1893, "eval_samples_per_second": 6.71, "eval_steps_per_second": 0.839, "step": 1500 }, { "epoch": 0.43, "grad_norm": 0.596713525589904, "learning_rate": 1.6857142857142858e-05, "loss": 1.2748, "step": 1600 }, { "epoch": 0.45, "grad_norm": 0.6872784300870175, "learning_rate": 1.6571428571428574e-05, "loss": 1.2971, "step": 1700 }, { "epoch": 0.48, "grad_norm": 0.5519473130229932, "learning_rate": 1.6285714285714287e-05, "loss": 1.2746, "step": 1800 }, { "epoch": 0.51, "grad_norm": 0.6633345934082727, "learning_rate": 1.6000000000000003e-05, "loss": 1.2906, "step": 1900 }, { "epoch": 0.53, "grad_norm": 0.6476833370052498, "learning_rate": 1.5714285714285715e-05, "loss": 1.2813, "step": 2000 }, { "epoch": 0.53, "eval_loss": 1.2873681783676147, "eval_runtime": 1954.0241, "eval_samples_per_second": 6.824, "eval_steps_per_second": 0.853, "step": 2000 }, { "epoch": 0.56, "grad_norm": 0.6831512981144761, "learning_rate": 1.542857142857143e-05, "loss": 1.2973, "step": 2100 }, { "epoch": 0.59, "grad_norm": 0.6111369458671693, "learning_rate": 1.5142857142857144e-05, "loss": 1.2847, "step": 2200 }, { "epoch": 0.61, "grad_norm": 0.6468631061437224, "learning_rate": 1.4857142857142858e-05, "loss": 1.2754, "step": 2300 }, { "epoch": 0.64, "grad_norm": 0.6799634395265591, "learning_rate": 1.4571428571428573e-05, "loss": 1.2857, "step": 2400 }, { "epoch": 0.67, "grad_norm": 0.7108542425010391, "learning_rate": 1.4285714285714287e-05, "loss": 1.289, "step": 2500 }, { "epoch": 0.67, "eval_loss": 1.283585786819458, "eval_runtime": 1870.2159, "eval_samples_per_second": 7.13, "eval_steps_per_second": 0.891, "step": 2500 }, { "epoch": 0.69, "grad_norm": 0.7253629154301096, "learning_rate": 1.4e-05, "loss": 1.2793, "step": 2600 }, { "epoch": 0.72, "grad_norm": 0.614991663938332, "learning_rate": 1.3714285714285716e-05, "loss": 1.2767, "step": 2700 }, { "epoch": 0.75, "grad_norm": 0.5699407855446496, "learning_rate": 1.3428571428571429e-05, "loss": 1.2719, "step": 2800 }, { "epoch": 0.77, "grad_norm": 0.763568295530126, "learning_rate": 1.3142857142857145e-05, "loss": 1.2733, "step": 2900 }, { "epoch": 0.8, "grad_norm": 0.6538876064837678, "learning_rate": 1.2857142857142859e-05, "loss": 1.2822, "step": 3000 }, { "epoch": 0.8, "eval_loss": 1.2801251411437988, "eval_runtime": 1889.0286, "eval_samples_per_second": 7.059, "eval_steps_per_second": 0.882, "step": 3000 }, { "epoch": 0.83, "grad_norm": 0.6413250796291871, "learning_rate": 1.2571428571428572e-05, "loss": 1.2709, "step": 3100 }, { "epoch": 0.85, "grad_norm": 0.7125599551490435, "learning_rate": 1.2285714285714288e-05, "loss": 1.2829, "step": 3200 }, { "epoch": 0.88, "grad_norm": 0.6792031678140497, "learning_rate": 1.2e-05, "loss": 1.2728, "step": 3300 }, { "epoch": 0.91, "grad_norm": 0.6989586202409576, "learning_rate": 1.1714285714285716e-05, "loss": 1.2835, "step": 3400 }, { "epoch": 0.93, "grad_norm": 0.7569471708203911, "learning_rate": 1.1428571428571429e-05, "loss": 1.2793, "step": 3500 }, { "epoch": 0.93, "eval_loss": 1.2773550748825073, "eval_runtime": 1896.8518, "eval_samples_per_second": 7.03, "eval_steps_per_second": 0.879, "step": 3500 }, { "epoch": 0.96, "grad_norm": 0.6242374298012554, "learning_rate": 1.1142857142857143e-05, "loss": 1.2805, "step": 3600 }, { "epoch": 0.99, "grad_norm": 0.6585613373985719, "learning_rate": 1.0857142857142858e-05, "loss": 1.265, "step": 3700 }, { "epoch": 1.01, "grad_norm": 0.7452400076297672, "learning_rate": 1.0571428571428572e-05, "loss": 1.2589, "step": 3800 }, { "epoch": 1.04, "grad_norm": 0.7861020289489161, "learning_rate": 1.0285714285714285e-05, "loss": 1.2616, "step": 3900 }, { "epoch": 1.07, "grad_norm": 0.7757094950303476, "learning_rate": 1e-05, "loss": 1.2721, "step": 4000 }, { "epoch": 1.07, "eval_loss": 1.2751243114471436, "eval_runtime": 1895.5553, "eval_samples_per_second": 7.035, "eval_steps_per_second": 0.879, "step": 4000 }, { "epoch": 1.09, "grad_norm": 0.6999523427612176, "learning_rate": 9.714285714285715e-06, "loss": 1.2649, "step": 4100 }, { "epoch": 1.12, "grad_norm": 0.6528269639884913, "learning_rate": 9.42857142857143e-06, "loss": 1.2675, "step": 4200 }, { "epoch": 1.15, "grad_norm": 0.8019204201326198, "learning_rate": 9.142857142857144e-06, "loss": 1.2775, "step": 4300 }, { "epoch": 1.17, "grad_norm": 0.8453041527524593, "learning_rate": 8.857142857142858e-06, "loss": 1.2682, "step": 4400 }, { "epoch": 1.2, "grad_norm": 0.6858153075780669, "learning_rate": 8.571428571428571e-06, "loss": 1.2753, "step": 4500 }, { "epoch": 1.2, "eval_loss": 1.2731624841690063, "eval_runtime": 2031.5498, "eval_samples_per_second": 6.564, "eval_steps_per_second": 0.821, "step": 4500 } ], "logging_steps": 100, "max_steps": 7500, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "total_flos": 6.185622662887244e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }