{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.8561484918793503, "eval_steps": 50, "global_step": 400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02320185614849188, "grad_norm": 26.745355095184312, "learning_rate": 5e-07, "loss": 1.6569, "step": 5 }, { "epoch": 0.04640371229698376, "grad_norm": 13.080188660738868, "learning_rate": 1e-06, "loss": 1.5099, "step": 10 }, { "epoch": 0.06960556844547564, "grad_norm": 7.775505635686351, "learning_rate": 9.998470286265414e-07, "loss": 1.2782, "step": 15 }, { "epoch": 0.09280742459396751, "grad_norm": 6.492859712313467, "learning_rate": 9.993882081071305e-07, "loss": 1.2359, "step": 20 }, { "epoch": 0.11600928074245939, "grad_norm": 5.355551984540261, "learning_rate": 9.986238191873872e-07, "loss": 1.204, "step": 25 }, { "epoch": 0.13921113689095127, "grad_norm": 5.125393165293572, "learning_rate": 9.975543295858033e-07, "loss": 1.1627, "step": 30 }, { "epoch": 0.16241299303944315, "grad_norm": 4.710071299991241, "learning_rate": 9.961803937075514e-07, "loss": 1.1463, "step": 35 }, { "epoch": 0.18561484918793503, "grad_norm": 5.029376799191572, "learning_rate": 9.945028522440653e-07, "loss": 1.1394, "step": 40 }, { "epoch": 0.2088167053364269, "grad_norm": 4.763291242870039, "learning_rate": 9.925227316586314e-07, "loss": 1.1371, "step": 45 }, { "epoch": 0.23201856148491878, "grad_norm": 4.903033982523367, "learning_rate": 9.902412435583125e-07, "loss": 1.1181, "step": 50 }, { "epoch": 0.23201856148491878, "eval_loss": 1.1214605569839478, "eval_runtime": 105.8329, "eval_samples_per_second": 57.912, "eval_steps_per_second": 0.907, "step": 50 }, { "epoch": 0.2552204176334107, "grad_norm": 4.785575700738186, "learning_rate": 9.876597839525813e-07, "loss": 1.1163, "step": 55 }, { "epoch": 0.27842227378190254, "grad_norm": 4.600727000401806, "learning_rate": 9.847799323991233e-07, "loss": 1.1232, "step": 60 }, { "epoch": 0.30162412993039445, "grad_norm": 4.8200416840356315, "learning_rate": 9.816034510373285e-07, "loss": 1.125, "step": 65 }, { "epoch": 0.3248259860788863, "grad_norm": 4.958997518359378, "learning_rate": 9.781322835100637e-07, "loss": 1.108, "step": 70 }, { "epoch": 0.3480278422273782, "grad_norm": 4.883541365508776, "learning_rate": 9.743685537743856e-07, "loss": 1.106, "step": 75 }, { "epoch": 0.37122969837587005, "grad_norm": 4.973507458353338, "learning_rate": 9.70314564801922e-07, "loss": 1.0973, "step": 80 }, { "epoch": 0.39443155452436196, "grad_norm": 4.704415990191669, "learning_rate": 9.659727971697173e-07, "loss": 1.0964, "step": 85 }, { "epoch": 0.4176334106728538, "grad_norm": 4.759885977268913, "learning_rate": 9.613459075424033e-07, "loss": 1.0956, "step": 90 }, { "epoch": 0.4408352668213457, "grad_norm": 4.868535908803129, "learning_rate": 9.564367270466245e-07, "loss": 1.0787, "step": 95 }, { "epoch": 0.46403712296983757, "grad_norm": 5.180286116736628, "learning_rate": 9.51248259538713e-07, "loss": 1.0765, "step": 100 }, { "epoch": 0.46403712296983757, "eval_loss": 1.0775035619735718, "eval_runtime": 105.5293, "eval_samples_per_second": 58.079, "eval_steps_per_second": 0.91, "step": 100 }, { "epoch": 0.4872389791183295, "grad_norm": 5.290465762761348, "learning_rate": 9.457836797666721e-07, "loss": 1.0903, "step": 105 }, { "epoch": 0.5104408352668214, "grad_norm": 4.81291157554945, "learning_rate": 9.400463314275941e-07, "loss": 1.0697, "step": 110 }, { "epoch": 0.5336426914153132, "grad_norm": 4.914554202012043, "learning_rate": 9.340397251217008e-07, "loss": 1.0668, "step": 115 }, { "epoch": 0.5568445475638051, "grad_norm": 5.240457841494325, "learning_rate": 9.27767536204258e-07, "loss": 1.0676, "step": 120 }, { "epoch": 0.580046403712297, "grad_norm": 4.957459385263701, "learning_rate": 9.212336025366787e-07, "loss": 1.0746, "step": 125 }, { "epoch": 0.6032482598607889, "grad_norm": 5.29032668711839, "learning_rate": 9.144419221381918e-07, "loss": 1.0724, "step": 130 }, { "epoch": 0.6264501160092807, "grad_norm": 4.908560953587426, "learning_rate": 9.073966507395121e-07, "loss": 1.0745, "step": 135 }, { "epoch": 0.6496519721577726, "grad_norm": 4.912842113728852, "learning_rate": 9.001020992400085e-07, "loss": 1.0559, "step": 140 }, { "epoch": 0.6728538283062645, "grad_norm": 5.088585906783296, "learning_rate": 8.925627310699274e-07, "loss": 1.0705, "step": 145 }, { "epoch": 0.6960556844547564, "grad_norm": 5.140684832177941, "learning_rate": 8.84783159459285e-07, "loss": 1.0639, "step": 150 }, { "epoch": 0.6960556844547564, "eval_loss": 1.0501643419265747, "eval_runtime": 105.4561, "eval_samples_per_second": 58.119, "eval_steps_per_second": 0.91, "step": 150 }, { "epoch": 0.7192575406032483, "grad_norm": 5.311257433234373, "learning_rate": 8.767681446150976e-07, "loss": 1.0472, "step": 155 }, { "epoch": 0.7424593967517401, "grad_norm": 5.091539509688025, "learning_rate": 8.68522590808682e-07, "loss": 1.0645, "step": 160 }, { "epoch": 0.765661252900232, "grad_norm": 5.132013982763288, "learning_rate": 8.600515433748001e-07, "loss": 1.0416, "step": 165 }, { "epoch": 0.7888631090487239, "grad_norm": 4.753354098230195, "learning_rate": 8.51360185624495e-07, "loss": 1.0478, "step": 170 }, { "epoch": 0.8120649651972158, "grad_norm": 5.029473978539478, "learning_rate": 8.424538356734956e-07, "loss": 1.0383, "step": 175 }, { "epoch": 0.8352668213457076, "grad_norm": 4.9588553004593345, "learning_rate": 8.333379431881397e-07, "loss": 1.0342, "step": 180 }, { "epoch": 0.8584686774941995, "grad_norm": 5.234591483099779, "learning_rate": 8.240180860508026e-07, "loss": 1.0413, "step": 185 }, { "epoch": 0.8816705336426914, "grad_norm": 5.121566469508508, "learning_rate": 8.144999669468713e-07, "loss": 1.0264, "step": 190 }, { "epoch": 0.9048723897911833, "grad_norm": 5.0479045768726305, "learning_rate": 8.047894098753539e-07, "loss": 1.028, "step": 195 }, { "epoch": 0.9280742459396751, "grad_norm": 5.0838098259091185, "learning_rate": 7.948923565852597e-07, "loss": 1.0308, "step": 200 }, { "epoch": 0.9280742459396751, "eval_loss": 1.0281875133514404, "eval_runtime": 105.8568, "eval_samples_per_second": 57.899, "eval_steps_per_second": 0.907, "step": 200 }, { "epoch": 0.951276102088167, "grad_norm": 5.3244675969022826, "learning_rate": 7.848148629399285e-07, "loss": 1.0262, "step": 205 }, { "epoch": 0.974477958236659, "grad_norm": 4.9307215762355305, "learning_rate": 7.745630952115363e-07, "loss": 1.0349, "step": 210 }, { "epoch": 0.9976798143851509, "grad_norm": 4.994203203030838, "learning_rate": 7.641433263080418e-07, "loss": 1.0216, "step": 215 }, { "epoch": 1.0208816705336428, "grad_norm": 5.112958880673586, "learning_rate": 7.535619319348865e-07, "loss": 0.9241, "step": 220 }, { "epoch": 1.0440835266821347, "grad_norm": 5.264187445397404, "learning_rate": 7.428253866937918e-07, "loss": 0.9001, "step": 225 }, { "epoch": 1.0672853828306264, "grad_norm": 5.645584402922182, "learning_rate": 7.319402601210447e-07, "loss": 0.8916, "step": 230 }, { "epoch": 1.0904872389791183, "grad_norm": 5.655360994963379, "learning_rate": 7.209132126676933e-07, "loss": 0.8876, "step": 235 }, { "epoch": 1.1136890951276102, "grad_norm": 5.3773890810778795, "learning_rate": 7.097509916241145e-07, "loss": 0.8931, "step": 240 }, { "epoch": 1.136890951276102, "grad_norm": 5.658881203794, "learning_rate": 6.984604269914436e-07, "loss": 0.905, "step": 245 }, { "epoch": 1.160092807424594, "grad_norm": 5.966282577193694, "learning_rate": 6.870484273023967e-07, "loss": 0.9038, "step": 250 }, { "epoch": 1.160092807424594, "eval_loss": 1.0220295190811157, "eval_runtime": 105.8362, "eval_samples_per_second": 57.91, "eval_steps_per_second": 0.907, "step": 250 }, { "epoch": 1.1832946635730859, "grad_norm": 5.794176185315156, "learning_rate": 6.755219753940388e-07, "loss": 0.8964, "step": 255 }, { "epoch": 1.2064965197215778, "grad_norm": 6.603391500331007, "learning_rate": 6.638881241350883e-07, "loss": 0.8898, "step": 260 }, { "epoch": 1.2296983758700697, "grad_norm": 5.5914639272443205, "learning_rate": 6.52153992110368e-07, "loss": 0.8951, "step": 265 }, { "epoch": 1.2529002320185616, "grad_norm": 5.339661007608592, "learning_rate": 6.403267592650466e-07, "loss": 0.8961, "step": 270 }, { "epoch": 1.2761020881670533, "grad_norm": 5.448280965038798, "learning_rate": 6.28413662511334e-07, "loss": 0.8919, "step": 275 }, { "epoch": 1.2993039443155452, "grad_norm": 5.476822697700394, "learning_rate": 6.164219913003207e-07, "loss": 0.8931, "step": 280 }, { "epoch": 1.322505800464037, "grad_norm": 5.783548079343189, "learning_rate": 6.043590831616676e-07, "loss": 0.8792, "step": 285 }, { "epoch": 1.345707656612529, "grad_norm": 5.59782698134665, "learning_rate": 5.92232319213878e-07, "loss": 0.8768, "step": 290 }, { "epoch": 1.368909512761021, "grad_norm": 5.193853086769952, "learning_rate": 5.800491196478988e-07, "loss": 0.8788, "step": 295 }, { "epoch": 1.3921113689095128, "grad_norm": 5.539347488257, "learning_rate": 5.678169391868127e-07, "loss": 0.8973, "step": 300 }, { "epoch": 1.3921113689095128, "eval_loss": 1.0114275217056274, "eval_runtime": 106.216, "eval_samples_per_second": 57.703, "eval_steps_per_second": 0.904, "step": 300 }, { "epoch": 1.4153132250580047, "grad_norm": 5.567338787725618, "learning_rate": 5.555432625244023e-07, "loss": 0.8831, "step": 305 }, { "epoch": 1.4385150812064964, "grad_norm": 5.412598997121907, "learning_rate": 5.432355997453728e-07, "loss": 0.8848, "step": 310 }, { "epoch": 1.4617169373549883, "grad_norm": 5.4356510240439775, "learning_rate": 5.309014817300421e-07, "loss": 0.8999, "step": 315 }, { "epoch": 1.4849187935034802, "grad_norm": 5.525607667062919, "learning_rate": 5.185484555463026e-07, "loss": 0.8901, "step": 320 }, { "epoch": 1.5081206496519721, "grad_norm": 5.583006624663847, "learning_rate": 5.061840798316814e-07, "loss": 0.8909, "step": 325 }, { "epoch": 1.531322505800464, "grad_norm": 5.822776934487761, "learning_rate": 4.938159201683186e-07, "loss": 0.8829, "step": 330 }, { "epoch": 1.554524361948956, "grad_norm": 5.427885443572571, "learning_rate": 4.814515444536974e-07, "loss": 0.8867, "step": 335 }, { "epoch": 1.5777262180974478, "grad_norm": 5.513594905050496, "learning_rate": 4.69098518269958e-07, "loss": 0.892, "step": 340 }, { "epoch": 1.6009280742459397, "grad_norm": 5.785273130658459, "learning_rate": 4.5676440025462726e-07, "loss": 0.8775, "step": 345 }, { "epoch": 1.6241299303944317, "grad_norm": 5.494906178164733, "learning_rate": 4.444567374755977e-07, "loss": 0.8747, "step": 350 }, { "epoch": 1.6241299303944317, "eval_loss": 1.0039345026016235, "eval_runtime": 105.8025, "eval_samples_per_second": 57.929, "eval_steps_per_second": 0.907, "step": 350 }, { "epoch": 1.6473317865429236, "grad_norm": 5.49251321009188, "learning_rate": 4.3218306081318713e-07, "loss": 0.884, "step": 355 }, { "epoch": 1.6705336426914155, "grad_norm": 5.618196571147986, "learning_rate": 4.199508803521012e-07, "loss": 0.8945, "step": 360 }, { "epoch": 1.6937354988399071, "grad_norm": 5.908619967180135, "learning_rate": 4.0776768078612207e-07, "loss": 0.8793, "step": 365 }, { "epoch": 1.716937354988399, "grad_norm": 5.822870470090775, "learning_rate": 3.9564091683833244e-07, "loss": 0.8785, "step": 370 }, { "epoch": 1.740139211136891, "grad_norm": 5.685661727934108, "learning_rate": 3.835780086996793e-07, "loss": 0.8772, "step": 375 }, { "epoch": 1.7633410672853829, "grad_norm": 5.692617224399981, "learning_rate": 3.7158633748866607e-07, "loss": 0.8701, "step": 380 }, { "epoch": 1.7865429234338746, "grad_norm": 5.6393286368292355, "learning_rate": 3.596732407349536e-07, "loss": 0.871, "step": 385 }, { "epoch": 1.8097447795823665, "grad_norm": 5.672110777773612, "learning_rate": 3.4784600788963193e-07, "loss": 0.8751, "step": 390 }, { "epoch": 1.8329466357308584, "grad_norm": 5.835389029793195, "learning_rate": 3.3611187586491157e-07, "loss": 0.8687, "step": 395 }, { "epoch": 1.8561484918793503, "grad_norm": 5.386053225084359, "learning_rate": 3.244780246059612e-07, "loss": 0.8818, "step": 400 }, { "epoch": 1.8561484918793503, "eval_loss": 0.996471107006073, "eval_runtime": 105.8008, "eval_samples_per_second": 57.93, "eval_steps_per_second": 0.907, "step": 400 } ], "logging_steps": 5, "max_steps": 645, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2358347349098496.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }