{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.4033333333333333, "eval_steps": 500, "global_step": 242, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008333333333333333, "grad_norm": 0.48900601267814636, "learning_rate": 4.999904807660428e-05, "loss": 0.4212, "num_input_tokens_seen": 30064, "step": 5 }, { "epoch": 0.016666666666666666, "grad_norm": 0.3595407009124756, "learning_rate": 4.9996192378909786e-05, "loss": 0.2803, "num_input_tokens_seen": 57376, "step": 10 }, { "epoch": 0.025, "grad_norm": 0.2623545527458191, "learning_rate": 4.999143312438893e-05, "loss": 0.1484, "num_input_tokens_seen": 86160, "step": 15 }, { "epoch": 0.03333333333333333, "grad_norm": 0.2743065357208252, "learning_rate": 4.99847706754774e-05, "loss": 0.1151, "num_input_tokens_seen": 110592, "step": 20 }, { "epoch": 0.041666666666666664, "grad_norm": 0.2651338577270508, "learning_rate": 4.997620553954645e-05, "loss": 0.0751, "num_input_tokens_seen": 138880, "step": 25 }, { "epoch": 0.05, "grad_norm": 0.10798148065805435, "learning_rate": 4.996573836886435e-05, "loss": 0.0471, "num_input_tokens_seen": 169600, "step": 30 }, { "epoch": 0.058333333333333334, "grad_norm": 0.1552838534116745, "learning_rate": 4.9953369960546676e-05, "loss": 0.0368, "num_input_tokens_seen": 201664, "step": 35 }, { "epoch": 0.06666666666666667, "grad_norm": 0.11832074075937271, "learning_rate": 4.993910125649561e-05, "loss": 0.0311, "num_input_tokens_seen": 229936, "step": 40 }, { "epoch": 0.075, "grad_norm": 0.10593917220830917, "learning_rate": 4.99229333433282e-05, "loss": 0.0262, "num_input_tokens_seen": 259088, "step": 45 }, { "epoch": 0.08333333333333333, "grad_norm": 0.09650842100381851, "learning_rate": 4.990486745229364e-05, "loss": 0.0244, "num_input_tokens_seen": 289824, "step": 50 }, { "epoch": 0.09166666666666666, "grad_norm": 0.14155137538909912, "learning_rate": 4.988490495917947e-05, "loss": 0.0266, "num_input_tokens_seen": 317760, "step": 55 }, { "epoch": 0.1, "grad_norm": 0.1490347981452942, "learning_rate": 4.9867570427929354e-05, "loss": 0.0283, "num_input_tokens_seen": 343120, "step": 60 }, { "epoch": 0.10833333333333334, "grad_norm": 0.17539122700691223, "learning_rate": 4.984419797901491e-05, "loss": 0.0265, "num_input_tokens_seen": 371600, "step": 65 }, { "epoch": 0.11666666666666667, "grad_norm": 0.1336919367313385, "learning_rate": 4.981893354823614e-05, "loss": 0.0235, "num_input_tokens_seen": 398720, "step": 70 }, { "epoch": 0.125, "grad_norm": 0.12203177809715271, "learning_rate": 4.979177905957726e-05, "loss": 0.0229, "num_input_tokens_seen": 426416, "step": 75 }, { "epoch": 0.13333333333333333, "grad_norm": 0.2557654082775116, "learning_rate": 4.976273658095772e-05, "loss": 0.0238, "num_input_tokens_seen": 452944, "step": 80 }, { "epoch": 0.14166666666666666, "grad_norm": 0.13327105343341827, "learning_rate": 4.9731808324074717e-05, "loss": 0.0208, "num_input_tokens_seen": 480896, "step": 85 }, { "epoch": 0.15, "grad_norm": 0.1262751668691635, "learning_rate": 4.969899664423473e-05, "loss": 0.0191, "num_input_tokens_seen": 510096, "step": 90 }, { "epoch": 0.15833333333333333, "grad_norm": 0.16084067523479462, "learning_rate": 4.966430404017424e-05, "loss": 0.0205, "num_input_tokens_seen": 536672, "step": 95 }, { "epoch": 0.16666666666666666, "grad_norm": 0.15870437026023865, "learning_rate": 4.962773315386935e-05, "loss": 0.0193, "num_input_tokens_seen": 566528, "step": 100 }, { "epoch": 0.175, "grad_norm": 0.07833187282085419, "learning_rate": 4.9589286770334654e-05, "loss": 0.024, "num_input_tokens_seen": 593584, "step": 105 }, { "epoch": 0.18333333333333332, "grad_norm": 0.09176061302423477, "learning_rate": 4.954896781741109e-05, "loss": 0.0206, "num_input_tokens_seen": 620992, "step": 110 }, { "epoch": 0.19166666666666668, "grad_norm": 0.08285439014434814, "learning_rate": 4.9506779365543046e-05, "loss": 0.0188, "num_input_tokens_seen": 647504, "step": 115 }, { "epoch": 0.2, "grad_norm": 0.13359522819519043, "learning_rate": 4.9462724627544466e-05, "loss": 0.0183, "num_input_tokens_seen": 676336, "step": 120 }, { "epoch": 0.20833333333333334, "grad_norm": 0.07570718228816986, "learning_rate": 4.94168069583542e-05, "loss": 0.0189, "num_input_tokens_seen": 701488, "step": 125 }, { "epoch": 0.21666666666666667, "grad_norm": 0.11286037415266037, "learning_rate": 4.936902985478055e-05, "loss": 0.0213, "num_input_tokens_seen": 731152, "step": 130 }, { "epoch": 0.225, "grad_norm": 0.17183929681777954, "learning_rate": 4.931939695523492e-05, "loss": 0.0214, "num_input_tokens_seen": 757792, "step": 135 }, { "epoch": 0.23333333333333334, "grad_norm": 0.11746834218502045, "learning_rate": 4.926791203945477e-05, "loss": 0.0214, "num_input_tokens_seen": 786064, "step": 140 }, { "epoch": 0.24166666666666667, "grad_norm": 0.09453292936086655, "learning_rate": 4.9214579028215776e-05, "loss": 0.0203, "num_input_tokens_seen": 814640, "step": 145 }, { "epoch": 0.25, "grad_norm": 0.2074723094701767, "learning_rate": 4.915940198303324e-05, "loss": 0.0194, "num_input_tokens_seen": 842656, "step": 150 }, { "epoch": 0.25833333333333336, "grad_norm": 0.15371476113796234, "learning_rate": 4.910238510585276e-05, "loss": 0.0158, "num_input_tokens_seen": 868512, "step": 155 }, { "epoch": 0.26666666666666666, "grad_norm": 0.07177837193012238, "learning_rate": 4.9043532738730284e-05, "loss": 0.0172, "num_input_tokens_seen": 896416, "step": 160 }, { "epoch": 0.275, "grad_norm": 0.07891824096441269, "learning_rate": 4.898284936350144e-05, "loss": 0.0193, "num_input_tokens_seen": 924720, "step": 165 }, { "epoch": 0.2833333333333333, "grad_norm": 0.08302944153547287, "learning_rate": 4.892033960144019e-05, "loss": 0.0189, "num_input_tokens_seen": 955584, "step": 170 }, { "epoch": 0.2916666666666667, "grad_norm": 0.25554358959198, "learning_rate": 4.8856008212906925e-05, "loss": 0.019, "num_input_tokens_seen": 985072, "step": 175 }, { "epoch": 0.3, "grad_norm": 0.1445908546447754, "learning_rate": 4.878986009698596e-05, "loss": 0.0168, "num_input_tokens_seen": 1013296, "step": 180 }, { "epoch": 0.30833333333333335, "grad_norm": 0.21417762339115143, "learning_rate": 4.8721900291112415e-05, "loss": 0.0194, "num_input_tokens_seen": 1043664, "step": 185 }, { "epoch": 0.31666666666666665, "grad_norm": 0.1550753116607666, "learning_rate": 4.8652133970688636e-05, "loss": 0.0155, "num_input_tokens_seen": 1070912, "step": 190 }, { "epoch": 0.325, "grad_norm": 0.132898211479187, "learning_rate": 4.858056644869002e-05, "loss": 0.0193, "num_input_tokens_seen": 1097856, "step": 195 }, { "epoch": 0.3333333333333333, "grad_norm": 0.09164389222860336, "learning_rate": 4.850720317526047e-05, "loss": 0.0164, "num_input_tokens_seen": 1126128, "step": 200 }, { "epoch": 0.3416666666666667, "grad_norm": 0.07462836056947708, "learning_rate": 4.843204973729729e-05, "loss": 0.014, "num_input_tokens_seen": 1157584, "step": 205 }, { "epoch": 0.35, "grad_norm": 0.08473635464906693, "learning_rate": 4.835511185802574e-05, "loss": 0.0154, "num_input_tokens_seen": 1184800, "step": 210 }, { "epoch": 0.35833333333333334, "grad_norm": 0.07371218502521515, "learning_rate": 4.827639539656321e-05, "loss": 0.0147, "num_input_tokens_seen": 1213104, "step": 215 }, { "epoch": 0.36666666666666664, "grad_norm": 0.1394609957933426, "learning_rate": 4.8195906347473e-05, "loss": 0.0185, "num_input_tokens_seen": 1241808, "step": 220 }, { "epoch": 0.375, "grad_norm": 0.11585802584886551, "learning_rate": 4.8113650840307834e-05, "loss": 0.016, "num_input_tokens_seen": 1270160, "step": 225 }, { "epoch": 0.38333333333333336, "grad_norm": 0.1538001149892807, "learning_rate": 4.802963513914304e-05, "loss": 0.0168, "num_input_tokens_seen": 1297456, "step": 230 }, { "epoch": 0.39166666666666666, "grad_norm": 0.09096246212720871, "learning_rate": 4.794386564209953e-05, "loss": 0.0154, "num_input_tokens_seen": 1323392, "step": 235 }, { "epoch": 0.4, "grad_norm": 0.11348401010036469, "learning_rate": 4.7856348880856595e-05, "loss": 0.0149, "num_input_tokens_seen": 1358576, "step": 240 }, { "epoch": 0.4033333333333333, "num_input_tokens_seen": 1371584, "step": 242, "total_flos": 4652852538507264.0, "train_loss": 0.040362141964849364, "train_runtime": 4583.2214, "train_samples_per_second": 6.284, "train_steps_per_second": 0.393 } ], "logging_steps": 5, "max_steps": 1800, "num_input_tokens_seen": 1371584, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 4652852538507264.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }