{ "best_metric": 1.3951435089111328, "best_model_checkpoint": "./results/checkpoint-18616", "epoch": 6.0, "eval_steps": 500, "global_step": 27924, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.10743446497636441, "grad_norm": 4.838443756103516, "learning_rate": 4.9462827675118175e-05, "loss": 3.2819, "step": 500 }, { "epoch": 0.21486892995272883, "grad_norm": 3.5501515865325928, "learning_rate": 4.892565535023636e-05, "loss": 2.3948, "step": 1000 }, { "epoch": 0.32230339492909327, "grad_norm": 3.361682653427124, "learning_rate": 4.8388483025354535e-05, "loss": 2.1166, "step": 1500 }, { "epoch": 0.42973785990545765, "grad_norm": 3.207756996154785, "learning_rate": 4.7851310700472715e-05, "loss": 1.983, "step": 2000 }, { "epoch": 0.5371723248818221, "grad_norm": 3.38034725189209, "learning_rate": 4.7314138375590894e-05, "loss": 1.8785, "step": 2500 }, { "epoch": 0.6446067898581865, "grad_norm": 3.3562798500061035, "learning_rate": 4.677696605070907e-05, "loss": 1.7957, "step": 3000 }, { "epoch": 0.752041254834551, "grad_norm": 2.897372007369995, "learning_rate": 4.623979372582725e-05, "loss": 1.7387, "step": 3500 }, { "epoch": 0.8594757198109153, "grad_norm": 2.9306259155273438, "learning_rate": 4.570262140094543e-05, "loss": 1.7103, "step": 4000 }, { "epoch": 0.9669101847872797, "grad_norm": 2.7418951988220215, "learning_rate": 4.5165449076063606e-05, "loss": 1.6637, "step": 4500 }, { "epoch": 1.0, "eval_loss": 1.5266185998916626, "eval_rouge2_fmeasure": 0.4101, "eval_rouge2_precision": 0.4101, "eval_rouge2_recall": 0.4101, "eval_runtime": 72.9922, "eval_samples_per_second": 65.582, "eval_steps_per_second": 1.028, "step": 4654 }, { "epoch": 1.0743446497636442, "grad_norm": 3.1563010215759277, "learning_rate": 4.462827675118178e-05, "loss": 1.4927, "step": 5000 }, { "epoch": 1.1817791147400085, "grad_norm": 2.845944404602051, "learning_rate": 4.409110442629996e-05, "loss": 1.4188, "step": 5500 }, { "epoch": 1.289213579716373, "grad_norm": 2.7506182193756104, "learning_rate": 4.355393210141814e-05, "loss": 1.418, "step": 6000 }, { "epoch": 1.3966480446927374, "grad_norm": 2.623769760131836, "learning_rate": 4.301675977653631e-05, "loss": 1.406, "step": 6500 }, { "epoch": 1.504082509669102, "grad_norm": 2.671093225479126, "learning_rate": 4.247958745165449e-05, "loss": 1.4021, "step": 7000 }, { "epoch": 1.6115169746454663, "grad_norm": 2.96425724029541, "learning_rate": 4.194241512677267e-05, "loss": 1.3883, "step": 7500 }, { "epoch": 1.7189514396218306, "grad_norm": 2.6702911853790283, "learning_rate": 4.140524280189085e-05, "loss": 1.3883, "step": 8000 }, { "epoch": 1.8263859045981952, "grad_norm": 2.6645655632019043, "learning_rate": 4.0868070477009024e-05, "loss": 1.3761, "step": 8500 }, { "epoch": 1.9338203695745595, "grad_norm": 3.023545265197754, "learning_rate": 4.0330898152127204e-05, "loss": 1.376, "step": 9000 }, { "epoch": 2.0, "eval_loss": 1.410889744758606, "eval_rouge2_fmeasure": 0.4238, "eval_rouge2_precision": 0.4238, "eval_rouge2_recall": 0.4238, "eval_runtime": 72.9371, "eval_samples_per_second": 65.632, "eval_steps_per_second": 1.028, "step": 9308 }, { "epoch": 2.041254834550924, "grad_norm": 2.6128861904144287, "learning_rate": 3.979372582724538e-05, "loss": 1.2778, "step": 9500 }, { "epoch": 2.1486892995272884, "grad_norm": 2.602187395095825, "learning_rate": 3.9256553502363556e-05, "loss": 1.1422, "step": 10000 }, { "epoch": 2.256123764503653, "grad_norm": 2.725907802581787, "learning_rate": 3.8719381177481736e-05, "loss": 1.1449, "step": 10500 }, { "epoch": 2.363558229480017, "grad_norm": 2.6348464488983154, "learning_rate": 3.8182208852599916e-05, "loss": 1.1529, "step": 11000 }, { "epoch": 2.4709926944563816, "grad_norm": 2.8584792613983154, "learning_rate": 3.764503652771809e-05, "loss": 1.1552, "step": 11500 }, { "epoch": 2.578427159432746, "grad_norm": 2.59346342086792, "learning_rate": 3.7107864202836275e-05, "loss": 1.1666, "step": 12000 }, { "epoch": 2.6858616244091102, "grad_norm": 2.655982255935669, "learning_rate": 3.657069187795445e-05, "loss": 1.1658, "step": 12500 }, { "epoch": 2.793296089385475, "grad_norm": 2.5818564891815186, "learning_rate": 3.603351955307263e-05, "loss": 1.1674, "step": 13000 }, { "epoch": 2.9007305543618394, "grad_norm": 2.8041298389434814, "learning_rate": 3.549634722819081e-05, "loss": 1.1684, "step": 13500 }, { "epoch": 3.0, "eval_loss": 1.3994433879852295, "eval_rouge2_fmeasure": 0.4311, "eval_rouge2_precision": 0.4311, "eval_rouge2_recall": 0.4311, "eval_runtime": 72.9642, "eval_samples_per_second": 65.607, "eval_steps_per_second": 1.028, "step": 13962 }, { "epoch": 3.008165019338204, "grad_norm": 2.646003246307373, "learning_rate": 3.495917490330898e-05, "loss": 1.1485, "step": 14000 }, { "epoch": 3.115599484314568, "grad_norm": 2.8257687091827393, "learning_rate": 3.442200257842716e-05, "loss": 0.9456, "step": 14500 }, { "epoch": 3.2230339492909326, "grad_norm": 2.9319422245025635, "learning_rate": 3.388483025354534e-05, "loss": 0.961, "step": 15000 }, { "epoch": 3.330468414267297, "grad_norm": 2.7773501873016357, "learning_rate": 3.334765792866352e-05, "loss": 0.9748, "step": 15500 }, { "epoch": 3.4379028792436612, "grad_norm": 2.673140048980713, "learning_rate": 3.281048560378169e-05, "loss": 0.9822, "step": 16000 }, { "epoch": 3.5453373442200258, "grad_norm": 2.7191991806030273, "learning_rate": 3.227331327889987e-05, "loss": 0.9808, "step": 16500 }, { "epoch": 3.6527718091963903, "grad_norm": 2.717005491256714, "learning_rate": 3.173614095401805e-05, "loss": 0.9832, "step": 17000 }, { "epoch": 3.760206274172755, "grad_norm": 2.696438789367676, "learning_rate": 3.1198968629136225e-05, "loss": 0.9954, "step": 17500 }, { "epoch": 3.867640739149119, "grad_norm": 2.5392231941223145, "learning_rate": 3.066179630425441e-05, "loss": 0.9978, "step": 18000 }, { "epoch": 3.9750752041254835, "grad_norm": 3.3882222175598145, "learning_rate": 3.0124623979372585e-05, "loss": 0.9957, "step": 18500 }, { "epoch": 4.0, "eval_loss": 1.3951435089111328, "eval_rouge2_fmeasure": 0.4323, "eval_rouge2_precision": 0.4323, "eval_rouge2_recall": 0.4323, "eval_runtime": 72.9673, "eval_samples_per_second": 65.605, "eval_steps_per_second": 1.028, "step": 18616 }, { "epoch": 4.082509669101848, "grad_norm": 2.655376672744751, "learning_rate": 2.958745165449076e-05, "loss": 0.8449, "step": 19000 }, { "epoch": 4.189944134078212, "grad_norm": 2.5333900451660156, "learning_rate": 2.9050279329608944e-05, "loss": 0.8125, "step": 19500 }, { "epoch": 4.297378599054577, "grad_norm": 2.4350857734680176, "learning_rate": 2.8513107004727117e-05, "loss": 0.8184, "step": 20000 }, { "epoch": 4.404813064030941, "grad_norm": 2.595292329788208, "learning_rate": 2.7975934679845293e-05, "loss": 0.8317, "step": 20500 }, { "epoch": 4.512247529007306, "grad_norm": 2.4971327781677246, "learning_rate": 2.743876235496347e-05, "loss": 0.833, "step": 21000 }, { "epoch": 4.61968199398367, "grad_norm": 2.7136857509613037, "learning_rate": 2.6901590030081653e-05, "loss": 0.8361, "step": 21500 }, { "epoch": 4.727116458960034, "grad_norm": 2.8456897735595703, "learning_rate": 2.636441770519983e-05, "loss": 0.8458, "step": 22000 }, { "epoch": 4.834550923936399, "grad_norm": 2.601877212524414, "learning_rate": 2.5827245380318005e-05, "loss": 0.846, "step": 22500 }, { "epoch": 4.941985388912763, "grad_norm": 2.6655149459838867, "learning_rate": 2.5290073055436185e-05, "loss": 0.848, "step": 23000 }, { "epoch": 5.0, "eval_loss": 1.42265784740448, "eval_rouge2_fmeasure": 0.4307, "eval_rouge2_precision": 0.4307, "eval_rouge2_recall": 0.4307, "eval_runtime": 73.0163, "eval_samples_per_second": 65.561, "eval_steps_per_second": 1.027, "step": 23270 }, { "epoch": 5.049419853889128, "grad_norm": 2.525017023086548, "learning_rate": 2.475290073055436e-05, "loss": 0.7735, "step": 23500 }, { "epoch": 5.156854318865492, "grad_norm": 2.2720000743865967, "learning_rate": 2.421572840567254e-05, "loss": 0.6797, "step": 24000 }, { "epoch": 5.264288783841857, "grad_norm": 2.633282423019409, "learning_rate": 2.3678556080790718e-05, "loss": 0.6955, "step": 24500 }, { "epoch": 5.3717232488182205, "grad_norm": 2.6699864864349365, "learning_rate": 2.3141383755908897e-05, "loss": 0.7026, "step": 25000 }, { "epoch": 5.479157713794585, "grad_norm": 2.765111207962036, "learning_rate": 2.2604211431027074e-05, "loss": 0.7045, "step": 25500 }, { "epoch": 5.58659217877095, "grad_norm": 2.8248813152313232, "learning_rate": 2.206703910614525e-05, "loss": 0.7148, "step": 26000 }, { "epoch": 5.694026643747314, "grad_norm": 2.7719056606292725, "learning_rate": 2.152986678126343e-05, "loss": 0.7218, "step": 26500 }, { "epoch": 5.801461108723679, "grad_norm": 2.8628671169281006, "learning_rate": 2.099269445638161e-05, "loss": 0.717, "step": 27000 }, { "epoch": 5.908895573700043, "grad_norm": 2.479224443435669, "learning_rate": 2.0455522131499786e-05, "loss": 0.7247, "step": 27500 }, { "epoch": 6.0, "eval_loss": 1.479453682899475, "eval_rouge2_fmeasure": 0.43, "eval_rouge2_precision": 0.43, "eval_rouge2_recall": 0.43, "eval_runtime": 73.0025, "eval_samples_per_second": 65.573, "eval_steps_per_second": 1.027, "step": 27924 } ], "logging_steps": 500, "max_steps": 46540, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 2 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.9760389250501837e+17, "train_batch_size": 64, "trial_name": null, "trial_params": null }