{ "best_metric": 0.9036697247706422, "best_model_checkpoint": "distilbert-base-uncased-finetuned-sst2/run-0/checkpoint-4210", "epoch": 5.0, "eval_steps": 500, "global_step": 21050, "is_hyper_param_search": true, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.1187648456057007, "grad_norm": 9.763530731201172, "learning_rate": 2.4130303407723998e-05, "loss": 0.3459, "step": 500 }, { "epoch": 0.2375296912114014, "grad_norm": 7.90731954574585, "learning_rate": 2.3543191402669885e-05, "loss": 0.2557, "step": 1000 }, { "epoch": 0.35629453681710216, "grad_norm": 7.188111305236816, "learning_rate": 2.2956079397615776e-05, "loss": 0.2242, "step": 1500 }, { "epoch": 0.4750593824228028, "grad_norm": 9.416199684143066, "learning_rate": 2.2368967392561663e-05, "loss": 0.2281, "step": 2000 }, { "epoch": 0.5938242280285035, "grad_norm": 6.982799053192139, "learning_rate": 2.1781855387507554e-05, "loss": 0.2196, "step": 2500 }, { "epoch": 0.7125890736342043, "grad_norm": 3.475909948348999, "learning_rate": 2.1194743382453438e-05, "loss": 0.199, "step": 3000 }, { "epoch": 0.831353919239905, "grad_norm": 18.930234909057617, "learning_rate": 2.0607631377399328e-05, "loss": 0.1988, "step": 3500 }, { "epoch": 0.9501187648456056, "grad_norm": 21.44248390197754, "learning_rate": 2.0020519372345215e-05, "loss": 0.1964, "step": 4000 }, { "epoch": 1.0, "eval_accuracy": 0.9036697247706422, "eval_loss": 0.279630184173584, "eval_runtime": 1.1981, "eval_samples_per_second": 727.831, "eval_steps_per_second": 45.907, "step": 4210 }, { "epoch": 1.0688836104513064, "grad_norm": 6.991998672485352, "learning_rate": 1.9433407367291103e-05, "loss": 0.1449, "step": 4500 }, { "epoch": 1.187648456057007, "grad_norm": 1.0682291984558105, "learning_rate": 1.8846295362236993e-05, "loss": 0.1307, "step": 5000 }, { "epoch": 1.3064133016627077, "grad_norm": 1.4475390911102295, "learning_rate": 1.825918335718288e-05, "loss": 0.1248, "step": 5500 }, { "epoch": 1.4251781472684084, "grad_norm": 0.4122050702571869, "learning_rate": 1.767207135212877e-05, "loss": 0.1288, "step": 6000 }, { "epoch": 1.5439429928741093, "grad_norm": 0.35570672154426575, "learning_rate": 1.7084959347074655e-05, "loss": 0.1353, "step": 6500 }, { "epoch": 1.66270783847981, "grad_norm": 0.9978006482124329, "learning_rate": 1.6497847342020545e-05, "loss": 0.1316, "step": 7000 }, { "epoch": 1.7814726840855108, "grad_norm": 7.568444728851318, "learning_rate": 1.5910735336966433e-05, "loss": 0.1363, "step": 7500 }, { "epoch": 1.9002375296912115, "grad_norm": 0.1252310872077942, "learning_rate": 1.532362333191232e-05, "loss": 0.1271, "step": 8000 }, { "epoch": 2.0, "eval_accuracy": 0.8887614678899083, "eval_loss": 0.37789419293403625, "eval_runtime": 1.2, "eval_samples_per_second": 726.651, "eval_steps_per_second": 45.832, "step": 8420 }, { "epoch": 2.019002375296912, "grad_norm": 0.2641623616218567, "learning_rate": 1.4736511326858209e-05, "loss": 0.114, "step": 8500 }, { "epoch": 2.137767220902613, "grad_norm": 16.91781234741211, "learning_rate": 1.4149399321804098e-05, "loss": 0.0814, "step": 9000 }, { "epoch": 2.2565320665083135, "grad_norm": 0.13691911101341248, "learning_rate": 1.3562287316749987e-05, "loss": 0.0807, "step": 9500 }, { "epoch": 2.375296912114014, "grad_norm": 14.652079582214355, "learning_rate": 1.2975175311695872e-05, "loss": 0.0909, "step": 10000 }, { "epoch": 2.494061757719715, "grad_norm": 0.23481616377830505, "learning_rate": 1.2388063306641761e-05, "loss": 0.0823, "step": 10500 }, { "epoch": 2.6128266033254155, "grad_norm": 0.16778022050857544, "learning_rate": 1.1800951301587648e-05, "loss": 0.0866, "step": 11000 }, { "epoch": 2.731591448931116, "grad_norm": 22.94564437866211, "learning_rate": 1.1213839296533537e-05, "loss": 0.0894, "step": 11500 }, { "epoch": 2.850356294536817, "grad_norm": 4.668305397033691, "learning_rate": 1.0626727291479424e-05, "loss": 0.0969, "step": 12000 }, { "epoch": 2.969121140142518, "grad_norm": 0.056590911000967026, "learning_rate": 1.0039615286425313e-05, "loss": 0.0932, "step": 12500 }, { "epoch": 3.0, "eval_accuracy": 0.8990825688073395, "eval_loss": 0.4282897412776947, "eval_runtime": 1.1944, "eval_samples_per_second": 730.051, "eval_steps_per_second": 46.047, "step": 12630 }, { "epoch": 3.0878859857482186, "grad_norm": 0.023897308856248856, "learning_rate": 9.452503281371202e-06, "loss": 0.0537, "step": 13000 }, { "epoch": 3.2066508313539193, "grad_norm": 0.7962842583656311, "learning_rate": 8.86539127631709e-06, "loss": 0.0473, "step": 13500 }, { "epoch": 3.32541567695962, "grad_norm": 0.47386619448661804, "learning_rate": 8.278279271262978e-06, "loss": 0.0608, "step": 14000 }, { "epoch": 3.4441805225653206, "grad_norm": 0.06599200516939163, "learning_rate": 7.691167266208865e-06, "loss": 0.0677, "step": 14500 }, { "epoch": 3.5629453681710213, "grad_norm": 0.18340125679969788, "learning_rate": 7.104055261154754e-06, "loss": 0.0579, "step": 15000 }, { "epoch": 3.6817102137767224, "grad_norm": 5.418150901794434, "learning_rate": 6.5169432561006415e-06, "loss": 0.0517, "step": 15500 }, { "epoch": 3.800475059382423, "grad_norm": 19.50031852722168, "learning_rate": 5.92983125104653e-06, "loss": 0.0617, "step": 16000 }, { "epoch": 3.9192399049881237, "grad_norm": 0.0843653604388237, "learning_rate": 5.3427192459924185e-06, "loss": 0.0599, "step": 16500 }, { "epoch": 4.0, "eval_accuracy": 0.9036697247706422, "eval_loss": 0.466791033744812, "eval_runtime": 1.2178, "eval_samples_per_second": 716.054, "eval_steps_per_second": 45.164, "step": 16840 }, { "epoch": 4.038004750593824, "grad_norm": 0.04484863579273224, "learning_rate": 4.7556072409383065e-06, "loss": 0.0424, "step": 17000 }, { "epoch": 4.156769596199525, "grad_norm": 0.0025116826873272657, "learning_rate": 4.1684952358841946e-06, "loss": 0.0335, "step": 17500 }, { "epoch": 4.275534441805226, "grad_norm": 0.4570842981338501, "learning_rate": 3.581383230830083e-06, "loss": 0.0395, "step": 18000 }, { "epoch": 4.394299287410926, "grad_norm": 0.7941423654556274, "learning_rate": 2.9942712257759707e-06, "loss": 0.0334, "step": 18500 }, { "epoch": 4.513064133016627, "grad_norm": 3.7694590091705322, "learning_rate": 2.4071592207218587e-06, "loss": 0.035, "step": 19000 }, { "epoch": 4.631828978622328, "grad_norm": 0.018865738064050674, "learning_rate": 1.8200472156677468e-06, "loss": 0.0373, "step": 19500 }, { "epoch": 4.750593824228028, "grad_norm": 0.03485831245779991, "learning_rate": 1.232935210613635e-06, "loss": 0.0282, "step": 20000 }, { "epoch": 4.869358669833729, "grad_norm": 0.014678980223834515, "learning_rate": 6.458232055595231e-07, "loss": 0.0369, "step": 20500 }, { "epoch": 4.98812351543943, "grad_norm": 0.06822077184915543, "learning_rate": 5.871120050541119e-08, "loss": 0.0299, "step": 21000 }, { "epoch": 5.0, "eval_accuracy": 0.9025229357798165, "eval_loss": 0.5252613425254822, "eval_runtime": 1.204, "eval_samples_per_second": 724.232, "eval_steps_per_second": 45.68, "step": 21050 } ], "logging_steps": 500, "max_steps": 21050, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "total_flos": 3061523906193576.0, "train_batch_size": 16, "trial_name": null, "trial_params": { "learning_rate": 2.471741541277811e-05, "num_train_epochs": 5, "per_device_train_batch_size": 16, "seed": 27 } }