|
{ |
|
"best_metric": 0.47253116965293884, |
|
"best_model_checkpoint": "Action_all_10_class/checkpoint-1500", |
|
"epoch": 10.0, |
|
"eval_steps": 100, |
|
"global_step": 2790, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.6752883195877075, |
|
"learning_rate": 9.96415770609319e-05, |
|
"loss": 2.247, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.8152681589126587, |
|
"learning_rate": 9.928315412186381e-05, |
|
"loss": 2.1602, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 2.164091110229492, |
|
"learning_rate": 9.892473118279571e-05, |
|
"loss": 2.0409, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 2.6417195796966553, |
|
"learning_rate": 9.85663082437276e-05, |
|
"loss": 1.8783, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 4.199918746948242, |
|
"learning_rate": 9.820788530465951e-05, |
|
"loss": 1.7363, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 2.9530537128448486, |
|
"learning_rate": 9.78494623655914e-05, |
|
"loss": 1.5894, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 2.777188301086426, |
|
"learning_rate": 9.74910394265233e-05, |
|
"loss": 1.4617, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 3.2515618801116943, |
|
"learning_rate": 9.713261648745519e-05, |
|
"loss": 1.4804, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 2.89544939994812, |
|
"learning_rate": 9.681003584229391e-05, |
|
"loss": 1.2876, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 2.372511625289917, |
|
"learning_rate": 9.645161290322581e-05, |
|
"loss": 1.2411, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_accuracy": 0.754601226993865, |
|
"eval_loss": 1.1517137289047241, |
|
"eval_runtime": 18.3534, |
|
"eval_samples_per_second": 53.287, |
|
"eval_steps_per_second": 6.702, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 3.142375946044922, |
|
"learning_rate": 9.609318996415772e-05, |
|
"loss": 1.2389, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 2.8880653381347656, |
|
"learning_rate": 9.573476702508962e-05, |
|
"loss": 1.179, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 3.925424575805664, |
|
"learning_rate": 9.541218637992831e-05, |
|
"loss": 1.1092, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 4.03208589553833, |
|
"learning_rate": 9.505376344086023e-05, |
|
"loss": 1.1389, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 2.35840106010437, |
|
"learning_rate": 9.469534050179211e-05, |
|
"loss": 1.0824, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 3.0924174785614014, |
|
"learning_rate": 9.433691756272402e-05, |
|
"loss": 0.957, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 4.47157096862793, |
|
"learning_rate": 9.397849462365592e-05, |
|
"loss": 1.0797, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 2.9026050567626953, |
|
"learning_rate": 9.362007168458782e-05, |
|
"loss": 1.0961, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 2.673151969909668, |
|
"learning_rate": 9.326164874551971e-05, |
|
"loss": 1.0421, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 1.5587886571884155, |
|
"learning_rate": 9.290322580645162e-05, |
|
"loss": 0.8932, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"eval_accuracy": 0.7975460122699386, |
|
"eval_loss": 0.785646915435791, |
|
"eval_runtime": 14.5537, |
|
"eval_samples_per_second": 67.199, |
|
"eval_steps_per_second": 8.451, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 3.5420029163360596, |
|
"learning_rate": 9.254480286738351e-05, |
|
"loss": 0.9145, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 3.8418989181518555, |
|
"learning_rate": 9.218637992831541e-05, |
|
"loss": 0.8473, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 3.9392898082733154, |
|
"learning_rate": 9.182795698924731e-05, |
|
"loss": 0.7867, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 5.856507778167725, |
|
"learning_rate": 9.146953405017922e-05, |
|
"loss": 0.9951, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 3.318254232406616, |
|
"learning_rate": 9.111111111111112e-05, |
|
"loss": 0.7852, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 2.2948904037475586, |
|
"learning_rate": 9.0752688172043e-05, |
|
"loss": 0.9565, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 4.516943454742432, |
|
"learning_rate": 9.039426523297492e-05, |
|
"loss": 0.7034, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.9273302555084229, |
|
"learning_rate": 9.003584229390681e-05, |
|
"loss": 0.8603, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 9.084336280822754, |
|
"learning_rate": 8.967741935483871e-05, |
|
"loss": 0.6914, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 4.446329116821289, |
|
"learning_rate": 8.931899641577061e-05, |
|
"loss": 0.6907, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"eval_accuracy": 0.8220858895705522, |
|
"eval_loss": 0.6636486649513245, |
|
"eval_runtime": 14.1733, |
|
"eval_samples_per_second": 69.003, |
|
"eval_steps_per_second": 8.678, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 5.061271667480469, |
|
"learning_rate": 8.896057347670252e-05, |
|
"loss": 0.7286, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 3.7218878269195557, |
|
"learning_rate": 8.86021505376344e-05, |
|
"loss": 0.7241, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 3.3088059425354004, |
|
"learning_rate": 8.824372759856632e-05, |
|
"loss": 0.7051, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 7.993207931518555, |
|
"learning_rate": 8.788530465949821e-05, |
|
"loss": 0.6652, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 3.7464661598205566, |
|
"learning_rate": 8.752688172043011e-05, |
|
"loss": 0.6898, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 4.751115798950195, |
|
"learning_rate": 8.716845878136201e-05, |
|
"loss": 0.8783, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 3.408888578414917, |
|
"learning_rate": 8.681003584229391e-05, |
|
"loss": 0.6584, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 4.679331302642822, |
|
"learning_rate": 8.645161290322581e-05, |
|
"loss": 0.7819, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 4.112971305847168, |
|
"learning_rate": 8.60931899641577e-05, |
|
"loss": 0.5751, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 3.214542865753174, |
|
"learning_rate": 8.573476702508962e-05, |
|
"loss": 0.5841, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"eval_accuracy": 0.8159509202453987, |
|
"eval_loss": 0.6388102173805237, |
|
"eval_runtime": 14.2283, |
|
"eval_samples_per_second": 68.736, |
|
"eval_steps_per_second": 8.645, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 8.963754653930664, |
|
"learning_rate": 8.53763440860215e-05, |
|
"loss": 0.7805, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 4.3538126945495605, |
|
"learning_rate": 8.501792114695341e-05, |
|
"loss": 0.6053, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 5.444653034210205, |
|
"learning_rate": 8.465949820788531e-05, |
|
"loss": 0.764, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 3.9525132179260254, |
|
"learning_rate": 8.430107526881721e-05, |
|
"loss": 0.684, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 3.4026594161987305, |
|
"learning_rate": 8.39426523297491e-05, |
|
"loss": 0.5835, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 3.843327522277832, |
|
"learning_rate": 8.358422939068102e-05, |
|
"loss": 0.5352, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 5.771996021270752, |
|
"learning_rate": 8.32258064516129e-05, |
|
"loss": 0.631, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 2.9874844551086426, |
|
"learning_rate": 8.28673835125448e-05, |
|
"loss": 0.611, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 2.693260431289673, |
|
"learning_rate": 8.250896057347671e-05, |
|
"loss": 0.6031, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 8.329994201660156, |
|
"learning_rate": 8.215053763440861e-05, |
|
"loss": 0.5425, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"eval_accuracy": 0.843558282208589, |
|
"eval_loss": 0.5870967507362366, |
|
"eval_runtime": 14.0184, |
|
"eval_samples_per_second": 69.765, |
|
"eval_steps_per_second": 8.774, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 3.7939178943634033, |
|
"learning_rate": 8.179211469534051e-05, |
|
"loss": 0.5645, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 4.04762077331543, |
|
"learning_rate": 8.14336917562724e-05, |
|
"loss": 0.5252, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 4.459852695465088, |
|
"learning_rate": 8.107526881720431e-05, |
|
"loss": 0.7369, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 7.254125118255615, |
|
"learning_rate": 8.07168458781362e-05, |
|
"loss": 0.6754, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 5.496582984924316, |
|
"learning_rate": 8.03584229390681e-05, |
|
"loss": 0.5504, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 7.471122741699219, |
|
"learning_rate": 8e-05, |
|
"loss": 0.643, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 4.464521408081055, |
|
"learning_rate": 7.964157706093191e-05, |
|
"loss": 0.575, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 1.7192500829696655, |
|
"learning_rate": 7.92831541218638e-05, |
|
"loss": 0.4709, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 6.228574275970459, |
|
"learning_rate": 7.892473118279571e-05, |
|
"loss": 0.5173, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 6.098475933074951, |
|
"learning_rate": 7.85663082437276e-05, |
|
"loss": 0.5929, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"eval_accuracy": 0.8210633946830266, |
|
"eval_loss": 0.5646098852157593, |
|
"eval_runtime": 14.2986, |
|
"eval_samples_per_second": 68.398, |
|
"eval_steps_per_second": 8.602, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 3.2586510181427, |
|
"learning_rate": 7.82078853046595e-05, |
|
"loss": 0.5392, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 6.211545944213867, |
|
"learning_rate": 7.784946236559139e-05, |
|
"loss": 0.5032, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 2.691046953201294, |
|
"learning_rate": 7.74910394265233e-05, |
|
"loss": 0.5918, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 3.3668406009674072, |
|
"learning_rate": 7.71326164874552e-05, |
|
"loss": 0.6054, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 3.9461565017700195, |
|
"learning_rate": 7.67741935483871e-05, |
|
"loss": 0.4755, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 4.586925506591797, |
|
"learning_rate": 7.6415770609319e-05, |
|
"loss": 0.4968, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 5.659130573272705, |
|
"learning_rate": 7.60573476702509e-05, |
|
"loss": 0.4894, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 4.458271503448486, |
|
"learning_rate": 7.56989247311828e-05, |
|
"loss": 0.4056, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 5.034984588623047, |
|
"learning_rate": 7.53405017921147e-05, |
|
"loss": 0.6026, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 4.018673419952393, |
|
"learning_rate": 7.49820788530466e-05, |
|
"loss": 0.4406, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"eval_accuracy": 0.8404907975460123, |
|
"eval_loss": 0.543875515460968, |
|
"eval_runtime": 14.0957, |
|
"eval_samples_per_second": 69.383, |
|
"eval_steps_per_second": 8.726, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 2.96732759475708, |
|
"learning_rate": 7.462365591397849e-05, |
|
"loss": 0.6851, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 3.6570851802825928, |
|
"learning_rate": 7.42652329749104e-05, |
|
"loss": 0.4738, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 3.4345834255218506, |
|
"learning_rate": 7.39068100358423e-05, |
|
"loss": 0.5739, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 5.19066858291626, |
|
"learning_rate": 7.35483870967742e-05, |
|
"loss": 0.4721, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 5.423111915588379, |
|
"learning_rate": 7.318996415770609e-05, |
|
"loss": 0.5271, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 5.806092262268066, |
|
"learning_rate": 7.2831541218638e-05, |
|
"loss": 0.5163, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 6.047975540161133, |
|
"learning_rate": 7.247311827956989e-05, |
|
"loss": 0.5061, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 7.956634044647217, |
|
"learning_rate": 7.211469534050179e-05, |
|
"loss": 0.5639, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 5.515818119049072, |
|
"learning_rate": 7.17562724014337e-05, |
|
"loss": 0.5214, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 4.654768943786621, |
|
"learning_rate": 7.13978494623656e-05, |
|
"loss": 0.4541, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"eval_accuracy": 0.8415132924335378, |
|
"eval_loss": 0.5317940711975098, |
|
"eval_runtime": 14.2499, |
|
"eval_samples_per_second": 68.632, |
|
"eval_steps_per_second": 8.632, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 4.857801914215088, |
|
"learning_rate": 7.10394265232975e-05, |
|
"loss": 0.415, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 4.023741722106934, |
|
"learning_rate": 7.06810035842294e-05, |
|
"loss": 0.3754, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 7.845776081085205, |
|
"learning_rate": 7.03225806451613e-05, |
|
"loss": 0.464, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"grad_norm": 10.843958854675293, |
|
"learning_rate": 6.996415770609319e-05, |
|
"loss": 0.655, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"grad_norm": 2.662417411804199, |
|
"learning_rate": 6.960573476702509e-05, |
|
"loss": 0.4086, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"grad_norm": 6.569747447967529, |
|
"learning_rate": 6.924731182795699e-05, |
|
"loss": 0.4273, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"grad_norm": 3.297159433364868, |
|
"learning_rate": 6.88888888888889e-05, |
|
"loss": 0.4426, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"grad_norm": 5.205211639404297, |
|
"learning_rate": 6.853046594982078e-05, |
|
"loss": 0.3445, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"grad_norm": 6.75899076461792, |
|
"learning_rate": 6.81720430107527e-05, |
|
"loss": 0.466, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"grad_norm": 4.749955654144287, |
|
"learning_rate": 6.781362007168459e-05, |
|
"loss": 0.3835, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"eval_accuracy": 0.8343558282208589, |
|
"eval_loss": 0.5224626660346985, |
|
"eval_runtime": 14.3694, |
|
"eval_samples_per_second": 68.062, |
|
"eval_steps_per_second": 8.56, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"grad_norm": 5.145068645477295, |
|
"learning_rate": 6.745519713261649e-05, |
|
"loss": 0.3779, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"grad_norm": 4.831909656524658, |
|
"learning_rate": 6.709677419354839e-05, |
|
"loss": 0.3744, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"grad_norm": 3.3703436851501465, |
|
"learning_rate": 6.673835125448029e-05, |
|
"loss": 0.3374, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"grad_norm": 10.194363594055176, |
|
"learning_rate": 6.63799283154122e-05, |
|
"loss": 0.4076, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"grad_norm": 1.0064858198165894, |
|
"learning_rate": 6.602150537634408e-05, |
|
"loss": 0.4434, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"grad_norm": 2.8392670154571533, |
|
"learning_rate": 6.5663082437276e-05, |
|
"loss": 0.318, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"grad_norm": 2.6734464168548584, |
|
"learning_rate": 6.530465949820789e-05, |
|
"loss": 0.5079, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"grad_norm": 2.2691242694854736, |
|
"learning_rate": 6.494623655913979e-05, |
|
"loss": 0.3849, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"grad_norm": 3.17016863822937, |
|
"learning_rate": 6.458781362007169e-05, |
|
"loss": 0.3427, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"grad_norm": 1.534572720527649, |
|
"learning_rate": 6.422939068100359e-05, |
|
"loss": 0.3924, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"eval_accuracy": 0.8302658486707567, |
|
"eval_loss": 0.5514972805976868, |
|
"eval_runtime": 14.2143, |
|
"eval_samples_per_second": 68.804, |
|
"eval_steps_per_second": 8.653, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"grad_norm": 1.171107292175293, |
|
"learning_rate": 6.387096774193548e-05, |
|
"loss": 0.3694, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"grad_norm": 4.934769630432129, |
|
"learning_rate": 6.35125448028674e-05, |
|
"loss": 0.3607, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"grad_norm": 7.391957759857178, |
|
"learning_rate": 6.315412186379928e-05, |
|
"loss": 0.3429, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"grad_norm": 7.1492156982421875, |
|
"learning_rate": 6.279569892473119e-05, |
|
"loss": 0.4211, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"grad_norm": 5.837622165679932, |
|
"learning_rate": 6.243727598566309e-05, |
|
"loss": 0.3857, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"grad_norm": 5.638124465942383, |
|
"learning_rate": 6.207885304659499e-05, |
|
"loss": 0.4379, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"grad_norm": 5.501093864440918, |
|
"learning_rate": 6.172043010752688e-05, |
|
"loss": 0.3903, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"grad_norm": 6.717010021209717, |
|
"learning_rate": 6.136200716845878e-05, |
|
"loss": 0.3844, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"grad_norm": 8.419461250305176, |
|
"learning_rate": 6.100358422939069e-05, |
|
"loss": 0.3705, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"grad_norm": 7.757752895355225, |
|
"learning_rate": 6.064516129032258e-05, |
|
"loss": 0.5741, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"eval_accuracy": 0.8251533742331288, |
|
"eval_loss": 0.5519386529922485, |
|
"eval_runtime": 13.907, |
|
"eval_samples_per_second": 70.324, |
|
"eval_steps_per_second": 8.844, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"grad_norm": 3.006796360015869, |
|
"learning_rate": 6.028673835125448e-05, |
|
"loss": 0.5208, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"grad_norm": 4.00032901763916, |
|
"learning_rate": 5.9928315412186386e-05, |
|
"loss": 0.3386, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"grad_norm": 1.4214438199996948, |
|
"learning_rate": 5.956989247311828e-05, |
|
"loss": 0.2971, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"grad_norm": 4.046026229858398, |
|
"learning_rate": 5.9211469534050176e-05, |
|
"loss": 0.4146, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"grad_norm": 2.4826581478118896, |
|
"learning_rate": 5.8853046594982085e-05, |
|
"loss": 0.4243, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"grad_norm": 4.898519515991211, |
|
"learning_rate": 5.849462365591398e-05, |
|
"loss": 0.3691, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"grad_norm": 7.121424198150635, |
|
"learning_rate": 5.813620071684588e-05, |
|
"loss": 0.4513, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"grad_norm": 5.212203025817871, |
|
"learning_rate": 5.7777777777777776e-05, |
|
"loss": 0.4, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"grad_norm": 11.669204711914062, |
|
"learning_rate": 5.7419354838709685e-05, |
|
"loss": 0.3429, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"grad_norm": 11.0433349609375, |
|
"learning_rate": 5.706093189964158e-05, |
|
"loss": 0.3991, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"eval_accuracy": 0.8445807770961146, |
|
"eval_loss": 0.49900951981544495, |
|
"eval_runtime": 13.9712, |
|
"eval_samples_per_second": 70.001, |
|
"eval_steps_per_second": 8.804, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"grad_norm": 3.952008008956909, |
|
"learning_rate": 5.6702508960573475e-05, |
|
"loss": 0.3259, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"grad_norm": 2.535245895385742, |
|
"learning_rate": 5.6344086021505384e-05, |
|
"loss": 0.2201, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"grad_norm": 5.427143096923828, |
|
"learning_rate": 5.598566308243728e-05, |
|
"loss": 0.3986, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"grad_norm": 6.882002353668213, |
|
"learning_rate": 5.5627240143369174e-05, |
|
"loss": 0.4718, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"grad_norm": 3.7098076343536377, |
|
"learning_rate": 5.526881720430108e-05, |
|
"loss": 0.3653, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"grad_norm": 3.8536739349365234, |
|
"learning_rate": 5.491039426523298e-05, |
|
"loss": 0.358, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"grad_norm": 6.399543762207031, |
|
"learning_rate": 5.455197132616487e-05, |
|
"loss": 0.4017, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"grad_norm": 5.091586112976074, |
|
"learning_rate": 5.419354838709678e-05, |
|
"loss": 0.493, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"grad_norm": 5.683386325836182, |
|
"learning_rate": 5.3835125448028676e-05, |
|
"loss": 0.5526, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"grad_norm": 3.3343377113342285, |
|
"learning_rate": 5.347670250896057e-05, |
|
"loss": 0.4732, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"eval_accuracy": 0.8302658486707567, |
|
"eval_loss": 0.5335860252380371, |
|
"eval_runtime": 14.2962, |
|
"eval_samples_per_second": 68.41, |
|
"eval_steps_per_second": 8.604, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"grad_norm": 5.0791521072387695, |
|
"learning_rate": 5.311827956989247e-05, |
|
"loss": 0.3433, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"grad_norm": 4.7593607902526855, |
|
"learning_rate": 5.2759856630824375e-05, |
|
"loss": 0.3617, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"grad_norm": 3.0220556259155273, |
|
"learning_rate": 5.2401433691756277e-05, |
|
"loss": 0.3243, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"grad_norm": 1.736549735069275, |
|
"learning_rate": 5.204301075268817e-05, |
|
"loss": 0.3488, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"grad_norm": 4.916406154632568, |
|
"learning_rate": 5.168458781362008e-05, |
|
"loss": 0.2955, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"grad_norm": 9.70033073425293, |
|
"learning_rate": 5.1326164874551975e-05, |
|
"loss": 0.392, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"grad_norm": 2.5204570293426514, |
|
"learning_rate": 5.096774193548387e-05, |
|
"loss": 0.2706, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"grad_norm": 4.673634052276611, |
|
"learning_rate": 5.060931899641578e-05, |
|
"loss": 0.3257, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"grad_norm": 5.239686012268066, |
|
"learning_rate": 5.0250896057347674e-05, |
|
"loss": 0.3738, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"grad_norm": 3.337782144546509, |
|
"learning_rate": 4.989247311827957e-05, |
|
"loss": 0.3324, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"eval_accuracy": 0.8282208588957055, |
|
"eval_loss": 0.5350744128227234, |
|
"eval_runtime": 14.1615, |
|
"eval_samples_per_second": 69.061, |
|
"eval_steps_per_second": 8.686, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"grad_norm": 2.9601023197174072, |
|
"learning_rate": 4.953405017921147e-05, |
|
"loss": 0.2966, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"grad_norm": 10.412677764892578, |
|
"learning_rate": 4.917562724014337e-05, |
|
"loss": 0.3838, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"grad_norm": 1.986757755279541, |
|
"learning_rate": 4.881720430107527e-05, |
|
"loss": 0.4096, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"grad_norm": 1.2032238245010376, |
|
"learning_rate": 4.845878136200717e-05, |
|
"loss": 0.3183, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"grad_norm": 1.7991982698440552, |
|
"learning_rate": 4.810035842293907e-05, |
|
"loss": 0.2713, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"grad_norm": 7.0565385818481445, |
|
"learning_rate": 4.774193548387097e-05, |
|
"loss": 0.2414, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"grad_norm": 3.8213694095611572, |
|
"learning_rate": 4.7383512544802875e-05, |
|
"loss": 0.363, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"grad_norm": 2.4153404235839844, |
|
"learning_rate": 4.702508960573477e-05, |
|
"loss": 0.349, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"grad_norm": 2.161435842514038, |
|
"learning_rate": 4.666666666666667e-05, |
|
"loss": 0.2626, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"grad_norm": 5.8719282150268555, |
|
"learning_rate": 4.630824372759857e-05, |
|
"loss": 0.3433, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"eval_accuracy": 0.8517382413087935, |
|
"eval_loss": 0.47253116965293884, |
|
"eval_runtime": 14.1174, |
|
"eval_samples_per_second": 69.276, |
|
"eval_steps_per_second": 8.713, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"grad_norm": 5.560650825500488, |
|
"learning_rate": 4.594982078853047e-05, |
|
"loss": 0.2775, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"grad_norm": 2.5716190338134766, |
|
"learning_rate": 4.559139784946237e-05, |
|
"loss": 0.2537, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"grad_norm": 5.602371692657471, |
|
"learning_rate": 4.5232974910394265e-05, |
|
"loss": 0.3127, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"grad_norm": 5.320743083953857, |
|
"learning_rate": 4.487455197132617e-05, |
|
"loss": 0.2878, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"grad_norm": 4.925574779510498, |
|
"learning_rate": 4.451612903225807e-05, |
|
"loss": 0.2842, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"grad_norm": 3.862708568572998, |
|
"learning_rate": 4.4157706093189964e-05, |
|
"loss": 0.3004, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"grad_norm": 3.0864365100860596, |
|
"learning_rate": 4.3799283154121866e-05, |
|
"loss": 0.2638, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"grad_norm": 3.5530340671539307, |
|
"learning_rate": 4.344086021505376e-05, |
|
"loss": 0.2439, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"grad_norm": 5.929757118225098, |
|
"learning_rate": 4.308243727598566e-05, |
|
"loss": 0.3235, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"grad_norm": 3.567291736602783, |
|
"learning_rate": 4.2724014336917564e-05, |
|
"loss": 0.2187, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"eval_accuracy": 0.8466257668711656, |
|
"eval_loss": 0.5041967034339905, |
|
"eval_runtime": 14.1345, |
|
"eval_samples_per_second": 69.193, |
|
"eval_steps_per_second": 8.702, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"grad_norm": 0.8105212450027466, |
|
"learning_rate": 4.2365591397849466e-05, |
|
"loss": 0.2792, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"grad_norm": 3.8311142921447754, |
|
"learning_rate": 4.200716845878137e-05, |
|
"loss": 0.2313, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"grad_norm": 6.618185043334961, |
|
"learning_rate": 4.164874551971326e-05, |
|
"loss": 0.2668, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"grad_norm": 7.174586772918701, |
|
"learning_rate": 4.1290322580645165e-05, |
|
"loss": 0.2684, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"grad_norm": 6.7463788986206055, |
|
"learning_rate": 4.093189964157707e-05, |
|
"loss": 0.2825, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 5.95, |
|
"grad_norm": 2.402876853942871, |
|
"learning_rate": 4.057347670250896e-05, |
|
"loss": 0.2986, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"grad_norm": 2.343830108642578, |
|
"learning_rate": 4.0215053763440864e-05, |
|
"loss": 0.3086, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"grad_norm": 3.3489632606506348, |
|
"learning_rate": 3.985663082437276e-05, |
|
"loss": 0.2844, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"grad_norm": 2.076068162918091, |
|
"learning_rate": 3.949820788530466e-05, |
|
"loss": 0.1609, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"grad_norm": 3.9822871685028076, |
|
"learning_rate": 3.913978494623656e-05, |
|
"loss": 0.2952, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"eval_accuracy": 0.8548057259713702, |
|
"eval_loss": 0.5240095257759094, |
|
"eval_runtime": 14.3375, |
|
"eval_samples_per_second": 68.213, |
|
"eval_steps_per_second": 8.579, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"grad_norm": 3.8955540657043457, |
|
"learning_rate": 3.878136200716846e-05, |
|
"loss": 0.3564, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"grad_norm": 2.374730348587036, |
|
"learning_rate": 3.842293906810036e-05, |
|
"loss": 0.2467, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"grad_norm": 6.376945495605469, |
|
"learning_rate": 3.8064516129032254e-05, |
|
"loss": 0.2584, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 6.24, |
|
"grad_norm": 5.645608901977539, |
|
"learning_rate": 3.7706093189964156e-05, |
|
"loss": 0.3091, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 6.27, |
|
"grad_norm": 2.882622241973877, |
|
"learning_rate": 3.734767025089606e-05, |
|
"loss": 0.2471, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"grad_norm": 8.7712984085083, |
|
"learning_rate": 3.698924731182796e-05, |
|
"loss": 0.2595, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"grad_norm": 2.5900111198425293, |
|
"learning_rate": 3.663082437275986e-05, |
|
"loss": 0.3474, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"grad_norm": 0.14616957306861877, |
|
"learning_rate": 3.627240143369176e-05, |
|
"loss": 0.2777, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"grad_norm": 5.120488166809082, |
|
"learning_rate": 3.591397849462366e-05, |
|
"loss": 0.2867, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"grad_norm": 4.957128524780273, |
|
"learning_rate": 3.555555555555556e-05, |
|
"loss": 0.2687, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"eval_accuracy": 0.83640081799591, |
|
"eval_loss": 0.5522516965866089, |
|
"eval_runtime": 14.281, |
|
"eval_samples_per_second": 68.482, |
|
"eval_steps_per_second": 8.613, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"grad_norm": 6.335272312164307, |
|
"learning_rate": 3.5197132616487455e-05, |
|
"loss": 0.2486, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"grad_norm": 3.3235371112823486, |
|
"learning_rate": 3.483870967741936e-05, |
|
"loss": 0.1797, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"grad_norm": 0.23497584462165833, |
|
"learning_rate": 3.448028673835126e-05, |
|
"loss": 0.2214, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 6.59, |
|
"grad_norm": 1.0797181129455566, |
|
"learning_rate": 3.4121863799283154e-05, |
|
"loss": 0.2369, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"grad_norm": 3.9156014919281006, |
|
"learning_rate": 3.3763440860215055e-05, |
|
"loss": 0.2826, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"grad_norm": 9.87369155883789, |
|
"learning_rate": 3.340501792114695e-05, |
|
"loss": 0.3399, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"grad_norm": 3.916306972503662, |
|
"learning_rate": 3.304659498207885e-05, |
|
"loss": 0.3758, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"grad_norm": 7.597679615020752, |
|
"learning_rate": 3.2688172043010754e-05, |
|
"loss": 0.2505, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"grad_norm": 4.102473735809326, |
|
"learning_rate": 3.2329749103942656e-05, |
|
"loss": 0.2692, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"grad_norm": 8.36500358581543, |
|
"learning_rate": 3.197132616487455e-05, |
|
"loss": 0.3111, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"eval_accuracy": 0.8496932515337423, |
|
"eval_loss": 0.53035569190979, |
|
"eval_runtime": 14.3586, |
|
"eval_samples_per_second": 68.112, |
|
"eval_steps_per_second": 8.566, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"grad_norm": 5.454440593719482, |
|
"learning_rate": 3.161290322580645e-05, |
|
"loss": 0.2373, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"grad_norm": 0.6554473638534546, |
|
"learning_rate": 3.1254480286738355e-05, |
|
"loss": 0.1937, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 6.92, |
|
"grad_norm": 0.24610257148742676, |
|
"learning_rate": 3.0896057347670256e-05, |
|
"loss": 0.2149, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 6.95, |
|
"grad_norm": 4.611593246459961, |
|
"learning_rate": 3.053763440860215e-05, |
|
"loss": 0.2791, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 6.99, |
|
"grad_norm": 8.88548755645752, |
|
"learning_rate": 3.0179211469534053e-05, |
|
"loss": 0.3238, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"grad_norm": 2.9050915241241455, |
|
"learning_rate": 2.9820788530465955e-05, |
|
"loss": 0.2527, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"grad_norm": 6.722965717315674, |
|
"learning_rate": 2.946236559139785e-05, |
|
"loss": 0.2451, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"grad_norm": 4.379875183105469, |
|
"learning_rate": 2.9103942652329752e-05, |
|
"loss": 0.269, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"grad_norm": 2.847017288208008, |
|
"learning_rate": 2.8745519713261647e-05, |
|
"loss": 0.2776, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 7.17, |
|
"grad_norm": 5.179224491119385, |
|
"learning_rate": 2.838709677419355e-05, |
|
"loss": 0.2431, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 7.17, |
|
"eval_accuracy": 0.8568507157464212, |
|
"eval_loss": 0.5104272365570068, |
|
"eval_runtime": 14.3416, |
|
"eval_samples_per_second": 68.193, |
|
"eval_steps_per_second": 8.576, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"grad_norm": 2.521162986755371, |
|
"learning_rate": 2.802867383512545e-05, |
|
"loss": 0.1631, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"grad_norm": 4.0884246826171875, |
|
"learning_rate": 2.767025089605735e-05, |
|
"loss": 0.2731, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 7.28, |
|
"grad_norm": 4.443497180938721, |
|
"learning_rate": 2.731182795698925e-05, |
|
"loss": 0.2594, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"grad_norm": 2.3641021251678467, |
|
"learning_rate": 2.6953405017921146e-05, |
|
"loss": 0.2477, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 7.35, |
|
"grad_norm": 0.09834872931241989, |
|
"learning_rate": 2.6594982078853048e-05, |
|
"loss": 0.2298, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 7.38, |
|
"grad_norm": 3.825798749923706, |
|
"learning_rate": 2.623655913978495e-05, |
|
"loss": 0.2955, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"grad_norm": 1.7917243242263794, |
|
"learning_rate": 2.5878136200716844e-05, |
|
"loss": 0.3034, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"grad_norm": 2.0665640830993652, |
|
"learning_rate": 2.5519713261648746e-05, |
|
"loss": 0.2204, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"grad_norm": 1.5993545055389404, |
|
"learning_rate": 2.5161290322580645e-05, |
|
"loss": 0.2114, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"grad_norm": 7.3012871742248535, |
|
"learning_rate": 2.4802867383512547e-05, |
|
"loss": 0.3265, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"eval_accuracy": 0.869120654396728, |
|
"eval_loss": 0.5085219740867615, |
|
"eval_runtime": 14.1317, |
|
"eval_samples_per_second": 69.206, |
|
"eval_steps_per_second": 8.704, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"grad_norm": 1.1089341640472412, |
|
"learning_rate": 2.4444444444444445e-05, |
|
"loss": 0.2762, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"grad_norm": 4.004220485687256, |
|
"learning_rate": 2.4086021505376347e-05, |
|
"loss": 0.2453, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"grad_norm": 7.738345623016357, |
|
"learning_rate": 2.3727598566308245e-05, |
|
"loss": 0.2757, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"grad_norm": 3.836663007736206, |
|
"learning_rate": 2.3405017921146955e-05, |
|
"loss": 0.298, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"grad_norm": 1.0338927507400513, |
|
"learning_rate": 2.3046594982078853e-05, |
|
"loss": 0.1816, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"grad_norm": 1.0525509119033813, |
|
"learning_rate": 2.268817204301075e-05, |
|
"loss": 0.145, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"grad_norm": 5.745477676391602, |
|
"learning_rate": 2.2329749103942653e-05, |
|
"loss": 0.2143, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"grad_norm": 4.428494930267334, |
|
"learning_rate": 2.1971326164874552e-05, |
|
"loss": 0.2301, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 7.85, |
|
"grad_norm": 5.054660320281982, |
|
"learning_rate": 2.1612903225806454e-05, |
|
"loss": 0.2317, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"grad_norm": 1.9087740182876587, |
|
"learning_rate": 2.1254480286738352e-05, |
|
"loss": 0.2595, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"eval_accuracy": 0.8568507157464212, |
|
"eval_loss": 0.5014929175376892, |
|
"eval_runtime": 14.0273, |
|
"eval_samples_per_second": 69.721, |
|
"eval_steps_per_second": 8.769, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"grad_norm": 3.2280988693237305, |
|
"learning_rate": 2.0896057347670254e-05, |
|
"loss": 0.1503, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 7.96, |
|
"grad_norm": 7.906033039093018, |
|
"learning_rate": 2.0537634408602152e-05, |
|
"loss": 0.2864, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"grad_norm": 8.353536605834961, |
|
"learning_rate": 2.017921146953405e-05, |
|
"loss": 0.2027, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"grad_norm": 6.651542663574219, |
|
"learning_rate": 1.982078853046595e-05, |
|
"loss": 0.2513, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"grad_norm": 4.423330307006836, |
|
"learning_rate": 1.9462365591397848e-05, |
|
"loss": 0.2591, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 8.1, |
|
"grad_norm": 7.127723217010498, |
|
"learning_rate": 1.910394265232975e-05, |
|
"loss": 0.2341, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"grad_norm": 1.9292550086975098, |
|
"learning_rate": 1.874551971326165e-05, |
|
"loss": 0.1744, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 8.17, |
|
"grad_norm": 3.043701648712158, |
|
"learning_rate": 1.838709677419355e-05, |
|
"loss": 0.3194, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"grad_norm": 5.514953136444092, |
|
"learning_rate": 1.8028673835125448e-05, |
|
"loss": 0.2454, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"grad_norm": 1.1831002235412598, |
|
"learning_rate": 1.767025089605735e-05, |
|
"loss": 0.1825, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"eval_accuracy": 0.8619631901840491, |
|
"eval_loss": 0.4919591546058655, |
|
"eval_runtime": 14.1344, |
|
"eval_samples_per_second": 69.193, |
|
"eval_steps_per_second": 8.702, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"grad_norm": 3.9401707649230957, |
|
"learning_rate": 1.7311827956989248e-05, |
|
"loss": 0.1456, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"grad_norm": 7.038700103759766, |
|
"learning_rate": 1.6989247311827958e-05, |
|
"loss": 0.1804, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"grad_norm": 9.696996688842773, |
|
"learning_rate": 1.6630824372759856e-05, |
|
"loss": 0.144, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"grad_norm": 1.7609977722167969, |
|
"learning_rate": 1.6272401433691755e-05, |
|
"loss": 0.1852, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 8.42, |
|
"grad_norm": 0.1569293588399887, |
|
"learning_rate": 1.5913978494623657e-05, |
|
"loss": 0.302, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"grad_norm": 6.627655506134033, |
|
"learning_rate": 1.5555555555555555e-05, |
|
"loss": 0.2418, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 8.49, |
|
"grad_norm": 4.680299282073975, |
|
"learning_rate": 1.5197132616487455e-05, |
|
"loss": 0.2042, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"grad_norm": 3.413140058517456, |
|
"learning_rate": 1.4838709677419355e-05, |
|
"loss": 0.2608, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"grad_norm": 0.6003228425979614, |
|
"learning_rate": 1.4480286738351257e-05, |
|
"loss": 0.2494, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"grad_norm": 9.677336692810059, |
|
"learning_rate": 1.4121863799283155e-05, |
|
"loss": 0.2602, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"eval_accuracy": 0.8619631901840491, |
|
"eval_loss": 0.5016282200813293, |
|
"eval_runtime": 14.4248, |
|
"eval_samples_per_second": 67.8, |
|
"eval_steps_per_second": 8.527, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"grad_norm": 3.607775926589966, |
|
"learning_rate": 1.3763440860215054e-05, |
|
"loss": 0.2171, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 8.67, |
|
"grad_norm": 0.4323180019855499, |
|
"learning_rate": 1.3405017921146954e-05, |
|
"loss": 0.1981, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 8.71, |
|
"grad_norm": 9.15799331665039, |
|
"learning_rate": 1.3046594982078852e-05, |
|
"loss": 0.2206, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"grad_norm": 3.574291229248047, |
|
"learning_rate": 1.2688172043010754e-05, |
|
"loss": 0.2388, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"grad_norm": 4.433730125427246, |
|
"learning_rate": 1.2329749103942653e-05, |
|
"loss": 0.1705, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"grad_norm": 9.609087944030762, |
|
"learning_rate": 1.1971326164874553e-05, |
|
"loss": 0.3498, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 8.85, |
|
"grad_norm": 5.335048675537109, |
|
"learning_rate": 1.1612903225806453e-05, |
|
"loss": 0.2864, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"grad_norm": 1.8026251792907715, |
|
"learning_rate": 1.1254480286738351e-05, |
|
"loss": 0.2135, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"grad_norm": 1.9891669750213623, |
|
"learning_rate": 1.0896057347670251e-05, |
|
"loss": 0.2214, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"grad_norm": 0.7803227305412292, |
|
"learning_rate": 1.0537634408602151e-05, |
|
"loss": 0.2628, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"eval_accuracy": 0.8680981595092024, |
|
"eval_loss": 0.47455406188964844, |
|
"eval_runtime": 14.3927, |
|
"eval_samples_per_second": 67.951, |
|
"eval_steps_per_second": 8.546, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"grad_norm": 0.4858655035495758, |
|
"learning_rate": 1.017921146953405e-05, |
|
"loss": 0.1861, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"grad_norm": 6.302251815795898, |
|
"learning_rate": 9.82078853046595e-06, |
|
"loss": 0.2009, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"grad_norm": 2.5692965984344482, |
|
"learning_rate": 9.46236559139785e-06, |
|
"loss": 0.1785, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"grad_norm": 6.564579010009766, |
|
"learning_rate": 9.10394265232975e-06, |
|
"loss": 0.1284, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"grad_norm": 5.155548572540283, |
|
"learning_rate": 8.745519713261649e-06, |
|
"loss": 0.2494, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"grad_norm": 4.653995990753174, |
|
"learning_rate": 8.387096774193549e-06, |
|
"loss": 0.1142, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"grad_norm": 5.491698741912842, |
|
"learning_rate": 8.028673835125449e-06, |
|
"loss": 0.1887, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"grad_norm": 1.841364860534668, |
|
"learning_rate": 7.670250896057347e-06, |
|
"loss": 0.2236, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"grad_norm": 4.787356376647949, |
|
"learning_rate": 7.3118279569892475e-06, |
|
"loss": 0.2155, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 9.32, |
|
"grad_norm": 0.1550242006778717, |
|
"learning_rate": 6.953405017921147e-06, |
|
"loss": 0.1024, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 9.32, |
|
"eval_accuracy": 0.869120654396728, |
|
"eval_loss": 0.48181313276290894, |
|
"eval_runtime": 14.3216, |
|
"eval_samples_per_second": 68.288, |
|
"eval_steps_per_second": 8.588, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 9.35, |
|
"grad_norm": 4.603553771972656, |
|
"learning_rate": 6.594982078853047e-06, |
|
"loss": 0.292, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 9.39, |
|
"grad_norm": 6.353521823883057, |
|
"learning_rate": 6.236559139784946e-06, |
|
"loss": 0.1462, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"grad_norm": 3.8513760566711426, |
|
"learning_rate": 5.878136200716846e-06, |
|
"loss": 0.228, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"grad_norm": 6.585672378540039, |
|
"learning_rate": 5.5197132616487455e-06, |
|
"loss": 0.1751, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"grad_norm": 0.19121739268302917, |
|
"learning_rate": 5.161290322580646e-06, |
|
"loss": 0.2495, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"grad_norm": 9.971940994262695, |
|
"learning_rate": 4.802867383512545e-06, |
|
"loss": 0.3073, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"grad_norm": 0.3317970037460327, |
|
"learning_rate": 4.444444444444445e-06, |
|
"loss": 0.1477, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"grad_norm": 4.959763526916504, |
|
"learning_rate": 4.086021505376344e-06, |
|
"loss": 0.2397, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"grad_norm": 1.1729767322540283, |
|
"learning_rate": 3.7275985663082444e-06, |
|
"loss": 0.2094, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"grad_norm": 5.1204118728637695, |
|
"learning_rate": 3.3691756272401432e-06, |
|
"loss": 0.1468, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"eval_accuracy": 0.8680981595092024, |
|
"eval_loss": 0.4764532148838043, |
|
"eval_runtime": 14.3404, |
|
"eval_samples_per_second": 68.199, |
|
"eval_steps_per_second": 8.577, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"grad_norm": 1.941149115562439, |
|
"learning_rate": 3.0107526881720433e-06, |
|
"loss": 0.1418, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 9.75, |
|
"grad_norm": 1.0805838108062744, |
|
"learning_rate": 2.6523297491039426e-06, |
|
"loss": 0.1762, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"grad_norm": 0.28982478380203247, |
|
"learning_rate": 2.2939068100358423e-06, |
|
"loss": 0.1462, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 9.82, |
|
"grad_norm": 0.11841225624084473, |
|
"learning_rate": 1.935483870967742e-06, |
|
"loss": 0.1495, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"grad_norm": 4.167394161224365, |
|
"learning_rate": 1.577060931899642e-06, |
|
"loss": 0.2596, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"grad_norm": 4.835020542144775, |
|
"learning_rate": 1.2186379928315414e-06, |
|
"loss": 0.1955, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"grad_norm": 7.901321887969971, |
|
"learning_rate": 8.602150537634409e-07, |
|
"loss": 0.2867, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"grad_norm": 8.663835525512695, |
|
"learning_rate": 5.017921146953406e-07, |
|
"loss": 0.1797, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 13.646035194396973, |
|
"learning_rate": 1.4336917562724014e-07, |
|
"loss": 0.2222, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 2790, |
|
"total_flos": 3.451740694569861e+18, |
|
"train_loss": 0.44341087110580935, |
|
"train_runtime": 1600.5784, |
|
"train_samples_per_second": 27.827, |
|
"train_steps_per_second": 1.743 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2790, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 100, |
|
"total_flos": 3.451740694569861e+18, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|