SLM_vs_LLM_experiments
/
LoRA
/Qwen
/Qwen1.5_7B_LoRA_coastalcph
/lex_glue_ledgar
/trainer_state.json
{ | |
"best_metric": 0.5025032162666321, | |
"best_model_checkpoint": "../experiments_checkpoints/LoRA/Qwen/Qwen1.5_7B_LoRA_coastalcph/lex_glue_ledgar/checkpoint-3700", | |
"epoch": 3.0, | |
"eval_steps": 100, | |
"global_step": 5625, | |
"is_hyper_param_search": false, | |
"is_local_process_zero": true, | |
"is_world_process_zero": true, | |
"log_history": [ | |
{ | |
"epoch": 0.01, | |
"grad_norm": 82.343017578125, | |
"learning_rate": 4.991111111111111e-05, | |
"loss": 11.1109, | |
"step": 10 | |
}, | |
{ | |
"epoch": 0.01, | |
"grad_norm": 83.22059631347656, | |
"learning_rate": 4.982222222222222e-05, | |
"loss": 8.525, | |
"step": 20 | |
}, | |
{ | |
"epoch": 0.02, | |
"grad_norm": 81.96794891357422, | |
"learning_rate": 4.973333333333334e-05, | |
"loss": 6.6453, | |
"step": 30 | |
}, | |
{ | |
"epoch": 0.02, | |
"grad_norm": 71.0719985961914, | |
"learning_rate": 4.964444444444445e-05, | |
"loss": 5.1352, | |
"step": 40 | |
}, | |
{ | |
"epoch": 0.03, | |
"grad_norm": 69.25579071044922, | |
"learning_rate": 4.955555555555556e-05, | |
"loss": 3.8117, | |
"step": 50 | |
}, | |
{ | |
"epoch": 0.03, | |
"grad_norm": 62.54920959472656, | |
"learning_rate": 4.9466666666666665e-05, | |
"loss": 2.7141, | |
"step": 60 | |
}, | |
{ | |
"epoch": 0.04, | |
"grad_norm": 51.00497055053711, | |
"learning_rate": 4.9377777777777776e-05, | |
"loss": 2.4369, | |
"step": 70 | |
}, | |
{ | |
"epoch": 0.04, | |
"grad_norm": 66.0743637084961, | |
"learning_rate": 4.928888888888889e-05, | |
"loss": 1.966, | |
"step": 80 | |
}, | |
{ | |
"epoch": 0.05, | |
"grad_norm": 55.161354064941406, | |
"learning_rate": 4.92e-05, | |
"loss": 1.8213, | |
"step": 90 | |
}, | |
{ | |
"epoch": 0.05, | |
"grad_norm": 50.841796875, | |
"learning_rate": 4.9111111111111114e-05, | |
"loss": 1.7995, | |
"step": 100 | |
}, | |
{ | |
"epoch": 0.05, | |
"eval_accuracy": 0.6512, | |
"eval_f1_macro": 0.4676003614917378, | |
"eval_f1_micro": 0.6512, | |
"eval_loss": 1.6894375085830688, | |
"eval_runtime": 126.3135, | |
"eval_samples_per_second": 79.168, | |
"eval_steps_per_second": 2.478, | |
"step": 100 | |
}, | |
{ | |
"epoch": 0.06, | |
"grad_norm": 44.830013275146484, | |
"learning_rate": 4.9022222222222224e-05, | |
"loss": 1.7258, | |
"step": 110 | |
}, | |
{ | |
"epoch": 0.06, | |
"grad_norm": 37.19465637207031, | |
"learning_rate": 4.8933333333333335e-05, | |
"loss": 1.4613, | |
"step": 120 | |
}, | |
{ | |
"epoch": 0.07, | |
"grad_norm": 50.72254943847656, | |
"learning_rate": 4.8844444444444445e-05, | |
"loss": 1.4131, | |
"step": 130 | |
}, | |
{ | |
"epoch": 0.07, | |
"grad_norm": 49.2353630065918, | |
"learning_rate": 4.875555555555556e-05, | |
"loss": 1.4369, | |
"step": 140 | |
}, | |
{ | |
"epoch": 0.08, | |
"grad_norm": 44.442604064941406, | |
"learning_rate": 4.866666666666667e-05, | |
"loss": 1.2461, | |
"step": 150 | |
}, | |
{ | |
"epoch": 0.09, | |
"grad_norm": 54.64115524291992, | |
"learning_rate": 4.8577777777777776e-05, | |
"loss": 1.2523, | |
"step": 160 | |
}, | |
{ | |
"epoch": 0.09, | |
"grad_norm": 46.350425720214844, | |
"learning_rate": 4.848888888888889e-05, | |
"loss": 1.2385, | |
"step": 170 | |
}, | |
{ | |
"epoch": 0.1, | |
"grad_norm": 45.80131912231445, | |
"learning_rate": 4.8400000000000004e-05, | |
"loss": 1.2934, | |
"step": 180 | |
}, | |
{ | |
"epoch": 0.1, | |
"grad_norm": 43.66839599609375, | |
"learning_rate": 4.8311111111111115e-05, | |
"loss": 1.1484, | |
"step": 190 | |
}, | |
{ | |
"epoch": 0.11, | |
"grad_norm": 45.98219680786133, | |
"learning_rate": 4.8222222222222225e-05, | |
"loss": 1.3922, | |
"step": 200 | |
}, | |
{ | |
"epoch": 0.11, | |
"eval_accuracy": 0.7076, | |
"eval_f1_macro": 0.5868340973418246, | |
"eval_f1_micro": 0.7076, | |
"eval_loss": 1.2208046913146973, | |
"eval_runtime": 125.6037, | |
"eval_samples_per_second": 79.616, | |
"eval_steps_per_second": 2.492, | |
"step": 200 | |
}, | |
{ | |
"epoch": 0.11, | |
"grad_norm": 53.35915756225586, | |
"learning_rate": 4.8133333333333336e-05, | |
"loss": 1.2534, | |
"step": 210 | |
}, | |
{ | |
"epoch": 0.12, | |
"grad_norm": 49.47136688232422, | |
"learning_rate": 4.8044444444444446e-05, | |
"loss": 1.2297, | |
"step": 220 | |
}, | |
{ | |
"epoch": 0.12, | |
"grad_norm": 41.21822738647461, | |
"learning_rate": 4.7955555555555556e-05, | |
"loss": 1.136, | |
"step": 230 | |
}, | |
{ | |
"epoch": 0.13, | |
"grad_norm": 49.57373809814453, | |
"learning_rate": 4.7866666666666674e-05, | |
"loss": 1.1995, | |
"step": 240 | |
}, | |
{ | |
"epoch": 0.13, | |
"grad_norm": 60.3203010559082, | |
"learning_rate": 4.7777777777777784e-05, | |
"loss": 1.1587, | |
"step": 250 | |
}, | |
{ | |
"epoch": 0.14, | |
"grad_norm": 44.72378158569336, | |
"learning_rate": 4.768888888888889e-05, | |
"loss": 1.1577, | |
"step": 260 | |
}, | |
{ | |
"epoch": 0.14, | |
"grad_norm": 45.23435974121094, | |
"learning_rate": 4.76e-05, | |
"loss": 1.0536, | |
"step": 270 | |
}, | |
{ | |
"epoch": 0.15, | |
"grad_norm": 34.0523567199707, | |
"learning_rate": 4.751111111111111e-05, | |
"loss": 1.0193, | |
"step": 280 | |
}, | |
{ | |
"epoch": 0.15, | |
"grad_norm": 43.04400634765625, | |
"learning_rate": 4.7422222222222226e-05, | |
"loss": 1.0108, | |
"step": 290 | |
}, | |
{ | |
"epoch": 0.16, | |
"grad_norm": 45.1166877746582, | |
"learning_rate": 4.7333333333333336e-05, | |
"loss": 1.0552, | |
"step": 300 | |
}, | |
{ | |
"epoch": 0.16, | |
"eval_accuracy": 0.7634, | |
"eval_f1_macro": 0.6328915988513745, | |
"eval_f1_micro": 0.7634, | |
"eval_loss": 0.9664570093154907, | |
"eval_runtime": 125.4379, | |
"eval_samples_per_second": 79.721, | |
"eval_steps_per_second": 2.495, | |
"step": 300 | |
}, | |
{ | |
"epoch": 0.17, | |
"grad_norm": 42.98120880126953, | |
"learning_rate": 4.724444444444445e-05, | |
"loss": 0.9302, | |
"step": 310 | |
}, | |
{ | |
"epoch": 0.17, | |
"grad_norm": 32.83863830566406, | |
"learning_rate": 4.715555555555556e-05, | |
"loss": 0.8612, | |
"step": 320 | |
}, | |
{ | |
"epoch": 0.18, | |
"grad_norm": 51.440391540527344, | |
"learning_rate": 4.706666666666667e-05, | |
"loss": 1.2096, | |
"step": 330 | |
}, | |
{ | |
"epoch": 0.18, | |
"grad_norm": 32.38331604003906, | |
"learning_rate": 4.6977777777777785e-05, | |
"loss": 0.939, | |
"step": 340 | |
}, | |
{ | |
"epoch": 0.19, | |
"grad_norm": 32.610382080078125, | |
"learning_rate": 4.6888888888888895e-05, | |
"loss": 0.9571, | |
"step": 350 | |
}, | |
{ | |
"epoch": 0.19, | |
"grad_norm": 49.74464797973633, | |
"learning_rate": 4.6800000000000006e-05, | |
"loss": 1.0897, | |
"step": 360 | |
}, | |
{ | |
"epoch": 0.2, | |
"grad_norm": 40.764678955078125, | |
"learning_rate": 4.671111111111111e-05, | |
"loss": 0.85, | |
"step": 370 | |
}, | |
{ | |
"epoch": 0.2, | |
"grad_norm": 55.797000885009766, | |
"learning_rate": 4.662222222222222e-05, | |
"loss": 0.9105, | |
"step": 380 | |
}, | |
{ | |
"epoch": 0.21, | |
"grad_norm": 41.21658706665039, | |
"learning_rate": 4.653333333333334e-05, | |
"loss": 0.8054, | |
"step": 390 | |
}, | |
{ | |
"epoch": 0.21, | |
"grad_norm": 39.946632385253906, | |
"learning_rate": 4.644444444444445e-05, | |
"loss": 0.8416, | |
"step": 400 | |
}, | |
{ | |
"epoch": 0.21, | |
"eval_accuracy": 0.767, | |
"eval_f1_macro": 0.6280418494347138, | |
"eval_f1_micro": 0.767, | |
"eval_loss": 0.9614554643630981, | |
"eval_runtime": 125.3428, | |
"eval_samples_per_second": 79.781, | |
"eval_steps_per_second": 2.497, | |
"step": 400 | |
}, | |
{ | |
"epoch": 0.22, | |
"grad_norm": 40.86934280395508, | |
"learning_rate": 4.635555555555556e-05, | |
"loss": 1.0155, | |
"step": 410 | |
}, | |
{ | |
"epoch": 0.22, | |
"grad_norm": 39.81444549560547, | |
"learning_rate": 4.626666666666667e-05, | |
"loss": 0.9998, | |
"step": 420 | |
}, | |
{ | |
"epoch": 0.23, | |
"grad_norm": 40.45600509643555, | |
"learning_rate": 4.617777777777778e-05, | |
"loss": 0.7995, | |
"step": 430 | |
}, | |
{ | |
"epoch": 0.23, | |
"grad_norm": 47.23322677612305, | |
"learning_rate": 4.608888888888889e-05, | |
"loss": 0.6886, | |
"step": 440 | |
}, | |
{ | |
"epoch": 0.24, | |
"grad_norm": 45.51860046386719, | |
"learning_rate": 4.600000000000001e-05, | |
"loss": 0.7401, | |
"step": 450 | |
}, | |
{ | |
"epoch": 0.25, | |
"grad_norm": 31.879140853881836, | |
"learning_rate": 4.591111111111112e-05, | |
"loss": 0.9815, | |
"step": 460 | |
}, | |
{ | |
"epoch": 0.25, | |
"grad_norm": 38.960289001464844, | |
"learning_rate": 4.582222222222222e-05, | |
"loss": 1.0263, | |
"step": 470 | |
}, | |
{ | |
"epoch": 0.26, | |
"grad_norm": 26.54925537109375, | |
"learning_rate": 4.573333333333333e-05, | |
"loss": 0.8351, | |
"step": 480 | |
}, | |
{ | |
"epoch": 0.26, | |
"grad_norm": 42.07994842529297, | |
"learning_rate": 4.564444444444444e-05, | |
"loss": 0.8951, | |
"step": 490 | |
}, | |
{ | |
"epoch": 0.27, | |
"grad_norm": 43.244110107421875, | |
"learning_rate": 4.555555555555556e-05, | |
"loss": 0.8204, | |
"step": 500 | |
}, | |
{ | |
"epoch": 0.27, | |
"eval_accuracy": 0.7892, | |
"eval_f1_macro": 0.6680353341510004, | |
"eval_f1_micro": 0.7892, | |
"eval_loss": 0.846860945224762, | |
"eval_runtime": 125.1499, | |
"eval_samples_per_second": 79.904, | |
"eval_steps_per_second": 2.501, | |
"step": 500 | |
}, | |
{ | |
"epoch": 0.27, | |
"grad_norm": 40.45793151855469, | |
"learning_rate": 4.546666666666667e-05, | |
"loss": 0.8416, | |
"step": 510 | |
}, | |
{ | |
"epoch": 0.28, | |
"grad_norm": 48.35597610473633, | |
"learning_rate": 4.537777777777778e-05, | |
"loss": 0.8184, | |
"step": 520 | |
}, | |
{ | |
"epoch": 0.28, | |
"grad_norm": 38.001399993896484, | |
"learning_rate": 4.528888888888889e-05, | |
"loss": 0.8784, | |
"step": 530 | |
}, | |
{ | |
"epoch": 0.29, | |
"grad_norm": 35.90673065185547, | |
"learning_rate": 4.52e-05, | |
"loss": 0.8764, | |
"step": 540 | |
}, | |
{ | |
"epoch": 0.29, | |
"grad_norm": 42.99267578125, | |
"learning_rate": 4.511111111111112e-05, | |
"loss": 0.7643, | |
"step": 550 | |
}, | |
{ | |
"epoch": 0.3, | |
"grad_norm": 35.354740142822266, | |
"learning_rate": 4.502222222222223e-05, | |
"loss": 0.7455, | |
"step": 560 | |
}, | |
{ | |
"epoch": 0.3, | |
"grad_norm": 46.95038986206055, | |
"learning_rate": 4.493333333333333e-05, | |
"loss": 1.0796, | |
"step": 570 | |
}, | |
{ | |
"epoch": 0.31, | |
"grad_norm": 36.5641975402832, | |
"learning_rate": 4.484444444444444e-05, | |
"loss": 0.7303, | |
"step": 580 | |
}, | |
{ | |
"epoch": 0.31, | |
"grad_norm": 31.755895614624023, | |
"learning_rate": 4.475555555555555e-05, | |
"loss": 0.6932, | |
"step": 590 | |
}, | |
{ | |
"epoch": 0.32, | |
"grad_norm": 41.90700149536133, | |
"learning_rate": 4.466666666666667e-05, | |
"loss": 0.7359, | |
"step": 600 | |
}, | |
{ | |
"epoch": 0.32, | |
"eval_accuracy": 0.8025, | |
"eval_f1_macro": 0.6858714672474995, | |
"eval_f1_micro": 0.8025, | |
"eval_loss": 0.7820109128952026, | |
"eval_runtime": 125.0637, | |
"eval_samples_per_second": 79.959, | |
"eval_steps_per_second": 2.503, | |
"step": 600 | |
}, | |
{ | |
"epoch": 0.33, | |
"grad_norm": 40.491641998291016, | |
"learning_rate": 4.457777777777778e-05, | |
"loss": 0.7398, | |
"step": 610 | |
}, | |
{ | |
"epoch": 0.33, | |
"grad_norm": 40.14454650878906, | |
"learning_rate": 4.448888888888889e-05, | |
"loss": 0.7974, | |
"step": 620 | |
}, | |
{ | |
"epoch": 0.34, | |
"grad_norm": 39.546302795410156, | |
"learning_rate": 4.44e-05, | |
"loss": 0.6925, | |
"step": 630 | |
}, | |
{ | |
"epoch": 0.34, | |
"grad_norm": 27.2608699798584, | |
"learning_rate": 4.431111111111111e-05, | |
"loss": 0.8556, | |
"step": 640 | |
}, | |
{ | |
"epoch": 0.35, | |
"grad_norm": 28.21966552734375, | |
"learning_rate": 4.422222222222222e-05, | |
"loss": 0.798, | |
"step": 650 | |
}, | |
{ | |
"epoch": 0.35, | |
"grad_norm": 26.93427276611328, | |
"learning_rate": 4.413333333333334e-05, | |
"loss": 0.8146, | |
"step": 660 | |
}, | |
{ | |
"epoch": 0.36, | |
"grad_norm": 38.73881912231445, | |
"learning_rate": 4.404444444444445e-05, | |
"loss": 0.7848, | |
"step": 670 | |
}, | |
{ | |
"epoch": 0.36, | |
"grad_norm": 29.951839447021484, | |
"learning_rate": 4.3955555555555554e-05, | |
"loss": 0.6993, | |
"step": 680 | |
}, | |
{ | |
"epoch": 0.37, | |
"grad_norm": 39.4987678527832, | |
"learning_rate": 4.3866666666666665e-05, | |
"loss": 0.9409, | |
"step": 690 | |
}, | |
{ | |
"epoch": 0.37, | |
"grad_norm": 44.70013427734375, | |
"learning_rate": 4.377777777777778e-05, | |
"loss": 0.7088, | |
"step": 700 | |
}, | |
{ | |
"epoch": 0.37, | |
"eval_accuracy": 0.7975, | |
"eval_f1_macro": 0.6808312841147881, | |
"eval_f1_micro": 0.7975, | |
"eval_loss": 0.7904818654060364, | |
"eval_runtime": 124.9315, | |
"eval_samples_per_second": 80.044, | |
"eval_steps_per_second": 2.505, | |
"step": 700 | |
}, | |
{ | |
"epoch": 0.38, | |
"grad_norm": 23.76036262512207, | |
"learning_rate": 4.368888888888889e-05, | |
"loss": 0.5632, | |
"step": 710 | |
}, | |
{ | |
"epoch": 0.38, | |
"grad_norm": 37.98194122314453, | |
"learning_rate": 4.36e-05, | |
"loss": 0.7349, | |
"step": 720 | |
}, | |
{ | |
"epoch": 0.39, | |
"grad_norm": 43.77289581298828, | |
"learning_rate": 4.351111111111111e-05, | |
"loss": 0.7132, | |
"step": 730 | |
}, | |
{ | |
"epoch": 0.39, | |
"grad_norm": 38.53476333618164, | |
"learning_rate": 4.3422222222222224e-05, | |
"loss": 0.6374, | |
"step": 740 | |
}, | |
{ | |
"epoch": 0.4, | |
"grad_norm": 36.89138412475586, | |
"learning_rate": 4.3333333333333334e-05, | |
"loss": 0.7341, | |
"step": 750 | |
}, | |
{ | |
"epoch": 0.41, | |
"grad_norm": 39.2047233581543, | |
"learning_rate": 4.324444444444445e-05, | |
"loss": 0.8566, | |
"step": 760 | |
}, | |
{ | |
"epoch": 0.41, | |
"grad_norm": 37.043670654296875, | |
"learning_rate": 4.315555555555556e-05, | |
"loss": 0.7392, | |
"step": 770 | |
}, | |
{ | |
"epoch": 0.42, | |
"grad_norm": 35.77628707885742, | |
"learning_rate": 4.3066666666666665e-05, | |
"loss": 0.7401, | |
"step": 780 | |
}, | |
{ | |
"epoch": 0.42, | |
"grad_norm": 32.939632415771484, | |
"learning_rate": 4.2977777777777776e-05, | |
"loss": 0.7112, | |
"step": 790 | |
}, | |
{ | |
"epoch": 0.43, | |
"grad_norm": 14.420007705688477, | |
"learning_rate": 4.2888888888888886e-05, | |
"loss": 0.6096, | |
"step": 800 | |
}, | |
{ | |
"epoch": 0.43, | |
"eval_accuracy": 0.8009, | |
"eval_f1_macro": 0.6822606117129679, | |
"eval_f1_micro": 0.8009, | |
"eval_loss": 0.7862304449081421, | |
"eval_runtime": 124.8581, | |
"eval_samples_per_second": 80.091, | |
"eval_steps_per_second": 2.507, | |
"step": 800 | |
}, | |
{ | |
"epoch": 0.43, | |
"grad_norm": 42.69435119628906, | |
"learning_rate": 4.2800000000000004e-05, | |
"loss": 0.6954, | |
"step": 810 | |
}, | |
{ | |
"epoch": 0.44, | |
"grad_norm": 45.64569091796875, | |
"learning_rate": 4.2711111111111114e-05, | |
"loss": 0.8801, | |
"step": 820 | |
}, | |
{ | |
"epoch": 0.44, | |
"grad_norm": 27.12781524658203, | |
"learning_rate": 4.2622222222222224e-05, | |
"loss": 0.8576, | |
"step": 830 | |
}, | |
{ | |
"epoch": 0.45, | |
"grad_norm": 48.820552825927734, | |
"learning_rate": 4.2533333333333335e-05, | |
"loss": 0.6565, | |
"step": 840 | |
}, | |
{ | |
"epoch": 0.45, | |
"grad_norm": 43.27635192871094, | |
"learning_rate": 4.2444444444444445e-05, | |
"loss": 0.7962, | |
"step": 850 | |
}, | |
{ | |
"epoch": 0.46, | |
"grad_norm": 39.85664367675781, | |
"learning_rate": 4.235555555555556e-05, | |
"loss": 0.9294, | |
"step": 860 | |
}, | |
{ | |
"epoch": 0.46, | |
"grad_norm": 45.196571350097656, | |
"learning_rate": 4.226666666666667e-05, | |
"loss": 0.9026, | |
"step": 870 | |
}, | |
{ | |
"epoch": 0.47, | |
"grad_norm": 29.43784523010254, | |
"learning_rate": 4.217777777777778e-05, | |
"loss": 0.7298, | |
"step": 880 | |
}, | |
{ | |
"epoch": 0.47, | |
"grad_norm": 36.50236129760742, | |
"learning_rate": 4.208888888888889e-05, | |
"loss": 0.787, | |
"step": 890 | |
}, | |
{ | |
"epoch": 0.48, | |
"grad_norm": 47.03664016723633, | |
"learning_rate": 4.2e-05, | |
"loss": 0.8682, | |
"step": 900 | |
}, | |
{ | |
"epoch": 0.48, | |
"eval_accuracy": 0.7987, | |
"eval_f1_macro": 0.6966502426266987, | |
"eval_f1_micro": 0.7987, | |
"eval_loss": 0.7767773270606995, | |
"eval_runtime": 125.0393, | |
"eval_samples_per_second": 79.975, | |
"eval_steps_per_second": 2.503, | |
"step": 900 | |
}, | |
{ | |
"epoch": 0.49, | |
"grad_norm": 39.72060012817383, | |
"learning_rate": 4.1911111111111115e-05, | |
"loss": 0.8498, | |
"step": 910 | |
}, | |
{ | |
"epoch": 0.49, | |
"grad_norm": 37.37692642211914, | |
"learning_rate": 4.1822222222222225e-05, | |
"loss": 0.718, | |
"step": 920 | |
}, | |
{ | |
"epoch": 0.5, | |
"grad_norm": 28.36815071105957, | |
"learning_rate": 4.1733333333333336e-05, | |
"loss": 0.6923, | |
"step": 930 | |
}, | |
{ | |
"epoch": 0.5, | |
"grad_norm": 41.33854675292969, | |
"learning_rate": 4.1644444444444446e-05, | |
"loss": 0.7234, | |
"step": 940 | |
}, | |
{ | |
"epoch": 0.51, | |
"grad_norm": 37.55774688720703, | |
"learning_rate": 4.155555555555556e-05, | |
"loss": 0.8102, | |
"step": 950 | |
}, | |
{ | |
"epoch": 0.51, | |
"grad_norm": 32.01072692871094, | |
"learning_rate": 4.146666666666667e-05, | |
"loss": 0.7338, | |
"step": 960 | |
}, | |
{ | |
"epoch": 0.52, | |
"grad_norm": 23.995277404785156, | |
"learning_rate": 4.1377777777777784e-05, | |
"loss": 0.7768, | |
"step": 970 | |
}, | |
{ | |
"epoch": 0.52, | |
"grad_norm": 37.772430419921875, | |
"learning_rate": 4.1288888888888895e-05, | |
"loss": 0.746, | |
"step": 980 | |
}, | |
{ | |
"epoch": 0.53, | |
"grad_norm": 45.859737396240234, | |
"learning_rate": 4.12e-05, | |
"loss": 0.7789, | |
"step": 990 | |
}, | |
{ | |
"epoch": 0.53, | |
"grad_norm": 41.380367279052734, | |
"learning_rate": 4.111111111111111e-05, | |
"loss": 0.6772, | |
"step": 1000 | |
}, | |
{ | |
"epoch": 0.53, | |
"eval_accuracy": 0.8094, | |
"eval_f1_macro": 0.6933975962115336, | |
"eval_f1_micro": 0.8094, | |
"eval_loss": 0.7299655079841614, | |
"eval_runtime": 125.167, | |
"eval_samples_per_second": 79.893, | |
"eval_steps_per_second": 2.501, | |
"step": 1000 | |
}, | |
{ | |
"epoch": 0.54, | |
"grad_norm": 29.1951904296875, | |
"learning_rate": 4.1022222222222226e-05, | |
"loss": 0.6658, | |
"step": 1010 | |
}, | |
{ | |
"epoch": 0.54, | |
"grad_norm": 32.311279296875, | |
"learning_rate": 4.093333333333334e-05, | |
"loss": 0.7022, | |
"step": 1020 | |
}, | |
{ | |
"epoch": 0.55, | |
"grad_norm": 30.058263778686523, | |
"learning_rate": 4.084444444444445e-05, | |
"loss": 0.8125, | |
"step": 1030 | |
}, | |
{ | |
"epoch": 0.55, | |
"grad_norm": 60.061588287353516, | |
"learning_rate": 4.075555555555556e-05, | |
"loss": 0.7732, | |
"step": 1040 | |
}, | |
{ | |
"epoch": 0.56, | |
"grad_norm": 40.17058181762695, | |
"learning_rate": 4.066666666666667e-05, | |
"loss": 0.6829, | |
"step": 1050 | |
}, | |
{ | |
"epoch": 0.57, | |
"grad_norm": 30.274438858032227, | |
"learning_rate": 4.057777777777778e-05, | |
"loss": 0.6306, | |
"step": 1060 | |
}, | |
{ | |
"epoch": 0.57, | |
"grad_norm": 32.10597229003906, | |
"learning_rate": 4.0488888888888896e-05, | |
"loss": 0.723, | |
"step": 1070 | |
}, | |
{ | |
"epoch": 0.58, | |
"grad_norm": 24.114013671875, | |
"learning_rate": 4.0400000000000006e-05, | |
"loss": 0.5363, | |
"step": 1080 | |
}, | |
{ | |
"epoch": 0.58, | |
"grad_norm": 34.77803421020508, | |
"learning_rate": 4.031111111111111e-05, | |
"loss": 0.6772, | |
"step": 1090 | |
}, | |
{ | |
"epoch": 0.59, | |
"grad_norm": 18.767980575561523, | |
"learning_rate": 4.022222222222222e-05, | |
"loss": 0.6224, | |
"step": 1100 | |
}, | |
{ | |
"epoch": 0.59, | |
"eval_accuracy": 0.8146, | |
"eval_f1_macro": 0.7190013669213452, | |
"eval_f1_micro": 0.8146, | |
"eval_loss": 0.675977349281311, | |
"eval_runtime": 125.1012, | |
"eval_samples_per_second": 79.935, | |
"eval_steps_per_second": 2.502, | |
"step": 1100 | |
}, | |
{ | |
"epoch": 0.59, | |
"grad_norm": 29.401121139526367, | |
"learning_rate": 4.013333333333333e-05, | |
"loss": 0.5357, | |
"step": 1110 | |
}, | |
{ | |
"epoch": 0.6, | |
"grad_norm": 34.86585235595703, | |
"learning_rate": 4.004444444444445e-05, | |
"loss": 0.7042, | |
"step": 1120 | |
}, | |
{ | |
"epoch": 0.6, | |
"grad_norm": 45.15092849731445, | |
"learning_rate": 3.995555555555556e-05, | |
"loss": 0.5536, | |
"step": 1130 | |
}, | |
{ | |
"epoch": 0.61, | |
"grad_norm": 49.75371551513672, | |
"learning_rate": 3.986666666666667e-05, | |
"loss": 0.7, | |
"step": 1140 | |
}, | |
{ | |
"epoch": 0.61, | |
"grad_norm": 35.728458404541016, | |
"learning_rate": 3.977777777777778e-05, | |
"loss": 0.6518, | |
"step": 1150 | |
}, | |
{ | |
"epoch": 0.62, | |
"grad_norm": 35.763145446777344, | |
"learning_rate": 3.968888888888889e-05, | |
"loss": 0.5926, | |
"step": 1160 | |
}, | |
{ | |
"epoch": 0.62, | |
"grad_norm": 34.241641998291016, | |
"learning_rate": 3.960000000000001e-05, | |
"loss": 0.7443, | |
"step": 1170 | |
}, | |
{ | |
"epoch": 0.63, | |
"grad_norm": 35.822608947753906, | |
"learning_rate": 3.951111111111112e-05, | |
"loss": 0.7653, | |
"step": 1180 | |
}, | |
{ | |
"epoch": 0.63, | |
"grad_norm": 37.08557891845703, | |
"learning_rate": 3.942222222222222e-05, | |
"loss": 0.6447, | |
"step": 1190 | |
}, | |
{ | |
"epoch": 0.64, | |
"grad_norm": 40.616458892822266, | |
"learning_rate": 3.933333333333333e-05, | |
"loss": 0.5875, | |
"step": 1200 | |
}, | |
{ | |
"epoch": 0.64, | |
"eval_accuracy": 0.8253, | |
"eval_f1_macro": 0.7442346170146863, | |
"eval_f1_micro": 0.8253, | |
"eval_loss": 0.6448789238929749, | |
"eval_runtime": 125.1421, | |
"eval_samples_per_second": 79.909, | |
"eval_steps_per_second": 2.501, | |
"step": 1200 | |
}, | |
{ | |
"epoch": 0.65, | |
"grad_norm": 43.901790618896484, | |
"learning_rate": 3.924444444444444e-05, | |
"loss": 0.6233, | |
"step": 1210 | |
}, | |
{ | |
"epoch": 0.65, | |
"grad_norm": 28.500553131103516, | |
"learning_rate": 3.915555555555556e-05, | |
"loss": 0.7282, | |
"step": 1220 | |
}, | |
{ | |
"epoch": 0.66, | |
"grad_norm": 41.72186279296875, | |
"learning_rate": 3.906666666666667e-05, | |
"loss": 0.6711, | |
"step": 1230 | |
}, | |
{ | |
"epoch": 0.66, | |
"grad_norm": 46.095245361328125, | |
"learning_rate": 3.897777777777778e-05, | |
"loss": 0.5736, | |
"step": 1240 | |
}, | |
{ | |
"epoch": 0.67, | |
"grad_norm": 41.30305862426758, | |
"learning_rate": 3.888888888888889e-05, | |
"loss": 0.6972, | |
"step": 1250 | |
}, | |
{ | |
"epoch": 0.67, | |
"grad_norm": 33.491661071777344, | |
"learning_rate": 3.88e-05, | |
"loss": 0.7147, | |
"step": 1260 | |
}, | |
{ | |
"epoch": 0.68, | |
"grad_norm": 40.78701400756836, | |
"learning_rate": 3.871111111111111e-05, | |
"loss": 0.5984, | |
"step": 1270 | |
}, | |
{ | |
"epoch": 0.68, | |
"grad_norm": 36.48966598510742, | |
"learning_rate": 3.862222222222223e-05, | |
"loss": 0.7407, | |
"step": 1280 | |
}, | |
{ | |
"epoch": 0.69, | |
"grad_norm": 39.34754180908203, | |
"learning_rate": 3.853333333333334e-05, | |
"loss": 0.5544, | |
"step": 1290 | |
}, | |
{ | |
"epoch": 0.69, | |
"grad_norm": 28.790693283081055, | |
"learning_rate": 3.844444444444444e-05, | |
"loss": 0.6147, | |
"step": 1300 | |
}, | |
{ | |
"epoch": 0.69, | |
"eval_accuracy": 0.8305, | |
"eval_f1_macro": 0.7207971849621393, | |
"eval_f1_micro": 0.8305, | |
"eval_loss": 0.6603367328643799, | |
"eval_runtime": 125.1561, | |
"eval_samples_per_second": 79.9, | |
"eval_steps_per_second": 2.501, | |
"step": 1300 | |
}, | |
{ | |
"epoch": 0.7, | |
"grad_norm": 23.91292381286621, | |
"learning_rate": 3.8355555555555553e-05, | |
"loss": 0.5826, | |
"step": 1310 | |
}, | |
{ | |
"epoch": 0.7, | |
"grad_norm": 39.718162536621094, | |
"learning_rate": 3.8266666666666664e-05, | |
"loss": 0.7496, | |
"step": 1320 | |
}, | |
{ | |
"epoch": 0.71, | |
"grad_norm": 28.492454528808594, | |
"learning_rate": 3.817777777777778e-05, | |
"loss": 0.611, | |
"step": 1330 | |
}, | |
{ | |
"epoch": 0.71, | |
"grad_norm": 34.915157318115234, | |
"learning_rate": 3.808888888888889e-05, | |
"loss": 0.7782, | |
"step": 1340 | |
}, | |
{ | |
"epoch": 0.72, | |
"grad_norm": 29.332368850708008, | |
"learning_rate": 3.8e-05, | |
"loss": 0.5906, | |
"step": 1350 | |
}, | |
{ | |
"epoch": 0.73, | |
"grad_norm": 25.36094856262207, | |
"learning_rate": 3.791111111111111e-05, | |
"loss": 0.5216, | |
"step": 1360 | |
}, | |
{ | |
"epoch": 0.73, | |
"grad_norm": 46.5117301940918, | |
"learning_rate": 3.782222222222222e-05, | |
"loss": 0.5596, | |
"step": 1370 | |
}, | |
{ | |
"epoch": 0.74, | |
"grad_norm": 35.714317321777344, | |
"learning_rate": 3.773333333333334e-05, | |
"loss": 0.6629, | |
"step": 1380 | |
}, | |
{ | |
"epoch": 0.74, | |
"grad_norm": 23.61018180847168, | |
"learning_rate": 3.764444444444445e-05, | |
"loss": 0.6112, | |
"step": 1390 | |
}, | |
{ | |
"epoch": 0.75, | |
"grad_norm": 26.34315299987793, | |
"learning_rate": 3.7555555555555554e-05, | |
"loss": 0.6355, | |
"step": 1400 | |
}, | |
{ | |
"epoch": 0.75, | |
"eval_accuracy": 0.8285, | |
"eval_f1_macro": 0.7294061203799771, | |
"eval_f1_micro": 0.8285, | |
"eval_loss": 0.6256009936332703, | |
"eval_runtime": 125.091, | |
"eval_samples_per_second": 79.942, | |
"eval_steps_per_second": 2.502, | |
"step": 1400 | |
}, | |
{ | |
"epoch": 0.75, | |
"grad_norm": 30.554851531982422, | |
"learning_rate": 3.7466666666666665e-05, | |
"loss": 0.6454, | |
"step": 1410 | |
}, | |
{ | |
"epoch": 0.76, | |
"grad_norm": 34.99703598022461, | |
"learning_rate": 3.7377777777777775e-05, | |
"loss": 0.5143, | |
"step": 1420 | |
}, | |
{ | |
"epoch": 0.76, | |
"grad_norm": 32.562931060791016, | |
"learning_rate": 3.728888888888889e-05, | |
"loss": 0.5926, | |
"step": 1430 | |
}, | |
{ | |
"epoch": 0.77, | |
"grad_norm": 35.89056396484375, | |
"learning_rate": 3.72e-05, | |
"loss": 0.5866, | |
"step": 1440 | |
}, | |
{ | |
"epoch": 0.77, | |
"grad_norm": 46.53374099731445, | |
"learning_rate": 3.7111111111111113e-05, | |
"loss": 0.6171, | |
"step": 1450 | |
}, | |
{ | |
"epoch": 0.78, | |
"grad_norm": 45.57334518432617, | |
"learning_rate": 3.7022222222222224e-05, | |
"loss": 0.6697, | |
"step": 1460 | |
}, | |
{ | |
"epoch": 0.78, | |
"grad_norm": 38.89950180053711, | |
"learning_rate": 3.6933333333333334e-05, | |
"loss": 0.7297, | |
"step": 1470 | |
}, | |
{ | |
"epoch": 0.79, | |
"grad_norm": 29.517240524291992, | |
"learning_rate": 3.6844444444444445e-05, | |
"loss": 0.6718, | |
"step": 1480 | |
}, | |
{ | |
"epoch": 0.79, | |
"grad_norm": 39.73881912231445, | |
"learning_rate": 3.675555555555556e-05, | |
"loss": 0.7189, | |
"step": 1490 | |
}, | |
{ | |
"epoch": 0.8, | |
"grad_norm": 32.796417236328125, | |
"learning_rate": 3.6666666666666666e-05, | |
"loss": 0.7076, | |
"step": 1500 | |
}, | |
{ | |
"epoch": 0.8, | |
"eval_accuracy": 0.8288, | |
"eval_f1_macro": 0.7290119185786764, | |
"eval_f1_micro": 0.8288, | |
"eval_loss": 0.6339758038520813, | |
"eval_runtime": 125.2044, | |
"eval_samples_per_second": 79.869, | |
"eval_steps_per_second": 2.5, | |
"step": 1500 | |
}, | |
{ | |
"epoch": 0.81, | |
"grad_norm": 46.705352783203125, | |
"learning_rate": 3.6577777777777776e-05, | |
"loss": 0.6376, | |
"step": 1510 | |
}, | |
{ | |
"epoch": 0.81, | |
"grad_norm": 37.39145278930664, | |
"learning_rate": 3.648888888888889e-05, | |
"loss": 0.6244, | |
"step": 1520 | |
}, | |
{ | |
"epoch": 0.82, | |
"grad_norm": 18.704368591308594, | |
"learning_rate": 3.6400000000000004e-05, | |
"loss": 0.5616, | |
"step": 1530 | |
}, | |
{ | |
"epoch": 0.82, | |
"grad_norm": 27.825958251953125, | |
"learning_rate": 3.6311111111111114e-05, | |
"loss": 0.5144, | |
"step": 1540 | |
}, | |
{ | |
"epoch": 0.83, | |
"grad_norm": 26.275487899780273, | |
"learning_rate": 3.6222222222222225e-05, | |
"loss": 0.703, | |
"step": 1550 | |
}, | |
{ | |
"epoch": 0.83, | |
"grad_norm": 39.88068771362305, | |
"learning_rate": 3.6133333333333335e-05, | |
"loss": 0.7873, | |
"step": 1560 | |
}, | |
{ | |
"epoch": 0.84, | |
"grad_norm": 37.45166015625, | |
"learning_rate": 3.6044444444444446e-05, | |
"loss": 0.6695, | |
"step": 1570 | |
}, | |
{ | |
"epoch": 0.84, | |
"grad_norm": 30.150066375732422, | |
"learning_rate": 3.5955555555555556e-05, | |
"loss": 0.6321, | |
"step": 1580 | |
}, | |
{ | |
"epoch": 0.85, | |
"grad_norm": 28.757957458496094, | |
"learning_rate": 3.586666666666667e-05, | |
"loss": 0.7031, | |
"step": 1590 | |
}, | |
{ | |
"epoch": 0.85, | |
"grad_norm": 18.669872283935547, | |
"learning_rate": 3.577777777777778e-05, | |
"loss": 0.4995, | |
"step": 1600 | |
}, | |
{ | |
"epoch": 0.85, | |
"eval_accuracy": 0.8315, | |
"eval_f1_macro": 0.7422265545358996, | |
"eval_f1_micro": 0.8315, | |
"eval_loss": 0.6185604333877563, | |
"eval_runtime": 125.1107, | |
"eval_samples_per_second": 79.929, | |
"eval_steps_per_second": 2.502, | |
"step": 1600 | |
}, | |
{ | |
"epoch": 0.86, | |
"grad_norm": 24.61341667175293, | |
"learning_rate": 3.568888888888889e-05, | |
"loss": 0.5862, | |
"step": 1610 | |
}, | |
{ | |
"epoch": 0.86, | |
"grad_norm": 39.579833984375, | |
"learning_rate": 3.56e-05, | |
"loss": 0.5583, | |
"step": 1620 | |
}, | |
{ | |
"epoch": 0.87, | |
"grad_norm": 35.16648864746094, | |
"learning_rate": 3.551111111111111e-05, | |
"loss": 0.6049, | |
"step": 1630 | |
}, | |
{ | |
"epoch": 0.87, | |
"grad_norm": 27.00229263305664, | |
"learning_rate": 3.5422222222222226e-05, | |
"loss": 0.7742, | |
"step": 1640 | |
}, | |
{ | |
"epoch": 0.88, | |
"grad_norm": 41.69742202758789, | |
"learning_rate": 3.5333333333333336e-05, | |
"loss": 0.754, | |
"step": 1650 | |
}, | |
{ | |
"epoch": 0.89, | |
"grad_norm": 39.063114166259766, | |
"learning_rate": 3.5244444444444447e-05, | |
"loss": 0.5125, | |
"step": 1660 | |
}, | |
{ | |
"epoch": 0.89, | |
"grad_norm": 36.81621170043945, | |
"learning_rate": 3.515555555555556e-05, | |
"loss": 0.5039, | |
"step": 1670 | |
}, | |
{ | |
"epoch": 0.9, | |
"grad_norm": 43.19419860839844, | |
"learning_rate": 3.506666666666667e-05, | |
"loss": 0.6167, | |
"step": 1680 | |
}, | |
{ | |
"epoch": 0.9, | |
"grad_norm": 36.92513656616211, | |
"learning_rate": 3.4977777777777785e-05, | |
"loss": 0.5754, | |
"step": 1690 | |
}, | |
{ | |
"epoch": 0.91, | |
"grad_norm": 35.30787658691406, | |
"learning_rate": 3.4888888888888895e-05, | |
"loss": 0.5754, | |
"step": 1700 | |
}, | |
{ | |
"epoch": 0.91, | |
"eval_accuracy": 0.8402, | |
"eval_f1_macro": 0.748230160429884, | |
"eval_f1_micro": 0.8402, | |
"eval_loss": 0.610479474067688, | |
"eval_runtime": 125.3127, | |
"eval_samples_per_second": 79.8, | |
"eval_steps_per_second": 2.498, | |
"step": 1700 | |
}, | |
{ | |
"epoch": 0.91, | |
"grad_norm": 29.194902420043945, | |
"learning_rate": 3.48e-05, | |
"loss": 0.6592, | |
"step": 1710 | |
}, | |
{ | |
"epoch": 0.92, | |
"grad_norm": 45.93604278564453, | |
"learning_rate": 3.471111111111111e-05, | |
"loss": 0.5685, | |
"step": 1720 | |
}, | |
{ | |
"epoch": 0.92, | |
"grad_norm": 29.713573455810547, | |
"learning_rate": 3.462222222222222e-05, | |
"loss": 0.6759, | |
"step": 1730 | |
}, | |
{ | |
"epoch": 0.93, | |
"grad_norm": 30.30419921875, | |
"learning_rate": 3.453333333333334e-05, | |
"loss": 0.5521, | |
"step": 1740 | |
}, | |
{ | |
"epoch": 0.93, | |
"grad_norm": 32.098819732666016, | |
"learning_rate": 3.444444444444445e-05, | |
"loss": 0.5322, | |
"step": 1750 | |
}, | |
{ | |
"epoch": 0.94, | |
"grad_norm": 36.17616271972656, | |
"learning_rate": 3.435555555555556e-05, | |
"loss": 0.5412, | |
"step": 1760 | |
}, | |
{ | |
"epoch": 0.94, | |
"grad_norm": 31.130359649658203, | |
"learning_rate": 3.426666666666667e-05, | |
"loss": 0.4915, | |
"step": 1770 | |
}, | |
{ | |
"epoch": 0.95, | |
"grad_norm": 29.99827003479004, | |
"learning_rate": 3.417777777777778e-05, | |
"loss": 0.5745, | |
"step": 1780 | |
}, | |
{ | |
"epoch": 0.95, | |
"grad_norm": 34.081581115722656, | |
"learning_rate": 3.408888888888889e-05, | |
"loss": 0.4867, | |
"step": 1790 | |
}, | |
{ | |
"epoch": 0.96, | |
"grad_norm": 27.07353401184082, | |
"learning_rate": 3.4000000000000007e-05, | |
"loss": 0.6775, | |
"step": 1800 | |
}, | |
{ | |
"epoch": 0.96, | |
"eval_accuracy": 0.8369, | |
"eval_f1_macro": 0.7531105549201089, | |
"eval_f1_micro": 0.8369, | |
"eval_loss": 0.5946730375289917, | |
"eval_runtime": 125.1987, | |
"eval_samples_per_second": 79.873, | |
"eval_steps_per_second": 2.5, | |
"step": 1800 | |
}, | |
{ | |
"epoch": 0.97, | |
"grad_norm": 30.427448272705078, | |
"learning_rate": 3.391111111111111e-05, | |
"loss": 0.6234, | |
"step": 1810 | |
}, | |
{ | |
"epoch": 0.97, | |
"grad_norm": 24.822948455810547, | |
"learning_rate": 3.382222222222222e-05, | |
"loss": 0.761, | |
"step": 1820 | |
}, | |
{ | |
"epoch": 0.98, | |
"grad_norm": 20.0987606048584, | |
"learning_rate": 3.373333333333333e-05, | |
"loss": 0.5981, | |
"step": 1830 | |
}, | |
{ | |
"epoch": 0.98, | |
"grad_norm": 29.777557373046875, | |
"learning_rate": 3.364444444444445e-05, | |
"loss": 0.5722, | |
"step": 1840 | |
}, | |
{ | |
"epoch": 0.99, | |
"grad_norm": 36.3715705871582, | |
"learning_rate": 3.355555555555556e-05, | |
"loss": 0.5465, | |
"step": 1850 | |
}, | |
{ | |
"epoch": 0.99, | |
"grad_norm": 36.25810623168945, | |
"learning_rate": 3.346666666666667e-05, | |
"loss": 0.5046, | |
"step": 1860 | |
}, | |
{ | |
"epoch": 1.0, | |
"grad_norm": 40.49335479736328, | |
"learning_rate": 3.337777777777778e-05, | |
"loss": 0.6467, | |
"step": 1870 | |
}, | |
{ | |
"epoch": 1.0, | |
"grad_norm": 24.9757022857666, | |
"learning_rate": 3.328888888888889e-05, | |
"loss": 0.5147, | |
"step": 1880 | |
}, | |
{ | |
"epoch": 1.01, | |
"grad_norm": 25.19569969177246, | |
"learning_rate": 3.32e-05, | |
"loss": 0.4116, | |
"step": 1890 | |
}, | |
{ | |
"epoch": 1.01, | |
"grad_norm": 18.321523666381836, | |
"learning_rate": 3.311111111111112e-05, | |
"loss": 0.3267, | |
"step": 1900 | |
}, | |
{ | |
"epoch": 1.01, | |
"eval_accuracy": 0.8528, | |
"eval_f1_macro": 0.7703652931864461, | |
"eval_f1_micro": 0.8528, | |
"eval_loss": 0.5678077936172485, | |
"eval_runtime": 125.1344, | |
"eval_samples_per_second": 79.914, | |
"eval_steps_per_second": 2.501, | |
"step": 1900 | |
}, | |
{ | |
"epoch": 1.02, | |
"grad_norm": 19.32207489013672, | |
"learning_rate": 3.302222222222222e-05, | |
"loss": 0.3264, | |
"step": 1910 | |
}, | |
{ | |
"epoch": 1.02, | |
"grad_norm": 27.11937141418457, | |
"learning_rate": 3.293333333333333e-05, | |
"loss": 0.3292, | |
"step": 1920 | |
}, | |
{ | |
"epoch": 1.03, | |
"grad_norm": 12.001670837402344, | |
"learning_rate": 3.284444444444444e-05, | |
"loss": 0.2931, | |
"step": 1930 | |
}, | |
{ | |
"epoch": 1.03, | |
"grad_norm": 43.393165588378906, | |
"learning_rate": 3.275555555555555e-05, | |
"loss": 0.3901, | |
"step": 1940 | |
}, | |
{ | |
"epoch": 1.04, | |
"grad_norm": 29.158109664916992, | |
"learning_rate": 3.266666666666667e-05, | |
"loss": 0.3262, | |
"step": 1950 | |
}, | |
{ | |
"epoch": 1.05, | |
"grad_norm": 26.838531494140625, | |
"learning_rate": 3.257777777777778e-05, | |
"loss": 0.337, | |
"step": 1960 | |
}, | |
{ | |
"epoch": 1.05, | |
"grad_norm": 32.79772186279297, | |
"learning_rate": 3.248888888888889e-05, | |
"loss": 0.312, | |
"step": 1970 | |
}, | |
{ | |
"epoch": 1.06, | |
"grad_norm": 18.537200927734375, | |
"learning_rate": 3.24e-05, | |
"loss": 0.2918, | |
"step": 1980 | |
}, | |
{ | |
"epoch": 1.06, | |
"grad_norm": 21.449663162231445, | |
"learning_rate": 3.231111111111111e-05, | |
"loss": 0.2633, | |
"step": 1990 | |
}, | |
{ | |
"epoch": 1.07, | |
"grad_norm": 21.72381591796875, | |
"learning_rate": 3.222222222222223e-05, | |
"loss": 0.2022, | |
"step": 2000 | |
}, | |
{ | |
"epoch": 1.07, | |
"eval_accuracy": 0.844, | |
"eval_f1_macro": 0.7638807977574182, | |
"eval_f1_micro": 0.844, | |
"eval_loss": 0.6360877752304077, | |
"eval_runtime": 125.2286, | |
"eval_samples_per_second": 79.854, | |
"eval_steps_per_second": 2.499, | |
"step": 2000 | |
}, | |
{ | |
"epoch": 1.07, | |
"grad_norm": 27.054370880126953, | |
"learning_rate": 3.213333333333334e-05, | |
"loss": 0.4437, | |
"step": 2010 | |
}, | |
{ | |
"epoch": 1.08, | |
"grad_norm": 22.364294052124023, | |
"learning_rate": 3.204444444444444e-05, | |
"loss": 0.2647, | |
"step": 2020 | |
}, | |
{ | |
"epoch": 1.08, | |
"grad_norm": 25.677640914916992, | |
"learning_rate": 3.1955555555555554e-05, | |
"loss": 0.3455, | |
"step": 2030 | |
}, | |
{ | |
"epoch": 1.09, | |
"grad_norm": 12.445222854614258, | |
"learning_rate": 3.1866666666666664e-05, | |
"loss": 0.2633, | |
"step": 2040 | |
}, | |
{ | |
"epoch": 1.09, | |
"grad_norm": 27.752634048461914, | |
"learning_rate": 3.177777777777778e-05, | |
"loss": 0.3091, | |
"step": 2050 | |
}, | |
{ | |
"epoch": 1.1, | |
"grad_norm": 34.70381546020508, | |
"learning_rate": 3.168888888888889e-05, | |
"loss": 0.3414, | |
"step": 2060 | |
}, | |
{ | |
"epoch": 1.1, | |
"grad_norm": 33.72677993774414, | |
"learning_rate": 3.16e-05, | |
"loss": 0.3628, | |
"step": 2070 | |
}, | |
{ | |
"epoch": 1.11, | |
"grad_norm": 19.364152908325195, | |
"learning_rate": 3.151111111111111e-05, | |
"loss": 0.3072, | |
"step": 2080 | |
}, | |
{ | |
"epoch": 1.11, | |
"grad_norm": 30.036209106445312, | |
"learning_rate": 3.142222222222222e-05, | |
"loss": 0.3389, | |
"step": 2090 | |
}, | |
{ | |
"epoch": 1.12, | |
"grad_norm": 19.264528274536133, | |
"learning_rate": 3.1333333333333334e-05, | |
"loss": 0.3831, | |
"step": 2100 | |
}, | |
{ | |
"epoch": 1.12, | |
"eval_accuracy": 0.8503, | |
"eval_f1_macro": 0.767154964941313, | |
"eval_f1_micro": 0.8503, | |
"eval_loss": 0.5956984162330627, | |
"eval_runtime": 125.0998, | |
"eval_samples_per_second": 79.936, | |
"eval_steps_per_second": 2.502, | |
"step": 2100 | |
}, | |
{ | |
"epoch": 1.13, | |
"grad_norm": 19.570907592773438, | |
"learning_rate": 3.124444444444445e-05, | |
"loss": 0.3448, | |
"step": 2110 | |
}, | |
{ | |
"epoch": 1.13, | |
"grad_norm": 30.030786514282227, | |
"learning_rate": 3.1155555555555555e-05, | |
"loss": 0.3049, | |
"step": 2120 | |
}, | |
{ | |
"epoch": 1.14, | |
"grad_norm": 25.751008987426758, | |
"learning_rate": 3.1066666666666665e-05, | |
"loss": 0.2604, | |
"step": 2130 | |
}, | |
{ | |
"epoch": 1.14, | |
"grad_norm": 29.36272430419922, | |
"learning_rate": 3.0977777777777776e-05, | |
"loss": 0.3661, | |
"step": 2140 | |
}, | |
{ | |
"epoch": 1.15, | |
"grad_norm": 33.18568801879883, | |
"learning_rate": 3.088888888888889e-05, | |
"loss": 0.2897, | |
"step": 2150 | |
}, | |
{ | |
"epoch": 1.15, | |
"grad_norm": 41.40974426269531, | |
"learning_rate": 3.08e-05, | |
"loss": 0.3758, | |
"step": 2160 | |
}, | |
{ | |
"epoch": 1.16, | |
"grad_norm": 21.188396453857422, | |
"learning_rate": 3.0711111111111114e-05, | |
"loss": 0.3537, | |
"step": 2170 | |
}, | |
{ | |
"epoch": 1.16, | |
"grad_norm": 20.362560272216797, | |
"learning_rate": 3.0622222222222224e-05, | |
"loss": 0.313, | |
"step": 2180 | |
}, | |
{ | |
"epoch": 1.17, | |
"grad_norm": 27.672611236572266, | |
"learning_rate": 3.0533333333333335e-05, | |
"loss": 0.3813, | |
"step": 2190 | |
}, | |
{ | |
"epoch": 1.17, | |
"grad_norm": 15.358328819274902, | |
"learning_rate": 3.044444444444445e-05, | |
"loss": 0.3235, | |
"step": 2200 | |
}, | |
{ | |
"epoch": 1.17, | |
"eval_accuracy": 0.8476, | |
"eval_f1_macro": 0.7685252040288097, | |
"eval_f1_micro": 0.8476, | |
"eval_loss": 0.6061830520629883, | |
"eval_runtime": 125.1594, | |
"eval_samples_per_second": 79.898, | |
"eval_steps_per_second": 2.501, | |
"step": 2200 | |
}, | |
{ | |
"epoch": 1.18, | |
"grad_norm": 20.829286575317383, | |
"learning_rate": 3.035555555555556e-05, | |
"loss": 0.4018, | |
"step": 2210 | |
}, | |
{ | |
"epoch": 1.18, | |
"grad_norm": 32.589256286621094, | |
"learning_rate": 3.0266666666666666e-05, | |
"loss": 0.4166, | |
"step": 2220 | |
}, | |
{ | |
"epoch": 1.19, | |
"grad_norm": 14.95271110534668, | |
"learning_rate": 3.0177777777777776e-05, | |
"loss": 0.3814, | |
"step": 2230 | |
}, | |
{ | |
"epoch": 1.19, | |
"grad_norm": 32.90339279174805, | |
"learning_rate": 3.008888888888889e-05, | |
"loss": 0.318, | |
"step": 2240 | |
}, | |
{ | |
"epoch": 1.2, | |
"grad_norm": 29.13666343688965, | |
"learning_rate": 3e-05, | |
"loss": 0.3152, | |
"step": 2250 | |
}, | |
{ | |
"epoch": 1.21, | |
"grad_norm": 20.47429847717285, | |
"learning_rate": 2.991111111111111e-05, | |
"loss": 0.4088, | |
"step": 2260 | |
}, | |
{ | |
"epoch": 1.21, | |
"grad_norm": 32.37903594970703, | |
"learning_rate": 2.9822222222222225e-05, | |
"loss": 0.3615, | |
"step": 2270 | |
}, | |
{ | |
"epoch": 1.22, | |
"grad_norm": 29.304582595825195, | |
"learning_rate": 2.9733333333333336e-05, | |
"loss": 0.2779, | |
"step": 2280 | |
}, | |
{ | |
"epoch": 1.22, | |
"grad_norm": 15.094827651977539, | |
"learning_rate": 2.9644444444444446e-05, | |
"loss": 0.2391, | |
"step": 2290 | |
}, | |
{ | |
"epoch": 1.23, | |
"grad_norm": 15.804471969604492, | |
"learning_rate": 2.955555555555556e-05, | |
"loss": 0.2279, | |
"step": 2300 | |
}, | |
{ | |
"epoch": 1.23, | |
"eval_accuracy": 0.847, | |
"eval_f1_macro": 0.7658476370782488, | |
"eval_f1_micro": 0.847, | |
"eval_loss": 0.6254826188087463, | |
"eval_runtime": 125.1512, | |
"eval_samples_per_second": 79.903, | |
"eval_steps_per_second": 2.501, | |
"step": 2300 | |
}, | |
{ | |
"epoch": 1.23, | |
"grad_norm": 12.60870361328125, | |
"learning_rate": 2.946666666666667e-05, | |
"loss": 0.4015, | |
"step": 2310 | |
}, | |
{ | |
"epoch": 1.24, | |
"grad_norm": 27.877355575561523, | |
"learning_rate": 2.937777777777778e-05, | |
"loss": 0.3093, | |
"step": 2320 | |
}, | |
{ | |
"epoch": 1.24, | |
"grad_norm": 15.59749698638916, | |
"learning_rate": 2.9288888888888888e-05, | |
"loss": 0.2629, | |
"step": 2330 | |
}, | |
{ | |
"epoch": 1.25, | |
"grad_norm": 28.317983627319336, | |
"learning_rate": 2.9199999999999998e-05, | |
"loss": 0.2197, | |
"step": 2340 | |
}, | |
{ | |
"epoch": 1.25, | |
"grad_norm": 27.348453521728516, | |
"learning_rate": 2.9111111111111112e-05, | |
"loss": 0.3047, | |
"step": 2350 | |
}, | |
{ | |
"epoch": 1.26, | |
"grad_norm": 17.648712158203125, | |
"learning_rate": 2.9022222222222223e-05, | |
"loss": 0.3159, | |
"step": 2360 | |
}, | |
{ | |
"epoch": 1.26, | |
"grad_norm": 20.40242576599121, | |
"learning_rate": 2.8933333333333333e-05, | |
"loss": 0.3453, | |
"step": 2370 | |
}, | |
{ | |
"epoch": 1.27, | |
"grad_norm": 24.179431915283203, | |
"learning_rate": 2.8844444444444447e-05, | |
"loss": 0.3431, | |
"step": 2380 | |
}, | |
{ | |
"epoch": 1.27, | |
"grad_norm": 32.673763275146484, | |
"learning_rate": 2.8755555555555557e-05, | |
"loss": 0.3027, | |
"step": 2390 | |
}, | |
{ | |
"epoch": 1.28, | |
"grad_norm": 28.0662841796875, | |
"learning_rate": 2.8666666666666668e-05, | |
"loss": 0.3224, | |
"step": 2400 | |
}, | |
{ | |
"epoch": 1.28, | |
"eval_accuracy": 0.8537, | |
"eval_f1_macro": 0.7772010038894511, | |
"eval_f1_micro": 0.8537, | |
"eval_loss": 0.5753965973854065, | |
"eval_runtime": 125.2268, | |
"eval_samples_per_second": 79.855, | |
"eval_steps_per_second": 2.499, | |
"step": 2400 | |
}, | |
{ | |
"epoch": 1.29, | |
"grad_norm": 20.53518295288086, | |
"learning_rate": 2.857777777777778e-05, | |
"loss": 0.2814, | |
"step": 2410 | |
}, | |
{ | |
"epoch": 1.29, | |
"grad_norm": 25.769210815429688, | |
"learning_rate": 2.8488888888888892e-05, | |
"loss": 0.369, | |
"step": 2420 | |
}, | |
{ | |
"epoch": 1.3, | |
"grad_norm": 13.653368949890137, | |
"learning_rate": 2.84e-05, | |
"loss": 0.2124, | |
"step": 2430 | |
}, | |
{ | |
"epoch": 1.3, | |
"grad_norm": 25.229494094848633, | |
"learning_rate": 2.831111111111111e-05, | |
"loss": 0.3308, | |
"step": 2440 | |
}, | |
{ | |
"epoch": 1.31, | |
"grad_norm": 27.03401756286621, | |
"learning_rate": 2.8222222222222223e-05, | |
"loss": 0.3702, | |
"step": 2450 | |
}, | |
{ | |
"epoch": 1.31, | |
"grad_norm": 28.564023971557617, | |
"learning_rate": 2.8133333333333334e-05, | |
"loss": 0.3641, | |
"step": 2460 | |
}, | |
{ | |
"epoch": 1.32, | |
"grad_norm": 30.772727966308594, | |
"learning_rate": 2.8044444444444444e-05, | |
"loss": 0.3255, | |
"step": 2470 | |
}, | |
{ | |
"epoch": 1.32, | |
"grad_norm": 33.72224426269531, | |
"learning_rate": 2.7955555555555558e-05, | |
"loss": 0.3417, | |
"step": 2480 | |
}, | |
{ | |
"epoch": 1.33, | |
"grad_norm": 6.024286270141602, | |
"learning_rate": 2.786666666666667e-05, | |
"loss": 0.2718, | |
"step": 2490 | |
}, | |
{ | |
"epoch": 1.33, | |
"grad_norm": 21.70349884033203, | |
"learning_rate": 2.777777777777778e-05, | |
"loss": 0.3281, | |
"step": 2500 | |
}, | |
{ | |
"epoch": 1.33, | |
"eval_accuracy": 0.8598, | |
"eval_f1_macro": 0.7768837492760268, | |
"eval_f1_micro": 0.8598, | |
"eval_loss": 0.5763229727745056, | |
"eval_runtime": 125.2018, | |
"eval_samples_per_second": 79.871, | |
"eval_steps_per_second": 2.5, | |
"step": 2500 | |
}, | |
{ | |
"epoch": 1.34, | |
"grad_norm": 24.865785598754883, | |
"learning_rate": 2.7688888888888893e-05, | |
"loss": 0.2607, | |
"step": 2510 | |
}, | |
{ | |
"epoch": 1.34, | |
"grad_norm": 21.175254821777344, | |
"learning_rate": 2.7600000000000003e-05, | |
"loss": 0.3339, | |
"step": 2520 | |
}, | |
{ | |
"epoch": 1.35, | |
"grad_norm": 23.783727645874023, | |
"learning_rate": 2.751111111111111e-05, | |
"loss": 0.261, | |
"step": 2530 | |
}, | |
{ | |
"epoch": 1.35, | |
"grad_norm": 28.94060707092285, | |
"learning_rate": 2.742222222222222e-05, | |
"loss": 0.3169, | |
"step": 2540 | |
}, | |
{ | |
"epoch": 1.36, | |
"grad_norm": 30.291284561157227, | |
"learning_rate": 2.733333333333333e-05, | |
"loss": 0.2878, | |
"step": 2550 | |
}, | |
{ | |
"epoch": 1.37, | |
"grad_norm": 26.320711135864258, | |
"learning_rate": 2.7244444444444445e-05, | |
"loss": 0.2883, | |
"step": 2560 | |
}, | |
{ | |
"epoch": 1.37, | |
"grad_norm": 35.11587905883789, | |
"learning_rate": 2.7155555555555556e-05, | |
"loss": 0.3464, | |
"step": 2570 | |
}, | |
{ | |
"epoch": 1.38, | |
"grad_norm": 21.037996292114258, | |
"learning_rate": 2.706666666666667e-05, | |
"loss": 0.3166, | |
"step": 2580 | |
}, | |
{ | |
"epoch": 1.38, | |
"grad_norm": 21.80624008178711, | |
"learning_rate": 2.697777777777778e-05, | |
"loss": 0.2901, | |
"step": 2590 | |
}, | |
{ | |
"epoch": 1.39, | |
"grad_norm": 33.16673278808594, | |
"learning_rate": 2.688888888888889e-05, | |
"loss": 0.3909, | |
"step": 2600 | |
}, | |
{ | |
"epoch": 1.39, | |
"eval_accuracy": 0.8545, | |
"eval_f1_macro": 0.7778240850537461, | |
"eval_f1_micro": 0.8545, | |
"eval_loss": 0.5518567562103271, | |
"eval_runtime": 125.2624, | |
"eval_samples_per_second": 79.832, | |
"eval_steps_per_second": 2.499, | |
"step": 2600 | |
}, | |
{ | |
"epoch": 1.39, | |
"grad_norm": 24.126609802246094, | |
"learning_rate": 2.6800000000000004e-05, | |
"loss": 0.3126, | |
"step": 2610 | |
}, | |
{ | |
"epoch": 1.4, | |
"grad_norm": 20.498868942260742, | |
"learning_rate": 2.6711111111111115e-05, | |
"loss": 0.2461, | |
"step": 2620 | |
}, | |
{ | |
"epoch": 1.4, | |
"grad_norm": 26.896896362304688, | |
"learning_rate": 2.6622222222222225e-05, | |
"loss": 0.2962, | |
"step": 2630 | |
}, | |
{ | |
"epoch": 1.41, | |
"grad_norm": 27.555248260498047, | |
"learning_rate": 2.6533333333333332e-05, | |
"loss": 0.3247, | |
"step": 2640 | |
}, | |
{ | |
"epoch": 1.41, | |
"grad_norm": 20.325254440307617, | |
"learning_rate": 2.6444444444444443e-05, | |
"loss": 0.3544, | |
"step": 2650 | |
}, | |
{ | |
"epoch": 1.42, | |
"grad_norm": 35.30345153808594, | |
"learning_rate": 2.6355555555555557e-05, | |
"loss": 0.355, | |
"step": 2660 | |
}, | |
{ | |
"epoch": 1.42, | |
"grad_norm": 19.69151496887207, | |
"learning_rate": 2.6266666666666667e-05, | |
"loss": 0.3337, | |
"step": 2670 | |
}, | |
{ | |
"epoch": 1.43, | |
"grad_norm": 21.770431518554688, | |
"learning_rate": 2.6177777777777777e-05, | |
"loss": 0.3063, | |
"step": 2680 | |
}, | |
{ | |
"epoch": 1.43, | |
"grad_norm": 20.847126007080078, | |
"learning_rate": 2.608888888888889e-05, | |
"loss": 0.27, | |
"step": 2690 | |
}, | |
{ | |
"epoch": 1.44, | |
"grad_norm": 31.529712677001953, | |
"learning_rate": 2.6000000000000002e-05, | |
"loss": 0.3064, | |
"step": 2700 | |
}, | |
{ | |
"epoch": 1.44, | |
"eval_accuracy": 0.8536, | |
"eval_f1_macro": 0.7790368774303406, | |
"eval_f1_micro": 0.8536, | |
"eval_loss": 0.5842456221580505, | |
"eval_runtime": 125.3409, | |
"eval_samples_per_second": 79.782, | |
"eval_steps_per_second": 2.497, | |
"step": 2700 | |
}, | |
{ | |
"epoch": 1.45, | |
"grad_norm": 39.8318977355957, | |
"learning_rate": 2.5911111111111112e-05, | |
"loss": 0.2705, | |
"step": 2710 | |
}, | |
{ | |
"epoch": 1.45, | |
"grad_norm": 29.740388870239258, | |
"learning_rate": 2.5822222222222226e-05, | |
"loss": 0.2761, | |
"step": 2720 | |
}, | |
{ | |
"epoch": 1.46, | |
"grad_norm": 33.67303466796875, | |
"learning_rate": 2.5733333333333337e-05, | |
"loss": 0.2903, | |
"step": 2730 | |
}, | |
{ | |
"epoch": 1.46, | |
"grad_norm": 18.60785484313965, | |
"learning_rate": 2.5644444444444444e-05, | |
"loss": 0.311, | |
"step": 2740 | |
}, | |
{ | |
"epoch": 1.47, | |
"grad_norm": 22.325334548950195, | |
"learning_rate": 2.5555555555555554e-05, | |
"loss": 0.3542, | |
"step": 2750 | |
}, | |
{ | |
"epoch": 1.47, | |
"grad_norm": 9.186369895935059, | |
"learning_rate": 2.5466666666666668e-05, | |
"loss": 0.2995, | |
"step": 2760 | |
}, | |
{ | |
"epoch": 1.48, | |
"grad_norm": 17.34896469116211, | |
"learning_rate": 2.537777777777778e-05, | |
"loss": 0.3195, | |
"step": 2770 | |
}, | |
{ | |
"epoch": 1.48, | |
"grad_norm": 33.827674865722656, | |
"learning_rate": 2.528888888888889e-05, | |
"loss": 0.3021, | |
"step": 2780 | |
}, | |
{ | |
"epoch": 1.49, | |
"grad_norm": 14.778377532958984, | |
"learning_rate": 2.5200000000000003e-05, | |
"loss": 0.2876, | |
"step": 2790 | |
}, | |
{ | |
"epoch": 1.49, | |
"grad_norm": 20.28546714782715, | |
"learning_rate": 2.5111111111111113e-05, | |
"loss": 0.2333, | |
"step": 2800 | |
}, | |
{ | |
"epoch": 1.49, | |
"eval_accuracy": 0.8447, | |
"eval_f1_macro": 0.7673871147650553, | |
"eval_f1_micro": 0.8447, | |
"eval_loss": 0.6084045767784119, | |
"eval_runtime": 125.269, | |
"eval_samples_per_second": 79.828, | |
"eval_steps_per_second": 2.499, | |
"step": 2800 | |
}, | |
{ | |
"epoch": 1.5, | |
"grad_norm": 26.260639190673828, | |
"learning_rate": 2.5022222222222224e-05, | |
"loss": 0.2639, | |
"step": 2810 | |
}, | |
{ | |
"epoch": 1.5, | |
"grad_norm": 22.30939292907715, | |
"learning_rate": 2.4933333333333334e-05, | |
"loss": 0.3508, | |
"step": 2820 | |
}, | |
{ | |
"epoch": 1.51, | |
"grad_norm": 29.393909454345703, | |
"learning_rate": 2.4844444444444444e-05, | |
"loss": 0.2569, | |
"step": 2830 | |
}, | |
{ | |
"epoch": 1.51, | |
"grad_norm": 26.341440200805664, | |
"learning_rate": 2.475555555555556e-05, | |
"loss": 0.2746, | |
"step": 2840 | |
}, | |
{ | |
"epoch": 1.52, | |
"grad_norm": 10.78348445892334, | |
"learning_rate": 2.466666666666667e-05, | |
"loss": 0.2224, | |
"step": 2850 | |
}, | |
{ | |
"epoch": 1.53, | |
"grad_norm": 23.771703720092773, | |
"learning_rate": 2.457777777777778e-05, | |
"loss": 0.2819, | |
"step": 2860 | |
}, | |
{ | |
"epoch": 1.53, | |
"grad_norm": 20.408435821533203, | |
"learning_rate": 2.448888888888889e-05, | |
"loss": 0.3146, | |
"step": 2870 | |
}, | |
{ | |
"epoch": 1.54, | |
"grad_norm": 18.23904800415039, | |
"learning_rate": 2.44e-05, | |
"loss": 0.3073, | |
"step": 2880 | |
}, | |
{ | |
"epoch": 1.54, | |
"grad_norm": 13.120834350585938, | |
"learning_rate": 2.431111111111111e-05, | |
"loss": 0.2746, | |
"step": 2890 | |
}, | |
{ | |
"epoch": 1.55, | |
"grad_norm": 26.703229904174805, | |
"learning_rate": 2.4222222222222224e-05, | |
"loss": 0.2361, | |
"step": 2900 | |
}, | |
{ | |
"epoch": 1.55, | |
"eval_accuracy": 0.8588, | |
"eval_f1_macro": 0.7853092454906397, | |
"eval_f1_micro": 0.8588, | |
"eval_loss": 0.5974697470664978, | |
"eval_runtime": 125.2009, | |
"eval_samples_per_second": 79.872, | |
"eval_steps_per_second": 2.5, | |
"step": 2900 | |
}, | |
{ | |
"epoch": 1.55, | |
"grad_norm": 22.55280876159668, | |
"learning_rate": 2.4133333333333335e-05, | |
"loss": 0.3899, | |
"step": 2910 | |
}, | |
{ | |
"epoch": 1.56, | |
"grad_norm": 18.248414993286133, | |
"learning_rate": 2.4044444444444445e-05, | |
"loss": 0.2995, | |
"step": 2920 | |
}, | |
{ | |
"epoch": 1.56, | |
"grad_norm": 26.95282554626465, | |
"learning_rate": 2.3955555555555556e-05, | |
"loss": 0.4188, | |
"step": 2930 | |
}, | |
{ | |
"epoch": 1.57, | |
"grad_norm": 18.83556365966797, | |
"learning_rate": 2.3866666666666666e-05, | |
"loss": 0.2215, | |
"step": 2940 | |
}, | |
{ | |
"epoch": 1.57, | |
"grad_norm": 21.471210479736328, | |
"learning_rate": 2.377777777777778e-05, | |
"loss": 0.3084, | |
"step": 2950 | |
}, | |
{ | |
"epoch": 1.58, | |
"grad_norm": 23.17391014099121, | |
"learning_rate": 2.368888888888889e-05, | |
"loss": 0.3674, | |
"step": 2960 | |
}, | |
{ | |
"epoch": 1.58, | |
"grad_norm": 31.083335876464844, | |
"learning_rate": 2.36e-05, | |
"loss": 0.286, | |
"step": 2970 | |
}, | |
{ | |
"epoch": 1.59, | |
"grad_norm": 27.996196746826172, | |
"learning_rate": 2.351111111111111e-05, | |
"loss": 0.3615, | |
"step": 2980 | |
}, | |
{ | |
"epoch": 1.59, | |
"grad_norm": 25.99715805053711, | |
"learning_rate": 2.3422222222222222e-05, | |
"loss": 0.2546, | |
"step": 2990 | |
}, | |
{ | |
"epoch": 1.6, | |
"grad_norm": 6.342403888702393, | |
"learning_rate": 2.3333333333333336e-05, | |
"loss": 0.3415, | |
"step": 3000 | |
}, | |
{ | |
"epoch": 1.6, | |
"eval_accuracy": 0.8572, | |
"eval_f1_macro": 0.7844301563007002, | |
"eval_f1_micro": 0.8572, | |
"eval_loss": 0.5700847506523132, | |
"eval_runtime": 125.2898, | |
"eval_samples_per_second": 79.815, | |
"eval_steps_per_second": 2.498, | |
"step": 3000 | |
}, | |
{ | |
"epoch": 1.61, | |
"grad_norm": 30.417648315429688, | |
"learning_rate": 2.3244444444444446e-05, | |
"loss": 0.3269, | |
"step": 3010 | |
}, | |
{ | |
"epoch": 1.61, | |
"grad_norm": 33.5994873046875, | |
"learning_rate": 2.3155555555555557e-05, | |
"loss": 0.342, | |
"step": 3020 | |
}, | |
{ | |
"epoch": 1.62, | |
"grad_norm": 30.278759002685547, | |
"learning_rate": 2.3066666666666667e-05, | |
"loss": 0.3424, | |
"step": 3030 | |
}, | |
{ | |
"epoch": 1.62, | |
"grad_norm": 28.47062873840332, | |
"learning_rate": 2.2977777777777778e-05, | |
"loss": 0.3254, | |
"step": 3040 | |
}, | |
{ | |
"epoch": 1.63, | |
"grad_norm": 27.17629051208496, | |
"learning_rate": 2.288888888888889e-05, | |
"loss": 0.4687, | |
"step": 3050 | |
}, | |
{ | |
"epoch": 1.63, | |
"grad_norm": 20.395219802856445, | |
"learning_rate": 2.2800000000000002e-05, | |
"loss": 0.2192, | |
"step": 3060 | |
}, | |
{ | |
"epoch": 1.64, | |
"grad_norm": 16.45476722717285, | |
"learning_rate": 2.2711111111111112e-05, | |
"loss": 0.5145, | |
"step": 3070 | |
}, | |
{ | |
"epoch": 1.64, | |
"grad_norm": 30.59330177307129, | |
"learning_rate": 2.2622222222222223e-05, | |
"loss": 0.2724, | |
"step": 3080 | |
}, | |
{ | |
"epoch": 1.65, | |
"grad_norm": 18.152570724487305, | |
"learning_rate": 2.2533333333333333e-05, | |
"loss": 0.319, | |
"step": 3090 | |
}, | |
{ | |
"epoch": 1.65, | |
"grad_norm": 26.22066879272461, | |
"learning_rate": 2.2444444444444447e-05, | |
"loss": 0.2535, | |
"step": 3100 | |
}, | |
{ | |
"epoch": 1.65, | |
"eval_accuracy": 0.8618, | |
"eval_f1_macro": 0.7827999074535285, | |
"eval_f1_micro": 0.8618, | |
"eval_loss": 0.5556691288948059, | |
"eval_runtime": 125.2171, | |
"eval_samples_per_second": 79.861, | |
"eval_steps_per_second": 2.5, | |
"step": 3100 | |
}, | |
{ | |
"epoch": 1.66, | |
"grad_norm": 12.443004608154297, | |
"learning_rate": 2.2355555555555558e-05, | |
"loss": 0.3178, | |
"step": 3110 | |
}, | |
{ | |
"epoch": 1.66, | |
"grad_norm": 27.130870819091797, | |
"learning_rate": 2.2266666666666668e-05, | |
"loss": 0.35, | |
"step": 3120 | |
}, | |
{ | |
"epoch": 1.67, | |
"grad_norm": 18.87975311279297, | |
"learning_rate": 2.217777777777778e-05, | |
"loss": 0.3123, | |
"step": 3130 | |
}, | |
{ | |
"epoch": 1.67, | |
"grad_norm": 25.946163177490234, | |
"learning_rate": 2.208888888888889e-05, | |
"loss": 0.3041, | |
"step": 3140 | |
}, | |
{ | |
"epoch": 1.68, | |
"grad_norm": 17.42608642578125, | |
"learning_rate": 2.2000000000000003e-05, | |
"loss": 0.2747, | |
"step": 3150 | |
}, | |
{ | |
"epoch": 1.69, | |
"grad_norm": 19.43211555480957, | |
"learning_rate": 2.1911111111111113e-05, | |
"loss": 0.2509, | |
"step": 3160 | |
}, | |
{ | |
"epoch": 1.69, | |
"grad_norm": 28.090959548950195, | |
"learning_rate": 2.1822222222222224e-05, | |
"loss": 0.2771, | |
"step": 3170 | |
}, | |
{ | |
"epoch": 1.7, | |
"grad_norm": 18.299936294555664, | |
"learning_rate": 2.1733333333333334e-05, | |
"loss": 0.3804, | |
"step": 3180 | |
}, | |
{ | |
"epoch": 1.7, | |
"grad_norm": 17.102720260620117, | |
"learning_rate": 2.1644444444444445e-05, | |
"loss": 0.2603, | |
"step": 3190 | |
}, | |
{ | |
"epoch": 1.71, | |
"grad_norm": 32.00794219970703, | |
"learning_rate": 2.1555555555555555e-05, | |
"loss": 0.2356, | |
"step": 3200 | |
}, | |
{ | |
"epoch": 1.71, | |
"eval_accuracy": 0.8612, | |
"eval_f1_macro": 0.782152177469759, | |
"eval_f1_micro": 0.8612, | |
"eval_loss": 0.5241914987564087, | |
"eval_runtime": 125.2745, | |
"eval_samples_per_second": 79.825, | |
"eval_steps_per_second": 2.499, | |
"step": 3200 | |
}, | |
{ | |
"epoch": 1.71, | |
"grad_norm": 31.098926544189453, | |
"learning_rate": 2.146666666666667e-05, | |
"loss": 0.257, | |
"step": 3210 | |
}, | |
{ | |
"epoch": 1.72, | |
"grad_norm": 22.57906723022461, | |
"learning_rate": 2.137777777777778e-05, | |
"loss": 0.2282, | |
"step": 3220 | |
}, | |
{ | |
"epoch": 1.72, | |
"grad_norm": 19.42692756652832, | |
"learning_rate": 2.128888888888889e-05, | |
"loss": 0.2567, | |
"step": 3230 | |
}, | |
{ | |
"epoch": 1.73, | |
"grad_norm": 22.812788009643555, | |
"learning_rate": 2.12e-05, | |
"loss": 0.2344, | |
"step": 3240 | |
}, | |
{ | |
"epoch": 1.73, | |
"grad_norm": 31.932159423828125, | |
"learning_rate": 2.111111111111111e-05, | |
"loss": 0.3359, | |
"step": 3250 | |
}, | |
{ | |
"epoch": 1.74, | |
"grad_norm": 21.950088500976562, | |
"learning_rate": 2.1022222222222225e-05, | |
"loss": 0.2967, | |
"step": 3260 | |
}, | |
{ | |
"epoch": 1.74, | |
"grad_norm": 31.876483917236328, | |
"learning_rate": 2.0933333333333335e-05, | |
"loss": 0.3029, | |
"step": 3270 | |
}, | |
{ | |
"epoch": 1.75, | |
"grad_norm": 28.16164779663086, | |
"learning_rate": 2.0844444444444446e-05, | |
"loss": 0.3843, | |
"step": 3280 | |
}, | |
{ | |
"epoch": 1.75, | |
"grad_norm": 18.941884994506836, | |
"learning_rate": 2.0755555555555556e-05, | |
"loss": 0.3108, | |
"step": 3290 | |
}, | |
{ | |
"epoch": 1.76, | |
"grad_norm": 15.317469596862793, | |
"learning_rate": 2.0666666666666666e-05, | |
"loss": 0.3383, | |
"step": 3300 | |
}, | |
{ | |
"epoch": 1.76, | |
"eval_accuracy": 0.8553, | |
"eval_f1_macro": 0.7873309798668017, | |
"eval_f1_micro": 0.8553, | |
"eval_loss": 0.5249765515327454, | |
"eval_runtime": 125.2578, | |
"eval_samples_per_second": 79.835, | |
"eval_steps_per_second": 2.499, | |
"step": 3300 | |
}, | |
{ | |
"epoch": 1.77, | |
"grad_norm": 21.06130027770996, | |
"learning_rate": 2.057777777777778e-05, | |
"loss": 0.2807, | |
"step": 3310 | |
}, | |
{ | |
"epoch": 1.77, | |
"grad_norm": 24.9218692779541, | |
"learning_rate": 2.048888888888889e-05, | |
"loss": 0.2634, | |
"step": 3320 | |
}, | |
{ | |
"epoch": 1.78, | |
"grad_norm": 16.426345825195312, | |
"learning_rate": 2.04e-05, | |
"loss": 0.2336, | |
"step": 3330 | |
}, | |
{ | |
"epoch": 1.78, | |
"grad_norm": 26.309751510620117, | |
"learning_rate": 2.031111111111111e-05, | |
"loss": 0.3588, | |
"step": 3340 | |
}, | |
{ | |
"epoch": 1.79, | |
"grad_norm": 16.83870506286621, | |
"learning_rate": 2.0222222222222222e-05, | |
"loss": 0.3155, | |
"step": 3350 | |
}, | |
{ | |
"epoch": 1.79, | |
"grad_norm": 19.22264862060547, | |
"learning_rate": 2.0133333333333336e-05, | |
"loss": 0.2739, | |
"step": 3360 | |
}, | |
{ | |
"epoch": 1.8, | |
"grad_norm": 19.02859878540039, | |
"learning_rate": 2.0044444444444446e-05, | |
"loss": 0.2537, | |
"step": 3370 | |
}, | |
{ | |
"epoch": 1.8, | |
"grad_norm": 23.24178695678711, | |
"learning_rate": 1.9955555555555557e-05, | |
"loss": 0.2639, | |
"step": 3380 | |
}, | |
{ | |
"epoch": 1.81, | |
"grad_norm": 13.695367813110352, | |
"learning_rate": 1.9866666666666667e-05, | |
"loss": 0.2067, | |
"step": 3390 | |
}, | |
{ | |
"epoch": 1.81, | |
"grad_norm": 20.969131469726562, | |
"learning_rate": 1.9777777777777778e-05, | |
"loss": 0.1886, | |
"step": 3400 | |
}, | |
{ | |
"epoch": 1.81, | |
"eval_accuracy": 0.8658, | |
"eval_f1_macro": 0.7923625631474337, | |
"eval_f1_micro": 0.8658, | |
"eval_loss": 0.5301083922386169, | |
"eval_runtime": 125.2412, | |
"eval_samples_per_second": 79.846, | |
"eval_steps_per_second": 2.499, | |
"step": 3400 | |
}, | |
{ | |
"epoch": 1.82, | |
"grad_norm": 30.716920852661133, | |
"learning_rate": 1.968888888888889e-05, | |
"loss": 0.3543, | |
"step": 3410 | |
}, | |
{ | |
"epoch": 1.82, | |
"grad_norm": 25.40995216369629, | |
"learning_rate": 1.9600000000000002e-05, | |
"loss": 0.3573, | |
"step": 3420 | |
}, | |
{ | |
"epoch": 1.83, | |
"grad_norm": 25.003276824951172, | |
"learning_rate": 1.9511111111111113e-05, | |
"loss": 0.2405, | |
"step": 3430 | |
}, | |
{ | |
"epoch": 1.83, | |
"grad_norm": 37.33772277832031, | |
"learning_rate": 1.9422222222222223e-05, | |
"loss": 0.3001, | |
"step": 3440 | |
}, | |
{ | |
"epoch": 1.84, | |
"grad_norm": 29.57843589782715, | |
"learning_rate": 1.9333333333333333e-05, | |
"loss": 0.2823, | |
"step": 3450 | |
}, | |
{ | |
"epoch": 1.85, | |
"grad_norm": 23.851917266845703, | |
"learning_rate": 1.9244444444444444e-05, | |
"loss": 0.235, | |
"step": 3460 | |
}, | |
{ | |
"epoch": 1.85, | |
"grad_norm": 25.79085922241211, | |
"learning_rate": 1.9155555555555558e-05, | |
"loss": 0.2136, | |
"step": 3470 | |
}, | |
{ | |
"epoch": 1.86, | |
"grad_norm": 18.118497848510742, | |
"learning_rate": 1.9066666666666668e-05, | |
"loss": 0.2425, | |
"step": 3480 | |
}, | |
{ | |
"epoch": 1.86, | |
"grad_norm": 31.972190856933594, | |
"learning_rate": 1.897777777777778e-05, | |
"loss": 0.2761, | |
"step": 3490 | |
}, | |
{ | |
"epoch": 1.87, | |
"grad_norm": 22.953575134277344, | |
"learning_rate": 1.888888888888889e-05, | |
"loss": 0.2468, | |
"step": 3500 | |
}, | |
{ | |
"epoch": 1.87, | |
"eval_accuracy": 0.8595, | |
"eval_f1_macro": 0.7812713849689391, | |
"eval_f1_micro": 0.8595, | |
"eval_loss": 0.5459020733833313, | |
"eval_runtime": 125.3137, | |
"eval_samples_per_second": 79.8, | |
"eval_steps_per_second": 2.498, | |
"step": 3500 | |
}, | |
{ | |
"epoch": 1.87, | |
"grad_norm": 16.223791122436523, | |
"learning_rate": 1.88e-05, | |
"loss": 0.2574, | |
"step": 3510 | |
}, | |
{ | |
"epoch": 1.88, | |
"grad_norm": 11.477206230163574, | |
"learning_rate": 1.8711111111111113e-05, | |
"loss": 0.1905, | |
"step": 3520 | |
}, | |
{ | |
"epoch": 1.88, | |
"grad_norm": 22.725454330444336, | |
"learning_rate": 1.8622222222222224e-05, | |
"loss": 0.2126, | |
"step": 3530 | |
}, | |
{ | |
"epoch": 1.89, | |
"grad_norm": 13.858516693115234, | |
"learning_rate": 1.8533333333333334e-05, | |
"loss": 0.2417, | |
"step": 3540 | |
}, | |
{ | |
"epoch": 1.89, | |
"grad_norm": 14.678023338317871, | |
"learning_rate": 1.8444444444444445e-05, | |
"loss": 0.2442, | |
"step": 3550 | |
}, | |
{ | |
"epoch": 1.9, | |
"grad_norm": 11.627692222595215, | |
"learning_rate": 1.8355555555555555e-05, | |
"loss": 0.2736, | |
"step": 3560 | |
}, | |
{ | |
"epoch": 1.9, | |
"grad_norm": 24.857446670532227, | |
"learning_rate": 1.826666666666667e-05, | |
"loss": 0.3312, | |
"step": 3570 | |
}, | |
{ | |
"epoch": 1.91, | |
"grad_norm": 22.696962356567383, | |
"learning_rate": 1.817777777777778e-05, | |
"loss": 0.2294, | |
"step": 3580 | |
}, | |
{ | |
"epoch": 1.91, | |
"grad_norm": 33.160343170166016, | |
"learning_rate": 1.808888888888889e-05, | |
"loss": 0.2481, | |
"step": 3590 | |
}, | |
{ | |
"epoch": 1.92, | |
"grad_norm": 21.321407318115234, | |
"learning_rate": 1.8e-05, | |
"loss": 0.2947, | |
"step": 3600 | |
}, | |
{ | |
"epoch": 1.92, | |
"eval_accuracy": 0.8688, | |
"eval_f1_macro": 0.7910490707304132, | |
"eval_f1_micro": 0.8688, | |
"eval_loss": 0.51407390832901, | |
"eval_runtime": 125.3442, | |
"eval_samples_per_second": 79.78, | |
"eval_steps_per_second": 2.497, | |
"step": 3600 | |
}, | |
{ | |
"epoch": 1.93, | |
"grad_norm": 24.3467960357666, | |
"learning_rate": 1.791111111111111e-05, | |
"loss": 0.2795, | |
"step": 3610 | |
}, | |
{ | |
"epoch": 1.93, | |
"grad_norm": 24.753925323486328, | |
"learning_rate": 1.7822222222222225e-05, | |
"loss": 0.2395, | |
"step": 3620 | |
}, | |
{ | |
"epoch": 1.94, | |
"grad_norm": 17.409852981567383, | |
"learning_rate": 1.7733333333333335e-05, | |
"loss": 0.2617, | |
"step": 3630 | |
}, | |
{ | |
"epoch": 1.94, | |
"grad_norm": 24.267776489257812, | |
"learning_rate": 1.7644444444444446e-05, | |
"loss": 0.2099, | |
"step": 3640 | |
}, | |
{ | |
"epoch": 1.95, | |
"grad_norm": 24.462526321411133, | |
"learning_rate": 1.7555555555555556e-05, | |
"loss": 0.2953, | |
"step": 3650 | |
}, | |
{ | |
"epoch": 1.95, | |
"grad_norm": 16.87688446044922, | |
"learning_rate": 1.7466666666666667e-05, | |
"loss": 0.364, | |
"step": 3660 | |
}, | |
{ | |
"epoch": 1.96, | |
"grad_norm": 16.433847427368164, | |
"learning_rate": 1.737777777777778e-05, | |
"loss": 0.2613, | |
"step": 3670 | |
}, | |
{ | |
"epoch": 1.96, | |
"grad_norm": 29.817991256713867, | |
"learning_rate": 1.728888888888889e-05, | |
"loss": 0.3216, | |
"step": 3680 | |
}, | |
{ | |
"epoch": 1.97, | |
"grad_norm": 26.87920379638672, | |
"learning_rate": 1.7199999999999998e-05, | |
"loss": 0.3593, | |
"step": 3690 | |
}, | |
{ | |
"epoch": 1.97, | |
"grad_norm": 20.05802345275879, | |
"learning_rate": 1.7111111111111112e-05, | |
"loss": 0.2625, | |
"step": 3700 | |
}, | |
{ | |
"epoch": 1.97, | |
"eval_accuracy": 0.867, | |
"eval_f1_macro": 0.7909960365885448, | |
"eval_f1_micro": 0.867, | |
"eval_loss": 0.5025032162666321, | |
"eval_runtime": 125.4107, | |
"eval_samples_per_second": 79.738, | |
"eval_steps_per_second": 2.496, | |
"step": 3700 | |
}, | |
{ | |
"epoch": 1.98, | |
"grad_norm": 21.102169036865234, | |
"learning_rate": 1.7022222222222222e-05, | |
"loss": 0.1997, | |
"step": 3710 | |
}, | |
{ | |
"epoch": 1.98, | |
"grad_norm": 20.016801834106445, | |
"learning_rate": 1.6933333333333333e-05, | |
"loss": 0.2763, | |
"step": 3720 | |
}, | |
{ | |
"epoch": 1.99, | |
"grad_norm": 23.357534408569336, | |
"learning_rate": 1.6844444444444447e-05, | |
"loss": 0.2474, | |
"step": 3730 | |
}, | |
{ | |
"epoch": 1.99, | |
"grad_norm": 16.342079162597656, | |
"learning_rate": 1.6755555555555557e-05, | |
"loss": 0.2428, | |
"step": 3740 | |
}, | |
{ | |
"epoch": 2.0, | |
"grad_norm": 24.250011444091797, | |
"learning_rate": 1.6666666666666667e-05, | |
"loss": 0.3232, | |
"step": 3750 | |
}, | |
{ | |
"epoch": 2.01, | |
"grad_norm": 5.438930511474609, | |
"learning_rate": 1.6577777777777778e-05, | |
"loss": 0.0914, | |
"step": 3760 | |
}, | |
{ | |
"epoch": 2.01, | |
"grad_norm": 4.791384696960449, | |
"learning_rate": 1.648888888888889e-05, | |
"loss": 0.0961, | |
"step": 3770 | |
}, | |
{ | |
"epoch": 2.02, | |
"grad_norm": 7.610979080200195, | |
"learning_rate": 1.6400000000000002e-05, | |
"loss": 0.0922, | |
"step": 3780 | |
}, | |
{ | |
"epoch": 2.02, | |
"grad_norm": 6.301875591278076, | |
"learning_rate": 1.6311111111111113e-05, | |
"loss": 0.0701, | |
"step": 3790 | |
}, | |
{ | |
"epoch": 2.03, | |
"grad_norm": 15.098999977111816, | |
"learning_rate": 1.6222222222222223e-05, | |
"loss": 0.0829, | |
"step": 3800 | |
}, | |
{ | |
"epoch": 2.03, | |
"eval_accuracy": 0.8697, | |
"eval_f1_macro": 0.8003818000945615, | |
"eval_f1_micro": 0.8697, | |
"eval_loss": 0.5624653100967407, | |
"eval_runtime": 125.2333, | |
"eval_samples_per_second": 79.851, | |
"eval_steps_per_second": 2.499, | |
"step": 3800 | |
}, | |
{ | |
"epoch": 2.03, | |
"grad_norm": 11.659759521484375, | |
"learning_rate": 1.6133333333333334e-05, | |
"loss": 0.0388, | |
"step": 3810 | |
}, | |
{ | |
"epoch": 2.04, | |
"grad_norm": 13.798972129821777, | |
"learning_rate": 1.6044444444444444e-05, | |
"loss": 0.1004, | |
"step": 3820 | |
}, | |
{ | |
"epoch": 2.04, | |
"grad_norm": 10.4561185836792, | |
"learning_rate": 1.5955555555555558e-05, | |
"loss": 0.0375, | |
"step": 3830 | |
}, | |
{ | |
"epoch": 2.05, | |
"grad_norm": 0.9646662473678589, | |
"learning_rate": 1.586666666666667e-05, | |
"loss": 0.0787, | |
"step": 3840 | |
}, | |
{ | |
"epoch": 2.05, | |
"grad_norm": 9.96677303314209, | |
"learning_rate": 1.577777777777778e-05, | |
"loss": 0.0557, | |
"step": 3850 | |
}, | |
{ | |
"epoch": 2.06, | |
"grad_norm": 18.659833908081055, | |
"learning_rate": 1.568888888888889e-05, | |
"loss": 0.0513, | |
"step": 3860 | |
}, | |
{ | |
"epoch": 2.06, | |
"grad_norm": 2.004011869430542, | |
"learning_rate": 1.56e-05, | |
"loss": 0.0427, | |
"step": 3870 | |
}, | |
{ | |
"epoch": 2.07, | |
"grad_norm": 20.428512573242188, | |
"learning_rate": 1.5511111111111114e-05, | |
"loss": 0.0475, | |
"step": 3880 | |
}, | |
{ | |
"epoch": 2.07, | |
"grad_norm": 14.12541389465332, | |
"learning_rate": 1.5422222222222224e-05, | |
"loss": 0.0703, | |
"step": 3890 | |
}, | |
{ | |
"epoch": 2.08, | |
"grad_norm": 11.519500732421875, | |
"learning_rate": 1.5333333333333334e-05, | |
"loss": 0.0297, | |
"step": 3900 | |
}, | |
{ | |
"epoch": 2.08, | |
"eval_accuracy": 0.8698, | |
"eval_f1_macro": 0.8018056240903841, | |
"eval_f1_micro": 0.8698, | |
"eval_loss": 0.6302907466888428, | |
"eval_runtime": 125.2533, | |
"eval_samples_per_second": 79.838, | |
"eval_steps_per_second": 2.499, | |
"step": 3900 | |
}, | |
{ | |
"epoch": 2.09, | |
"grad_norm": 1.5828900337219238, | |
"learning_rate": 1.5244444444444445e-05, | |
"loss": 0.0656, | |
"step": 3910 | |
}, | |
{ | |
"epoch": 2.09, | |
"grad_norm": 2.4143779277801514, | |
"learning_rate": 1.5155555555555555e-05, | |
"loss": 0.0343, | |
"step": 3920 | |
}, | |
{ | |
"epoch": 2.1, | |
"grad_norm": 1.4197189807891846, | |
"learning_rate": 1.5066666666666668e-05, | |
"loss": 0.0378, | |
"step": 3930 | |
}, | |
{ | |
"epoch": 2.1, | |
"grad_norm": 0.3790437579154968, | |
"learning_rate": 1.497777777777778e-05, | |
"loss": 0.1217, | |
"step": 3940 | |
}, | |
{ | |
"epoch": 2.11, | |
"grad_norm": 17.306047439575195, | |
"learning_rate": 1.4888888888888888e-05, | |
"loss": 0.0532, | |
"step": 3950 | |
}, | |
{ | |
"epoch": 2.11, | |
"grad_norm": 15.603998184204102, | |
"learning_rate": 1.48e-05, | |
"loss": 0.0296, | |
"step": 3960 | |
}, | |
{ | |
"epoch": 2.12, | |
"grad_norm": 12.390862464904785, | |
"learning_rate": 1.4711111111111111e-05, | |
"loss": 0.0204, | |
"step": 3970 | |
}, | |
{ | |
"epoch": 2.12, | |
"grad_norm": 15.381484985351562, | |
"learning_rate": 1.4622222222222223e-05, | |
"loss": 0.0664, | |
"step": 3980 | |
}, | |
{ | |
"epoch": 2.13, | |
"grad_norm": 1.980068564414978, | |
"learning_rate": 1.4533333333333335e-05, | |
"loss": 0.0307, | |
"step": 3990 | |
}, | |
{ | |
"epoch": 2.13, | |
"grad_norm": 4.541590690612793, | |
"learning_rate": 1.4444444444444444e-05, | |
"loss": 0.0474, | |
"step": 4000 | |
}, | |
{ | |
"epoch": 2.13, | |
"eval_accuracy": 0.8713, | |
"eval_f1_macro": 0.8046329981487478, | |
"eval_f1_micro": 0.8713, | |
"eval_loss": 0.6244204640388489, | |
"eval_runtime": 125.3647, | |
"eval_samples_per_second": 79.767, | |
"eval_steps_per_second": 2.497, | |
"step": 4000 | |
}, | |
{ | |
"epoch": 2.14, | |
"grad_norm": 15.43924617767334, | |
"learning_rate": 1.4355555555555556e-05, | |
"loss": 0.0256, | |
"step": 4010 | |
}, | |
{ | |
"epoch": 2.14, | |
"grad_norm": 13.164155960083008, | |
"learning_rate": 1.4266666666666667e-05, | |
"loss": 0.084, | |
"step": 4020 | |
}, | |
{ | |
"epoch": 2.15, | |
"grad_norm": 16.708253860473633, | |
"learning_rate": 1.4177777777777779e-05, | |
"loss": 0.0789, | |
"step": 4030 | |
}, | |
{ | |
"epoch": 2.15, | |
"grad_norm": 8.556520462036133, | |
"learning_rate": 1.4088888888888891e-05, | |
"loss": 0.0366, | |
"step": 4040 | |
}, | |
{ | |
"epoch": 2.16, | |
"grad_norm": 22.674955368041992, | |
"learning_rate": 1.4000000000000001e-05, | |
"loss": 0.1215, | |
"step": 4050 | |
}, | |
{ | |
"epoch": 2.17, | |
"grad_norm": 16.388124465942383, | |
"learning_rate": 1.391111111111111e-05, | |
"loss": 0.0926, | |
"step": 4060 | |
}, | |
{ | |
"epoch": 2.17, | |
"grad_norm": 9.75473690032959, | |
"learning_rate": 1.3822222222222222e-05, | |
"loss": 0.0504, | |
"step": 4070 | |
}, | |
{ | |
"epoch": 2.18, | |
"grad_norm": 6.83505392074585, | |
"learning_rate": 1.3733333333333335e-05, | |
"loss": 0.0523, | |
"step": 4080 | |
}, | |
{ | |
"epoch": 2.18, | |
"grad_norm": 2.356135606765747, | |
"learning_rate": 1.3644444444444445e-05, | |
"loss": 0.0628, | |
"step": 4090 | |
}, | |
{ | |
"epoch": 2.19, | |
"grad_norm": 7.485025882720947, | |
"learning_rate": 1.3555555555555557e-05, | |
"loss": 0.0267, | |
"step": 4100 | |
}, | |
{ | |
"epoch": 2.19, | |
"eval_accuracy": 0.8737, | |
"eval_f1_macro": 0.806064854698872, | |
"eval_f1_micro": 0.8737, | |
"eval_loss": 0.5801290273666382, | |
"eval_runtime": 125.3816, | |
"eval_samples_per_second": 79.757, | |
"eval_steps_per_second": 2.496, | |
"step": 4100 | |
}, | |
{ | |
"epoch": 2.19, | |
"grad_norm": 15.373751640319824, | |
"learning_rate": 1.3466666666666666e-05, | |
"loss": 0.0548, | |
"step": 4110 | |
}, | |
{ | |
"epoch": 2.2, | |
"grad_norm": 1.997728943824768, | |
"learning_rate": 1.3377777777777778e-05, | |
"loss": 0.0744, | |
"step": 4120 | |
}, | |
{ | |
"epoch": 2.2, | |
"grad_norm": 3.4920177459716797, | |
"learning_rate": 1.328888888888889e-05, | |
"loss": 0.0474, | |
"step": 4130 | |
}, | |
{ | |
"epoch": 2.21, | |
"grad_norm": 17.398141860961914, | |
"learning_rate": 1.32e-05, | |
"loss": 0.0382, | |
"step": 4140 | |
}, | |
{ | |
"epoch": 2.21, | |
"grad_norm": 4.560288429260254, | |
"learning_rate": 1.3111111111111113e-05, | |
"loss": 0.0432, | |
"step": 4150 | |
}, | |
{ | |
"epoch": 2.22, | |
"grad_norm": 1.7191858291625977, | |
"learning_rate": 1.3022222222222222e-05, | |
"loss": 0.0593, | |
"step": 4160 | |
}, | |
{ | |
"epoch": 2.22, | |
"grad_norm": 1.9446855783462524, | |
"learning_rate": 1.2933333333333334e-05, | |
"loss": 0.0374, | |
"step": 4170 | |
}, | |
{ | |
"epoch": 2.23, | |
"grad_norm": 13.725001335144043, | |
"learning_rate": 1.2844444444444446e-05, | |
"loss": 0.0758, | |
"step": 4180 | |
}, | |
{ | |
"epoch": 2.23, | |
"grad_norm": 10.993112564086914, | |
"learning_rate": 1.2755555555555556e-05, | |
"loss": 0.039, | |
"step": 4190 | |
}, | |
{ | |
"epoch": 2.24, | |
"grad_norm": 3.023743152618408, | |
"learning_rate": 1.2666666666666668e-05, | |
"loss": 0.0487, | |
"step": 4200 | |
}, | |
{ | |
"epoch": 2.24, | |
"eval_accuracy": 0.8745, | |
"eval_f1_macro": 0.8018343299864858, | |
"eval_f1_micro": 0.8745, | |
"eval_loss": 0.5914608240127563, | |
"eval_runtime": 125.4055, | |
"eval_samples_per_second": 79.741, | |
"eval_steps_per_second": 2.496, | |
"step": 4200 | |
}, | |
{ | |
"epoch": 2.25, | |
"grad_norm": 10.14614200592041, | |
"learning_rate": 1.2577777777777777e-05, | |
"loss": 0.0456, | |
"step": 4210 | |
}, | |
{ | |
"epoch": 2.25, | |
"grad_norm": 8.039652824401855, | |
"learning_rate": 1.248888888888889e-05, | |
"loss": 0.0765, | |
"step": 4220 | |
}, | |
{ | |
"epoch": 2.26, | |
"grad_norm": 0.6049771308898926, | |
"learning_rate": 1.24e-05, | |
"loss": 0.0577, | |
"step": 4230 | |
}, | |
{ | |
"epoch": 2.26, | |
"grad_norm": 2.0384902954101562, | |
"learning_rate": 1.2311111111111112e-05, | |
"loss": 0.0487, | |
"step": 4240 | |
}, | |
{ | |
"epoch": 2.27, | |
"grad_norm": 10.251029014587402, | |
"learning_rate": 1.2222222222222222e-05, | |
"loss": 0.0435, | |
"step": 4250 | |
}, | |
{ | |
"epoch": 2.27, | |
"grad_norm": 4.818918704986572, | |
"learning_rate": 1.2133333333333335e-05, | |
"loss": 0.0354, | |
"step": 4260 | |
}, | |
{ | |
"epoch": 2.28, | |
"grad_norm": 12.102091789245605, | |
"learning_rate": 1.2044444444444445e-05, | |
"loss": 0.0525, | |
"step": 4270 | |
}, | |
{ | |
"epoch": 2.28, | |
"grad_norm": 1.7436389923095703, | |
"learning_rate": 1.1955555555555556e-05, | |
"loss": 0.0288, | |
"step": 4280 | |
}, | |
{ | |
"epoch": 2.29, | |
"grad_norm": 5.876868724822998, | |
"learning_rate": 1.1866666666666668e-05, | |
"loss": 0.0388, | |
"step": 4290 | |
}, | |
{ | |
"epoch": 2.29, | |
"grad_norm": 1.1634790897369385, | |
"learning_rate": 1.1777777777777778e-05, | |
"loss": 0.0272, | |
"step": 4300 | |
}, | |
{ | |
"epoch": 2.29, | |
"eval_accuracy": 0.8764, | |
"eval_f1_macro": 0.8043166746272719, | |
"eval_f1_micro": 0.8764, | |
"eval_loss": 0.6174055933952332, | |
"eval_runtime": 125.4129, | |
"eval_samples_per_second": 79.737, | |
"eval_steps_per_second": 2.496, | |
"step": 4300 | |
}, | |
{ | |
"epoch": 2.3, | |
"grad_norm": 3.459371328353882, | |
"learning_rate": 1.168888888888889e-05, | |
"loss": 0.0292, | |
"step": 4310 | |
}, | |
{ | |
"epoch": 2.3, | |
"grad_norm": 5.280309677124023, | |
"learning_rate": 1.16e-05, | |
"loss": 0.05, | |
"step": 4320 | |
}, | |
{ | |
"epoch": 2.31, | |
"grad_norm": 15.677001953125, | |
"learning_rate": 1.1511111111111111e-05, | |
"loss": 0.0971, | |
"step": 4330 | |
}, | |
{ | |
"epoch": 2.31, | |
"grad_norm": 17.34125518798828, | |
"learning_rate": 1.1422222222222223e-05, | |
"loss": 0.0679, | |
"step": 4340 | |
}, | |
{ | |
"epoch": 2.32, | |
"grad_norm": 12.430140495300293, | |
"learning_rate": 1.1333333333333334e-05, | |
"loss": 0.0536, | |
"step": 4350 | |
}, | |
{ | |
"epoch": 2.33, | |
"grad_norm": 1.634233832359314, | |
"learning_rate": 1.1244444444444444e-05, | |
"loss": 0.0215, | |
"step": 4360 | |
}, | |
{ | |
"epoch": 2.33, | |
"grad_norm": 0.8702821135520935, | |
"learning_rate": 1.1155555555555556e-05, | |
"loss": 0.0634, | |
"step": 4370 | |
}, | |
{ | |
"epoch": 2.34, | |
"grad_norm": 4.945422649383545, | |
"learning_rate": 1.1066666666666667e-05, | |
"loss": 0.0815, | |
"step": 4380 | |
}, | |
{ | |
"epoch": 2.34, | |
"grad_norm": 13.194092750549316, | |
"learning_rate": 1.0977777777777779e-05, | |
"loss": 0.0314, | |
"step": 4390 | |
}, | |
{ | |
"epoch": 2.35, | |
"grad_norm": 1.432800531387329, | |
"learning_rate": 1.088888888888889e-05, | |
"loss": 0.02, | |
"step": 4400 | |
}, | |
{ | |
"epoch": 2.35, | |
"eval_accuracy": 0.87, | |
"eval_f1_macro": 0.7985763244299169, | |
"eval_f1_micro": 0.87, | |
"eval_loss": 0.6261060833930969, | |
"eval_runtime": 125.4273, | |
"eval_samples_per_second": 79.727, | |
"eval_steps_per_second": 2.495, | |
"step": 4400 | |
}, | |
{ | |
"epoch": 2.35, | |
"grad_norm": 12.279364585876465, | |
"learning_rate": 1.08e-05, | |
"loss": 0.0547, | |
"step": 4410 | |
}, | |
{ | |
"epoch": 2.36, | |
"grad_norm": 12.83892822265625, | |
"learning_rate": 1.0711111111111112e-05, | |
"loss": 0.0482, | |
"step": 4420 | |
}, | |
{ | |
"epoch": 2.36, | |
"grad_norm": 16.726667404174805, | |
"learning_rate": 1.0622222222222223e-05, | |
"loss": 0.0504, | |
"step": 4430 | |
}, | |
{ | |
"epoch": 2.37, | |
"grad_norm": 32.52857208251953, | |
"learning_rate": 1.0533333333333335e-05, | |
"loss": 0.0836, | |
"step": 4440 | |
}, | |
{ | |
"epoch": 2.37, | |
"grad_norm": 6.495666980743408, | |
"learning_rate": 1.0444444444444445e-05, | |
"loss": 0.0251, | |
"step": 4450 | |
}, | |
{ | |
"epoch": 2.38, | |
"grad_norm": 0.357742041349411, | |
"learning_rate": 1.0355555555555556e-05, | |
"loss": 0.0645, | |
"step": 4460 | |
}, | |
{ | |
"epoch": 2.38, | |
"grad_norm": 15.541790962219238, | |
"learning_rate": 1.0266666666666668e-05, | |
"loss": 0.0365, | |
"step": 4470 | |
}, | |
{ | |
"epoch": 2.39, | |
"grad_norm": 8.922079086303711, | |
"learning_rate": 1.0177777777777778e-05, | |
"loss": 0.0473, | |
"step": 4480 | |
}, | |
{ | |
"epoch": 2.39, | |
"grad_norm": 13.641845703125, | |
"learning_rate": 1.0088888888888889e-05, | |
"loss": 0.0702, | |
"step": 4490 | |
}, | |
{ | |
"epoch": 2.4, | |
"grad_norm": 0.6197155714035034, | |
"learning_rate": 1e-05, | |
"loss": 0.0414, | |
"step": 4500 | |
}, | |
{ | |
"epoch": 2.4, | |
"eval_accuracy": 0.8748, | |
"eval_f1_macro": 0.8035611999756955, | |
"eval_f1_micro": 0.8748, | |
"eval_loss": 0.6157392859458923, | |
"eval_runtime": 125.3937, | |
"eval_samples_per_second": 79.749, | |
"eval_steps_per_second": 2.496, | |
"step": 4500 | |
}, | |
{ | |
"epoch": 2.41, | |
"grad_norm": 1.329261302947998, | |
"learning_rate": 9.911111111111111e-06, | |
"loss": 0.0358, | |
"step": 4510 | |
}, | |
{ | |
"epoch": 2.41, | |
"grad_norm": 15.336833953857422, | |
"learning_rate": 9.822222222222223e-06, | |
"loss": 0.0478, | |
"step": 4520 | |
}, | |
{ | |
"epoch": 2.42, | |
"grad_norm": 10.690662384033203, | |
"learning_rate": 9.733333333333334e-06, | |
"loss": 0.1247, | |
"step": 4530 | |
}, | |
{ | |
"epoch": 2.42, | |
"grad_norm": 16.75906753540039, | |
"learning_rate": 9.644444444444444e-06, | |
"loss": 0.0898, | |
"step": 4540 | |
}, | |
{ | |
"epoch": 2.43, | |
"grad_norm": 7.825833320617676, | |
"learning_rate": 9.555555555555556e-06, | |
"loss": 0.0455, | |
"step": 4550 | |
}, | |
{ | |
"epoch": 2.43, | |
"grad_norm": 21.768123626708984, | |
"learning_rate": 9.466666666666667e-06, | |
"loss": 0.058, | |
"step": 4560 | |
}, | |
{ | |
"epoch": 2.44, | |
"grad_norm": 11.219696998596191, | |
"learning_rate": 9.377777777777779e-06, | |
"loss": 0.03, | |
"step": 4570 | |
}, | |
{ | |
"epoch": 2.44, | |
"grad_norm": 0.5012785792350769, | |
"learning_rate": 9.288888888888888e-06, | |
"loss": 0.038, | |
"step": 4580 | |
}, | |
{ | |
"epoch": 2.45, | |
"grad_norm": 24.533123016357422, | |
"learning_rate": 9.2e-06, | |
"loss": 0.0551, | |
"step": 4590 | |
}, | |
{ | |
"epoch": 2.45, | |
"grad_norm": 3.5600054264068604, | |
"learning_rate": 9.111111111111112e-06, | |
"loss": 0.0394, | |
"step": 4600 | |
}, | |
{ | |
"epoch": 2.45, | |
"eval_accuracy": 0.8755, | |
"eval_f1_macro": 0.8076102905129778, | |
"eval_f1_micro": 0.8755, | |
"eval_loss": 0.6051170825958252, | |
"eval_runtime": 125.3948, | |
"eval_samples_per_second": 79.748, | |
"eval_steps_per_second": 2.496, | |
"step": 4600 | |
}, | |
{ | |
"epoch": 2.46, | |
"grad_norm": 12.15698528289795, | |
"learning_rate": 9.022222222222223e-06, | |
"loss": 0.0483, | |
"step": 4610 | |
}, | |
{ | |
"epoch": 2.46, | |
"grad_norm": 25.406009674072266, | |
"learning_rate": 8.933333333333333e-06, | |
"loss": 0.1181, | |
"step": 4620 | |
}, | |
{ | |
"epoch": 2.47, | |
"grad_norm": 14.99547004699707, | |
"learning_rate": 8.844444444444445e-06, | |
"loss": 0.069, | |
"step": 4630 | |
}, | |
{ | |
"epoch": 2.47, | |
"grad_norm": 16.374536514282227, | |
"learning_rate": 8.755555555555556e-06, | |
"loss": 0.0705, | |
"step": 4640 | |
}, | |
{ | |
"epoch": 2.48, | |
"grad_norm": 8.475914001464844, | |
"learning_rate": 8.666666666666668e-06, | |
"loss": 0.0479, | |
"step": 4650 | |
}, | |
{ | |
"epoch": 2.49, | |
"grad_norm": 14.819987297058105, | |
"learning_rate": 8.577777777777778e-06, | |
"loss": 0.0403, | |
"step": 4660 | |
}, | |
{ | |
"epoch": 2.49, | |
"grad_norm": 13.341416358947754, | |
"learning_rate": 8.488888888888889e-06, | |
"loss": 0.0305, | |
"step": 4670 | |
}, | |
{ | |
"epoch": 2.5, | |
"grad_norm": 14.73620891571045, | |
"learning_rate": 8.400000000000001e-06, | |
"loss": 0.0719, | |
"step": 4680 | |
}, | |
{ | |
"epoch": 2.5, | |
"grad_norm": 12.851972579956055, | |
"learning_rate": 8.311111111111111e-06, | |
"loss": 0.0499, | |
"step": 4690 | |
}, | |
{ | |
"epoch": 2.51, | |
"grad_norm": 17.08949851989746, | |
"learning_rate": 8.222222222222223e-06, | |
"loss": 0.0513, | |
"step": 4700 | |
}, | |
{ | |
"epoch": 2.51, | |
"eval_accuracy": 0.874, | |
"eval_f1_macro": 0.807227860552302, | |
"eval_f1_micro": 0.874, | |
"eval_loss": 0.6078110933303833, | |
"eval_runtime": 125.4439, | |
"eval_samples_per_second": 79.717, | |
"eval_steps_per_second": 2.495, | |
"step": 4700 | |
}, | |
{ | |
"epoch": 2.51, | |
"grad_norm": 2.959608793258667, | |
"learning_rate": 8.133333333333332e-06, | |
"loss": 0.0433, | |
"step": 4710 | |
}, | |
{ | |
"epoch": 2.52, | |
"grad_norm": 17.176597595214844, | |
"learning_rate": 8.044444444444444e-06, | |
"loss": 0.0431, | |
"step": 4720 | |
}, | |
{ | |
"epoch": 2.52, | |
"grad_norm": 0.21210907399654388, | |
"learning_rate": 7.955555555555557e-06, | |
"loss": 0.0382, | |
"step": 4730 | |
}, | |
{ | |
"epoch": 2.53, | |
"grad_norm": 19.490718841552734, | |
"learning_rate": 7.866666666666667e-06, | |
"loss": 0.0335, | |
"step": 4740 | |
}, | |
{ | |
"epoch": 2.53, | |
"grad_norm": 9.648574829101562, | |
"learning_rate": 7.777777777777777e-06, | |
"loss": 0.0401, | |
"step": 4750 | |
}, | |
{ | |
"epoch": 2.54, | |
"grad_norm": 17.405658721923828, | |
"learning_rate": 7.68888888888889e-06, | |
"loss": 0.0642, | |
"step": 4760 | |
}, | |
{ | |
"epoch": 2.54, | |
"grad_norm": 3.015732526779175, | |
"learning_rate": 7.6e-06, | |
"loss": 0.0464, | |
"step": 4770 | |
}, | |
{ | |
"epoch": 2.55, | |
"grad_norm": 11.392130851745605, | |
"learning_rate": 7.511111111111112e-06, | |
"loss": 0.0541, | |
"step": 4780 | |
}, | |
{ | |
"epoch": 2.55, | |
"grad_norm": 19.27250862121582, | |
"learning_rate": 7.422222222222222e-06, | |
"loss": 0.0577, | |
"step": 4790 | |
}, | |
{ | |
"epoch": 2.56, | |
"grad_norm": 9.552457809448242, | |
"learning_rate": 7.333333333333334e-06, | |
"loss": 0.0553, | |
"step": 4800 | |
}, | |
{ | |
"epoch": 2.56, | |
"eval_accuracy": 0.8734, | |
"eval_f1_macro": 0.8022937956798373, | |
"eval_f1_micro": 0.8734, | |
"eval_loss": 0.6020609140396118, | |
"eval_runtime": 125.3761, | |
"eval_samples_per_second": 79.76, | |
"eval_steps_per_second": 2.496, | |
"step": 4800 | |
}, | |
{ | |
"epoch": 2.57, | |
"grad_norm": 24.78680419921875, | |
"learning_rate": 7.244444444444445e-06, | |
"loss": 0.0884, | |
"step": 4810 | |
}, | |
{ | |
"epoch": 2.57, | |
"grad_norm": 12.720375061035156, | |
"learning_rate": 7.155555555555556e-06, | |
"loss": 0.0457, | |
"step": 4820 | |
}, | |
{ | |
"epoch": 2.58, | |
"grad_norm": 11.253555297851562, | |
"learning_rate": 7.066666666666667e-06, | |
"loss": 0.0444, | |
"step": 4830 | |
}, | |
{ | |
"epoch": 2.58, | |
"grad_norm": 9.232181549072266, | |
"learning_rate": 6.9777777777777775e-06, | |
"loss": 0.0632, | |
"step": 4840 | |
}, | |
{ | |
"epoch": 2.59, | |
"grad_norm": 15.697772979736328, | |
"learning_rate": 6.888888888888889e-06, | |
"loss": 0.0551, | |
"step": 4850 | |
}, | |
{ | |
"epoch": 2.59, | |
"grad_norm": 8.457032203674316, | |
"learning_rate": 6.800000000000001e-06, | |
"loss": 0.022, | |
"step": 4860 | |
}, | |
{ | |
"epoch": 2.6, | |
"grad_norm": 18.339475631713867, | |
"learning_rate": 6.711111111111111e-06, | |
"loss": 0.0325, | |
"step": 4870 | |
}, | |
{ | |
"epoch": 2.6, | |
"grad_norm": 12.341594696044922, | |
"learning_rate": 6.622222222222223e-06, | |
"loss": 0.0359, | |
"step": 4880 | |
}, | |
{ | |
"epoch": 2.61, | |
"grad_norm": 0.5485441088676453, | |
"learning_rate": 6.533333333333333e-06, | |
"loss": 0.0223, | |
"step": 4890 | |
}, | |
{ | |
"epoch": 2.61, | |
"grad_norm": 1.2296168804168701, | |
"learning_rate": 6.4444444444444445e-06, | |
"loss": 0.0843, | |
"step": 4900 | |
}, | |
{ | |
"epoch": 2.61, | |
"eval_accuracy": 0.8766, | |
"eval_f1_macro": 0.8096414854477127, | |
"eval_f1_micro": 0.8766, | |
"eval_loss": 0.6084060668945312, | |
"eval_runtime": 125.3934, | |
"eval_samples_per_second": 79.749, | |
"eval_steps_per_second": 2.496, | |
"step": 4900 | |
}, | |
{ | |
"epoch": 2.62, | |
"grad_norm": 15.707212448120117, | |
"learning_rate": 6.355555555555557e-06, | |
"loss": 0.0422, | |
"step": 4910 | |
}, | |
{ | |
"epoch": 2.62, | |
"grad_norm": 8.062418937683105, | |
"learning_rate": 6.266666666666666e-06, | |
"loss": 0.0301, | |
"step": 4920 | |
}, | |
{ | |
"epoch": 2.63, | |
"grad_norm": 1.2240488529205322, | |
"learning_rate": 6.177777777777778e-06, | |
"loss": 0.033, | |
"step": 4930 | |
}, | |
{ | |
"epoch": 2.63, | |
"grad_norm": 2.520549774169922, | |
"learning_rate": 6.088888888888889e-06, | |
"loss": 0.0305, | |
"step": 4940 | |
}, | |
{ | |
"epoch": 2.64, | |
"grad_norm": 5.890824794769287, | |
"learning_rate": 6e-06, | |
"loss": 0.0265, | |
"step": 4950 | |
}, | |
{ | |
"epoch": 2.65, | |
"grad_norm": 13.103384971618652, | |
"learning_rate": 5.9111111111111115e-06, | |
"loss": 0.0311, | |
"step": 4960 | |
}, | |
{ | |
"epoch": 2.65, | |
"grad_norm": 20.12167739868164, | |
"learning_rate": 5.822222222222223e-06, | |
"loss": 0.0631, | |
"step": 4970 | |
}, | |
{ | |
"epoch": 2.66, | |
"grad_norm": 1.561964750289917, | |
"learning_rate": 5.733333333333333e-06, | |
"loss": 0.0279, | |
"step": 4980 | |
}, | |
{ | |
"epoch": 2.66, | |
"grad_norm": 4.472821235656738, | |
"learning_rate": 5.6444444444444445e-06, | |
"loss": 0.0338, | |
"step": 4990 | |
}, | |
{ | |
"epoch": 2.67, | |
"grad_norm": 13.154914855957031, | |
"learning_rate": 5.555555555555556e-06, | |
"loss": 0.0361, | |
"step": 5000 | |
}, | |
{ | |
"epoch": 2.67, | |
"eval_accuracy": 0.8764, | |
"eval_f1_macro": 0.8090536790170773, | |
"eval_f1_micro": 0.8764, | |
"eval_loss": 0.612912118434906, | |
"eval_runtime": 125.3911, | |
"eval_samples_per_second": 79.75, | |
"eval_steps_per_second": 2.496, | |
"step": 5000 | |
}, | |
{ | |
"epoch": 2.67, | |
"grad_norm": 20.26685905456543, | |
"learning_rate": 5.466666666666667e-06, | |
"loss": 0.0539, | |
"step": 5010 | |
}, | |
{ | |
"epoch": 2.68, | |
"grad_norm": 13.797248840332031, | |
"learning_rate": 5.3777777777777784e-06, | |
"loss": 0.0264, | |
"step": 5020 | |
}, | |
{ | |
"epoch": 2.68, | |
"grad_norm": 0.8838375806808472, | |
"learning_rate": 5.288888888888889e-06, | |
"loss": 0.0275, | |
"step": 5030 | |
}, | |
{ | |
"epoch": 2.69, | |
"grad_norm": 0.7623136639595032, | |
"learning_rate": 5.2e-06, | |
"loss": 0.0577, | |
"step": 5040 | |
}, | |
{ | |
"epoch": 2.69, | |
"grad_norm": 2.851203680038452, | |
"learning_rate": 5.1111111111111115e-06, | |
"loss": 0.0342, | |
"step": 5050 | |
}, | |
{ | |
"epoch": 2.7, | |
"grad_norm": 16.084320068359375, | |
"learning_rate": 5.022222222222223e-06, | |
"loss": 0.0355, | |
"step": 5060 | |
}, | |
{ | |
"epoch": 2.7, | |
"grad_norm": 1.148029088973999, | |
"learning_rate": 4.933333333333333e-06, | |
"loss": 0.0349, | |
"step": 5070 | |
}, | |
{ | |
"epoch": 2.71, | |
"grad_norm": 1.4330673217773438, | |
"learning_rate": 4.8444444444444446e-06, | |
"loss": 0.0557, | |
"step": 5080 | |
}, | |
{ | |
"epoch": 2.71, | |
"grad_norm": 1.7061699628829956, | |
"learning_rate": 4.755555555555556e-06, | |
"loss": 0.0247, | |
"step": 5090 | |
}, | |
{ | |
"epoch": 2.72, | |
"grad_norm": 6.559273719787598, | |
"learning_rate": 4.666666666666667e-06, | |
"loss": 0.0485, | |
"step": 5100 | |
}, | |
{ | |
"epoch": 2.72, | |
"eval_accuracy": 0.8789, | |
"eval_f1_macro": 0.8096045063363193, | |
"eval_f1_micro": 0.8789, | |
"eval_loss": 0.6214478611946106, | |
"eval_runtime": 125.3994, | |
"eval_samples_per_second": 79.745, | |
"eval_steps_per_second": 2.496, | |
"step": 5100 | |
}, | |
{ | |
"epoch": 2.73, | |
"grad_norm": 14.104253768920898, | |
"learning_rate": 4.5777777777777785e-06, | |
"loss": 0.0492, | |
"step": 5110 | |
}, | |
{ | |
"epoch": 2.73, | |
"grad_norm": 4.382899284362793, | |
"learning_rate": 4.488888888888889e-06, | |
"loss": 0.0387, | |
"step": 5120 | |
}, | |
{ | |
"epoch": 2.74, | |
"grad_norm": 7.168239593505859, | |
"learning_rate": 4.4e-06, | |
"loss": 0.02, | |
"step": 5130 | |
}, | |
{ | |
"epoch": 2.74, | |
"grad_norm": 0.057392317801713943, | |
"learning_rate": 4.3111111111111115e-06, | |
"loss": 0.0673, | |
"step": 5140 | |
}, | |
{ | |
"epoch": 2.75, | |
"grad_norm": 8.1379976272583, | |
"learning_rate": 4.222222222222223e-06, | |
"loss": 0.043, | |
"step": 5150 | |
}, | |
{ | |
"epoch": 2.75, | |
"grad_norm": 6.515874862670898, | |
"learning_rate": 4.133333333333333e-06, | |
"loss": 0.1161, | |
"step": 5160 | |
}, | |
{ | |
"epoch": 2.76, | |
"grad_norm": 8.1536226272583, | |
"learning_rate": 4.044444444444445e-06, | |
"loss": 0.0377, | |
"step": 5170 | |
}, | |
{ | |
"epoch": 2.76, | |
"grad_norm": 1.2644639015197754, | |
"learning_rate": 3.955555555555555e-06, | |
"loss": 0.0544, | |
"step": 5180 | |
}, | |
{ | |
"epoch": 2.77, | |
"grad_norm": 2.3972737789154053, | |
"learning_rate": 3.866666666666667e-06, | |
"loss": 0.0341, | |
"step": 5190 | |
}, | |
{ | |
"epoch": 2.77, | |
"grad_norm": 3.3570358753204346, | |
"learning_rate": 3.777777777777778e-06, | |
"loss": 0.0209, | |
"step": 5200 | |
}, | |
{ | |
"epoch": 2.77, | |
"eval_accuracy": 0.8795, | |
"eval_f1_macro": 0.810206057447178, | |
"eval_f1_micro": 0.8795, | |
"eval_loss": 0.5887149572372437, | |
"eval_runtime": 125.3769, | |
"eval_samples_per_second": 79.759, | |
"eval_steps_per_second": 2.496, | |
"step": 5200 | |
}, | |
{ | |
"epoch": 2.78, | |
"grad_norm": 0.38933780789375305, | |
"learning_rate": 3.688888888888889e-06, | |
"loss": 0.0141, | |
"step": 5210 | |
}, | |
{ | |
"epoch": 2.78, | |
"grad_norm": 3.201914072036743, | |
"learning_rate": 3.6e-06, | |
"loss": 0.0249, | |
"step": 5220 | |
}, | |
{ | |
"epoch": 2.79, | |
"grad_norm": 14.43453598022461, | |
"learning_rate": 3.5111111111111116e-06, | |
"loss": 0.0292, | |
"step": 5230 | |
}, | |
{ | |
"epoch": 2.79, | |
"grad_norm": 2.8055851459503174, | |
"learning_rate": 3.4222222222222224e-06, | |
"loss": 0.0317, | |
"step": 5240 | |
}, | |
{ | |
"epoch": 2.8, | |
"grad_norm": 5.319042682647705, | |
"learning_rate": 3.3333333333333333e-06, | |
"loss": 0.0527, | |
"step": 5250 | |
}, | |
{ | |
"epoch": 2.81, | |
"grad_norm": 2.3870460987091064, | |
"learning_rate": 3.244444444444444e-06, | |
"loss": 0.0263, | |
"step": 5260 | |
}, | |
{ | |
"epoch": 2.81, | |
"grad_norm": 17.512996673583984, | |
"learning_rate": 3.155555555555556e-06, | |
"loss": 0.0546, | |
"step": 5270 | |
}, | |
{ | |
"epoch": 2.82, | |
"grad_norm": 12.532732009887695, | |
"learning_rate": 3.066666666666667e-06, | |
"loss": 0.0642, | |
"step": 5280 | |
}, | |
{ | |
"epoch": 2.82, | |
"grad_norm": 7.901776313781738, | |
"learning_rate": 2.977777777777778e-06, | |
"loss": 0.0471, | |
"step": 5290 | |
}, | |
{ | |
"epoch": 2.83, | |
"grad_norm": 0.39620164036750793, | |
"learning_rate": 2.888888888888889e-06, | |
"loss": 0.028, | |
"step": 5300 | |
}, | |
{ | |
"epoch": 2.83, | |
"eval_accuracy": 0.8798, | |
"eval_f1_macro": 0.8131591273209479, | |
"eval_f1_micro": 0.8798, | |
"eval_loss": 0.5953407883644104, | |
"eval_runtime": 125.3854, | |
"eval_samples_per_second": 79.754, | |
"eval_steps_per_second": 2.496, | |
"step": 5300 | |
}, | |
{ | |
"epoch": 2.83, | |
"grad_norm": 1.7311279773712158, | |
"learning_rate": 2.8000000000000003e-06, | |
"loss": 0.0613, | |
"step": 5310 | |
}, | |
{ | |
"epoch": 2.84, | |
"grad_norm": 6.2115583419799805, | |
"learning_rate": 2.711111111111111e-06, | |
"loss": 0.0229, | |
"step": 5320 | |
}, | |
{ | |
"epoch": 2.84, | |
"grad_norm": 1.4274439811706543, | |
"learning_rate": 2.6222222222222225e-06, | |
"loss": 0.04, | |
"step": 5330 | |
}, | |
{ | |
"epoch": 2.85, | |
"grad_norm": 0.795637309551239, | |
"learning_rate": 2.5333333333333334e-06, | |
"loss": 0.0469, | |
"step": 5340 | |
}, | |
{ | |
"epoch": 2.85, | |
"grad_norm": 1.7568246126174927, | |
"learning_rate": 2.4444444444444447e-06, | |
"loss": 0.038, | |
"step": 5350 | |
}, | |
{ | |
"epoch": 2.86, | |
"grad_norm": 14.119540214538574, | |
"learning_rate": 2.3555555555555555e-06, | |
"loss": 0.0252, | |
"step": 5360 | |
}, | |
{ | |
"epoch": 2.86, | |
"grad_norm": 0.14776268601417542, | |
"learning_rate": 2.266666666666667e-06, | |
"loss": 0.034, | |
"step": 5370 | |
}, | |
{ | |
"epoch": 2.87, | |
"grad_norm": 14.7776460647583, | |
"learning_rate": 2.1777777777777777e-06, | |
"loss": 0.0343, | |
"step": 5380 | |
}, | |
{ | |
"epoch": 2.87, | |
"grad_norm": 19.50803565979004, | |
"learning_rate": 2.088888888888889e-06, | |
"loss": 0.0459, | |
"step": 5390 | |
}, | |
{ | |
"epoch": 2.88, | |
"grad_norm": 11.52786636352539, | |
"learning_rate": 2.0000000000000003e-06, | |
"loss": 0.0513, | |
"step": 5400 | |
}, | |
{ | |
"epoch": 2.88, | |
"eval_accuracy": 0.8818, | |
"eval_f1_macro": 0.8154458340510421, | |
"eval_f1_micro": 0.8818, | |
"eval_loss": 0.5944418907165527, | |
"eval_runtime": 125.3484, | |
"eval_samples_per_second": 79.778, | |
"eval_steps_per_second": 2.497, | |
"step": 5400 | |
}, | |
{ | |
"epoch": 2.89, | |
"grad_norm": 0.1808658242225647, | |
"learning_rate": 1.9111111111111112e-06, | |
"loss": 0.0206, | |
"step": 5410 | |
}, | |
{ | |
"epoch": 2.89, | |
"grad_norm": 0.8445340394973755, | |
"learning_rate": 1.8222222222222223e-06, | |
"loss": 0.0221, | |
"step": 5420 | |
}, | |
{ | |
"epoch": 2.9, | |
"grad_norm": 4.306654453277588, | |
"learning_rate": 1.7333333333333334e-06, | |
"loss": 0.032, | |
"step": 5430 | |
}, | |
{ | |
"epoch": 2.9, | |
"grad_norm": 3.2306783199310303, | |
"learning_rate": 1.6444444444444447e-06, | |
"loss": 0.0204, | |
"step": 5440 | |
}, | |
{ | |
"epoch": 2.91, | |
"grad_norm": 6.173311233520508, | |
"learning_rate": 1.5555555555555556e-06, | |
"loss": 0.0538, | |
"step": 5450 | |
}, | |
{ | |
"epoch": 2.91, | |
"grad_norm": 4.854103088378906, | |
"learning_rate": 1.4666666666666667e-06, | |
"loss": 0.0197, | |
"step": 5460 | |
}, | |
{ | |
"epoch": 2.92, | |
"grad_norm": 14.71440315246582, | |
"learning_rate": 1.3777777777777778e-06, | |
"loss": 0.0372, | |
"step": 5470 | |
}, | |
{ | |
"epoch": 2.92, | |
"grad_norm": 14.649943351745605, | |
"learning_rate": 1.2888888888888889e-06, | |
"loss": 0.0193, | |
"step": 5480 | |
}, | |
{ | |
"epoch": 2.93, | |
"grad_norm": 9.308882713317871, | |
"learning_rate": 1.2000000000000002e-06, | |
"loss": 0.019, | |
"step": 5490 | |
}, | |
{ | |
"epoch": 2.93, | |
"grad_norm": 0.6163176894187927, | |
"learning_rate": 1.1111111111111112e-06, | |
"loss": 0.0073, | |
"step": 5500 | |
}, | |
{ | |
"epoch": 2.93, | |
"eval_accuracy": 0.8794, | |
"eval_f1_macro": 0.8136104418639312, | |
"eval_f1_micro": 0.8794, | |
"eval_loss": 0.6020895838737488, | |
"eval_runtime": 125.3554, | |
"eval_samples_per_second": 79.773, | |
"eval_steps_per_second": 2.497, | |
"step": 5500 | |
}, | |
{ | |
"epoch": 2.94, | |
"grad_norm": 1.0901674032211304, | |
"learning_rate": 1.0222222222222223e-06, | |
"loss": 0.0336, | |
"step": 5510 | |
}, | |
{ | |
"epoch": 2.94, | |
"grad_norm": 0.9365220665931702, | |
"learning_rate": 9.333333333333334e-07, | |
"loss": 0.0096, | |
"step": 5520 | |
}, | |
{ | |
"epoch": 2.95, | |
"grad_norm": 17.01777458190918, | |
"learning_rate": 8.444444444444444e-07, | |
"loss": 0.0689, | |
"step": 5530 | |
}, | |
{ | |
"epoch": 2.95, | |
"grad_norm": 14.930281639099121, | |
"learning_rate": 7.555555555555556e-07, | |
"loss": 0.0414, | |
"step": 5540 | |
}, | |
{ | |
"epoch": 2.96, | |
"grad_norm": 1.0289640426635742, | |
"learning_rate": 6.666666666666667e-07, | |
"loss": 0.0151, | |
"step": 5550 | |
}, | |
{ | |
"epoch": 2.97, | |
"grad_norm": 11.737308502197266, | |
"learning_rate": 5.777777777777778e-07, | |
"loss": 0.0237, | |
"step": 5560 | |
}, | |
{ | |
"epoch": 2.97, | |
"grad_norm": 14.464305877685547, | |
"learning_rate": 4.888888888888889e-07, | |
"loss": 0.0249, | |
"step": 5570 | |
}, | |
{ | |
"epoch": 2.98, | |
"grad_norm": 1.2712557315826416, | |
"learning_rate": 4.0000000000000003e-07, | |
"loss": 0.026, | |
"step": 5580 | |
}, | |
{ | |
"epoch": 2.98, | |
"grad_norm": 4.292672634124756, | |
"learning_rate": 3.111111111111111e-07, | |
"loss": 0.026, | |
"step": 5590 | |
}, | |
{ | |
"epoch": 2.99, | |
"grad_norm": 0.36005184054374695, | |
"learning_rate": 2.2222222222222224e-07, | |
"loss": 0.0398, | |
"step": 5600 | |
}, | |
{ | |
"epoch": 2.99, | |
"eval_accuracy": 0.88, | |
"eval_f1_macro": 0.8123656641031487, | |
"eval_f1_micro": 0.88, | |
"eval_loss": 0.6064128875732422, | |
"eval_runtime": 125.3778, | |
"eval_samples_per_second": 79.759, | |
"eval_steps_per_second": 2.496, | |
"step": 5600 | |
}, | |
{ | |
"epoch": 2.99, | |
"grad_norm": 1.3966044187545776, | |
"learning_rate": 1.3333333333333334e-07, | |
"loss": 0.0279, | |
"step": 5610 | |
}, | |
{ | |
"epoch": 3.0, | |
"grad_norm": 2.1456825733184814, | |
"learning_rate": 4.444444444444445e-08, | |
"loss": 0.0397, | |
"step": 5620 | |
}, | |
{ | |
"epoch": 3.0, | |
"step": 5625, | |
"total_flos": 9.39658749935616e+17, | |
"train_loss": 0.44225363720787897, | |
"train_runtime": 13430.0481, | |
"train_samples_per_second": 13.403, | |
"train_steps_per_second": 0.419 | |
} | |
], | |
"logging_steps": 10, | |
"max_steps": 5625, | |
"num_input_tokens_seen": 0, | |
"num_train_epochs": 3, | |
"save_steps": 100, | |
"total_flos": 9.39658749935616e+17, | |
"train_batch_size": 16, | |
"trial_name": null, | |
"trial_params": null | |
} | |