|
{ |
|
"best_metric": 0.9275591969490051, |
|
"best_model_checkpoint": "../../experiments_checkpoints/MAdAiLab/google_t5/t5_base_patent/checkpoint-1550", |
|
"epoch": 1.9820971867007673, |
|
"eval_steps": 50, |
|
"global_step": 1550, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.4974615573883057, |
|
"learning_rate": 0.0004978687127024723, |
|
"loss": 2.0423, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.5172781944274902, |
|
"learning_rate": 0.0004957374254049446, |
|
"loss": 1.9857, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.7396634817123413, |
|
"learning_rate": 0.0004936061381074169, |
|
"loss": 1.7715, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 2.822936773300171, |
|
"learning_rate": 0.0004914748508098892, |
|
"loss": 1.4156, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 2.627573013305664, |
|
"learning_rate": 0.0004893435635123615, |
|
"loss": 1.3522, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_accuracy": 0.5254, |
|
"eval_f1_macro": 0.36088387014645285, |
|
"eval_f1_micro": 0.5254, |
|
"eval_loss": 1.4201537370681763, |
|
"eval_runtime": 12.9425, |
|
"eval_samples_per_second": 386.324, |
|
"eval_steps_per_second": 12.131, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 2.4126358032226562, |
|
"learning_rate": 0.00048721227621483377, |
|
"loss": 1.2477, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.7662773132324219, |
|
"learning_rate": 0.00048508098891730605, |
|
"loss": 1.2976, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 2.7678210735321045, |
|
"learning_rate": 0.0004829497016197784, |
|
"loss": 1.2331, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 2.4279048442840576, |
|
"learning_rate": 0.00048081841432225065, |
|
"loss": 1.2387, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.8737666606903076, |
|
"learning_rate": 0.0004786871270247229, |
|
"loss": 1.1693, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_accuracy": 0.597, |
|
"eval_f1_macro": 0.4695490866463984, |
|
"eval_f1_micro": 0.597, |
|
"eval_loss": 1.1673717498779297, |
|
"eval_runtime": 13.0292, |
|
"eval_samples_per_second": 383.754, |
|
"eval_steps_per_second": 12.05, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 2.6654303073883057, |
|
"learning_rate": 0.00047655583972719526, |
|
"loss": 1.2759, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 2.226668119430542, |
|
"learning_rate": 0.00047442455242966753, |
|
"loss": 1.2904, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 2.8846795558929443, |
|
"learning_rate": 0.0004722932651321398, |
|
"loss": 1.1975, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 2.5968539714813232, |
|
"learning_rate": 0.00047016197783461214, |
|
"loss": 1.2166, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.5721819400787354, |
|
"learning_rate": 0.0004680306905370844, |
|
"loss": 1.171, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_accuracy": 0.6052, |
|
"eval_f1_macro": 0.4713385676807627, |
|
"eval_f1_micro": 0.6052, |
|
"eval_loss": 1.1373018026351929, |
|
"eval_runtime": 13.0547, |
|
"eval_samples_per_second": 383.005, |
|
"eval_steps_per_second": 12.026, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 2.7147936820983887, |
|
"learning_rate": 0.0004658994032395567, |
|
"loss": 1.3234, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.9919277429580688, |
|
"learning_rate": 0.000463768115942029, |
|
"loss": 1.2193, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.9794032573699951, |
|
"learning_rate": 0.0004616368286445013, |
|
"loss": 1.1321, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 2.1101603507995605, |
|
"learning_rate": 0.00045950554134697357, |
|
"loss": 1.2112, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 2.165870428085327, |
|
"learning_rate": 0.0004573742540494459, |
|
"loss": 1.048, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_accuracy": 0.6286, |
|
"eval_f1_macro": 0.5499463679348023, |
|
"eval_f1_micro": 0.6286, |
|
"eval_loss": 1.0826044082641602, |
|
"eval_runtime": 13.0087, |
|
"eval_samples_per_second": 384.358, |
|
"eval_steps_per_second": 12.069, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 2.1708426475524902, |
|
"learning_rate": 0.0004552429667519182, |
|
"loss": 1.089, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 2.311868190765381, |
|
"learning_rate": 0.00045311167945439045, |
|
"loss": 1.1744, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 2.2947585582733154, |
|
"learning_rate": 0.0004509803921568628, |
|
"loss": 1.0987, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 2.115337371826172, |
|
"learning_rate": 0.00044884910485933505, |
|
"loss": 1.1407, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 2.1188411712646484, |
|
"learning_rate": 0.00044671781756180733, |
|
"loss": 0.9991, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_accuracy": 0.638, |
|
"eval_f1_macro": 0.5422074962269037, |
|
"eval_f1_micro": 0.638, |
|
"eval_loss": 1.0598992109298706, |
|
"eval_runtime": 13.0177, |
|
"eval_samples_per_second": 384.093, |
|
"eval_steps_per_second": 12.061, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 2.213646411895752, |
|
"learning_rate": 0.00044458653026427966, |
|
"loss": 1.0853, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 2.5469391345977783, |
|
"learning_rate": 0.00044245524296675193, |
|
"loss": 1.0904, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 2.0771143436431885, |
|
"learning_rate": 0.0004403239556692242, |
|
"loss": 1.04, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 5.050015926361084, |
|
"learning_rate": 0.00043819266837169654, |
|
"loss": 1.1204, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 2.272465705871582, |
|
"learning_rate": 0.0004360613810741688, |
|
"loss": 1.1814, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_accuracy": 0.6332, |
|
"eval_f1_macro": 0.559347721629516, |
|
"eval_f1_micro": 0.6332, |
|
"eval_loss": 1.0633331537246704, |
|
"eval_runtime": 13.0185, |
|
"eval_samples_per_second": 384.069, |
|
"eval_steps_per_second": 12.06, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 2.2277400493621826, |
|
"learning_rate": 0.0004339300937766411, |
|
"loss": 1.1177, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 1.6775270700454712, |
|
"learning_rate": 0.0004317988064791134, |
|
"loss": 0.9599, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 2.208353042602539, |
|
"learning_rate": 0.0004296675191815857, |
|
"loss": 1.1534, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1.7430713176727295, |
|
"learning_rate": 0.00042753623188405797, |
|
"loss": 1.1505, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 2.3776443004608154, |
|
"learning_rate": 0.0004254049445865303, |
|
"loss": 1.0864, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_accuracy": 0.6392, |
|
"eval_f1_macro": 0.5678118184018813, |
|
"eval_f1_micro": 0.6392, |
|
"eval_loss": 1.040041208267212, |
|
"eval_runtime": 12.9954, |
|
"eval_samples_per_second": 384.752, |
|
"eval_steps_per_second": 12.081, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 1.8053480386734009, |
|
"learning_rate": 0.0004232736572890026, |
|
"loss": 1.0725, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 1.8648344278335571, |
|
"learning_rate": 0.00042114236999147485, |
|
"loss": 1.1635, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 1.467507243156433, |
|
"learning_rate": 0.0004190110826939472, |
|
"loss": 1.1063, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 2.082329750061035, |
|
"learning_rate": 0.00041687979539641946, |
|
"loss": 1.176, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 1.4149022102355957, |
|
"learning_rate": 0.00041474850809889173, |
|
"loss": 0.9748, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"eval_accuracy": 0.6424, |
|
"eval_f1_macro": 0.5612605717682496, |
|
"eval_f1_micro": 0.6424, |
|
"eval_loss": 1.0440382957458496, |
|
"eval_runtime": 13.0068, |
|
"eval_samples_per_second": 384.415, |
|
"eval_steps_per_second": 12.071, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 2.5410144329071045, |
|
"learning_rate": 0.00041261722080136406, |
|
"loss": 1.1015, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 1.6288481950759888, |
|
"learning_rate": 0.00041048593350383634, |
|
"loss": 1.1052, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 2.03472638130188, |
|
"learning_rate": 0.0004083546462063086, |
|
"loss": 1.073, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 1.3407670259475708, |
|
"learning_rate": 0.00040622335890878094, |
|
"loss": 1.0857, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 2.194918155670166, |
|
"learning_rate": 0.0004040920716112532, |
|
"loss": 1.0267, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"eval_accuracy": 0.6526, |
|
"eval_f1_macro": 0.5818441324147113, |
|
"eval_f1_micro": 0.6526, |
|
"eval_loss": 1.01156485080719, |
|
"eval_runtime": 13.0045, |
|
"eval_samples_per_second": 384.482, |
|
"eval_steps_per_second": 12.073, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 2.4339776039123535, |
|
"learning_rate": 0.0004019607843137255, |
|
"loss": 1.0199, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 1.824545979499817, |
|
"learning_rate": 0.0003998294970161978, |
|
"loss": 1.0241, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 1.3973139524459839, |
|
"learning_rate": 0.0003976982097186701, |
|
"loss": 1.0537, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 1.7667598724365234, |
|
"learning_rate": 0.00039556692242114237, |
|
"loss": 1.1928, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 1.8740298748016357, |
|
"learning_rate": 0.0003934356351236147, |
|
"loss": 1.0052, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"eval_accuracy": 0.657, |
|
"eval_f1_macro": 0.5787057373173895, |
|
"eval_f1_micro": 0.657, |
|
"eval_loss": 0.9947913289070129, |
|
"eval_runtime": 13.0161, |
|
"eval_samples_per_second": 384.139, |
|
"eval_steps_per_second": 12.062, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 2.1658332347869873, |
|
"learning_rate": 0.000391304347826087, |
|
"loss": 1.0857, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 1.7591208219528198, |
|
"learning_rate": 0.00038917306052855925, |
|
"loss": 0.9992, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 2.5212056636810303, |
|
"learning_rate": 0.0003870417732310316, |
|
"loss": 1.0708, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 1.8266304731369019, |
|
"learning_rate": 0.00038491048593350386, |
|
"loss": 1.0144, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 2.446922540664673, |
|
"learning_rate": 0.00038277919863597613, |
|
"loss": 0.9244, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"eval_accuracy": 0.657, |
|
"eval_f1_macro": 0.5869643883356389, |
|
"eval_f1_micro": 0.657, |
|
"eval_loss": 1.000241994857788, |
|
"eval_runtime": 12.9976, |
|
"eval_samples_per_second": 384.686, |
|
"eval_steps_per_second": 12.079, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 2.180079936981201, |
|
"learning_rate": 0.00038064791133844846, |
|
"loss": 0.9825, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 2.3864924907684326, |
|
"learning_rate": 0.00037851662404092074, |
|
"loss": 0.9943, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 1.930967092514038, |
|
"learning_rate": 0.000376385336743393, |
|
"loss": 0.9622, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 1.6336820125579834, |
|
"learning_rate": 0.00037425404944586534, |
|
"loss": 1.0352, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.4508672952651978, |
|
"learning_rate": 0.0003721227621483376, |
|
"loss": 1.0172, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"eval_accuracy": 0.661, |
|
"eval_f1_macro": 0.5889167117692362, |
|
"eval_f1_micro": 0.661, |
|
"eval_loss": 0.9869310259819031, |
|
"eval_runtime": 13.005, |
|
"eval_samples_per_second": 384.466, |
|
"eval_steps_per_second": 12.072, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 2.1621804237365723, |
|
"learning_rate": 0.0003699914748508099, |
|
"loss": 1.0703, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 1.213054895401001, |
|
"learning_rate": 0.0003678601875532822, |
|
"loss": 0.9779, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 1.6633962392807007, |
|
"learning_rate": 0.0003657289002557545, |
|
"loss": 1.0609, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.6113487482070923, |
|
"learning_rate": 0.0003635976129582268, |
|
"loss": 0.9406, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 1.8522602319717407, |
|
"learning_rate": 0.0003614663256606991, |
|
"loss": 1.032, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"eval_accuracy": 0.658, |
|
"eval_f1_macro": 0.5967445966950204, |
|
"eval_f1_micro": 0.658, |
|
"eval_loss": 0.9922341108322144, |
|
"eval_runtime": 13.0065, |
|
"eval_samples_per_second": 384.423, |
|
"eval_steps_per_second": 12.071, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 2.109052896499634, |
|
"learning_rate": 0.0003593350383631714, |
|
"loss": 0.9791, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 2.2362546920776367, |
|
"learning_rate": 0.00035720375106564365, |
|
"loss": 1.0476, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1.6640779972076416, |
|
"learning_rate": 0.000355072463768116, |
|
"loss": 1.0723, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 2.161720037460327, |
|
"learning_rate": 0.00035294117647058826, |
|
"loss": 1.1059, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 1.3153353929519653, |
|
"learning_rate": 0.00035080988917306053, |
|
"loss": 0.9623, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"eval_accuracy": 0.6488, |
|
"eval_f1_macro": 0.5862935110976637, |
|
"eval_f1_micro": 0.6488, |
|
"eval_loss": 0.9955071210861206, |
|
"eval_runtime": 13.0077, |
|
"eval_samples_per_second": 384.387, |
|
"eval_steps_per_second": 12.07, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 1.304362416267395, |
|
"learning_rate": 0.00034867860187553286, |
|
"loss": 1.0294, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 1.4024779796600342, |
|
"learning_rate": 0.00034654731457800514, |
|
"loss": 1.0708, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 2.148181676864624, |
|
"learning_rate": 0.0003444160272804774, |
|
"loss": 0.9677, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 2.531114101409912, |
|
"learning_rate": 0.00034228473998294974, |
|
"loss": 0.966, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 1.9326214790344238, |
|
"learning_rate": 0.000340153452685422, |
|
"loss": 0.9257, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"eval_accuracy": 0.6556, |
|
"eval_f1_macro": 0.5884168435088513, |
|
"eval_f1_micro": 0.6556, |
|
"eval_loss": 0.9993236064910889, |
|
"eval_runtime": 13.0548, |
|
"eval_samples_per_second": 383.002, |
|
"eval_steps_per_second": 12.026, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 2.2656257152557373, |
|
"learning_rate": 0.0003380221653878943, |
|
"loss": 0.991, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 5.2825469970703125, |
|
"learning_rate": 0.0003358908780903666, |
|
"loss": 1.0599, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.6359708309173584, |
|
"learning_rate": 0.0003337595907928389, |
|
"loss": 1.0755, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.8754523992538452, |
|
"learning_rate": 0.0003316283034953112, |
|
"loss": 0.8551, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 1.687317132949829, |
|
"learning_rate": 0.0003294970161977835, |
|
"loss": 0.7956, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"eval_accuracy": 0.6662, |
|
"eval_f1_macro": 0.6147575900411174, |
|
"eval_f1_micro": 0.6662, |
|
"eval_loss": 0.9737330079078674, |
|
"eval_runtime": 12.9893, |
|
"eval_samples_per_second": 384.933, |
|
"eval_steps_per_second": 12.087, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 2.0937979221343994, |
|
"learning_rate": 0.0003273657289002558, |
|
"loss": 0.8608, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 1.6474498510360718, |
|
"learning_rate": 0.00032523444160272806, |
|
"loss": 0.6793, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 2.0066182613372803, |
|
"learning_rate": 0.0003231031543052004, |
|
"loss": 0.8549, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 2.7142579555511475, |
|
"learning_rate": 0.0003209718670076726, |
|
"loss": 0.8081, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 2.2779908180236816, |
|
"learning_rate": 0.0003188405797101449, |
|
"loss": 0.8475, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"eval_accuracy": 0.6544, |
|
"eval_f1_macro": 0.5728923304855623, |
|
"eval_f1_micro": 0.6544, |
|
"eval_loss": 1.0124825239181519, |
|
"eval_runtime": 13.0224, |
|
"eval_samples_per_second": 383.955, |
|
"eval_steps_per_second": 12.056, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 2.1091928482055664, |
|
"learning_rate": 0.0003167092924126172, |
|
"loss": 0.8423, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 1.7248471975326538, |
|
"learning_rate": 0.0003145780051150895, |
|
"loss": 0.8673, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 1.7542129755020142, |
|
"learning_rate": 0.00031244671781756176, |
|
"loss": 0.7861, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 2.300936222076416, |
|
"learning_rate": 0.0003103154305200341, |
|
"loss": 0.8717, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 2.793142318725586, |
|
"learning_rate": 0.00030818414322250637, |
|
"loss": 0.8527, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"eval_accuracy": 0.6524, |
|
"eval_f1_macro": 0.5897355833668082, |
|
"eval_f1_micro": 0.6524, |
|
"eval_loss": 0.9999291300773621, |
|
"eval_runtime": 12.9931, |
|
"eval_samples_per_second": 384.819, |
|
"eval_steps_per_second": 12.083, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 1.7186204195022583, |
|
"learning_rate": 0.00030605285592497864, |
|
"loss": 0.899, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 1.7719347476959229, |
|
"learning_rate": 0.00030392156862745097, |
|
"loss": 0.8529, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 1.558523416519165, |
|
"learning_rate": 0.00030179028132992325, |
|
"loss": 0.7822, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 1.474290132522583, |
|
"learning_rate": 0.0002996589940323955, |
|
"loss": 0.7782, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 2.070176839828491, |
|
"learning_rate": 0.00029752770673486785, |
|
"loss": 0.8587, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"eval_accuracy": 0.6576, |
|
"eval_f1_macro": 0.5872663998332279, |
|
"eval_f1_micro": 0.6576, |
|
"eval_loss": 1.0072367191314697, |
|
"eval_runtime": 13.0045, |
|
"eval_samples_per_second": 384.481, |
|
"eval_steps_per_second": 12.073, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 1.9397375583648682, |
|
"learning_rate": 0.00029539641943734013, |
|
"loss": 0.7791, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 2.1534318923950195, |
|
"learning_rate": 0.0002932651321398124, |
|
"loss": 0.9695, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 1.9184337854385376, |
|
"learning_rate": 0.00029113384484228473, |
|
"loss": 0.8261, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 1.854306697845459, |
|
"learning_rate": 0.000289002557544757, |
|
"loss": 0.7905, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 2.377044200897217, |
|
"learning_rate": 0.0002868712702472293, |
|
"loss": 0.8855, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"eval_accuracy": 0.6592, |
|
"eval_f1_macro": 0.6034965481160524, |
|
"eval_f1_micro": 0.6592, |
|
"eval_loss": 0.9839973449707031, |
|
"eval_runtime": 13.0007, |
|
"eval_samples_per_second": 384.595, |
|
"eval_steps_per_second": 12.076, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 2.329939603805542, |
|
"learning_rate": 0.0002847399829497016, |
|
"loss": 0.873, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 2.4969725608825684, |
|
"learning_rate": 0.0002826086956521739, |
|
"loss": 0.8279, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 1.957000970840454, |
|
"learning_rate": 0.00028047740835464616, |
|
"loss": 0.7704, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 2.533355951309204, |
|
"learning_rate": 0.0002783461210571185, |
|
"loss": 0.8184, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 2.0572900772094727, |
|
"learning_rate": 0.00027621483375959077, |
|
"loss": 0.7015, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"eval_accuracy": 0.6682, |
|
"eval_f1_macro": 0.5993496469081598, |
|
"eval_f1_micro": 0.6682, |
|
"eval_loss": 0.9846757650375366, |
|
"eval_runtime": 13.0022, |
|
"eval_samples_per_second": 384.551, |
|
"eval_steps_per_second": 12.075, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 2.5160746574401855, |
|
"learning_rate": 0.0002740835464620631, |
|
"loss": 0.9059, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 1.8105578422546387, |
|
"learning_rate": 0.0002719522591645354, |
|
"loss": 0.9492, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 1.9851925373077393, |
|
"learning_rate": 0.00026982097186700765, |
|
"loss": 0.8391, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 1.477924108505249, |
|
"learning_rate": 0.00026768968456948, |
|
"loss": 0.8203, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 2.1417877674102783, |
|
"learning_rate": 0.00026555839727195225, |
|
"loss": 0.8116, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"eval_accuracy": 0.6678, |
|
"eval_f1_macro": 0.6079453841442157, |
|
"eval_f1_micro": 0.6678, |
|
"eval_loss": 0.9702128171920776, |
|
"eval_runtime": 12.99, |
|
"eval_samples_per_second": 384.912, |
|
"eval_steps_per_second": 12.086, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 2.5949513912200928, |
|
"learning_rate": 0.00026342710997442453, |
|
"loss": 0.8707, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 2.1411094665527344, |
|
"learning_rate": 0.00026129582267689686, |
|
"loss": 0.7374, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 3.001037836074829, |
|
"learning_rate": 0.00025916453537936913, |
|
"loss": 0.8355, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 1.8605945110321045, |
|
"learning_rate": 0.0002570332480818414, |
|
"loss": 0.8346, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 2.481954574584961, |
|
"learning_rate": 0.00025490196078431374, |
|
"loss": 0.8409, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"eval_accuracy": 0.6606, |
|
"eval_f1_macro": 0.6017292151887733, |
|
"eval_f1_micro": 0.6606, |
|
"eval_loss": 0.978935182094574, |
|
"eval_runtime": 12.9978, |
|
"eval_samples_per_second": 384.68, |
|
"eval_steps_per_second": 12.079, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 2.708108425140381, |
|
"learning_rate": 0.000252770673486786, |
|
"loss": 0.9453, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 2.0015769004821777, |
|
"learning_rate": 0.0002506393861892583, |
|
"loss": 0.9272, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 1.576611042022705, |
|
"learning_rate": 0.0002485080988917306, |
|
"loss": 0.7801, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 1.3962684869766235, |
|
"learning_rate": 0.0002463768115942029, |
|
"loss": 0.8375, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 2.231125593185425, |
|
"learning_rate": 0.0002442455242966752, |
|
"loss": 0.7889, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"eval_accuracy": 0.6818, |
|
"eval_f1_macro": 0.6125360407387105, |
|
"eval_f1_micro": 0.6818, |
|
"eval_loss": 0.9462108612060547, |
|
"eval_runtime": 13.0042, |
|
"eval_samples_per_second": 384.492, |
|
"eval_steps_per_second": 12.073, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 3.0127203464508057, |
|
"learning_rate": 0.0002421142369991475, |
|
"loss": 0.7764, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 1.7407495975494385, |
|
"learning_rate": 0.0002399829497016198, |
|
"loss": 0.9385, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 1.6206921339035034, |
|
"learning_rate": 0.00023785166240409208, |
|
"loss": 0.8436, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 1.9031857252120972, |
|
"learning_rate": 0.00023572037510656438, |
|
"loss": 0.8037, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 1.3563182353973389, |
|
"learning_rate": 0.00023358908780903668, |
|
"loss": 0.8059, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"eval_accuracy": 0.6694, |
|
"eval_f1_macro": 0.6092531381348701, |
|
"eval_f1_micro": 0.6694, |
|
"eval_loss": 0.9374752044677734, |
|
"eval_runtime": 13.0242, |
|
"eval_samples_per_second": 383.9, |
|
"eval_steps_per_second": 12.054, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 1.6967196464538574, |
|
"learning_rate": 0.00023145780051150893, |
|
"loss": 0.7743, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 2.4736151695251465, |
|
"learning_rate": 0.00022932651321398123, |
|
"loss": 0.7983, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 1.9957785606384277, |
|
"learning_rate": 0.00022719522591645354, |
|
"loss": 0.8545, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 1.7612760066986084, |
|
"learning_rate": 0.0002250639386189258, |
|
"loss": 0.7164, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 1.766614317893982, |
|
"learning_rate": 0.00022293265132139811, |
|
"loss": 0.7893, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"eval_accuracy": 0.6762, |
|
"eval_f1_macro": 0.6102136225362762, |
|
"eval_f1_micro": 0.6762, |
|
"eval_loss": 0.946709394454956, |
|
"eval_runtime": 13.0006, |
|
"eval_samples_per_second": 384.598, |
|
"eval_steps_per_second": 12.076, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 1.87382972240448, |
|
"learning_rate": 0.00022080136402387042, |
|
"loss": 0.8124, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 2.0260634422302246, |
|
"learning_rate": 0.0002186700767263427, |
|
"loss": 0.8122, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 2.4003918170928955, |
|
"learning_rate": 0.000216538789428815, |
|
"loss": 0.7986, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 1.366548776626587, |
|
"learning_rate": 0.0002144075021312873, |
|
"loss": 0.8306, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 1.6609718799591064, |
|
"learning_rate": 0.00021227621483375957, |
|
"loss": 0.8152, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"eval_accuracy": 0.6822, |
|
"eval_f1_macro": 0.6157614528042974, |
|
"eval_f1_micro": 0.6822, |
|
"eval_loss": 0.9396414160728455, |
|
"eval_runtime": 12.9993, |
|
"eval_samples_per_second": 384.637, |
|
"eval_steps_per_second": 12.078, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 1.8561837673187256, |
|
"learning_rate": 0.00021014492753623187, |
|
"loss": 0.8248, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 1.7248742580413818, |
|
"learning_rate": 0.00020801364023870418, |
|
"loss": 0.7829, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 1.480776071548462, |
|
"learning_rate": 0.00020588235294117645, |
|
"loss": 0.7657, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 1.869985580444336, |
|
"learning_rate": 0.00020375106564364876, |
|
"loss": 0.8593, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 2.7019195556640625, |
|
"learning_rate": 0.00020161977834612106, |
|
"loss": 0.7644, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"eval_accuracy": 0.6798, |
|
"eval_f1_macro": 0.619022399994399, |
|
"eval_f1_micro": 0.6798, |
|
"eval_loss": 0.9445300698280334, |
|
"eval_runtime": 13.0377, |
|
"eval_samples_per_second": 383.502, |
|
"eval_steps_per_second": 12.042, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 2.2178022861480713, |
|
"learning_rate": 0.00019948849104859333, |
|
"loss": 0.8636, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 2.1172451972961426, |
|
"learning_rate": 0.00019735720375106564, |
|
"loss": 0.7812, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 2.43135142326355, |
|
"learning_rate": 0.00019522591645353794, |
|
"loss": 0.7948, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 3.7343592643737793, |
|
"learning_rate": 0.0001930946291560102, |
|
"loss": 0.8153, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 1.6375885009765625, |
|
"learning_rate": 0.00019096334185848252, |
|
"loss": 0.7252, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"eval_accuracy": 0.688, |
|
"eval_f1_macro": 0.6208786695728835, |
|
"eval_f1_micro": 0.688, |
|
"eval_loss": 0.9284820556640625, |
|
"eval_runtime": 12.989, |
|
"eval_samples_per_second": 384.94, |
|
"eval_steps_per_second": 12.087, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 2.259469509124756, |
|
"learning_rate": 0.00018883205456095482, |
|
"loss": 0.7284, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 2.0817081928253174, |
|
"learning_rate": 0.0001867007672634271, |
|
"loss": 0.8395, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 1.963901400566101, |
|
"learning_rate": 0.0001845694799658994, |
|
"loss": 0.8532, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 2.1315038204193115, |
|
"learning_rate": 0.0001824381926683717, |
|
"loss": 0.8127, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 1.7359952926635742, |
|
"learning_rate": 0.000180306905370844, |
|
"loss": 1.0028, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"eval_accuracy": 0.6702, |
|
"eval_f1_macro": 0.6079183899918916, |
|
"eval_f1_micro": 0.6702, |
|
"eval_loss": 0.9379138946533203, |
|
"eval_runtime": 13.0102, |
|
"eval_samples_per_second": 384.315, |
|
"eval_steps_per_second": 12.067, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 1.7253365516662598, |
|
"learning_rate": 0.00017817561807331628, |
|
"loss": 0.7684, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 1.6444343328475952, |
|
"learning_rate": 0.00017604433077578858, |
|
"loss": 0.7371, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 2.163722038269043, |
|
"learning_rate": 0.00017391304347826088, |
|
"loss": 0.8029, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 1.5761432647705078, |
|
"learning_rate": 0.00017178175618073316, |
|
"loss": 0.8265, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 2.286616563796997, |
|
"learning_rate": 0.00016965046888320546, |
|
"loss": 0.8056, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"eval_accuracy": 0.6776, |
|
"eval_f1_macro": 0.6236906518711645, |
|
"eval_f1_micro": 0.6776, |
|
"eval_loss": 0.9275591969490051, |
|
"eval_runtime": 13.0046, |
|
"eval_samples_per_second": 384.479, |
|
"eval_steps_per_second": 12.073, |
|
"step": 1550 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2346, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 50, |
|
"total_flos": 7573839093432320.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|