|
{ |
|
"best_metric": 0.6829969882965088, |
|
"best_model_checkpoint": "./renovation/checkpoint-2600", |
|
"epoch": 4.0, |
|
"eval_steps": 100, |
|
"global_step": 5844, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.7498056888580322, |
|
"learning_rate": 0.0001996577686516085, |
|
"loss": 1.6691, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 4.621170997619629, |
|
"learning_rate": 0.00019931553730321699, |
|
"loss": 1.5728, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 2.670748233795166, |
|
"learning_rate": 0.00019897330595482546, |
|
"loss": 1.3843, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 2.2553956508636475, |
|
"learning_rate": 0.00019863107460643396, |
|
"loss": 1.3548, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 2.815242290496826, |
|
"learning_rate": 0.00019828884325804244, |
|
"loss": 1.2678, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 3.8017427921295166, |
|
"learning_rate": 0.00019794661190965094, |
|
"loss": 1.1998, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 2.7234277725219727, |
|
"learning_rate": 0.00019760438056125944, |
|
"loss": 1.1949, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 2.4754185676574707, |
|
"learning_rate": 0.0001972621492128679, |
|
"loss": 1.0971, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.840914249420166, |
|
"learning_rate": 0.0001969199178644764, |
|
"loss": 1.2097, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 2.4401698112487793, |
|
"learning_rate": 0.0001965776865160849, |
|
"loss": 1.0475, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_accuracy": 0.582421340629275, |
|
"eval_loss": 1.0331681966781616, |
|
"eval_runtime": 118.3782, |
|
"eval_samples_per_second": 24.701, |
|
"eval_steps_per_second": 3.092, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.5946625471115112, |
|
"learning_rate": 0.00019623545516769337, |
|
"loss": 1.0628, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 2.1281728744506836, |
|
"learning_rate": 0.00019589322381930187, |
|
"loss": 1.0603, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 3.139054298400879, |
|
"learning_rate": 0.00019555099247091034, |
|
"loss": 1.0245, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 2.660050392150879, |
|
"learning_rate": 0.00019520876112251882, |
|
"loss": 0.97, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 2.315675735473633, |
|
"learning_rate": 0.00019486652977412732, |
|
"loss": 0.97, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 2.5216097831726074, |
|
"learning_rate": 0.00019452429842573582, |
|
"loss": 0.979, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.6023298501968384, |
|
"learning_rate": 0.0001941820670773443, |
|
"loss": 1.0199, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 3.1774656772613525, |
|
"learning_rate": 0.0001938398357289528, |
|
"loss": 0.9622, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 4.399940490722656, |
|
"learning_rate": 0.00019349760438056127, |
|
"loss": 0.9323, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 3.0077695846557617, |
|
"learning_rate": 0.00019315537303216975, |
|
"loss": 0.8651, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_accuracy": 0.6203830369357045, |
|
"eval_loss": 0.9321824908256531, |
|
"eval_runtime": 99.3808, |
|
"eval_samples_per_second": 29.422, |
|
"eval_steps_per_second": 3.683, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 2.711514472961426, |
|
"learning_rate": 0.00019281314168377825, |
|
"loss": 0.9065, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 3.137427568435669, |
|
"learning_rate": 0.00019247091033538672, |
|
"loss": 0.8798, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 2.13639235496521, |
|
"learning_rate": 0.00019212867898699522, |
|
"loss": 0.9009, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 3.441429615020752, |
|
"learning_rate": 0.0001917864476386037, |
|
"loss": 0.895, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 2.388650894165039, |
|
"learning_rate": 0.00019144421629021217, |
|
"loss": 0.9888, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 3.304922103881836, |
|
"learning_rate": 0.00019110198494182067, |
|
"loss": 0.7772, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 3.080766439437866, |
|
"learning_rate": 0.00019075975359342918, |
|
"loss": 0.9646, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 2.9031128883361816, |
|
"learning_rate": 0.00019041752224503765, |
|
"loss": 0.817, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 3.220935106277466, |
|
"learning_rate": 0.00019007529089664615, |
|
"loss": 0.7329, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 3.5123424530029297, |
|
"learning_rate": 0.00018973305954825463, |
|
"loss": 1.0022, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_accuracy": 0.5147058823529411, |
|
"eval_loss": 1.2150331735610962, |
|
"eval_runtime": 99.731, |
|
"eval_samples_per_second": 29.319, |
|
"eval_steps_per_second": 3.67, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 3.335115671157837, |
|
"learning_rate": 0.0001893908281998631, |
|
"loss": 1.0487, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 3.9753947257995605, |
|
"learning_rate": 0.0001890485968514716, |
|
"loss": 0.9879, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 3.008366346359253, |
|
"learning_rate": 0.0001887063655030801, |
|
"loss": 1.063, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.9694970846176147, |
|
"learning_rate": 0.00018836413415468858, |
|
"loss": 0.9867, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 3.496166944503784, |
|
"learning_rate": 0.00018802190280629705, |
|
"loss": 0.9286, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 4.208122730255127, |
|
"learning_rate": 0.00018767967145790556, |
|
"loss": 1.2562, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 3.086482048034668, |
|
"learning_rate": 0.00018733744010951403, |
|
"loss": 0.8493, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.9667807817459106, |
|
"learning_rate": 0.00018699520876112253, |
|
"loss": 0.8258, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 5.695191860198975, |
|
"learning_rate": 0.00018665297741273103, |
|
"loss": 0.8873, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 1.56025230884552, |
|
"learning_rate": 0.00018631074606433948, |
|
"loss": 1.0636, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_accuracy": 0.625170998632011, |
|
"eval_loss": 0.9522940516471863, |
|
"eval_runtime": 97.3502, |
|
"eval_samples_per_second": 30.036, |
|
"eval_steps_per_second": 3.76, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 2.2179884910583496, |
|
"learning_rate": 0.00018596851471594798, |
|
"loss": 0.9573, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 2.3255908489227295, |
|
"learning_rate": 0.00018562628336755649, |
|
"loss": 0.9596, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 1.8122656345367432, |
|
"learning_rate": 0.00018528405201916496, |
|
"loss": 0.9204, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 2.713568687438965, |
|
"learning_rate": 0.00018494182067077346, |
|
"loss": 1.005, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 2.269219160079956, |
|
"learning_rate": 0.00018459958932238194, |
|
"loss": 0.8583, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 2.7371773719787598, |
|
"learning_rate": 0.0001842573579739904, |
|
"loss": 0.8506, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 3.109008550643921, |
|
"learning_rate": 0.0001839151266255989, |
|
"loss": 0.9721, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 3.8040051460266113, |
|
"learning_rate": 0.00018357289527720741, |
|
"loss": 0.8799, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 2.8032755851745605, |
|
"learning_rate": 0.0001832306639288159, |
|
"loss": 0.8998, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 5.312169551849365, |
|
"learning_rate": 0.0001828884325804244, |
|
"loss": 0.8311, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_accuracy": 0.655608755129959, |
|
"eval_loss": 0.8439633846282959, |
|
"eval_runtime": 97.8414, |
|
"eval_samples_per_second": 29.885, |
|
"eval_steps_per_second": 3.741, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 3.493340253829956, |
|
"learning_rate": 0.00018254620123203286, |
|
"loss": 0.8658, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 2.711038827896118, |
|
"learning_rate": 0.00018220396988364134, |
|
"loss": 0.8064, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 2.2599730491638184, |
|
"learning_rate": 0.00018186173853524984, |
|
"loss": 0.8902, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 1.6086162328720093, |
|
"learning_rate": 0.00018151950718685832, |
|
"loss": 0.7395, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 3.698824405670166, |
|
"learning_rate": 0.00018117727583846682, |
|
"loss": 0.8971, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 3.232140064239502, |
|
"learning_rate": 0.0001808350444900753, |
|
"loss": 0.8815, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 2.3350541591644287, |
|
"learning_rate": 0.0001804928131416838, |
|
"loss": 0.7075, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 2.1570167541503906, |
|
"learning_rate": 0.00018015058179329227, |
|
"loss": 0.873, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 3.0529322624206543, |
|
"learning_rate": 0.00017980835044490077, |
|
"loss": 0.8208, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 2.9495370388031006, |
|
"learning_rate": 0.00017946611909650924, |
|
"loss": 0.88, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_accuracy": 0.649452804377565, |
|
"eval_loss": 0.8707343935966492, |
|
"eval_runtime": 97.0815, |
|
"eval_samples_per_second": 30.119, |
|
"eval_steps_per_second": 3.77, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 4.997549533843994, |
|
"learning_rate": 0.00017912388774811775, |
|
"loss": 0.915, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 3.1065518856048584, |
|
"learning_rate": 0.00017878165639972622, |
|
"loss": 0.8154, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 2.30731201171875, |
|
"learning_rate": 0.0001784394250513347, |
|
"loss": 0.8252, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 3.0879149436950684, |
|
"learning_rate": 0.0001780971937029432, |
|
"loss": 0.8409, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 3.711212158203125, |
|
"learning_rate": 0.0001777549623545517, |
|
"loss": 0.7977, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 2.27301287651062, |
|
"learning_rate": 0.00017741273100616017, |
|
"loss": 0.8718, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 3.153876781463623, |
|
"learning_rate": 0.00017707049965776865, |
|
"loss": 0.8712, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 2.7592906951904297, |
|
"learning_rate": 0.00017672826830937715, |
|
"loss": 0.9281, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 6.171240329742432, |
|
"learning_rate": 0.00017638603696098562, |
|
"loss": 1.0087, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 3.861326217651367, |
|
"learning_rate": 0.00017604380561259413, |
|
"loss": 0.8881, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"eval_accuracy": 0.6333789329685362, |
|
"eval_loss": 0.8902959227561951, |
|
"eval_runtime": 98.1212, |
|
"eval_samples_per_second": 29.8, |
|
"eval_steps_per_second": 3.73, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 5.067281246185303, |
|
"learning_rate": 0.00017570157426420263, |
|
"loss": 0.8387, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 3.4782888889312744, |
|
"learning_rate": 0.00017535934291581108, |
|
"loss": 0.8257, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 3.553341865539551, |
|
"learning_rate": 0.00017501711156741958, |
|
"loss": 0.7631, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 2.3364970684051514, |
|
"learning_rate": 0.00017467488021902808, |
|
"loss": 0.8531, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 2.0276718139648438, |
|
"learning_rate": 0.00017433264887063655, |
|
"loss": 0.7493, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 5.019596576690674, |
|
"learning_rate": 0.00017399041752224506, |
|
"loss": 0.7348, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 3.7119674682617188, |
|
"learning_rate": 0.00017364818617385356, |
|
"loss": 0.8455, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 2.028977870941162, |
|
"learning_rate": 0.000173305954825462, |
|
"loss": 0.8198, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 1.9546798467636108, |
|
"learning_rate": 0.0001729637234770705, |
|
"loss": 0.754, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 1.9083542823791504, |
|
"learning_rate": 0.000172621492128679, |
|
"loss": 0.7522, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"eval_accuracy": 0.6576607387140903, |
|
"eval_loss": 0.8479247093200684, |
|
"eval_runtime": 98.2024, |
|
"eval_samples_per_second": 29.775, |
|
"eval_steps_per_second": 3.727, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 2.516714572906494, |
|
"learning_rate": 0.00017227926078028748, |
|
"loss": 0.7789, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 2.8401355743408203, |
|
"learning_rate": 0.00017193702943189598, |
|
"loss": 0.7046, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 2.477896213531494, |
|
"learning_rate": 0.00017159479808350446, |
|
"loss": 0.879, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 3.136587381362915, |
|
"learning_rate": 0.00017125256673511293, |
|
"loss": 0.8485, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 1.215061068534851, |
|
"learning_rate": 0.00017091033538672144, |
|
"loss": 0.8007, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 2.920921564102173, |
|
"learning_rate": 0.00017056810403832994, |
|
"loss": 0.8547, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 2.390735626220703, |
|
"learning_rate": 0.0001702258726899384, |
|
"loss": 0.7409, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 2.2901852130889893, |
|
"learning_rate": 0.00016988364134154689, |
|
"loss": 0.9101, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 2.337519884109497, |
|
"learning_rate": 0.0001695414099931554, |
|
"loss": 0.7682, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 2.016847848892212, |
|
"learning_rate": 0.00016919917864476386, |
|
"loss": 0.798, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"eval_accuracy": 0.6843365253077975, |
|
"eval_loss": 0.7738928198814392, |
|
"eval_runtime": 98.3162, |
|
"eval_samples_per_second": 29.741, |
|
"eval_steps_per_second": 3.723, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 2.343900203704834, |
|
"learning_rate": 0.00016885694729637236, |
|
"loss": 0.7555, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 1.9535636901855469, |
|
"learning_rate": 0.00016851471594798084, |
|
"loss": 0.8414, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 2.9519262313842773, |
|
"learning_rate": 0.00016820670773442847, |
|
"loss": 0.6694, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 2.4800071716308594, |
|
"learning_rate": 0.00016786447638603697, |
|
"loss": 0.6643, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 2.3085594177246094, |
|
"learning_rate": 0.00016752224503764547, |
|
"loss": 0.7348, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 4.01909065246582, |
|
"learning_rate": 0.00016718001368925392, |
|
"loss": 0.8339, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 1.9890522956848145, |
|
"learning_rate": 0.00016683778234086242, |
|
"loss": 0.7035, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 2.7115402221679688, |
|
"learning_rate": 0.00016649555099247093, |
|
"loss": 0.8243, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 3.829646348953247, |
|
"learning_rate": 0.0001661533196440794, |
|
"loss": 0.9374, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 2.90616774559021, |
|
"learning_rate": 0.0001658110882956879, |
|
"loss": 0.7317, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"eval_accuracy": 0.6795485636114911, |
|
"eval_loss": 0.7855507731437683, |
|
"eval_runtime": 98.3649, |
|
"eval_samples_per_second": 29.726, |
|
"eval_steps_per_second": 3.721, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 2.388803482055664, |
|
"learning_rate": 0.0001654688569472964, |
|
"loss": 0.8345, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 2.2373769283294678, |
|
"learning_rate": 0.00016512662559890485, |
|
"loss": 0.712, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 2.6666200160980225, |
|
"learning_rate": 0.00016478439425051335, |
|
"loss": 0.8304, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 2.198072910308838, |
|
"learning_rate": 0.00016444216290212185, |
|
"loss": 0.6709, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 2.4663429260253906, |
|
"learning_rate": 0.00016409993155373033, |
|
"loss": 0.8756, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 3.6292316913604736, |
|
"learning_rate": 0.00016375770020533883, |
|
"loss": 0.8577, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 2.9756243228912354, |
|
"learning_rate": 0.0001634154688569473, |
|
"loss": 0.6306, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 3.7847914695739746, |
|
"learning_rate": 0.00016307323750855578, |
|
"loss": 0.7334, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 2.1593246459960938, |
|
"learning_rate": 0.00016273100616016428, |
|
"loss": 0.7105, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 3.798877239227295, |
|
"learning_rate": 0.00016238877481177278, |
|
"loss": 0.8372, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"eval_accuracy": 0.6354309165526676, |
|
"eval_loss": 0.888414204120636, |
|
"eval_runtime": 98.4587, |
|
"eval_samples_per_second": 29.698, |
|
"eval_steps_per_second": 3.717, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 3.8040831089019775, |
|
"learning_rate": 0.00016204654346338126, |
|
"loss": 0.8219, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.5006623268127441, |
|
"learning_rate": 0.0001617385352498289, |
|
"loss": 0.683, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 2.44973087310791, |
|
"learning_rate": 0.0001613963039014374, |
|
"loss": 0.7756, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 3.5681519508361816, |
|
"learning_rate": 0.00016105407255304587, |
|
"loss": 0.9079, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 2.513470411300659, |
|
"learning_rate": 0.00016071184120465434, |
|
"loss": 0.7569, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 5.322854518890381, |
|
"learning_rate": 0.00016036960985626284, |
|
"loss": 0.8551, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 2.9448812007904053, |
|
"learning_rate": 0.00016002737850787132, |
|
"loss": 0.8706, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 3.101590394973755, |
|
"learning_rate": 0.00015968514715947982, |
|
"loss": 0.8553, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 5.033396244049072, |
|
"learning_rate": 0.00015934291581108832, |
|
"loss": 0.7715, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.8060014247894287, |
|
"learning_rate": 0.00015900068446269677, |
|
"loss": 0.6629, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"eval_accuracy": 0.6870725034199726, |
|
"eval_loss": 0.7572759389877319, |
|
"eval_runtime": 97.9392, |
|
"eval_samples_per_second": 29.855, |
|
"eval_steps_per_second": 3.737, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 3.305678129196167, |
|
"learning_rate": 0.00015865845311430527, |
|
"loss": 0.6609, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 1.3885236978530884, |
|
"learning_rate": 0.00015831622176591377, |
|
"loss": 0.8702, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 1.796901822090149, |
|
"learning_rate": 0.00015797399041752225, |
|
"loss": 0.6958, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 2.182668685913086, |
|
"learning_rate": 0.00015763175906913075, |
|
"loss": 0.7975, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 2.096924066543579, |
|
"learning_rate": 0.00015728952772073925, |
|
"loss": 0.5707, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 2.922907590866089, |
|
"learning_rate": 0.0001569472963723477, |
|
"loss": 0.6806, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1.7715729475021362, |
|
"learning_rate": 0.0001566050650239562, |
|
"loss": 0.7296, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 6.666139602661133, |
|
"learning_rate": 0.0001562628336755647, |
|
"loss": 0.6637, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 2.5472826957702637, |
|
"learning_rate": 0.00015592060232717318, |
|
"loss": 0.7683, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 3.5583953857421875, |
|
"learning_rate": 0.00015557837097878168, |
|
"loss": 0.7767, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"eval_accuracy": 0.686046511627907, |
|
"eval_loss": 0.7543095946311951, |
|
"eval_runtime": 97.3791, |
|
"eval_samples_per_second": 30.027, |
|
"eval_steps_per_second": 3.759, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 3.853813886642456, |
|
"learning_rate": 0.00015523613963039015, |
|
"loss": 0.8368, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 2.1980948448181152, |
|
"learning_rate": 0.00015489390828199863, |
|
"loss": 0.8148, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 2.6982269287109375, |
|
"learning_rate": 0.00015455167693360713, |
|
"loss": 0.7574, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 1.415897011756897, |
|
"learning_rate": 0.00015420944558521563, |
|
"loss": 0.7289, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 1.5153343677520752, |
|
"learning_rate": 0.0001538672142368241, |
|
"loss": 0.7146, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 4.069477558135986, |
|
"learning_rate": 0.00015352498288843258, |
|
"loss": 0.7833, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 2.2534329891204834, |
|
"learning_rate": 0.00015318275154004108, |
|
"loss": 0.7169, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 2.8250768184661865, |
|
"learning_rate": 0.00015284052019164956, |
|
"loss": 0.7735, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 4.815442085266113, |
|
"learning_rate": 0.00015249828884325806, |
|
"loss": 0.8695, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 1.844257116317749, |
|
"learning_rate": 0.00015215605749486653, |
|
"loss": 0.9246, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"eval_accuracy": 0.6634746922024624, |
|
"eval_loss": 0.7895568013191223, |
|
"eval_runtime": 97.6845, |
|
"eval_samples_per_second": 29.933, |
|
"eval_steps_per_second": 3.747, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 3.4265382289886475, |
|
"learning_rate": 0.000151813826146475, |
|
"loss": 0.8821, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 1.8195841312408447, |
|
"learning_rate": 0.0001514715947980835, |
|
"loss": 0.7777, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 2.0240731239318848, |
|
"learning_rate": 0.000151129363449692, |
|
"loss": 0.8862, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 2.748037338256836, |
|
"learning_rate": 0.00015078713210130048, |
|
"loss": 0.7216, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 2.5891735553741455, |
|
"learning_rate": 0.00015044490075290899, |
|
"loss": 0.8374, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.7044845819473267, |
|
"learning_rate": 0.00015010266940451746, |
|
"loss": 0.739, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.914048194885254, |
|
"learning_rate": 0.00014976043805612594, |
|
"loss": 0.6031, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 1.34175443649292, |
|
"learning_rate": 0.00014941820670773444, |
|
"loss": 0.5003, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 2.1494643688201904, |
|
"learning_rate": 0.0001490759753593429, |
|
"loss": 0.5636, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 1.9443460702896118, |
|
"learning_rate": 0.0001487337440109514, |
|
"loss": 0.5026, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"eval_accuracy": 0.6812585499316005, |
|
"eval_loss": 0.787219226360321, |
|
"eval_runtime": 98.8516, |
|
"eval_samples_per_second": 29.58, |
|
"eval_steps_per_second": 3.703, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 5.107473373413086, |
|
"learning_rate": 0.00014839151266255992, |
|
"loss": 0.6356, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 1.6247891187667847, |
|
"learning_rate": 0.00014804928131416836, |
|
"loss": 0.54, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 3.002816677093506, |
|
"learning_rate": 0.00014770704996577686, |
|
"loss": 0.5619, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 3.7361197471618652, |
|
"learning_rate": 0.00014736481861738537, |
|
"loss": 0.6403, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 2.038827896118164, |
|
"learning_rate": 0.00014702258726899384, |
|
"loss": 0.5458, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 2.4605889320373535, |
|
"learning_rate": 0.00014668035592060234, |
|
"loss": 0.4914, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 3.549699544906616, |
|
"learning_rate": 0.00014633812457221082, |
|
"loss": 0.5927, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 2.2531206607818604, |
|
"learning_rate": 0.0001459958932238193, |
|
"loss": 0.5065, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 2.869708299636841, |
|
"learning_rate": 0.0001456536618754278, |
|
"loss": 0.6245, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 1.7271431684494019, |
|
"learning_rate": 0.0001453114305270363, |
|
"loss": 0.7599, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"eval_accuracy": 0.6757865937072504, |
|
"eval_loss": 0.7860671281814575, |
|
"eval_runtime": 98.0642, |
|
"eval_samples_per_second": 29.817, |
|
"eval_steps_per_second": 3.732, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 3.388765573501587, |
|
"learning_rate": 0.00014496919917864477, |
|
"loss": 0.6075, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 1.6970986127853394, |
|
"learning_rate": 0.00014462696783025327, |
|
"loss": 0.506, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 3.88580584526062, |
|
"learning_rate": 0.00014428473648186175, |
|
"loss": 0.6463, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 4.629645347595215, |
|
"learning_rate": 0.00014394250513347022, |
|
"loss": 0.5743, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 3.3085250854492188, |
|
"learning_rate": 0.00014360027378507872, |
|
"loss": 0.5702, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 1.6527128219604492, |
|
"learning_rate": 0.00014325804243668722, |
|
"loss": 0.3941, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 4.697700500488281, |
|
"learning_rate": 0.0001429158110882957, |
|
"loss": 0.6121, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 2.6221749782562256, |
|
"learning_rate": 0.00014257357973990417, |
|
"loss": 0.5243, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 2.580265522003174, |
|
"learning_rate": 0.00014223134839151267, |
|
"loss": 0.5563, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 2.5967156887054443, |
|
"learning_rate": 0.00014188911704312115, |
|
"loss": 0.5764, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"eval_accuracy": 0.6802325581395349, |
|
"eval_loss": 0.808827817440033, |
|
"eval_runtime": 97.9955, |
|
"eval_samples_per_second": 29.838, |
|
"eval_steps_per_second": 3.735, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 2.171257257461548, |
|
"learning_rate": 0.00014154688569472965, |
|
"loss": 0.6382, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 1.1206802129745483, |
|
"learning_rate": 0.00014120465434633813, |
|
"loss": 0.5243, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 1.483469009399414, |
|
"learning_rate": 0.0001408624229979466, |
|
"loss": 0.7305, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 2.7231271266937256, |
|
"learning_rate": 0.0001405201916495551, |
|
"loss": 0.4636, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 3.154658794403076, |
|
"learning_rate": 0.0001401779603011636, |
|
"loss": 0.5783, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 3.6419148445129395, |
|
"learning_rate": 0.00013983572895277208, |
|
"loss": 0.5313, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 2.149718999862671, |
|
"learning_rate": 0.00013949349760438058, |
|
"loss": 0.5444, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 3.819136142730713, |
|
"learning_rate": 0.00013915126625598905, |
|
"loss": 0.7068, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 1.5383284091949463, |
|
"learning_rate": 0.00013880903490759753, |
|
"loss": 0.5585, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 2.2872893810272217, |
|
"learning_rate": 0.00013846680355920603, |
|
"loss": 0.4329, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"eval_accuracy": 0.7058823529411765, |
|
"eval_loss": 0.7281420826911926, |
|
"eval_runtime": 98.0691, |
|
"eval_samples_per_second": 29.816, |
|
"eval_steps_per_second": 3.732, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 4.656331539154053, |
|
"learning_rate": 0.0001381245722108145, |
|
"loss": 0.4951, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 3.128675937652588, |
|
"learning_rate": 0.000137782340862423, |
|
"loss": 0.6682, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 2.384826898574829, |
|
"learning_rate": 0.0001374401095140315, |
|
"loss": 0.5825, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 2.245954990386963, |
|
"learning_rate": 0.00013709787816563998, |
|
"loss": 0.48, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 2.726475715637207, |
|
"learning_rate": 0.00013675564681724846, |
|
"loss": 0.4232, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 2.0738930702209473, |
|
"learning_rate": 0.00013641341546885696, |
|
"loss": 0.5, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 1.9418065547943115, |
|
"learning_rate": 0.00013607118412046543, |
|
"loss": 0.4846, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 3.351672410964966, |
|
"learning_rate": 0.00013572895277207394, |
|
"loss": 0.4581, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 3.308870792388916, |
|
"learning_rate": 0.0001353867214236824, |
|
"loss": 0.5721, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 3.86739182472229, |
|
"learning_rate": 0.00013504449007529089, |
|
"loss": 0.6271, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"eval_accuracy": 0.7116963064295485, |
|
"eval_loss": 0.7290834784507751, |
|
"eval_runtime": 98.3785, |
|
"eval_samples_per_second": 29.722, |
|
"eval_steps_per_second": 3.72, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 2.4816877841949463, |
|
"learning_rate": 0.0001347022587268994, |
|
"loss": 0.4315, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 2.6154468059539795, |
|
"learning_rate": 0.0001343600273785079, |
|
"loss": 0.5494, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 4.586129188537598, |
|
"learning_rate": 0.00013401779603011636, |
|
"loss": 0.5922, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 3.945017099380493, |
|
"learning_rate": 0.00013367556468172487, |
|
"loss": 0.7348, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 9.25570297241211, |
|
"learning_rate": 0.00013333333333333334, |
|
"loss": 0.4907, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 2.919928789138794, |
|
"learning_rate": 0.00013299110198494181, |
|
"loss": 0.5759, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 3.0628156661987305, |
|
"learning_rate": 0.00013264887063655032, |
|
"loss": 0.5712, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 3.5648550987243652, |
|
"learning_rate": 0.00013230663928815882, |
|
"loss": 0.6428, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 3.348015546798706, |
|
"learning_rate": 0.0001319644079397673, |
|
"loss": 0.5054, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 3.1563289165496826, |
|
"learning_rate": 0.00013162217659137577, |
|
"loss": 0.5498, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"eval_accuracy": 0.7058823529411765, |
|
"eval_loss": 0.7744925618171692, |
|
"eval_runtime": 98.5948, |
|
"eval_samples_per_second": 29.657, |
|
"eval_steps_per_second": 3.712, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 2.594055414199829, |
|
"learning_rate": 0.00013127994524298427, |
|
"loss": 0.5086, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 5.243160724639893, |
|
"learning_rate": 0.00013093771389459274, |
|
"loss": 0.6349, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 3.0656027793884277, |
|
"learning_rate": 0.00013059548254620124, |
|
"loss": 0.5671, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 2.7779109477996826, |
|
"learning_rate": 0.00013025325119780975, |
|
"loss": 0.5813, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 3.042433023452759, |
|
"learning_rate": 0.0001299110198494182, |
|
"loss": 0.6665, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 2.6497788429260254, |
|
"learning_rate": 0.0001295687885010267, |
|
"loss": 0.517, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 2.5200939178466797, |
|
"learning_rate": 0.0001292265571526352, |
|
"loss": 0.6524, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 3.1486799716949463, |
|
"learning_rate": 0.00012888432580424367, |
|
"loss": 0.4255, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 3.680635929107666, |
|
"learning_rate": 0.00012854209445585217, |
|
"loss": 0.499, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 4.004785060882568, |
|
"learning_rate": 0.00012819986310746065, |
|
"loss": 0.5247, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"eval_accuracy": 0.689124487004104, |
|
"eval_loss": 0.8001527190208435, |
|
"eval_runtime": 98.0128, |
|
"eval_samples_per_second": 29.833, |
|
"eval_steps_per_second": 3.734, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 3.6504130363464355, |
|
"learning_rate": 0.00012785763175906912, |
|
"loss": 0.575, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 2.601299285888672, |
|
"learning_rate": 0.00012751540041067762, |
|
"loss": 0.4903, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 3.950484275817871, |
|
"learning_rate": 0.0001271731690622861, |
|
"loss": 0.5872, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 4.166860103607178, |
|
"learning_rate": 0.0001268309377138946, |
|
"loss": 0.5385, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 1.910089373588562, |
|
"learning_rate": 0.0001264887063655031, |
|
"loss": 0.6209, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 14.525965690612793, |
|
"learning_rate": 0.00012614647501711158, |
|
"loss": 0.6604, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 6.641749858856201, |
|
"learning_rate": 0.00012580424366872005, |
|
"loss": 0.5131, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 3.671076536178589, |
|
"learning_rate": 0.00012546201232032855, |
|
"loss": 0.5391, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 3.4304141998291016, |
|
"learning_rate": 0.00012511978097193703, |
|
"loss": 0.4564, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 2.004720687866211, |
|
"learning_rate": 0.00012477754962354553, |
|
"loss": 0.4891, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"eval_accuracy": 0.7099863201094391, |
|
"eval_loss": 0.7013882994651794, |
|
"eval_runtime": 98.2469, |
|
"eval_samples_per_second": 29.762, |
|
"eval_steps_per_second": 3.725, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 4.185796737670898, |
|
"learning_rate": 0.000124435318275154, |
|
"loss": 0.5397, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 4.123441696166992, |
|
"learning_rate": 0.00012409308692676248, |
|
"loss": 0.5452, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 4.169721603393555, |
|
"learning_rate": 0.00012375085557837098, |
|
"loss": 0.658, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 1.8797214031219482, |
|
"learning_rate": 0.00012340862422997948, |
|
"loss": 0.5156, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 2.8339829444885254, |
|
"learning_rate": 0.00012306639288158796, |
|
"loss": 0.5107, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 3.297180414199829, |
|
"learning_rate": 0.00012272416153319643, |
|
"loss": 0.7244, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 2.3934600353240967, |
|
"learning_rate": 0.00012238193018480493, |
|
"loss": 0.6198, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 3.875546932220459, |
|
"learning_rate": 0.00012203969883641342, |
|
"loss": 0.5995, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 4.619594097137451, |
|
"learning_rate": 0.00012169746748802191, |
|
"loss": 0.6293, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 2.215853452682495, |
|
"learning_rate": 0.0001213552361396304, |
|
"loss": 0.5211, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"eval_accuracy": 0.6863885088919288, |
|
"eval_loss": 0.7724740505218506, |
|
"eval_runtime": 99.41, |
|
"eval_samples_per_second": 29.414, |
|
"eval_steps_per_second": 3.682, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 1.7644644975662231, |
|
"learning_rate": 0.0001210130047912389, |
|
"loss": 0.5552, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 2.6976282596588135, |
|
"learning_rate": 0.00012067077344284736, |
|
"loss": 0.4784, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 2.179863691329956, |
|
"learning_rate": 0.00012032854209445585, |
|
"loss": 0.5625, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 1.2474085092544556, |
|
"learning_rate": 0.00011998631074606435, |
|
"loss": 0.5838, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 3.6004438400268555, |
|
"learning_rate": 0.00011964407939767284, |
|
"loss": 0.5386, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 2.642122268676758, |
|
"learning_rate": 0.00011930184804928133, |
|
"loss": 0.5415, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 6.505972862243652, |
|
"learning_rate": 0.0001189596167008898, |
|
"loss": 0.5824, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 4.343991279602051, |
|
"learning_rate": 0.00011861738535249829, |
|
"loss": 0.484, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 3.2172842025756836, |
|
"learning_rate": 0.00011827515400410678, |
|
"loss": 0.5003, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 3.127413272857666, |
|
"learning_rate": 0.00011793292265571527, |
|
"loss": 0.659, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"eval_accuracy": 0.7086183310533516, |
|
"eval_loss": 0.7477269172668457, |
|
"eval_runtime": 98.4799, |
|
"eval_samples_per_second": 29.691, |
|
"eval_steps_per_second": 3.716, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 2.628113269805908, |
|
"learning_rate": 0.00011759069130732377, |
|
"loss": 0.5786, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 1.421924114227295, |
|
"learning_rate": 0.00011724845995893223, |
|
"loss": 0.5878, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 3.8365235328674316, |
|
"learning_rate": 0.00011690622861054073, |
|
"loss": 0.4624, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 3.5438849925994873, |
|
"learning_rate": 0.00011656399726214922, |
|
"loss": 0.5347, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 2.891608238220215, |
|
"learning_rate": 0.00011622176591375771, |
|
"loss": 0.4244, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 3.9522762298583984, |
|
"learning_rate": 0.0001158795345653662, |
|
"loss": 0.5261, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 3.749552011489868, |
|
"learning_rate": 0.0001155373032169747, |
|
"loss": 0.5521, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 3.110929250717163, |
|
"learning_rate": 0.00011519507186858316, |
|
"loss": 0.5003, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 2.78469181060791, |
|
"learning_rate": 0.00011485284052019165, |
|
"loss": 0.5088, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 4.185746669769287, |
|
"learning_rate": 0.00011451060917180015, |
|
"loss": 0.4878, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"eval_accuracy": 0.7051983584131327, |
|
"eval_loss": 0.7128526568412781, |
|
"eval_runtime": 98.1754, |
|
"eval_samples_per_second": 29.783, |
|
"eval_steps_per_second": 3.728, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 2.7198736667633057, |
|
"learning_rate": 0.00011416837782340864, |
|
"loss": 0.4998, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 1.5983054637908936, |
|
"learning_rate": 0.00011382614647501712, |
|
"loss": 0.509, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 3.75059175491333, |
|
"learning_rate": 0.0001134839151266256, |
|
"loss": 0.6075, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 3.323063611984253, |
|
"learning_rate": 0.00011314168377823409, |
|
"loss": 0.5783, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 5.036363124847412, |
|
"learning_rate": 0.00011279945242984257, |
|
"loss": 0.647, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 2.75390887260437, |
|
"learning_rate": 0.00011245722108145106, |
|
"loss": 0.5227, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 2.0095183849334717, |
|
"learning_rate": 0.00011211498973305956, |
|
"loss": 0.4561, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 3.252497911453247, |
|
"learning_rate": 0.00011177275838466803, |
|
"loss": 0.5389, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 2.183528184890747, |
|
"learning_rate": 0.00011143052703627653, |
|
"loss": 0.4228, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 2.2591006755828857, |
|
"learning_rate": 0.00011108829568788502, |
|
"loss": 0.4941, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"eval_accuracy": 0.7219562243502052, |
|
"eval_loss": 0.6829969882965088, |
|
"eval_runtime": 99.2532, |
|
"eval_samples_per_second": 29.46, |
|
"eval_steps_per_second": 3.688, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 2.982792377471924, |
|
"learning_rate": 0.0001107460643394935, |
|
"loss": 0.4714, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 2.341184616088867, |
|
"learning_rate": 0.00011040383299110199, |
|
"loss": 0.4921, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 4.106105327606201, |
|
"learning_rate": 0.0001100616016427105, |
|
"loss": 0.5576, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 2.408310651779175, |
|
"learning_rate": 0.00010971937029431895, |
|
"loss": 0.5945, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 2.786745071411133, |
|
"learning_rate": 0.00010937713894592744, |
|
"loss": 0.411, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 2.012185573577881, |
|
"learning_rate": 0.00010903490759753594, |
|
"loss": 0.5151, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 4.303542137145996, |
|
"learning_rate": 0.00010869267624914443, |
|
"loss": 0.5431, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 3.5029568672180176, |
|
"learning_rate": 0.00010835044490075292, |
|
"loss": 0.62, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 2.5191617012023926, |
|
"learning_rate": 0.0001080082135523614, |
|
"loss": 0.5678, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 2.9402916431427, |
|
"learning_rate": 0.00010766598220396988, |
|
"loss": 0.4648, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"eval_accuracy": 0.7028043775649795, |
|
"eval_loss": 0.7181953191757202, |
|
"eval_runtime": 97.5335, |
|
"eval_samples_per_second": 29.979, |
|
"eval_steps_per_second": 3.753, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 2.9497809410095215, |
|
"learning_rate": 0.00010732375085557837, |
|
"loss": 0.4891, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 4.476220607757568, |
|
"learning_rate": 0.00010698151950718687, |
|
"loss": 0.4989, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 3.7788009643554688, |
|
"learning_rate": 0.00010663928815879536, |
|
"loss": 0.649, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 3.8084425926208496, |
|
"learning_rate": 0.00010629705681040382, |
|
"loss": 0.6365, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 2.09145188331604, |
|
"learning_rate": 0.00010595482546201232, |
|
"loss": 0.621, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 1.95127272605896, |
|
"learning_rate": 0.00010561259411362081, |
|
"loss": 0.5456, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 3.267646074295044, |
|
"learning_rate": 0.0001052703627652293, |
|
"loss": 0.529, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 3.5166428089141846, |
|
"learning_rate": 0.00010492813141683779, |
|
"loss": 0.442, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 2.4923620223999023, |
|
"learning_rate": 0.00010458590006844629, |
|
"loss": 0.4446, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 4.28298282623291, |
|
"learning_rate": 0.00010424366872005475, |
|
"loss": 0.5501, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"eval_accuracy": 0.7144322845417237, |
|
"eval_loss": 0.7190529108047485, |
|
"eval_runtime": 97.7263, |
|
"eval_samples_per_second": 29.92, |
|
"eval_steps_per_second": 3.745, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 3.6538245677948, |
|
"learning_rate": 0.00010390143737166324, |
|
"loss": 0.484, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 4.382143974304199, |
|
"learning_rate": 0.00010355920602327174, |
|
"loss": 0.4907, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 4.952005386352539, |
|
"learning_rate": 0.00010321697467488023, |
|
"loss": 0.6108, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 5.33458948135376, |
|
"learning_rate": 0.00010287474332648872, |
|
"loss": 0.5257, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"grad_norm": 1.6532644033432007, |
|
"learning_rate": 0.00010253251197809719, |
|
"loss": 0.5796, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 3.1267786026000977, |
|
"learning_rate": 0.00010219028062970568, |
|
"loss": 0.5385, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 2.2853283882141113, |
|
"learning_rate": 0.00010184804928131417, |
|
"loss": 0.4369, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 4.428330421447754, |
|
"learning_rate": 0.00010150581793292267, |
|
"loss": 0.5144, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 2.0124640464782715, |
|
"learning_rate": 0.00010116358658453116, |
|
"loss": 0.427, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 5.213257789611816, |
|
"learning_rate": 0.00010082135523613962, |
|
"loss": 0.5491, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"eval_accuracy": 0.7154582763337893, |
|
"eval_loss": 0.7131847143173218, |
|
"eval_runtime": 97.9351, |
|
"eval_samples_per_second": 29.857, |
|
"eval_steps_per_second": 3.737, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 4.522421836853027, |
|
"learning_rate": 0.00010047912388774812, |
|
"loss": 0.4969, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 2.8199572563171387, |
|
"learning_rate": 0.00010013689253935661, |
|
"loss": 0.535, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 3.0851550102233887, |
|
"learning_rate": 9.97946611909651e-05, |
|
"loss": 0.3725, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 2.9317426681518555, |
|
"learning_rate": 9.945242984257359e-05, |
|
"loss": 0.2253, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 2.0250513553619385, |
|
"learning_rate": 9.911019849418207e-05, |
|
"loss": 0.226, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 2.064840316772461, |
|
"learning_rate": 9.876796714579056e-05, |
|
"loss": 0.2686, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 2.1055028438568115, |
|
"learning_rate": 9.842573579739904e-05, |
|
"loss": 0.26, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 1.3024044036865234, |
|
"learning_rate": 9.808350444900754e-05, |
|
"loss": 0.2297, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 5.22949743270874, |
|
"learning_rate": 9.774127310061603e-05, |
|
"loss": 0.2884, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 3.9687552452087402, |
|
"learning_rate": 9.73990417522245e-05, |
|
"loss": 0.2373, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"eval_accuracy": 0.7096443228454172, |
|
"eval_loss": 0.7831349968910217, |
|
"eval_runtime": 98.1821, |
|
"eval_samples_per_second": 29.781, |
|
"eval_steps_per_second": 3.728, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 2.5891599655151367, |
|
"learning_rate": 9.7056810403833e-05, |
|
"loss": 0.1881, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 1.736016035079956, |
|
"learning_rate": 9.671457905544148e-05, |
|
"loss": 0.2452, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 2.477384567260742, |
|
"learning_rate": 9.637234770704997e-05, |
|
"loss": 0.3086, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 2.031851053237915, |
|
"learning_rate": 9.603011635865847e-05, |
|
"loss": 0.2753, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 2.2902073860168457, |
|
"learning_rate": 9.568788501026694e-05, |
|
"loss": 0.2279, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 3.447719097137451, |
|
"learning_rate": 9.534565366187543e-05, |
|
"loss": 0.3219, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 3.44710373878479, |
|
"learning_rate": 9.500342231348393e-05, |
|
"loss": 0.2313, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 2.4824321269989014, |
|
"learning_rate": 9.46611909650924e-05, |
|
"loss": 0.2762, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 2.154989719390869, |
|
"learning_rate": 9.43189596167009e-05, |
|
"loss": 0.2523, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 6.111990928649902, |
|
"learning_rate": 9.397672826830938e-05, |
|
"loss": 0.2756, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"eval_accuracy": 0.7246922024623803, |
|
"eval_loss": 0.7965042591094971, |
|
"eval_runtime": 97.9768, |
|
"eval_samples_per_second": 29.844, |
|
"eval_steps_per_second": 3.736, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 3.067180633544922, |
|
"learning_rate": 9.363449691991787e-05, |
|
"loss": 0.3076, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 1.2516257762908936, |
|
"learning_rate": 9.329226557152636e-05, |
|
"loss": 0.308, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 0.4879588484764099, |
|
"learning_rate": 9.295003422313485e-05, |
|
"loss": 0.2395, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 1.9442752599716187, |
|
"learning_rate": 9.260780287474334e-05, |
|
"loss": 0.1651, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 1.563724160194397, |
|
"learning_rate": 9.226557152635181e-05, |
|
"loss": 0.1976, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 2.747708320617676, |
|
"learning_rate": 9.19233401779603e-05, |
|
"loss": 0.1894, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 1.413163185119629, |
|
"learning_rate": 9.15811088295688e-05, |
|
"loss": 0.257, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 5.1449875831604, |
|
"learning_rate": 9.123887748117727e-05, |
|
"loss": 0.2783, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 8.056723594665527, |
|
"learning_rate": 9.089664613278576e-05, |
|
"loss": 0.4055, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 3.5667428970336914, |
|
"learning_rate": 9.058863791923341e-05, |
|
"loss": 0.2299, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"eval_accuracy": 0.7219562243502052, |
|
"eval_loss": 0.8241410851478577, |
|
"eval_runtime": 97.6979, |
|
"eval_samples_per_second": 29.929, |
|
"eval_steps_per_second": 3.746, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 0.5471103191375732, |
|
"learning_rate": 9.02464065708419e-05, |
|
"loss": 0.2466, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 3.4794716835021973, |
|
"learning_rate": 8.990417522245039e-05, |
|
"loss": 0.3227, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 4.952136993408203, |
|
"learning_rate": 8.956194387405887e-05, |
|
"loss": 0.2804, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 3.02123761177063, |
|
"learning_rate": 8.921971252566735e-05, |
|
"loss": 0.3655, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 3.419013500213623, |
|
"learning_rate": 8.887748117727585e-05, |
|
"loss": 0.2, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 2.6587626934051514, |
|
"learning_rate": 8.853524982888432e-05, |
|
"loss": 0.1915, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 6.789358615875244, |
|
"learning_rate": 8.819301848049281e-05, |
|
"loss": 0.2843, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 4.29807710647583, |
|
"learning_rate": 8.785078713210131e-05, |
|
"loss": 0.314, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 3.908632755279541, |
|
"learning_rate": 8.750855578370979e-05, |
|
"loss": 0.1721, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 2.6210243701934814, |
|
"learning_rate": 8.716632443531828e-05, |
|
"loss": 0.2323, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"eval_accuracy": 0.7110123119015048, |
|
"eval_loss": 0.8286489248275757, |
|
"eval_runtime": 97.6813, |
|
"eval_samples_per_second": 29.934, |
|
"eval_steps_per_second": 3.747, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 2.0647199153900146, |
|
"learning_rate": 8.682409308692678e-05, |
|
"loss": 0.2073, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 3.1741795539855957, |
|
"learning_rate": 8.648186173853525e-05, |
|
"loss": 0.3861, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 2.711451530456543, |
|
"learning_rate": 8.613963039014374e-05, |
|
"loss": 0.2986, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 3.5253520011901855, |
|
"learning_rate": 8.579739904175223e-05, |
|
"loss": 0.1992, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 2.7021403312683105, |
|
"learning_rate": 8.545516769336072e-05, |
|
"loss": 0.2105, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 1.193359136581421, |
|
"learning_rate": 8.51129363449692e-05, |
|
"loss": 0.3453, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 4.054200649261475, |
|
"learning_rate": 8.47707049965777e-05, |
|
"loss": 0.2351, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 5.733345985412598, |
|
"learning_rate": 8.442847364818618e-05, |
|
"loss": 0.3263, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.9603079557418823, |
|
"learning_rate": 8.408624229979466e-05, |
|
"loss": 0.2073, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 3.1017444133758545, |
|
"learning_rate": 8.374401095140314e-05, |
|
"loss": 0.1979, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"eval_accuracy": 0.7301641586867305, |
|
"eval_loss": 0.7993032336235046, |
|
"eval_runtime": 98.5314, |
|
"eval_samples_per_second": 29.676, |
|
"eval_steps_per_second": 3.715, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 4.0206146240234375, |
|
"learning_rate": 8.340177960301165e-05, |
|
"loss": 0.277, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 5.1184797286987305, |
|
"learning_rate": 8.305954825462012e-05, |
|
"loss": 0.2841, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 1.4796267747879028, |
|
"learning_rate": 8.271731690622861e-05, |
|
"loss": 0.1861, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 3.755347967147827, |
|
"learning_rate": 8.237508555783711e-05, |
|
"loss": 0.2651, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 1.0390390157699585, |
|
"learning_rate": 8.203285420944559e-05, |
|
"loss": 0.2076, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"grad_norm": 3.164752244949341, |
|
"learning_rate": 8.169062286105407e-05, |
|
"loss": 0.2688, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 1.338983416557312, |
|
"learning_rate": 8.134839151266256e-05, |
|
"loss": 0.1923, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 3.2453787326812744, |
|
"learning_rate": 8.100616016427105e-05, |
|
"loss": 0.3519, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 1.3863258361816406, |
|
"learning_rate": 8.066392881587954e-05, |
|
"loss": 0.3351, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 3.6055376529693604, |
|
"learning_rate": 8.032169746748803e-05, |
|
"loss": 0.2507, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"eval_accuracy": 0.7188782489740082, |
|
"eval_loss": 0.8477036952972412, |
|
"eval_runtime": 98.8022, |
|
"eval_samples_per_second": 29.594, |
|
"eval_steps_per_second": 3.704, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 3.500161647796631, |
|
"learning_rate": 7.997946611909651e-05, |
|
"loss": 0.2426, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 2.6914896965026855, |
|
"learning_rate": 7.9637234770705e-05, |
|
"loss": 0.2767, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 3.296560287475586, |
|
"learning_rate": 7.929500342231349e-05, |
|
"loss": 0.3303, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 3.575965642929077, |
|
"learning_rate": 7.895277207392198e-05, |
|
"loss": 0.2548, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 3.445319175720215, |
|
"learning_rate": 7.861054072553045e-05, |
|
"loss": 0.2209, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 3.6920342445373535, |
|
"learning_rate": 7.826830937713896e-05, |
|
"loss": 0.3507, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 4.31784200668335, |
|
"learning_rate": 7.792607802874744e-05, |
|
"loss": 0.339, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 3.469020128250122, |
|
"learning_rate": 7.758384668035592e-05, |
|
"loss": 0.297, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 2.89150333404541, |
|
"learning_rate": 7.72416153319644e-05, |
|
"loss": 0.2303, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 2.1436691284179688, |
|
"learning_rate": 7.689938398357291e-05, |
|
"loss": 0.205, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"eval_accuracy": 0.7123803009575923, |
|
"eval_loss": 0.8196815252304077, |
|
"eval_runtime": 98.7766, |
|
"eval_samples_per_second": 29.602, |
|
"eval_steps_per_second": 3.705, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 3.9782419204711914, |
|
"learning_rate": 7.655715263518138e-05, |
|
"loss": 0.2987, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 2.255920648574829, |
|
"learning_rate": 7.621492128678987e-05, |
|
"loss": 0.2195, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 1.3451207876205444, |
|
"learning_rate": 7.587268993839836e-05, |
|
"loss": 0.1647, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 3.4238128662109375, |
|
"learning_rate": 7.553045859000685e-05, |
|
"loss": 0.2629, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 4.091916084289551, |
|
"learning_rate": 7.518822724161534e-05, |
|
"loss": 0.2447, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 5.296712398529053, |
|
"learning_rate": 7.484599589322382e-05, |
|
"loss": 0.2414, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 1.0360019207000732, |
|
"learning_rate": 7.450376454483231e-05, |
|
"loss": 0.1963, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 3.331775665283203, |
|
"learning_rate": 7.41615331964408e-05, |
|
"loss": 0.275, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 1.4381409883499146, |
|
"learning_rate": 7.381930184804929e-05, |
|
"loss": 0.2396, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 2.7507870197296143, |
|
"learning_rate": 7.347707049965778e-05, |
|
"loss": 0.35, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"eval_accuracy": 0.7127222982216143, |
|
"eval_loss": 0.8347957730293274, |
|
"eval_runtime": 98.1633, |
|
"eval_samples_per_second": 29.787, |
|
"eval_steps_per_second": 3.728, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 5.777230262756348, |
|
"learning_rate": 7.313483915126625e-05, |
|
"loss": 0.2073, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 2.3496434688568115, |
|
"learning_rate": 7.279260780287475e-05, |
|
"loss": 0.1783, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 0.5175051689147949, |
|
"learning_rate": 7.245037645448324e-05, |
|
"loss": 0.3318, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 3.0439980030059814, |
|
"learning_rate": 7.210814510609171e-05, |
|
"loss": 0.1888, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 3.7360622882843018, |
|
"learning_rate": 7.17659137577002e-05, |
|
"loss": 0.1365, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 2.4673309326171875, |
|
"learning_rate": 7.14236824093087e-05, |
|
"loss": 0.3447, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 5.723246097564697, |
|
"learning_rate": 7.108145106091718e-05, |
|
"loss": 0.256, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 1.58307683467865, |
|
"learning_rate": 7.073921971252567e-05, |
|
"loss": 0.2825, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 5.220044136047363, |
|
"learning_rate": 7.039698836413416e-05, |
|
"loss": 0.2681, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 4.61745548248291, |
|
"learning_rate": 7.005475701574264e-05, |
|
"loss": 0.3372, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"eval_accuracy": 0.7199042407660738, |
|
"eval_loss": 0.8999379277229309, |
|
"eval_runtime": 97.7794, |
|
"eval_samples_per_second": 29.904, |
|
"eval_steps_per_second": 3.743, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 3.353637218475342, |
|
"learning_rate": 6.971252566735113e-05, |
|
"loss": 0.3247, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 3.8344199657440186, |
|
"learning_rate": 6.937029431895962e-05, |
|
"loss": 0.1697, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 6.490875244140625, |
|
"learning_rate": 6.902806297056811e-05, |
|
"loss": 0.1713, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 5.620298862457275, |
|
"learning_rate": 6.86858316221766e-05, |
|
"loss": 0.204, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 2.180828094482422, |
|
"learning_rate": 6.834360027378508e-05, |
|
"loss": 0.2048, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 4.349495887756348, |
|
"learning_rate": 6.800136892539357e-05, |
|
"loss": 0.189, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 4.728153705596924, |
|
"learning_rate": 6.765913757700205e-05, |
|
"loss": 0.2495, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 0.8773788809776306, |
|
"learning_rate": 6.731690622861055e-05, |
|
"loss": 0.2749, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 5.594605922698975, |
|
"learning_rate": 6.697467488021904e-05, |
|
"loss": 0.2288, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 2.9209389686584473, |
|
"learning_rate": 6.663244353182751e-05, |
|
"loss": 0.1968, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"eval_accuracy": 0.7274281805745554, |
|
"eval_loss": 0.8263109922409058, |
|
"eval_runtime": 97.795, |
|
"eval_samples_per_second": 29.899, |
|
"eval_steps_per_second": 3.743, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 5.26107120513916, |
|
"learning_rate": 6.6290212183436e-05, |
|
"loss": 0.2062, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 3.8810832500457764, |
|
"learning_rate": 6.59479808350445e-05, |
|
"loss": 0.2644, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 6.908627510070801, |
|
"learning_rate": 6.560574948665298e-05, |
|
"loss": 0.234, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 3.4059367179870605, |
|
"learning_rate": 6.526351813826146e-05, |
|
"loss": 0.2296, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 4.1873698234558105, |
|
"learning_rate": 6.492128678986995e-05, |
|
"loss": 0.2907, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 3.698305606842041, |
|
"learning_rate": 6.457905544147844e-05, |
|
"loss": 0.2561, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 6.283819198608398, |
|
"learning_rate": 6.423682409308693e-05, |
|
"loss": 0.1497, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 3.8906710147857666, |
|
"learning_rate": 6.389459274469542e-05, |
|
"loss": 0.2858, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 3.4375922679901123, |
|
"learning_rate": 6.35523613963039e-05, |
|
"loss": 0.1774, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 0.2142799347639084, |
|
"learning_rate": 6.32101300479124e-05, |
|
"loss": 0.1443, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"eval_accuracy": 0.7243502051983585, |
|
"eval_loss": 0.8703867793083191, |
|
"eval_runtime": 98.045, |
|
"eval_samples_per_second": 29.823, |
|
"eval_steps_per_second": 3.733, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 3.87772798538208, |
|
"learning_rate": 6.286789869952088e-05, |
|
"loss": 0.2021, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 0.43029889464378357, |
|
"learning_rate": 6.252566735112937e-05, |
|
"loss": 0.2097, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 7.595574855804443, |
|
"learning_rate": 6.2217659137577e-05, |
|
"loss": 0.2467, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 4.542736053466797, |
|
"learning_rate": 6.187542778918549e-05, |
|
"loss": 0.2685, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 3.0954976081848145, |
|
"learning_rate": 6.153319644079398e-05, |
|
"loss": 0.2651, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 0.6948625445365906, |
|
"learning_rate": 6.119096509240247e-05, |
|
"loss": 0.3066, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 2.493579387664795, |
|
"learning_rate": 6.0848733744010955e-05, |
|
"loss": 0.286, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 4.651486873626709, |
|
"learning_rate": 6.050650239561945e-05, |
|
"loss": 0.3193, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 2.718686103820801, |
|
"learning_rate": 6.0164271047227924e-05, |
|
"loss": 0.2494, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 3.578540563583374, |
|
"learning_rate": 5.982203969883642e-05, |
|
"loss": 0.1933, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"eval_accuracy": 0.7243502051983585, |
|
"eval_loss": 0.8270108103752136, |
|
"eval_runtime": 97.3019, |
|
"eval_samples_per_second": 30.051, |
|
"eval_steps_per_second": 3.761, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 5.739526271820068, |
|
"learning_rate": 5.94798083504449e-05, |
|
"loss": 0.2929, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 5.148767471313477, |
|
"learning_rate": 5.913757700205339e-05, |
|
"loss": 0.245, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 3.4480843544006348, |
|
"learning_rate": 5.8795345653661884e-05, |
|
"loss": 0.1627, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 8.554824829101562, |
|
"learning_rate": 5.8453114305270365e-05, |
|
"loss": 0.2578, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 4.859767436981201, |
|
"learning_rate": 5.8110882956878853e-05, |
|
"loss": 0.1952, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 3.225494146347046, |
|
"learning_rate": 5.776865160848735e-05, |
|
"loss": 0.2609, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 6.27465295791626, |
|
"learning_rate": 5.742642026009582e-05, |
|
"loss": 0.3345, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 3.6918678283691406, |
|
"learning_rate": 5.708418891170432e-05, |
|
"loss": 0.2109, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 5.811365604400635, |
|
"learning_rate": 5.67419575633128e-05, |
|
"loss": 0.1586, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 10.013077735900879, |
|
"learning_rate": 5.639972621492129e-05, |
|
"loss": 0.2044, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"eval_accuracy": 0.7274281805745554, |
|
"eval_loss": 0.8322764039039612, |
|
"eval_runtime": 97.5717, |
|
"eval_samples_per_second": 29.968, |
|
"eval_steps_per_second": 3.751, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.11126231402158737, |
|
"learning_rate": 5.605749486652978e-05, |
|
"loss": 0.2304, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 8.398926734924316, |
|
"learning_rate": 5.5715263518138264e-05, |
|
"loss": 0.259, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 4.514709949493408, |
|
"learning_rate": 5.537303216974675e-05, |
|
"loss": 0.1818, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 5.3074235916137695, |
|
"learning_rate": 5.503080082135525e-05, |
|
"loss": 0.2162, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"grad_norm": 6.365301132202148, |
|
"learning_rate": 5.468856947296372e-05, |
|
"loss": 0.3011, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 1.199822187423706, |
|
"learning_rate": 5.4346338124572216e-05, |
|
"loss": 0.2197, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 5.970396995544434, |
|
"learning_rate": 5.40041067761807e-05, |
|
"loss": 0.2257, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 4.547550678253174, |
|
"learning_rate": 5.3661875427789186e-05, |
|
"loss": 0.1782, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 1.4566155672073364, |
|
"learning_rate": 5.331964407939768e-05, |
|
"loss": 0.1697, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 0.7294021248817444, |
|
"learning_rate": 5.297741273100616e-05, |
|
"loss": 0.2709, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"eval_accuracy": 0.7294801641586868, |
|
"eval_loss": 0.8493737578392029, |
|
"eval_runtime": 98.5093, |
|
"eval_samples_per_second": 29.682, |
|
"eval_steps_per_second": 3.715, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 1.6976951360702515, |
|
"learning_rate": 5.263518138261465e-05, |
|
"loss": 0.1856, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 2.718456268310547, |
|
"learning_rate": 5.2292950034223145e-05, |
|
"loss": 0.148, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 2.8870649337768555, |
|
"learning_rate": 5.195071868583162e-05, |
|
"loss": 0.2673, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 8.342497825622559, |
|
"learning_rate": 5.1608487337440115e-05, |
|
"loss": 0.1964, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 4.078546524047852, |
|
"learning_rate": 5.1266255989048596e-05, |
|
"loss": 0.2781, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 4.1737799644470215, |
|
"learning_rate": 5.0924024640657084e-05, |
|
"loss": 0.2913, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 4.929452896118164, |
|
"learning_rate": 5.058179329226558e-05, |
|
"loss": 0.1411, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 3.994762420654297, |
|
"learning_rate": 5.023956194387406e-05, |
|
"loss": 0.2904, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.9369414448738098, |
|
"learning_rate": 4.989733059548255e-05, |
|
"loss": 0.1364, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"grad_norm": 2.4886672496795654, |
|
"learning_rate": 4.955509924709104e-05, |
|
"loss": 0.1021, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"eval_accuracy": 0.7335841313269493, |
|
"eval_loss": 0.857344388961792, |
|
"eval_runtime": 97.6576, |
|
"eval_samples_per_second": 29.941, |
|
"eval_steps_per_second": 3.748, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 1.9758846759796143, |
|
"learning_rate": 4.921286789869952e-05, |
|
"loss": 0.0595, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"grad_norm": 0.5239822268486023, |
|
"learning_rate": 4.887063655030801e-05, |
|
"loss": 0.0655, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"grad_norm": 3.8521106243133545, |
|
"learning_rate": 4.85284052019165e-05, |
|
"loss": 0.0548, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"grad_norm": 0.09303455054759979, |
|
"learning_rate": 4.818617385352498e-05, |
|
"loss": 0.0626, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"grad_norm": 0.5870689749717712, |
|
"learning_rate": 4.784394250513347e-05, |
|
"loss": 0.0421, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"grad_norm": 2.024603843688965, |
|
"learning_rate": 4.7501711156741966e-05, |
|
"loss": 0.0241, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"grad_norm": 1.6237764358520508, |
|
"learning_rate": 4.715947980835045e-05, |
|
"loss": 0.092, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"grad_norm": 0.9192093014717102, |
|
"learning_rate": 4.6817248459958935e-05, |
|
"loss": 0.0629, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"grad_norm": 0.39932507276535034, |
|
"learning_rate": 4.6475017111567424e-05, |
|
"loss": 0.0191, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"grad_norm": 0.11515524983406067, |
|
"learning_rate": 4.6132785763175905e-05, |
|
"loss": 0.0393, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"eval_accuracy": 0.737688098495212, |
|
"eval_loss": 0.9333314299583435, |
|
"eval_runtime": 100.3956, |
|
"eval_samples_per_second": 29.125, |
|
"eval_steps_per_second": 3.646, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"grad_norm": 1.743781566619873, |
|
"learning_rate": 4.57905544147844e-05, |
|
"loss": 0.0314, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"grad_norm": 0.39528560638427734, |
|
"learning_rate": 4.544832306639288e-05, |
|
"loss": 0.0695, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"grad_norm": 6.3506340980529785, |
|
"learning_rate": 4.510609171800137e-05, |
|
"loss": 0.0587, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"grad_norm": 4.544747352600098, |
|
"learning_rate": 4.476386036960986e-05, |
|
"loss": 0.0302, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"grad_norm": 1.1574302911758423, |
|
"learning_rate": 4.4421629021218346e-05, |
|
"loss": 0.1469, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"grad_norm": 9.069709777832031, |
|
"learning_rate": 4.4079397672826834e-05, |
|
"loss": 0.0678, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"grad_norm": 0.6530300974845886, |
|
"learning_rate": 4.373716632443532e-05, |
|
"loss": 0.0596, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"grad_norm": 0.2227853536605835, |
|
"learning_rate": 4.3394934976043803e-05, |
|
"loss": 0.0933, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"grad_norm": 4.514962673187256, |
|
"learning_rate": 4.30527036276523e-05, |
|
"loss": 0.0967, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"grad_norm": 3.422731399536133, |
|
"learning_rate": 4.271047227926078e-05, |
|
"loss": 0.0973, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"eval_accuracy": 0.7335841313269493, |
|
"eval_loss": 0.964632511138916, |
|
"eval_runtime": 98.6581, |
|
"eval_samples_per_second": 29.638, |
|
"eval_steps_per_second": 3.71, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"grad_norm": 2.566582202911377, |
|
"learning_rate": 4.236824093086927e-05, |
|
"loss": 0.0826, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"grad_norm": 0.1476365625858307, |
|
"learning_rate": 4.2026009582477756e-05, |
|
"loss": 0.0658, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"grad_norm": 0.049972761422395706, |
|
"learning_rate": 4.1683778234086244e-05, |
|
"loss": 0.0615, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"grad_norm": 0.12682102620601654, |
|
"learning_rate": 4.134154688569473e-05, |
|
"loss": 0.0509, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"grad_norm": 5.897772789001465, |
|
"learning_rate": 4.099931553730322e-05, |
|
"loss": 0.0915, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"grad_norm": 0.03461030498147011, |
|
"learning_rate": 4.06570841889117e-05, |
|
"loss": 0.0837, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"grad_norm": 0.14550870656967163, |
|
"learning_rate": 4.03148528405202e-05, |
|
"loss": 0.0638, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"grad_norm": 0.03923880308866501, |
|
"learning_rate": 3.997262149212868e-05, |
|
"loss": 0.0367, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"grad_norm": 3.9474973678588867, |
|
"learning_rate": 3.9630390143737166e-05, |
|
"loss": 0.1589, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"grad_norm": 0.4555126428604126, |
|
"learning_rate": 3.9288158795345655e-05, |
|
"loss": 0.0317, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"eval_accuracy": 0.7335841313269493, |
|
"eval_loss": 0.9820448160171509, |
|
"eval_runtime": 98.4684, |
|
"eval_samples_per_second": 29.695, |
|
"eval_steps_per_second": 3.717, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"grad_norm": 0.07397227734327316, |
|
"learning_rate": 3.894592744695414e-05, |
|
"loss": 0.0509, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"grad_norm": 0.12771746516227722, |
|
"learning_rate": 3.860369609856263e-05, |
|
"loss": 0.0432, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"grad_norm": 0.5672203302383423, |
|
"learning_rate": 3.826146475017112e-05, |
|
"loss": 0.0284, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"grad_norm": 0.2664686441421509, |
|
"learning_rate": 3.79192334017796e-05, |
|
"loss": 0.0858, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"grad_norm": 6.073988437652588, |
|
"learning_rate": 3.7577002053388095e-05, |
|
"loss": 0.0706, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"grad_norm": 0.13147962093353271, |
|
"learning_rate": 3.723477070499658e-05, |
|
"loss": 0.0774, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"grad_norm": 6.215001106262207, |
|
"learning_rate": 3.6892539356605065e-05, |
|
"loss": 0.039, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"grad_norm": 0.5598773956298828, |
|
"learning_rate": 3.655030800821355e-05, |
|
"loss": 0.0236, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"grad_norm": 0.738316535949707, |
|
"learning_rate": 3.620807665982204e-05, |
|
"loss": 0.0366, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"grad_norm": 1.859127402305603, |
|
"learning_rate": 3.586584531143053e-05, |
|
"loss": 0.0458, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"eval_accuracy": 0.7325581395348837, |
|
"eval_loss": 1.0716344118118286, |
|
"eval_runtime": 98.9214, |
|
"eval_samples_per_second": 29.559, |
|
"eval_steps_per_second": 3.7, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"grad_norm": 3.993727684020996, |
|
"learning_rate": 3.552361396303902e-05, |
|
"loss": 0.0542, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"grad_norm": 3.287825345993042, |
|
"learning_rate": 3.51813826146475e-05, |
|
"loss": 0.0408, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"grad_norm": 5.9396071434021, |
|
"learning_rate": 3.4839151266255994e-05, |
|
"loss": 0.0919, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"grad_norm": 0.04825394228100777, |
|
"learning_rate": 3.449691991786448e-05, |
|
"loss": 0.0754, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"grad_norm": 1.768473744392395, |
|
"learning_rate": 3.415468856947296e-05, |
|
"loss": 0.0924, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"grad_norm": 0.08893775939941406, |
|
"learning_rate": 3.381245722108145e-05, |
|
"loss": 0.0526, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"grad_norm": 10.308042526245117, |
|
"learning_rate": 3.347022587268994e-05, |
|
"loss": 0.1114, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"grad_norm": 7.975988388061523, |
|
"learning_rate": 3.312799452429843e-05, |
|
"loss": 0.0368, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"grad_norm": 1.1892056465148926, |
|
"learning_rate": 3.2785763175906916e-05, |
|
"loss": 0.0976, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"grad_norm": 4.882401466369629, |
|
"learning_rate": 3.24435318275154e-05, |
|
"loss": 0.164, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"eval_accuracy": 0.7311901504787962, |
|
"eval_loss": 1.088935375213623, |
|
"eval_runtime": 99.102, |
|
"eval_samples_per_second": 29.505, |
|
"eval_steps_per_second": 3.693, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"grad_norm": 0.2973441481590271, |
|
"learning_rate": 3.210130047912389e-05, |
|
"loss": 0.0566, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"grad_norm": 2.9655909538269043, |
|
"learning_rate": 3.175906913073238e-05, |
|
"loss": 0.0773, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"grad_norm": 0.0812312662601471, |
|
"learning_rate": 3.141683778234086e-05, |
|
"loss": 0.0484, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"grad_norm": 0.9421347975730896, |
|
"learning_rate": 3.107460643394935e-05, |
|
"loss": 0.044, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"grad_norm": 0.39745262265205383, |
|
"learning_rate": 3.073237508555784e-05, |
|
"loss": 0.0686, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"grad_norm": 5.0486159324646, |
|
"learning_rate": 3.0390143737166326e-05, |
|
"loss": 0.0724, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"grad_norm": 0.15137074887752533, |
|
"learning_rate": 3.004791238877481e-05, |
|
"loss": 0.0659, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"grad_norm": 0.8398496508598328, |
|
"learning_rate": 2.97056810403833e-05, |
|
"loss": 0.0367, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"grad_norm": 0.19229641556739807, |
|
"learning_rate": 2.936344969199179e-05, |
|
"loss": 0.0558, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"grad_norm": 1.1081570386886597, |
|
"learning_rate": 2.9021218343600276e-05, |
|
"loss": 0.0578, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"eval_accuracy": 0.7311901504787962, |
|
"eval_loss": 1.1010822057724, |
|
"eval_runtime": 98.8332, |
|
"eval_samples_per_second": 29.585, |
|
"eval_steps_per_second": 3.703, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"grad_norm": 0.5777124166488647, |
|
"learning_rate": 2.867898699520876e-05, |
|
"loss": 0.0801, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"grad_norm": 0.2859288156032562, |
|
"learning_rate": 2.833675564681725e-05, |
|
"loss": 0.0107, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"grad_norm": 0.21665354073047638, |
|
"learning_rate": 2.799452429842574e-05, |
|
"loss": 0.0714, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"grad_norm": 0.04660337045788765, |
|
"learning_rate": 2.7652292950034225e-05, |
|
"loss": 0.033, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"grad_norm": 0.08590657263994217, |
|
"learning_rate": 2.7310061601642713e-05, |
|
"loss": 0.0928, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"grad_norm": 4.4041032791137695, |
|
"learning_rate": 2.6967830253251198e-05, |
|
"loss": 0.0976, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"grad_norm": 4.6001081466674805, |
|
"learning_rate": 2.662559890485969e-05, |
|
"loss": 0.0908, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"grad_norm": 1.22696852684021, |
|
"learning_rate": 2.6283367556468174e-05, |
|
"loss": 0.0581, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"grad_norm": 5.576196193695068, |
|
"learning_rate": 2.5941136208076662e-05, |
|
"loss": 0.0522, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"grad_norm": 0.21262912452220917, |
|
"learning_rate": 2.5598904859685147e-05, |
|
"loss": 0.0563, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"eval_accuracy": 0.7356361149110807, |
|
"eval_loss": 1.1010431051254272, |
|
"eval_runtime": 98.8855, |
|
"eval_samples_per_second": 29.57, |
|
"eval_steps_per_second": 3.701, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 3.5604753494262695, |
|
"learning_rate": 2.525667351129364e-05, |
|
"loss": 0.0959, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 0.3624362349510193, |
|
"learning_rate": 2.4914442162902123e-05, |
|
"loss": 0.0614, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"grad_norm": 0.05665013566613197, |
|
"learning_rate": 2.457221081451061e-05, |
|
"loss": 0.078, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"grad_norm": 0.3010866641998291, |
|
"learning_rate": 2.42299794661191e-05, |
|
"loss": 0.0446, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"grad_norm": 2.0640757083892822, |
|
"learning_rate": 2.3887748117727584e-05, |
|
"loss": 0.0422, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"grad_norm": 0.3488443195819855, |
|
"learning_rate": 2.3545516769336072e-05, |
|
"loss": 0.0837, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"grad_norm": 3.6810433864593506, |
|
"learning_rate": 2.320328542094456e-05, |
|
"loss": 0.1436, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"grad_norm": 6.979311943054199, |
|
"learning_rate": 2.286105407255305e-05, |
|
"loss": 0.0283, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"grad_norm": 0.41332653164863586, |
|
"learning_rate": 2.2518822724161534e-05, |
|
"loss": 0.0409, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"grad_norm": 4.3381853103637695, |
|
"learning_rate": 2.2176591375770022e-05, |
|
"loss": 0.0318, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"eval_accuracy": 0.7342681258549931, |
|
"eval_loss": 1.0923420190811157, |
|
"eval_runtime": 98.234, |
|
"eval_samples_per_second": 29.766, |
|
"eval_steps_per_second": 3.726, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"grad_norm": 5.9086761474609375, |
|
"learning_rate": 2.183436002737851e-05, |
|
"loss": 0.0669, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"grad_norm": 0.3409196734428406, |
|
"learning_rate": 2.1492128678986998e-05, |
|
"loss": 0.0446, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"grad_norm": 1.891818642616272, |
|
"learning_rate": 2.1149897330595483e-05, |
|
"loss": 0.0644, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"grad_norm": 3.4832639694213867, |
|
"learning_rate": 2.080766598220397e-05, |
|
"loss": 0.0489, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"grad_norm": 0.3685603141784668, |
|
"learning_rate": 2.046543463381246e-05, |
|
"loss": 0.064, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"grad_norm": 0.036173015832901, |
|
"learning_rate": 2.0123203285420947e-05, |
|
"loss": 0.0618, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"grad_norm": 0.8872077465057373, |
|
"learning_rate": 1.9780971937029432e-05, |
|
"loss": 0.088, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"grad_norm": 15.102840423583984, |
|
"learning_rate": 1.943874058863792e-05, |
|
"loss": 0.0912, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"grad_norm": 0.5182992815971375, |
|
"learning_rate": 1.909650924024641e-05, |
|
"loss": 0.0427, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"grad_norm": 4.398647308349609, |
|
"learning_rate": 1.8754277891854896e-05, |
|
"loss": 0.0255, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"eval_accuracy": 0.7332421340629275, |
|
"eval_loss": 1.1156063079833984, |
|
"eval_runtime": 97.9604, |
|
"eval_samples_per_second": 29.849, |
|
"eval_steps_per_second": 3.736, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"grad_norm": 0.02750484086573124, |
|
"learning_rate": 1.841204654346338e-05, |
|
"loss": 0.0359, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"grad_norm": 0.08833522349596024, |
|
"learning_rate": 1.806981519507187e-05, |
|
"loss": 0.0361, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"grad_norm": 5.376247882843018, |
|
"learning_rate": 1.7727583846680358e-05, |
|
"loss": 0.0454, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"grad_norm": 2.7907521724700928, |
|
"learning_rate": 1.7385352498288842e-05, |
|
"loss": 0.0519, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"grad_norm": 3.718421220779419, |
|
"learning_rate": 1.704312114989733e-05, |
|
"loss": 0.0496, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"grad_norm": 4.242332458496094, |
|
"learning_rate": 1.670088980150582e-05, |
|
"loss": 0.0148, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"grad_norm": 1.1361379623413086, |
|
"learning_rate": 1.6358658453114307e-05, |
|
"loss": 0.0713, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"grad_norm": 1.4509526491165161, |
|
"learning_rate": 1.601642710472279e-05, |
|
"loss": 0.1173, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"grad_norm": 0.025651389732956886, |
|
"learning_rate": 1.567419575633128e-05, |
|
"loss": 0.0497, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"grad_norm": 0.9164614081382751, |
|
"learning_rate": 1.5331964407939768e-05, |
|
"loss": 0.0169, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"eval_accuracy": 0.7414500683994528, |
|
"eval_loss": 1.1049845218658447, |
|
"eval_runtime": 97.5294, |
|
"eval_samples_per_second": 29.981, |
|
"eval_steps_per_second": 3.753, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"grad_norm": 0.16985657811164856, |
|
"learning_rate": 1.4989733059548256e-05, |
|
"loss": 0.0156, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"grad_norm": 0.046321526169776917, |
|
"learning_rate": 1.464750171115674e-05, |
|
"loss": 0.0675, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"grad_norm": 0.12195578962564468, |
|
"learning_rate": 1.430527036276523e-05, |
|
"loss": 0.0291, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"grad_norm": 0.8293215036392212, |
|
"learning_rate": 1.3963039014373715e-05, |
|
"loss": 0.0309, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"grad_norm": 0.5577486753463745, |
|
"learning_rate": 1.3620807665982205e-05, |
|
"loss": 0.0484, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"grad_norm": 0.04991504177451134, |
|
"learning_rate": 1.327857631759069e-05, |
|
"loss": 0.0577, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"grad_norm": 9.814260482788086, |
|
"learning_rate": 1.293634496919918e-05, |
|
"loss": 0.0549, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"grad_norm": 2.222853183746338, |
|
"learning_rate": 1.2594113620807665e-05, |
|
"loss": 0.0377, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"grad_norm": 0.03162392973899841, |
|
"learning_rate": 1.2251882272416155e-05, |
|
"loss": 0.0259, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"grad_norm": 0.04703206941485405, |
|
"learning_rate": 1.1909650924024641e-05, |
|
"loss": 0.0629, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"eval_accuracy": 0.7373461012311902, |
|
"eval_loss": 1.1132022142410278, |
|
"eval_runtime": 97.6002, |
|
"eval_samples_per_second": 29.959, |
|
"eval_steps_per_second": 3.75, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"grad_norm": 0.29484233260154724, |
|
"learning_rate": 1.1567419575633129e-05, |
|
"loss": 0.028, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"grad_norm": 0.05921239033341408, |
|
"learning_rate": 1.1225188227241616e-05, |
|
"loss": 0.004, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"grad_norm": 0.025789450854063034, |
|
"learning_rate": 1.0882956878850104e-05, |
|
"loss": 0.0729, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"grad_norm": 2.618316411972046, |
|
"learning_rate": 1.054072553045859e-05, |
|
"loss": 0.133, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"grad_norm": 0.7046251893043518, |
|
"learning_rate": 1.023271731690623e-05, |
|
"loss": 0.0963, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"grad_norm": 0.7423825263977051, |
|
"learning_rate": 9.890485968514716e-06, |
|
"loss": 0.1037, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"grad_norm": 0.02404310554265976, |
|
"learning_rate": 9.548254620123204e-06, |
|
"loss": 0.0922, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"grad_norm": 0.8627146482467651, |
|
"learning_rate": 9.20602327173169e-06, |
|
"loss": 0.0687, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"grad_norm": 0.3027166426181793, |
|
"learning_rate": 8.863791923340179e-06, |
|
"loss": 0.0347, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"grad_norm": 5.652677059173584, |
|
"learning_rate": 8.521560574948665e-06, |
|
"loss": 0.0627, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"eval_accuracy": 0.738030095759234, |
|
"eval_loss": 1.1109539270401, |
|
"eval_runtime": 98.2791, |
|
"eval_samples_per_second": 29.752, |
|
"eval_steps_per_second": 3.724, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"grad_norm": 0.16831336915493011, |
|
"learning_rate": 8.179329226557153e-06, |
|
"loss": 0.0305, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"grad_norm": 0.02433227188885212, |
|
"learning_rate": 7.83709787816564e-06, |
|
"loss": 0.0803, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"grad_norm": 0.018608298152685165, |
|
"learning_rate": 7.494866529774128e-06, |
|
"loss": 0.0106, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"grad_norm": 0.16128766536712646, |
|
"learning_rate": 7.152635181382615e-06, |
|
"loss": 0.0339, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"grad_norm": 5.512686729431152, |
|
"learning_rate": 6.810403832991103e-06, |
|
"loss": 0.1132, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"grad_norm": 0.03263100981712341, |
|
"learning_rate": 6.46817248459959e-06, |
|
"loss": 0.1002, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"grad_norm": 0.07603902369737625, |
|
"learning_rate": 6.125941136208077e-06, |
|
"loss": 0.0517, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"grad_norm": 0.04678818956017494, |
|
"learning_rate": 5.7837097878165646e-06, |
|
"loss": 0.0745, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"grad_norm": 0.03259843587875366, |
|
"learning_rate": 5.441478439425052e-06, |
|
"loss": 0.0306, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"grad_norm": 0.06205374002456665, |
|
"learning_rate": 5.099247091033539e-06, |
|
"loss": 0.0078, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"eval_accuracy": 0.7349521203830369, |
|
"eval_loss": 1.1116923093795776, |
|
"eval_runtime": 98.2862, |
|
"eval_samples_per_second": 29.75, |
|
"eval_steps_per_second": 3.724, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"grad_norm": 12.98844051361084, |
|
"learning_rate": 4.7570157426420265e-06, |
|
"loss": 0.0493, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"grad_norm": 3.6260194778442383, |
|
"learning_rate": 4.414784394250514e-06, |
|
"loss": 0.1285, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"grad_norm": 1.3537558317184448, |
|
"learning_rate": 4.072553045859001e-06, |
|
"loss": 0.055, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"grad_norm": 6.886358737945557, |
|
"learning_rate": 3.7303216974674884e-06, |
|
"loss": 0.0996, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"grad_norm": 6.763581275939941, |
|
"learning_rate": 3.3880903490759757e-06, |
|
"loss": 0.0322, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"grad_norm": 0.09468955546617508, |
|
"learning_rate": 3.0458590006844626e-06, |
|
"loss": 0.0433, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"grad_norm": 0.07523488253355026, |
|
"learning_rate": 2.70362765229295e-06, |
|
"loss": 0.0596, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"grad_norm": 6.736556529998779, |
|
"learning_rate": 2.3613963039014377e-06, |
|
"loss": 0.0475, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"grad_norm": 13.317132949829102, |
|
"learning_rate": 2.019164955509925e-06, |
|
"loss": 0.0395, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"grad_norm": 0.16380663216114044, |
|
"learning_rate": 1.676933607118412e-06, |
|
"loss": 0.027, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"eval_accuracy": 0.7342681258549931, |
|
"eval_loss": 1.1201249361038208, |
|
"eval_runtime": 97.9757, |
|
"eval_samples_per_second": 29.844, |
|
"eval_steps_per_second": 3.736, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"grad_norm": 0.9811205267906189, |
|
"learning_rate": 1.3347022587268994e-06, |
|
"loss": 0.0991, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"grad_norm": 0.2990233898162842, |
|
"learning_rate": 9.924709103353869e-07, |
|
"loss": 0.0435, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"grad_norm": 0.09798883646726608, |
|
"learning_rate": 6.502395619438741e-07, |
|
"loss": 0.03, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 0.43798768520355225, |
|
"learning_rate": 3.080082135523614e-07, |
|
"loss": 0.0463, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"step": 5844, |
|
"total_flos": 7.243900911073935e+18, |
|
"train_loss": 0.43194858603623126, |
|
"train_runtime": 10540.446, |
|
"train_samples_per_second": 8.868, |
|
"train_steps_per_second": 0.554 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 5844, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 100, |
|
"total_flos": 7.243900911073935e+18, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|