|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.7232, |
|
"eval_steps": 500, |
|
"global_step": 9040, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 62.10089111328125, |
|
"learning_rate": 1e-06, |
|
"loss": 4.5777, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 39.39016342163086, |
|
"learning_rate": 2e-06, |
|
"loss": 4.4077, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 54.24020767211914, |
|
"learning_rate": 3e-06, |
|
"loss": 4.4807, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 30.161609649658203, |
|
"learning_rate": 4e-06, |
|
"loss": 4.5756, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 40.131675720214844, |
|
"learning_rate": 4.9999999999999996e-06, |
|
"loss": 4.4352, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 52.3621940612793, |
|
"learning_rate": 6e-06, |
|
"loss": 4.5096, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 49.86561584472656, |
|
"learning_rate": 7e-06, |
|
"loss": 4.493, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 20.034923553466797, |
|
"learning_rate": 8e-06, |
|
"loss": 4.4088, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 50.790679931640625, |
|
"learning_rate": 9e-06, |
|
"loss": 4.4901, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 48.5693473815918, |
|
"learning_rate": 9.999999999999999e-06, |
|
"loss": 4.3628, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 37.95353698730469, |
|
"learning_rate": 1.1e-05, |
|
"loss": 4.3298, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 35.7153434753418, |
|
"learning_rate": 1.2e-05, |
|
"loss": 4.2839, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 91.47773742675781, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"loss": 4.1238, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 23.16193389892578, |
|
"learning_rate": 1.4e-05, |
|
"loss": 4.1245, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 28.304485321044922, |
|
"learning_rate": 1.5e-05, |
|
"loss": 4.2198, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 34.03230285644531, |
|
"learning_rate": 1.6e-05, |
|
"loss": 4.0958, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 29.786975860595703, |
|
"learning_rate": 1.7e-05, |
|
"loss": 4.024, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 33.04754638671875, |
|
"learning_rate": 1.8e-05, |
|
"loss": 4.0832, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 28.68460464477539, |
|
"learning_rate": 1.9e-05, |
|
"loss": 3.9827, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 26.463253021240234, |
|
"learning_rate": 1.9999999999999998e-05, |
|
"loss": 3.9454, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 19.407127380371094, |
|
"learning_rate": 2.1e-05, |
|
"loss": 4.0119, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 26.383380889892578, |
|
"learning_rate": 2.2e-05, |
|
"loss": 3.9554, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 33.225223541259766, |
|
"learning_rate": 2.3000000000000003e-05, |
|
"loss": 3.8172, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 26.000978469848633, |
|
"learning_rate": 2.4e-05, |
|
"loss": 3.8934, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 28.714366912841797, |
|
"learning_rate": 2.5e-05, |
|
"loss": 3.9194, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 28.721248626708984, |
|
"learning_rate": 2.6000000000000002e-05, |
|
"loss": 3.8144, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 24.934555053710938, |
|
"learning_rate": 2.7000000000000002e-05, |
|
"loss": 3.9166, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 23.113840103149414, |
|
"learning_rate": 2.8e-05, |
|
"loss": 3.8248, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 21.58758544921875, |
|
"learning_rate": 2.9e-05, |
|
"loss": 3.7538, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 22.339618682861328, |
|
"learning_rate": 3e-05, |
|
"loss": 3.726, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 49.25693893432617, |
|
"learning_rate": 2.999992132854894e-05, |
|
"loss": 3.8692, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 45.1494026184082, |
|
"learning_rate": 2.999968531502098e-05, |
|
"loss": 3.7374, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 72.25853729248047, |
|
"learning_rate": 2.99992919618918e-05, |
|
"loss": 3.7735, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 39.445220947265625, |
|
"learning_rate": 2.999874127328748e-05, |
|
"loss": 3.759, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 21.18370246887207, |
|
"learning_rate": 2.9998033254984483e-05, |
|
"loss": 3.7841, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 24.310373306274414, |
|
"learning_rate": 2.999716791440959e-05, |
|
"loss": 3.679, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 36.432350158691406, |
|
"learning_rate": 2.9996145260639812e-05, |
|
"loss": 3.6796, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 32.12275314331055, |
|
"learning_rate": 2.9994965304402304e-05, |
|
"loss": 3.7613, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 38.32442855834961, |
|
"learning_rate": 2.999362805807425e-05, |
|
"loss": 3.7586, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 30.289432525634766, |
|
"learning_rate": 2.9992133535682725e-05, |
|
"loss": 3.6919, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 32.69138717651367, |
|
"learning_rate": 2.9990481752904566e-05, |
|
"loss": 3.6855, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 46.554874420166016, |
|
"learning_rate": 2.9988672727066197e-05, |
|
"loss": 3.7201, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 28.671123504638672, |
|
"learning_rate": 2.9986706477143436e-05, |
|
"loss": 3.6594, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 49.44480895996094, |
|
"learning_rate": 2.9984583023761318e-05, |
|
"loss": 3.7271, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 26.61457061767578, |
|
"learning_rate": 2.998230238919386e-05, |
|
"loss": 3.7376, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 27.453275680541992, |
|
"learning_rate": 2.9979864597363846e-05, |
|
"loss": 3.6716, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 22.791175842285156, |
|
"learning_rate": 2.9977269673842554e-05, |
|
"loss": 3.6172, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 58.2718620300293, |
|
"learning_rate": 2.997451764584951e-05, |
|
"loss": 3.7494, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 33.610286712646484, |
|
"learning_rate": 2.9971608542252175e-05, |
|
"loss": 3.7077, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 25.48147201538086, |
|
"learning_rate": 2.9968542393565674e-05, |
|
"loss": 3.6721, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 27.07135581970215, |
|
"learning_rate": 2.996531923195246e-05, |
|
"loss": 3.7106, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 37.24673843383789, |
|
"learning_rate": 2.996193909122197e-05, |
|
"loss": 3.7447, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 26.41890525817871, |
|
"learning_rate": 2.995840200683028e-05, |
|
"loss": 3.5839, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 32.88002014160156, |
|
"learning_rate": 2.995470801587973e-05, |
|
"loss": 3.6606, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 32.0895881652832, |
|
"learning_rate": 2.9950857157118544e-05, |
|
"loss": 3.677, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 37.726783752441406, |
|
"learning_rate": 2.9946849470940395e-05, |
|
"loss": 3.5546, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 56.246299743652344, |
|
"learning_rate": 2.9942684999384034e-05, |
|
"loss": 3.6391, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 35.675662994384766, |
|
"learning_rate": 2.993836378613278e-05, |
|
"loss": 3.5918, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 26.685134887695312, |
|
"learning_rate": 2.993388587651412e-05, |
|
"loss": 3.6331, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 27.400333404541016, |
|
"learning_rate": 2.992925131749921e-05, |
|
"loss": 3.6214, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 28.501314163208008, |
|
"learning_rate": 2.9924460157702378e-05, |
|
"loss": 3.6619, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 30.773778915405273, |
|
"learning_rate": 2.991951244738063e-05, |
|
"loss": 3.6453, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 24.701374053955078, |
|
"learning_rate": 2.9914408238433095e-05, |
|
"loss": 3.7282, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 27.605117797851562, |
|
"learning_rate": 2.990914758440052e-05, |
|
"loss": 3.6635, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 27.829086303710938, |
|
"learning_rate": 2.9903730540464668e-05, |
|
"loss": 3.5293, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 40.916263580322266, |
|
"learning_rate": 2.9898157163447767e-05, |
|
"loss": 3.6976, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 33.31068420410156, |
|
"learning_rate": 2.9892427511811912e-05, |
|
"loss": 3.548, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 29.932533264160156, |
|
"learning_rate": 2.9886541645658435e-05, |
|
"loss": 3.7486, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 35.59455490112305, |
|
"learning_rate": 2.9880499626727284e-05, |
|
"loss": 3.6342, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 29.93869400024414, |
|
"learning_rate": 2.9874301518396377e-05, |
|
"loss": 3.6615, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 43.417213439941406, |
|
"learning_rate": 2.986794738568094e-05, |
|
"loss": 3.607, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 52.483917236328125, |
|
"learning_rate": 2.9861437295232825e-05, |
|
"loss": 3.5937, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 30.312334060668945, |
|
"learning_rate": 2.9854771315339787e-05, |
|
"loss": 3.5991, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 49.459136962890625, |
|
"learning_rate": 2.984794951592481e-05, |
|
"loss": 3.5261, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 34.81111526489258, |
|
"learning_rate": 2.984097196854534e-05, |
|
"loss": 3.6818, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 34.721946716308594, |
|
"learning_rate": 2.9833838746392544e-05, |
|
"loss": 3.5636, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 31.46621322631836, |
|
"learning_rate": 2.982654992429056e-05, |
|
"loss": 3.5597, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 38.78512191772461, |
|
"learning_rate": 2.981910557869566e-05, |
|
"loss": 3.661, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 27.38837432861328, |
|
"learning_rate": 2.981150578769553e-05, |
|
"loss": 3.6173, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 45.619632720947266, |
|
"learning_rate": 2.980375063100836e-05, |
|
"loss": 3.6632, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 30.708433151245117, |
|
"learning_rate": 2.979584018998209e-05, |
|
"loss": 3.5165, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 35.472938537597656, |
|
"learning_rate": 2.97877745475935e-05, |
|
"loss": 3.5157, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 39.029415130615234, |
|
"learning_rate": 2.9779553788447358e-05, |
|
"loss": 3.6259, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 57.90769577026367, |
|
"learning_rate": 2.977117799877554e-05, |
|
"loss": 3.6378, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 36.95255661010742, |
|
"learning_rate": 2.9762647266436115e-05, |
|
"loss": 3.5845, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 27.456787109375, |
|
"learning_rate": 2.9753961680912432e-05, |
|
"loss": 3.6647, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 27.383285522460938, |
|
"learning_rate": 2.9745121333312166e-05, |
|
"loss": 3.6668, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 26.555049896240234, |
|
"learning_rate": 2.9736126316366385e-05, |
|
"loss": 3.6617, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 34.009620666503906, |
|
"learning_rate": 2.9726976724428563e-05, |
|
"loss": 3.572, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 45.44181823730469, |
|
"learning_rate": 2.9717672653473588e-05, |
|
"loss": 3.6354, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 30.79588508605957, |
|
"learning_rate": 2.9708214201096758e-05, |
|
"loss": 3.6953, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 46.61872482299805, |
|
"learning_rate": 2.9698601466512767e-05, |
|
"loss": 3.5373, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 42.86500930786133, |
|
"learning_rate": 2.9688834550554647e-05, |
|
"loss": 3.5982, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 33.480289459228516, |
|
"learning_rate": 2.9678913555672733e-05, |
|
"loss": 3.6024, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 36.41415786743164, |
|
"learning_rate": 2.966883858593356e-05, |
|
"loss": 3.4843, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 41.39873123168945, |
|
"learning_rate": 2.9658609747018796e-05, |
|
"loss": 3.5257, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 31.24024200439453, |
|
"learning_rate": 2.964822714622412e-05, |
|
"loss": 3.5927, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 52.78026580810547, |
|
"learning_rate": 2.9637690892458103e-05, |
|
"loss": 3.4678, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 27.40117835998535, |
|
"learning_rate": 2.962700109624106e-05, |
|
"loss": 3.5541, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 23.172683715820312, |
|
"learning_rate": 2.961615786970389e-05, |
|
"loss": 3.5713, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 24.177541732788086, |
|
"learning_rate": 2.960516132658692e-05, |
|
"loss": 3.585, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 44.673912048339844, |
|
"learning_rate": 2.9594011582238672e-05, |
|
"loss": 3.5035, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 35.91664505004883, |
|
"learning_rate": 2.95827087536147e-05, |
|
"loss": 3.6404, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 27.3450870513916, |
|
"learning_rate": 2.9571252959276313e-05, |
|
"loss": 3.5121, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 25.66405487060547, |
|
"learning_rate": 2.955964431938939e-05, |
|
"loss": 3.5009, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 24.1674861907959, |
|
"learning_rate": 2.9547882955723052e-05, |
|
"loss": 3.5482, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 49.72268295288086, |
|
"learning_rate": 2.953596899164846e-05, |
|
"loss": 3.4969, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 26.238168716430664, |
|
"learning_rate": 2.9523902552137436e-05, |
|
"loss": 3.5541, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 30.524545669555664, |
|
"learning_rate": 2.951168376376124e-05, |
|
"loss": 3.6343, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 38.179908752441406, |
|
"learning_rate": 2.9499312754689168e-05, |
|
"loss": 3.4795, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 32.98453903198242, |
|
"learning_rate": 2.9486789654687256e-05, |
|
"loss": 3.6333, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 26.77848243713379, |
|
"learning_rate": 2.94741145951169e-05, |
|
"loss": 3.5654, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 27.737852096557617, |
|
"learning_rate": 2.9461287708933475e-05, |
|
"loss": 3.5044, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 34.2584342956543, |
|
"learning_rate": 2.9448309130684944e-05, |
|
"loss": 3.5979, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 57.86616897583008, |
|
"learning_rate": 2.9435178996510456e-05, |
|
"loss": 3.5726, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 37.64597702026367, |
|
"learning_rate": 2.9421897444138902e-05, |
|
"loss": 3.5913, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 35.22037124633789, |
|
"learning_rate": 2.9408464612887484e-05, |
|
"loss": 3.5959, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 31.878395080566406, |
|
"learning_rate": 2.9394880643660242e-05, |
|
"loss": 3.5974, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 43.614994049072266, |
|
"learning_rate": 2.938114567894659e-05, |
|
"loss": 3.4834, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 27.587766647338867, |
|
"learning_rate": 2.9367259862819805e-05, |
|
"loss": 3.6154, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 30.223772048950195, |
|
"learning_rate": 2.9353223340935533e-05, |
|
"loss": 3.4871, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 34.057884216308594, |
|
"learning_rate": 2.933903626053024e-05, |
|
"loss": 3.605, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 39.219242095947266, |
|
"learning_rate": 2.932469877041969e-05, |
|
"loss": 3.6091, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 33.33955001831055, |
|
"learning_rate": 2.931021102099737e-05, |
|
"loss": 3.4862, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 37.07484436035156, |
|
"learning_rate": 2.9295573164232913e-05, |
|
"loss": 3.5267, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 27.145864486694336, |
|
"learning_rate": 2.9280785353670514e-05, |
|
"loss": 3.4369, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 30.31035041809082, |
|
"learning_rate": 2.9265847744427305e-05, |
|
"loss": 3.6056, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 40.823490142822266, |
|
"learning_rate": 2.925076049319174e-05, |
|
"loss": 3.5916, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 44.224796295166016, |
|
"learning_rate": 2.9235523758221944e-05, |
|
"loss": 3.5881, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 33.34773254394531, |
|
"learning_rate": 2.922013769934406e-05, |
|
"loss": 3.5315, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 25.755775451660156, |
|
"learning_rate": 2.920460247795056e-05, |
|
"loss": 3.621, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 29.75731086730957, |
|
"learning_rate": 2.918891825699857e-05, |
|
"loss": 3.6067, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 40.04263687133789, |
|
"learning_rate": 2.9173085201008144e-05, |
|
"loss": 3.5365, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 26.394916534423828, |
|
"learning_rate": 2.9157103476060547e-05, |
|
"loss": 3.5944, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 23.64137077331543, |
|
"learning_rate": 2.914097324979651e-05, |
|
"loss": 3.5048, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 35.18840408325195, |
|
"learning_rate": 2.9124694691414485e-05, |
|
"loss": 3.5769, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 25.55812644958496, |
|
"learning_rate": 2.9108267971668828e-05, |
|
"loss": 3.5082, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 26.25998878479004, |
|
"learning_rate": 2.909169326286807e-05, |
|
"loss": 3.4967, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 41.78467559814453, |
|
"learning_rate": 2.9074970738873054e-05, |
|
"loss": 3.6409, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 27.53672218322754, |
|
"learning_rate": 2.9058100575095156e-05, |
|
"loss": 3.5903, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 27.329477310180664, |
|
"learning_rate": 2.90410829484944e-05, |
|
"loss": 3.5148, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 34.077083587646484, |
|
"learning_rate": 2.902391803757764e-05, |
|
"loss": 3.5561, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 33.75153732299805, |
|
"learning_rate": 2.900660602239667e-05, |
|
"loss": 3.498, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 28.681190490722656, |
|
"learning_rate": 2.8989147084546335e-05, |
|
"loss": 3.501, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 41.04441833496094, |
|
"learning_rate": 2.8971541407162637e-05, |
|
"loss": 3.5707, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 31.303403854370117, |
|
"learning_rate": 2.8953789174920795e-05, |
|
"loss": 3.5748, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 36.61361312866211, |
|
"learning_rate": 2.8935890574033325e-05, |
|
"loss": 3.6372, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 24.437808990478516, |
|
"learning_rate": 2.8917845792248085e-05, |
|
"loss": 3.5737, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 47.12533187866211, |
|
"learning_rate": 2.8899655018846297e-05, |
|
"loss": 3.5125, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 23.667312622070312, |
|
"learning_rate": 2.8881318444640564e-05, |
|
"loss": 3.5043, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 26.047521591186523, |
|
"learning_rate": 2.8862836261972873e-05, |
|
"loss": 3.6236, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 39.91059494018555, |
|
"learning_rate": 2.8844208664712577e-05, |
|
"loss": 3.4851, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 57.92033004760742, |
|
"learning_rate": 2.882543584825435e-05, |
|
"loss": 3.5578, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 99.40699005126953, |
|
"learning_rate": 2.880651800951616e-05, |
|
"loss": 3.577, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 30.7738094329834, |
|
"learning_rate": 2.8787455346937182e-05, |
|
"loss": 3.5683, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 55.61745834350586, |
|
"learning_rate": 2.876824806047573e-05, |
|
"loss": 3.4959, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 23.431480407714844, |
|
"learning_rate": 2.8748896351607145e-05, |
|
"loss": 3.5882, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 20.944149017333984, |
|
"learning_rate": 2.8729400423321693e-05, |
|
"loss": 3.6096, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 46.77301788330078, |
|
"learning_rate": 2.8709760480122443e-05, |
|
"loss": 3.5665, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 36.645328521728516, |
|
"learning_rate": 2.8689976728023103e-05, |
|
"loss": 3.5087, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 32.61063003540039, |
|
"learning_rate": 2.8670049374545873e-05, |
|
"loss": 3.5054, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 47.47910690307617, |
|
"learning_rate": 2.8649978628719256e-05, |
|
"loss": 3.5674, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 40.817115783691406, |
|
"learning_rate": 2.8629764701075885e-05, |
|
"loss": 3.4504, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 32.290626525878906, |
|
"learning_rate": 2.8609407803650295e-05, |
|
"loss": 3.4699, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 26.942697525024414, |
|
"learning_rate": 2.8588908149976702e-05, |
|
"loss": 3.5642, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 30.925710678100586, |
|
"learning_rate": 2.856826595508678e-05, |
|
"loss": 3.6097, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 42.991546630859375, |
|
"learning_rate": 2.8547481435507382e-05, |
|
"loss": 3.5292, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 23.902685165405273, |
|
"learning_rate": 2.852655480925828e-05, |
|
"loss": 3.5195, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 36.46931076049805, |
|
"learning_rate": 2.8505486295849884e-05, |
|
"loss": 3.5451, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 35.51163101196289, |
|
"learning_rate": 2.848427611628093e-05, |
|
"loss": 3.4868, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 28.35199546813965, |
|
"learning_rate": 2.8462924493036168e-05, |
|
"loss": 3.4984, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 40.32368850708008, |
|
"learning_rate": 2.8441431650084018e-05, |
|
"loss": 3.4839, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 46.23302459716797, |
|
"learning_rate": 2.841979781287424e-05, |
|
"loss": 3.5411, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 36.41411209106445, |
|
"learning_rate": 2.8398023208335537e-05, |
|
"loss": 3.5701, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 29.031347274780273, |
|
"learning_rate": 2.8376108064873216e-05, |
|
"loss": 3.5228, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 32.50440216064453, |
|
"learning_rate": 2.835405261236676e-05, |
|
"loss": 3.479, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 30.749858856201172, |
|
"learning_rate": 2.833185708216743e-05, |
|
"loss": 3.53, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 27.148874282836914, |
|
"learning_rate": 2.8309521707095835e-05, |
|
"loss": 3.4933, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 28.534822463989258, |
|
"learning_rate": 2.8287046721439487e-05, |
|
"loss": 3.5881, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 53.12272644042969, |
|
"learning_rate": 2.8264432360950355e-05, |
|
"loss": 3.5626, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 27.4322509765625, |
|
"learning_rate": 2.8241678862842374e-05, |
|
"loss": 3.4831, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 31.194787979125977, |
|
"learning_rate": 2.8218786465788984e-05, |
|
"loss": 3.6001, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 27.951440811157227, |
|
"learning_rate": 2.8195755409920584e-05, |
|
"loss": 3.4387, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 43.26001739501953, |
|
"learning_rate": 2.8172585936822056e-05, |
|
"loss": 3.5127, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 30.359540939331055, |
|
"learning_rate": 2.814927828953022e-05, |
|
"loss": 3.4761, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 29.243453979492188, |
|
"learning_rate": 2.812583271253125e-05, |
|
"loss": 3.6265, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 25.85703468322754, |
|
"learning_rate": 2.8102249451758162e-05, |
|
"loss": 3.5619, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 32.1122932434082, |
|
"learning_rate": 2.8078528754588207e-05, |
|
"loss": 3.5173, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 28.02115821838379, |
|
"learning_rate": 2.805467086984027e-05, |
|
"loss": 3.4905, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 41.776817321777344, |
|
"learning_rate": 2.803067604777227e-05, |
|
"loss": 3.5289, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 33.876312255859375, |
|
"learning_rate": 2.8006544540078535e-05, |
|
"loss": 3.5414, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 33.06633758544922, |
|
"learning_rate": 2.798227659988717e-05, |
|
"loss": 3.5586, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 24.628082275390625, |
|
"learning_rate": 2.7957872481757377e-05, |
|
"loss": 3.5455, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 23.779598236083984, |
|
"learning_rate": 2.793333244167681e-05, |
|
"loss": 3.5345, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 29.29547119140625, |
|
"learning_rate": 2.790865673705888e-05, |
|
"loss": 3.4588, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 29.236722946166992, |
|
"learning_rate": 2.7883845626740046e-05, |
|
"loss": 3.5286, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 36.6258430480957, |
|
"learning_rate": 2.7858899370977123e-05, |
|
"loss": 3.589, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 48.60535430908203, |
|
"learning_rate": 2.783381823144452e-05, |
|
"loss": 3.6398, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 28.959075927734375, |
|
"learning_rate": 2.780860247123153e-05, |
|
"loss": 3.4484, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 30.727558135986328, |
|
"learning_rate": 2.778325235483954e-05, |
|
"loss": 3.5112, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 32.5091667175293, |
|
"learning_rate": 2.775776814817928e-05, |
|
"loss": 3.4266, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 37.957637786865234, |
|
"learning_rate": 2.7732150118568016e-05, |
|
"loss": 3.6768, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 42.639320373535156, |
|
"learning_rate": 2.770639853472676e-05, |
|
"loss": 3.5102, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 29.558870315551758, |
|
"learning_rate": 2.768051366677744e-05, |
|
"loss": 3.5354, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 23.12784194946289, |
|
"learning_rate": 2.765449578624007e-05, |
|
"loss": 3.5432, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 27.601444244384766, |
|
"learning_rate": 2.7628345166029907e-05, |
|
"loss": 3.5672, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 26.26235008239746, |
|
"learning_rate": 2.760206208045458e-05, |
|
"loss": 3.5635, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 57.84916305541992, |
|
"learning_rate": 2.7575646805211224e-05, |
|
"loss": 3.5254, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 25.554025650024414, |
|
"learning_rate": 2.7549099617383573e-05, |
|
"loss": 3.5142, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 38.82815170288086, |
|
"learning_rate": 2.7522420795439067e-05, |
|
"loss": 3.6104, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 28.093948364257812, |
|
"learning_rate": 2.7495610619225925e-05, |
|
"loss": 3.5265, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 26.187891006469727, |
|
"learning_rate": 2.746866936997021e-05, |
|
"loss": 3.4307, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 35.642738342285156, |
|
"learning_rate": 2.7441597330272874e-05, |
|
"loss": 3.5501, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 32.99201965332031, |
|
"learning_rate": 2.7414394784106812e-05, |
|
"loss": 3.4463, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 28.848899841308594, |
|
"learning_rate": 2.7387062016813845e-05, |
|
"loss": 3.5128, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 30.502288818359375, |
|
"learning_rate": 2.7359599315101788e-05, |
|
"loss": 3.4909, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 27.91356658935547, |
|
"learning_rate": 2.7332006967041373e-05, |
|
"loss": 3.53, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 47.296627044677734, |
|
"learning_rate": 2.7304285262063274e-05, |
|
"loss": 3.4793, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 33.32771682739258, |
|
"learning_rate": 2.7276434490955074e-05, |
|
"loss": 3.4695, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 36.65375518798828, |
|
"learning_rate": 2.7248454945858164e-05, |
|
"loss": 3.4502, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 62.65798568725586, |
|
"learning_rate": 2.7220346920264743e-05, |
|
"loss": 3.5893, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 26.921863555908203, |
|
"learning_rate": 2.71921107090147e-05, |
|
"loss": 3.4381, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 35.91081619262695, |
|
"learning_rate": 2.7163746608292525e-05, |
|
"loss": 3.5292, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 42.172306060791016, |
|
"learning_rate": 2.7135254915624213e-05, |
|
"loss": 3.5314, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 65.17137908935547, |
|
"learning_rate": 2.710663592987414e-05, |
|
"loss": 3.518, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 32.53944396972656, |
|
"learning_rate": 2.7077889951241924e-05, |
|
"loss": 3.5562, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 25.663211822509766, |
|
"learning_rate": 2.704901728125928e-05, |
|
"loss": 3.5537, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 23.626951217651367, |
|
"learning_rate": 2.702001822278685e-05, |
|
"loss": 3.5525, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 30.527162551879883, |
|
"learning_rate": 2.699089308001104e-05, |
|
"loss": 3.4913, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 37.62814712524414, |
|
"learning_rate": 2.696164215844081e-05, |
|
"loss": 3.5342, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 26.47550392150879, |
|
"learning_rate": 2.6932265764904494e-05, |
|
"loss": 3.4708, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 30.779155731201172, |
|
"learning_rate": 2.6902764207546553e-05, |
|
"loss": 3.5078, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 34.16841506958008, |
|
"learning_rate": 2.6873137795824367e-05, |
|
"loss": 3.4754, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 36.18644714355469, |
|
"learning_rate": 2.6843386840504972e-05, |
|
"loss": 3.4413, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 34.17078399658203, |
|
"learning_rate": 2.6813511653661817e-05, |
|
"loss": 3.4916, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 24.693265914916992, |
|
"learning_rate": 2.678351254867147e-05, |
|
"loss": 3.4072, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 27.831270217895508, |
|
"learning_rate": 2.675338984021035e-05, |
|
"loss": 3.5353, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 18.52642059326172, |
|
"learning_rate": 2.672314384425142e-05, |
|
"loss": 3.4582, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 44.86159133911133, |
|
"learning_rate": 2.669277487806085e-05, |
|
"loss": 3.4384, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 28.123258590698242, |
|
"learning_rate": 2.6662283260194743e-05, |
|
"loss": 3.5766, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 27.150848388671875, |
|
"learning_rate": 2.6631669310495725e-05, |
|
"loss": 3.5095, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 43.018043518066406, |
|
"learning_rate": 2.660093335008966e-05, |
|
"loss": 3.4795, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 29.27479362487793, |
|
"learning_rate": 2.6570075701382213e-05, |
|
"loss": 3.5236, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 19.501262664794922, |
|
"learning_rate": 2.653909668805553e-05, |
|
"loss": 3.5479, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 57.28257369995117, |
|
"learning_rate": 2.6507996635064792e-05, |
|
"loss": 3.5156, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 27.764036178588867, |
|
"learning_rate": 2.647677586863484e-05, |
|
"loss": 3.5222, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 33.74861526489258, |
|
"learning_rate": 2.644543471625675e-05, |
|
"loss": 3.4773, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 25.404314041137695, |
|
"learning_rate": 2.6413973506684366e-05, |
|
"loss": 3.4646, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 33.307674407958984, |
|
"learning_rate": 2.63823925699309e-05, |
|
"loss": 3.4975, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 28.22442054748535, |
|
"learning_rate": 2.6350692237265428e-05, |
|
"loss": 3.4797, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 26.52558135986328, |
|
"learning_rate": 2.6318872841209446e-05, |
|
"loss": 3.4309, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 36.679386138916016, |
|
"learning_rate": 2.6286934715533353e-05, |
|
"loss": 3.585, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 42.78778839111328, |
|
"learning_rate": 2.6254878195252985e-05, |
|
"loss": 3.4239, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 34.719482421875, |
|
"learning_rate": 2.622270361662606e-05, |
|
"loss": 3.4777, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 33.207427978515625, |
|
"learning_rate": 2.619041131714869e-05, |
|
"loss": 3.5593, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 35.62514877319336, |
|
"learning_rate": 2.6158001635551818e-05, |
|
"loss": 3.5606, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 31.691574096679688, |
|
"learning_rate": 2.6125474911797664e-05, |
|
"loss": 3.4959, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 34.012420654296875, |
|
"learning_rate": 2.6092831487076163e-05, |
|
"loss": 3.57, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 36.75544357299805, |
|
"learning_rate": 2.6060071703801406e-05, |
|
"loss": 3.4718, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 37.18219757080078, |
|
"learning_rate": 2.6027195905608006e-05, |
|
"loss": 3.5332, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 32.344398498535156, |
|
"learning_rate": 2.599420443734754e-05, |
|
"loss": 3.5154, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 26.169748306274414, |
|
"learning_rate": 2.596109764508489e-05, |
|
"loss": 3.5462, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 33.38447570800781, |
|
"learning_rate": 2.592787587609465e-05, |
|
"loss": 3.5658, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 43.4962158203125, |
|
"learning_rate": 2.589453947885745e-05, |
|
"loss": 3.5018, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 34.955291748046875, |
|
"learning_rate": 2.5861088803056324e-05, |
|
"loss": 3.4988, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 29.175064086914062, |
|
"learning_rate": 2.5827524199573033e-05, |
|
"loss": 3.475, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 40.04597091674805, |
|
"learning_rate": 2.5793846020484383e-05, |
|
"loss": 3.4903, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 27.712465286254883, |
|
"learning_rate": 2.5760054619058537e-05, |
|
"loss": 3.5108, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 43.66648864746094, |
|
"learning_rate": 2.5726150349751306e-05, |
|
"loss": 3.4656, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 28.305545806884766, |
|
"learning_rate": 2.569213356820244e-05, |
|
"loss": 3.5766, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 34.897857666015625, |
|
"learning_rate": 2.565800463123187e-05, |
|
"loss": 3.4286, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 43.083229064941406, |
|
"learning_rate": 2.5623763896835997e-05, |
|
"loss": 3.4292, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 35.71794128417969, |
|
"learning_rate": 2.5589411724183926e-05, |
|
"loss": 3.5542, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 23.70340347290039, |
|
"learning_rate": 2.555494847361369e-05, |
|
"loss": 3.5276, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 55.299556732177734, |
|
"learning_rate": 2.552037450662849e-05, |
|
"loss": 3.5644, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 22.433879852294922, |
|
"learning_rate": 2.5485690185892864e-05, |
|
"loss": 3.4475, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 26.397438049316406, |
|
"learning_rate": 2.545089587522893e-05, |
|
"loss": 3.507, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 25.061750411987305, |
|
"learning_rate": 2.5415991939612545e-05, |
|
"loss": 3.4521, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 35.5067138671875, |
|
"learning_rate": 2.5380978745169473e-05, |
|
"loss": 3.4894, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 27.13253402709961, |
|
"learning_rate": 2.5345856659171567e-05, |
|
"loss": 3.4033, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 34.580299377441406, |
|
"learning_rate": 2.5310626050032873e-05, |
|
"loss": 3.5598, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 21.44734764099121, |
|
"learning_rate": 2.527528728730582e-05, |
|
"loss": 3.5189, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 26.825183868408203, |
|
"learning_rate": 2.5239840741677307e-05, |
|
"loss": 3.6052, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 29.19519805908203, |
|
"learning_rate": 2.5204286784964823e-05, |
|
"loss": 3.4724, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 37.9902229309082, |
|
"learning_rate": 2.516862579011255e-05, |
|
"loss": 3.3665, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 28.95624542236328, |
|
"learning_rate": 2.5132858131187446e-05, |
|
"loss": 3.4688, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 34.07016372680664, |
|
"learning_rate": 2.509698418337534e-05, |
|
"loss": 3.4938, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 32.694766998291016, |
|
"learning_rate": 2.5061004322976953e-05, |
|
"loss": 3.4351, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 92.71965789794922, |
|
"learning_rate": 2.5024918927404005e-05, |
|
"loss": 3.5194, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 59.62504959106445, |
|
"learning_rate": 2.4988728375175216e-05, |
|
"loss": 3.5436, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 40.504127502441406, |
|
"learning_rate": 2.495243304591236e-05, |
|
"loss": 3.561, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 43.67660140991211, |
|
"learning_rate": 2.4916033320336263e-05, |
|
"loss": 3.3979, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 38.79692077636719, |
|
"learning_rate": 2.487952958026282e-05, |
|
"loss": 3.5581, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 34.00471878051758, |
|
"learning_rate": 2.4842922208598996e-05, |
|
"loss": 3.3891, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 51.836395263671875, |
|
"learning_rate": 2.480621158933879e-05, |
|
"loss": 3.5146, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 34.47429656982422, |
|
"learning_rate": 2.476939810755923e-05, |
|
"loss": 3.4988, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 27.388093948364258, |
|
"learning_rate": 2.4732482149416325e-05, |
|
"loss": 3.4577, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 22.31077766418457, |
|
"learning_rate": 2.4695464102141002e-05, |
|
"loss": 3.4816, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 33.66743469238281, |
|
"learning_rate": 2.4658344354035063e-05, |
|
"loss": 3.5148, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 30.200748443603516, |
|
"learning_rate": 2.46211232944671e-05, |
|
"loss": 3.6094, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 36.24076843261719, |
|
"learning_rate": 2.4583801313868417e-05, |
|
"loss": 3.4601, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 36.80402374267578, |
|
"learning_rate": 2.4546378803728922e-05, |
|
"loss": 3.5053, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 25.655963897705078, |
|
"learning_rate": 2.450885615659305e-05, |
|
"loss": 3.4791, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 47.66796112060547, |
|
"learning_rate": 2.447123376605561e-05, |
|
"loss": 3.4535, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 39.040924072265625, |
|
"learning_rate": 2.4433512026757668e-05, |
|
"loss": 3.5125, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 27.91790008544922, |
|
"learning_rate": 2.439569133438243e-05, |
|
"loss": 3.5131, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 42.502357482910156, |
|
"learning_rate": 2.435777208565106e-05, |
|
"loss": 3.5424, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 26.83557891845703, |
|
"learning_rate": 2.431975467831853e-05, |
|
"loss": 3.4445, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 39.8357048034668, |
|
"learning_rate": 2.4281639511169457e-05, |
|
"loss": 3.5702, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 31.746124267578125, |
|
"learning_rate": 2.424342698401391e-05, |
|
"loss": 3.4539, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 53.038482666015625, |
|
"learning_rate": 2.4205117497683213e-05, |
|
"loss": 3.5491, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 28.155752182006836, |
|
"learning_rate": 2.4166711454025754e-05, |
|
"loss": 3.4353, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 42.623130798339844, |
|
"learning_rate": 2.4128209255902753e-05, |
|
"loss": 3.4348, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 44.352596282958984, |
|
"learning_rate": 2.408961130718405e-05, |
|
"loss": 3.4637, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 35.380767822265625, |
|
"learning_rate": 2.405091801274387e-05, |
|
"loss": 3.4403, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 67.38970184326172, |
|
"learning_rate": 2.4012129778456556e-05, |
|
"loss": 3.5062, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 36.066741943359375, |
|
"learning_rate": 2.397324701119233e-05, |
|
"loss": 3.6017, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 29.4312744140625, |
|
"learning_rate": 2.3934270118813024e-05, |
|
"loss": 3.4738, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 24.581098556518555, |
|
"learning_rate": 2.3895199510167793e-05, |
|
"loss": 3.4775, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 22.933914184570312, |
|
"learning_rate": 2.385603559508884e-05, |
|
"loss": 3.5718, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 30.199399948120117, |
|
"learning_rate": 2.3816778784387097e-05, |
|
"loss": 3.4446, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 53.968971252441406, |
|
"learning_rate": 2.3777429489847935e-05, |
|
"loss": 3.5161, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 34.33303451538086, |
|
"learning_rate": 2.3737988124226834e-05, |
|
"loss": 3.422, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 28.162084579467773, |
|
"learning_rate": 2.3698455101245052e-05, |
|
"loss": 3.5403, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 28.261327743530273, |
|
"learning_rate": 2.3658830835585294e-05, |
|
"loss": 3.5093, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 37.4576530456543, |
|
"learning_rate": 2.361911574288736e-05, |
|
"loss": 3.5952, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 41.03891372680664, |
|
"learning_rate": 2.3579310239743776e-05, |
|
"loss": 3.5194, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 26.61548614501953, |
|
"learning_rate": 2.353941474369544e-05, |
|
"loss": 3.4933, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 33.990203857421875, |
|
"learning_rate": 2.3499429673227224e-05, |
|
"loss": 3.4745, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 39.59190368652344, |
|
"learning_rate": 2.3459355447763596e-05, |
|
"loss": 3.4875, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 28.182842254638672, |
|
"learning_rate": 2.341919248766422e-05, |
|
"loss": 3.5085, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 27.481103897094727, |
|
"learning_rate": 2.3378941214219545e-05, |
|
"loss": 3.5401, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 29.300914764404297, |
|
"learning_rate": 2.3338602049646372e-05, |
|
"loss": 3.4601, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 27.341121673583984, |
|
"learning_rate": 2.329817541708346e-05, |
|
"loss": 3.4163, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 21.337862014770508, |
|
"learning_rate": 2.3257661740587055e-05, |
|
"loss": 3.3443, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 38.06822204589844, |
|
"learning_rate": 2.3217061445126444e-05, |
|
"loss": 3.4762, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 27.79935646057129, |
|
"learning_rate": 2.3176374956579525e-05, |
|
"loss": 3.4748, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 41.864219665527344, |
|
"learning_rate": 2.3135602701728302e-05, |
|
"loss": 3.4859, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 48.69860076904297, |
|
"learning_rate": 2.3094745108254437e-05, |
|
"loss": 3.5111, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 26.23602867126465, |
|
"learning_rate": 2.305380260473476e-05, |
|
"loss": 3.4637, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 32.71681213378906, |
|
"learning_rate": 2.3012775620636747e-05, |
|
"loss": 3.4752, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 55.637813568115234, |
|
"learning_rate": 2.2971664586314055e-05, |
|
"loss": 3.531, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 29.17162322998047, |
|
"learning_rate": 2.293046993300198e-05, |
|
"loss": 3.567, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 21.349533081054688, |
|
"learning_rate": 2.288919209281294e-05, |
|
"loss": 3.4752, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 25.281597137451172, |
|
"learning_rate": 2.284783149873195e-05, |
|
"loss": 3.4934, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 32.13029861450195, |
|
"learning_rate": 2.2806388584612067e-05, |
|
"loss": 3.4793, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 21.672082901000977, |
|
"learning_rate": 2.2764863785169857e-05, |
|
"loss": 3.4366, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 29.802305221557617, |
|
"learning_rate": 2.2723257535980804e-05, |
|
"loss": 3.5174, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 43.310577392578125, |
|
"learning_rate": 2.2681570273474783e-05, |
|
"loss": 3.4745, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 32.417236328125, |
|
"learning_rate": 2.2639802434931447e-05, |
|
"loss": 3.438, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 42.29374313354492, |
|
"learning_rate": 2.259795445847566e-05, |
|
"loss": 3.5194, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 51.19217300415039, |
|
"learning_rate": 2.2556026783072896e-05, |
|
"loss": 3.496, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 24.400171279907227, |
|
"learning_rate": 2.251401984852463e-05, |
|
"loss": 3.4013, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 26.298309326171875, |
|
"learning_rate": 2.2471934095463724e-05, |
|
"loss": 3.5607, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 27.419946670532227, |
|
"learning_rate": 2.2429769965349818e-05, |
|
"loss": 3.4593, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 29.470266342163086, |
|
"learning_rate": 2.2387527900464676e-05, |
|
"loss": 3.4388, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 21.410829544067383, |
|
"learning_rate": 2.2345208343907577e-05, |
|
"loss": 3.5141, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 32.501766204833984, |
|
"learning_rate": 2.2302811739590642e-05, |
|
"loss": 3.4647, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 24.69274139404297, |
|
"learning_rate": 2.2260338532234194e-05, |
|
"loss": 3.4781, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 25.721759796142578, |
|
"learning_rate": 2.2217789167362078e-05, |
|
"loss": 3.4405, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 34.66562271118164, |
|
"learning_rate": 2.217516409129699e-05, |
|
"loss": 3.5408, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 42.050819396972656, |
|
"learning_rate": 2.2132463751155815e-05, |
|
"loss": 3.4422, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 26.5496883392334, |
|
"learning_rate": 2.2089688594844917e-05, |
|
"loss": 3.4953, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 28.221229553222656, |
|
"learning_rate": 2.2046839071055436e-05, |
|
"loss": 3.3627, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 33.21296310424805, |
|
"learning_rate": 2.2003915629258607e-05, |
|
"loss": 3.383, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 34.35666275024414, |
|
"learning_rate": 2.196091871970103e-05, |
|
"loss": 3.5508, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 25.02892303466797, |
|
"learning_rate": 2.1917848793399926e-05, |
|
"loss": 3.395, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 25.45415687561035, |
|
"learning_rate": 2.187470630213845e-05, |
|
"loss": 3.4692, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 30.72098159790039, |
|
"learning_rate": 2.1831491698460923e-05, |
|
"loss": 3.5714, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 32.58151626586914, |
|
"learning_rate": 2.1788205435668086e-05, |
|
"loss": 3.4358, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 32.444644927978516, |
|
"learning_rate": 2.1744847967812352e-05, |
|
"loss": 3.5389, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 39.33094787597656, |
|
"learning_rate": 2.1701419749693036e-05, |
|
"loss": 3.4003, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 24.18759536743164, |
|
"learning_rate": 2.1657921236851607e-05, |
|
"loss": 3.4501, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 24.81824493408203, |
|
"learning_rate": 2.1614352885566874e-05, |
|
"loss": 3.5004, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 94.60846710205078, |
|
"learning_rate": 2.1570715152850237e-05, |
|
"loss": 3.3695, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 39.35719299316406, |
|
"learning_rate": 2.1527008496440848e-05, |
|
"loss": 3.4058, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 23.86483383178711, |
|
"learning_rate": 2.1483233374800863e-05, |
|
"loss": 3.403, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 36.54322052001953, |
|
"learning_rate": 2.143939024711059e-05, |
|
"loss": 3.504, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 32.654876708984375, |
|
"learning_rate": 2.139547957326369e-05, |
|
"loss": 3.421, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 31.50335121154785, |
|
"learning_rate": 2.1351501813862358e-05, |
|
"loss": 3.5079, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 27.86897850036621, |
|
"learning_rate": 2.130745743021247e-05, |
|
"loss": 3.5605, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 29.914051055908203, |
|
"learning_rate": 2.1263346884318778e-05, |
|
"loss": 3.4826, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 30.09800910949707, |
|
"learning_rate": 2.121917063888004e-05, |
|
"loss": 3.5523, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 37.76126480102539, |
|
"learning_rate": 2.117492915728416e-05, |
|
"loss": 3.6012, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 23.360584259033203, |
|
"learning_rate": 2.1130622903603344e-05, |
|
"loss": 3.45, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 31.356124877929688, |
|
"learning_rate": 2.1086252342589235e-05, |
|
"loss": 3.4666, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 36.73623275756836, |
|
"learning_rate": 2.1041817939668006e-05, |
|
"loss": 3.5774, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 29.439634323120117, |
|
"learning_rate": 2.0997320160935536e-05, |
|
"loss": 3.3953, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 22.339067459106445, |
|
"learning_rate": 2.095275947315246e-05, |
|
"loss": 3.4628, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 16.739225387573242, |
|
"learning_rate": 2.0908136343739308e-05, |
|
"loss": 3.4708, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 27.203229904174805, |
|
"learning_rate": 2.0863451240771592e-05, |
|
"loss": 3.4326, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 30.963396072387695, |
|
"learning_rate": 2.0818704632974896e-05, |
|
"loss": 3.445, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 23.80015754699707, |
|
"learning_rate": 2.0773896989719967e-05, |
|
"loss": 3.5519, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 39.84937286376953, |
|
"learning_rate": 2.072902878101778e-05, |
|
"loss": 3.4384, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 25.7357120513916, |
|
"learning_rate": 2.06841004775146e-05, |
|
"loss": 3.4322, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 25.275423049926758, |
|
"learning_rate": 2.0639112550487085e-05, |
|
"loss": 3.4426, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 28.167940139770508, |
|
"learning_rate": 2.0594065471837302e-05, |
|
"loss": 3.5217, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 44.70423889160156, |
|
"learning_rate": 2.0548959714087783e-05, |
|
"loss": 3.5207, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 30.5069580078125, |
|
"learning_rate": 2.0503795750376583e-05, |
|
"loss": 3.4576, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 24.197166442871094, |
|
"learning_rate": 2.0458574054452316e-05, |
|
"loss": 3.4993, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 31.20002555847168, |
|
"learning_rate": 2.0413295100669167e-05, |
|
"loss": 3.4158, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 33.78310012817383, |
|
"learning_rate": 2.036795936398194e-05, |
|
"loss": 3.4395, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 28.749357223510742, |
|
"learning_rate": 2.0322567319941062e-05, |
|
"loss": 3.451, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 19.698383331298828, |
|
"learning_rate": 2.0277119444687586e-05, |
|
"loss": 3.5069, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 25.00635528564453, |
|
"learning_rate": 2.0231616214948232e-05, |
|
"loss": 3.3704, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 31.135547637939453, |
|
"learning_rate": 2.0186058108030343e-05, |
|
"loss": 3.4219, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 29.921775817871094, |
|
"learning_rate": 2.01404456018169e-05, |
|
"loss": 3.431, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 20.908395767211914, |
|
"learning_rate": 2.0094779174761507e-05, |
|
"loss": 3.4647, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 17.694931030273438, |
|
"learning_rate": 2.0049059305883383e-05, |
|
"loss": 3.4588, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 29.230735778808594, |
|
"learning_rate": 2.000328647476231e-05, |
|
"loss": 3.4788, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 41.20121383666992, |
|
"learning_rate": 1.995746116153363e-05, |
|
"loss": 3.4812, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 39.98432540893555, |
|
"learning_rate": 1.99115838468832e-05, |
|
"loss": 3.4172, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 41.378902435302734, |
|
"learning_rate": 1.9865655012042337e-05, |
|
"loss": 3.5478, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 42.58726501464844, |
|
"learning_rate": 1.9819675138782785e-05, |
|
"loss": 3.4849, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 22.349281311035156, |
|
"learning_rate": 1.9773644709411662e-05, |
|
"loss": 3.4835, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 31.974403381347656, |
|
"learning_rate": 1.9727564206766382e-05, |
|
"loss": 3.3919, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 20.418291091918945, |
|
"learning_rate": 1.9681434114209617e-05, |
|
"loss": 3.4415, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 30.5143985748291, |
|
"learning_rate": 1.963525491562421e-05, |
|
"loss": 3.4806, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 27.770360946655273, |
|
"learning_rate": 1.958902709540811e-05, |
|
"loss": 3.4379, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 24.266944885253906, |
|
"learning_rate": 1.954275113846926e-05, |
|
"loss": 3.4933, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 43.798301696777344, |
|
"learning_rate": 1.9496427530220567e-05, |
|
"loss": 3.4107, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 32.54145431518555, |
|
"learning_rate": 1.9450056756574753e-05, |
|
"loss": 3.507, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 29.06185531616211, |
|
"learning_rate": 1.9403639303939293e-05, |
|
"loss": 3.4434, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 26.419170379638672, |
|
"learning_rate": 1.93571756592113e-05, |
|
"loss": 3.4684, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 31.831510543823242, |
|
"learning_rate": 1.9310666309772426e-05, |
|
"loss": 3.4565, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 33.18935775756836, |
|
"learning_rate": 1.926411174348373e-05, |
|
"loss": 3.4568, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 25.60289192199707, |
|
"learning_rate": 1.9217512448680586e-05, |
|
"loss": 3.3513, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 27.03973960876465, |
|
"learning_rate": 1.917086891416755e-05, |
|
"loss": 3.4551, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 39.20319366455078, |
|
"learning_rate": 1.9124181629213228e-05, |
|
"loss": 3.4217, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 25.287826538085938, |
|
"learning_rate": 1.9077451083545144e-05, |
|
"loss": 3.4172, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 51.33893585205078, |
|
"learning_rate": 1.903067776734461e-05, |
|
"loss": 3.4487, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 23.41849708557129, |
|
"learning_rate": 1.8983862171241577e-05, |
|
"loss": 3.4675, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 37.373104095458984, |
|
"learning_rate": 1.8937004786309504e-05, |
|
"loss": 3.4782, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 26.743480682373047, |
|
"learning_rate": 1.8890106104060177e-05, |
|
"loss": 3.432, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 39.051124572753906, |
|
"learning_rate": 1.8843166616438585e-05, |
|
"loss": 3.4937, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 32.51453399658203, |
|
"learning_rate": 1.8796186815817743e-05, |
|
"loss": 3.4618, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 22.23604393005371, |
|
"learning_rate": 1.874916719499353e-05, |
|
"loss": 3.5259, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 27.168733596801758, |
|
"learning_rate": 1.8702108247179512e-05, |
|
"loss": 3.4829, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 28.35675621032715, |
|
"learning_rate": 1.8655010466001794e-05, |
|
"loss": 3.4133, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 30.800825119018555, |
|
"learning_rate": 1.8607874345493806e-05, |
|
"loss": 3.4527, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 26.275304794311523, |
|
"learning_rate": 1.856070038009115e-05, |
|
"loss": 3.4386, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 27.828401565551758, |
|
"learning_rate": 1.85134890646264e-05, |
|
"loss": 3.4806, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 31.448450088500977, |
|
"learning_rate": 1.846624089432392e-05, |
|
"loss": 3.4659, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 30.396820068359375, |
|
"learning_rate": 1.8418956364794655e-05, |
|
"loss": 3.4717, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 65.90203857421875, |
|
"learning_rate": 1.8371635972030942e-05, |
|
"loss": 3.4938, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 25.6701602935791, |
|
"learning_rate": 1.8324280212401316e-05, |
|
"loss": 3.3931, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 27.28550148010254, |
|
"learning_rate": 1.8276889582645278e-05, |
|
"loss": 3.426, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 26.73450469970703, |
|
"learning_rate": 1.8229464579868124e-05, |
|
"loss": 3.4074, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 30.16777992248535, |
|
"learning_rate": 1.818200570153568e-05, |
|
"loss": 3.4204, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 30.022031784057617, |
|
"learning_rate": 1.813451344546913e-05, |
|
"loss": 3.4905, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 34.67860412597656, |
|
"learning_rate": 1.8086988309839755e-05, |
|
"loss": 3.4165, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 26.23653793334961, |
|
"learning_rate": 1.8039430793163753e-05, |
|
"loss": 3.5014, |
|
"step": 9040 |
|
} |
|
], |
|
"logging_steps": 20, |
|
"max_steps": 20000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 20, |
|
"total_flos": 2.1373183266914304e+16, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|