|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.9946018893387314, |
|
"eval_steps": 500, |
|
"global_step": 1480, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 5172.492790419565, |
|
"learning_rate": 6.756756756756757e-08, |
|
"loss": 16.2669, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 4320.729018206391, |
|
"learning_rate": 3.378378378378379e-07, |
|
"loss": 15.6969, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 488.5525501221014, |
|
"learning_rate": 6.756756756756758e-07, |
|
"loss": 12.0521, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 289.24969349599496, |
|
"learning_rate": 1.0135135135135136e-06, |
|
"loss": 9.2401, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 126.31621322294927, |
|
"learning_rate": 1.3513513513513515e-06, |
|
"loss": 7.9461, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 71.89469484569835, |
|
"learning_rate": 1.6891891891891894e-06, |
|
"loss": 7.2699, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 103.51185464247453, |
|
"learning_rate": 2.0270270270270273e-06, |
|
"loss": 6.9743, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 47.053277376605855, |
|
"learning_rate": 2.364864864864865e-06, |
|
"loss": 6.7672, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 49.644931428662844, |
|
"learning_rate": 2.702702702702703e-06, |
|
"loss": 6.5877, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 35.00554720125077, |
|
"learning_rate": 3.040540540540541e-06, |
|
"loss": 6.4428, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 33.14232577992853, |
|
"learning_rate": 3.3783783783783788e-06, |
|
"loss": 6.2824, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 35.40029424621722, |
|
"learning_rate": 3.7162162162162162e-06, |
|
"loss": 6.0897, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 40.23393814067933, |
|
"learning_rate": 4.0540540540540545e-06, |
|
"loss": 5.8782, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 55.427875975850924, |
|
"learning_rate": 4.391891891891892e-06, |
|
"loss": 5.5613, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 96.7762735033588, |
|
"learning_rate": 4.72972972972973e-06, |
|
"loss": 4.8791, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 88.58451501801524, |
|
"learning_rate": 5.067567567567568e-06, |
|
"loss": 3.2214, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 22.221611879190654, |
|
"learning_rate": 5.405405405405406e-06, |
|
"loss": 1.7005, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 14.404618337150714, |
|
"learning_rate": 5.743243243243244e-06, |
|
"loss": 1.55, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 14.657001400244646, |
|
"learning_rate": 6.081081081081082e-06, |
|
"loss": 1.5228, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 6.295090886362698, |
|
"learning_rate": 6.41891891891892e-06, |
|
"loss": 1.4585, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 5.24494216356606, |
|
"learning_rate": 6.7567567567567575e-06, |
|
"loss": 1.4326, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 6.440473193782044, |
|
"learning_rate": 7.0945945945945946e-06, |
|
"loss": 1.3865, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 4.8147339321014, |
|
"learning_rate": 7.4324324324324324e-06, |
|
"loss": 1.3655, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 5.442428248209423, |
|
"learning_rate": 7.77027027027027e-06, |
|
"loss": 1.3394, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 5.78770666444895, |
|
"learning_rate": 8.108108108108109e-06, |
|
"loss": 1.3354, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 7.28821915344424, |
|
"learning_rate": 8.445945945945948e-06, |
|
"loss": 1.3065, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 6.492851696765312, |
|
"learning_rate": 8.783783783783785e-06, |
|
"loss": 1.2902, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 3.1773499503060454, |
|
"learning_rate": 9.121621621621622e-06, |
|
"loss": 1.2769, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 2.9874866899468238, |
|
"learning_rate": 9.45945945945946e-06, |
|
"loss": 1.266, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 9.94355578144089, |
|
"learning_rate": 9.797297297297298e-06, |
|
"loss": 1.2624, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 3.0193035270709077, |
|
"learning_rate": 9.999944372378571e-06, |
|
"loss": 1.2452, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 5.139515176735739, |
|
"learning_rate": 9.999318575852451e-06, |
|
"loss": 1.2341, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 3.156285780849088, |
|
"learning_rate": 9.99799753559161e-06, |
|
"loss": 1.2106, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 5.924005437849429, |
|
"learning_rate": 9.995981435310078e-06, |
|
"loss": 1.221, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 4.165437348359594, |
|
"learning_rate": 9.993270555382283e-06, |
|
"loss": 1.2105, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 2.342058479511419, |
|
"learning_rate": 9.989865272804064e-06, |
|
"loss": 1.2067, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 2.6456490809423467, |
|
"learning_rate": 9.985766061140233e-06, |
|
"loss": 1.196, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 6.465974100807255, |
|
"learning_rate": 9.980973490458728e-06, |
|
"loss": 1.1881, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 2.808324774416455, |
|
"learning_rate": 9.97548822725133e-06, |
|
"loss": 1.1852, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 2.834201027736285, |
|
"learning_rate": 9.969311034340977e-06, |
|
"loss": 1.1839, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 4.502666083698819, |
|
"learning_rate": 9.962442770775675e-06, |
|
"loss": 1.1648, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 5.780648711057674, |
|
"learning_rate": 9.954884391709043e-06, |
|
"loss": 1.1619, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 3.498462904619314, |
|
"learning_rate": 9.946636948267468e-06, |
|
"loss": 1.1483, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 3.0642292894829213, |
|
"learning_rate": 9.937701587403941e-06, |
|
"loss": 1.1412, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 4.727096769429053, |
|
"learning_rate": 9.928079551738542e-06, |
|
"loss": 1.1446, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 2.206091891548497, |
|
"learning_rate": 9.91777217938564e-06, |
|
"loss": 1.1354, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 3.242676582016566, |
|
"learning_rate": 9.906780903767799e-06, |
|
"loss": 1.137, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 2.5806553543973396, |
|
"learning_rate": 9.895107253416434e-06, |
|
"loss": 1.134, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 2.305243444378998, |
|
"learning_rate": 9.882752851759247e-06, |
|
"loss": 1.1411, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 3.911161316221469, |
|
"learning_rate": 9.869719416894462e-06, |
|
"loss": 1.1144, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 4.2780174060875025, |
|
"learning_rate": 9.856008761351882e-06, |
|
"loss": 1.109, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 2.3947970947073425, |
|
"learning_rate": 9.841622791840839e-06, |
|
"loss": 1.1143, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 5.2640331483926595, |
|
"learning_rate": 9.826563508985017e-06, |
|
"loss": 1.1158, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 2.8813671261206033, |
|
"learning_rate": 9.810833007044247e-06, |
|
"loss": 1.1038, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 1.7947375217746502, |
|
"learning_rate": 9.794433473623249e-06, |
|
"loss": 1.1063, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 2.239529292108957, |
|
"learning_rate": 9.777367189367412e-06, |
|
"loss": 1.1034, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 2.8141790841216405, |
|
"learning_rate": 9.759636527645633e-06, |
|
"loss": 1.094, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 3.3580672092709722, |
|
"learning_rate": 9.74124395422025e-06, |
|
"loss": 1.0989, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 3.116597962727994, |
|
"learning_rate": 9.722192026904145e-06, |
|
"loss": 1.0927, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 8.108195152123194, |
|
"learning_rate": 9.702483395205023e-06, |
|
"loss": 1.0785, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 2.4053331358570564, |
|
"learning_rate": 9.682120799956961e-06, |
|
"loss": 1.0939, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 3.036700462595253, |
|
"learning_rate": 9.661107072939244e-06, |
|
"loss": 1.0957, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 1.9808850854861142, |
|
"learning_rate": 9.639445136482549e-06, |
|
"loss": 1.0718, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 1.7199945370090353, |
|
"learning_rate": 9.61713800306255e-06, |
|
"loss": 1.0908, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 3.416121355561205, |
|
"learning_rate": 9.594188774880981e-06, |
|
"loss": 1.0683, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 2.6221732481646085, |
|
"learning_rate": 9.570600643434217e-06, |
|
"loss": 1.0812, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 3.0081014994296655, |
|
"learning_rate": 9.546376889069442e-06, |
|
"loss": 1.0672, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 1.928715018400264, |
|
"learning_rate": 9.521520880528453e-06, |
|
"loss": 1.0696, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 3.510557690197284, |
|
"learning_rate": 9.496036074479184e-06, |
|
"loss": 1.0748, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 2.1531622758856206, |
|
"learning_rate": 9.46992601503499e-06, |
|
"loss": 1.072, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.5600473945363453, |
|
"learning_rate": 9.44319433326178e-06, |
|
"loss": 1.0674, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 1.9316399234342516, |
|
"learning_rate": 9.415844746673047e-06, |
|
"loss": 1.0715, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 3.1641780363359624, |
|
"learning_rate": 9.387881058712888e-06, |
|
"loss": 1.0607, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 2.5389167608685477, |
|
"learning_rate": 9.359307158227067e-06, |
|
"loss": 1.0571, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 3.209876430466923, |
|
"learning_rate": 9.330127018922195e-06, |
|
"loss": 1.0656, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"grad_norm": 2.0112001074770856, |
|
"learning_rate": 9.300344698813124e-06, |
|
"loss": 1.0473, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 2.3621990368530037, |
|
"learning_rate": 9.269964339658605e-06, |
|
"loss": 1.0479, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 1.9555447312938368, |
|
"learning_rate": 9.238990166385304e-06, |
|
"loss": 1.0481, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 1.9979169106404795, |
|
"learning_rate": 9.207426486500252e-06, |
|
"loss": 1.0489, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 2.7432607054958145, |
|
"learning_rate": 9.175277689491804e-06, |
|
"loss": 1.0553, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 1.9273849661076683, |
|
"learning_rate": 9.142548246219212e-06, |
|
"loss": 1.0398, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 2.0451128390125235, |
|
"learning_rate": 9.109242708290864e-06, |
|
"loss": 1.0417, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 4.496881794147535, |
|
"learning_rate": 9.075365707431311e-06, |
|
"loss": 1.0348, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 3.7112676473960327, |
|
"learning_rate": 9.040921954837139e-06, |
|
"loss": 1.0475, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 1.7843942716305177, |
|
"learning_rate": 9.005916240521788e-06, |
|
"loss": 1.0435, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 3.688231246483607, |
|
"learning_rate": 8.97035343264943e-06, |
|
"loss": 1.0316, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 3.0133406140771672, |
|
"learning_rate": 8.93423847685795e-06, |
|
"loss": 1.0363, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 3.65260559397669, |
|
"learning_rate": 8.89757639557118e-06, |
|
"loss": 1.0295, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 2.0183294533927083, |
|
"learning_rate": 8.860372287300432e-06, |
|
"loss": 1.033, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 1.8479490054275187, |
|
"learning_rate": 8.822631325935463e-06, |
|
"loss": 1.0472, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 2.114757094329874, |
|
"learning_rate": 8.78435876002496e-06, |
|
"loss": 1.0295, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"grad_norm": 1.9082626009926926, |
|
"learning_rate": 8.745559912046625e-06, |
|
"loss": 1.033, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 1.730783726030946, |
|
"learning_rate": 8.706240177667003e-06, |
|
"loss": 1.0385, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 2.5320093097552614, |
|
"learning_rate": 8.666405024991105e-06, |
|
"loss": 1.0375, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 3.400617294472977, |
|
"learning_rate": 8.626059993801988e-06, |
|
"loss": 1.0253, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 2.7269138208449157, |
|
"learning_rate": 8.585210694790333e-06, |
|
"loss": 1.0191, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 1.943069131491427, |
|
"learning_rate": 8.543862808774193e-06, |
|
"loss": 1.021, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 1.8339225500743817, |
|
"learning_rate": 8.502022085908963e-06, |
|
"loss": 1.0164, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 2.137449239722819, |
|
"learning_rate": 8.459694344887732e-06, |
|
"loss": 1.0367, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"grad_norm": 2.581812285540177, |
|
"learning_rate": 8.416885472132077e-06, |
|
"loss": 1.0283, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 1.8160848127506812, |
|
"learning_rate": 8.373601420973464e-06, |
|
"loss": 1.017, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 2.346923224689188, |
|
"learning_rate": 8.329848210825322e-06, |
|
"loss": 1.0237, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 2.024470398972999, |
|
"learning_rate": 8.285631926345943e-06, |
|
"loss": 1.028, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 2.382418457888275, |
|
"learning_rate": 8.240958716592304e-06, |
|
"loss": 1.0103, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 1.754061139243149, |
|
"learning_rate": 8.195834794164925e-06, |
|
"loss": 1.0179, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 1.6884398821609177, |
|
"learning_rate": 8.150266434343904e-06, |
|
"loss": 1.0218, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"grad_norm": 1.8408143701714872, |
|
"learning_rate": 8.104259974216219e-06, |
|
"loss": 1.0109, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 1.8898424458167393, |
|
"learning_rate": 8.057821811794457e-06, |
|
"loss": 1.0113, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 2.0040005987182923, |
|
"learning_rate": 8.010958405127048e-06, |
|
"loss": 1.0205, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 1.8031516449231006, |
|
"learning_rate": 7.963676271400158e-06, |
|
"loss": 1.0102, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 2.0776879759963798, |
|
"learning_rate": 7.915981986031367e-06, |
|
"loss": 1.0089, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 2.4545184199390695, |
|
"learning_rate": 7.86788218175523e-06, |
|
"loss": 1.021, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 1.9664285212166526, |
|
"learning_rate": 7.819383547700889e-06, |
|
"loss": 1.008, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 2.7035372351508404, |
|
"learning_rate": 7.770492828461824e-06, |
|
"loss": 1.013, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 2.737641626508961, |
|
"learning_rate": 7.721216823157896e-06, |
|
"loss": 1.0125, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 1.7709332048302728, |
|
"learning_rate": 7.671562384489819e-06, |
|
"loss": 1.0197, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 1.592514417019854, |
|
"learning_rate": 7.621536417786159e-06, |
|
"loss": 1.006, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 1.8116371870435586, |
|
"learning_rate": 7.571145880043036e-06, |
|
"loss": 1.0021, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 2.742484137764987, |
|
"learning_rate": 7.520397778956623e-06, |
|
"loss": 0.9942, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 1.8390385024848412, |
|
"learning_rate": 7.469299171948608e-06, |
|
"loss": 1.0085, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 1.8627501795637935, |
|
"learning_rate": 7.417857165184723e-06, |
|
"loss": 1.0065, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 1.8662148392820241, |
|
"learning_rate": 7.366078912586523e-06, |
|
"loss": 1.0069, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"grad_norm": 1.9898176799731553, |
|
"learning_rate": 7.313971614836496e-06, |
|
"loss": 1.0075, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 1.6374351854744644, |
|
"learning_rate": 7.261542518376677e-06, |
|
"loss": 0.9954, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 2.3713553804160554, |
|
"learning_rate": 7.208798914400916e-06, |
|
"loss": 1.0041, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"grad_norm": 2.43390000834423, |
|
"learning_rate": 7.155748137840892e-06, |
|
"loss": 1.0043, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 1.9046613363144396, |
|
"learning_rate": 7.102397566346073e-06, |
|
"loss": 0.9984, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"grad_norm": 1.667190267378518, |
|
"learning_rate": 7.048754619257716e-06, |
|
"loss": 1.0029, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 1.638081070696857, |
|
"learning_rate": 6.994826756577082e-06, |
|
"loss": 1.0032, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 2.2506476935835407, |
|
"learning_rate": 6.940621477927988e-06, |
|
"loss": 1.0023, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 2.0559944356090116, |
|
"learning_rate": 6.88614632151385e-06, |
|
"loss": 1.0036, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 1.7722705407021138, |
|
"learning_rate": 6.831408863069364e-06, |
|
"loss": 0.9981, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"grad_norm": 1.510560677423063, |
|
"learning_rate": 6.7764167148069695e-06, |
|
"loss": 0.9845, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 1.6159682830961077, |
|
"learning_rate": 6.721177524358226e-06, |
|
"loss": 1.0014, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 2.547134856734226, |
|
"learning_rate": 6.665698973710289e-06, |
|
"loss": 0.9955, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 2.6937900779118524, |
|
"learning_rate": 6.609988778137582e-06, |
|
"loss": 0.9856, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 2.0660116621981937, |
|
"learning_rate": 6.554054685128857e-06, |
|
"loss": 0.987, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 1.7256065453185816, |
|
"learning_rate": 6.497904473309766e-06, |
|
"loss": 0.9907, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 1.9848736194116567, |
|
"learning_rate": 6.44154595136111e-06, |
|
"loss": 0.9938, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 2.226796297825019, |
|
"learning_rate": 6.384986956932897e-06, |
|
"loss": 0.9891, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 1.7756206272082837, |
|
"learning_rate": 6.328235355554382e-06, |
|
"loss": 0.989, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 2.055029436503266, |
|
"learning_rate": 6.271299039540228e-06, |
|
"loss": 0.9854, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 2.1572868944232733, |
|
"learning_rate": 6.214185926892936e-06, |
|
"loss": 0.991, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 1.705088008506106, |
|
"learning_rate": 6.156903960201709e-06, |
|
"loss": 0.99, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 2.394966448163764, |
|
"learning_rate": 6.099461105537889e-06, |
|
"loss": 0.9866, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 2.016768798949277, |
|
"learning_rate": 6.041865351347146e-06, |
|
"loss": 0.9919, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 1.8476093711911672, |
|
"learning_rate": 5.984124707338528e-06, |
|
"loss": 0.9932, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 1.9896940215311267, |
|
"learning_rate": 5.926247203370583e-06, |
|
"loss": 0.9958, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 2.5258042306430637, |
|
"learning_rate": 5.8682408883346535e-06, |
|
"loss": 0.9915, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 2.300171514270174, |
|
"learning_rate": 5.810113829035544e-06, |
|
"loss": 0.9681, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 1.9721303641863295, |
|
"learning_rate": 5.751874109069685e-06, |
|
"loss": 0.9597, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"grad_norm": 2.2390559024844086, |
|
"learning_rate": 5.693529827700967e-06, |
|
"loss": 0.9578, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 2.0408626045391167, |
|
"learning_rate": 5.635089098734394e-06, |
|
"loss": 0.9759, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 2.3474388556665353, |
|
"learning_rate": 5.576560049387717e-06, |
|
"loss": 0.9692, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 1.577362362698819, |
|
"learning_rate": 5.517950819161196e-06, |
|
"loss": 0.9693, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 1.798565847037349, |
|
"learning_rate": 5.459269558705667e-06, |
|
"loss": 0.9583, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"grad_norm": 1.9304589279981041, |
|
"learning_rate": 5.400524428689035e-06, |
|
"loss": 0.9717, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 1.721891020322892, |
|
"learning_rate": 5.341723598661409e-06, |
|
"loss": 0.9639, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"grad_norm": 1.7487840479558292, |
|
"learning_rate": 5.282875245918963e-06, |
|
"loss": 0.9705, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 1.8716448470255413, |
|
"learning_rate": 5.2239875543667465e-06, |
|
"loss": 0.9714, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 1.5692767824863278, |
|
"learning_rate": 5.165068713380568e-06, |
|
"loss": 0.9728, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 1.6666960593346472, |
|
"learning_rate": 5.106126916668118e-06, |
|
"loss": 0.9634, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 1.6602656394819693, |
|
"learning_rate": 5.047170361129484e-06, |
|
"loss": 0.9665, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 1.5606922426383971, |
|
"learning_rate": 4.988207245717232e-06, |
|
"loss": 0.9615, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 1.8903502303369755, |
|
"learning_rate": 4.929245770296191e-06, |
|
"loss": 0.9517, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 2.0904161072997076, |
|
"learning_rate": 4.870294134503123e-06, |
|
"loss": 0.9578, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 1.8320825036562747, |
|
"learning_rate": 4.811360536606416e-06, |
|
"loss": 0.9695, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 1.9133665529378556, |
|
"learning_rate": 4.752453172365966e-06, |
|
"loss": 0.9585, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 1.8758941343800197, |
|
"learning_rate": 4.69358023389342e-06, |
|
"loss": 0.9608, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"grad_norm": 2.133289277786595, |
|
"learning_rate": 4.634749908512907e-06, |
|
"loss": 0.9619, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 1.8131410749776957, |
|
"learning_rate": 4.575970377622456e-06, |
|
"loss": 0.9623, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 2.0361433234498514, |
|
"learning_rate": 4.517249815556219e-06, |
|
"loss": 0.9572, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 1.8450224454341402, |
|
"learning_rate": 4.458596388447691e-06, |
|
"loss": 0.9588, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 1.5855628731202065, |
|
"learning_rate": 4.400018253094065e-06, |
|
"loss": 0.9544, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"grad_norm": 1.7015227109739095, |
|
"learning_rate": 4.341523555821882e-06, |
|
"loss": 0.9629, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 1.533925044445312, |
|
"learning_rate": 4.283120431354137e-06, |
|
"loss": 0.9571, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 1.5138597264361546, |
|
"learning_rate": 4.224817001679011e-06, |
|
"loss": 0.9642, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 2.3034815379141795, |
|
"learning_rate": 4.1666213749203545e-06, |
|
"loss": 0.9562, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 1.6982866609702905, |
|
"learning_rate": 4.1085416442101205e-06, |
|
"loss": 0.9628, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 1.8780386107869884, |
|
"learning_rate": 4.050585886562858e-06, |
|
"loss": 0.9549, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 1.683670163601615, |
|
"learning_rate": 3.992762161752474e-06, |
|
"loss": 0.9615, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 2.001515746557216, |
|
"learning_rate": 3.935078511191368e-06, |
|
"loss": 0.9617, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 2.4485180814449925, |
|
"learning_rate": 3.877542956812137e-06, |
|
"loss": 0.9599, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 2.2535920157325373, |
|
"learning_rate": 3.820163499951984e-06, |
|
"loss": 0.9485, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"grad_norm": 1.6920809388339335, |
|
"learning_rate": 3.7629481202399886e-06, |
|
"loss": 0.9464, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 1.9251860711568731, |
|
"learning_rate": 3.705904774487396e-06, |
|
"loss": 0.9565, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 1.6963234663767837, |
|
"learning_rate": 3.64904139558109e-06, |
|
"loss": 0.9657, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 1.723192642702987, |
|
"learning_rate": 3.5923658913803726e-06, |
|
"loss": 0.9545, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 1.857223252079437, |
|
"learning_rate": 3.5358861436172487e-06, |
|
"loss": 0.951, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 1.9463652794821473, |
|
"learning_rate": 3.47961000680032e-06, |
|
"loss": 0.9545, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 1.6275095264739814, |
|
"learning_rate": 3.4235453071224882e-06, |
|
"loss": 0.956, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 2.132166681197278, |
|
"learning_rate": 3.3676998413725726e-06, |
|
"loss": 0.9597, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 2.389342012426533, |
|
"learning_rate": 3.3120813758510385e-06, |
|
"loss": 0.9444, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 2.1360938817924424, |
|
"learning_rate": 3.2566976452899507e-06, |
|
"loss": 0.9488, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 2.1039499268212323, |
|
"learning_rate": 3.2015563517773214e-06, |
|
"loss": 0.953, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 2.310285226796488, |
|
"learning_rate": 3.1466651636860025e-06, |
|
"loss": 0.9617, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 2.0128464455024817, |
|
"learning_rate": 3.0920317146072577e-06, |
|
"loss": 0.9516, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 1.4735182975439758, |
|
"learning_rate": 3.0376636022891813e-06, |
|
"loss": 0.9575, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 2.318196631109318, |
|
"learning_rate": 2.983568387580094e-06, |
|
"loss": 0.9536, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 1.616843986566008, |
|
"learning_rate": 2.9297535933770732e-06, |
|
"loss": 0.9474, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 1.7693581907423008, |
|
"learning_rate": 2.8762267035797607e-06, |
|
"loss": 0.95, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 2.122336403201617, |
|
"learning_rate": 2.822995162049599e-06, |
|
"loss": 0.9505, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 1.5679869228768337, |
|
"learning_rate": 2.7700663715746213e-06, |
|
"loss": 0.9474, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 1.5424846799708574, |
|
"learning_rate": 2.7174476928399685e-06, |
|
"loss": 0.9483, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 1.654933401000696, |
|
"learning_rate": 2.66514644340426e-06, |
|
"loss": 0.9521, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 1.5993372025284307, |
|
"learning_rate": 2.613169896681949e-06, |
|
"loss": 0.9597, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 1.780666444469016, |
|
"learning_rate": 2.5615252809318287e-06, |
|
"loss": 0.9488, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 1.5279733243149005, |
|
"learning_rate": 2.5102197782518145e-06, |
|
"loss": 0.9438, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 1.8194536821858915, |
|
"learning_rate": 2.4592605235801544e-06, |
|
"loss": 0.943, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 1.6480818333921619, |
|
"learning_rate": 2.4086546037031734e-06, |
|
"loss": 0.9437, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 1.9968939308919122, |
|
"learning_rate": 2.3584090562697427e-06, |
|
"loss": 0.949, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 1.712960437774201, |
|
"learning_rate": 2.3085308688125695e-06, |
|
"loss": 0.9522, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 1.7812514943208981, |
|
"learning_rate": 2.2590269777764516e-06, |
|
"loss": 0.9503, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 1.4783532304720237, |
|
"learning_rate": 2.2099042675536437e-06, |
|
"loss": 0.9565, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 1.4892687284353463, |
|
"learning_rate": 2.161169569526461e-06, |
|
"loss": 0.9481, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 1.5738512766968247, |
|
"learning_rate": 2.1128296611172593e-06, |
|
"loss": 0.9483, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"grad_norm": 1.7245957459593824, |
|
"learning_rate": 2.0648912648459072e-06, |
|
"loss": 0.9424, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"grad_norm": 1.6116002624575987, |
|
"learning_rate": 2.0173610473949048e-06, |
|
"loss": 0.9462, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 1.5813885175238174, |
|
"learning_rate": 1.9702456186822595e-06, |
|
"loss": 0.9443, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 1.6220303849517923, |
|
"learning_rate": 1.9235515309422685e-06, |
|
"loss": 0.9428, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 1.847653265677833, |
|
"learning_rate": 1.8772852778143064e-06, |
|
"loss": 0.9431, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 1.7766441613041704, |
|
"learning_rate": 1.831453293439771e-06, |
|
"loss": 0.9371, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 1.7572717020990576, |
|
"learning_rate": 1.7860619515673034e-06, |
|
"loss": 0.9455, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"grad_norm": 2.0366564271261396, |
|
"learning_rate": 1.7411175646664103e-06, |
|
"loss": 0.9415, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 1.641890714375309, |
|
"learning_rate": 1.6966263830495939e-06, |
|
"loss": 0.9245, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"grad_norm": 1.7389532361568043, |
|
"learning_rate": 1.6525945940031407e-06, |
|
"loss": 0.9387, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"grad_norm": 1.6927577102388072, |
|
"learning_rate": 1.6090283209266682e-06, |
|
"loss": 0.9358, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"grad_norm": 1.6810193706370176, |
|
"learning_rate": 1.5659336224815642e-06, |
|
"loss": 0.9302, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"grad_norm": 1.5836446992301734, |
|
"learning_rate": 1.5233164917484117e-06, |
|
"loss": 0.9339, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"grad_norm": 1.4925374056585887, |
|
"learning_rate": 1.4811828553935498e-06, |
|
"loss": 0.9359, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"grad_norm": 1.562133820438609, |
|
"learning_rate": 1.439538572844873e-06, |
|
"loss": 0.9265, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"grad_norm": 1.508715392161331, |
|
"learning_rate": 1.3983894354769616e-06, |
|
"loss": 0.94, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"grad_norm": 1.4833918032553421, |
|
"learning_rate": 1.3577411658056965e-06, |
|
"loss": 0.9359, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"grad_norm": 1.7381479633485923, |
|
"learning_rate": 1.3175994166924394e-06, |
|
"loss": 0.9292, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"grad_norm": 2.4773729170453276, |
|
"learning_rate": 1.2779697705579058e-06, |
|
"loss": 0.935, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"grad_norm": 1.6423953500818698, |
|
"learning_rate": 1.2388577386058248e-06, |
|
"loss": 0.9281, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"grad_norm": 1.5747180314126341, |
|
"learning_rate": 1.2002687600565138e-06, |
|
"loss": 0.9251, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"grad_norm": 1.4796953478538386, |
|
"learning_rate": 1.1622082013904535e-06, |
|
"loss": 0.9309, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"grad_norm": 1.9087071208773028, |
|
"learning_rate": 1.1246813556019925e-06, |
|
"loss": 0.9302, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"grad_norm": 1.441416955211069, |
|
"learning_rate": 1.0876934414632523e-06, |
|
"loss": 0.9222, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"grad_norm": 1.4528421452118758, |
|
"learning_rate": 1.0512496027983715e-06, |
|
"loss": 0.9358, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"grad_norm": 1.4391309054269417, |
|
"learning_rate": 1.0153549077681617e-06, |
|
"loss": 0.9294, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"grad_norm": 1.5484810522299401, |
|
"learning_rate": 9.80014348165298e-07, |
|
"loss": 0.9258, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"grad_norm": 1.5093687112295802, |
|
"learning_rate": 9.452328387201104e-07, |
|
"loss": 0.9294, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"grad_norm": 1.4811632926742155, |
|
"learning_rate": 9.110152164171127e-07, |
|
"loss": 0.9291, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"grad_norm": 1.6851965695398232, |
|
"learning_rate": 8.773662398223276e-07, |
|
"loss": 0.9315, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"grad_norm": 1.5680181843321963, |
|
"learning_rate": 8.44290588421533e-07, |
|
"loss": 0.9245, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"grad_norm": 1.501501151688511, |
|
"learning_rate": 8.117928619694848e-07, |
|
"loss": 0.9312, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"grad_norm": 1.5069447398439229, |
|
"learning_rate": 7.798775798502484e-07, |
|
"loss": 0.9256, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"grad_norm": 1.4179041700918098, |
|
"learning_rate": 7.485491804486972e-07, |
|
"loss": 0.9209, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"grad_norm": 1.388023240525021, |
|
"learning_rate": 7.178120205332717e-07, |
|
"loss": 0.9311, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"grad_norm": 1.6540883448168693, |
|
"learning_rate": 6.876703746500984e-07, |
|
"loss": 0.9398, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"grad_norm": 1.3848004961206983, |
|
"learning_rate": 6.581284345285371e-07, |
|
"loss": 0.9316, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"grad_norm": 1.3665324107772734, |
|
"learning_rate": 6.291903084982481e-07, |
|
"loss": 0.9301, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"grad_norm": 1.4952001545175686, |
|
"learning_rate": 6.008600209178539e-07, |
|
"loss": 0.9375, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"grad_norm": 1.4437175986820594, |
|
"learning_rate": 5.7314151161528e-07, |
|
"loss": 0.9324, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"grad_norm": 1.3632138123352606, |
|
"learning_rate": 5.460386353398583e-07, |
|
"loss": 0.9355, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"grad_norm": 1.4355913142058574, |
|
"learning_rate": 5.195551612262478e-07, |
|
"loss": 0.9369, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"grad_norm": 1.3593932016675367, |
|
"learning_rate": 4.936947722702762e-07, |
|
"loss": 0.9248, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 1.5313367509871634, |
|
"learning_rate": 4.6846106481675035e-07, |
|
"loss": 0.9313, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"grad_norm": 1.5679931446767343, |
|
"learning_rate": 4.43857548059321e-07, |
|
"loss": 0.9241, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"grad_norm": 1.4231649052268165, |
|
"learning_rate": 4.198876435524718e-07, |
|
"loss": 0.9288, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"grad_norm": 1.5147664043720925, |
|
"learning_rate": 3.9655468473568436e-07, |
|
"loss": 0.937, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"grad_norm": 1.505384912577037, |
|
"learning_rate": 3.7386191646987094e-07, |
|
"loss": 0.9362, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"grad_norm": 1.7241739930760707, |
|
"learning_rate": 3.51812494586114e-07, |
|
"loss": 0.9209, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"grad_norm": 1.498479993175917, |
|
"learning_rate": 3.3040948544679817e-07, |
|
"loss": 0.9323, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"grad_norm": 1.464257082461974, |
|
"learning_rate": 3.096558655191706e-07, |
|
"loss": 0.9365, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"grad_norm": 1.4384497143831931, |
|
"learning_rate": 2.895545209614176e-07, |
|
"loss": 0.939, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"grad_norm": 1.4541548130541968, |
|
"learning_rate": 2.701082472212879e-07, |
|
"loss": 0.9277, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"grad_norm": 1.5175629961859558, |
|
"learning_rate": 2.5131974864734063e-07, |
|
"loss": 0.9279, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"grad_norm": 1.4611673173663144, |
|
"learning_rate": 2.331916381128535e-07, |
|
"loss": 0.9268, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"grad_norm": 1.3907174934591902, |
|
"learning_rate": 2.157264366524603e-07, |
|
"loss": 0.9297, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"grad_norm": 1.6602489752653944, |
|
"learning_rate": 1.989265731115525e-07, |
|
"loss": 0.9279, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"grad_norm": 1.4335900513136002, |
|
"learning_rate": 1.827943838085111e-07, |
|
"loss": 0.9261, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"grad_norm": 1.4382877939912813, |
|
"learning_rate": 1.6733211220979316e-07, |
|
"loss": 0.9286, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"grad_norm": 1.5442925793363935, |
|
"learning_rate": 1.5254190861794415e-07, |
|
"loss": 0.931, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"grad_norm": 1.474749185950876, |
|
"learning_rate": 1.3842582987255494e-07, |
|
"loss": 0.9237, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"grad_norm": 1.3949775304121632, |
|
"learning_rate": 1.2498583906422779e-07, |
|
"loss": 0.9264, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"grad_norm": 1.4864944753529736, |
|
"learning_rate": 1.1222380526156929e-07, |
|
"loss": 0.9337, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"grad_norm": 1.4352950913827418, |
|
"learning_rate": 1.0014150325126315e-07, |
|
"loss": 0.9296, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"grad_norm": 1.54647479187705, |
|
"learning_rate": 8.874061329125939e-08, |
|
"loss": 0.9284, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"grad_norm": 1.5426799677559493, |
|
"learning_rate": 7.802272087709951e-08, |
|
"loss": 0.9267, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"grad_norm": 1.4345720592945528, |
|
"learning_rate": 6.798931652142738e-08, |
|
"loss": 0.9412, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"grad_norm": 1.4666166802740588, |
|
"learning_rate": 5.864179554670724e-08, |
|
"loss": 0.9274, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"grad_norm": 1.37454276267336, |
|
"learning_rate": 4.998145789118114e-08, |
|
"loss": 0.9347, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"grad_norm": 1.3537939594957207, |
|
"learning_rate": 4.2009507928084116e-08, |
|
"loss": 0.9374, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"grad_norm": 1.5201092358761166, |
|
"learning_rate": 3.4727054298161475e-08, |
|
"loss": 0.9278, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"grad_norm": 1.3548157698669074, |
|
"learning_rate": 2.8135109755487723e-08, |
|
"loss": 0.9229, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"grad_norm": 1.413302347919726, |
|
"learning_rate": 2.223459102662695e-08, |
|
"loss": 0.9341, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"grad_norm": 1.3765482397349185, |
|
"learning_rate": 1.7026318683147082e-08, |
|
"loss": 0.9289, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"grad_norm": 1.3538302103790776, |
|
"learning_rate": 1.2511017027501682e-08, |
|
"loss": 0.9233, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"grad_norm": 1.466960745073582, |
|
"learning_rate": 8.689313992306104e-09, |
|
"loss": 0.9303, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"grad_norm": 1.3714566713014886, |
|
"learning_rate": 5.561741053010661e-09, |
|
"loss": 0.936, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"grad_norm": 1.4859645171436624, |
|
"learning_rate": 3.1287331539903155e-09, |
|
"loss": 0.9247, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"grad_norm": 1.386772570990769, |
|
"learning_rate": 1.3906286480563913e-09, |
|
"loss": 0.9195, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"grad_norm": 1.3873824123701393, |
|
"learning_rate": 3.4766924940476954e-10, |
|
"loss": 0.9374, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"grad_norm": 1.492034850579766, |
|
"learning_rate": 0.0, |
|
"loss": 0.9213, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"step": 1480, |
|
"total_flos": 1630317777518592.0, |
|
"train_loss": 1.3370767467730753, |
|
"train_runtime": 16871.6289, |
|
"train_samples_per_second": 44.965, |
|
"train_steps_per_second": 0.088 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 1480, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 500, |
|
"total_flos": 1630317777518592.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|