|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 972, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.030864197530864196, |
|
"grad_norm": 5.446850666927686, |
|
"learning_rate": 5e-06, |
|
"loss": 0.9199, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06172839506172839, |
|
"grad_norm": 2.663126142630872, |
|
"learning_rate": 5e-06, |
|
"loss": 0.821, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.09259259259259259, |
|
"grad_norm": 1.1642696107304236, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7706, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.12345679012345678, |
|
"grad_norm": 4.280110588222716, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7596, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.15432098765432098, |
|
"grad_norm": 4.412787627263132, |
|
"learning_rate": 5e-06, |
|
"loss": 0.746, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.18518518518518517, |
|
"grad_norm": 1.5711972402992538, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7267, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.21604938271604937, |
|
"grad_norm": 0.9451756673135364, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7186, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.24691358024691357, |
|
"grad_norm": 0.9020860720269491, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7099, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.2777777777777778, |
|
"grad_norm": 0.7223983788610128, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7001, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.30864197530864196, |
|
"grad_norm": 0.8632286972903407, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6965, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3395061728395062, |
|
"grad_norm": 0.6600509272889379, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6878, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.37037037037037035, |
|
"grad_norm": 0.8445466577407515, |
|
"learning_rate": 5e-06, |
|
"loss": 0.685, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.4012345679012346, |
|
"grad_norm": 0.7606146778204088, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6826, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.43209876543209874, |
|
"grad_norm": 0.6093510086614801, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6828, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.46296296296296297, |
|
"grad_norm": 0.7249517817925787, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6808, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.49382716049382713, |
|
"grad_norm": 0.559508936091321, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6806, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5246913580246914, |
|
"grad_norm": 0.6485112723957187, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6755, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 0.5734053780909245, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6676, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.5864197530864198, |
|
"grad_norm": 0.7330429838481911, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6695, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.6172839506172839, |
|
"grad_norm": 0.7031639101048125, |
|
"learning_rate": 5e-06, |
|
"loss": 0.669, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.6481481481481481, |
|
"grad_norm": 0.5618292117245931, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6759, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.6790123456790124, |
|
"grad_norm": 0.5384920681614226, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6626, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.7098765432098766, |
|
"grad_norm": 0.8695843119793333, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6613, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.7407407407407407, |
|
"grad_norm": 0.8530303198305609, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6622, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.7716049382716049, |
|
"grad_norm": 0.7246183829809985, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6585, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.8024691358024691, |
|
"grad_norm": 0.6467615123805251, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6663, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.8333333333333334, |
|
"grad_norm": 0.6051023696593527, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6578, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.8641975308641975, |
|
"grad_norm": 0.5325735601636417, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6628, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.8950617283950617, |
|
"grad_norm": 0.5624534504885511, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6551, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.9259259259259259, |
|
"grad_norm": 0.5105392081865389, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6585, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.9567901234567902, |
|
"grad_norm": 0.7882251080908214, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6578, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.9876543209876543, |
|
"grad_norm": 0.7218610961797266, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6542, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.6550179123878479, |
|
"eval_runtime": 31.8484, |
|
"eval_samples_per_second": 273.641, |
|
"eval_steps_per_second": 1.099, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.0185185185185186, |
|
"grad_norm": 0.7148119037457841, |
|
"learning_rate": 5e-06, |
|
"loss": 0.624, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.0493827160493827, |
|
"grad_norm": 0.8285481607039845, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6138, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.0802469135802468, |
|
"grad_norm": 0.5405908730469038, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6154, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.1111111111111112, |
|
"grad_norm": 0.6837013751545337, |
|
"learning_rate": 5e-06, |
|
"loss": 0.613, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.1419753086419753, |
|
"grad_norm": 0.8086734597067436, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6051, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.1728395061728394, |
|
"grad_norm": 0.6168727254851533, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6047, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.2037037037037037, |
|
"grad_norm": 0.6222564529681911, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6135, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.2345679012345678, |
|
"grad_norm": 0.6502199448782471, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6092, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.2654320987654322, |
|
"grad_norm": 0.642957942043697, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6151, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.2962962962962963, |
|
"grad_norm": 0.6973190504371732, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6087, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.3271604938271606, |
|
"grad_norm": 0.5681720826825847, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6058, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.3580246913580247, |
|
"grad_norm": 0.5560347945291988, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6152, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.3888888888888888, |
|
"grad_norm": 0.5202152575422632, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6112, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.4197530864197532, |
|
"grad_norm": 0.6652453548193578, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6084, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.4506172839506173, |
|
"grad_norm": 0.5959726615804369, |
|
"learning_rate": 5e-06, |
|
"loss": 0.611, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.4814814814814814, |
|
"grad_norm": 0.5753042580268376, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6099, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.5123456790123457, |
|
"grad_norm": 0.5657589973831328, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6158, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.5432098765432098, |
|
"grad_norm": 0.48962281372551314, |
|
"learning_rate": 5e-06, |
|
"loss": 0.612, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.574074074074074, |
|
"grad_norm": 0.5867122795097997, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6092, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.6049382716049383, |
|
"grad_norm": 0.567887162579392, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6049, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.6358024691358026, |
|
"grad_norm": 0.5846370116192817, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6102, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.6666666666666665, |
|
"grad_norm": 0.7372794706158421, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6121, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.6975308641975309, |
|
"grad_norm": 0.5679832644741175, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6039, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.7283950617283952, |
|
"grad_norm": 0.4951982048137007, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6097, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.7592592592592593, |
|
"grad_norm": 0.5947044750562988, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6108, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.7901234567901234, |
|
"grad_norm": 0.5672737096963988, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6095, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.8209876543209877, |
|
"grad_norm": 0.5215432816515835, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6096, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.8518518518518519, |
|
"grad_norm": 0.5356125614039279, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6037, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.882716049382716, |
|
"grad_norm": 0.5290482585426443, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6062, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.9135802469135803, |
|
"grad_norm": 0.6241816332212873, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6106, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.9444444444444444, |
|
"grad_norm": 0.4959658493368521, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6067, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.9753086419753085, |
|
"grad_norm": 0.5282675010462219, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6069, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.6462317109107971, |
|
"eval_runtime": 32.1855, |
|
"eval_samples_per_second": 270.774, |
|
"eval_steps_per_second": 1.087, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 2.006172839506173, |
|
"grad_norm": 0.7935581137695162, |
|
"learning_rate": 5e-06, |
|
"loss": 0.596, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.037037037037037, |
|
"grad_norm": 0.6261816641543, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5645, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.067901234567901, |
|
"grad_norm": 0.5849160839297688, |
|
"learning_rate": 5e-06, |
|
"loss": 0.558, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.0987654320987654, |
|
"grad_norm": 0.6105130235265663, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5591, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.1296296296296298, |
|
"grad_norm": 0.6033282280363377, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5573, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.1604938271604937, |
|
"grad_norm": 0.7739145588445309, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5571, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.191358024691358, |
|
"grad_norm": 0.626140801689634, |
|
"learning_rate": 5e-06, |
|
"loss": 0.56, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.2222222222222223, |
|
"grad_norm": 0.5916889343710152, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5571, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.253086419753086, |
|
"grad_norm": 0.636584234517333, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5638, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.2839506172839505, |
|
"grad_norm": 0.5507349524263677, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5724, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.314814814814815, |
|
"grad_norm": 0.6732342797910674, |
|
"learning_rate": 5e-06, |
|
"loss": 0.567, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.3456790123456788, |
|
"grad_norm": 0.5395015214035337, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5569, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.376543209876543, |
|
"grad_norm": 0.5992087208696053, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5647, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.4074074074074074, |
|
"grad_norm": 0.5102699543572234, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5622, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.4382716049382718, |
|
"grad_norm": 0.6025610989122395, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5634, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.4691358024691357, |
|
"grad_norm": 0.6231368998767735, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5594, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.5707370285498646, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5685, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.5308641975308643, |
|
"grad_norm": 0.5549089552658856, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5629, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.5617283950617287, |
|
"grad_norm": 0.542471712528822, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5681, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.5925925925925926, |
|
"grad_norm": 0.5566494869370542, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5663, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.623456790123457, |
|
"grad_norm": 0.5552433471790524, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5635, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.6543209876543212, |
|
"grad_norm": 0.5049248358370862, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5654, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.685185185185185, |
|
"grad_norm": 0.6071017582058569, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5669, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.7160493827160495, |
|
"grad_norm": 0.5625590442980464, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5642, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.746913580246914, |
|
"grad_norm": 0.5506548153090438, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5638, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.7777777777777777, |
|
"grad_norm": 0.6470830808425209, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5629, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.808641975308642, |
|
"grad_norm": 0.6202475264816772, |
|
"learning_rate": 5e-06, |
|
"loss": 0.566, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.8395061728395063, |
|
"grad_norm": 0.6200010347166321, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5672, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.8703703703703702, |
|
"grad_norm": 0.5382597238916978, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5674, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.9012345679012346, |
|
"grad_norm": 0.6294871904049958, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5714, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 2.932098765432099, |
|
"grad_norm": 0.5413425077416242, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5677, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.962962962962963, |
|
"grad_norm": 0.5307261656256946, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5674, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 2.993827160493827, |
|
"grad_norm": 0.519858507326187, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5688, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.6513996720314026, |
|
"eval_runtime": 31.4148, |
|
"eval_samples_per_second": 277.417, |
|
"eval_steps_per_second": 1.114, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 972, |
|
"total_flos": 1628136202567680.0, |
|
"train_loss": 0.6230609294318368, |
|
"train_runtime": 6074.7502, |
|
"train_samples_per_second": 81.773, |
|
"train_steps_per_second": 0.16 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 972, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1628136202567680.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|