|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 9.984301412872842, |
|
"eval_steps": 500, |
|
"global_step": 1590, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06279434850863422, |
|
"grad_norm": 0.7715030908584595, |
|
"learning_rate": 0.00019998048082884911, |
|
"loss": 1.3075, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.12558869701726844, |
|
"grad_norm": 0.7010838389396667, |
|
"learning_rate": 0.0001999219309353572, |
|
"loss": 1.0321, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.18838304552590268, |
|
"grad_norm": 0.5988379120826721, |
|
"learning_rate": 0.00019982437317643217, |
|
"loss": 0.9725, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.25117739403453687, |
|
"grad_norm": 0.5993463397026062, |
|
"learning_rate": 0.00019968784563700586, |
|
"loss": 0.9429, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.3139717425431711, |
|
"grad_norm": 0.5684201121330261, |
|
"learning_rate": 0.0001995124016151664, |
|
"loss": 0.9221, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.37676609105180536, |
|
"grad_norm": 0.5648465752601624, |
|
"learning_rate": 0.00019929810960135172, |
|
"loss": 0.957, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.43956043956043955, |
|
"grad_norm": 0.5728427767753601, |
|
"learning_rate": 0.00019904505325161158, |
|
"loss": 0.8807, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.5023547880690737, |
|
"grad_norm": 0.5611833333969116, |
|
"learning_rate": 0.00019875333135495002, |
|
"loss": 0.8894, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.565149136577708, |
|
"grad_norm": 0.5440887808799744, |
|
"learning_rate": 0.00019842305779475968, |
|
"loss": 0.9077, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.6279434850863422, |
|
"grad_norm": 0.5734382271766663, |
|
"learning_rate": 0.00019805436150436352, |
|
"loss": 0.872, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.6907378335949764, |
|
"grad_norm": 0.5940066576004028, |
|
"learning_rate": 0.00019764738641668138, |
|
"loss": 0.8535, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.7535321821036107, |
|
"grad_norm": 0.5004596710205078, |
|
"learning_rate": 0.0001972022914080411, |
|
"loss": 0.8571, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.8163265306122449, |
|
"grad_norm": 0.49201518297195435, |
|
"learning_rate": 0.00019671925023615573, |
|
"loss": 0.8738, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.8791208791208791, |
|
"grad_norm": 0.5316975116729736, |
|
"learning_rate": 0.00019619845147229138, |
|
"loss": 0.8571, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.9419152276295133, |
|
"grad_norm": 0.4743081033229828, |
|
"learning_rate": 0.00019564009842765225, |
|
"loss": 0.8552, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.0047095761381475, |
|
"grad_norm": 0.488436758518219, |
|
"learning_rate": 0.0001950444090740111, |
|
"loss": 0.8686, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.0675039246467817, |
|
"grad_norm": 0.552605390548706, |
|
"learning_rate": 0.0001944116159586169, |
|
"loss": 0.7721, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.130298273155416, |
|
"grad_norm": 0.5751229524612427, |
|
"learning_rate": 0.0001937419661134121, |
|
"loss": 0.7748, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.1930926216640503, |
|
"grad_norm": 0.4798067808151245, |
|
"learning_rate": 0.00019303572095859546, |
|
"loss": 0.78, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.2558869701726845, |
|
"grad_norm": 0.5503143072128296, |
|
"learning_rate": 0.00019229315620056803, |
|
"loss": 0.7998, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.3186813186813187, |
|
"grad_norm": 0.5721454620361328, |
|
"learning_rate": 0.00019151456172430183, |
|
"loss": 0.787, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.3814756671899528, |
|
"grad_norm": 0.5621303915977478, |
|
"learning_rate": 0.00019070024148017375, |
|
"loss": 0.7816, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.4442700156985873, |
|
"grad_norm": 0.5912421941757202, |
|
"learning_rate": 0.00018985051336530798, |
|
"loss": 0.7806, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.5070643642072215, |
|
"grad_norm": 0.5405421257019043, |
|
"learning_rate": 0.00018896570909947475, |
|
"loss": 0.7548, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.5698587127158556, |
|
"grad_norm": 0.47901952266693115, |
|
"learning_rate": 0.00018804617409559198, |
|
"loss": 0.7776, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.6326530612244898, |
|
"grad_norm": 0.5640778541564941, |
|
"learning_rate": 0.00018709226732488215, |
|
"loss": 0.7683, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.695447409733124, |
|
"grad_norm": 0.5678220987319946, |
|
"learning_rate": 0.00018610436117673555, |
|
"loss": 0.7773, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.7582417582417582, |
|
"grad_norm": 0.5400906801223755, |
|
"learning_rate": 0.00018508284131333602, |
|
"loss": 0.7797, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.8210361067503924, |
|
"grad_norm": 0.5115251541137695, |
|
"learning_rate": 0.00018402810651910442, |
|
"loss": 0.78, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.8838304552590266, |
|
"grad_norm": 0.5241789221763611, |
|
"learning_rate": 0.0001829405685450202, |
|
"loss": 0.7978, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.9466248037676608, |
|
"grad_norm": 0.540064811706543, |
|
"learning_rate": 0.00018182065194788025, |
|
"loss": 0.7754, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.009419152276295, |
|
"grad_norm": 0.5090541243553162, |
|
"learning_rate": 0.0001806687939245593, |
|
"loss": 0.7707, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.072213500784929, |
|
"grad_norm": 0.6285854578018188, |
|
"learning_rate": 0.00017948544414133534, |
|
"loss": 0.6989, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.1350078492935634, |
|
"grad_norm": 0.5556398630142212, |
|
"learning_rate": 0.00017827106455834733, |
|
"loss": 0.681, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.197802197802198, |
|
"grad_norm": 0.6137024760246277, |
|
"learning_rate": 0.00017702612924925376, |
|
"loss": 0.7002, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.260596546310832, |
|
"grad_norm": 0.5647903680801392, |
|
"learning_rate": 0.00017575112421616202, |
|
"loss": 0.6796, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.3233908948194664, |
|
"grad_norm": 0.5665746927261353, |
|
"learning_rate": 0.00017444654719990128, |
|
"loss": 0.7247, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.3861852433281006, |
|
"grad_norm": 0.5704404711723328, |
|
"learning_rate": 0.00017311290748571275, |
|
"loss": 0.6982, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.4489795918367347, |
|
"grad_norm": 0.6268453001976013, |
|
"learning_rate": 0.00017175072570443312, |
|
"loss": 0.6788, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.511773940345369, |
|
"grad_norm": 0.5893650054931641, |
|
"learning_rate": 0.00017036053362924896, |
|
"loss": 0.7071, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.574568288854003, |
|
"grad_norm": 0.5590313673019409, |
|
"learning_rate": 0.0001689428739681012, |
|
"loss": 0.6977, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.6373626373626373, |
|
"grad_norm": 0.5971296429634094, |
|
"learning_rate": 0.00016749830015182107, |
|
"loss": 0.7079, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.7001569858712715, |
|
"grad_norm": 0.60088050365448, |
|
"learning_rate": 0.00016602737611807976, |
|
"loss": 0.6985, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.7629513343799057, |
|
"grad_norm": 0.5298681855201721, |
|
"learning_rate": 0.00016453067609123654, |
|
"loss": 0.7271, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.82574568288854, |
|
"grad_norm": 0.6201738119125366, |
|
"learning_rate": 0.00016300878435817113, |
|
"loss": 0.6905, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.8885400313971745, |
|
"grad_norm": 0.605073869228363, |
|
"learning_rate": 0.00016146229504018775, |
|
"loss": 0.7071, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.9513343799058083, |
|
"grad_norm": 0.5479874610900879, |
|
"learning_rate": 0.00015989181186108002, |
|
"loss": 0.7233, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 3.014128728414443, |
|
"grad_norm": 0.5628566145896912, |
|
"learning_rate": 0.0001582979479114472, |
|
"loss": 0.6769, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 3.076923076923077, |
|
"grad_norm": 0.6384280920028687, |
|
"learning_rate": 0.0001566813254093538, |
|
"loss": 0.6165, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 3.1397174254317113, |
|
"grad_norm": 0.6049516797065735, |
|
"learning_rate": 0.00015504257545742584, |
|
"loss": 0.6414, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 3.2025117739403455, |
|
"grad_norm": 0.6833707690238953, |
|
"learning_rate": 0.0001533823377964791, |
|
"loss": 0.6559, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 3.2653061224489797, |
|
"grad_norm": 0.6257244944572449, |
|
"learning_rate": 0.00015170126055577462, |
|
"loss": 0.6465, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 3.328100470957614, |
|
"grad_norm": 0.6548301577568054, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 0.6332, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 3.390894819466248, |
|
"grad_norm": 0.6910219788551331, |
|
"learning_rate": 0.00014827922027307451, |
|
"loss": 0.6331, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 3.4536891679748822, |
|
"grad_norm": 0.5907390713691711, |
|
"learning_rate": 0.00014653959313887813, |
|
"loss": 0.6106, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 3.5164835164835164, |
|
"grad_norm": 0.6401733756065369, |
|
"learning_rate": 0.00014478179771900632, |
|
"loss": 0.642, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 3.5792778649921506, |
|
"grad_norm": 0.5502592325210571, |
|
"learning_rate": 0.00014300652022765207, |
|
"loss": 0.6392, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 3.642072213500785, |
|
"grad_norm": 0.6283037066459656, |
|
"learning_rate": 0.00014121445370371923, |
|
"loss": 0.612, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 3.704866562009419, |
|
"grad_norm": 0.6005645990371704, |
|
"learning_rate": 0.0001394062977402717, |
|
"loss": 0.6289, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 3.767660910518053, |
|
"grad_norm": 0.6629378199577332, |
|
"learning_rate": 0.00013758275821142382, |
|
"loss": 0.6281, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 3.830455259026688, |
|
"grad_norm": 0.6070135235786438, |
|
"learning_rate": 0.00013574454699677893, |
|
"loss": 0.6308, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 3.8932496075353216, |
|
"grad_norm": 0.6706437468528748, |
|
"learning_rate": 0.00013389238170352318, |
|
"loss": 0.6474, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 3.956043956043956, |
|
"grad_norm": 0.591901957988739, |
|
"learning_rate": 0.00013202698538628376, |
|
"loss": 0.6418, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 4.01883830455259, |
|
"grad_norm": 0.5693519711494446, |
|
"learning_rate": 0.0001301490862648603, |
|
"loss": 0.6092, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 4.081632653061225, |
|
"grad_norm": 0.6287245154380798, |
|
"learning_rate": 0.0001282594174399399, |
|
"loss": 0.5511, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 4.144427001569858, |
|
"grad_norm": 0.6552153825759888, |
|
"learning_rate": 0.00012635871660690676, |
|
"loss": 0.5828, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 4.207221350078493, |
|
"grad_norm": 0.664696216583252, |
|
"learning_rate": 0.00012444772576785827, |
|
"loss": 0.5643, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 4.270015698587127, |
|
"grad_norm": 0.6447893381118774, |
|
"learning_rate": 0.0001225271909419395, |
|
"loss": 0.5883, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 4.332810047095761, |
|
"grad_norm": 0.6885414123535156, |
|
"learning_rate": 0.00012059786187410984, |
|
"loss": 0.5826, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 4.395604395604396, |
|
"grad_norm": 0.608504593372345, |
|
"learning_rate": 0.00011866049174245491, |
|
"loss": 0.5499, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 4.45839874411303, |
|
"grad_norm": 0.6774938702583313, |
|
"learning_rate": 0.00011671583686415832, |
|
"loss": 0.5785, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 4.521193092621664, |
|
"grad_norm": 0.6806432604789734, |
|
"learning_rate": 0.00011476465640024814, |
|
"loss": 0.5678, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 4.583987441130298, |
|
"grad_norm": 0.7246624827384949, |
|
"learning_rate": 0.00011280771205923268, |
|
"loss": 0.5797, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 4.646781789638933, |
|
"grad_norm": 0.6730565428733826, |
|
"learning_rate": 0.00011084576779974257, |
|
"loss": 0.565, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 4.7095761381475665, |
|
"grad_norm": 0.6672644019126892, |
|
"learning_rate": 0.00010887958953229349, |
|
"loss": 0.5624, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 4.772370486656201, |
|
"grad_norm": 0.7056964039802551, |
|
"learning_rate": 0.0001069099448202878, |
|
"loss": 0.587, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 4.835164835164835, |
|
"grad_norm": 0.7322887182235718, |
|
"learning_rate": 0.00010493760258037031, |
|
"loss": 0.5712, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 4.8979591836734695, |
|
"grad_norm": 0.7399300336837769, |
|
"learning_rate": 0.00010296333278225599, |
|
"loss": 0.5842, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 4.960753532182103, |
|
"grad_norm": 0.6447321176528931, |
|
"learning_rate": 0.00010098790614814658, |
|
"loss": 0.559, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 5.023547880690738, |
|
"grad_norm": 0.7017489075660706, |
|
"learning_rate": 9.901209385185345e-05, |
|
"loss": 0.5439, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 5.086342229199372, |
|
"grad_norm": 0.7099691033363342, |
|
"learning_rate": 9.703666721774402e-05, |
|
"loss": 0.5144, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 5.149136577708006, |
|
"grad_norm": 0.6762498617172241, |
|
"learning_rate": 9.506239741962971e-05, |
|
"loss": 0.5159, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 5.211930926216641, |
|
"grad_norm": 0.722254753112793, |
|
"learning_rate": 9.309005517971222e-05, |
|
"loss": 0.5174, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 5.274725274725275, |
|
"grad_norm": 0.7049859166145325, |
|
"learning_rate": 9.112041046770653e-05, |
|
"loss": 0.5067, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 5.337519623233909, |
|
"grad_norm": 0.6614891886711121, |
|
"learning_rate": 8.915423220025747e-05, |
|
"loss": 0.5178, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 5.400313971742543, |
|
"grad_norm": 0.7367340326309204, |
|
"learning_rate": 8.719228794076733e-05, |
|
"loss": 0.5158, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 5.463108320251178, |
|
"grad_norm": 0.7544198632240295, |
|
"learning_rate": 8.523534359975189e-05, |
|
"loss": 0.5181, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 5.525902668759811, |
|
"grad_norm": 0.7748332023620605, |
|
"learning_rate": 8.328416313584169e-05, |
|
"loss": 0.5211, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 5.588697017268446, |
|
"grad_norm": 0.7265179753303528, |
|
"learning_rate": 8.13395082575451e-05, |
|
"loss": 0.5005, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 5.65149136577708, |
|
"grad_norm": 0.7204436659812927, |
|
"learning_rate": 7.940213812589018e-05, |
|
"loss": 0.5034, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 5.714285714285714, |
|
"grad_norm": 0.7397501468658447, |
|
"learning_rate": 7.747280905806052e-05, |
|
"loss": 0.5204, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 5.777080062794348, |
|
"grad_norm": 0.7541511654853821, |
|
"learning_rate": 7.555227423214174e-05, |
|
"loss": 0.4947, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 5.839874411302983, |
|
"grad_norm": 0.7153120636940002, |
|
"learning_rate": 7.364128339309326e-05, |
|
"loss": 0.5189, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 5.9026687598116165, |
|
"grad_norm": 0.7152422070503235, |
|
"learning_rate": 7.174058256006012e-05, |
|
"loss": 0.5254, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 5.965463108320251, |
|
"grad_norm": 0.7143374681472778, |
|
"learning_rate": 6.985091373513972e-05, |
|
"loss": 0.5044, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 6.028257456828886, |
|
"grad_norm": 0.6690304279327393, |
|
"learning_rate": 6.797301461371625e-05, |
|
"loss": 0.482, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 6.0910518053375196, |
|
"grad_norm": 0.7355020642280579, |
|
"learning_rate": 6.610761829647685e-05, |
|
"loss": 0.4561, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 6.153846153846154, |
|
"grad_norm": 0.7926068902015686, |
|
"learning_rate": 6.425545300322112e-05, |
|
"loss": 0.4708, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 6.216640502354788, |
|
"grad_norm": 0.7191044688224792, |
|
"learning_rate": 6.24172417885762e-05, |
|
"loss": 0.4586, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 6.279434850863423, |
|
"grad_norm": 0.7463113069534302, |
|
"learning_rate": 6.0593702259728336e-05, |
|
"loss": 0.4674, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 6.342229199372056, |
|
"grad_norm": 0.743307888507843, |
|
"learning_rate": 5.8785546296280816e-05, |
|
"loss": 0.471, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 6.405023547880691, |
|
"grad_norm": 0.7751266956329346, |
|
"learning_rate": 5.699347977234799e-05, |
|
"loss": 0.4655, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 6.467817896389325, |
|
"grad_norm": 0.8425696492195129, |
|
"learning_rate": 5.5218202280993725e-05, |
|
"loss": 0.476, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 6.530612244897959, |
|
"grad_norm": 0.778333842754364, |
|
"learning_rate": 5.3460406861121894e-05, |
|
"loss": 0.4667, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 6.593406593406593, |
|
"grad_norm": 0.7623078227043152, |
|
"learning_rate": 5.172077972692553e-05, |
|
"loss": 0.4799, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 6.656200941915228, |
|
"grad_norm": 0.7711922526359558, |
|
"learning_rate": 5.000000000000002e-05, |
|
"loss": 0.4621, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 6.718995290423862, |
|
"grad_norm": 0.812324583530426, |
|
"learning_rate": 4.829873944422544e-05, |
|
"loss": 0.4612, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 6.781789638932496, |
|
"grad_norm": 0.7873876094818115, |
|
"learning_rate": 4.661766220352097e-05, |
|
"loss": 0.4794, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 6.84458398744113, |
|
"grad_norm": 0.8011544942855835, |
|
"learning_rate": 4.495742454257418e-05, |
|
"loss": 0.4608, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 6.9073783359497645, |
|
"grad_norm": 0.77611243724823, |
|
"learning_rate": 4.3318674590646237e-05, |
|
"loss": 0.471, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 6.970172684458399, |
|
"grad_norm": 0.8160433769226074, |
|
"learning_rate": 4.170205208855281e-05, |
|
"loss": 0.4717, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 7.032967032967033, |
|
"grad_norm": 0.7271689176559448, |
|
"learning_rate": 4.010818813892e-05, |
|
"loss": 0.4322, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 7.0957613814756675, |
|
"grad_norm": 0.8266540765762329, |
|
"learning_rate": 3.8537704959812294e-05, |
|
"loss": 0.4184, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 7.158555729984301, |
|
"grad_norm": 0.8125666975975037, |
|
"learning_rate": 3.69912156418289e-05, |
|
"loss": 0.4284, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 7.221350078492936, |
|
"grad_norm": 0.839213490486145, |
|
"learning_rate": 3.546932390876351e-05, |
|
"loss": 0.4389, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 7.28414442700157, |
|
"grad_norm": 0.89272141456604, |
|
"learning_rate": 3.397262388192029e-05, |
|
"loss": 0.4477, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 7.346938775510204, |
|
"grad_norm": 0.7432073950767517, |
|
"learning_rate": 3.250169984817897e-05, |
|
"loss": 0.43, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 7.409733124018838, |
|
"grad_norm": 0.7860047817230225, |
|
"learning_rate": 3.105712603189884e-05, |
|
"loss": 0.4269, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 7.472527472527473, |
|
"grad_norm": 0.832013726234436, |
|
"learning_rate": 2.9639466370751068e-05, |
|
"loss": 0.4256, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 7.535321821036106, |
|
"grad_norm": 0.7955825924873352, |
|
"learning_rate": 2.8249274295566864e-05, |
|
"loss": 0.4282, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 7.598116169544741, |
|
"grad_norm": 0.7972998023033142, |
|
"learning_rate": 2.688709251428725e-05, |
|
"loss": 0.4388, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 7.660910518053376, |
|
"grad_norm": 0.8105794191360474, |
|
"learning_rate": 2.555345280009872e-05, |
|
"loss": 0.4381, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 7.723704866562009, |
|
"grad_norm": 0.7821062207221985, |
|
"learning_rate": 2.4248875783837987e-05, |
|
"loss": 0.4276, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 7.786499215070644, |
|
"grad_norm": 0.805131733417511, |
|
"learning_rate": 2.297387075074625e-05, |
|
"loss": 0.4307, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 7.849293563579278, |
|
"grad_norm": 0.797322690486908, |
|
"learning_rate": 2.1728935441652686e-05, |
|
"loss": 0.4295, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 7.912087912087912, |
|
"grad_norm": 0.7943819761276245, |
|
"learning_rate": 2.0514555858664663e-05, |
|
"loss": 0.4346, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 7.974882260596546, |
|
"grad_norm": 0.814616322517395, |
|
"learning_rate": 1.93312060754407e-05, |
|
"loss": 0.4331, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 8.03767660910518, |
|
"grad_norm": 0.79474937915802, |
|
"learning_rate": 1.817934805211976e-05, |
|
"loss": 0.4153, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 8.100470957613815, |
|
"grad_norm": 0.8072588443756104, |
|
"learning_rate": 1.7059431454979824e-05, |
|
"loss": 0.4072, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 8.16326530612245, |
|
"grad_norm": 0.8833736777305603, |
|
"learning_rate": 1.5971893480895583e-05, |
|
"loss": 0.3996, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 8.226059654631083, |
|
"grad_norm": 0.8510453104972839, |
|
"learning_rate": 1.491715868666399e-05, |
|
"loss": 0.4132, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 8.288854003139717, |
|
"grad_norm": 0.8392331004142761, |
|
"learning_rate": 1.3895638823264446e-05, |
|
"loss": 0.4181, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 8.351648351648352, |
|
"grad_norm": 0.7935534119606018, |
|
"learning_rate": 1.290773267511788e-05, |
|
"loss": 0.3973, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 8.414442700156986, |
|
"grad_norm": 0.821882963180542, |
|
"learning_rate": 1.1953825904408034e-05, |
|
"loss": 0.4069, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 8.47723704866562, |
|
"grad_norm": 0.8700546026229858, |
|
"learning_rate": 1.103429090052528e-05, |
|
"loss": 0.4144, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 8.540031397174253, |
|
"grad_norm": 0.822404146194458, |
|
"learning_rate": 1.0149486634692018e-05, |
|
"loss": 0.4052, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 8.602825745682889, |
|
"grad_norm": 0.8342509269714355, |
|
"learning_rate": 9.299758519826273e-06, |
|
"loss": 0.4076, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 8.665620094191523, |
|
"grad_norm": 0.8416010141372681, |
|
"learning_rate": 8.485438275698154e-06, |
|
"loss": 0.4161, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 8.728414442700156, |
|
"grad_norm": 0.7646155953407288, |
|
"learning_rate": 7.706843799431984e-06, |
|
"loss": 0.4153, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 8.791208791208792, |
|
"grad_norm": 0.7787067294120789, |
|
"learning_rate": 6.964279041404553e-06, |
|
"loss": 0.3995, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 8.854003139717426, |
|
"grad_norm": 0.9011712074279785, |
|
"learning_rate": 6.258033886587911e-06, |
|
"loss": 0.4144, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 8.91679748822606, |
|
"grad_norm": 0.8679221272468567, |
|
"learning_rate": 5.588384041383088e-06, |
|
"loss": 0.4147, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 8.979591836734693, |
|
"grad_norm": 0.8739622831344604, |
|
"learning_rate": 4.955590925988895e-06, |
|
"loss": 0.4011, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 9.042386185243329, |
|
"grad_norm": 0.824476420879364, |
|
"learning_rate": 4.359901572347758e-06, |
|
"loss": 0.4138, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 9.105180533751962, |
|
"grad_norm": 0.8952732682228088, |
|
"learning_rate": 3.8015485277086205e-06, |
|
"loss": 0.4088, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 9.167974882260596, |
|
"grad_norm": 0.7407020330429077, |
|
"learning_rate": 3.280749763844293e-06, |
|
"loss": 0.4018, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 9.23076923076923, |
|
"grad_norm": 0.8719793558120728, |
|
"learning_rate": 2.7977085919589254e-06, |
|
"loss": 0.3925, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 9.293563579277865, |
|
"grad_norm": 0.8144113421440125, |
|
"learning_rate": 2.3526135833186525e-06, |
|
"loss": 0.4012, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 9.3563579277865, |
|
"grad_norm": 0.8647878170013428, |
|
"learning_rate": 1.945638495636515e-06, |
|
"loss": 0.4006, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 9.419152276295133, |
|
"grad_norm": 0.827995240688324, |
|
"learning_rate": 1.576942205240317e-06, |
|
"loss": 0.3984, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 9.481946624803768, |
|
"grad_norm": 0.8608419895172119, |
|
"learning_rate": 1.2466686450499865e-06, |
|
"loss": 0.3956, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 9.544740973312402, |
|
"grad_norm": 0.8144489526748657, |
|
"learning_rate": 9.549467483884412e-07, |
|
"loss": 0.3967, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 9.607535321821036, |
|
"grad_norm": 0.906897783279419, |
|
"learning_rate": 7.018903986483083e-07, |
|
"loss": 0.4039, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 9.67032967032967, |
|
"grad_norm": 0.8140434622764587, |
|
"learning_rate": 4.875983848335874e-07, |
|
"loss": 0.3924, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 9.733124018838305, |
|
"grad_norm": 0.8026575446128845, |
|
"learning_rate": 3.12154362994177e-07, |
|
"loss": 0.4028, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 9.795918367346939, |
|
"grad_norm": 0.8668120503425598, |
|
"learning_rate": 1.7562682356786487e-07, |
|
"loss": 0.404, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 9.858712715855573, |
|
"grad_norm": 0.8131958842277527, |
|
"learning_rate": 7.806906464281615e-08, |
|
"loss": 0.4016, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 9.921507064364206, |
|
"grad_norm": 0.820686399936676, |
|
"learning_rate": 1.951917115091684e-08, |
|
"loss": 0.3993, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 9.984301412872842, |
|
"grad_norm": 0.8288590312004089, |
|
"learning_rate": 0.0, |
|
"loss": 0.389, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 9.984301412872842, |
|
"step": 1590, |
|
"total_flos": 1.04418647212032e+17, |
|
"train_loss": 0.5835858842861728, |
|
"train_runtime": 7264.6263, |
|
"train_samples_per_second": 0.877, |
|
"train_steps_per_second": 0.219 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1590, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"total_flos": 1.04418647212032e+17, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|