{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9996678844237795, "eval_steps": 500, "global_step": 1505, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0006642311524410495, "grad_norm": 1.015625, "learning_rate": 1.3245033112582782e-06, "loss": 1.8253, "step": 1 }, { "epoch": 0.0033211557622052474, "grad_norm": 0.7421875, "learning_rate": 6.622516556291391e-06, "loss": 1.838, "step": 5 }, { "epoch": 0.006642311524410495, "grad_norm": 0.87109375, "learning_rate": 1.3245033112582782e-05, "loss": 1.823, "step": 10 }, { "epoch": 0.009963467286615742, "grad_norm": 0.58984375, "learning_rate": 1.9867549668874173e-05, "loss": 1.7917, "step": 15 }, { "epoch": 0.01328462304882099, "grad_norm": 0.353515625, "learning_rate": 2.6490066225165565e-05, "loss": 1.7849, "step": 20 }, { "epoch": 0.016605778811026237, "grad_norm": 0.375, "learning_rate": 3.311258278145696e-05, "loss": 1.7819, "step": 25 }, { "epoch": 0.019926934573231483, "grad_norm": 0.4609375, "learning_rate": 3.9735099337748346e-05, "loss": 1.7091, "step": 30 }, { "epoch": 0.023248090335436733, "grad_norm": 0.490234375, "learning_rate": 4.635761589403974e-05, "loss": 1.6498, "step": 35 }, { "epoch": 0.02656924609764198, "grad_norm": 2.0, "learning_rate": 5.298013245033113e-05, "loss": 1.6126, "step": 40 }, { "epoch": 0.029890401859847225, "grad_norm": 0.5078125, "learning_rate": 5.960264900662252e-05, "loss": 1.5847, "step": 45 }, { "epoch": 0.033211557622052475, "grad_norm": 0.515625, "learning_rate": 6.622516556291392e-05, "loss": 1.4876, "step": 50 }, { "epoch": 0.036532713384257724, "grad_norm": 0.33203125, "learning_rate": 7.284768211920529e-05, "loss": 1.4771, "step": 55 }, { "epoch": 0.03985386914646297, "grad_norm": 0.234375, "learning_rate": 7.947019867549669e-05, "loss": 1.4546, "step": 60 }, { "epoch": 0.043175024908668216, "grad_norm": 0.6015625, "learning_rate": 8.609271523178808e-05, "loss": 1.4066, "step": 65 }, { "epoch": 0.046496180670873466, "grad_norm": 0.1611328125, "learning_rate": 9.271523178807948e-05, "loss": 1.4112, "step": 70 }, { "epoch": 0.04981733643307871, "grad_norm": 0.189453125, "learning_rate": 9.933774834437086e-05, "loss": 1.3762, "step": 75 }, { "epoch": 0.05313849219528396, "grad_norm": 0.150390625, "learning_rate": 0.00010596026490066226, "loss": 1.3657, "step": 80 }, { "epoch": 0.05645964795748921, "grad_norm": 0.1484375, "learning_rate": 0.00011258278145695364, "loss": 1.3309, "step": 85 }, { "epoch": 0.05978080371969445, "grad_norm": 0.11865234375, "learning_rate": 0.00011920529801324504, "loss": 1.3321, "step": 90 }, { "epoch": 0.0631019594818997, "grad_norm": 0.1083984375, "learning_rate": 0.00012582781456953643, "loss": 1.3179, "step": 95 }, { "epoch": 0.06642311524410495, "grad_norm": 0.095703125, "learning_rate": 0.00013245033112582784, "loss": 1.3108, "step": 100 }, { "epoch": 0.0697442710063102, "grad_norm": 0.09033203125, "learning_rate": 0.0001390728476821192, "loss": 1.2985, "step": 105 }, { "epoch": 0.07306542676851545, "grad_norm": 0.09814453125, "learning_rate": 0.00014569536423841059, "loss": 1.2727, "step": 110 }, { "epoch": 0.07638658253072068, "grad_norm": 0.11328125, "learning_rate": 0.000152317880794702, "loss": 1.2735, "step": 115 }, { "epoch": 0.07970773829292593, "grad_norm": 0.06787109375, "learning_rate": 0.00015894039735099338, "loss": 1.2625, "step": 120 }, { "epoch": 0.08302889405513118, "grad_norm": 0.08056640625, "learning_rate": 0.00016556291390728477, "loss": 1.2354, "step": 125 }, { "epoch": 0.08635004981733643, "grad_norm": 0.08837890625, "learning_rate": 0.00017218543046357615, "loss": 1.2534, "step": 130 }, { "epoch": 0.08967120557954168, "grad_norm": 0.0888671875, "learning_rate": 0.00017880794701986757, "loss": 1.2281, "step": 135 }, { "epoch": 0.09299236134174693, "grad_norm": 0.146484375, "learning_rate": 0.00018543046357615895, "loss": 1.2436, "step": 140 }, { "epoch": 0.09631351710395218, "grad_norm": 0.0986328125, "learning_rate": 0.00019205298013245034, "loss": 1.2399, "step": 145 }, { "epoch": 0.09963467286615742, "grad_norm": 0.1142578125, "learning_rate": 0.00019867549668874172, "loss": 1.2157, "step": 150 }, { "epoch": 0.10295582862836267, "grad_norm": 0.107421875, "learning_rate": 0.00019999569325372926, "loss": 1.2191, "step": 155 }, { "epoch": 0.10627698439056792, "grad_norm": 0.08154296875, "learning_rate": 0.0001999781977327807, "loss": 1.2421, "step": 160 }, { "epoch": 0.10959814015277317, "grad_norm": 0.0927734375, "learning_rate": 0.00019994724661834065, "loss": 1.2209, "step": 165 }, { "epoch": 0.11291929591497842, "grad_norm": 0.11572265625, "learning_rate": 0.0001999028440759734, "loss": 1.1945, "step": 170 }, { "epoch": 0.11624045167718366, "grad_norm": 0.11083984375, "learning_rate": 0.00019984499608160746, "loss": 1.2276, "step": 175 }, { "epoch": 0.1195616074393889, "grad_norm": 0.07177734375, "learning_rate": 0.00019977371042073115, "loss": 1.2141, "step": 180 }, { "epoch": 0.12288276320159415, "grad_norm": 0.126953125, "learning_rate": 0.00019968899668734502, "loss": 1.2088, "step": 185 }, { "epoch": 0.1262039189637994, "grad_norm": 0.083984375, "learning_rate": 0.00019959086628267033, "loss": 1.2116, "step": 190 }, { "epoch": 0.12952507472600466, "grad_norm": 0.09033203125, "learning_rate": 0.00019947933241361483, "loss": 1.2059, "step": 195 }, { "epoch": 0.1328462304882099, "grad_norm": 0.0859375, "learning_rate": 0.00019935441009099524, "loss": 1.2049, "step": 200 }, { "epoch": 0.13616738625041513, "grad_norm": 0.130859375, "learning_rate": 0.00019921611612751708, "loss": 1.1862, "step": 205 }, { "epoch": 0.1394885420126204, "grad_norm": 0.091796875, "learning_rate": 0.00019906446913551173, "loss": 1.1948, "step": 210 }, { "epoch": 0.14280969777482563, "grad_norm": 0.0830078125, "learning_rate": 0.00019889948952443173, "loss": 1.1951, "step": 215 }, { "epoch": 0.1461308535370309, "grad_norm": 0.07763671875, "learning_rate": 0.0001987211994981039, "loss": 1.1761, "step": 220 }, { "epoch": 0.14945200929923613, "grad_norm": 0.08056640625, "learning_rate": 0.00019852962305174078, "loss": 1.2026, "step": 225 }, { "epoch": 0.15277316506144137, "grad_norm": 0.2421875, "learning_rate": 0.00019832478596871166, "loss": 1.1896, "step": 230 }, { "epoch": 0.15609432082364663, "grad_norm": 0.0888671875, "learning_rate": 0.00019810671581707223, "loss": 1.1745, "step": 235 }, { "epoch": 0.15941547658585187, "grad_norm": 0.09228515625, "learning_rate": 0.00019787544194585434, "loss": 1.1878, "step": 240 }, { "epoch": 0.16273663234805713, "grad_norm": 0.08203125, "learning_rate": 0.00019763099548111615, "loss": 1.1851, "step": 245 }, { "epoch": 0.16605778811026237, "grad_norm": 0.10009765625, "learning_rate": 0.00019737340932175296, "loss": 1.1765, "step": 250 }, { "epoch": 0.16937894387246763, "grad_norm": 0.1044921875, "learning_rate": 0.00019710271813506953, "loss": 1.1818, "step": 255 }, { "epoch": 0.17270009963467287, "grad_norm": 0.09521484375, "learning_rate": 0.00019681895835211438, "loss": 1.1489, "step": 260 }, { "epoch": 0.1760212553968781, "grad_norm": 0.134765625, "learning_rate": 0.00019652216816277656, "loss": 1.1565, "step": 265 }, { "epoch": 0.17934241115908336, "grad_norm": 0.09912109375, "learning_rate": 0.0001962123875106462, "loss": 1.1663, "step": 270 }, { "epoch": 0.1826635669212886, "grad_norm": 0.1005859375, "learning_rate": 0.0001958896580876383, "loss": 1.187, "step": 275 }, { "epoch": 0.18598472268349386, "grad_norm": 0.0791015625, "learning_rate": 0.00019555402332838178, "loss": 1.1769, "step": 280 }, { "epoch": 0.1893058784456991, "grad_norm": 0.0791015625, "learning_rate": 0.00019520552840437394, "loss": 1.1765, "step": 285 }, { "epoch": 0.19262703420790436, "grad_norm": 0.1005859375, "learning_rate": 0.00019484422021790083, "loss": 1.1769, "step": 290 }, { "epoch": 0.1959481899701096, "grad_norm": 0.1123046875, "learning_rate": 0.00019447014739572502, "loss": 1.1654, "step": 295 }, { "epoch": 0.19926934573231483, "grad_norm": 0.1396484375, "learning_rate": 0.0001940833602825411, "loss": 1.1695, "step": 300 }, { "epoch": 0.2025905014945201, "grad_norm": 0.11376953125, "learning_rate": 0.00019368391093420003, "loss": 1.1685, "step": 305 }, { "epoch": 0.20591165725672533, "grad_norm": 0.09423828125, "learning_rate": 0.0001932718531107033, "loss": 1.1502, "step": 310 }, { "epoch": 0.2092328130189306, "grad_norm": 0.09716796875, "learning_rate": 0.00019284724226896737, "loss": 1.1694, "step": 315 }, { "epoch": 0.21255396878113583, "grad_norm": 0.12158203125, "learning_rate": 0.0001924101355553603, "loss": 1.1701, "step": 320 }, { "epoch": 0.2158751245433411, "grad_norm": 0.0908203125, "learning_rate": 0.0001919605917980104, "loss": 1.1501, "step": 325 }, { "epoch": 0.21919628030554633, "grad_norm": 0.1572265625, "learning_rate": 0.00019149867149888906, "loss": 1.1645, "step": 330 }, { "epoch": 0.22251743606775157, "grad_norm": 0.0908203125, "learning_rate": 0.00019102443682566793, "loss": 1.1432, "step": 335 }, { "epoch": 0.22583859182995683, "grad_norm": 0.126953125, "learning_rate": 0.00019053795160335214, "loss": 1.1632, "step": 340 }, { "epoch": 0.22915974759216207, "grad_norm": 0.09423828125, "learning_rate": 0.00019003928130569039, "loss": 1.1464, "step": 345 }, { "epoch": 0.23248090335436733, "grad_norm": 0.1064453125, "learning_rate": 0.00018952849304636307, "loss": 1.1568, "step": 350 }, { "epoch": 0.23580205911657257, "grad_norm": 0.107421875, "learning_rate": 0.00018900565556994984, "loss": 1.156, "step": 355 }, { "epoch": 0.2391232148787778, "grad_norm": 0.1123046875, "learning_rate": 0.0001884708392426776, "loss": 1.1471, "step": 360 }, { "epoch": 0.24244437064098306, "grad_norm": 0.09521484375, "learning_rate": 0.00018792411604295014, "loss": 1.1302, "step": 365 }, { "epoch": 0.2457655264031883, "grad_norm": 0.10107421875, "learning_rate": 0.0001873655595516611, "loss": 1.1276, "step": 370 }, { "epoch": 0.24908668216539356, "grad_norm": 0.279296875, "learning_rate": 0.0001867952449422909, "loss": 1.165, "step": 375 }, { "epoch": 0.2524078379275988, "grad_norm": 0.19921875, "learning_rate": 0.0001862132489707895, "loss": 1.1524, "step": 380 }, { "epoch": 0.25572899368980406, "grad_norm": 0.435546875, "learning_rate": 0.00018561964996524627, "loss": 1.1459, "step": 385 }, { "epoch": 0.2590501494520093, "grad_norm": 0.08935546875, "learning_rate": 0.00018501452781534812, "loss": 1.1375, "step": 390 }, { "epoch": 0.26237130521421453, "grad_norm": 0.1083984375, "learning_rate": 0.00018439796396162755, "loss": 1.1394, "step": 395 }, { "epoch": 0.2656924609764198, "grad_norm": 0.109375, "learning_rate": 0.00018377004138450196, "loss": 1.146, "step": 400 }, { "epoch": 0.26901361673862506, "grad_norm": 0.0947265625, "learning_rate": 0.00018313084459310568, "loss": 1.1391, "step": 405 }, { "epoch": 0.27233477250083027, "grad_norm": 0.421875, "learning_rate": 0.0001824804596139163, "loss": 1.1368, "step": 410 }, { "epoch": 0.27565592826303553, "grad_norm": 0.095703125, "learning_rate": 0.00018181897397917673, "loss": 1.1573, "step": 415 }, { "epoch": 0.2789770840252408, "grad_norm": 0.09521484375, "learning_rate": 0.00018114647671511473, "loss": 1.1502, "step": 420 }, { "epoch": 0.282298239787446, "grad_norm": 0.09765625, "learning_rate": 0.00018046305832996126, "loss": 1.1461, "step": 425 }, { "epoch": 0.28561939554965127, "grad_norm": 0.1044921875, "learning_rate": 0.00017976881080176938, "loss": 1.163, "step": 430 }, { "epoch": 0.28894055131185653, "grad_norm": 0.10546875, "learning_rate": 0.00017906382756603534, "loss": 1.155, "step": 435 }, { "epoch": 0.2922617070740618, "grad_norm": 0.1123046875, "learning_rate": 0.0001783482035031236, "loss": 1.1576, "step": 440 }, { "epoch": 0.295582862836267, "grad_norm": 0.1142578125, "learning_rate": 0.00017762203492549728, "loss": 1.1548, "step": 445 }, { "epoch": 0.29890401859847227, "grad_norm": 0.1259765625, "learning_rate": 0.000176885419564756, "loss": 1.1501, "step": 450 }, { "epoch": 0.30222517436067753, "grad_norm": 0.0869140625, "learning_rate": 0.00017613845655848248, "loss": 1.1265, "step": 455 }, { "epoch": 0.30554633012288274, "grad_norm": 0.0966796875, "learning_rate": 0.00017538124643690033, "loss": 1.1424, "step": 460 }, { "epoch": 0.308867485885088, "grad_norm": 0.12158203125, "learning_rate": 0.00017461389110934382, "loss": 1.1632, "step": 465 }, { "epoch": 0.31218864164729326, "grad_norm": 0.1279296875, "learning_rate": 0.00017383649385054276, "loss": 1.1293, "step": 470 }, { "epoch": 0.3155097974094985, "grad_norm": 0.103515625, "learning_rate": 0.000173049159286723, "loss": 1.1472, "step": 475 }, { "epoch": 0.31883095317170373, "grad_norm": 0.1025390625, "learning_rate": 0.0001722519933815253, "loss": 1.1435, "step": 480 }, { "epoch": 0.322152108933909, "grad_norm": 0.0927734375, "learning_rate": 0.0001714451034217443, "loss": 1.1612, "step": 485 }, { "epoch": 0.32547326469611426, "grad_norm": 0.0859375, "learning_rate": 0.0001706285980028892, "loss": 1.1242, "step": 490 }, { "epoch": 0.32879442045831947, "grad_norm": 0.09228515625, "learning_rate": 0.00016980258701456845, "loss": 1.1492, "step": 495 }, { "epoch": 0.33211557622052473, "grad_norm": 0.08642578125, "learning_rate": 0.00016896718162570009, "loss": 1.1566, "step": 500 }, { "epoch": 0.33543673198273, "grad_norm": 0.09912109375, "learning_rate": 0.00016812249426955032, "loss": 1.1261, "step": 505 }, { "epoch": 0.33875788774493526, "grad_norm": 0.125, "learning_rate": 0.00016726863862860146, "loss": 1.1389, "step": 510 }, { "epoch": 0.34207904350714047, "grad_norm": 0.09130859375, "learning_rate": 0.00016640572961925182, "loss": 1.1283, "step": 515 }, { "epoch": 0.34540019926934573, "grad_norm": 0.091796875, "learning_rate": 0.00016553388337635, "loss": 1.1372, "step": 520 }, { "epoch": 0.348721355031551, "grad_norm": 0.09130859375, "learning_rate": 0.00016465321723756464, "loss": 1.1392, "step": 525 }, { "epoch": 0.3520425107937562, "grad_norm": 0.09619140625, "learning_rate": 0.00016376384972759236, "loss": 1.1359, "step": 530 }, { "epoch": 0.35536366655596147, "grad_norm": 0.08837890625, "learning_rate": 0.00016286590054220642, "loss": 1.129, "step": 535 }, { "epoch": 0.35868482231816673, "grad_norm": 0.10400390625, "learning_rate": 0.00016195949053214709, "loss": 1.1348, "step": 540 }, { "epoch": 0.362005978080372, "grad_norm": 0.1005859375, "learning_rate": 0.00016104474168685725, "loss": 1.1489, "step": 545 }, { "epoch": 0.3653271338425772, "grad_norm": 0.1005859375, "learning_rate": 0.000160121777118064, "loss": 1.1367, "step": 550 }, { "epoch": 0.36864828960478246, "grad_norm": 0.09375, "learning_rate": 0.0001591907210432102, "loss": 1.1236, "step": 555 }, { "epoch": 0.3719694453669877, "grad_norm": 0.0888671875, "learning_rate": 0.00015825169876873592, "loss": 1.1313, "step": 560 }, { "epoch": 0.37529060112919294, "grad_norm": 0.1328125, "learning_rate": 0.0001573048366732147, "loss": 1.1452, "step": 565 }, { "epoch": 0.3786117568913982, "grad_norm": 0.1572265625, "learning_rate": 0.00015635026219034445, "loss": 1.1364, "step": 570 }, { "epoch": 0.38193291265360346, "grad_norm": 0.09326171875, "learning_rate": 0.00015538810379179694, "loss": 1.1348, "step": 575 }, { "epoch": 0.3852540684158087, "grad_norm": 0.10595703125, "learning_rate": 0.0001544184909699272, "loss": 1.1367, "step": 580 }, { "epoch": 0.38857522417801393, "grad_norm": 0.12451171875, "learning_rate": 0.00015344155422034609, "loss": 1.1194, "step": 585 }, { "epoch": 0.3918963799402192, "grad_norm": 0.11376953125, "learning_rate": 0.00015245742502435712, "loss": 1.1346, "step": 590 }, { "epoch": 0.39521753570242446, "grad_norm": 0.0986328125, "learning_rate": 0.00015146623583126134, "loss": 1.1223, "step": 595 }, { "epoch": 0.39853869146462967, "grad_norm": 0.0869140625, "learning_rate": 0.00015046812004053134, "loss": 1.1088, "step": 600 }, { "epoch": 0.40185984722683493, "grad_norm": 0.08935546875, "learning_rate": 0.00014946321198385795, "loss": 1.1381, "step": 605 }, { "epoch": 0.4051810029890402, "grad_norm": 0.09375, "learning_rate": 0.00014845164690707087, "loss": 1.1354, "step": 610 }, { "epoch": 0.40850215875124546, "grad_norm": 0.10400390625, "learning_rate": 0.00014743356095193666, "loss": 1.1204, "step": 615 }, { "epoch": 0.41182331451345067, "grad_norm": 0.1533203125, "learning_rate": 0.00014640909113783622, "loss": 1.1322, "step": 620 }, { "epoch": 0.41514447027565593, "grad_norm": 0.10693359375, "learning_rate": 0.00014537837534332385, "loss": 1.1271, "step": 625 }, { "epoch": 0.4184656260378612, "grad_norm": 0.09228515625, "learning_rate": 0.0001443415522875708, "loss": 1.1458, "step": 630 }, { "epoch": 0.4217867818000664, "grad_norm": 0.091796875, "learning_rate": 0.0001432987615116958, "loss": 1.133, "step": 635 }, { "epoch": 0.42510793756227166, "grad_norm": 0.08349609375, "learning_rate": 0.00014225014335998492, "loss": 1.1346, "step": 640 }, { "epoch": 0.42842909332447693, "grad_norm": 0.0869140625, "learning_rate": 0.0001411958389610031, "loss": 1.1061, "step": 645 }, { "epoch": 0.4317502490866822, "grad_norm": 0.09912109375, "learning_rate": 0.00014013599020860047, "loss": 1.1341, "step": 650 }, { "epoch": 0.4350714048488874, "grad_norm": 0.08837890625, "learning_rate": 0.00013907073974281562, "loss": 1.1186, "step": 655 }, { "epoch": 0.43839256061109266, "grad_norm": 0.08642578125, "learning_rate": 0.00013800023093067814, "loss": 1.1225, "step": 660 }, { "epoch": 0.4417137163732979, "grad_norm": 0.10400390625, "learning_rate": 0.00013692460784691356, "loss": 1.125, "step": 665 }, { "epoch": 0.44503487213550313, "grad_norm": 0.09375, "learning_rate": 0.0001358440152545533, "loss": 1.1509, "step": 670 }, { "epoch": 0.4483560278977084, "grad_norm": 0.08837890625, "learning_rate": 0.0001347585985854512, "loss": 1.1305, "step": 675 }, { "epoch": 0.45167718365991366, "grad_norm": 0.08642578125, "learning_rate": 0.000133668503920711, "loss": 1.1143, "step": 680 }, { "epoch": 0.4549983394221189, "grad_norm": 0.0927734375, "learning_rate": 0.0001325738779710257, "loss": 1.1207, "step": 685 }, { "epoch": 0.45831949518432413, "grad_norm": 0.091796875, "learning_rate": 0.00013147486805693256, "loss": 1.1355, "step": 690 }, { "epoch": 0.4616406509465294, "grad_norm": 0.10205078125, "learning_rate": 0.0001303716220889859, "loss": 1.1294, "step": 695 }, { "epoch": 0.46496180670873466, "grad_norm": 0.09326171875, "learning_rate": 0.00012926428854785052, "loss": 1.1368, "step": 700 }, { "epoch": 0.46828296247093987, "grad_norm": 0.08984375, "learning_rate": 0.00012815301646431845, "loss": 1.1347, "step": 705 }, { "epoch": 0.47160411823314513, "grad_norm": 0.10302734375, "learning_rate": 0.00012703795539925142, "loss": 1.129, "step": 710 }, { "epoch": 0.4749252739953504, "grad_norm": 0.10791015625, "learning_rate": 0.00012591925542345243, "loss": 1.1162, "step": 715 }, { "epoch": 0.4782464297575556, "grad_norm": 0.1025390625, "learning_rate": 0.0001247970670974682, "loss": 1.147, "step": 720 }, { "epoch": 0.48156758551976087, "grad_norm": 0.10546875, "learning_rate": 0.00012367154145132608, "loss": 1.1165, "step": 725 }, { "epoch": 0.48488874128196613, "grad_norm": 0.1103515625, "learning_rate": 0.00012254282996420753, "loss": 1.1372, "step": 730 }, { "epoch": 0.4882098970441714, "grad_norm": 0.107421875, "learning_rate": 0.0001214110845440613, "loss": 1.1144, "step": 735 }, { "epoch": 0.4915310528063766, "grad_norm": 0.1044921875, "learning_rate": 0.00012027645750715892, "loss": 1.1397, "step": 740 }, { "epoch": 0.49485220856858186, "grad_norm": 0.1025390625, "learning_rate": 0.0001191391015575951, "loss": 1.1364, "step": 745 }, { "epoch": 0.4981733643307871, "grad_norm": 0.087890625, "learning_rate": 0.00011799916976673589, "loss": 1.1234, "step": 750 }, { "epoch": 0.5014945200929923, "grad_norm": 0.111328125, "learning_rate": 0.00011685681555261786, "loss": 1.1223, "step": 755 }, { "epoch": 0.5048156758551976, "grad_norm": 0.0966796875, "learning_rate": 0.00011571219265929998, "loss": 1.1357, "step": 760 }, { "epoch": 0.5081368316174029, "grad_norm": 0.0947265625, "learning_rate": 0.00011456545513617198, "loss": 1.133, "step": 765 }, { "epoch": 0.5114579873796081, "grad_norm": 0.12890625, "learning_rate": 0.00011341675731722175, "loss": 1.1346, "step": 770 }, { "epoch": 0.5147791431418134, "grad_norm": 0.1611328125, "learning_rate": 0.00011226625380026407, "loss": 1.1094, "step": 775 }, { "epoch": 0.5181002989040187, "grad_norm": 0.1376953125, "learning_rate": 0.0001111140994261341, "loss": 1.1246, "step": 780 }, { "epoch": 0.5214214546662238, "grad_norm": 0.09521484375, "learning_rate": 0.00010996044925784806, "loss": 1.0992, "step": 785 }, { "epoch": 0.5247426104284291, "grad_norm": 0.1025390625, "learning_rate": 0.00010880545855973405, "loss": 1.1271, "step": 790 }, { "epoch": 0.5280637661906343, "grad_norm": 0.087890625, "learning_rate": 0.00010764928277653576, "loss": 1.1272, "step": 795 }, { "epoch": 0.5313849219528396, "grad_norm": 0.0927734375, "learning_rate": 0.00010649207751249188, "loss": 1.1211, "step": 800 }, { "epoch": 0.5347060777150449, "grad_norm": 0.09716796875, "learning_rate": 0.0001053339985103941, "loss": 1.1108, "step": 805 }, { "epoch": 0.5380272334772501, "grad_norm": 0.09814453125, "learning_rate": 0.00010417520163062626, "loss": 1.1259, "step": 810 }, { "epoch": 0.5413483892394554, "grad_norm": 0.0849609375, "learning_rate": 0.00010301584283018813, "loss": 1.1131, "step": 815 }, { "epoch": 0.5446695450016605, "grad_norm": 0.09228515625, "learning_rate": 0.00010185607814170561, "loss": 1.1278, "step": 820 }, { "epoch": 0.5479907007638658, "grad_norm": 0.09765625, "learning_rate": 0.00010069606365243123, "loss": 1.1259, "step": 825 }, { "epoch": 0.5513118565260711, "grad_norm": 0.08349609375, "learning_rate": 9.953595548323699e-05, "loss": 1.1277, "step": 830 }, { "epoch": 0.5546330122882763, "grad_norm": 0.11083984375, "learning_rate": 9.837590976760282e-05, "loss": 1.125, "step": 835 }, { "epoch": 0.5579541680504816, "grad_norm": 0.10302734375, "learning_rate": 9.721608263060341e-05, "loss": 1.1217, "step": 840 }, { "epoch": 0.5612753238126869, "grad_norm": 0.109375, "learning_rate": 9.605663016789583e-05, "loss": 1.1234, "step": 845 }, { "epoch": 0.564596479574892, "grad_norm": 0.08544921875, "learning_rate": 9.489770842471159e-05, "loss": 1.1118, "step": 850 }, { "epoch": 0.5679176353370973, "grad_norm": 0.1005859375, "learning_rate": 9.373947337485522e-05, "loss": 1.1207, "step": 855 }, { "epoch": 0.5712387910993025, "grad_norm": 0.2119140625, "learning_rate": 9.258208089971232e-05, "loss": 1.1324, "step": 860 }, { "epoch": 0.5745599468615078, "grad_norm": 0.09423828125, "learning_rate": 9.142568676727043e-05, "loss": 1.1263, "step": 865 }, { "epoch": 0.5778811026237131, "grad_norm": 0.10498046875, "learning_rate": 9.027044661115485e-05, "loss": 1.1263, "step": 870 }, { "epoch": 0.5812022583859183, "grad_norm": 0.10498046875, "learning_rate": 8.911651590968259e-05, "loss": 1.1125, "step": 875 }, { "epoch": 0.5845234141481236, "grad_norm": 0.08740234375, "learning_rate": 8.79640499649374e-05, "loss": 1.1256, "step": 880 }, { "epoch": 0.5878445699103287, "grad_norm": 0.0888671875, "learning_rate": 8.68132038818684e-05, "loss": 1.1185, "step": 885 }, { "epoch": 0.591165725672534, "grad_norm": 0.08447265625, "learning_rate": 8.5664132547415e-05, "loss": 1.146, "step": 890 }, { "epoch": 0.5944868814347393, "grad_norm": 0.0791015625, "learning_rate": 8.451699060966173e-05, "loss": 1.105, "step": 895 }, { "epoch": 0.5978080371969445, "grad_norm": 0.1630859375, "learning_rate": 8.33719324570247e-05, "loss": 1.1334, "step": 900 }, { "epoch": 0.6011291929591498, "grad_norm": 0.0888671875, "learning_rate": 8.222911219747316e-05, "loss": 1.1373, "step": 905 }, { "epoch": 0.6044503487213551, "grad_norm": 0.1044921875, "learning_rate": 8.108868363778894e-05, "loss": 1.1089, "step": 910 }, { "epoch": 0.6077715044835603, "grad_norm": 0.1240234375, "learning_rate": 7.995080026286631e-05, "loss": 1.1296, "step": 915 }, { "epoch": 0.6110926602457655, "grad_norm": 0.0986328125, "learning_rate": 7.881561521505514e-05, "loss": 1.1318, "step": 920 }, { "epoch": 0.6144138160079707, "grad_norm": 0.09521484375, "learning_rate": 7.768328127355007e-05, "loss": 1.1231, "step": 925 }, { "epoch": 0.617734971770176, "grad_norm": 0.0859375, "learning_rate": 7.655395083382899e-05, "loss": 1.13, "step": 930 }, { "epoch": 0.6210561275323813, "grad_norm": 0.0966796875, "learning_rate": 7.542777588714256e-05, "loss": 1.1109, "step": 935 }, { "epoch": 0.6243772832945865, "grad_norm": 0.09423828125, "learning_rate": 7.430490800005854e-05, "loss": 1.1166, "step": 940 }, { "epoch": 0.6276984390567918, "grad_norm": 0.09619140625, "learning_rate": 7.318549829406318e-05, "loss": 1.1117, "step": 945 }, { "epoch": 0.631019594818997, "grad_norm": 0.08740234375, "learning_rate": 7.206969742522252e-05, "loss": 1.121, "step": 950 }, { "epoch": 0.6343407505812022, "grad_norm": 0.08740234375, "learning_rate": 7.095765556390606e-05, "loss": 1.1349, "step": 955 }, { "epoch": 0.6376619063434075, "grad_norm": 0.08349609375, "learning_rate": 6.984952237457647e-05, "loss": 1.0884, "step": 960 }, { "epoch": 0.6409830621056127, "grad_norm": 0.08642578125, "learning_rate": 6.874544699564663e-05, "loss": 1.1405, "step": 965 }, { "epoch": 0.644304217867818, "grad_norm": 0.203125, "learning_rate": 6.764557801940771e-05, "loss": 1.1103, "step": 970 }, { "epoch": 0.6476253736300233, "grad_norm": 0.0927734375, "learning_rate": 6.655006347203128e-05, "loss": 1.1222, "step": 975 }, { "epoch": 0.6509465293922285, "grad_norm": 0.0927734375, "learning_rate": 6.545905079364678e-05, "loss": 1.1323, "step": 980 }, { "epoch": 0.6542676851544338, "grad_norm": 0.08544921875, "learning_rate": 6.437268681849825e-05, "loss": 1.1185, "step": 985 }, { "epoch": 0.6575888409166389, "grad_norm": 0.09033203125, "learning_rate": 6.329111775518284e-05, "loss": 1.1153, "step": 990 }, { "epoch": 0.6609099966788442, "grad_norm": 0.09375, "learning_rate": 6.221448916697323e-05, "loss": 1.1356, "step": 995 }, { "epoch": 0.6642311524410495, "grad_norm": 0.09716796875, "learning_rate": 6.114294595222667e-05, "loss": 1.1316, "step": 1000 }, { "epoch": 0.6675523082032547, "grad_norm": 0.1484375, "learning_rate": 6.0076632324884184e-05, "loss": 1.1045, "step": 1005 }, { "epoch": 0.67087346396546, "grad_norm": 0.083984375, "learning_rate": 5.9015691795061276e-05, "loss": 1.1136, "step": 1010 }, { "epoch": 0.6741946197276653, "grad_norm": 0.08740234375, "learning_rate": 5.796026714973358e-05, "loss": 1.1233, "step": 1015 }, { "epoch": 0.6775157754898705, "grad_norm": 0.083984375, "learning_rate": 5.691050043352e-05, "loss": 1.1125, "step": 1020 }, { "epoch": 0.6808369312520757, "grad_norm": 0.08837890625, "learning_rate": 5.586653292956536e-05, "loss": 1.1257, "step": 1025 }, { "epoch": 0.6841580870142809, "grad_norm": 0.0947265625, "learning_rate": 5.482850514052593e-05, "loss": 1.112, "step": 1030 }, { "epoch": 0.6874792427764862, "grad_norm": 0.0966796875, "learning_rate": 5.379655676965985e-05, "loss": 1.1022, "step": 1035 }, { "epoch": 0.6908003985386915, "grad_norm": 0.0986328125, "learning_rate": 5.2770826702025024e-05, "loss": 1.1031, "step": 1040 }, { "epoch": 0.6941215543008967, "grad_norm": 0.09326171875, "learning_rate": 5.1751452985787294e-05, "loss": 1.1314, "step": 1045 }, { "epoch": 0.697442710063102, "grad_norm": 0.09326171875, "learning_rate": 5.073857281364116e-05, "loss": 1.115, "step": 1050 }, { "epoch": 0.7007638658253073, "grad_norm": 0.087890625, "learning_rate": 4.9732322504345784e-05, "loss": 1.1162, "step": 1055 }, { "epoch": 0.7040850215875124, "grad_norm": 0.08740234375, "learning_rate": 4.873283748437832e-05, "loss": 1.1398, "step": 1060 }, { "epoch": 0.7074061773497177, "grad_norm": 0.0849609375, "learning_rate": 4.7740252269707545e-05, "loss": 1.1165, "step": 1065 }, { "epoch": 0.7107273331119229, "grad_norm": 0.09521484375, "learning_rate": 4.6754700447690145e-05, "loss": 1.1134, "step": 1070 }, { "epoch": 0.7140484888741282, "grad_norm": 0.087890625, "learning_rate": 4.57763146590916e-05, "loss": 1.1101, "step": 1075 }, { "epoch": 0.7173696446363335, "grad_norm": 0.08056640625, "learning_rate": 4.4805226580234794e-05, "loss": 1.1058, "step": 1080 }, { "epoch": 0.7206908003985387, "grad_norm": 0.10009765625, "learning_rate": 4.384156690527842e-05, "loss": 1.1228, "step": 1085 }, { "epoch": 0.724011956160744, "grad_norm": 0.08740234375, "learning_rate": 4.2885465328627275e-05, "loss": 1.1173, "step": 1090 }, { "epoch": 0.7273331119229491, "grad_norm": 0.10498046875, "learning_rate": 4.193705052747737e-05, "loss": 1.1233, "step": 1095 }, { "epoch": 0.7306542676851544, "grad_norm": 0.08251953125, "learning_rate": 4.0996450144498044e-05, "loss": 1.1163, "step": 1100 }, { "epoch": 0.7339754234473597, "grad_norm": 0.08349609375, "learning_rate": 4.0063790770652876e-05, "loss": 1.1264, "step": 1105 }, { "epoch": 0.7372965792095649, "grad_norm": 0.078125, "learning_rate": 3.913919792816252e-05, "loss": 1.1334, "step": 1110 }, { "epoch": 0.7406177349717702, "grad_norm": 0.076171875, "learning_rate": 3.822279605361138e-05, "loss": 1.1133, "step": 1115 }, { "epoch": 0.7439388907339755, "grad_norm": 0.0810546875, "learning_rate": 3.731470848120005e-05, "loss": 1.1068, "step": 1120 }, { "epoch": 0.7472600464961807, "grad_norm": 0.0947265625, "learning_rate": 3.64150574261465e-05, "loss": 1.1204, "step": 1125 }, { "epoch": 0.7505812022583859, "grad_norm": 0.0869140625, "learning_rate": 3.552396396823774e-05, "loss": 1.1446, "step": 1130 }, { "epoch": 0.7539023580205911, "grad_norm": 0.0986328125, "learning_rate": 3.464154803553408e-05, "loss": 1.0895, "step": 1135 }, { "epoch": 0.7572235137827964, "grad_norm": 0.09326171875, "learning_rate": 3.376792838822873e-05, "loss": 1.1094, "step": 1140 }, { "epoch": 0.7605446695450017, "grad_norm": 0.08544921875, "learning_rate": 3.2903222602664464e-05, "loss": 1.1178, "step": 1145 }, { "epoch": 0.7638658253072069, "grad_norm": 0.10205078125, "learning_rate": 3.20475470555094e-05, "loss": 1.1149, "step": 1150 }, { "epoch": 0.7671869810694122, "grad_norm": 0.08349609375, "learning_rate": 3.1201016908094514e-05, "loss": 1.1282, "step": 1155 }, { "epoch": 0.7705081368316175, "grad_norm": 0.0986328125, "learning_rate": 3.0363746090914724e-05, "loss": 1.1392, "step": 1160 }, { "epoch": 0.7738292925938226, "grad_norm": 0.08251953125, "learning_rate": 2.9535847288295325e-05, "loss": 1.0927, "step": 1165 }, { "epoch": 0.7771504483560279, "grad_norm": 0.08642578125, "learning_rate": 2.8717431923226422e-05, "loss": 1.1235, "step": 1170 }, { "epoch": 0.7804716041182331, "grad_norm": 0.08740234375, "learning_rate": 2.7908610142367143e-05, "loss": 1.109, "step": 1175 }, { "epoch": 0.7837927598804384, "grad_norm": 0.10009765625, "learning_rate": 2.7109490801221327e-05, "loss": 1.1214, "step": 1180 }, { "epoch": 0.7871139156426437, "grad_norm": 0.08935546875, "learning_rate": 2.632018144948727e-05, "loss": 1.1099, "step": 1185 }, { "epoch": 0.7904350714048489, "grad_norm": 0.0830078125, "learning_rate": 2.5540788316583207e-05, "loss": 1.1275, "step": 1190 }, { "epoch": 0.7937562271670542, "grad_norm": 0.10205078125, "learning_rate": 2.4771416297350247e-05, "loss": 1.1306, "step": 1195 }, { "epoch": 0.7970773829292593, "grad_norm": 0.1630859375, "learning_rate": 2.4012168937935096e-05, "loss": 1.0969, "step": 1200 }, { "epoch": 0.8003985386914646, "grad_norm": 0.0810546875, "learning_rate": 2.326314842185443e-05, "loss": 1.1186, "step": 1205 }, { "epoch": 0.8037196944536699, "grad_norm": 0.0830078125, "learning_rate": 2.252445555624225e-05, "loss": 1.1222, "step": 1210 }, { "epoch": 0.8070408502158751, "grad_norm": 0.0966796875, "learning_rate": 2.1796189758282915e-05, "loss": 1.1112, "step": 1215 }, { "epoch": 0.8103620059780804, "grad_norm": 0.08251953125, "learning_rate": 2.107844904183096e-05, "loss": 1.0992, "step": 1220 }, { "epoch": 0.8136831617402857, "grad_norm": 0.08935546875, "learning_rate": 2.0371330004219967e-05, "loss": 1.1067, "step": 1225 }, { "epoch": 0.8170043175024909, "grad_norm": 0.0859375, "learning_rate": 1.9674927813261856e-05, "loss": 1.135, "step": 1230 }, { "epoch": 0.8203254732646961, "grad_norm": 0.0849609375, "learning_rate": 1.8989336194438757e-05, "loss": 1.1272, "step": 1235 }, { "epoch": 0.8236466290269013, "grad_norm": 0.09228515625, "learning_rate": 1.831464741828903e-05, "loss": 1.11, "step": 1240 }, { "epoch": 0.8269677847891066, "grad_norm": 0.08154296875, "learning_rate": 1.7650952287988864e-05, "loss": 1.123, "step": 1245 }, { "epoch": 0.8302889405513119, "grad_norm": 0.08203125, "learning_rate": 1.699834012713155e-05, "loss": 1.1179, "step": 1250 }, { "epoch": 0.8336100963135171, "grad_norm": 0.08447265625, "learning_rate": 1.6356898767705954e-05, "loss": 1.1216, "step": 1255 }, { "epoch": 0.8369312520757224, "grad_norm": 0.078125, "learning_rate": 1.5726714538275423e-05, "loss": 1.1349, "step": 1260 }, { "epoch": 0.8402524078379277, "grad_norm": 0.0791015625, "learning_rate": 1.5107872252359389e-05, "loss": 1.1137, "step": 1265 }, { "epoch": 0.8435735636001328, "grad_norm": 0.09033203125, "learning_rate": 1.4500455197018703e-05, "loss": 1.1134, "step": 1270 }, { "epoch": 0.8468947193623381, "grad_norm": 0.08056640625, "learning_rate": 1.3904545121646317e-05, "loss": 1.1245, "step": 1275 }, { "epoch": 0.8502158751245433, "grad_norm": 0.08056640625, "learning_rate": 1.3320222226965118e-05, "loss": 1.1231, "step": 1280 }, { "epoch": 0.8535370308867486, "grad_norm": 0.08349609375, "learning_rate": 1.2747565154234109e-05, "loss": 1.1352, "step": 1285 }, { "epoch": 0.8568581866489539, "grad_norm": 0.0859375, "learning_rate": 1.2186650974664337e-05, "loss": 1.1065, "step": 1290 }, { "epoch": 0.8601793424111591, "grad_norm": 0.0869140625, "learning_rate": 1.1637555179046345e-05, "loss": 1.1235, "step": 1295 }, { "epoch": 0.8635004981733644, "grad_norm": 0.08447265625, "learning_rate": 1.1100351667590204e-05, "loss": 1.1458, "step": 1300 }, { "epoch": 0.8668216539355695, "grad_norm": 0.080078125, "learning_rate": 1.057511273997962e-05, "loss": 1.1097, "step": 1305 }, { "epoch": 0.8701428096977748, "grad_norm": 0.08203125, "learning_rate": 1.0061909085641396e-05, "loss": 1.1038, "step": 1310 }, { "epoch": 0.8734639654599801, "grad_norm": 0.10107421875, "learning_rate": 9.560809774231871e-06, "loss": 1.1226, "step": 1315 }, { "epoch": 0.8767851212221853, "grad_norm": 0.08251953125, "learning_rate": 9.0718822463409e-06, "loss": 1.1157, "step": 1320 }, { "epoch": 0.8801062769843906, "grad_norm": 0.08447265625, "learning_rate": 8.595192304415534e-06, "loss": 1.0975, "step": 1325 }, { "epoch": 0.8834274327465959, "grad_norm": 0.0888671875, "learning_rate": 8.130804103903956e-06, "loss": 1.1056, "step": 1330 }, { "epoch": 0.8867485885088011, "grad_norm": 0.123046875, "learning_rate": 7.678780144620956e-06, "loss": 1.1328, "step": 1335 }, { "epoch": 0.8900697442710063, "grad_norm": 0.083984375, "learning_rate": 7.2391812623366025e-06, "loss": 1.1226, "step": 1340 }, { "epoch": 0.8933909000332115, "grad_norm": 0.10400390625, "learning_rate": 6.812066620588519e-06, "loss": 1.1112, "step": 1345 }, { "epoch": 0.8967120557954168, "grad_norm": 0.0888671875, "learning_rate": 6.397493702719226e-06, "loss": 1.121, "step": 1350 }, { "epoch": 0.9000332115576221, "grad_norm": 0.08544921875, "learning_rate": 5.995518304139991e-06, "loss": 1.1104, "step": 1355 }, { "epoch": 0.9033543673198273, "grad_norm": 0.08984375, "learning_rate": 5.606194524821429e-06, "loss": 1.1058, "step": 1360 }, { "epoch": 0.9066755230820326, "grad_norm": 0.08935546875, "learning_rate": 5.229574762012379e-06, "loss": 1.1069, "step": 1365 }, { "epoch": 0.9099966788442378, "grad_norm": 0.08984375, "learning_rate": 4.865709703188193e-06, "loss": 1.1414, "step": 1370 }, { "epoch": 0.913317834606443, "grad_norm": 0.08251953125, "learning_rate": 4.514648319228798e-06, "loss": 1.1161, "step": 1375 }, { "epoch": 0.9166389903686483, "grad_norm": 0.07763671875, "learning_rate": 4.1764378578279865e-06, "loss": 1.1083, "step": 1380 }, { "epoch": 0.9199601461308535, "grad_norm": 0.08447265625, "learning_rate": 3.8511238371345845e-06, "loss": 1.1135, "step": 1385 }, { "epoch": 0.9232813018930588, "grad_norm": 0.08544921875, "learning_rate": 3.5387500396263483e-06, "loss": 1.1137, "step": 1390 }, { "epoch": 0.9266024576552641, "grad_norm": 0.09228515625, "learning_rate": 3.2393585062175493e-06, "loss": 1.1097, "step": 1395 }, { "epoch": 0.9299236134174693, "grad_norm": 0.08349609375, "learning_rate": 2.9529895306007804e-06, "loss": 1.1307, "step": 1400 }, { "epoch": 0.9332447691796745, "grad_norm": 0.09619140625, "learning_rate": 2.6796816538241065e-06, "loss": 1.1214, "step": 1405 }, { "epoch": 0.9365659249418797, "grad_norm": 0.083984375, "learning_rate": 2.419471659104e-06, "loss": 1.1157, "step": 1410 }, { "epoch": 0.939887080704085, "grad_norm": 0.08544921875, "learning_rate": 2.1723945668748247e-06, "loss": 1.1047, "step": 1415 }, { "epoch": 0.9432082364662903, "grad_norm": 0.10205078125, "learning_rate": 1.9384836300755937e-06, "loss": 1.1367, "step": 1420 }, { "epoch": 0.9465293922284955, "grad_norm": 0.08837890625, "learning_rate": 1.717770329674684e-06, "loss": 1.0773, "step": 1425 }, { "epoch": 0.9498505479907008, "grad_norm": 0.0771484375, "learning_rate": 1.5102843704328684e-06, "loss": 1.1348, "step": 1430 }, { "epoch": 0.953171703752906, "grad_norm": 0.080078125, "learning_rate": 1.3160536769055708e-06, "loss": 1.1134, "step": 1435 }, { "epoch": 0.9564928595151112, "grad_norm": 0.08544921875, "learning_rate": 1.1351043896846047e-06, "loss": 1.1216, "step": 1440 }, { "epoch": 0.9598140152773165, "grad_norm": 0.09375, "learning_rate": 9.674608618799985e-07, "loss": 1.1085, "step": 1445 }, { "epoch": 0.9631351710395217, "grad_norm": 0.0830078125, "learning_rate": 8.131456558424622e-07, "loss": 1.1218, "step": 1450 }, { "epoch": 0.966456326801727, "grad_norm": 0.09326171875, "learning_rate": 6.721795401268493e-07, "loss": 1.1021, "step": 1455 }, { "epoch": 0.9697774825639323, "grad_norm": 0.1259765625, "learning_rate": 5.445814866969712e-07, "loss": 1.1071, "step": 1460 }, { "epoch": 0.9730986383261375, "grad_norm": 0.08544921875, "learning_rate": 4.3036866837224964e-07, "loss": 1.0951, "step": 1465 }, { "epoch": 0.9764197940883428, "grad_norm": 0.08837890625, "learning_rate": 3.295564565165554e-07, "loss": 1.1135, "step": 1470 }, { "epoch": 0.9797409498505479, "grad_norm": 0.0869140625, "learning_rate": 2.4215841896938487e-07, "loss": 1.1155, "step": 1475 }, { "epoch": 0.9830621056127532, "grad_norm": 0.1533203125, "learning_rate": 1.681863182198984e-07, "loss": 1.1137, "step": 1480 }, { "epoch": 0.9863832613749585, "grad_norm": 0.083984375, "learning_rate": 1.0765010982378699e-07, "loss": 1.1328, "step": 1485 }, { "epoch": 0.9897044171371637, "grad_norm": 0.0810546875, "learning_rate": 6.055794106347712e-08, "loss": 1.1114, "step": 1490 }, { "epoch": 0.993025572899369, "grad_norm": 0.0869140625, "learning_rate": 2.6916149851563543e-08, "loss": 1.1065, "step": 1495 }, { "epoch": 0.9963467286615743, "grad_norm": 0.08251953125, "learning_rate": 6.729263877847114e-09, "loss": 1.1432, "step": 1500 }, { "epoch": 0.9996678844237795, "grad_norm": 0.08203125, "learning_rate": 0.0, "loss": 1.1166, "step": 1505 }, { "epoch": 0.9996678844237795, "eval_loss": 1.3468068838119507, "eval_runtime": 2438.0312, "eval_samples_per_second": 5.613, "eval_steps_per_second": 5.613, "step": 1505 }, { "epoch": 0.9996678844237795, "step": 1505, "total_flos": 4.73786206333698e+18, "train_loss": 0.5388378948072262, "train_runtime": 65741.8056, "train_samples_per_second": 2.931, "train_steps_per_second": 0.023 } ], "logging_steps": 5, "max_steps": 1505, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 20, "total_flos": 4.73786206333698e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }