{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.035964035964036, "global_step": 2038, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "learning_rate": 0.0012782749814122844, "loss": 1.6279, "step": 5 }, { "epoch": 0.01, "learning_rate": 0.0018077538151554684, "loss": 1.2509, "step": 10 }, { "epoch": 0.01, "learning_rate": 0.0022140372138502386, "loss": 1.4568, "step": 15 }, { "epoch": 0.02, "learning_rate": 0.0025565499628245687, "loss": 1.4557, "step": 20 }, { "epoch": 0.02, "learning_rate": 0.0028583097523751473, "loss": 1.1046, "step": 25 }, { "epoch": 0.03, "learning_rate": 0.0031311214554257476, "loss": 1.3079, "step": 30 }, { "epoch": 0.03, "learning_rate": 0.003381997707972616, "loss": 1.1913, "step": 35 }, { "epoch": 0.04, "learning_rate": 0.0036155076303109367, "loss": 1.415, "step": 40 }, { "epoch": 0.04, "learning_rate": 0.003834824944236853, "loss": 1.2117, "step": 45 }, { "epoch": 0.05, "learning_rate": 0.0040422604172722166, "loss": 1.3286, "step": 50 }, { "epoch": 0.05, "learning_rate": 0.004239558492243069, "loss": 1.2645, "step": 55 }, { "epoch": 0.06, "learning_rate": 0.004428074427700477, "loss": 1.3262, "step": 60 }, { "epoch": 0.06, "learning_rate": 0.0046088859896247685, "loss": 1.3905, "step": 65 }, { "epoch": 0.07, "learning_rate": 0.004782867026529596, "loss": 1.2985, "step": 70 }, { "epoch": 0.07, "learning_rate": 0.004950737714883372, "loss": 1.1367, "step": 75 }, { "epoch": 0.08, "learning_rate": 0.0051130999256491375, "loss": 1.3175, "step": 80 }, { "epoch": 0.08, "learning_rate": 0.005270462766947299, "loss": 1.4444, "step": 85 }, { "epoch": 0.09, "learning_rate": 0.0054232614454664055, "loss": 1.2882, "step": 90 }, { "epoch": 0.09, "learning_rate": 0.005571871466032479, "loss": 1.4907, "step": 95 }, { "epoch": 0.1, "learning_rate": 0.0057166195047502946, "loss": 1.4369, "step": 100 }, { "epoch": 0.1, "learning_rate": 0.005857791861290061, "loss": 1.2819, "step": 105 }, { "epoch": 0.11, "learning_rate": 0.005995641118204179, "loss": 1.3525, "step": 110 }, { "epoch": 0.11, "learning_rate": 0.006130391451319013, "loss": 1.3266, "step": 115 }, { "epoch": 0.12, "learning_rate": 0.006262242910851495, "loss": 1.4452, "step": 120 }, { "epoch": 0.12, "learning_rate": 0.006391374907061422, "loss": 1.5723, "step": 125 }, { "epoch": 0.13, "learning_rate": 0.006517949073958691, "loss": 1.4178, "step": 130 }, { "epoch": 0.13, "learning_rate": 0.0066421116415507145, "loss": 1.4234, "step": 135 }, { "epoch": 0.14, "learning_rate": 0.006763995415945232, "loss": 1.5096, "step": 140 }, { "epoch": 0.14, "learning_rate": 0.006883721443741945, "loss": 1.3332, "step": 145 }, { "epoch": 0.15, "learning_rate": 0.007001400420140049, "loss": 1.5342, "step": 150 }, { "epoch": 0.15, "learning_rate": 0.007117133887404524, "loss": 1.5825, "step": 155 }, { "epoch": 0.16, "learning_rate": 0.007231015260621873, "loss": 1.3028, "step": 160 }, { "epoch": 0.16, "learning_rate": 0.007343130710225101, "loss": 1.5338, "step": 165 }, { "epoch": 0.17, "learning_rate": 0.0074535599249993005, "loss": 1.4564, "step": 170 }, { "epoch": 0.17, "learning_rate": 0.007562376774775252, "loss": 1.319, "step": 175 }, { "epoch": 0.18, "learning_rate": 0.007669649888473706, "loss": 1.6363, "step": 180 }, { "epoch": 0.18, "learning_rate": 0.007775443160352296, "loss": 1.4293, "step": 185 }, { "epoch": 0.19, "learning_rate": 0.007879816195062792, "loss": 1.4431, "step": 190 }, { "epoch": 0.19, "learning_rate": 0.007982824700322464, "loss": 1.3983, "step": 195 }, { "epoch": 0.2, "learning_rate": 0.008084520834544433, "loss": 1.448, "step": 200 }, { "epoch": 0.2, "learning_rate": 0.008184953515585021, "loss": 1.6799, "step": 205 }, { "epoch": 0.21, "learning_rate": 0.008284168695795142, "loss": 1.5823, "step": 210 }, { "epoch": 0.21, "learning_rate": 0.008382209607764337, "loss": 1.4564, "step": 215 }, { "epoch": 0.22, "learning_rate": 0.008479116984486139, "loss": 1.576, "step": 220 }, { "epoch": 0.22, "learning_rate": 0.008574929257125443, "loss": 1.5307, "step": 225 }, { "epoch": 0.23, "learning_rate": 0.00866968273311143, "loss": 1.4519, "step": 230 }, { "epoch": 0.23, "learning_rate": 0.008763411756896685, "loss": 1.6421, "step": 235 }, { "epoch": 0.24, "learning_rate": 0.008856148855400954, "loss": 1.321, "step": 240 }, { "epoch": 0.24, "learning_rate": 0.008947924869885989, "loss": 1.3645, "step": 245 }, { "epoch": 0.25, "learning_rate": 0.009038769075777341, "loss": 1.6528, "step": 250 }, { "epoch": 0.25, "learning_rate": 0.00912870929175277, "loss": 1.5315, "step": 255 }, { "epoch": 0.26, "learning_rate": 0.009217771979249537, "loss": 1.6313, "step": 260 }, { "epoch": 0.26, "learning_rate": 0.00930598233339944, "loss": 1.6657, "step": 265 }, { "epoch": 0.27, "learning_rate": 0.009393364366277242, "loss": 1.602, "step": 270 }, { "epoch": 0.27, "learning_rate": 0.00947994098324202, "loss": 1.5175, "step": 275 }, { "epoch": 0.28, "learning_rate": 0.009565734053059192, "loss": 1.4379, "step": 280 }, { "epoch": 0.28, "learning_rate": 0.009650764472411541, "loss": 1.5895, "step": 285 }, { "epoch": 0.29, "learning_rate": 0.009735052225338362, "loss": 1.9034, "step": 290 }, { "epoch": 0.29, "learning_rate": 0.00981861643808179, "loss": 1.5616, "step": 295 }, { "epoch": 0.3, "learning_rate": 0.009901475429766745, "loss": 1.5318, "step": 300 }, { "epoch": 0.3, "learning_rate": 0.009983646759294876, "loss": 1.6532, "step": 305 }, { "epoch": 0.31, "learning_rate": 0.009976905542725173, "loss": 1.6601, "step": 310 }, { "epoch": 0.31, "learning_rate": 0.00994803747113164, "loss": 1.9683, "step": 315 }, { "epoch": 0.32, "learning_rate": 0.009919169399538106, "loss": 1.694, "step": 320 }, { "epoch": 0.32, "learning_rate": 0.009890301327944573, "loss": 1.4985, "step": 325 }, { "epoch": 0.33, "learning_rate": 0.009861433256351039, "loss": 1.7896, "step": 330 }, { "epoch": 0.33, "learning_rate": 0.009832565184757506, "loss": 1.6075, "step": 335 }, { "epoch": 0.34, "learning_rate": 0.009803697113163973, "loss": 1.7854, "step": 340 }, { "epoch": 0.34, "learning_rate": 0.009774829041570439, "loss": 1.6752, "step": 345 }, { "epoch": 0.35, "learning_rate": 0.009745960969976906, "loss": 1.6439, "step": 350 }, { "epoch": 0.35, "learning_rate": 0.009717092898383371, "loss": 1.9989, "step": 355 }, { "epoch": 0.36, "learning_rate": 0.009688224826789838, "loss": 1.8198, "step": 360 }, { "epoch": 0.36, "learning_rate": 0.009659356755196306, "loss": 1.8466, "step": 365 }, { "epoch": 0.37, "learning_rate": 0.009630488683602771, "loss": 1.7034, "step": 370 }, { "epoch": 0.37, "learning_rate": 0.009601620612009238, "loss": 1.6235, "step": 375 }, { "epoch": 0.38, "learning_rate": 0.009572752540415704, "loss": 1.5958, "step": 380 }, { "epoch": 0.38, "learning_rate": 0.009543884468822171, "loss": 1.7292, "step": 385 }, { "epoch": 0.39, "learning_rate": 0.009515016397228638, "loss": 1.6004, "step": 390 }, { "epoch": 0.39, "learning_rate": 0.009486148325635104, "loss": 1.8302, "step": 395 }, { "epoch": 0.4, "learning_rate": 0.009457280254041571, "loss": 1.6322, "step": 400 }, { "epoch": 0.4, "learning_rate": 0.009428412182448036, "loss": 1.4802, "step": 405 }, { "epoch": 0.41, "learning_rate": 0.009399544110854504, "loss": 1.5544, "step": 410 }, { "epoch": 0.41, "learning_rate": 0.009370676039260971, "loss": 1.6265, "step": 415 }, { "epoch": 0.42, "learning_rate": 0.009341807967667436, "loss": 1.7169, "step": 420 }, { "epoch": 0.42, "learning_rate": 0.009312939896073904, "loss": 1.511, "step": 425 }, { "epoch": 0.43, "learning_rate": 0.009284071824480369, "loss": 1.63, "step": 430 }, { "epoch": 0.43, "learning_rate": 0.009255203752886836, "loss": 1.6076, "step": 435 }, { "epoch": 0.44, "learning_rate": 0.009226335681293303, "loss": 1.5973, "step": 440 }, { "epoch": 0.44, "learning_rate": 0.009197467609699769, "loss": 1.6757, "step": 445 }, { "epoch": 0.45, "learning_rate": 0.009168599538106236, "loss": 1.5146, "step": 450 }, { "epoch": 0.45, "learning_rate": 0.009139731466512702, "loss": 1.459, "step": 455 }, { "epoch": 0.46, "learning_rate": 0.009110863394919169, "loss": 1.4292, "step": 460 }, { "epoch": 0.46, "learning_rate": 0.009081995323325636, "loss": 1.7206, "step": 465 }, { "epoch": 0.47, "learning_rate": 0.009053127251732102, "loss": 1.491, "step": 470 }, { "epoch": 0.47, "learning_rate": 0.009024259180138567, "loss": 1.5426, "step": 475 }, { "epoch": 0.48, "learning_rate": 0.008995391108545034, "loss": 1.6884, "step": 480 }, { "epoch": 0.48, "learning_rate": 0.008966523036951501, "loss": 1.6555, "step": 485 }, { "epoch": 0.49, "learning_rate": 0.008937654965357967, "loss": 1.6036, "step": 490 }, { "epoch": 0.49, "learning_rate": 0.008908786893764434, "loss": 1.4405, "step": 495 }, { "epoch": 0.5, "learning_rate": 0.008879918822170901, "loss": 1.5134, "step": 500 }, { "epoch": 0.5, "learning_rate": 0.008851050750577369, "loss": 1.6087, "step": 505 }, { "epoch": 0.51, "learning_rate": 0.008822182678983834, "loss": 1.6898, "step": 510 }, { "epoch": 0.51, "learning_rate": 0.0087933146073903, "loss": 1.7172, "step": 515 }, { "epoch": 0.52, "learning_rate": 0.008764446535796767, "loss": 1.6621, "step": 520 }, { "epoch": 0.52, "learning_rate": 0.008735578464203234, "loss": 1.4144, "step": 525 }, { "epoch": 0.53, "learning_rate": 0.008706710392609701, "loss": 1.685, "step": 530 }, { "epoch": 0.53, "learning_rate": 0.008677842321016167, "loss": 1.6112, "step": 535 }, { "epoch": 0.54, "learning_rate": 0.008648974249422632, "loss": 1.5041, "step": 540 }, { "epoch": 0.54, "learning_rate": 0.0086201061778291, "loss": 1.5515, "step": 545 }, { "epoch": 0.55, "learning_rate": 0.008591238106235567, "loss": 1.711, "step": 550 }, { "epoch": 0.55, "learning_rate": 0.008562370034642034, "loss": 1.6166, "step": 555 }, { "epoch": 0.56, "learning_rate": 0.0085335019630485, "loss": 1.5324, "step": 560 }, { "epoch": 0.56, "learning_rate": 0.008504633891454965, "loss": 1.4422, "step": 565 }, { "epoch": 0.57, "learning_rate": 0.008475765819861432, "loss": 1.6069, "step": 570 }, { "epoch": 0.57, "learning_rate": 0.008446897748267899, "loss": 1.6231, "step": 575 }, { "epoch": 0.58, "learning_rate": 0.008418029676674366, "loss": 1.3719, "step": 580 }, { "epoch": 0.58, "learning_rate": 0.008389161605080832, "loss": 1.6578, "step": 585 }, { "epoch": 0.59, "learning_rate": 0.008360293533487297, "loss": 1.6525, "step": 590 }, { "epoch": 0.59, "learning_rate": 0.008331425461893764, "loss": 1.6094, "step": 595 }, { "epoch": 0.6, "learning_rate": 0.008302557390300232, "loss": 1.633, "step": 600 }, { "epoch": 0.6, "learning_rate": 0.008273689318706697, "loss": 1.7865, "step": 605 }, { "epoch": 0.61, "learning_rate": 0.008244821247113164, "loss": 1.602, "step": 610 }, { "epoch": 0.61, "learning_rate": 0.00821595317551963, "loss": 1.6723, "step": 615 }, { "epoch": 0.62, "learning_rate": 0.008187085103926097, "loss": 1.5495, "step": 620 }, { "epoch": 0.62, "learning_rate": 0.008158217032332564, "loss": 1.5716, "step": 625 }, { "epoch": 0.63, "learning_rate": 0.00812934896073903, "loss": 1.6036, "step": 630 }, { "epoch": 0.63, "learning_rate": 0.008100480889145497, "loss": 1.4638, "step": 635 }, { "epoch": 0.64, "learning_rate": 0.008071612817551962, "loss": 1.6912, "step": 640 }, { "epoch": 0.64, "learning_rate": 0.00804274474595843, "loss": 1.6037, "step": 645 }, { "epoch": 0.65, "learning_rate": 0.008013876674364897, "loss": 1.5457, "step": 650 }, { "epoch": 0.65, "learning_rate": 0.007985008602771362, "loss": 1.4561, "step": 655 }, { "epoch": 0.66, "learning_rate": 0.00795614053117783, "loss": 1.4871, "step": 660 }, { "epoch": 0.66, "learning_rate": 0.007927272459584295, "loss": 1.5253, "step": 665 }, { "epoch": 0.67, "learning_rate": 0.007898404387990762, "loss": 1.5775, "step": 670 }, { "epoch": 0.67, "learning_rate": 0.00786953631639723, "loss": 1.724, "step": 675 }, { "epoch": 0.68, "learning_rate": 0.007840668244803695, "loss": 1.6629, "step": 680 }, { "epoch": 0.68, "learning_rate": 0.007811800173210163, "loss": 1.519, "step": 685 }, { "epoch": 0.69, "learning_rate": 0.00778293210161663, "loss": 1.5684, "step": 690 }, { "epoch": 0.69, "learning_rate": 0.007754064030023096, "loss": 1.6109, "step": 695 }, { "epoch": 0.7, "learning_rate": 0.007725195958429562, "loss": 1.6378, "step": 700 }, { "epoch": 0.7, "learning_rate": 0.007696327886836028, "loss": 1.4673, "step": 705 }, { "epoch": 0.71, "learning_rate": 0.007667459815242496, "loss": 1.3671, "step": 710 }, { "epoch": 0.71, "learning_rate": 0.007638591743648962, "loss": 1.6971, "step": 715 }, { "epoch": 0.72, "learning_rate": 0.007609723672055428, "loss": 1.5656, "step": 720 }, { "epoch": 0.72, "learning_rate": 0.007580855600461894, "loss": 1.6135, "step": 725 }, { "epoch": 0.73, "learning_rate": 0.007551987528868362, "loss": 1.6247, "step": 730 }, { "epoch": 0.73, "learning_rate": 0.007523119457274827, "loss": 1.4691, "step": 735 }, { "epoch": 0.74, "learning_rate": 0.007494251385681295, "loss": 1.5108, "step": 740 }, { "epoch": 0.74, "learning_rate": 0.007465383314087762, "loss": 1.7138, "step": 745 }, { "epoch": 0.75, "learning_rate": 0.007436515242494226, "loss": 1.4466, "step": 750 }, { "epoch": 0.75, "learning_rate": 0.007407647170900693, "loss": 1.6503, "step": 755 }, { "epoch": 0.76, "learning_rate": 0.00737877909930716, "loss": 1.5161, "step": 760 }, { "epoch": 0.76, "learning_rate": 0.007349911027713627, "loss": 1.5638, "step": 765 }, { "epoch": 0.77, "learning_rate": 0.007321042956120093, "loss": 1.6321, "step": 770 }, { "epoch": 0.77, "learning_rate": 0.007292174884526559, "loss": 1.5957, "step": 775 }, { "epoch": 0.78, "learning_rate": 0.0072633068129330245, "loss": 1.4813, "step": 780 }, { "epoch": 0.78, "learning_rate": 0.0072344387413394926, "loss": 1.6003, "step": 785 }, { "epoch": 0.79, "learning_rate": 0.00720557066974596, "loss": 1.474, "step": 790 }, { "epoch": 0.79, "learning_rate": 0.007176702598152426, "loss": 1.6341, "step": 795 }, { "epoch": 0.8, "learning_rate": 0.007147834526558892, "loss": 1.6502, "step": 800 }, { "epoch": 0.8, "learning_rate": 0.007118966454965358, "loss": 1.4565, "step": 805 }, { "epoch": 0.81, "learning_rate": 0.007090098383371824, "loss": 1.4588, "step": 810 }, { "epoch": 0.81, "learning_rate": 0.007061230311778292, "loss": 1.5407, "step": 815 }, { "epoch": 0.82, "learning_rate": 0.007032362240184759, "loss": 1.3778, "step": 820 }, { "epoch": 0.82, "learning_rate": 0.007003494168591224, "loss": 1.5602, "step": 825 }, { "epoch": 0.83, "learning_rate": 0.0069746260969976905, "loss": 1.6861, "step": 830 }, { "epoch": 0.83, "learning_rate": 0.006945758025404158, "loss": 1.5511, "step": 835 }, { "epoch": 0.84, "learning_rate": 0.006916889953810624, "loss": 1.44, "step": 840 }, { "epoch": 0.84, "learning_rate": 0.006888021882217091, "loss": 1.4159, "step": 845 }, { "epoch": 0.85, "learning_rate": 0.006859153810623558, "loss": 1.6642, "step": 850 }, { "epoch": 0.85, "learning_rate": 0.006830285739030023, "loss": 1.461, "step": 855 }, { "epoch": 0.86, "learning_rate": 0.00680141766743649, "loss": 1.6532, "step": 860 }, { "epoch": 0.86, "learning_rate": 0.0067725495958429575, "loss": 1.5106, "step": 865 }, { "epoch": 0.87, "learning_rate": 0.006743681524249424, "loss": 1.564, "step": 870 }, { "epoch": 0.87, "learning_rate": 0.00671481345265589, "loss": 1.5954, "step": 875 }, { "epoch": 0.88, "learning_rate": 0.006685945381062356, "loss": 1.5075, "step": 880 }, { "epoch": 0.88, "learning_rate": 0.006657077309468823, "loss": 1.4821, "step": 885 }, { "epoch": 0.89, "learning_rate": 0.006628209237875289, "loss": 1.6993, "step": 890 }, { "epoch": 0.89, "learning_rate": 0.006599341166281756, "loss": 1.3872, "step": 895 }, { "epoch": 0.9, "learning_rate": 0.006570473094688223, "loss": 1.684, "step": 900 }, { "epoch": 0.9, "learning_rate": 0.006541605023094689, "loss": 1.3419, "step": 905 }, { "epoch": 0.91, "learning_rate": 0.006512736951501155, "loss": 1.3974, "step": 910 }, { "epoch": 0.91, "learning_rate": 0.006483868879907622, "loss": 1.5703, "step": 915 }, { "epoch": 0.92, "learning_rate": 0.006455000808314089, "loss": 1.5572, "step": 920 }, { "epoch": 0.92, "learning_rate": 0.006426132736720555, "loss": 1.5941, "step": 925 }, { "epoch": 0.93, "learning_rate": 0.006397264665127021, "loss": 1.706, "step": 930 }, { "epoch": 0.93, "learning_rate": 0.006368396593533487, "loss": 1.5107, "step": 935 }, { "epoch": 0.94, "learning_rate": 0.0063395285219399545, "loss": 1.4807, "step": 940 }, { "epoch": 0.94, "learning_rate": 0.006310660450346422, "loss": 1.3222, "step": 945 }, { "epoch": 0.95, "learning_rate": 0.006281792378752889, "loss": 1.4944, "step": 950 }, { "epoch": 0.95, "learning_rate": 0.006252924307159354, "loss": 1.4851, "step": 955 }, { "epoch": 0.96, "learning_rate": 0.006224056235565821, "loss": 1.3356, "step": 960 }, { "epoch": 0.96, "learning_rate": 0.006195188163972287, "loss": 1.5368, "step": 965 }, { "epoch": 0.97, "learning_rate": 0.006166320092378754, "loss": 1.4879, "step": 970 }, { "epoch": 0.97, "learning_rate": 0.006137452020785221, "loss": 1.7248, "step": 975 }, { "epoch": 0.98, "learning_rate": 0.006108583949191687, "loss": 1.5369, "step": 980 }, { "epoch": 0.98, "learning_rate": 0.006079715877598152, "loss": 1.6353, "step": 985 }, { "epoch": 0.99, "learning_rate": 0.0060508478060046205, "loss": 1.4272, "step": 990 }, { "epoch": 0.99, "learning_rate": 0.006021979734411086, "loss": 1.4956, "step": 995 }, { "epoch": 1.0, "learning_rate": 0.005993111662817553, "loss": 1.6992, "step": 1000 }, { "epoch": 1.0, "learning_rate": 0.00596424359122402, "loss": 1.49, "step": 1005 }, { "epoch": 1.01, "learning_rate": 0.005935375519630485, "loss": 1.5019, "step": 1010 }, { "epoch": 1.01, "learning_rate": 0.005906507448036952, "loss": 1.4958, "step": 1015 }, { "epoch": 1.02, "learning_rate": 0.005877639376443419, "loss": 1.5754, "step": 1020 }, { "epoch": 1.02, "learning_rate": 0.005848771304849886, "loss": 1.5251, "step": 1025 }, { "epoch": 1.03, "learning_rate": 0.005819903233256352, "loss": 1.4752, "step": 1030 }, { "epoch": 1.03, "learning_rate": 0.005791035161662818, "loss": 1.4888, "step": 1035 }, { "epoch": 1.04, "learning_rate": 0.005762167090069283, "loss": 1.7421, "step": 1040 }, { "epoch": 1.04, "learning_rate": 0.005733299018475752, "loss": 1.4161, "step": 1045 }, { "epoch": 1.05, "learning_rate": 0.005704430946882218, "loss": 1.5144, "step": 1050 }, { "epoch": 1.05, "learning_rate": 0.005675562875288685, "loss": 1.4593, "step": 1055 }, { "epoch": 1.06, "learning_rate": 0.00564669480369515, "loss": 1.5569, "step": 1060 }, { "epoch": 1.06, "learning_rate": 0.0056178267321016166, "loss": 1.699, "step": 1065 }, { "epoch": 1.07, "learning_rate": 0.005588958660508083, "loss": 1.382, "step": 1070 }, { "epoch": 1.07, "learning_rate": 0.005560090588914552, "loss": 1.5344, "step": 1075 }, { "epoch": 1.08, "learning_rate": 0.005531222517321017, "loss": 1.5147, "step": 1080 }, { "epoch": 1.08, "learning_rate": 0.005502354445727483, "loss": 1.483, "step": 1085 }, { "epoch": 1.09, "learning_rate": 0.005473486374133949, "loss": 1.4432, "step": 1090 }, { "epoch": 1.09, "learning_rate": 0.005444618302540416, "loss": 1.4741, "step": 1095 }, { "epoch": 1.1, "learning_rate": 0.005415750230946883, "loss": 1.3889, "step": 1100 }, { "epoch": 1.1, "learning_rate": 0.00538688215935335, "loss": 1.5193, "step": 1105 }, { "epoch": 1.11, "learning_rate": 0.005358014087759816, "loss": 1.3495, "step": 1110 }, { "epoch": 1.11, "learning_rate": 0.005329146016166282, "loss": 1.4738, "step": 1115 }, { "epoch": 1.12, "learning_rate": 0.005300277944572749, "loss": 1.4875, "step": 1120 }, { "epoch": 1.12, "learning_rate": 0.005271409872979214, "loss": 1.5795, "step": 1125 }, { "epoch": 1.13, "learning_rate": 0.0052425418013856825, "loss": 1.5649, "step": 1130 }, { "epoch": 1.13, "learning_rate": 0.005213673729792149, "loss": 1.3385, "step": 1135 }, { "epoch": 1.14, "learning_rate": 0.005184805658198614, "loss": 1.6503, "step": 1140 }, { "epoch": 1.14, "learning_rate": 0.005155937586605081, "loss": 1.4296, "step": 1145 }, { "epoch": 1.15, "learning_rate": 0.005127069515011548, "loss": 1.4508, "step": 1150 }, { "epoch": 1.15, "learning_rate": 0.005098201443418014, "loss": 1.6112, "step": 1155 }, { "epoch": 1.16, "learning_rate": 0.0050693333718244814, "loss": 1.4412, "step": 1160 }, { "epoch": 1.16, "learning_rate": 0.005040465300230946, "loss": 1.3971, "step": 1165 }, { "epoch": 1.17, "learning_rate": 0.005011597228637413, "loss": 1.5099, "step": 1170 }, { "epoch": 1.17, "learning_rate": 0.0049827291570438805, "loss": 1.5568, "step": 1175 }, { "epoch": 1.18, "learning_rate": 0.004953861085450347, "loss": 1.5243, "step": 1180 }, { "epoch": 1.18, "learning_rate": 0.004924993013856814, "loss": 1.2718, "step": 1185 }, { "epoch": 1.19, "learning_rate": 0.0048961249422632795, "loss": 1.6473, "step": 1190 }, { "epoch": 1.19, "learning_rate": 0.004867256870669747, "loss": 1.5639, "step": 1195 }, { "epoch": 1.2, "learning_rate": 0.004838388799076213, "loss": 1.3558, "step": 1200 }, { "epoch": 1.2, "learning_rate": 0.004809520727482679, "loss": 1.4211, "step": 1205 }, { "epoch": 1.21, "learning_rate": 0.004780652655889147, "loss": 1.5578, "step": 1210 }, { "epoch": 1.21, "learning_rate": 0.004751784584295613, "loss": 1.6661, "step": 1215 }, { "epoch": 1.22, "learning_rate": 0.0047229165127020785, "loss": 1.3678, "step": 1220 }, { "epoch": 1.22, "learning_rate": 0.004694048441108546, "loss": 1.4871, "step": 1225 }, { "epoch": 1.23, "learning_rate": 0.004665180369515012, "loss": 1.4524, "step": 1230 }, { "epoch": 1.23, "learning_rate": 0.004636312297921479, "loss": 1.4723, "step": 1235 }, { "epoch": 1.24, "learning_rate": 0.004607444226327946, "loss": 1.3289, "step": 1240 }, { "epoch": 1.24, "learning_rate": 0.004578576154734411, "loss": 1.5643, "step": 1245 }, { "epoch": 1.25, "learning_rate": 0.004549708083140878, "loss": 1.2773, "step": 1250 }, { "epoch": 1.25, "learning_rate": 0.004520840011547345, "loss": 1.4601, "step": 1255 }, { "epoch": 1.26, "learning_rate": 0.004491971939953812, "loss": 1.2864, "step": 1260 }, { "epoch": 1.26, "learning_rate": 0.004463103868360278, "loss": 1.3144, "step": 1265 }, { "epoch": 1.27, "learning_rate": 0.004434235796766744, "loss": 1.4182, "step": 1270 }, { "epoch": 1.27, "learning_rate": 0.004405367725173211, "loss": 1.5354, "step": 1275 }, { "epoch": 1.28, "learning_rate": 0.004376499653579677, "loss": 1.4776, "step": 1280 }, { "epoch": 1.28, "learning_rate": 0.0043476315819861435, "loss": 1.5618, "step": 1285 }, { "epoch": 1.29, "learning_rate": 0.004318763510392611, "loss": 1.3853, "step": 1290 }, { "epoch": 1.29, "learning_rate": 0.004289895438799076, "loss": 1.4081, "step": 1295 }, { "epoch": 1.3, "learning_rate": 0.0042610273672055434, "loss": 1.2231, "step": 1300 }, { "epoch": 1.3, "learning_rate": 0.00423215929561201, "loss": 1.4096, "step": 1305 }, { "epoch": 1.31, "learning_rate": 0.004203291224018476, "loss": 1.724, "step": 1310 }, { "epoch": 1.31, "learning_rate": 0.004174423152424943, "loss": 1.2597, "step": 1315 }, { "epoch": 1.32, "learning_rate": 0.004145555080831409, "loss": 1.3268, "step": 1320 }, { "epoch": 1.32, "learning_rate": 0.004116687009237876, "loss": 1.3372, "step": 1325 }, { "epoch": 1.33, "learning_rate": 0.004087818937644342, "loss": 1.4024, "step": 1330 }, { "epoch": 1.33, "learning_rate": 0.004058950866050809, "loss": 1.3837, "step": 1335 }, { "epoch": 1.34, "learning_rate": 0.004030082794457276, "loss": 1.4427, "step": 1340 }, { "epoch": 1.34, "learning_rate": 0.004001214722863742, "loss": 1.5162, "step": 1345 }, { "epoch": 1.35, "learning_rate": 0.003972346651270208, "loss": 1.5025, "step": 1350 }, { "epoch": 1.35, "learning_rate": 0.003943478579676675, "loss": 1.3713, "step": 1355 }, { "epoch": 1.36, "learning_rate": 0.003914610508083141, "loss": 1.2259, "step": 1360 }, { "epoch": 1.36, "learning_rate": 0.003885742436489608, "loss": 1.4999, "step": 1365 }, { "epoch": 1.37, "learning_rate": 0.003856874364896074, "loss": 1.3068, "step": 1370 }, { "epoch": 1.37, "learning_rate": 0.003828006293302541, "loss": 1.4383, "step": 1375 }, { "epoch": 1.38, "learning_rate": 0.003799138221709007, "loss": 1.3846, "step": 1380 }, { "epoch": 1.38, "learning_rate": 0.0037702701501154735, "loss": 1.4892, "step": 1385 }, { "epoch": 1.39, "learning_rate": 0.0037414020785219403, "loss": 1.4041, "step": 1390 }, { "epoch": 1.39, "learning_rate": 0.003712534006928406, "loss": 1.6152, "step": 1395 }, { "epoch": 1.4, "learning_rate": 0.003683665935334873, "loss": 1.3959, "step": 1400 }, { "epoch": 1.4, "learning_rate": 0.00365479786374134, "loss": 1.2466, "step": 1405 }, { "epoch": 1.41, "learning_rate": 0.003625929792147806, "loss": 1.4626, "step": 1410 }, { "epoch": 1.41, "learning_rate": 0.003597061720554272, "loss": 1.1847, "step": 1415 }, { "epoch": 1.42, "learning_rate": 0.003568193648960739, "loss": 1.3083, "step": 1420 }, { "epoch": 1.42, "learning_rate": 0.0035393255773672055, "loss": 1.3253, "step": 1425 }, { "epoch": 1.43, "learning_rate": 0.0035104575057736728, "loss": 1.372, "step": 1430 }, { "epoch": 1.43, "learning_rate": 0.0034815894341801387, "loss": 1.4528, "step": 1435 }, { "epoch": 1.44, "learning_rate": 0.003452721362586605, "loss": 1.5115, "step": 1440 }, { "epoch": 1.44, "learning_rate": 0.003423853290993072, "loss": 1.3855, "step": 1445 }, { "epoch": 1.45, "learning_rate": 0.0033949852193995377, "loss": 1.2547, "step": 1450 }, { "epoch": 1.45, "learning_rate": 0.0033661171478060045, "loss": 1.4584, "step": 1455 }, { "epoch": 1.46, "learning_rate": 0.0033372490762124713, "loss": 1.2172, "step": 1460 }, { "epoch": 1.46, "learning_rate": 0.0033083810046189376, "loss": 1.3224, "step": 1465 }, { "epoch": 1.47, "learning_rate": 0.003279512933025405, "loss": 1.5014, "step": 1470 }, { "epoch": 1.47, "learning_rate": 0.0032506448614318707, "loss": 1.4566, "step": 1475 }, { "epoch": 1.48, "learning_rate": 0.0032217767898383366, "loss": 1.5488, "step": 1480 }, { "epoch": 1.48, "learning_rate": 0.0031929087182448043, "loss": 1.3555, "step": 1485 }, { "epoch": 1.49, "learning_rate": 0.00316404064665127, "loss": 1.488, "step": 1490 }, { "epoch": 1.49, "learning_rate": 0.0031351725750577374, "loss": 1.3927, "step": 1495 }, { "epoch": 1.5, "learning_rate": 0.0031063045034642033, "loss": 1.1759, "step": 1500 }, { "epoch": 1.5, "learning_rate": 0.00307743643187067, "loss": 1.4156, "step": 1505 }, { "epoch": 1.51, "learning_rate": 0.0030485683602771364, "loss": 1.3284, "step": 1510 }, { "epoch": 1.51, "learning_rate": 0.003019700288683603, "loss": 1.5301, "step": 1515 }, { "epoch": 1.52, "learning_rate": 0.0029908322170900687, "loss": 1.4688, "step": 1520 }, { "epoch": 1.52, "learning_rate": 0.002961964145496536, "loss": 1.3861, "step": 1525 }, { "epoch": 1.53, "learning_rate": 0.0029330960739030023, "loss": 1.3078, "step": 1530 }, { "epoch": 1.53, "learning_rate": 0.00290422800230947, "loss": 1.5033, "step": 1535 }, { "epoch": 1.54, "learning_rate": 0.0028753599307159354, "loss": 1.3179, "step": 1540 }, { "epoch": 1.54, "learning_rate": 0.0028464918591224013, "loss": 1.3151, "step": 1545 }, { "epoch": 1.55, "learning_rate": 0.0028176237875288685, "loss": 1.2228, "step": 1550 }, { "epoch": 1.55, "learning_rate": 0.002788755715935335, "loss": 1.4193, "step": 1555 }, { "epoch": 1.56, "learning_rate": 0.002759887644341802, "loss": 1.5102, "step": 1560 }, { "epoch": 1.56, "learning_rate": 0.002731019572748268, "loss": 1.4108, "step": 1565 }, { "epoch": 1.57, "learning_rate": 0.0027021515011547343, "loss": 1.6702, "step": 1570 }, { "epoch": 1.57, "learning_rate": 0.002673283429561201, "loss": 1.4893, "step": 1575 }, { "epoch": 1.58, "learning_rate": 0.002644415357967667, "loss": 1.5393, "step": 1580 }, { "epoch": 1.58, "learning_rate": 0.002615547286374133, "loss": 1.3174, "step": 1585 }, { "epoch": 1.59, "learning_rate": 0.0025866792147806006, "loss": 1.5737, "step": 1590 }, { "epoch": 1.59, "learning_rate": 0.002557811143187067, "loss": 1.3413, "step": 1595 }, { "epoch": 1.6, "learning_rate": 0.002528943071593534, "loss": 1.4321, "step": 1600 }, { "epoch": 1.6, "learning_rate": 0.002500075, "loss": 1.5292, "step": 1605 }, { "epoch": 1.61, "learning_rate": 0.002471206928406466, "loss": 1.1747, "step": 1610 }, { "epoch": 1.61, "learning_rate": 0.002442338856812933, "loss": 1.3613, "step": 1615 }, { "epoch": 1.62, "learning_rate": 0.0024134707852193995, "loss": 1.4922, "step": 1620 }, { "epoch": 1.62, "learning_rate": 0.0023846027136258667, "loss": 1.4311, "step": 1625 }, { "epoch": 1.63, "learning_rate": 0.0023557346420323326, "loss": 1.4672, "step": 1630 }, { "epoch": 1.63, "learning_rate": 0.002326866570438799, "loss": 1.3726, "step": 1635 }, { "epoch": 1.64, "learning_rate": 0.0022979984988452658, "loss": 1.2646, "step": 1640 }, { "epoch": 1.64, "learning_rate": 0.002269130427251732, "loss": 1.2819, "step": 1645 }, { "epoch": 1.65, "learning_rate": 0.002240262355658198, "loss": 1.3917, "step": 1650 }, { "epoch": 1.65, "learning_rate": 0.0022113942840646652, "loss": 1.3339, "step": 1655 }, { "epoch": 1.66, "learning_rate": 0.0021825262124711316, "loss": 1.439, "step": 1660 }, { "epoch": 1.66, "learning_rate": 0.002153658140877599, "loss": 1.3474, "step": 1665 }, { "epoch": 1.67, "learning_rate": 0.0021247900692840647, "loss": 1.1514, "step": 1670 }, { "epoch": 1.67, "learning_rate": 0.0020959219976905306, "loss": 1.3472, "step": 1675 }, { "epoch": 1.68, "learning_rate": 0.002067053926096998, "loss": 1.3059, "step": 1680 }, { "epoch": 1.68, "learning_rate": 0.002038185854503464, "loss": 1.4306, "step": 1685 }, { "epoch": 1.69, "learning_rate": 0.0020093177829099314, "loss": 1.4762, "step": 1690 }, { "epoch": 1.69, "learning_rate": 0.0019804497113163973, "loss": 1.2309, "step": 1695 }, { "epoch": 1.7, "learning_rate": 0.0019515816397228636, "loss": 1.4477, "step": 1700 }, { "epoch": 1.7, "learning_rate": 0.0019227135681293308, "loss": 1.3584, "step": 1705 }, { "epoch": 1.71, "learning_rate": 0.001893845496535797, "loss": 1.2721, "step": 1710 }, { "epoch": 1.71, "learning_rate": 0.0018649774249422629, "loss": 1.3599, "step": 1715 }, { "epoch": 1.72, "learning_rate": 0.0018361093533487299, "loss": 1.3039, "step": 1720 }, { "epoch": 1.72, "learning_rate": 0.0018072412817551962, "loss": 1.4414, "step": 1725 }, { "epoch": 1.73, "learning_rate": 0.0017783732101616634, "loss": 1.3903, "step": 1730 }, { "epoch": 1.73, "learning_rate": 0.0017495051385681296, "loss": 1.4272, "step": 1735 }, { "epoch": 1.74, "learning_rate": 0.0017206370669745957, "loss": 1.3649, "step": 1740 }, { "epoch": 1.74, "learning_rate": 0.0016917689953810627, "loss": 1.2328, "step": 1745 }, { "epoch": 1.75, "learning_rate": 0.001662900923787529, "loss": 1.2621, "step": 1750 }, { "epoch": 1.75, "learning_rate": 0.001634032852193996, "loss": 1.2631, "step": 1755 }, { "epoch": 1.76, "learning_rate": 0.0016051647806004622, "loss": 1.3229, "step": 1760 }, { "epoch": 1.76, "learning_rate": 0.0015762967090069283, "loss": 1.4429, "step": 1765 }, { "epoch": 1.77, "learning_rate": 0.0015474286374133955, "loss": 1.3552, "step": 1770 }, { "epoch": 1.77, "learning_rate": 0.0015185605658198618, "loss": 1.3257, "step": 1775 }, { "epoch": 1.78, "learning_rate": 0.0014896924942263275, "loss": 1.1749, "step": 1780 }, { "epoch": 1.78, "learning_rate": 0.0014608244226327945, "loss": 1.2705, "step": 1785 }, { "epoch": 1.79, "learning_rate": 0.0014319563510392609, "loss": 1.4362, "step": 1790 }, { "epoch": 1.79, "learning_rate": 0.001403088279445728, "loss": 1.1888, "step": 1795 }, { "epoch": 1.8, "learning_rate": 0.0013742202078521942, "loss": 1.3183, "step": 1800 }, { "epoch": 1.8, "learning_rate": 0.0013453521362586604, "loss": 1.4149, "step": 1805 }, { "epoch": 1.81, "learning_rate": 0.0013164840646651273, "loss": 1.2567, "step": 1810 }, { "epoch": 1.81, "learning_rate": 0.0012876159930715937, "loss": 1.251, "step": 1815 }, { "epoch": 1.82, "learning_rate": 0.0012587479214780596, "loss": 1.1756, "step": 1820 }, { "epoch": 1.82, "learning_rate": 0.0012298798498845268, "loss": 1.4292, "step": 1825 }, { "epoch": 1.83, "learning_rate": 0.001201011778290993, "loss": 1.354, "step": 1830 }, { "epoch": 1.83, "learning_rate": 0.0011721437066974602, "loss": 1.2577, "step": 1835 }, { "epoch": 1.84, "learning_rate": 0.0011432756351039263, "loss": 1.3255, "step": 1840 }, { "epoch": 1.84, "learning_rate": 0.0011144075635103924, "loss": 1.1484, "step": 1845 }, { "epoch": 1.85, "learning_rate": 0.0010855394919168594, "loss": 1.3225, "step": 1850 }, { "epoch": 1.85, "learning_rate": 0.0010566714203233255, "loss": 1.4918, "step": 1855 }, { "epoch": 1.86, "learning_rate": 0.0010278033487297927, "loss": 1.2976, "step": 1860 }, { "epoch": 1.86, "learning_rate": 0.0009989352771362589, "loss": 1.2532, "step": 1865 }, { "epoch": 1.87, "learning_rate": 0.000970067205542725, "loss": 1.3491, "step": 1870 }, { "epoch": 1.87, "learning_rate": 0.000941199133949192, "loss": 1.3206, "step": 1875 }, { "epoch": 1.88, "learning_rate": 0.0009123310623556581, "loss": 1.4938, "step": 1880 }, { "epoch": 1.88, "learning_rate": 0.0008834629907621243, "loss": 1.424, "step": 1885 }, { "epoch": 1.89, "learning_rate": 0.0008545949191685913, "loss": 1.3984, "step": 1890 }, { "epoch": 1.89, "learning_rate": 0.0008257268475750575, "loss": 1.1661, "step": 1895 }, { "epoch": 1.9, "learning_rate": 0.0007968587759815248, "loss": 1.4248, "step": 1900 }, { "epoch": 1.9, "learning_rate": 0.0007679907043879908, "loss": 1.357, "step": 1905 }, { "epoch": 1.91, "learning_rate": 0.0007391226327944569, "loss": 1.269, "step": 1910 }, { "epoch": 1.91, "learning_rate": 0.0007102545612009241, "loss": 1.3927, "step": 1915 }, { "epoch": 1.92, "learning_rate": 0.0006813864896073901, "loss": 1.2803, "step": 1920 }, { "epoch": 1.92, "learning_rate": 0.0006525184180138573, "loss": 1.5049, "step": 1925 }, { "epoch": 1.93, "learning_rate": 0.0006236503464203234, "loss": 1.3776, "step": 1930 }, { "epoch": 1.93, "learning_rate": 0.0005947822748267896, "loss": 1.3658, "step": 1935 }, { "epoch": 1.94, "learning_rate": 0.0005659142032332567, "loss": 1.0963, "step": 1940 }, { "epoch": 1.94, "learning_rate": 0.0005370461316397228, "loss": 1.4584, "step": 1945 }, { "epoch": 1.95, "learning_rate": 0.0005081780600461889, "loss": 1.2742, "step": 1950 }, { "epoch": 1.95, "learning_rate": 0.0004793099884526561, "loss": 1.1752, "step": 1955 }, { "epoch": 1.96, "learning_rate": 0.0004504419168591222, "loss": 1.1821, "step": 1960 }, { "epoch": 1.96, "learning_rate": 0.0004215738452655894, "loss": 1.3403, "step": 1965 }, { "epoch": 1.97, "learning_rate": 0.00039270577367205543, "loss": 1.449, "step": 1970 }, { "epoch": 1.97, "learning_rate": 0.00036383770207852156, "loss": 1.1671, "step": 1975 }, { "epoch": 1.98, "learning_rate": 0.00033496963048498877, "loss": 1.4356, "step": 1980 }, { "epoch": 1.98, "learning_rate": 0.00030610155889145485, "loss": 1.2352, "step": 1985 }, { "epoch": 1.99, "learning_rate": 0.00027723348729792206, "loss": 1.3243, "step": 1990 }, { "epoch": 1.99, "learning_rate": 0.00024836541570438813, "loss": 1.2791, "step": 1995 }, { "epoch": 2.0, "learning_rate": 0.0002194973441108542, "loss": 1.2112, "step": 2000 }, { "epoch": 2.0, "learning_rate": 0.0001906292725173214, "loss": 1.1661, "step": 2005 }, { "epoch": 2.01, "learning_rate": 0.0001617612009237875, "loss": 1.4046, "step": 2010 }, { "epoch": 2.01, "learning_rate": 0.00013289312933025357, "loss": 1.2514, "step": 2015 }, { "epoch": 2.02, "learning_rate": 0.00010402505773672078, "loss": 1.1908, "step": 2020 }, { "epoch": 2.02, "learning_rate": 7.515698614318685e-05, "loss": 1.3698, "step": 2025 }, { "epoch": 2.03, "learning_rate": 4.628891454965405e-05, "loss": 1.5172, "step": 2030 }, { "epoch": 2.03, "learning_rate": 1.7420842956120128e-05, "loss": 1.3045, "step": 2035 }, { "epoch": 2.04, "step": 2038, "total_flos": 1.1512671589937643e+23, "train_loss": 1.4677173125030247, "train_runtime": 22959.9641, "train_samples_per_second": 1454.296, "train_steps_per_second": 0.089 } ], "max_steps": 2038, "num_train_epochs": 3, "start_time": 1669316150.4760864, "total_flos": 1.1512671589937643e+23, "trial_name": null, "trial_params": null }