{ "best_metric": 0.3091997504234314, "best_model_checkpoint": "../../saves/LLaMA3-70B-qlora-bnb/lora/sft/AG_16000-2/checkpoint-1900", "epoch": 1.3071895424836601, "eval_steps": 100, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.006535947712418301, "grad_norm": 13.670208930969238, "learning_rate": 8.714596949891069e-07, "loss": 2.4115, "step": 10 }, { "epoch": 0.013071895424836602, "grad_norm": 31.536911010742188, "learning_rate": 2.6143790849673204e-06, "loss": 2.6282, "step": 20 }, { "epoch": 0.0196078431372549, "grad_norm": 14.474563598632812, "learning_rate": 4.7930283224400875e-06, "loss": 2.5659, "step": 30 }, { "epoch": 0.026143790849673203, "grad_norm": 14.182409286499023, "learning_rate": 6.971677559912855e-06, "loss": 2.3685, "step": 40 }, { "epoch": 0.032679738562091505, "grad_norm": 34.63951873779297, "learning_rate": 8.932461873638345e-06, "loss": 2.351, "step": 50 }, { "epoch": 0.0392156862745098, "grad_norm": 23.15449333190918, "learning_rate": 1.1111111111111112e-05, "loss": 2.1125, "step": 60 }, { "epoch": 0.0457516339869281, "grad_norm": 32.01618576049805, "learning_rate": 1.328976034858388e-05, "loss": 1.5225, "step": 70 }, { "epoch": 0.05228758169934641, "grad_norm": 8.76276969909668, "learning_rate": 1.5468409586056645e-05, "loss": 1.1141, "step": 80 }, { "epoch": 0.058823529411764705, "grad_norm": 12.724370956420898, "learning_rate": 1.7647058823529414e-05, "loss": 0.654, "step": 90 }, { "epoch": 0.06535947712418301, "grad_norm": 13.92119312286377, "learning_rate": 1.982570806100218e-05, "loss": 0.4776, "step": 100 }, { "epoch": 0.06535947712418301, "eval_loss": 0.460627019405365, "eval_runtime": 7467.6581, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 100 }, { "epoch": 0.0718954248366013, "grad_norm": 13.412103652954102, "learning_rate": 2.2004357298474944e-05, "loss": 0.4607, "step": 110 }, { "epoch": 0.0784313725490196, "grad_norm": 16.269054412841797, "learning_rate": 2.4183006535947712e-05, "loss": 0.3912, "step": 120 }, { "epoch": 0.08496732026143791, "grad_norm": 5.597168922424316, "learning_rate": 2.636165577342048e-05, "loss": 0.3446, "step": 130 }, { "epoch": 0.0915032679738562, "grad_norm": 7.308394432067871, "learning_rate": 2.854030501089325e-05, "loss": 0.3572, "step": 140 }, { "epoch": 0.09803921568627451, "grad_norm": 8.47480583190918, "learning_rate": 3.0718954248366014e-05, "loss": 0.3224, "step": 150 }, { "epoch": 0.10457516339869281, "grad_norm": 6.073367595672607, "learning_rate": 3.289760348583878e-05, "loss": 0.4026, "step": 160 }, { "epoch": 0.1111111111111111, "grad_norm": 6.276689052581787, "learning_rate": 3.507625272331155e-05, "loss": 0.367, "step": 170 }, { "epoch": 0.11764705882352941, "grad_norm": 11.574933052062988, "learning_rate": 3.725490196078432e-05, "loss": 0.3921, "step": 180 }, { "epoch": 0.12418300653594772, "grad_norm": 11.751296043395996, "learning_rate": 3.943355119825709e-05, "loss": 0.3898, "step": 190 }, { "epoch": 0.13071895424836602, "grad_norm": 9.689138412475586, "learning_rate": 4.161220043572985e-05, "loss": 0.3675, "step": 200 }, { "epoch": 0.13071895424836602, "eval_loss": 0.3592655658721924, "eval_runtime": 7465.6804, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 200 }, { "epoch": 0.13725490196078433, "grad_norm": 5.775482177734375, "learning_rate": 4.379084967320262e-05, "loss": 0.3693, "step": 210 }, { "epoch": 0.1437908496732026, "grad_norm": 12.238544464111328, "learning_rate": 4.5969498910675387e-05, "loss": 0.4207, "step": 220 }, { "epoch": 0.1503267973856209, "grad_norm": 6.162591457366943, "learning_rate": 4.814814814814815e-05, "loss": 0.3702, "step": 230 }, { "epoch": 0.1568627450980392, "grad_norm": 5.743127346038818, "learning_rate": 5.032679738562092e-05, "loss": 0.3505, "step": 240 }, { "epoch": 0.16339869281045752, "grad_norm": 12.115300178527832, "learning_rate": 5.250544662309368e-05, "loss": 0.3654, "step": 250 }, { "epoch": 0.16993464052287582, "grad_norm": 12.006166458129883, "learning_rate": 5.4684095860566454e-05, "loss": 0.352, "step": 260 }, { "epoch": 0.17647058823529413, "grad_norm": 3.973567247390747, "learning_rate": 5.6862745098039215e-05, "loss": 0.36, "step": 270 }, { "epoch": 0.1830065359477124, "grad_norm": 5.307390213012695, "learning_rate": 5.904139433551199e-05, "loss": 0.3475, "step": 280 }, { "epoch": 0.1895424836601307, "grad_norm": 5.818578243255615, "learning_rate": 6.122004357298475e-05, "loss": 0.4053, "step": 290 }, { "epoch": 0.19607843137254902, "grad_norm": 5.835134506225586, "learning_rate": 6.339869281045751e-05, "loss": 0.3761, "step": 300 }, { "epoch": 0.19607843137254902, "eval_loss": 0.4034684896469116, "eval_runtime": 7466.244, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 300 }, { "epoch": 0.20261437908496732, "grad_norm": 3.4767305850982666, "learning_rate": 6.557734204793029e-05, "loss": 0.3709, "step": 310 }, { "epoch": 0.20915032679738563, "grad_norm": 5.347959995269775, "learning_rate": 6.775599128540305e-05, "loss": 0.35, "step": 320 }, { "epoch": 0.21568627450980393, "grad_norm": 5.1962480545043945, "learning_rate": 6.993464052287581e-05, "loss": 0.3892, "step": 330 }, { "epoch": 0.2222222222222222, "grad_norm": 3.1436469554901123, "learning_rate": 7.211328976034859e-05, "loss": 0.3538, "step": 340 }, { "epoch": 0.22875816993464052, "grad_norm": 2.677011489868164, "learning_rate": 7.429193899782135e-05, "loss": 0.3533, "step": 350 }, { "epoch": 0.23529411764705882, "grad_norm": 11.516694068908691, "learning_rate": 7.647058823529411e-05, "loss": 0.4018, "step": 360 }, { "epoch": 0.24183006535947713, "grad_norm": 10.96320629119873, "learning_rate": 7.864923747276689e-05, "loss": 0.43, "step": 370 }, { "epoch": 0.24836601307189543, "grad_norm": 6.594156265258789, "learning_rate": 8.082788671023965e-05, "loss": 0.3737, "step": 380 }, { "epoch": 0.2549019607843137, "grad_norm": 4.763483047485352, "learning_rate": 8.300653594771242e-05, "loss": 0.3629, "step": 390 }, { "epoch": 0.26143790849673204, "grad_norm": 3.6535167694091797, "learning_rate": 8.518518518518518e-05, "loss": 0.3465, "step": 400 }, { "epoch": 0.26143790849673204, "eval_loss": 0.35558873414993286, "eval_runtime": 7465.9696, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 400 }, { "epoch": 0.2679738562091503, "grad_norm": 10.275704383850098, "learning_rate": 8.736383442265795e-05, "loss": 0.3435, "step": 410 }, { "epoch": 0.27450980392156865, "grad_norm": 6.834077835083008, "learning_rate": 8.954248366013072e-05, "loss": 0.3828, "step": 420 }, { "epoch": 0.28104575163398693, "grad_norm": 13.24137020111084, "learning_rate": 9.172113289760348e-05, "loss": 0.3704, "step": 430 }, { "epoch": 0.2875816993464052, "grad_norm": 9.123649597167969, "learning_rate": 9.389978213507626e-05, "loss": 0.3569, "step": 440 }, { "epoch": 0.29411764705882354, "grad_norm": 6.183751106262207, "learning_rate": 9.607843137254903e-05, "loss": 0.4278, "step": 450 }, { "epoch": 0.3006535947712418, "grad_norm": 7.0605645179748535, "learning_rate": 9.82570806100218e-05, "loss": 0.3747, "step": 460 }, { "epoch": 0.30718954248366015, "grad_norm": 4.283732891082764, "learning_rate": 9.999994216519553e-05, "loss": 0.3289, "step": 470 }, { "epoch": 0.3137254901960784, "grad_norm": 12.275751113891602, "learning_rate": 9.999791796108715e-05, "loss": 0.4004, "step": 480 }, { "epoch": 0.3202614379084967, "grad_norm": 6.7879958152771, "learning_rate": 9.999300215054801e-05, "loss": 0.3768, "step": 490 }, { "epoch": 0.32679738562091504, "grad_norm": 3.2305805683135986, "learning_rate": 9.998519501788174e-05, "loss": 0.394, "step": 500 }, { "epoch": 0.32679738562091504, "eval_loss": 0.3434564173221588, "eval_runtime": 7464.9222, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 500 }, { "epoch": 0.3333333333333333, "grad_norm": 3.8374526500701904, "learning_rate": 9.997449701461023e-05, "loss": 0.3625, "step": 510 }, { "epoch": 0.33986928104575165, "grad_norm": 5.077846050262451, "learning_rate": 9.996090875944755e-05, "loss": 0.359, "step": 520 }, { "epoch": 0.3464052287581699, "grad_norm": 3.4323623180389404, "learning_rate": 9.994443103826414e-05, "loss": 0.356, "step": 530 }, { "epoch": 0.35294117647058826, "grad_norm": 4.085783004760742, "learning_rate": 9.992506480404138e-05, "loss": 0.3463, "step": 540 }, { "epoch": 0.35947712418300654, "grad_norm": 7.318523406982422, "learning_rate": 9.990281117681645e-05, "loss": 0.3896, "step": 550 }, { "epoch": 0.3660130718954248, "grad_norm": 16.51464080810547, "learning_rate": 9.987767144361759e-05, "loss": 0.343, "step": 560 }, { "epoch": 0.37254901960784315, "grad_norm": 2.0446882247924805, "learning_rate": 9.98496470583896e-05, "loss": 0.3291, "step": 570 }, { "epoch": 0.3790849673202614, "grad_norm": 2.331265687942505, "learning_rate": 9.981873964190987e-05, "loss": 0.3571, "step": 580 }, { "epoch": 0.38562091503267976, "grad_norm": 6.242280006408691, "learning_rate": 9.978495098169445e-05, "loss": 0.3624, "step": 590 }, { "epoch": 0.39215686274509803, "grad_norm": 1.5557068586349487, "learning_rate": 9.974828303189491e-05, "loss": 0.3579, "step": 600 }, { "epoch": 0.39215686274509803, "eval_loss": 0.341349333524704, "eval_runtime": 7465.0518, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 600 }, { "epoch": 0.39869281045751637, "grad_norm": 4.937715530395508, "learning_rate": 9.970873791318512e-05, "loss": 0.3576, "step": 610 }, { "epoch": 0.40522875816993464, "grad_norm": 4.85018253326416, "learning_rate": 9.966631791263872e-05, "loss": 0.3567, "step": 620 }, { "epoch": 0.4117647058823529, "grad_norm": 4.347261905670166, "learning_rate": 9.96210254835968e-05, "loss": 0.3372, "step": 630 }, { "epoch": 0.41830065359477125, "grad_norm": 4.017812728881836, "learning_rate": 9.9572863245526e-05, "loss": 0.3363, "step": 640 }, { "epoch": 0.42483660130718953, "grad_norm": 7.451604843139648, "learning_rate": 9.952183398386706e-05, "loss": 0.3269, "step": 650 }, { "epoch": 0.43137254901960786, "grad_norm": 12.005084991455078, "learning_rate": 9.946794064987371e-05, "loss": 0.3242, "step": 660 }, { "epoch": 0.43790849673202614, "grad_norm": 3.397099494934082, "learning_rate": 9.941118636044193e-05, "loss": 0.383, "step": 670 }, { "epoch": 0.4444444444444444, "grad_norm": 2.081789493560791, "learning_rate": 9.935157439792982e-05, "loss": 0.3354, "step": 680 }, { "epoch": 0.45098039215686275, "grad_norm": 3.097064733505249, "learning_rate": 9.928910820996756e-05, "loss": 0.3281, "step": 690 }, { "epoch": 0.45751633986928103, "grad_norm": 5.233646869659424, "learning_rate": 9.922379140925826e-05, "loss": 0.3524, "step": 700 }, { "epoch": 0.45751633986928103, "eval_loss": 0.3694455921649933, "eval_runtime": 7465.0105, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 700 }, { "epoch": 0.46405228758169936, "grad_norm": 4.417613983154297, "learning_rate": 9.915562777336879e-05, "loss": 0.3924, "step": 710 }, { "epoch": 0.47058823529411764, "grad_norm": 2.472074508666992, "learning_rate": 9.908462124451152e-05, "loss": 0.3488, "step": 720 }, { "epoch": 0.477124183006536, "grad_norm": 2.744296073913574, "learning_rate": 9.901077592931612e-05, "loss": 0.3291, "step": 730 }, { "epoch": 0.48366013071895425, "grad_norm": 3.440641403198242, "learning_rate": 9.893409609859222e-05, "loss": 0.3901, "step": 740 }, { "epoch": 0.49019607843137253, "grad_norm": 2.749112844467163, "learning_rate": 9.88545861870823e-05, "loss": 0.3535, "step": 750 }, { "epoch": 0.49673202614379086, "grad_norm": 4.953378200531006, "learning_rate": 9.877225079320526e-05, "loss": 0.3361, "step": 760 }, { "epoch": 0.5032679738562091, "grad_norm": 2.971996307373047, "learning_rate": 9.868709467879051e-05, "loss": 0.3772, "step": 770 }, { "epoch": 0.5098039215686274, "grad_norm": 3.6259024143218994, "learning_rate": 9.859912276880247e-05, "loss": 0.3476, "step": 780 }, { "epoch": 0.5163398692810458, "grad_norm": 3.4774110317230225, "learning_rate": 9.850834015105583e-05, "loss": 0.3622, "step": 790 }, { "epoch": 0.5228758169934641, "grad_norm": 5.781759262084961, "learning_rate": 9.841475207592122e-05, "loss": 0.3554, "step": 800 }, { "epoch": 0.5228758169934641, "eval_loss": 0.351721853017807, "eval_runtime": 7466.3074, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 800 }, { "epoch": 0.5294117647058824, "grad_norm": 2.1952409744262695, "learning_rate": 9.831836395602163e-05, "loss": 0.3345, "step": 810 }, { "epoch": 0.5359477124183006, "grad_norm": 5.838998317718506, "learning_rate": 9.821918136591934e-05, "loss": 0.3387, "step": 820 }, { "epoch": 0.5424836601307189, "grad_norm": 2.8472695350646973, "learning_rate": 9.811721004179352e-05, "loss": 0.3351, "step": 830 }, { "epoch": 0.5490196078431373, "grad_norm": 4.2800068855285645, "learning_rate": 9.801245588110848e-05, "loss": 0.3373, "step": 840 }, { "epoch": 0.5555555555555556, "grad_norm": 2.831066370010376, "learning_rate": 9.790492494227258e-05, "loss": 0.3718, "step": 850 }, { "epoch": 0.5620915032679739, "grad_norm": 2.9960384368896484, "learning_rate": 9.779462344428789e-05, "loss": 0.3391, "step": 860 }, { "epoch": 0.5686274509803921, "grad_norm": 2.505591869354248, "learning_rate": 9.768155776639044e-05, "loss": 0.356, "step": 870 }, { "epoch": 0.5751633986928104, "grad_norm": 5.350845813751221, "learning_rate": 9.756573444768133e-05, "loss": 0.3377, "step": 880 }, { "epoch": 0.5816993464052288, "grad_norm": 3.700345516204834, "learning_rate": 9.744716018674862e-05, "loss": 0.3306, "step": 890 }, { "epoch": 0.5882352941176471, "grad_norm": 3.254387140274048, "learning_rate": 9.732584184127973e-05, "loss": 0.3378, "step": 900 }, { "epoch": 0.5882352941176471, "eval_loss": 0.3639741837978363, "eval_runtime": 7465.3231, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 900 }, { "epoch": 0.5947712418300654, "grad_norm": 5.645087242126465, "learning_rate": 9.7201786427665e-05, "loss": 0.3604, "step": 910 }, { "epoch": 0.6013071895424836, "grad_norm": 6.0515456199646, "learning_rate": 9.707500112059183e-05, "loss": 0.3789, "step": 920 }, { "epoch": 0.6078431372549019, "grad_norm": 7.369515419006348, "learning_rate": 9.694549325262974e-05, "loss": 0.3571, "step": 930 }, { "epoch": 0.6143790849673203, "grad_norm": 7.788597106933594, "learning_rate": 9.681327031380629e-05, "loss": 0.3424, "step": 940 }, { "epoch": 0.6209150326797386, "grad_norm": 4.71685791015625, "learning_rate": 9.667833995117391e-05, "loss": 0.3474, "step": 950 }, { "epoch": 0.6274509803921569, "grad_norm": 4.404577255249023, "learning_rate": 9.654070996836765e-05, "loss": 0.3431, "step": 960 }, { "epoch": 0.6339869281045751, "grad_norm": 12.022799491882324, "learning_rate": 9.640038832515381e-05, "loss": 0.3246, "step": 970 }, { "epoch": 0.6405228758169934, "grad_norm": 3.9037022590637207, "learning_rate": 9.625738313696966e-05, "loss": 0.3394, "step": 980 }, { "epoch": 0.6470588235294118, "grad_norm": 3.2752866744995117, "learning_rate": 9.611170267445401e-05, "loss": 0.376, "step": 990 }, { "epoch": 0.6535947712418301, "grad_norm": 4.711609840393066, "learning_rate": 9.596335536296897e-05, "loss": 0.3245, "step": 1000 }, { "epoch": 0.6535947712418301, "eval_loss": 0.32997554540634155, "eval_runtime": 7465.3556, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1000 }, { "epoch": 0.6601307189542484, "grad_norm": 1.1813262701034546, "learning_rate": 9.581234978211257e-05, "loss": 0.3446, "step": 1010 }, { "epoch": 0.6666666666666666, "grad_norm": 3.696199417114258, "learning_rate": 9.565869466522265e-05, "loss": 0.3484, "step": 1020 }, { "epoch": 0.673202614379085, "grad_norm": 1.4563554525375366, "learning_rate": 9.550239889887179e-05, "loss": 0.3155, "step": 1030 }, { "epoch": 0.6797385620915033, "grad_norm": 1.6707898378372192, "learning_rate": 9.534347152235317e-05, "loss": 0.3364, "step": 1040 }, { "epoch": 0.6862745098039216, "grad_norm": 3.146296977996826, "learning_rate": 9.518192172715807e-05, "loss": 0.3286, "step": 1050 }, { "epoch": 0.6928104575163399, "grad_norm": 8.924692153930664, "learning_rate": 9.501775885644405e-05, "loss": 0.3337, "step": 1060 }, { "epoch": 0.6993464052287581, "grad_norm": 4.0824480056762695, "learning_rate": 9.485099240449474e-05, "loss": 0.328, "step": 1070 }, { "epoch": 0.7058823529411765, "grad_norm": 2.718278408050537, "learning_rate": 9.468163201617062e-05, "loss": 0.3186, "step": 1080 }, { "epoch": 0.7124183006535948, "grad_norm": 3.190894603729248, "learning_rate": 9.450968748635133e-05, "loss": 0.3397, "step": 1090 }, { "epoch": 0.7189542483660131, "grad_norm": 3.299405813217163, "learning_rate": 9.433516875936916e-05, "loss": 0.4178, "step": 1100 }, { "epoch": 0.7189542483660131, "eval_loss": 0.3399566113948822, "eval_runtime": 7464.7626, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1100 }, { "epoch": 0.7254901960784313, "grad_norm": 2.694213628768921, "learning_rate": 9.415808592843383e-05, "loss": 0.3442, "step": 1110 }, { "epoch": 0.7320261437908496, "grad_norm": 1.782759428024292, "learning_rate": 9.397844923504885e-05, "loss": 0.2938, "step": 1120 }, { "epoch": 0.738562091503268, "grad_norm": 2.5469982624053955, "learning_rate": 9.37962690684192e-05, "loss": 0.3099, "step": 1130 }, { "epoch": 0.7450980392156863, "grad_norm": 4.2174906730651855, "learning_rate": 9.361155596485046e-05, "loss": 0.3695, "step": 1140 }, { "epoch": 0.7516339869281046, "grad_norm": 2.9205563068389893, "learning_rate": 9.342432060713942e-05, "loss": 0.3253, "step": 1150 }, { "epoch": 0.7581699346405228, "grad_norm": 1.3839212656021118, "learning_rate": 9.323457382395628e-05, "loss": 0.3306, "step": 1160 }, { "epoch": 0.7647058823529411, "grad_norm": 3.993194341659546, "learning_rate": 9.304232658921839e-05, "loss": 0.3048, "step": 1170 }, { "epoch": 0.7712418300653595, "grad_norm": 3.2434442043304443, "learning_rate": 9.284759002145552e-05, "loss": 0.3235, "step": 1180 }, { "epoch": 0.7777777777777778, "grad_norm": 6.1412153244018555, "learning_rate": 9.26503753831669e-05, "loss": 0.2977, "step": 1190 }, { "epoch": 0.7843137254901961, "grad_norm": 6.04712438583374, "learning_rate": 9.245069408016977e-05, "loss": 0.2912, "step": 1200 }, { "epoch": 0.7843137254901961, "eval_loss": 0.34279727935791016, "eval_runtime": 7466.0217, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1200 }, { "epoch": 0.7908496732026143, "grad_norm": 1.9226710796356201, "learning_rate": 9.224855766093985e-05, "loss": 0.3681, "step": 1210 }, { "epoch": 0.7973856209150327, "grad_norm": 2.343475341796875, "learning_rate": 9.204397781594331e-05, "loss": 0.3081, "step": 1220 }, { "epoch": 0.803921568627451, "grad_norm": 4.996776103973389, "learning_rate": 9.183696637696077e-05, "loss": 0.3268, "step": 1230 }, { "epoch": 0.8104575163398693, "grad_norm": 3.761845111846924, "learning_rate": 9.162753531640292e-05, "loss": 1.001, "step": 1240 }, { "epoch": 0.8169934640522876, "grad_norm": 13.737248420715332, "learning_rate": 9.141569674661817e-05, "loss": 0.3769, "step": 1250 }, { "epoch": 0.8235294117647058, "grad_norm": 8.417398452758789, "learning_rate": 9.120146291919204e-05, "loss": 0.402, "step": 1260 }, { "epoch": 0.8300653594771242, "grad_norm": 3.1971797943115234, "learning_rate": 9.098484622423882e-05, "loss": 0.3705, "step": 1270 }, { "epoch": 0.8366013071895425, "grad_norm": 8.737996101379395, "learning_rate": 9.076585918968468e-05, "loss": 0.3536, "step": 1280 }, { "epoch": 0.8431372549019608, "grad_norm": 3.097640037536621, "learning_rate": 9.054451448054335e-05, "loss": 0.3847, "step": 1290 }, { "epoch": 0.8496732026143791, "grad_norm": 3.883129596710205, "learning_rate": 9.03208248981836e-05, "loss": 0.3905, "step": 1300 }, { "epoch": 0.8496732026143791, "eval_loss": 0.36324241757392883, "eval_runtime": 7459.1059, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1300 }, { "epoch": 0.8562091503267973, "grad_norm": 3.2262656688690186, "learning_rate": 9.009480337958883e-05, "loss": 0.351, "step": 1310 }, { "epoch": 0.8627450980392157, "grad_norm": 2.727170705795288, "learning_rate": 8.986646299660889e-05, "loss": 0.3583, "step": 1320 }, { "epoch": 0.869281045751634, "grad_norm": 6.688016414642334, "learning_rate": 8.963581695520408e-05, "loss": 0.3513, "step": 1330 }, { "epoch": 0.8758169934640523, "grad_norm": 1.6296730041503906, "learning_rate": 8.940287859468139e-05, "loss": 0.3563, "step": 1340 }, { "epoch": 0.8823529411764706, "grad_norm": 1.2997907400131226, "learning_rate": 8.916766138692303e-05, "loss": 0.3741, "step": 1350 }, { "epoch": 0.8888888888888888, "grad_norm": 7.427231788635254, "learning_rate": 8.893017893560727e-05, "loss": 0.3419, "step": 1360 }, { "epoch": 0.8954248366013072, "grad_norm": 1.4134552478790283, "learning_rate": 8.869044497542172e-05, "loss": 0.3585, "step": 1370 }, { "epoch": 0.9019607843137255, "grad_norm": 1.9037864208221436, "learning_rate": 8.844847337126893e-05, "loss": 0.3563, "step": 1380 }, { "epoch": 0.9084967320261438, "grad_norm": 2.263465404510498, "learning_rate": 8.820427811746456e-05, "loss": 0.3429, "step": 1390 }, { "epoch": 0.9150326797385621, "grad_norm": 3.114661455154419, "learning_rate": 8.795787333692807e-05, "loss": 0.3389, "step": 1400 }, { "epoch": 0.9150326797385621, "eval_loss": 0.35517746210098267, "eval_runtime": 7467.6922, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1400 }, { "epoch": 0.9215686274509803, "grad_norm": 2.826345682144165, "learning_rate": 8.770927328036575e-05, "loss": 0.3097, "step": 1410 }, { "epoch": 0.9281045751633987, "grad_norm": 2.172060012817383, "learning_rate": 8.745849232544681e-05, "loss": 0.3014, "step": 1420 }, { "epoch": 0.934640522875817, "grad_norm": 2.7672016620635986, "learning_rate": 8.720554497597159e-05, "loss": 0.3057, "step": 1430 }, { "epoch": 0.9411764705882353, "grad_norm": 6.170320510864258, "learning_rate": 8.695044586103296e-05, "loss": 0.3334, "step": 1440 }, { "epoch": 0.9477124183006536, "grad_norm": 3.521531581878662, "learning_rate": 8.669320973417006e-05, "loss": 0.3057, "step": 1450 }, { "epoch": 0.954248366013072, "grad_norm": 3.1676554679870605, "learning_rate": 8.643385147251515e-05, "loss": 0.2932, "step": 1460 }, { "epoch": 0.9607843137254902, "grad_norm": 2.741800546646118, "learning_rate": 8.617238607593319e-05, "loss": 0.352, "step": 1470 }, { "epoch": 0.9673202614379085, "grad_norm": 3.750753164291382, "learning_rate": 8.590882866615432e-05, "loss": 0.329, "step": 1480 }, { "epoch": 0.9738562091503268, "grad_norm": 4.675954818725586, "learning_rate": 8.564319448589926e-05, "loss": 0.2993, "step": 1490 }, { "epoch": 0.9803921568627451, "grad_norm": 3.0706264972686768, "learning_rate": 8.537549889799781e-05, "loss": 0.3422, "step": 1500 }, { "epoch": 0.9803921568627451, "eval_loss": 0.35487601161003113, "eval_runtime": 7465.2366, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1500 }, { "epoch": 0.9869281045751634, "grad_norm": 4.867375373840332, "learning_rate": 8.510575738450032e-05, "loss": 0.326, "step": 1510 }, { "epoch": 0.9934640522875817, "grad_norm": 4.509498119354248, "learning_rate": 8.483398554578232e-05, "loss": 0.2879, "step": 1520 }, { "epoch": 1.0, "grad_norm": 5.889903545379639, "learning_rate": 8.456019909964224e-05, "loss": 0.2913, "step": 1530 }, { "epoch": 1.0065359477124183, "grad_norm": 3.161592960357666, "learning_rate": 8.428441388039238e-05, "loss": 0.3207, "step": 1540 }, { "epoch": 1.0130718954248366, "grad_norm": 6.485883712768555, "learning_rate": 8.400664583794319e-05, "loss": 0.3892, "step": 1550 }, { "epoch": 1.0196078431372548, "grad_norm": 2.0741546154022217, "learning_rate": 8.372691103688079e-05, "loss": 0.2883, "step": 1560 }, { "epoch": 1.026143790849673, "grad_norm": 1.5439857244491577, "learning_rate": 8.34452256555378e-05, "loss": 0.3029, "step": 1570 }, { "epoch": 1.0326797385620916, "grad_norm": 4.525278568267822, "learning_rate": 8.316160598505784e-05, "loss": 0.2775, "step": 1580 }, { "epoch": 1.0392156862745099, "grad_norm": 3.212484121322632, "learning_rate": 8.28760684284532e-05, "loss": 0.3194, "step": 1590 }, { "epoch": 1.0457516339869282, "grad_norm": 3.7706832885742188, "learning_rate": 8.25886294996562e-05, "loss": 0.2597, "step": 1600 }, { "epoch": 1.0457516339869282, "eval_loss": 0.35163000226020813, "eval_runtime": 7465.7169, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1600 }, { "epoch": 1.0522875816993464, "grad_norm": 3.6558139324188232, "learning_rate": 8.22993058225642e-05, "loss": 0.3301, "step": 1610 }, { "epoch": 1.0588235294117647, "grad_norm": 2.815463066101074, "learning_rate": 8.200811413007807e-05, "loss": 0.2903, "step": 1620 }, { "epoch": 1.065359477124183, "grad_norm": 6.215457439422607, "learning_rate": 8.171507126313451e-05, "loss": 0.3399, "step": 1630 }, { "epoch": 1.0718954248366013, "grad_norm": 1.9962519407272339, "learning_rate": 8.142019416973199e-05, "loss": 0.3062, "step": 1640 }, { "epoch": 1.0784313725490196, "grad_norm": 1.8813626766204834, "learning_rate": 8.112349990395065e-05, "loss": 0.3419, "step": 1650 }, { "epoch": 1.0849673202614378, "grad_norm": 5.976492881774902, "learning_rate": 8.082500562496596e-05, "loss": 0.3135, "step": 1660 }, { "epoch": 1.091503267973856, "grad_norm": 2.2381093502044678, "learning_rate": 8.052472859605631e-05, "loss": 0.3222, "step": 1670 }, { "epoch": 1.0980392156862746, "grad_norm": 3.5524046421051025, "learning_rate": 8.02226861836046e-05, "loss": 0.2798, "step": 1680 }, { "epoch": 1.1045751633986929, "grad_norm": 5.400121688842773, "learning_rate": 7.991889585609387e-05, "loss": 0.3218, "step": 1690 }, { "epoch": 1.1111111111111112, "grad_norm": 37.609107971191406, "learning_rate": 7.961337518309704e-05, "loss": 0.3235, "step": 1700 }, { "epoch": 1.1111111111111112, "eval_loss": 0.325325071811676, "eval_runtime": 7465.0207, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1700 }, { "epoch": 1.1176470588235294, "grad_norm": 2.548970937728882, "learning_rate": 7.930614183426074e-05, "loss": 0.325, "step": 1710 }, { "epoch": 1.1241830065359477, "grad_norm": 2.6757702827453613, "learning_rate": 7.89972135782834e-05, "loss": 0.2955, "step": 1720 }, { "epoch": 1.130718954248366, "grad_norm": 2.190845012664795, "learning_rate": 7.868660828188765e-05, "loss": 0.3062, "step": 1730 }, { "epoch": 1.1372549019607843, "grad_norm": 2.709989547729492, "learning_rate": 7.837434390878698e-05, "loss": 0.3231, "step": 1740 }, { "epoch": 1.1437908496732025, "grad_norm": 2.8716742992401123, "learning_rate": 7.806043851864674e-05, "loss": 0.2853, "step": 1750 }, { "epoch": 1.1503267973856208, "grad_norm": 4.456597805023193, "learning_rate": 7.774491026603985e-05, "loss": 0.3226, "step": 1760 }, { "epoch": 1.156862745098039, "grad_norm": 13.620016098022461, "learning_rate": 7.742777739939666e-05, "loss": 0.2897, "step": 1770 }, { "epoch": 1.1633986928104576, "grad_norm": 2.6009891033172607, "learning_rate": 7.710905825994962e-05, "loss": 0.2482, "step": 1780 }, { "epoch": 1.1699346405228759, "grad_norm": 7.119050025939941, "learning_rate": 7.678877128067261e-05, "loss": 0.3552, "step": 1790 }, { "epoch": 1.1764705882352942, "grad_norm": 2.4954652786254883, "learning_rate": 7.646693498521471e-05, "loss": 0.3148, "step": 1800 }, { "epoch": 1.1764705882352942, "eval_loss": 0.31464409828186035, "eval_runtime": 7466.0649, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1800 }, { "epoch": 1.1830065359477124, "grad_norm": 2.542064905166626, "learning_rate": 7.614356798682904e-05, "loss": 0.2973, "step": 1810 }, { "epoch": 1.1895424836601307, "grad_norm": 2.8179709911346436, "learning_rate": 7.581868898729618e-05, "loss": 0.3049, "step": 1820 }, { "epoch": 1.196078431372549, "grad_norm": 1.8893694877624512, "learning_rate": 7.549231677584262e-05, "loss": 0.2867, "step": 1830 }, { "epoch": 1.2026143790849673, "grad_norm": 3.376964807510376, "learning_rate": 7.516447022805407e-05, "loss": 0.3599, "step": 1840 }, { "epoch": 1.2091503267973855, "grad_norm": 1.8863316774368286, "learning_rate": 7.483516830478379e-05, "loss": 0.3283, "step": 1850 }, { "epoch": 1.215686274509804, "grad_norm": 5.017312049865723, "learning_rate": 7.450443005105601e-05, "loss": 0.3335, "step": 1860 }, { "epoch": 1.2222222222222223, "grad_norm": 1.3233927488327026, "learning_rate": 7.417227459496445e-05, "loss": 0.3031, "step": 1870 }, { "epoch": 1.2287581699346406, "grad_norm": 2.100111484527588, "learning_rate": 7.383872114656611e-05, "loss": 0.3251, "step": 1880 }, { "epoch": 1.2352941176470589, "grad_norm": 3.839590311050415, "learning_rate": 7.35037889967702e-05, "loss": 0.2726, "step": 1890 }, { "epoch": 1.2418300653594772, "grad_norm": 3.1495704650878906, "learning_rate": 7.31674975162225e-05, "loss": 0.3502, "step": 1900 }, { "epoch": 1.2418300653594772, "eval_loss": 0.3091997504234314, "eval_runtime": 7466.4145, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 1900 }, { "epoch": 1.2483660130718954, "grad_norm": 5.134574890136719, "learning_rate": 7.282986615418503e-05, "loss": 0.2902, "step": 1910 }, { "epoch": 1.2549019607843137, "grad_norm": 2.6781771183013916, "learning_rate": 7.249091443741126e-05, "loss": 0.2741, "step": 1920 }, { "epoch": 1.261437908496732, "grad_norm": 6.247425079345703, "learning_rate": 7.215066196901676e-05, "loss": 0.3026, "step": 1930 }, { "epoch": 1.2679738562091503, "grad_norm": 3.1725962162017822, "learning_rate": 7.180912842734548e-05, "loss": 0.3574, "step": 1940 }, { "epoch": 1.2745098039215685, "grad_norm": 5.736270427703857, "learning_rate": 7.146633356483161e-05, "loss": 0.3166, "step": 1950 }, { "epoch": 1.2810457516339868, "grad_norm": 1.851462960243225, "learning_rate": 7.11222972068573e-05, "loss": 0.2737, "step": 1960 }, { "epoch": 1.287581699346405, "grad_norm": 2.9439609050750732, "learning_rate": 7.077703925060594e-05, "loss": 0.3556, "step": 1970 }, { "epoch": 1.2941176470588236, "grad_norm": 3.4385366439819336, "learning_rate": 7.043057966391157e-05, "loss": 0.3269, "step": 1980 }, { "epoch": 1.3006535947712419, "grad_norm": 3.6838247776031494, "learning_rate": 7.008293848410396e-05, "loss": 0.3342, "step": 1990 }, { "epoch": 1.3071895424836601, "grad_norm": 1.6612685918807983, "learning_rate": 6.973413581684972e-05, "loss": 0.3009, "step": 2000 }, { "epoch": 1.3071895424836601, "eval_loss": 0.33713921904563904, "eval_runtime": 7466.3969, "eval_samples_per_second": 0.182, "eval_steps_per_second": 0.182, "step": 2000 } ], "logging_steps": 10, "max_steps": 4590, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "total_flos": 6.542090621219321e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }