{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0323434527256183, "eval_steps": 780, "global_step": 7800, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 220.54913330078125, "learning_rate": 1.282051282051282e-07, "loss": 2.8332, "step": 10 }, { "epoch": 0.0, "grad_norm": 136.59881591796875, "learning_rate": 2.564102564102564e-07, "loss": 2.6208, "step": 20 }, { "epoch": 0.0, "grad_norm": 40.77531814575195, "learning_rate": 3.846153846153847e-07, "loss": 2.0014, "step": 30 }, { "epoch": 0.01, "grad_norm": 16.872848510742188, "learning_rate": 5.128205128205128e-07, "loss": 1.4716, "step": 40 }, { "epoch": 0.01, "grad_norm": 15.107361793518066, "learning_rate": 6.41025641025641e-07, "loss": 1.1896, "step": 50 }, { "epoch": 0.01, "grad_norm": 43.72957229614258, "learning_rate": 7.692307692307694e-07, "loss": 1.0214, "step": 60 }, { "epoch": 0.01, "grad_norm": 22.193782806396484, "learning_rate": 8.974358974358975e-07, "loss": 0.9242, "step": 70 }, { "epoch": 0.01, "grad_norm": 39.77405548095703, "learning_rate": 1.0256410256410257e-06, "loss": 0.8996, "step": 80 }, { "epoch": 0.01, "grad_norm": 11.561063766479492, "learning_rate": 1.153846153846154e-06, "loss": 0.874, "step": 90 }, { "epoch": 0.01, "grad_norm": 15.518475532531738, "learning_rate": 1.282051282051282e-06, "loss": 0.8376, "step": 100 }, { "epoch": 0.01, "grad_norm": 7.103811740875244, "learning_rate": 1.4102564102564104e-06, "loss": 0.8272, "step": 110 }, { "epoch": 0.02, "grad_norm": 9.517041206359863, "learning_rate": 1.5384615384615387e-06, "loss": 0.8133, "step": 120 }, { "epoch": 0.02, "grad_norm": 8.397263526916504, "learning_rate": 1.6666666666666667e-06, "loss": 0.7843, "step": 130 }, { "epoch": 0.02, "grad_norm": 8.737699508666992, "learning_rate": 1.794871794871795e-06, "loss": 0.7893, "step": 140 }, { "epoch": 0.02, "grad_norm": 8.244195938110352, "learning_rate": 1.9230769230769234e-06, "loss": 0.775, "step": 150 }, { "epoch": 0.02, "grad_norm": 7.860988140106201, "learning_rate": 2.0512820512820513e-06, "loss": 0.7527, "step": 160 }, { "epoch": 0.02, "grad_norm": 9.824834823608398, "learning_rate": 2.1794871794871797e-06, "loss": 0.739, "step": 170 }, { "epoch": 0.02, "grad_norm": 8.463064193725586, "learning_rate": 2.307692307692308e-06, "loss": 0.7432, "step": 180 }, { "epoch": 0.03, "grad_norm": 7.5682692527771, "learning_rate": 2.435897435897436e-06, "loss": 0.7301, "step": 190 }, { "epoch": 0.03, "grad_norm": 6.610253810882568, "learning_rate": 2.564102564102564e-06, "loss": 0.7238, "step": 200 }, { "epoch": 0.03, "grad_norm": 8.918222427368164, "learning_rate": 2.6923076923076923e-06, "loss": 0.7205, "step": 210 }, { "epoch": 0.03, "grad_norm": 8.915346145629883, "learning_rate": 2.8205128205128207e-06, "loss": 0.7171, "step": 220 }, { "epoch": 0.03, "grad_norm": 7.223239421844482, "learning_rate": 2.948717948717949e-06, "loss": 0.7259, "step": 230 }, { "epoch": 0.03, "grad_norm": 8.313645362854004, "learning_rate": 3.0769230769230774e-06, "loss": 0.7207, "step": 240 }, { "epoch": 0.03, "grad_norm": 8.011472702026367, "learning_rate": 3.205128205128206e-06, "loss": 0.7195, "step": 250 }, { "epoch": 0.03, "grad_norm": 163.78948974609375, "learning_rate": 3.3333333333333333e-06, "loss": 0.8906, "step": 260 }, { "epoch": 0.04, "grad_norm": 7.948330402374268, "learning_rate": 3.4615384615384617e-06, "loss": 0.9354, "step": 270 }, { "epoch": 0.04, "grad_norm": 8.010615348815918, "learning_rate": 3.58974358974359e-06, "loss": 0.7462, "step": 280 }, { "epoch": 0.04, "grad_norm": 7.337215900421143, "learning_rate": 3.7179487179487184e-06, "loss": 0.7211, "step": 290 }, { "epoch": 0.04, "grad_norm": 7.3255696296691895, "learning_rate": 3.846153846153847e-06, "loss": 0.7085, "step": 300 }, { "epoch": 0.04, "grad_norm": 5.787943363189697, "learning_rate": 3.974358974358974e-06, "loss": 0.7035, "step": 310 }, { "epoch": 0.04, "grad_norm": 5.677683353424072, "learning_rate": 4.102564102564103e-06, "loss": 0.7026, "step": 320 }, { "epoch": 0.04, "grad_norm": 4.641010761260986, "learning_rate": 4.230769230769231e-06, "loss": 0.6932, "step": 330 }, { "epoch": 0.04, "grad_norm": 7.157845497131348, "learning_rate": 4.358974358974359e-06, "loss": 0.6837, "step": 340 }, { "epoch": 0.05, "grad_norm": 5.9800801277160645, "learning_rate": 4.487179487179488e-06, "loss": 0.6946, "step": 350 }, { "epoch": 0.05, "grad_norm": 6.784861087799072, "learning_rate": 4.615384615384616e-06, "loss": 0.7001, "step": 360 }, { "epoch": 0.05, "grad_norm": 7.393420696258545, "learning_rate": 4.743589743589744e-06, "loss": 0.6867, "step": 370 }, { "epoch": 0.05, "grad_norm": 7.812753677368164, "learning_rate": 4.871794871794872e-06, "loss": 0.685, "step": 380 }, { "epoch": 0.05, "grad_norm": 5.967596054077148, "learning_rate": 5e-06, "loss": 0.6793, "step": 390 }, { "epoch": 0.05, "grad_norm": 5.9974517822265625, "learning_rate": 5.128205128205128e-06, "loss": 0.6826, "step": 400 }, { "epoch": 0.05, "grad_norm": 5.928895950317383, "learning_rate": 5.256410256410257e-06, "loss": 0.6774, "step": 410 }, { "epoch": 0.06, "grad_norm": 7.543839931488037, "learning_rate": 5.384615384615385e-06, "loss": 0.6804, "step": 420 }, { "epoch": 0.06, "grad_norm": 6.843835830688477, "learning_rate": 5.512820512820514e-06, "loss": 0.6798, "step": 430 }, { "epoch": 0.06, "grad_norm": 5.476939678192139, "learning_rate": 5.641025641025641e-06, "loss": 0.68, "step": 440 }, { "epoch": 0.06, "grad_norm": 6.189926624298096, "learning_rate": 5.769230769230769e-06, "loss": 0.6836, "step": 450 }, { "epoch": 0.06, "grad_norm": 6.458014965057373, "learning_rate": 5.897435897435898e-06, "loss": 0.6773, "step": 460 }, { "epoch": 0.06, "grad_norm": 5.884819030761719, "learning_rate": 6.025641025641026e-06, "loss": 0.6694, "step": 470 }, { "epoch": 0.06, "grad_norm": 5.87222957611084, "learning_rate": 6.153846153846155e-06, "loss": 0.6632, "step": 480 }, { "epoch": 0.06, "grad_norm": 5.767669200897217, "learning_rate": 6.282051282051282e-06, "loss": 0.6831, "step": 490 }, { "epoch": 0.07, "grad_norm": 6.337037563323975, "learning_rate": 6.410256410256412e-06, "loss": 0.6703, "step": 500 }, { "epoch": 0.07, "grad_norm": 7.12393856048584, "learning_rate": 6.538461538461539e-06, "loss": 0.6655, "step": 510 }, { "epoch": 0.07, "grad_norm": 6.461047172546387, "learning_rate": 6.666666666666667e-06, "loss": 0.6691, "step": 520 }, { "epoch": 0.07, "grad_norm": 5.647514820098877, "learning_rate": 6.794871794871796e-06, "loss": 0.6766, "step": 530 }, { "epoch": 0.07, "grad_norm": 5.566062927246094, "learning_rate": 6.923076923076923e-06, "loss": 0.677, "step": 540 }, { "epoch": 0.07, "grad_norm": 5.5864129066467285, "learning_rate": 7.051282051282053e-06, "loss": 0.6494, "step": 550 }, { "epoch": 0.07, "grad_norm": 5.37799596786499, "learning_rate": 7.17948717948718e-06, "loss": 0.6745, "step": 560 }, { "epoch": 0.08, "grad_norm": 6.159377098083496, "learning_rate": 7.307692307692308e-06, "loss": 0.6645, "step": 570 }, { "epoch": 0.08, "grad_norm": 5.143188953399658, "learning_rate": 7.435897435897437e-06, "loss": 0.6643, "step": 580 }, { "epoch": 0.08, "grad_norm": 6.492277145385742, "learning_rate": 7.564102564102564e-06, "loss": 0.6646, "step": 590 }, { "epoch": 0.08, "grad_norm": 5.831822872161865, "learning_rate": 7.692307692307694e-06, "loss": 0.6717, "step": 600 }, { "epoch": 0.08, "grad_norm": 6.297182083129883, "learning_rate": 7.820512820512822e-06, "loss": 0.6806, "step": 610 }, { "epoch": 0.08, "grad_norm": 6.297981262207031, "learning_rate": 7.948717948717949e-06, "loss": 0.6578, "step": 620 }, { "epoch": 0.08, "grad_norm": 5.828994274139404, "learning_rate": 8.076923076923077e-06, "loss": 0.659, "step": 630 }, { "epoch": 0.08, "grad_norm": 5.524103164672852, "learning_rate": 8.205128205128205e-06, "loss": 0.6615, "step": 640 }, { "epoch": 0.09, "grad_norm": 5.9705729484558105, "learning_rate": 8.333333333333334e-06, "loss": 0.6473, "step": 650 }, { "epoch": 0.09, "grad_norm": 6.803117275238037, "learning_rate": 8.461538461538462e-06, "loss": 0.6658, "step": 660 }, { "epoch": 0.09, "grad_norm": 6.929677486419678, "learning_rate": 8.58974358974359e-06, "loss": 0.6685, "step": 670 }, { "epoch": 0.09, "grad_norm": 7.229376792907715, "learning_rate": 8.717948717948719e-06, "loss": 0.6586, "step": 680 }, { "epoch": 0.09, "grad_norm": 6.211330413818359, "learning_rate": 8.846153846153847e-06, "loss": 0.6705, "step": 690 }, { "epoch": 0.09, "grad_norm": 8.68479061126709, "learning_rate": 8.974358974358976e-06, "loss": 0.6634, "step": 700 }, { "epoch": 0.09, "grad_norm": 6.314587593078613, "learning_rate": 9.102564102564104e-06, "loss": 0.6595, "step": 710 }, { "epoch": 0.1, "grad_norm": 24.571941375732422, "learning_rate": 9.230769230769232e-06, "loss": 0.7864, "step": 720 }, { "epoch": 0.1, "grad_norm": 82.88648986816406, "learning_rate": 9.358974358974359e-06, "loss": 0.7688, "step": 730 }, { "epoch": 0.1, "grad_norm": 16.867046356201172, "learning_rate": 9.487179487179487e-06, "loss": 3.5579, "step": 740 }, { "epoch": 0.1, "grad_norm": 46.3172721862793, "learning_rate": 9.615384615384616e-06, "loss": 1.5314, "step": 750 }, { "epoch": 0.1, "grad_norm": 4635.1376953125, "learning_rate": 9.743589743589744e-06, "loss": 1.3744, "step": 760 }, { "epoch": 0.1, "grad_norm": 8.688736915588379, "learning_rate": 9.871794871794872e-06, "loss": 0.9862, "step": 770 }, { "epoch": 0.1, "grad_norm": 7.020205497741699, "learning_rate": 1e-05, "loss": 0.6837, "step": 780 }, { "epoch": 0.1, "eval_loss": 0.6935335397720337, "eval_runtime": 169.5292, "eval_samples_per_second": 64.886, "eval_steps_per_second": 8.111, "step": 780 }, { "epoch": 0.1, "grad_norm": 9.475116729736328, "learning_rate": 9.99994993147413e-06, "loss": 0.6691, "step": 790 }, { "epoch": 0.11, "grad_norm": 6.155986785888672, "learning_rate": 9.999799726899261e-06, "loss": 0.6729, "step": 800 }, { "epoch": 0.11, "grad_norm": 6.841563701629639, "learning_rate": 9.999549389283605e-06, "loss": 0.6576, "step": 810 }, { "epoch": 0.11, "grad_norm": 6.626774311065674, "learning_rate": 9.999198923640774e-06, "loss": 0.6493, "step": 820 }, { "epoch": 0.11, "grad_norm": 7.392406940460205, "learning_rate": 9.998748336989687e-06, "loss": 0.6409, "step": 830 }, { "epoch": 0.11, "grad_norm": 7.925800800323486, "learning_rate": 9.998197638354428e-06, "loss": 0.6716, "step": 840 }, { "epoch": 0.11, "grad_norm": 6.4082231521606445, "learning_rate": 9.997546838764066e-06, "loss": 0.6415, "step": 850 }, { "epoch": 0.11, "grad_norm": 5.968785762786865, "learning_rate": 9.996795951252427e-06, "loss": 0.6516, "step": 860 }, { "epoch": 0.12, "grad_norm": 6.396018981933594, "learning_rate": 9.995944990857848e-06, "loss": 0.6449, "step": 870 }, { "epoch": 0.12, "grad_norm": 6.221564292907715, "learning_rate": 9.994993974622863e-06, "loss": 0.6453, "step": 880 }, { "epoch": 0.12, "grad_norm": 6.979955673217773, "learning_rate": 9.993942921593858e-06, "loss": 0.6659, "step": 890 }, { "epoch": 0.12, "grad_norm": 5.904052734375, "learning_rate": 9.992791852820709e-06, "loss": 0.66, "step": 900 }, { "epoch": 0.12, "grad_norm": 17.975086212158203, "learning_rate": 9.991540791356342e-06, "loss": 0.6866, "step": 910 }, { "epoch": 0.12, "grad_norm": 6.759973049163818, "learning_rate": 9.990189762256275e-06, "loss": 0.7491, "step": 920 }, { "epoch": 0.12, "grad_norm": 13.584983825683594, "learning_rate": 9.988738792578126e-06, "loss": 0.6689, "step": 930 }, { "epoch": 0.12, "grad_norm": 6.420701503753662, "learning_rate": 9.987187911381059e-06, "loss": 0.6539, "step": 940 }, { "epoch": 0.13, "grad_norm": 6.14314603805542, "learning_rate": 9.985537149725207e-06, "loss": 0.6546, "step": 950 }, { "epoch": 0.13, "grad_norm": 6.1515679359436035, "learning_rate": 9.983786540671052e-06, "loss": 0.6539, "step": 960 }, { "epoch": 0.13, "grad_norm": 6.025599956512451, "learning_rate": 9.981936119278758e-06, "loss": 0.6487, "step": 970 }, { "epoch": 0.13, "grad_norm": 7.110809803009033, "learning_rate": 9.979985922607476e-06, "loss": 0.6494, "step": 980 }, { "epoch": 0.13, "grad_norm": 6.404927730560303, "learning_rate": 9.977935989714594e-06, "loss": 0.6504, "step": 990 }, { "epoch": 0.13, "grad_norm": 5.934031963348389, "learning_rate": 9.975786361654959e-06, "loss": 0.637, "step": 1000 }, { "epoch": 0.13, "grad_norm": 5.632777690887451, "learning_rate": 9.973537081480056e-06, "loss": 0.635, "step": 1010 }, { "epoch": 0.13, "grad_norm": 5.7026777267456055, "learning_rate": 9.971188194237141e-06, "loss": 0.6388, "step": 1020 }, { "epoch": 0.14, "grad_norm": 4.8541388511657715, "learning_rate": 9.96873974696834e-06, "loss": 0.6225, "step": 1030 }, { "epoch": 0.14, "grad_norm": 6.6458210945129395, "learning_rate": 9.966191788709716e-06, "loss": 0.6373, "step": 1040 }, { "epoch": 0.14, "grad_norm": 17.634096145629883, "learning_rate": 9.96354437049027e-06, "loss": 0.645, "step": 1050 }, { "epoch": 0.14, "grad_norm": 9.264060974121094, "learning_rate": 9.960797545330936e-06, "loss": 0.6341, "step": 1060 }, { "epoch": 0.14, "grad_norm": 6.132219314575195, "learning_rate": 9.95795136824351e-06, "loss": 0.6302, "step": 1070 }, { "epoch": 0.14, "grad_norm": 5.637242794036865, "learning_rate": 9.955005896229543e-06, "loss": 0.6363, "step": 1080 }, { "epoch": 0.14, "grad_norm": 5.741671085357666, "learning_rate": 9.951961188279216e-06, "loss": 0.6415, "step": 1090 }, { "epoch": 0.15, "grad_norm": 6.340778827667236, "learning_rate": 9.948817305370145e-06, "loss": 0.6324, "step": 1100 }, { "epoch": 0.15, "grad_norm": 6.51924467086792, "learning_rate": 9.94557431046616e-06, "loss": 0.6301, "step": 1110 }, { "epoch": 0.15, "grad_norm": 82.26183319091797, "learning_rate": 9.942232268516051e-06, "loss": 0.6402, "step": 1120 }, { "epoch": 0.15, "grad_norm": 7.635751247406006, "learning_rate": 9.938791246452267e-06, "loss": 0.6295, "step": 1130 }, { "epoch": 0.15, "grad_norm": 6.102968215942383, "learning_rate": 9.935251313189564e-06, "loss": 0.6322, "step": 1140 }, { "epoch": 0.15, "grad_norm": 5.711689472198486, "learning_rate": 9.931612539623643e-06, "loss": 0.6433, "step": 1150 }, { "epoch": 0.15, "grad_norm": 7.467568874359131, "learning_rate": 9.927874998629714e-06, "loss": 0.6415, "step": 1160 }, { "epoch": 0.15, "grad_norm": 5.802877426147461, "learning_rate": 9.924038765061042e-06, "loss": 0.6469, "step": 1170 }, { "epoch": 0.16, "grad_norm": 5.440192699432373, "learning_rate": 9.920103915747452e-06, "loss": 0.6352, "step": 1180 }, { "epoch": 0.16, "grad_norm": 5.665472984313965, "learning_rate": 9.916070529493785e-06, "loss": 0.628, "step": 1190 }, { "epoch": 0.16, "grad_norm": 5.986988067626953, "learning_rate": 9.911938687078324e-06, "loss": 0.6292, "step": 1200 }, { "epoch": 0.16, "grad_norm": 5.383033752441406, "learning_rate": 9.907708471251173e-06, "loss": 0.6214, "step": 1210 }, { "epoch": 0.16, "grad_norm": 49.42042541503906, "learning_rate": 9.9033799667326e-06, "loss": 0.6292, "step": 1220 }, { "epoch": 0.16, "grad_norm": 5.570804595947266, "learning_rate": 9.89895326021134e-06, "loss": 0.6371, "step": 1230 }, { "epoch": 0.16, "grad_norm": 6.779324531555176, "learning_rate": 9.89442844034286e-06, "loss": 0.6396, "step": 1240 }, { "epoch": 0.17, "grad_norm": 6.269560813903809, "learning_rate": 9.889805597747588e-06, "loss": 0.6405, "step": 1250 }, { "epoch": 0.17, "grad_norm": 5.74788236618042, "learning_rate": 9.885084825009085e-06, "loss": 0.6302, "step": 1260 }, { "epoch": 0.17, "grad_norm": 6.128660202026367, "learning_rate": 9.88026621667221e-06, "loss": 0.6307, "step": 1270 }, { "epoch": 0.17, "grad_norm": 6.70562744140625, "learning_rate": 9.875349869241202e-06, "loss": 0.6275, "step": 1280 }, { "epoch": 0.17, "grad_norm": 6.813882827758789, "learning_rate": 9.870335881177774e-06, "loss": 0.6238, "step": 1290 }, { "epoch": 0.17, "grad_norm": 6.28445291519165, "learning_rate": 9.86522435289912e-06, "loss": 0.6433, "step": 1300 }, { "epoch": 0.17, "grad_norm": 6.097434043884277, "learning_rate": 9.860015386775915e-06, "loss": 0.6237, "step": 1310 }, { "epoch": 0.17, "grad_norm": 5.640145778656006, "learning_rate": 9.854709087130261e-06, "loss": 0.6294, "step": 1320 }, { "epoch": 0.18, "grad_norm": 5.536041259765625, "learning_rate": 9.849305560233598e-06, "loss": 0.6262, "step": 1330 }, { "epoch": 0.18, "grad_norm": 13.29723834991455, "learning_rate": 9.843804914304578e-06, "loss": 0.6426, "step": 1340 }, { "epoch": 0.18, "grad_norm": 5.7919230461120605, "learning_rate": 9.838207259506891e-06, "loss": 0.6376, "step": 1350 }, { "epoch": 0.18, "grad_norm": 6.46877908706665, "learning_rate": 9.83251270794707e-06, "loss": 0.6111, "step": 1360 }, { "epoch": 0.18, "grad_norm": 6.7596516609191895, "learning_rate": 9.826721373672235e-06, "loss": 0.6405, "step": 1370 }, { "epoch": 0.18, "grad_norm": 5.5958428382873535, "learning_rate": 9.820833372667813e-06, "loss": 0.6311, "step": 1380 }, { "epoch": 0.18, "grad_norm": 6.962647438049316, "learning_rate": 9.814848822855216e-06, "loss": 0.6336, "step": 1390 }, { "epoch": 0.19, "grad_norm": 6.83704948425293, "learning_rate": 9.80876784408948e-06, "loss": 0.6271, "step": 1400 }, { "epoch": 0.19, "grad_norm": 5.530167579650879, "learning_rate": 9.802590558156863e-06, "loss": 0.6261, "step": 1410 }, { "epoch": 0.19, "grad_norm": 5.446717739105225, "learning_rate": 9.796317088772402e-06, "loss": 0.618, "step": 1420 }, { "epoch": 0.19, "grad_norm": 5.7587080001831055, "learning_rate": 9.789947561577445e-06, "loss": 0.6295, "step": 1430 }, { "epoch": 0.19, "grad_norm": 6.391341686248779, "learning_rate": 9.783482104137127e-06, "loss": 0.6232, "step": 1440 }, { "epoch": 0.19, "grad_norm": 5.596224308013916, "learning_rate": 9.776920845937816e-06, "loss": 0.6175, "step": 1450 }, { "epoch": 0.19, "grad_norm": 5.266319751739502, "learning_rate": 9.770263918384523e-06, "loss": 0.6237, "step": 1460 }, { "epoch": 0.19, "grad_norm": 5.594700813293457, "learning_rate": 9.763511454798268e-06, "loss": 0.6164, "step": 1470 }, { "epoch": 0.2, "grad_norm": 6.352781772613525, "learning_rate": 9.75666359041341e-06, "loss": 0.6521, "step": 1480 }, { "epoch": 0.2, "grad_norm": 5.978342533111572, "learning_rate": 9.749720462374939e-06, "loss": 0.6247, "step": 1490 }, { "epoch": 0.2, "grad_norm": 6.177164554595947, "learning_rate": 9.742682209735727e-06, "loss": 0.6387, "step": 1500 }, { "epoch": 0.2, "grad_norm": 5.946715354919434, "learning_rate": 9.735548973453753e-06, "loss": 0.6445, "step": 1510 }, { "epoch": 0.2, "grad_norm": 6.191423416137695, "learning_rate": 9.728320896389263e-06, "loss": 0.6353, "step": 1520 }, { "epoch": 0.2, "grad_norm": 5.473761558532715, "learning_rate": 9.720998123301924e-06, "loss": 0.6223, "step": 1530 }, { "epoch": 0.2, "grad_norm": 6.137279033660889, "learning_rate": 9.713580800847917e-06, "loss": 0.6345, "step": 1540 }, { "epoch": 0.21, "grad_norm": 6.810376167297363, "learning_rate": 9.706069077577e-06, "loss": 0.6096, "step": 1550 }, { "epoch": 0.21, "grad_norm": 5.427652835845947, "learning_rate": 9.698463103929542e-06, "loss": 0.6133, "step": 1560 }, { "epoch": 0.21, "eval_loss": 0.6688335537910461, "eval_runtime": 169.0314, "eval_samples_per_second": 65.077, "eval_steps_per_second": 8.135, "step": 1560 }, { "epoch": 0.21, "grad_norm": 5.631854057312012, "learning_rate": 9.690763032233498e-06, "loss": 0.629, "step": 1570 }, { "epoch": 0.21, "grad_norm": 6.538657188415527, "learning_rate": 9.682969016701357e-06, "loss": 0.6359, "step": 1580 }, { "epoch": 0.21, "grad_norm": 10.069457054138184, "learning_rate": 9.675081213427076e-06, "loss": 0.6302, "step": 1590 }, { "epoch": 0.21, "grad_norm": 12.996840476989746, "learning_rate": 9.66709978038292e-06, "loss": 0.6616, "step": 1600 }, { "epoch": 0.21, "grad_norm": 5.648403644561768, "learning_rate": 9.659024877416328e-06, "loss": 0.6615, "step": 1610 }, { "epoch": 0.21, "grad_norm": 5.897425174713135, "learning_rate": 9.650856666246693e-06, "loss": 0.6362, "step": 1620 }, { "epoch": 0.22, "grad_norm": 5.605139255523682, "learning_rate": 9.642595310462133e-06, "loss": 0.6448, "step": 1630 }, { "epoch": 0.22, "grad_norm": 5.556352138519287, "learning_rate": 9.63424097551621e-06, "loss": 0.6246, "step": 1640 }, { "epoch": 0.22, "grad_norm": 6.256978511810303, "learning_rate": 9.62579382872462e-06, "loss": 0.6282, "step": 1650 }, { "epoch": 0.22, "grad_norm": 5.5795793533325195, "learning_rate": 9.617254039261835e-06, "loss": 0.6331, "step": 1660 }, { "epoch": 0.22, "grad_norm": 5.844163417816162, "learning_rate": 9.608621778157722e-06, "loss": 0.624, "step": 1670 }, { "epoch": 0.22, "grad_norm": 6.517988204956055, "learning_rate": 9.599897218294122e-06, "loss": 0.6327, "step": 1680 }, { "epoch": 0.22, "grad_norm": 5.43684720993042, "learning_rate": 9.591080534401371e-06, "loss": 0.6252, "step": 1690 }, { "epoch": 0.22, "grad_norm": 5.179018020629883, "learning_rate": 9.582171903054815e-06, "loss": 0.6149, "step": 1700 }, { "epoch": 0.23, "grad_norm": 5.8681793212890625, "learning_rate": 9.573171502671273e-06, "loss": 0.6292, "step": 1710 }, { "epoch": 0.23, "grad_norm": 5.400745391845703, "learning_rate": 9.564079513505455e-06, "loss": 0.6207, "step": 1720 }, { "epoch": 0.23, "grad_norm": 5.303137302398682, "learning_rate": 9.554896117646357e-06, "loss": 0.6162, "step": 1730 }, { "epoch": 0.23, "grad_norm": 5.112619876861572, "learning_rate": 9.54562149901362e-06, "loss": 0.6301, "step": 1740 }, { "epoch": 0.23, "grad_norm": 5.014323711395264, "learning_rate": 9.536255843353832e-06, "loss": 0.6144, "step": 1750 }, { "epoch": 0.23, "grad_norm": 5.302499294281006, "learning_rate": 9.526799338236828e-06, "loss": 0.637, "step": 1760 }, { "epoch": 0.23, "grad_norm": 5.237521171569824, "learning_rate": 9.517252173051912e-06, "loss": 0.6092, "step": 1770 }, { "epoch": 0.24, "grad_norm": 5.327495098114014, "learning_rate": 9.507614539004082e-06, "loss": 0.6364, "step": 1780 }, { "epoch": 0.24, "grad_norm": 5.794159412384033, "learning_rate": 9.497886629110187e-06, "loss": 0.6382, "step": 1790 }, { "epoch": 0.24, "grad_norm": 6.391251564025879, "learning_rate": 9.488068638195072e-06, "loss": 0.62, "step": 1800 }, { "epoch": 0.24, "grad_norm": 5.799778461456299, "learning_rate": 9.478160762887668e-06, "loss": 0.6196, "step": 1810 }, { "epoch": 0.24, "grad_norm": 5.956018924713135, "learning_rate": 9.468163201617063e-06, "loss": 0.6286, "step": 1820 }, { "epoch": 0.24, "grad_norm": 6.173859596252441, "learning_rate": 9.458076154608515e-06, "loss": 0.6208, "step": 1830 }, { "epoch": 0.24, "grad_norm": 6.374012470245361, "learning_rate": 9.447899823879456e-06, "loss": 0.625, "step": 1840 }, { "epoch": 0.24, "grad_norm": 5.741364479064941, "learning_rate": 9.437634413235437e-06, "loss": 0.6869, "step": 1850 }, { "epoch": 0.25, "grad_norm": 5.758192539215088, "learning_rate": 9.427280128266049e-06, "loss": 0.6102, "step": 1860 }, { "epoch": 0.25, "grad_norm": 5.713798522949219, "learning_rate": 9.416837176340807e-06, "loss": 0.6215, "step": 1870 }, { "epoch": 0.25, "grad_norm": 5.849081516265869, "learning_rate": 9.406305766604996e-06, "loss": 0.6158, "step": 1880 }, { "epoch": 0.25, "grad_norm": 6.048788547515869, "learning_rate": 9.395686109975475e-06, "loss": 0.6282, "step": 1890 }, { "epoch": 0.25, "grad_norm": 9.530996322631836, "learning_rate": 9.384978419136469e-06, "loss": 0.6135, "step": 1900 }, { "epoch": 0.25, "grad_norm": 6.369543075561523, "learning_rate": 9.374182908535293e-06, "loss": 0.6104, "step": 1910 }, { "epoch": 0.25, "grad_norm": 6.199366569519043, "learning_rate": 9.363299794378072e-06, "loss": 0.6199, "step": 1920 }, { "epoch": 0.26, "grad_norm": 5.299448013305664, "learning_rate": 9.352329294625397e-06, "loss": 0.6286, "step": 1930 }, { "epoch": 0.26, "grad_norm": 5.6097917556762695, "learning_rate": 9.34127162898797e-06, "loss": 0.6196, "step": 1940 }, { "epoch": 0.26, "grad_norm": 5.62219762802124, "learning_rate": 9.330127018922195e-06, "loss": 0.6244, "step": 1950 }, { "epoch": 0.26, "grad_norm": 5.205328464508057, "learning_rate": 9.318895687625752e-06, "loss": 0.6089, "step": 1960 }, { "epoch": 0.26, "grad_norm": 5.849984169006348, "learning_rate": 9.307577860033123e-06, "loss": 0.6213, "step": 1970 }, { "epoch": 0.26, "grad_norm": 6.027492523193359, "learning_rate": 9.296173762811084e-06, "loss": 0.6192, "step": 1980 }, { "epoch": 0.26, "grad_norm": 7.360931873321533, "learning_rate": 9.284683624354172e-06, "loss": 0.6313, "step": 1990 }, { "epoch": 0.26, "grad_norm": 5.995561599731445, "learning_rate": 9.273107674780102e-06, "loss": 0.6273, "step": 2000 }, { "epoch": 0.27, "grad_norm": 6.6998701095581055, "learning_rate": 9.261446145925167e-06, "loss": 0.624, "step": 2010 }, { "epoch": 0.27, "grad_norm": 7.960203647613525, "learning_rate": 9.249699271339594e-06, "loss": 0.6119, "step": 2020 }, { "epoch": 0.27, "grad_norm": 5.833003997802734, "learning_rate": 9.237867286282855e-06, "loss": 0.6615, "step": 2030 }, { "epoch": 0.27, "grad_norm": 5.727327346801758, "learning_rate": 9.225950427718974e-06, "loss": 0.6363, "step": 2040 }, { "epoch": 0.27, "grad_norm": 6.32903528213501, "learning_rate": 9.213948934311767e-06, "loss": 0.6382, "step": 2050 }, { "epoch": 0.27, "grad_norm": 6.892453670501709, "learning_rate": 9.201863046420065e-06, "loss": 0.6196, "step": 2060 }, { "epoch": 0.27, "grad_norm": 5.348752498626709, "learning_rate": 9.189693006092907e-06, "loss": 0.6104, "step": 2070 }, { "epoch": 0.28, "grad_norm": 5.757750034332275, "learning_rate": 9.177439057064684e-06, "loss": 0.6153, "step": 2080 }, { "epoch": 0.28, "grad_norm": 22.043140411376953, "learning_rate": 9.165101444750259e-06, "loss": 0.6178, "step": 2090 }, { "epoch": 0.28, "grad_norm": 5.820219993591309, "learning_rate": 9.152680416240059e-06, "loss": 2.4575, "step": 2100 }, { "epoch": 0.28, "grad_norm": 6.317713737487793, "learning_rate": 9.140176220295118e-06, "loss": 1.9869, "step": 2110 }, { "epoch": 0.28, "grad_norm": 5.32202672958374, "learning_rate": 9.1275891073421e-06, "loss": 0.6366, "step": 2120 }, { "epoch": 0.28, "grad_norm": 6.378438949584961, "learning_rate": 9.114919329468283e-06, "loss": 0.6138, "step": 2130 }, { "epoch": 0.28, "grad_norm": 6.984612941741943, "learning_rate": 9.102167140416503e-06, "loss": 0.6239, "step": 2140 }, { "epoch": 0.28, "grad_norm": 9.644604682922363, "learning_rate": 9.089332795580085e-06, "loss": 0.6094, "step": 2150 }, { "epoch": 0.29, "grad_norm": 6.99331521987915, "learning_rate": 9.076416551997721e-06, "loss": 0.6358, "step": 2160 }, { "epoch": 0.29, "grad_norm": 6.56105899810791, "learning_rate": 9.063418668348323e-06, "loss": 0.6118, "step": 2170 }, { "epoch": 0.29, "grad_norm": 5.86183500289917, "learning_rate": 9.050339404945834e-06, "loss": 0.6402, "step": 2180 }, { "epoch": 0.29, "grad_norm": 7.894756317138672, "learning_rate": 9.037179023734036e-06, "loss": 0.6323, "step": 2190 }, { "epoch": 0.29, "grad_norm": 6.089447975158691, "learning_rate": 9.023937788281278e-06, "loss": 0.6279, "step": 2200 }, { "epoch": 0.29, "grad_norm": 5.405038833618164, "learning_rate": 9.01061596377522e-06, "loss": 0.6088, "step": 2210 }, { "epoch": 0.29, "grad_norm": 6.3252949714660645, "learning_rate": 8.997213817017508e-06, "loss": 0.618, "step": 2220 }, { "epoch": 0.3, "grad_norm": 6.129386901855469, "learning_rate": 8.98373161641843e-06, "loss": 0.6124, "step": 2230 }, { "epoch": 0.3, "grad_norm": 7.108587265014648, "learning_rate": 8.970169631991556e-06, "loss": 0.627, "step": 2240 }, { "epoch": 0.3, "grad_norm": 5.341206073760986, "learning_rate": 8.95652813534831e-06, "loss": 0.6217, "step": 2250 }, { "epoch": 0.3, "grad_norm": 7.05128812789917, "learning_rate": 8.942807399692543e-06, "loss": 0.6335, "step": 2260 }, { "epoch": 0.3, "grad_norm": 5.670904636383057, "learning_rate": 8.92900769981506e-06, "loss": 0.6246, "step": 2270 }, { "epoch": 0.3, "grad_norm": 6.543381690979004, "learning_rate": 8.915129312088112e-06, "loss": 0.6193, "step": 2280 }, { "epoch": 0.3, "grad_norm": 5.513144016265869, "learning_rate": 8.901172514459864e-06, "loss": 0.5986, "step": 2290 }, { "epoch": 0.3, "grad_norm": 6.7181854248046875, "learning_rate": 8.88713758644883e-06, "loss": 0.5983, "step": 2300 }, { "epoch": 0.31, "grad_norm": 6.5360002517700195, "learning_rate": 8.873024809138272e-06, "loss": 0.6202, "step": 2310 }, { "epoch": 0.31, "grad_norm": 6.717379093170166, "learning_rate": 8.858834465170576e-06, "loss": 0.6232, "step": 2320 }, { "epoch": 0.31, "grad_norm": 8.741230964660645, "learning_rate": 8.84456683874158e-06, "loss": 0.6154, "step": 2330 }, { "epoch": 0.31, "grad_norm": 6.7567338943481445, "learning_rate": 8.83022221559489e-06, "loss": 0.607, "step": 2340 }, { "epoch": 0.31, "eval_loss": 0.6728433966636658, "eval_runtime": 169.7753, "eval_samples_per_second": 64.792, "eval_steps_per_second": 8.099, "step": 2340 }, { "epoch": 0.31, "grad_norm": 5.661694526672363, "learning_rate": 8.815800883016168e-06, "loss": 0.6176, "step": 2350 }, { "epoch": 0.31, "grad_norm": 6.330115795135498, "learning_rate": 8.801303129827352e-06, "loss": 0.6089, "step": 2360 }, { "epoch": 0.31, "grad_norm": 5.605167388916016, "learning_rate": 8.786729246380901e-06, "loss": 0.6127, "step": 2370 }, { "epoch": 0.31, "grad_norm": 5.510767459869385, "learning_rate": 8.772079524553951e-06, "loss": 0.6199, "step": 2380 }, { "epoch": 0.32, "grad_norm": 5.994042873382568, "learning_rate": 8.757354257742501e-06, "loss": 0.6245, "step": 2390 }, { "epoch": 0.32, "grad_norm": 6.51716947555542, "learning_rate": 8.742553740855507e-06, "loss": 0.6305, "step": 2400 }, { "epoch": 0.32, "grad_norm": 6.130660533905029, "learning_rate": 8.727678270308994e-06, "loss": 0.616, "step": 2410 }, { "epoch": 0.32, "grad_norm": 6.110134601593018, "learning_rate": 8.712728144020118e-06, "loss": 0.6277, "step": 2420 }, { "epoch": 0.32, "grad_norm": 6.650820255279541, "learning_rate": 8.697703661401187e-06, "loss": 0.6219, "step": 2430 }, { "epoch": 0.32, "grad_norm": 6.013575553894043, "learning_rate": 8.682605123353685e-06, "loss": 0.6268, "step": 2440 }, { "epoch": 0.32, "grad_norm": 7.728272914886475, "learning_rate": 8.66743283226223e-06, "loss": 0.6186, "step": 2450 }, { "epoch": 0.33, "grad_norm": 6.504591464996338, "learning_rate": 8.652187091988516e-06, "loss": 0.6007, "step": 2460 }, { "epoch": 0.33, "grad_norm": 6.214788436889648, "learning_rate": 8.636868207865244e-06, "loss": 0.6032, "step": 2470 }, { "epoch": 0.33, "grad_norm": 6.597141265869141, "learning_rate": 8.621476486689991e-06, "loss": 0.6086, "step": 2480 }, { "epoch": 0.33, "grad_norm": 7.287806034088135, "learning_rate": 8.606012236719073e-06, "loss": 0.6171, "step": 2490 }, { "epoch": 0.33, "grad_norm": 6.4462714195251465, "learning_rate": 8.590475767661371e-06, "loss": 0.6078, "step": 2500 }, { "epoch": 0.33, "grad_norm": 5.836272716522217, "learning_rate": 8.574867390672124e-06, "loss": 0.6125, "step": 2510 }, { "epoch": 0.33, "grad_norm": 8.247157096862793, "learning_rate": 8.559187418346703e-06, "loss": 0.6307, "step": 2520 }, { "epoch": 0.33, "grad_norm": 6.056578159332275, "learning_rate": 8.543436164714351e-06, "loss": 0.6274, "step": 2530 }, { "epoch": 0.34, "grad_norm": 5.924829483032227, "learning_rate": 8.527613945231886e-06, "loss": 0.6155, "step": 2540 }, { "epoch": 0.34, "grad_norm": 5.2663469314575195, "learning_rate": 8.511721076777388e-06, "loss": 0.6092, "step": 2550 }, { "epoch": 0.34, "grad_norm": 5.494378566741943, "learning_rate": 8.495757877643857e-06, "loss": 0.6178, "step": 2560 }, { "epoch": 0.34, "grad_norm": 6.344980239868164, "learning_rate": 8.479724667532836e-06, "loss": 0.6144, "step": 2570 }, { "epoch": 0.34, "grad_norm": 5.618322849273682, "learning_rate": 8.463621767547998e-06, "loss": 0.6056, "step": 2580 }, { "epoch": 0.34, "grad_norm": 5.677299499511719, "learning_rate": 8.447449500188731e-06, "loss": 0.6094, "step": 2590 }, { "epoch": 0.34, "grad_norm": 5.477785110473633, "learning_rate": 8.43120818934367e-06, "loss": 0.6279, "step": 2600 }, { "epoch": 0.35, "grad_norm": 5.992845058441162, "learning_rate": 8.414898160284208e-06, "loss": 0.6112, "step": 2610 }, { "epoch": 0.35, "grad_norm": 5.6498847007751465, "learning_rate": 8.398519739657997e-06, "loss": 0.6196, "step": 2620 }, { "epoch": 0.35, "grad_norm": 5.752150535583496, "learning_rate": 8.382073255482381e-06, "loss": 0.606, "step": 2630 }, { "epoch": 0.35, "grad_norm": 5.367984294891357, "learning_rate": 8.36555903713785e-06, "loss": 0.6123, "step": 2640 }, { "epoch": 0.35, "grad_norm": 5.714753150939941, "learning_rate": 8.348977415361435e-06, "loss": 0.6064, "step": 2650 }, { "epoch": 0.35, "grad_norm": 7.572054386138916, "learning_rate": 8.332328722240072e-06, "loss": 0.6109, "step": 2660 }, { "epoch": 0.35, "grad_norm": 5.684463024139404, "learning_rate": 8.315613291203977e-06, "loss": 0.6341, "step": 2670 }, { "epoch": 0.35, "grad_norm": 5.910693645477295, "learning_rate": 8.298831457019943e-06, "loss": 0.6156, "step": 2680 }, { "epoch": 0.36, "grad_norm": 5.144363880157471, "learning_rate": 8.28198355578465e-06, "loss": 0.6085, "step": 2690 }, { "epoch": 0.36, "grad_norm": 5.353940963745117, "learning_rate": 8.265069924917925e-06, "loss": 0.6048, "step": 2700 }, { "epoch": 0.36, "grad_norm": 5.588951587677002, "learning_rate": 8.248090903156003e-06, "loss": 0.6272, "step": 2710 }, { "epoch": 0.36, "grad_norm": 5.791905879974365, "learning_rate": 8.231046830544716e-06, "loss": 0.6183, "step": 2720 }, { "epoch": 0.36, "grad_norm": 6.0394744873046875, "learning_rate": 8.213938048432697e-06, "loss": 0.6069, "step": 2730 }, { "epoch": 0.36, "grad_norm": 8.518630027770996, "learning_rate": 8.196764899464552e-06, "loss": 0.6111, "step": 2740 }, { "epoch": 0.36, "grad_norm": 5.609602451324463, "learning_rate": 8.179527727573975e-06, "loss": 0.6239, "step": 2750 }, { "epoch": 0.37, "grad_norm": 10.055850982666016, "learning_rate": 8.162226877976886e-06, "loss": 0.6115, "step": 2760 }, { "epoch": 0.37, "grad_norm": 5.720135688781738, "learning_rate": 8.144862697164499e-06, "loss": 0.6025, "step": 2770 }, { "epoch": 0.37, "grad_norm": 7.587475299835205, "learning_rate": 8.127435532896388e-06, "loss": 0.6135, "step": 2780 }, { "epoch": 0.37, "grad_norm": 8.063986778259277, "learning_rate": 8.10994573419352e-06, "loss": 0.5999, "step": 2790 }, { "epoch": 0.37, "grad_norm": 7.293218612670898, "learning_rate": 8.092393651331275e-06, "loss": 0.6108, "step": 2800 }, { "epoch": 0.37, "grad_norm": 5.801172733306885, "learning_rate": 8.074779635832417e-06, "loss": 0.608, "step": 2810 }, { "epoch": 0.37, "grad_norm": 6.789213180541992, "learning_rate": 8.057104040460062e-06, "loss": 0.6105, "step": 2820 }, { "epoch": 0.37, "grad_norm": 5.721503734588623, "learning_rate": 8.03936721921061e-06, "loss": 0.6214, "step": 2830 }, { "epoch": 0.38, "grad_norm": 6.258848190307617, "learning_rate": 8.021569527306663e-06, "loss": 0.6174, "step": 2840 }, { "epoch": 0.38, "grad_norm": 5.338409900665283, "learning_rate": 8.003711321189895e-06, "loss": 0.6159, "step": 2850 }, { "epoch": 0.38, "grad_norm": 6.426341533660889, "learning_rate": 7.985792958513932e-06, "loss": 0.6152, "step": 2860 }, { "epoch": 0.38, "grad_norm": 7.031759262084961, "learning_rate": 7.967814798137173e-06, "loss": 0.6261, "step": 2870 }, { "epoch": 0.38, "grad_norm": 5.78716516494751, "learning_rate": 7.949777200115617e-06, "loss": 0.6173, "step": 2880 }, { "epoch": 0.38, "grad_norm": 20.726688385009766, "learning_rate": 7.931680525695634e-06, "loss": 0.6178, "step": 2890 }, { "epoch": 0.38, "grad_norm": 6.066954612731934, "learning_rate": 7.913525137306756e-06, "loss": 0.6176, "step": 2900 }, { "epoch": 0.39, "grad_norm": 5.461245536804199, "learning_rate": 7.895311398554395e-06, "loss": 0.6169, "step": 2910 }, { "epoch": 0.39, "grad_norm": 6.37878942489624, "learning_rate": 7.877039674212569e-06, "loss": 0.6105, "step": 2920 }, { "epoch": 0.39, "grad_norm": 47.42095947265625, "learning_rate": 7.8587103302166e-06, "loss": 0.617, "step": 2930 }, { "epoch": 0.39, "grad_norm": 6.993823528289795, "learning_rate": 7.84032373365578e-06, "loss": 0.6286, "step": 2940 }, { "epoch": 0.39, "grad_norm": 5.397260665893555, "learning_rate": 7.821880252766025e-06, "loss": 0.6031, "step": 2950 }, { "epoch": 0.39, "grad_norm": 6.4610915184021, "learning_rate": 7.803380256922495e-06, "loss": 0.6303, "step": 2960 }, { "epoch": 0.39, "grad_norm": 6.266001224517822, "learning_rate": 7.784824116632198e-06, "loss": 0.613, "step": 2970 }, { "epoch": 0.39, "grad_norm": 5.821497917175293, "learning_rate": 7.76621220352657e-06, "loss": 0.6156, "step": 2980 }, { "epoch": 0.4, "grad_norm": 6.601454257965088, "learning_rate": 7.747544890354031e-06, "loss": 0.6169, "step": 2990 }, { "epoch": 0.4, "grad_norm": 6.371436595916748, "learning_rate": 7.728822550972523e-06, "loss": 0.6099, "step": 3000 }, { "epoch": 0.4, "grad_norm": 5.537062168121338, "learning_rate": 7.710045560342021e-06, "loss": 0.6175, "step": 3010 }, { "epoch": 0.4, "grad_norm": 6.325994491577148, "learning_rate": 7.69121429451702e-06, "loss": 0.6005, "step": 3020 }, { "epoch": 0.4, "grad_norm": 5.711528778076172, "learning_rate": 7.672329130639007e-06, "loss": 0.612, "step": 3030 }, { "epoch": 0.4, "grad_norm": 6.956279754638672, "learning_rate": 7.65339044692891e-06, "loss": 0.6123, "step": 3040 }, { "epoch": 0.4, "grad_norm": 6.838433265686035, "learning_rate": 7.634398622679517e-06, "loss": 0.5872, "step": 3050 }, { "epoch": 0.4, "grad_norm": 10.948911666870117, "learning_rate": 7.615354038247889e-06, "loss": 0.6257, "step": 3060 }, { "epoch": 0.41, "grad_norm": 7.123111248016357, "learning_rate": 7.59625707504773e-06, "loss": 0.619, "step": 3070 }, { "epoch": 0.41, "grad_norm": 7.197707653045654, "learning_rate": 7.577108115541761e-06, "loss": 0.6077, "step": 3080 }, { "epoch": 0.41, "grad_norm": 5471.7568359375, "learning_rate": 7.557907543234051e-06, "loss": 0.6283, "step": 3090 }, { "epoch": 0.41, "grad_norm": 6.704062461853027, "learning_rate": 7.53865574266234e-06, "loss": 0.6213, "step": 3100 }, { "epoch": 0.41, "grad_norm": 5.542397499084473, "learning_rate": 7.519353099390336e-06, "loss": 0.6223, "step": 3110 }, { "epoch": 0.41, "grad_norm": 5.533992767333984, "learning_rate": 7.500000000000001e-06, "loss": 0.6051, "step": 3120 }, { "epoch": 0.41, "eval_loss": 0.6693721413612366, "eval_runtime": 169.48, "eval_samples_per_second": 64.904, "eval_steps_per_second": 8.113, "step": 3120 }, { "epoch": 0.41, "grad_norm": 5.968539714813232, "learning_rate": 7.480596832083795e-06, "loss": 0.6143, "step": 3130 }, { "epoch": 0.42, "grad_norm": 5.992849349975586, "learning_rate": 7.461143984236925e-06, "loss": 0.616, "step": 3140 }, { "epoch": 0.42, "grad_norm": 5.622570037841797, "learning_rate": 7.441641846049557e-06, "loss": 0.609, "step": 3150 }, { "epoch": 0.42, "grad_norm": 5.672114849090576, "learning_rate": 7.422090808099014e-06, "loss": 0.6175, "step": 3160 }, { "epoch": 0.42, "grad_norm": 5.839463710784912, "learning_rate": 7.402491261941958e-06, "loss": 0.6085, "step": 3170 }, { "epoch": 0.42, "grad_norm": 6.008167266845703, "learning_rate": 7.382843600106539e-06, "loss": 0.6069, "step": 3180 }, { "epoch": 0.42, "grad_norm": 6.840440273284912, "learning_rate": 7.363148216084548e-06, "loss": 0.6132, "step": 3190 }, { "epoch": 0.42, "grad_norm": 5.5630693435668945, "learning_rate": 7.343405504323519e-06, "loss": 0.6065, "step": 3200 }, { "epoch": 0.42, "grad_norm": 5.013092994689941, "learning_rate": 7.323615860218844e-06, "loss": 0.6123, "step": 3210 }, { "epoch": 0.43, "grad_norm": 48.6692008972168, "learning_rate": 7.303779680105844e-06, "loss": 0.6109, "step": 3220 }, { "epoch": 0.43, "grad_norm": 5.752770900726318, "learning_rate": 7.28389736125184e-06, "loss": 0.606, "step": 3230 }, { "epoch": 0.43, "grad_norm": 5.404486656188965, "learning_rate": 7.263969301848188e-06, "loss": 0.6109, "step": 3240 }, { "epoch": 0.43, "grad_norm": 5.718954563140869, "learning_rate": 7.243995901002312e-06, "loss": 0.6046, "step": 3250 }, { "epoch": 0.43, "grad_norm": 6.1195831298828125, "learning_rate": 7.223977558729707e-06, "loss": 0.6072, "step": 3260 }, { "epoch": 0.43, "grad_norm": 5.4051337242126465, "learning_rate": 7.203914675945929e-06, "loss": 0.6212, "step": 3270 }, { "epoch": 0.43, "grad_norm": 5.423203945159912, "learning_rate": 7.183807654458565e-06, "loss": 0.6204, "step": 3280 }, { "epoch": 0.44, "grad_norm": 5.746906757354736, "learning_rate": 7.163656896959181e-06, "loss": 0.6023, "step": 3290 }, { "epoch": 0.44, "grad_norm": 6.6502156257629395, "learning_rate": 7.143462807015271e-06, "loss": 0.6123, "step": 3300 }, { "epoch": 0.44, "grad_norm": 7.021644592285156, "learning_rate": 7.1232257890621605e-06, "loss": 0.6094, "step": 3310 }, { "epoch": 0.44, "grad_norm": 5.808534145355225, "learning_rate": 7.102946248394908e-06, "loss": 0.628, "step": 3320 }, { "epoch": 0.44, "grad_norm": 5.422878265380859, "learning_rate": 7.082624591160201e-06, "loss": 0.5978, "step": 3330 }, { "epoch": 0.44, "grad_norm": 5.411374568939209, "learning_rate": 7.0622612243482035e-06, "loss": 0.611, "step": 3340 }, { "epoch": 0.44, "grad_norm": 6.385402679443359, "learning_rate": 7.041856555784421e-06, "loss": 0.5982, "step": 3350 }, { "epoch": 0.44, "grad_norm": 4.959177494049072, "learning_rate": 7.021410994121525e-06, "loss": 0.5913, "step": 3360 }, { "epoch": 0.45, "grad_norm": 7.832031726837158, "learning_rate": 7.0009249488311685e-06, "loss": 0.615, "step": 3370 }, { "epoch": 0.45, "grad_norm": 7.2229485511779785, "learning_rate": 6.980398830195785e-06, "loss": 0.619, "step": 3380 }, { "epoch": 0.45, "grad_norm": 5.132372856140137, "learning_rate": 6.959833049300376e-06, "loss": 0.6236, "step": 3390 }, { "epoch": 0.45, "grad_norm": 5.803369045257568, "learning_rate": 6.939228018024275e-06, "loss": 0.6091, "step": 3400 }, { "epoch": 0.45, "grad_norm": 5.913552284240723, "learning_rate": 6.918584149032899e-06, "loss": 0.61, "step": 3410 }, { "epoch": 0.45, "grad_norm": 5.712104797363281, "learning_rate": 6.897901855769483e-06, "loss": 0.6139, "step": 3420 }, { "epoch": 0.45, "grad_norm": 6.360735893249512, "learning_rate": 6.8771815524468e-06, "loss": 0.606, "step": 3430 }, { "epoch": 0.46, "grad_norm": 7.527002334594727, "learning_rate": 6.856423654038868e-06, "loss": 0.5892, "step": 3440 }, { "epoch": 0.46, "grad_norm": 6.143304824829102, "learning_rate": 6.8356285762726385e-06, "loss": 0.6041, "step": 3450 }, { "epoch": 0.46, "grad_norm": 5.828638553619385, "learning_rate": 6.814796735619664e-06, "loss": 0.5932, "step": 3460 }, { "epoch": 0.46, "grad_norm": 5.889917373657227, "learning_rate": 6.793928549287767e-06, "loss": 0.6117, "step": 3470 }, { "epoch": 0.46, "grad_norm": 6.243914604187012, "learning_rate": 6.773024435212678e-06, "loss": 0.6034, "step": 3480 }, { "epoch": 0.46, "grad_norm": 6.9411420822143555, "learning_rate": 6.75208481204967e-06, "loss": 0.6003, "step": 3490 }, { "epoch": 0.46, "grad_norm": 5.287111282348633, "learning_rate": 6.731110099165165e-06, "loss": 0.6208, "step": 3500 }, { "epoch": 0.46, "grad_norm": 5.86116886138916, "learning_rate": 6.710100716628345e-06, "loss": 0.597, "step": 3510 }, { "epoch": 0.47, "grad_norm": 5.526810646057129, "learning_rate": 6.689057085202737e-06, "loss": 0.6052, "step": 3520 }, { "epoch": 0.47, "grad_norm": 6.855119228363037, "learning_rate": 6.667979626337783e-06, "loss": 0.6147, "step": 3530 }, { "epoch": 0.47, "grad_norm": 6.467052936553955, "learning_rate": 6.646868762160399e-06, "loss": 0.6221, "step": 3540 }, { "epoch": 0.47, "grad_norm": 5.706518650054932, "learning_rate": 6.625724915466526e-06, "loss": 0.6083, "step": 3550 }, { "epoch": 0.47, "grad_norm": 5.94938325881958, "learning_rate": 6.6045485097126585e-06, "loss": 0.6067, "step": 3560 }, { "epoch": 0.47, "grad_norm": 6.971024990081787, "learning_rate": 6.583339969007364e-06, "loss": 0.6172, "step": 3570 }, { "epoch": 0.47, "grad_norm": 6.673549652099609, "learning_rate": 6.562099718102788e-06, "loss": 0.5947, "step": 3580 }, { "epoch": 0.48, "grad_norm": 6.783125400543213, "learning_rate": 6.540828182386154e-06, "loss": 0.6101, "step": 3590 }, { "epoch": 0.48, "grad_norm": 6.100213527679443, "learning_rate": 6.519525787871235e-06, "loss": 0.6201, "step": 3600 }, { "epoch": 0.48, "grad_norm": 5.297542095184326, "learning_rate": 6.498192961189829e-06, "loss": 0.6125, "step": 3610 }, { "epoch": 0.48, "grad_norm": 7.120528697967529, "learning_rate": 6.476830129583207e-06, "loss": 0.6224, "step": 3620 }, { "epoch": 0.48, "grad_norm": 6.501312255859375, "learning_rate": 6.455437720893565e-06, "loss": 0.6167, "step": 3630 }, { "epoch": 0.48, "grad_norm": 5.54337739944458, "learning_rate": 6.434016163555452e-06, "loss": 0.6052, "step": 3640 }, { "epoch": 0.48, "grad_norm": 7.030393123626709, "learning_rate": 6.412565886587186e-06, "loss": 0.6056, "step": 3650 }, { "epoch": 0.48, "grad_norm": 5.686173915863037, "learning_rate": 6.391087319582264e-06, "loss": 0.6069, "step": 3660 }, { "epoch": 0.49, "grad_norm": 6.340301990509033, "learning_rate": 6.369580892700763e-06, "loss": 0.6117, "step": 3670 }, { "epoch": 0.49, "grad_norm": 7.223570823669434, "learning_rate": 6.34804703666072e-06, "loss": 0.6086, "step": 3680 }, { "epoch": 0.49, "grad_norm": 6.784290790557861, "learning_rate": 6.326486182729504e-06, "loss": 0.603, "step": 3690 }, { "epoch": 0.49, "grad_norm": 6.71796178817749, "learning_rate": 6.304898762715187e-06, "loss": 0.6162, "step": 3700 }, { "epoch": 0.49, "grad_norm": 5.618252277374268, "learning_rate": 6.283285208957884e-06, "loss": 0.6067, "step": 3710 }, { "epoch": 0.49, "grad_norm": 5.947731971740723, "learning_rate": 6.261645954321109e-06, "loss": 0.5991, "step": 3720 }, { "epoch": 0.49, "grad_norm": 5.969636917114258, "learning_rate": 6.239981432183093e-06, "loss": 0.5944, "step": 3730 }, { "epoch": 0.49, "grad_norm": 5.365227222442627, "learning_rate": 6.21829207642811e-06, "loss": 0.5966, "step": 3740 }, { "epoch": 0.5, "grad_norm": 6.247452735900879, "learning_rate": 6.1965783214377895e-06, "loss": 0.611, "step": 3750 }, { "epoch": 0.5, "grad_norm": 6.950301170349121, "learning_rate": 6.1748406020824115e-06, "loss": 0.6248, "step": 3760 }, { "epoch": 0.5, "grad_norm": 6.410281658172607, "learning_rate": 6.153079353712201e-06, "loss": 0.6153, "step": 3770 }, { "epoch": 0.5, "grad_norm": 5.590805530548096, "learning_rate": 6.131295012148613e-06, "loss": 0.5987, "step": 3780 }, { "epoch": 0.5, "grad_norm": 5.457326412200928, "learning_rate": 6.1094880136755886e-06, "loss": 0.6026, "step": 3790 }, { "epoch": 0.5, "grad_norm": 6.030434608459473, "learning_rate": 6.087658795030838e-06, "loss": 0.6042, "step": 3800 }, { "epoch": 0.5, "grad_norm": 6.162191390991211, "learning_rate": 6.065807793397081e-06, "loss": 0.6068, "step": 3810 }, { "epoch": 0.51, "grad_norm": 5.414748191833496, "learning_rate": 6.043935446393294e-06, "loss": 0.6108, "step": 3820 }, { "epoch": 0.51, "grad_norm": 6.97568941116333, "learning_rate": 6.022042192065946e-06, "loss": 0.6181, "step": 3830 }, { "epoch": 0.51, "grad_norm": 5.456526756286621, "learning_rate": 6.000128468880223e-06, "loss": 0.602, "step": 3840 }, { "epoch": 0.51, "grad_norm": 5.599926471710205, "learning_rate": 5.978194715711254e-06, "loss": 0.6007, "step": 3850 }, { "epoch": 0.51, "grad_norm": 6.381017684936523, "learning_rate": 5.956241371835312e-06, "loss": 0.6228, "step": 3860 }, { "epoch": 0.51, "grad_norm": 6.424213409423828, "learning_rate": 5.934268876921026e-06, "loss": 0.6139, "step": 3870 }, { "epoch": 0.51, "grad_norm": 6.539239406585693, "learning_rate": 5.912277671020564e-06, "loss": 0.6137, "step": 3880 }, { "epoch": 0.51, "grad_norm": 5.962428569793701, "learning_rate": 5.890268194560834e-06, "loss": 0.6146, "step": 3890 }, { "epoch": 0.52, "grad_norm": 7.252692222595215, "learning_rate": 5.8682408883346535e-06, "loss": 0.6106, "step": 3900 }, { "epoch": 0.52, "eval_loss": 0.665640652179718, "eval_runtime": 169.5554, "eval_samples_per_second": 64.876, "eval_steps_per_second": 8.109, "step": 3900 }, { "epoch": 0.52, "grad_norm": 5.630374431610107, "learning_rate": 5.846196193491919e-06, "loss": 0.6113, "step": 3910 }, { "epoch": 0.52, "grad_norm": 6.424310684204102, "learning_rate": 5.824134551530783e-06, "loss": 0.6183, "step": 3920 }, { "epoch": 0.52, "grad_norm": 6.524292469024658, "learning_rate": 5.8020564042888015e-06, "loss": 0.5956, "step": 3930 }, { "epoch": 0.52, "grad_norm": 6.760373592376709, "learning_rate": 5.77996219393409e-06, "loss": 0.6046, "step": 3940 }, { "epoch": 0.52, "grad_norm": 17.418819427490234, "learning_rate": 5.757852362956463e-06, "loss": 0.5965, "step": 3950 }, { "epoch": 0.52, "grad_norm": 6.237477779388428, "learning_rate": 5.735727354158581e-06, "loss": 0.608, "step": 3960 }, { "epoch": 0.53, "grad_norm": 6.2491536140441895, "learning_rate": 5.713587610647073e-06, "loss": 0.6216, "step": 3970 }, { "epoch": 0.53, "grad_norm": 5.632169723510742, "learning_rate": 5.6914335758236665e-06, "loss": 0.6086, "step": 3980 }, { "epoch": 0.53, "grad_norm": 10.672119140625, "learning_rate": 5.669265693376309e-06, "loss": 0.6306, "step": 3990 }, { "epoch": 0.53, "grad_norm": 7.666195869445801, "learning_rate": 5.647084407270277e-06, "loss": 0.6212, "step": 4000 }, { "epoch": 0.53, "grad_norm": 7.483991622924805, "learning_rate": 5.624890161739292e-06, "loss": 0.6169, "step": 4010 }, { "epoch": 0.53, "grad_norm": 6.328516960144043, "learning_rate": 5.6026834012766155e-06, "loss": 0.6156, "step": 4020 }, { "epoch": 0.53, "grad_norm": 6.012834072113037, "learning_rate": 5.5804645706261515e-06, "loss": 0.5989, "step": 4030 }, { "epoch": 0.53, "grad_norm": 6.906214714050293, "learning_rate": 5.5582341147735396e-06, "loss": 0.5973, "step": 4040 }, { "epoch": 0.54, "grad_norm": 5.286527156829834, "learning_rate": 5.53599247893724e-06, "loss": 0.6118, "step": 4050 }, { "epoch": 0.54, "grad_norm": 6.227565288543701, "learning_rate": 5.5137401085596224e-06, "loss": 0.6164, "step": 4060 }, { "epoch": 0.54, "grad_norm": 6.331086158752441, "learning_rate": 5.491477449298036e-06, "loss": 0.599, "step": 4070 }, { "epoch": 0.54, "grad_norm": 5.511313438415527, "learning_rate": 5.469204947015897e-06, "loss": 0.6144, "step": 4080 }, { "epoch": 0.54, "grad_norm": 6.1719136238098145, "learning_rate": 5.4469230477737466e-06, "loss": 0.6062, "step": 4090 }, { "epoch": 0.54, "grad_norm": 6.096289157867432, "learning_rate": 5.424632197820325e-06, "loss": 0.5993, "step": 4100 }, { "epoch": 0.54, "grad_norm": 6.381209850311279, "learning_rate": 5.402332843583631e-06, "loss": 0.5882, "step": 4110 }, { "epoch": 0.55, "grad_norm": 5.978997230529785, "learning_rate": 5.380025431661981e-06, "loss": 0.6173, "step": 4120 }, { "epoch": 0.55, "grad_norm": 7.058841705322266, "learning_rate": 5.3577104088150685e-06, "loss": 0.6216, "step": 4130 }, { "epoch": 0.55, "grad_norm": 29.12560272216797, "learning_rate": 5.335388221955012e-06, "loss": 0.6031, "step": 4140 }, { "epoch": 0.55, "grad_norm": 5.8936357498168945, "learning_rate": 5.31305931813741e-06, "loss": 0.609, "step": 4150 }, { "epoch": 0.55, "grad_norm": 5.908172130584717, "learning_rate": 5.290724144552379e-06, "loss": 0.5983, "step": 4160 }, { "epoch": 0.55, "grad_norm": 5.777021408081055, "learning_rate": 5.268383148515608e-06, "loss": 0.6044, "step": 4170 }, { "epoch": 0.55, "grad_norm": 6.296756744384766, "learning_rate": 5.246036777459391e-06, "loss": 0.5939, "step": 4180 }, { "epoch": 0.55, "grad_norm": 6.158642292022705, "learning_rate": 5.223685478923671e-06, "loss": 0.6138, "step": 4190 }, { "epoch": 0.56, "grad_norm": 6.421185493469238, "learning_rate": 5.201329700547077e-06, "loss": 0.6052, "step": 4200 }, { "epoch": 0.56, "grad_norm": 6.061400890350342, "learning_rate": 5.178969890057953e-06, "loss": 0.5955, "step": 4210 }, { "epoch": 0.56, "grad_norm": 6.597458362579346, "learning_rate": 5.156606495265402e-06, "loss": 0.602, "step": 4220 }, { "epoch": 0.56, "grad_norm": 6.535460948944092, "learning_rate": 5.1342399640503074e-06, "loss": 0.5859, "step": 4230 }, { "epoch": 0.56, "grad_norm": 7.226145267486572, "learning_rate": 5.111870744356366e-06, "loss": 0.623, "step": 4240 }, { "epoch": 0.56, "grad_norm": 5.355672836303711, "learning_rate": 5.0894992841811216e-06, "loss": 0.6009, "step": 4250 }, { "epoch": 0.56, "grad_norm": 9.383895874023438, "learning_rate": 5.067126031566988e-06, "loss": 0.6037, "step": 4260 }, { "epoch": 0.57, "grad_norm": 5.733826160430908, "learning_rate": 5.044751434592274e-06, "loss": 0.5935, "step": 4270 }, { "epoch": 0.57, "grad_norm": 6.465234279632568, "learning_rate": 5.022375941362218e-06, "loss": 0.6264, "step": 4280 }, { "epoch": 0.57, "grad_norm": 5.9884514808654785, "learning_rate": 5e-06, "loss": 0.605, "step": 4290 }, { "epoch": 0.57, "grad_norm": 5.8463029861450195, "learning_rate": 4.977624058637783e-06, "loss": 0.6032, "step": 4300 }, { "epoch": 0.57, "grad_norm": 6.030627250671387, "learning_rate": 4.955248565407727e-06, "loss": 0.5974, "step": 4310 }, { "epoch": 0.57, "grad_norm": 7.388688564300537, "learning_rate": 4.932873968433014e-06, "loss": 0.6036, "step": 4320 }, { "epoch": 0.57, "grad_norm": 11.483485221862793, "learning_rate": 4.910500715818879e-06, "loss": 0.5931, "step": 4330 }, { "epoch": 0.57, "grad_norm": 12.173521995544434, "learning_rate": 4.8881292556436355e-06, "loss": 0.6069, "step": 4340 }, { "epoch": 0.58, "grad_norm": 5.8268866539001465, "learning_rate": 4.865760035949695e-06, "loss": 0.6134, "step": 4350 }, { "epoch": 0.58, "grad_norm": 7.893507957458496, "learning_rate": 4.8433935047346e-06, "loss": 0.6227, "step": 4360 }, { "epoch": 0.58, "grad_norm": 5.854642391204834, "learning_rate": 4.821030109942048e-06, "loss": 0.5971, "step": 4370 }, { "epoch": 0.58, "grad_norm": 5.798580169677734, "learning_rate": 4.798670299452926e-06, "loss": 0.6025, "step": 4380 }, { "epoch": 0.58, "grad_norm": 5.5879011154174805, "learning_rate": 4.77631452107633e-06, "loss": 0.6108, "step": 4390 }, { "epoch": 0.58, "grad_norm": 5.21212911605835, "learning_rate": 4.75396322254061e-06, "loss": 0.6054, "step": 4400 }, { "epoch": 0.58, "grad_norm": 6.118653774261475, "learning_rate": 4.731616851484392e-06, "loss": 0.6178, "step": 4410 }, { "epoch": 0.58, "grad_norm": 6.629573345184326, "learning_rate": 4.7092758554476215e-06, "loss": 0.5872, "step": 4420 }, { "epoch": 0.59, "grad_norm": 6.795328617095947, "learning_rate": 4.686940681862591e-06, "loss": 0.6138, "step": 4430 }, { "epoch": 0.59, "grad_norm": 7.000875473022461, "learning_rate": 4.664611778044988e-06, "loss": 0.5995, "step": 4440 }, { "epoch": 0.59, "grad_norm": 6.718314170837402, "learning_rate": 4.642289591184934e-06, "loss": 0.6128, "step": 4450 }, { "epoch": 0.59, "grad_norm": 7.343639373779297, "learning_rate": 4.619974568338021e-06, "loss": 0.5903, "step": 4460 }, { "epoch": 0.59, "grad_norm": 6.8176069259643555, "learning_rate": 4.597667156416371e-06, "loss": 0.6094, "step": 4470 }, { "epoch": 0.59, "grad_norm": 6.549642562866211, "learning_rate": 4.575367802179675e-06, "loss": 0.6127, "step": 4480 }, { "epoch": 0.59, "grad_norm": 6.30284309387207, "learning_rate": 4.553076952226255e-06, "loss": 0.6077, "step": 4490 }, { "epoch": 0.6, "grad_norm": 6.47008752822876, "learning_rate": 4.530795052984104e-06, "loss": 0.6188, "step": 4500 }, { "epoch": 0.6, "grad_norm": 6.808812618255615, "learning_rate": 4.508522550701965e-06, "loss": 0.5933, "step": 4510 }, { "epoch": 0.6, "grad_norm": 8.047310829162598, "learning_rate": 4.48625989144038e-06, "loss": 0.615, "step": 4520 }, { "epoch": 0.6, "grad_norm": 5.543693542480469, "learning_rate": 4.464007521062761e-06, "loss": 0.6225, "step": 4530 }, { "epoch": 0.6, "grad_norm": 5.247865200042725, "learning_rate": 4.441765885226462e-06, "loss": 0.6078, "step": 4540 }, { "epoch": 0.6, "grad_norm": 14.305082321166992, "learning_rate": 4.4195354293738484e-06, "loss": 0.6163, "step": 4550 }, { "epoch": 0.6, "grad_norm": 8.802489280700684, "learning_rate": 4.397316598723385e-06, "loss": 0.5915, "step": 4560 }, { "epoch": 0.6, "grad_norm": 5.3068952560424805, "learning_rate": 4.3751098382607084e-06, "loss": 0.6135, "step": 4570 }, { "epoch": 0.61, "grad_norm": 6.380529403686523, "learning_rate": 4.352915592729723e-06, "loss": 0.6061, "step": 4580 }, { "epoch": 0.61, "grad_norm": 6.269029140472412, "learning_rate": 4.330734306623694e-06, "loss": 0.5981, "step": 4590 }, { "epoch": 0.61, "grad_norm": 6.364799499511719, "learning_rate": 4.308566424176336e-06, "loss": 0.5953, "step": 4600 }, { "epoch": 0.61, "grad_norm": 5.895086288452148, "learning_rate": 4.286412389352929e-06, "loss": 0.613, "step": 4610 }, { "epoch": 0.61, "grad_norm": 7.223175048828125, "learning_rate": 4.264272645841419e-06, "loss": 0.6082, "step": 4620 }, { "epoch": 0.61, "grad_norm": 5.315123558044434, "learning_rate": 4.242147637043539e-06, "loss": 0.5997, "step": 4630 }, { "epoch": 0.61, "grad_norm": 6.662210941314697, "learning_rate": 4.220037806065911e-06, "loss": 0.6159, "step": 4640 }, { "epoch": 0.62, "grad_norm": 8.199719429016113, "learning_rate": 4.1979435957111984e-06, "loss": 0.6022, "step": 4650 }, { "epoch": 0.62, "grad_norm": 7.3503499031066895, "learning_rate": 4.175865448469219e-06, "loss": 0.6111, "step": 4660 }, { "epoch": 0.62, "grad_norm": 5.498391151428223, "learning_rate": 4.153803806508083e-06, "loss": 0.5987, "step": 4670 }, { "epoch": 0.62, "grad_norm": 6.277636528015137, "learning_rate": 4.131759111665349e-06, "loss": 0.5946, "step": 4680 }, { "epoch": 0.62, "eval_loss": 0.6720156073570251, "eval_runtime": 169.1493, "eval_samples_per_second": 65.031, "eval_steps_per_second": 8.129, "step": 4680 }, { "epoch": 0.62, "grad_norm": 7.248624801635742, "learning_rate": 4.109731805439168e-06, "loss": 0.6196, "step": 4690 }, { "epoch": 0.62, "grad_norm": 7.02220344543457, "learning_rate": 4.087722328979437e-06, "loss": 0.6122, "step": 4700 }, { "epoch": 0.62, "grad_norm": 6.665443420410156, "learning_rate": 4.065731123078977e-06, "loss": 0.5931, "step": 4710 }, { "epoch": 0.62, "grad_norm": 7.180152416229248, "learning_rate": 4.043758628164688e-06, "loss": 0.6053, "step": 4720 }, { "epoch": 0.63, "grad_norm": 5.866796970367432, "learning_rate": 4.021805284288749e-06, "loss": 0.6113, "step": 4730 }, { "epoch": 0.63, "grad_norm": 6.837026596069336, "learning_rate": 3.999871531119779e-06, "loss": 0.5973, "step": 4740 }, { "epoch": 0.63, "grad_norm": 7.780344486236572, "learning_rate": 3.977957807934055e-06, "loss": 0.6065, "step": 4750 }, { "epoch": 0.63, "grad_norm": 7.0064568519592285, "learning_rate": 3.956064553606708e-06, "loss": 0.5954, "step": 4760 }, { "epoch": 0.63, "grad_norm": 5.387651443481445, "learning_rate": 3.934192206602921e-06, "loss": 0.6239, "step": 4770 }, { "epoch": 0.63, "grad_norm": 5.932905673980713, "learning_rate": 3.912341204969164e-06, "loss": 0.6012, "step": 4780 }, { "epoch": 0.63, "grad_norm": 6.533616065979004, "learning_rate": 3.890511986324413e-06, "loss": 0.6034, "step": 4790 }, { "epoch": 0.64, "grad_norm": 6.714369297027588, "learning_rate": 3.86870498785139e-06, "loss": 0.6067, "step": 4800 }, { "epoch": 0.64, "grad_norm": 6.622659206390381, "learning_rate": 3.8469206462878e-06, "loss": 0.6018, "step": 4810 }, { "epoch": 0.64, "grad_norm": 6.005035400390625, "learning_rate": 3.825159397917589e-06, "loss": 0.6064, "step": 4820 }, { "epoch": 0.64, "grad_norm": 5.99045991897583, "learning_rate": 3.803421678562213e-06, "loss": 0.5974, "step": 4830 }, { "epoch": 0.64, "grad_norm": 6.072996616363525, "learning_rate": 3.781707923571891e-06, "loss": 0.5925, "step": 4840 }, { "epoch": 0.64, "grad_norm": 5.99989652633667, "learning_rate": 3.7600185678169083e-06, "loss": 0.6085, "step": 4850 }, { "epoch": 0.64, "grad_norm": 7.072287082672119, "learning_rate": 3.7383540456788915e-06, "loss": 0.5999, "step": 4860 }, { "epoch": 0.64, "grad_norm": 28.138442993164062, "learning_rate": 3.7167147910421165e-06, "loss": 0.6033, "step": 4870 }, { "epoch": 0.65, "grad_norm": 6.020947456359863, "learning_rate": 3.695101237284815e-06, "loss": 0.6274, "step": 4880 }, { "epoch": 0.65, "grad_norm": 6.451099872589111, "learning_rate": 3.6735138172704967e-06, "loss": 0.6164, "step": 4890 }, { "epoch": 0.65, "grad_norm": 6.501095294952393, "learning_rate": 3.6519529633392825e-06, "loss": 0.6001, "step": 4900 }, { "epoch": 0.65, "grad_norm": 6.591226100921631, "learning_rate": 3.6304191072992376e-06, "loss": 0.6006, "step": 4910 }, { "epoch": 0.65, "grad_norm": 6.160434246063232, "learning_rate": 3.6089126804177373e-06, "loss": 0.6086, "step": 4920 }, { "epoch": 0.65, "grad_norm": 5.883294105529785, "learning_rate": 3.5874341134128156e-06, "loss": 0.5847, "step": 4930 }, { "epoch": 0.65, "grad_norm": 6.181455135345459, "learning_rate": 3.5659838364445505e-06, "loss": 0.6045, "step": 4940 }, { "epoch": 0.66, "grad_norm": 7.934025764465332, "learning_rate": 3.544562279106436e-06, "loss": 0.5986, "step": 4950 }, { "epoch": 0.66, "grad_norm": 5.755185127258301, "learning_rate": 3.523169870416795e-06, "loss": 0.5843, "step": 4960 }, { "epoch": 0.66, "grad_norm": 6.632129669189453, "learning_rate": 3.501807038810174e-06, "loss": 0.6017, "step": 4970 }, { "epoch": 0.66, "grad_norm": 6.323507785797119, "learning_rate": 3.480474212128766e-06, "loss": 0.6, "step": 4980 }, { "epoch": 0.66, "grad_norm": 6.898191928863525, "learning_rate": 3.459171817613847e-06, "loss": 0.6072, "step": 4990 }, { "epoch": 0.66, "grad_norm": 6.425208568572998, "learning_rate": 3.4379002818972122e-06, "loss": 0.6057, "step": 5000 }, { "epoch": 0.66, "grad_norm": 6.319818019866943, "learning_rate": 3.416660030992639e-06, "loss": 0.5943, "step": 5010 }, { "epoch": 0.66, "grad_norm": 6.483379364013672, "learning_rate": 3.3954514902873427e-06, "loss": 0.606, "step": 5020 }, { "epoch": 0.67, "grad_norm": 5.618423938751221, "learning_rate": 3.3742750845334748e-06, "loss": 0.5912, "step": 5030 }, { "epoch": 0.67, "grad_norm": 6.536773681640625, "learning_rate": 3.3531312378396026e-06, "loss": 0.5996, "step": 5040 }, { "epoch": 0.67, "grad_norm": 5.867932319641113, "learning_rate": 3.3320203736622185e-06, "loss": 0.6041, "step": 5050 }, { "epoch": 0.67, "grad_norm": 5.637106895446777, "learning_rate": 3.310942914797265e-06, "loss": 0.6047, "step": 5060 }, { "epoch": 0.67, "grad_norm": 5.646607875823975, "learning_rate": 3.289899283371657e-06, "loss": 0.5951, "step": 5070 }, { "epoch": 0.67, "grad_norm": 5.731027603149414, "learning_rate": 3.2688899008348386e-06, "loss": 0.583, "step": 5080 }, { "epoch": 0.67, "grad_norm": 6.196643352508545, "learning_rate": 3.2479151879503324e-06, "loss": 0.6097, "step": 5090 }, { "epoch": 0.67, "grad_norm": 5.765705108642578, "learning_rate": 3.226975564787322e-06, "loss": 0.6014, "step": 5100 }, { "epoch": 0.68, "grad_norm": 6.380090713500977, "learning_rate": 3.206071450712235e-06, "loss": 0.5932, "step": 5110 }, { "epoch": 0.68, "grad_norm": 5.803713798522949, "learning_rate": 3.1852032643803377e-06, "loss": 0.5956, "step": 5120 }, { "epoch": 0.68, "grad_norm": 5.434231758117676, "learning_rate": 3.1643714237273628e-06, "loss": 0.6111, "step": 5130 }, { "epoch": 0.68, "grad_norm": 6.616121292114258, "learning_rate": 3.143576345961132e-06, "loss": 0.5966, "step": 5140 }, { "epoch": 0.68, "grad_norm": 5.209160804748535, "learning_rate": 3.1228184475532015e-06, "loss": 0.5979, "step": 5150 }, { "epoch": 0.68, "grad_norm": 6.537261486053467, "learning_rate": 3.1020981442305187e-06, "loss": 0.6166, "step": 5160 }, { "epoch": 0.68, "grad_norm": 7.26732063293457, "learning_rate": 3.0814158509671015e-06, "loss": 0.6068, "step": 5170 }, { "epoch": 0.69, "grad_norm": 6.632195472717285, "learning_rate": 3.0607719819757264e-06, "loss": 0.598, "step": 5180 }, { "epoch": 0.69, "grad_norm": 6.387509346008301, "learning_rate": 3.040166950699626e-06, "loss": 0.5923, "step": 5190 }, { "epoch": 0.69, "grad_norm": 6.4542975425720215, "learning_rate": 3.019601169804216e-06, "loss": 0.5946, "step": 5200 }, { "epoch": 0.69, "grad_norm": 5.96135139465332, "learning_rate": 2.9990750511688323e-06, "loss": 0.604, "step": 5210 }, { "epoch": 0.69, "grad_norm": 5.886032581329346, "learning_rate": 2.978589005878476e-06, "loss": 0.6048, "step": 5220 }, { "epoch": 0.69, "grad_norm": 6.059714317321777, "learning_rate": 2.9581434442155798e-06, "loss": 0.6122, "step": 5230 }, { "epoch": 0.69, "grad_norm": 6.174469947814941, "learning_rate": 2.937738775651798e-06, "loss": 0.5996, "step": 5240 }, { "epoch": 0.69, "grad_norm": 6.120980262756348, "learning_rate": 2.917375408839803e-06, "loss": 0.6096, "step": 5250 }, { "epoch": 0.7, "grad_norm": 6.153501510620117, "learning_rate": 2.8970537516050935e-06, "loss": 0.6014, "step": 5260 }, { "epoch": 0.7, "grad_norm": 6.2347211837768555, "learning_rate": 2.876774210937843e-06, "loss": 0.5917, "step": 5270 }, { "epoch": 0.7, "grad_norm": 7.352357864379883, "learning_rate": 2.8565371929847286e-06, "loss": 0.6026, "step": 5280 }, { "epoch": 0.7, "grad_norm": 6.723049163818359, "learning_rate": 2.836343103040819e-06, "loss": 0.6073, "step": 5290 }, { "epoch": 0.7, "grad_norm": 6.070889472961426, "learning_rate": 2.816192345541437e-06, "loss": 0.6028, "step": 5300 }, { "epoch": 0.7, "grad_norm": 5.253773212432861, "learning_rate": 2.7960853240540703e-06, "loss": 0.6038, "step": 5310 }, { "epoch": 0.7, "grad_norm": 7.381737232208252, "learning_rate": 2.776022441270295e-06, "loss": 0.6045, "step": 5320 }, { "epoch": 0.71, "grad_norm": 6.627200126647949, "learning_rate": 2.7560040989976894e-06, "loss": 0.5919, "step": 5330 }, { "epoch": 0.71, "grad_norm": 6.173203945159912, "learning_rate": 2.736030698151815e-06, "loss": 0.582, "step": 5340 }, { "epoch": 0.71, "grad_norm": 7.563076972961426, "learning_rate": 2.7161026387481636e-06, "loss": 0.5951, "step": 5350 }, { "epoch": 0.71, "grad_norm": 7.46902322769165, "learning_rate": 2.6962203198941587e-06, "loss": 0.6025, "step": 5360 }, { "epoch": 0.71, "grad_norm": 5.640381336212158, "learning_rate": 2.6763841397811576e-06, "loss": 0.6019, "step": 5370 }, { "epoch": 0.71, "grad_norm": 7.339272975921631, "learning_rate": 2.656594495676482e-06, "loss": 0.5874, "step": 5380 }, { "epoch": 0.71, "grad_norm": 8.082894325256348, "learning_rate": 2.636851783915454e-06, "loss": 0.6085, "step": 5390 }, { "epoch": 0.71, "grad_norm": 6.488770008087158, "learning_rate": 2.6171563998934605e-06, "loss": 0.5936, "step": 5400 }, { "epoch": 0.72, "grad_norm": 7.645140647888184, "learning_rate": 2.5975087380580445e-06, "loss": 0.6101, "step": 5410 }, { "epoch": 0.72, "grad_norm": 6.867657661437988, "learning_rate": 2.577909191900988e-06, "loss": 0.619, "step": 5420 }, { "epoch": 0.72, "grad_norm": 6.3017258644104, "learning_rate": 2.5583581539504464e-06, "loss": 0.5912, "step": 5430 }, { "epoch": 0.72, "grad_norm": 7.013880252838135, "learning_rate": 2.5388560157630765e-06, "loss": 0.604, "step": 5440 }, { "epoch": 0.72, "grad_norm": 6.177152633666992, "learning_rate": 2.519403167916207e-06, "loss": 0.5953, "step": 5450 }, { "epoch": 0.72, "grad_norm": 9.21536922454834, "learning_rate": 2.5000000000000015e-06, "loss": 0.6057, "step": 5460 }, { "epoch": 0.72, "eval_loss": 0.6694371104240417, "eval_runtime": 169.1763, "eval_samples_per_second": 65.021, "eval_steps_per_second": 8.128, "step": 5460 }, { "epoch": 0.72, "grad_norm": 9.104439735412598, "learning_rate": 2.480646900609664e-06, "loss": 0.597, "step": 5470 }, { "epoch": 0.73, "grad_norm": 6.437440395355225, "learning_rate": 2.4613442573376625e-06, "loss": 0.5985, "step": 5480 }, { "epoch": 0.73, "grad_norm": 7.327053546905518, "learning_rate": 2.4420924567659508e-06, "loss": 0.6095, "step": 5490 }, { "epoch": 0.73, "grad_norm": 6.946102142333984, "learning_rate": 2.422891884458241e-06, "loss": 0.5944, "step": 5500 }, { "epoch": 0.73, "grad_norm": 7.699533462524414, "learning_rate": 2.4037429249522702e-06, "loss": 0.6156, "step": 5510 }, { "epoch": 0.73, "grad_norm": 6.800520896911621, "learning_rate": 2.384645961752113e-06, "loss": 0.6075, "step": 5520 }, { "epoch": 0.73, "grad_norm": 6.476116180419922, "learning_rate": 2.3656013773204843e-06, "loss": 0.5989, "step": 5530 }, { "epoch": 0.73, "grad_norm": 7.190838813781738, "learning_rate": 2.346609553071093e-06, "loss": 0.5993, "step": 5540 }, { "epoch": 0.73, "grad_norm": 6.840661525726318, "learning_rate": 2.3276708693609947e-06, "loss": 0.6068, "step": 5550 }, { "epoch": 0.74, "grad_norm": 6.123993396759033, "learning_rate": 2.308785705482982e-06, "loss": 0.6089, "step": 5560 }, { "epoch": 0.74, "grad_norm": 6.094503402709961, "learning_rate": 2.289954439657981e-06, "loss": 0.6055, "step": 5570 }, { "epoch": 0.74, "grad_norm": 7.03631067276001, "learning_rate": 2.2711774490274767e-06, "loss": 0.6056, "step": 5580 }, { "epoch": 0.74, "grad_norm": 7.555922508239746, "learning_rate": 2.2524551096459703e-06, "loss": 0.6018, "step": 5590 }, { "epoch": 0.74, "grad_norm": 10.68952465057373, "learning_rate": 2.2337877964734324e-06, "loss": 0.5999, "step": 5600 }, { "epoch": 0.74, "grad_norm": 10.688403129577637, "learning_rate": 2.2151758833678044e-06, "loss": 0.6061, "step": 5610 }, { "epoch": 0.74, "grad_norm": 8.527271270751953, "learning_rate": 2.1966197430775056e-06, "loss": 0.5914, "step": 5620 }, { "epoch": 0.75, "grad_norm": 8.061843872070312, "learning_rate": 2.178119747233976e-06, "loss": 0.5957, "step": 5630 }, { "epoch": 0.75, "grad_norm": 7.056491851806641, "learning_rate": 2.159676266344222e-06, "loss": 0.5993, "step": 5640 }, { "epoch": 0.75, "grad_norm": 6.01476526260376, "learning_rate": 2.141289669783401e-06, "loss": 0.5988, "step": 5650 }, { "epoch": 0.75, "grad_norm": 7.841024875640869, "learning_rate": 2.122960325787432e-06, "loss": 0.6008, "step": 5660 }, { "epoch": 0.75, "grad_norm": 6.125739574432373, "learning_rate": 2.104688601445606e-06, "loss": 0.5924, "step": 5670 }, { "epoch": 0.75, "grad_norm": 5.347300052642822, "learning_rate": 2.086474862693244e-06, "loss": 0.6024, "step": 5680 }, { "epoch": 0.75, "grad_norm": 8.490260124206543, "learning_rate": 2.068319474304365e-06, "loss": 0.5842, "step": 5690 }, { "epoch": 0.75, "grad_norm": 5.8698344230651855, "learning_rate": 2.050222799884387e-06, "loss": 0.6001, "step": 5700 }, { "epoch": 0.76, "grad_norm": 5.577418327331543, "learning_rate": 2.0321852018628278e-06, "loss": 0.5981, "step": 5710 }, { "epoch": 0.76, "grad_norm": 6.239425182342529, "learning_rate": 2.0142070414860704e-06, "loss": 0.5936, "step": 5720 }, { "epoch": 0.76, "grad_norm": 5.704605579376221, "learning_rate": 1.996288678810105e-06, "loss": 0.5973, "step": 5730 }, { "epoch": 0.76, "grad_norm": 6.664093017578125, "learning_rate": 1.9784304726933384e-06, "loss": 0.5979, "step": 5740 }, { "epoch": 0.76, "grad_norm": 5.986905574798584, "learning_rate": 1.9606327807893905e-06, "loss": 0.6006, "step": 5750 }, { "epoch": 0.76, "grad_norm": 6.272210121154785, "learning_rate": 1.942895959539939e-06, "loss": 0.5991, "step": 5760 }, { "epoch": 0.76, "grad_norm": 6.319902420043945, "learning_rate": 1.9252203641675854e-06, "loss": 0.5921, "step": 5770 }, { "epoch": 0.76, "grad_norm": 6.060082912445068, "learning_rate": 1.9076063486687256e-06, "loss": 0.5872, "step": 5780 }, { "epoch": 0.77, "grad_norm": 6.724019527435303, "learning_rate": 1.8900542658064807e-06, "loss": 0.5845, "step": 5790 }, { "epoch": 0.77, "grad_norm": 5.889808654785156, "learning_rate": 1.8725644671036125e-06, "loss": 0.6051, "step": 5800 }, { "epoch": 0.77, "grad_norm": 8.664203643798828, "learning_rate": 1.8551373028355013e-06, "loss": 0.609, "step": 5810 }, { "epoch": 0.77, "grad_norm": 6.822065830230713, "learning_rate": 1.8377731220231144e-06, "loss": 0.6033, "step": 5820 }, { "epoch": 0.77, "grad_norm": 6.636890888214111, "learning_rate": 1.8204722724260266e-06, "loss": 0.6068, "step": 5830 }, { "epoch": 0.77, "grad_norm": 7.8858256340026855, "learning_rate": 1.803235100535452e-06, "loss": 0.5931, "step": 5840 }, { "epoch": 0.77, "grad_norm": 7.670845985412598, "learning_rate": 1.7860619515673034e-06, "loss": 0.6111, "step": 5850 }, { "epoch": 0.78, "grad_norm": 6.426079750061035, "learning_rate": 1.7689531694552863e-06, "loss": 0.5929, "step": 5860 }, { "epoch": 0.78, "grad_norm": 5.926182270050049, "learning_rate": 1.7519090968439966e-06, "loss": 0.587, "step": 5870 }, { "epoch": 0.78, "grad_norm": 5.617947101593018, "learning_rate": 1.7349300750820758e-06, "loss": 0.5958, "step": 5880 }, { "epoch": 0.78, "grad_norm": 5.876589775085449, "learning_rate": 1.7180164442153529e-06, "loss": 0.6037, "step": 5890 }, { "epoch": 0.78, "grad_norm": 6.038319110870361, "learning_rate": 1.7011685429800596e-06, "loss": 0.6023, "step": 5900 }, { "epoch": 0.78, "grad_norm": 7.0782318115234375, "learning_rate": 1.6843867087960252e-06, "loss": 0.6068, "step": 5910 }, { "epoch": 0.78, "grad_norm": 5.8368144035339355, "learning_rate": 1.6676712777599275e-06, "loss": 0.5919, "step": 5920 }, { "epoch": 0.78, "grad_norm": 6.66851806640625, "learning_rate": 1.6510225846385668e-06, "loss": 0.5912, "step": 5930 }, { "epoch": 0.79, "grad_norm": 6.366353988647461, "learning_rate": 1.6344409628621482e-06, "loss": 0.6087, "step": 5940 }, { "epoch": 0.79, "grad_norm": 7.111102104187012, "learning_rate": 1.6179267445176206e-06, "loss": 0.6059, "step": 5950 }, { "epoch": 0.79, "grad_norm": 6.061978816986084, "learning_rate": 1.6014802603420044e-06, "loss": 0.5938, "step": 5960 }, { "epoch": 0.79, "grad_norm": 6.1793341636657715, "learning_rate": 1.5851018397157918e-06, "loss": 0.6087, "step": 5970 }, { "epoch": 0.79, "grad_norm": 6.220329761505127, "learning_rate": 1.5687918106563326e-06, "loss": 0.5961, "step": 5980 }, { "epoch": 0.79, "grad_norm": 6.557837009429932, "learning_rate": 1.5525504998112717e-06, "loss": 0.5922, "step": 5990 }, { "epoch": 0.79, "grad_norm": 6.448672771453857, "learning_rate": 1.5363782324520033e-06, "loss": 0.5969, "step": 6000 }, { "epoch": 0.8, "grad_norm": 6.809442043304443, "learning_rate": 1.520275332467166e-06, "loss": 0.6038, "step": 6010 }, { "epoch": 0.8, "grad_norm": 5.7299041748046875, "learning_rate": 1.504242122356143e-06, "loss": 0.5929, "step": 6020 }, { "epoch": 0.8, "grad_norm": 5.356698989868164, "learning_rate": 1.4882789232226124e-06, "loss": 0.5916, "step": 6030 }, { "epoch": 0.8, "grad_norm": 6.809737682342529, "learning_rate": 1.4723860547681163e-06, "loss": 0.6108, "step": 6040 }, { "epoch": 0.8, "grad_norm": 6.65049409866333, "learning_rate": 1.4565638352856504e-06, "loss": 0.5958, "step": 6050 }, { "epoch": 0.8, "grad_norm": 6.739075660705566, "learning_rate": 1.4408125816532981e-06, "loss": 0.5805, "step": 6060 }, { "epoch": 0.8, "grad_norm": 6.686561584472656, "learning_rate": 1.4251326093278773e-06, "loss": 0.6039, "step": 6070 }, { "epoch": 0.8, "grad_norm": 6.082182884216309, "learning_rate": 1.4095242323386305e-06, "loss": 0.5949, "step": 6080 }, { "epoch": 0.81, "grad_norm": 5.692592620849609, "learning_rate": 1.3939877632809279e-06, "loss": 0.5944, "step": 6090 }, { "epoch": 0.81, "grad_norm": 5.9873480796813965, "learning_rate": 1.3785235133100088e-06, "loss": 0.5853, "step": 6100 }, { "epoch": 0.81, "grad_norm": 34.229915618896484, "learning_rate": 1.3631317921347564e-06, "loss": 0.5978, "step": 6110 }, { "epoch": 0.81, "grad_norm": 7.1477952003479, "learning_rate": 1.347812908011485e-06, "loss": 0.5803, "step": 6120 }, { "epoch": 0.81, "grad_norm": 6.827530384063721, "learning_rate": 1.3325671677377727e-06, "loss": 0.5955, "step": 6130 }, { "epoch": 0.81, "grad_norm": 6.848330974578857, "learning_rate": 1.3173948766463146e-06, "loss": 0.5997, "step": 6140 }, { "epoch": 0.81, "grad_norm": 6.4514288902282715, "learning_rate": 1.3022963385988153e-06, "loss": 0.59, "step": 6150 }, { "epoch": 0.82, "grad_norm": 6.169240951538086, "learning_rate": 1.2872718559798852e-06, "loss": 0.602, "step": 6160 }, { "epoch": 0.82, "grad_norm": 6.365018367767334, "learning_rate": 1.2723217296910078e-06, "loss": 0.6034, "step": 6170 }, { "epoch": 0.82, "grad_norm": 6.3110785484313965, "learning_rate": 1.257446259144494e-06, "loss": 0.5973, "step": 6180 }, { "epoch": 0.82, "grad_norm": 6.271914958953857, "learning_rate": 1.2426457422575e-06, "loss": 0.5988, "step": 6190 }, { "epoch": 0.82, "grad_norm": 6.755878448486328, "learning_rate": 1.2279204754460494e-06, "loss": 0.5896, "step": 6200 }, { "epoch": 0.82, "grad_norm": 5.7621684074401855, "learning_rate": 1.2132707536191008e-06, "loss": 0.5961, "step": 6210 }, { "epoch": 0.82, "grad_norm": 5.900276184082031, "learning_rate": 1.1986968701726492e-06, "loss": 0.5944, "step": 6220 }, { "epoch": 0.82, "grad_norm": 6.725846290588379, "learning_rate": 1.184199116983834e-06, "loss": 0.5897, "step": 6230 }, { "epoch": 0.83, "grad_norm": 8.138350486755371, "learning_rate": 1.1697777844051105e-06, "loss": 0.6036, "step": 6240 }, { "epoch": 0.83, "eval_loss": 0.6633522510528564, "eval_runtime": 169.2613, "eval_samples_per_second": 64.988, "eval_steps_per_second": 8.124, "step": 6240 }, { "epoch": 0.83, "grad_norm": 6.216264247894287, "learning_rate": 1.1554331612584218e-06, "loss": 0.5926, "step": 6250 }, { "epoch": 0.83, "grad_norm": 6.344448566436768, "learning_rate": 1.141165534829425e-06, "loss": 0.5855, "step": 6260 }, { "epoch": 0.83, "grad_norm": 7.457535743713379, "learning_rate": 1.1269751908617277e-06, "loss": 0.5912, "step": 6270 }, { "epoch": 0.83, "grad_norm": 6.311199188232422, "learning_rate": 1.1128624135511712e-06, "loss": 0.5983, "step": 6280 }, { "epoch": 0.83, "grad_norm": 6.246806621551514, "learning_rate": 1.0988274855401377e-06, "loss": 0.5875, "step": 6290 }, { "epoch": 0.83, "grad_norm": 6.010527610778809, "learning_rate": 1.0848706879118893e-06, "loss": 0.5905, "step": 6300 }, { "epoch": 0.84, "grad_norm": 5.997354984283447, "learning_rate": 1.070992300184941e-06, "loss": 0.608, "step": 6310 }, { "epoch": 0.84, "grad_norm": 6.129363536834717, "learning_rate": 1.057192600307456e-06, "loss": 0.6013, "step": 6320 }, { "epoch": 0.84, "grad_norm": 6.758306503295898, "learning_rate": 1.0434718646516917e-06, "loss": 0.5961, "step": 6330 }, { "epoch": 0.84, "grad_norm": 6.165789604187012, "learning_rate": 1.0298303680084448e-06, "loss": 0.5952, "step": 6340 }, { "epoch": 0.84, "grad_norm": 6.356081008911133, "learning_rate": 1.0162683835815706e-06, "loss": 0.595, "step": 6350 }, { "epoch": 0.84, "grad_norm": 5.993287563323975, "learning_rate": 1.0027861829824953e-06, "loss": 0.6059, "step": 6360 }, { "epoch": 0.84, "grad_norm": 6.2351226806640625, "learning_rate": 9.893840362247809e-07, "loss": 0.5966, "step": 6370 }, { "epoch": 0.84, "grad_norm": 6.520230293273926, "learning_rate": 9.760622117187234e-07, "loss": 0.6169, "step": 6380 }, { "epoch": 0.85, "grad_norm": 6.322525978088379, "learning_rate": 9.628209762659658e-07, "loss": 0.6066, "step": 6390 }, { "epoch": 0.85, "grad_norm": 8.321556091308594, "learning_rate": 9.496605950541676e-07, "loss": 0.5976, "step": 6400 }, { "epoch": 0.85, "grad_norm": 7.012268543243408, "learning_rate": 9.365813316516787e-07, "loss": 0.6028, "step": 6410 }, { "epoch": 0.85, "grad_norm": 7.8040771484375, "learning_rate": 9.235834480022788e-07, "loss": 0.611, "step": 6420 }, { "epoch": 0.85, "grad_norm": 5.389317035675049, "learning_rate": 9.10667204419915e-07, "loss": 0.5917, "step": 6430 }, { "epoch": 0.85, "grad_norm": 7.818176746368408, "learning_rate": 8.978328595834984e-07, "loss": 0.6056, "step": 6440 }, { "epoch": 0.85, "grad_norm": 6.365832328796387, "learning_rate": 8.850806705317183e-07, "loss": 0.5949, "step": 6450 }, { "epoch": 0.85, "grad_norm": 6.176076412200928, "learning_rate": 8.724108926579e-07, "loss": 0.6034, "step": 6460 }, { "epoch": 0.86, "grad_norm": 6.508719444274902, "learning_rate": 8.598237797048825e-07, "loss": 0.5851, "step": 6470 }, { "epoch": 0.86, "grad_norm": 6.904651165008545, "learning_rate": 8.473195837599419e-07, "loss": 0.5949, "step": 6480 }, { "epoch": 0.86, "grad_norm": 6.214872360229492, "learning_rate": 8.348985552497424e-07, "loss": 0.5826, "step": 6490 }, { "epoch": 0.86, "grad_norm": 9.990653991699219, "learning_rate": 8.225609429353187e-07, "loss": 0.6072, "step": 6500 }, { "epoch": 0.86, "grad_norm": 7.119464874267578, "learning_rate": 8.103069939070945e-07, "loss": 0.6097, "step": 6510 }, { "epoch": 0.86, "grad_norm": 7.176183700561523, "learning_rate": 7.981369535799354e-07, "loss": 0.5989, "step": 6520 }, { "epoch": 0.86, "grad_norm": 6.2535014152526855, "learning_rate": 7.860510656882342e-07, "loss": 0.602, "step": 6530 }, { "epoch": 0.87, "grad_norm": 6.955148696899414, "learning_rate": 7.740495722810271e-07, "loss": 0.6027, "step": 6540 }, { "epoch": 0.87, "grad_norm": 6.898377895355225, "learning_rate": 7.621327137171447e-07, "loss": 0.5999, "step": 6550 }, { "epoch": 0.87, "grad_norm": 6.001263618469238, "learning_rate": 7.50300728660407e-07, "loss": 0.5719, "step": 6560 }, { "epoch": 0.87, "grad_norm": 6.911612510681152, "learning_rate": 7.385538540748327e-07, "loss": 0.5802, "step": 6570 }, { "epoch": 0.87, "grad_norm": 5.093260765075684, "learning_rate": 7.26892325219899e-07, "loss": 0.5966, "step": 6580 }, { "epoch": 0.87, "grad_norm": 6.112546920776367, "learning_rate": 7.153163756458287e-07, "loss": 0.5949, "step": 6590 }, { "epoch": 0.87, "grad_norm": 6.923007011413574, "learning_rate": 7.03826237188916e-07, "loss": 0.5967, "step": 6600 }, { "epoch": 0.87, "grad_norm": 8.302557945251465, "learning_rate": 6.924221399668785e-07, "loss": 0.5879, "step": 6610 }, { "epoch": 0.88, "grad_norm": 6.832093238830566, "learning_rate": 6.811043123742494e-07, "loss": 0.596, "step": 6620 }, { "epoch": 0.88, "grad_norm": 6.365548133850098, "learning_rate": 6.698729810778065e-07, "loss": 0.5975, "step": 6630 }, { "epoch": 0.88, "grad_norm": 6.902804851531982, "learning_rate": 6.587283710120324e-07, "loss": 0.5882, "step": 6640 }, { "epoch": 0.88, "grad_norm": 7.0722808837890625, "learning_rate": 6.47670705374604e-07, "loss": 0.5945, "step": 6650 }, { "epoch": 0.88, "grad_norm": 9.358778953552246, "learning_rate": 6.367002056219285e-07, "loss": 0.5891, "step": 6660 }, { "epoch": 0.88, "grad_norm": 7.326229572296143, "learning_rate": 6.258170914647077e-07, "loss": 0.6016, "step": 6670 }, { "epoch": 0.88, "grad_norm": 5.735571384429932, "learning_rate": 6.150215808635334e-07, "loss": 0.6066, "step": 6680 }, { "epoch": 0.89, "grad_norm": 5.679074764251709, "learning_rate": 6.043138900245277e-07, "loss": 0.5944, "step": 6690 }, { "epoch": 0.89, "grad_norm": 7.554742813110352, "learning_rate": 5.936942333950063e-07, "loss": 0.5904, "step": 6700 }, { "epoch": 0.89, "grad_norm": 7.517423629760742, "learning_rate": 5.831628236591929e-07, "loss": 0.5885, "step": 6710 }, { "epoch": 0.89, "grad_norm": 7.0763959884643555, "learning_rate": 5.727198717339511e-07, "loss": 0.6007, "step": 6720 }, { "epoch": 0.89, "grad_norm": 6.555006504058838, "learning_rate": 5.623655867645628e-07, "loss": 0.6059, "step": 6730 }, { "epoch": 0.89, "grad_norm": 6.526163101196289, "learning_rate": 5.521001761205441e-07, "loss": 0.5923, "step": 6740 }, { "epoch": 0.89, "grad_norm": 7.336793899536133, "learning_rate": 5.41923845391486e-07, "loss": 0.5961, "step": 6750 }, { "epoch": 0.89, "grad_norm": 7.833865642547607, "learning_rate": 5.318367983829393e-07, "loss": 0.6139, "step": 6760 }, { "epoch": 0.9, "grad_norm": 6.014789581298828, "learning_rate": 5.218392371123326e-07, "loss": 0.5936, "step": 6770 }, { "epoch": 0.9, "grad_norm": 5.68282413482666, "learning_rate": 5.119313618049309e-07, "loss": 0.6043, "step": 6780 }, { "epoch": 0.9, "grad_norm": 7.465061664581299, "learning_rate": 5.021133708898146e-07, "loss": 0.6057, "step": 6790 }, { "epoch": 0.9, "grad_norm": 6.3078083992004395, "learning_rate": 4.9238546099592e-07, "loss": 0.5766, "step": 6800 }, { "epoch": 0.9, "grad_norm": 6.257416725158691, "learning_rate": 4.827478269480895e-07, "loss": 0.6008, "step": 6810 }, { "epoch": 0.9, "grad_norm": 6.170396327972412, "learning_rate": 4.732006617631729e-07, "loss": 0.588, "step": 6820 }, { "epoch": 0.9, "grad_norm": 6.257867813110352, "learning_rate": 4.637441566461681e-07, "loss": 0.5986, "step": 6830 }, { "epoch": 0.91, "grad_norm": 6.7348952293396, "learning_rate": 4.54378500986381e-07, "loss": 0.6115, "step": 6840 }, { "epoch": 0.91, "grad_norm": 6.10768461227417, "learning_rate": 4.451038823536441e-07, "loss": 0.6063, "step": 6850 }, { "epoch": 0.91, "grad_norm": 6.2565226554870605, "learning_rate": 4.35920486494546e-07, "loss": 0.5842, "step": 6860 }, { "epoch": 0.91, "grad_norm": 7.341768741607666, "learning_rate": 4.268284973287273e-07, "loss": 0.5969, "step": 6870 }, { "epoch": 0.91, "grad_norm": 5.864892959594727, "learning_rate": 4.1782809694518533e-07, "loss": 0.597, "step": 6880 }, { "epoch": 0.91, "grad_norm": 5.842800617218018, "learning_rate": 4.089194655986306e-07, "loss": 0.5968, "step": 6890 }, { "epoch": 0.91, "grad_norm": 6.370209217071533, "learning_rate": 4.001027817058789e-07, "loss": 0.6031, "step": 6900 }, { "epoch": 0.91, "grad_norm": 6.6684489250183105, "learning_rate": 3.9137822184227845e-07, "loss": 0.5959, "step": 6910 }, { "epoch": 0.92, "grad_norm": 6.804083824157715, "learning_rate": 3.8274596073816784e-07, "loss": 0.5985, "step": 6920 }, { "epoch": 0.92, "grad_norm": 6.737246990203857, "learning_rate": 3.742061712753825e-07, "loss": 0.5926, "step": 6930 }, { "epoch": 0.92, "grad_norm": 7.578712463378906, "learning_rate": 3.657590244837911e-07, "loss": 0.5812, "step": 6940 }, { "epoch": 0.92, "grad_norm": 7.795099258422852, "learning_rate": 3.5740468953786854e-07, "loss": 0.5836, "step": 6950 }, { "epoch": 0.92, "grad_norm": 8.640656471252441, "learning_rate": 3.49143333753309e-07, "loss": 0.598, "step": 6960 }, { "epoch": 0.92, "grad_norm": 7.3872971534729, "learning_rate": 3.4097512258367385e-07, "loss": 0.5771, "step": 6970 }, { "epoch": 0.92, "grad_norm": 6.033993244171143, "learning_rate": 3.3290021961708163e-07, "loss": 0.5866, "step": 6980 }, { "epoch": 0.93, "grad_norm": 7.115429401397705, "learning_rate": 3.2491878657292643e-07, "loss": 0.5888, "step": 6990 }, { "epoch": 0.93, "grad_norm": 7.127076148986816, "learning_rate": 3.1703098329864237e-07, "loss": 0.5962, "step": 7000 }, { "epoch": 0.93, "grad_norm": 5.615909576416016, "learning_rate": 3.0923696776650414e-07, "loss": 0.5977, "step": 7010 }, { "epoch": 0.93, "grad_norm": 7.339127540588379, "learning_rate": 3.015368960704584e-07, "loss": 0.5952, "step": 7020 }, { "epoch": 0.93, "eval_loss": 0.6624879240989685, "eval_runtime": 169.2633, "eval_samples_per_second": 64.988, "eval_steps_per_second": 8.123, "step": 7020 }, { "epoch": 0.93, "grad_norm": 6.424869060516357, "learning_rate": 2.9393092242300026e-07, "loss": 0.6031, "step": 7030 }, { "epoch": 0.93, "grad_norm": 7.278805255889893, "learning_rate": 2.864191991520848e-07, "loss": 0.582, "step": 7040 }, { "epoch": 0.93, "grad_norm": 7.123143672943115, "learning_rate": 2.790018766980773e-07, "loss": 0.6012, "step": 7050 }, { "epoch": 0.93, "grad_norm": 7.571608543395996, "learning_rate": 2.71679103610738e-07, "loss": 0.5986, "step": 7060 }, { "epoch": 0.94, "grad_norm": 9.16704273223877, "learning_rate": 2.64451026546248e-07, "loss": 0.5973, "step": 7070 }, { "epoch": 0.94, "grad_norm": 21.925724029541016, "learning_rate": 2.573177902642726e-07, "loss": 0.6081, "step": 7080 }, { "epoch": 0.94, "grad_norm": 8.148933410644531, "learning_rate": 2.502795376250622e-07, "loss": 0.5902, "step": 7090 }, { "epoch": 0.94, "grad_norm": 7.818910598754883, "learning_rate": 2.4333640958659144e-07, "loss": 0.6025, "step": 7100 }, { "epoch": 0.94, "grad_norm": 8.272276878356934, "learning_rate": 2.3648854520173237e-07, "loss": 0.5759, "step": 7110 }, { "epoch": 0.94, "grad_norm": 7.721363544464111, "learning_rate": 2.2973608161547755e-07, "loss": 0.5904, "step": 7120 }, { "epoch": 0.94, "grad_norm": 6.874349117279053, "learning_rate": 2.2307915406218517e-07, "loss": 0.5935, "step": 7130 }, { "epoch": 0.94, "grad_norm": 7.524652481079102, "learning_rate": 2.1651789586287442e-07, "loss": 0.5943, "step": 7140 }, { "epoch": 0.95, "grad_norm": 7.76678991317749, "learning_rate": 2.1005243842255552e-07, "loss": 0.5978, "step": 7150 }, { "epoch": 0.95, "grad_norm": 7.988404750823975, "learning_rate": 2.0368291122759898e-07, "loss": 0.6036, "step": 7160 }, { "epoch": 0.95, "grad_norm": 10.661617279052734, "learning_rate": 1.9740944184313882e-07, "loss": 0.5818, "step": 7170 }, { "epoch": 0.95, "grad_norm": 8.164312362670898, "learning_rate": 1.9123215591052014e-07, "loss": 0.5947, "step": 7180 }, { "epoch": 0.95, "grad_norm": 6.9098639488220215, "learning_rate": 1.8515117714478447e-07, "loss": 0.5905, "step": 7190 }, { "epoch": 0.95, "grad_norm": 6.581024646759033, "learning_rate": 1.7916662733218848e-07, "loss": 0.5863, "step": 7200 }, { "epoch": 0.95, "grad_norm": 7.17993688583374, "learning_rate": 1.7327862632776638e-07, "loss": 0.5896, "step": 7210 }, { "epoch": 0.96, "grad_norm": 8.833417892456055, "learning_rate": 1.6748729205293024e-07, "loss": 0.5968, "step": 7220 }, { "epoch": 0.96, "grad_norm": 9.489375114440918, "learning_rate": 1.6179274049310966e-07, "loss": 0.5915, "step": 7230 }, { "epoch": 0.96, "grad_norm": 9.514772415161133, "learning_rate": 1.5619508569542363e-07, "loss": 0.6042, "step": 7240 }, { "epoch": 0.96, "grad_norm": 8.74714183807373, "learning_rate": 1.5069443976640287e-07, "loss": 0.5893, "step": 7250 }, { "epoch": 0.96, "grad_norm": 9.094064712524414, "learning_rate": 1.4529091286973994e-07, "loss": 0.5952, "step": 7260 }, { "epoch": 0.96, "grad_norm": 7.5646748542785645, "learning_rate": 1.3998461322408563e-07, "loss": 0.5858, "step": 7270 }, { "epoch": 0.96, "grad_norm": 6.778071880340576, "learning_rate": 1.3477564710088097e-07, "loss": 0.5909, "step": 7280 }, { "epoch": 0.96, "grad_norm": 8.356637954711914, "learning_rate": 1.2966411882222695e-07, "loss": 0.587, "step": 7290 }, { "epoch": 0.97, "grad_norm": 30.982200622558594, "learning_rate": 1.2465013075879884e-07, "loss": 0.5991, "step": 7300 }, { "epoch": 0.97, "grad_norm": 9.689143180847168, "learning_rate": 1.1973378332779229e-07, "loss": 0.5941, "step": 7310 }, { "epoch": 0.97, "grad_norm": 9.626836776733398, "learning_rate": 1.1491517499091498e-07, "loss": 0.6101, "step": 7320 }, { "epoch": 0.97, "grad_norm": 8.071940422058105, "learning_rate": 1.1019440225241317e-07, "loss": 0.6048, "step": 7330 }, { "epoch": 0.97, "grad_norm": 8.815977096557617, "learning_rate": 1.055715596571405e-07, "loss": 0.5978, "step": 7340 }, { "epoch": 0.97, "grad_norm": 8.037093162536621, "learning_rate": 1.0104673978866164e-07, "loss": 0.5957, "step": 7350 }, { "epoch": 0.97, "grad_norm": 10.43423080444336, "learning_rate": 9.662003326740166e-08, "loss": 0.5852, "step": 7360 }, { "epoch": 0.98, "grad_norm": 11.2993803024292, "learning_rate": 9.22915287488274e-08, "loss": 0.5991, "step": 7370 }, { "epoch": 0.98, "grad_norm": 8.106484413146973, "learning_rate": 8.80613129216762e-08, "loss": 0.5826, "step": 7380 }, { "epoch": 0.98, "grad_norm": 7.548207759857178, "learning_rate": 8.392947050621603e-08, "loss": 0.6091, "step": 7390 }, { "epoch": 0.98, "grad_norm": 9.398821830749512, "learning_rate": 7.989608425254924e-08, "loss": 0.6062, "step": 7400 }, { "epoch": 0.98, "grad_norm": 9.462056159973145, "learning_rate": 7.59612349389599e-08, "loss": 0.5883, "step": 7410 }, { "epoch": 0.98, "grad_norm": 7.41684627532959, "learning_rate": 7.212500137028789e-08, "loss": 0.5713, "step": 7420 }, { "epoch": 0.98, "grad_norm": 8.549308776855469, "learning_rate": 6.838746037635735e-08, "loss": 0.5913, "step": 7430 }, { "epoch": 0.98, "grad_norm": 9.199472427368164, "learning_rate": 6.474868681043578e-08, "loss": 0.5937, "step": 7440 }, { "epoch": 0.99, "grad_norm": 9.204155921936035, "learning_rate": 6.120875354773459e-08, "loss": 0.6123, "step": 7450 }, { "epoch": 0.99, "grad_norm": 11.123385429382324, "learning_rate": 5.776773148394976e-08, "loss": 0.601, "step": 7460 }, { "epoch": 0.99, "grad_norm": 10.552239418029785, "learning_rate": 5.4425689533841864e-08, "loss": 0.5879, "step": 7470 }, { "epoch": 0.99, "grad_norm": 9.744135856628418, "learning_rate": 5.1182694629857145e-08, "loss": 0.5935, "step": 7480 }, { "epoch": 0.99, "grad_norm": 8.253293991088867, "learning_rate": 4.803881172078473e-08, "loss": 0.5794, "step": 7490 }, { "epoch": 0.99, "grad_norm": 10.403715133666992, "learning_rate": 4.499410377045765e-08, "loss": 0.5859, "step": 7500 }, { "epoch": 0.99, "grad_norm": 10.5465669631958, "learning_rate": 4.2048631756492206e-08, "loss": 0.5915, "step": 7510 }, { "epoch": 1.0, "grad_norm": 10.897085189819336, "learning_rate": 3.9202454669063915e-08, "loss": 0.5948, "step": 7520 }, { "epoch": 1.0, "grad_norm": 10.889774322509766, "learning_rate": 3.645562950973014e-08, "loss": 0.5855, "step": 7530 }, { "epoch": 1.0, "grad_norm": 14.552692413330078, "learning_rate": 3.3808211290284886e-08, "loss": 0.5792, "step": 7540 }, { "epoch": 1.0, "grad_norm": 14.617317199707031, "learning_rate": 3.1260253031660247e-08, "loss": 0.5728, "step": 7550 }, { "epoch": 1.0, "grad_norm": 9.893068313598633, "learning_rate": 2.8811805762860578e-08, "loss": 0.5764, "step": 7560 }, { "epoch": 1.0, "grad_norm": 12.069340705871582, "learning_rate": 2.6462918519944425e-08, "loss": 0.5868, "step": 7570 }, { "epoch": 1.0, "grad_norm": 11.886162757873535, "learning_rate": 2.4213638345040868e-08, "loss": 0.5765, "step": 7580 }, { "epoch": 1.0, "grad_norm": 15.018707275390625, "learning_rate": 2.206401028540639e-08, "loss": 0.5941, "step": 7590 }, { "epoch": 1.01, "grad_norm": 13.571678161621094, "learning_rate": 2.0014077392525035e-08, "loss": 0.5735, "step": 7600 }, { "epoch": 1.01, "grad_norm": 10.342925071716309, "learning_rate": 1.8063880721242986e-08, "loss": 0.5858, "step": 7610 }, { "epoch": 1.01, "grad_norm": 12.00433349609375, "learning_rate": 1.6213459328950355e-08, "loss": 0.5805, "step": 7620 }, { "epoch": 1.01, "grad_norm": 11.714844703674316, "learning_rate": 1.4462850274794548e-08, "loss": 0.5769, "step": 7630 }, { "epoch": 1.01, "grad_norm": 14.100560188293457, "learning_rate": 1.2812088618942009e-08, "loss": 0.5837, "step": 7640 }, { "epoch": 1.01, "grad_norm": 11.162514686584473, "learning_rate": 1.1261207421874309e-08, "loss": 0.5901, "step": 7650 }, { "epoch": 1.01, "grad_norm": 13.700465202331543, "learning_rate": 9.810237743724805e-09, "loss": 0.6027, "step": 7660 }, { "epoch": 1.02, "grad_norm": 10.646740913391113, "learning_rate": 8.459208643659122e-09, "loss": 0.5797, "step": 7670 }, { "epoch": 1.02, "grad_norm": 14.71796703338623, "learning_rate": 7.2081471792911914e-09, "loss": 0.5724, "step": 7680 }, { "epoch": 1.02, "grad_norm": 14.192370414733887, "learning_rate": 6.057078406142003e-09, "loss": 0.5751, "step": 7690 }, { "epoch": 1.02, "grad_norm": 13.943014144897461, "learning_rate": 5.006025377138901e-09, "loss": 0.5801, "step": 7700 }, { "epoch": 1.02, "grad_norm": 16.19605827331543, "learning_rate": 4.055009142152066e-09, "loss": 0.5887, "step": 7710 }, { "epoch": 1.02, "grad_norm": 15.496548652648926, "learning_rate": 3.204048747573185e-09, "loss": 0.5885, "step": 7720 }, { "epoch": 1.02, "grad_norm": 14.580850601196289, "learning_rate": 2.4531612359363077e-09, "loss": 0.5806, "step": 7730 }, { "epoch": 1.02, "grad_norm": 11.053496360778809, "learning_rate": 1.8023616455731253e-09, "loss": 0.5916, "step": 7740 }, { "epoch": 1.03, "grad_norm": 12.858407020568848, "learning_rate": 1.2516630103137638e-09, "loss": 0.5805, "step": 7750 }, { "epoch": 1.03, "grad_norm": 12.221976280212402, "learning_rate": 8.010763592264381e-10, "loss": 0.5747, "step": 7760 }, { "epoch": 1.03, "grad_norm": 14.73697566986084, "learning_rate": 4.506107163948503e-10, "loss": 0.5811, "step": 7770 }, { "epoch": 1.03, "grad_norm": 11.176998138427734, "learning_rate": 2.0027310073833516e-10, "loss": 0.5929, "step": 7780 }, { "epoch": 1.03, "grad_norm": 17.99811363220215, "learning_rate": 5.0068525870305974e-11, "loss": 0.5805, "step": 7790 }, { "epoch": 1.03, "grad_norm": 12.911456108093262, "learning_rate": 0.0, "loss": 0.5872, "step": 7800 }, { "epoch": 1.03, "eval_loss": 0.6635136008262634, "eval_runtime": 169.3272, "eval_samples_per_second": 64.963, "eval_steps_per_second": 8.12, "step": 7800 } ], "logging_steps": 10, "max_steps": 7800, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 1, "total_flos": 6.410944662073875e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }