{ "best_metric": null, "best_model_checkpoint": null, "epoch": 49.01960784313726, "eval_steps": 500, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.1, "learning_rate": 0.0003996, "loss": 9.2736, "step": 10 }, { "epoch": 0.2, "learning_rate": 0.0003992, "loss": 7.7626, "step": 20 }, { "epoch": 0.29, "learning_rate": 0.00039880000000000004, "loss": 7.5202, "step": 30 }, { "epoch": 0.39, "learning_rate": 0.00039840000000000003, "loss": 7.371, "step": 40 }, { "epoch": 0.49, "learning_rate": 0.000398, "loss": 7.2352, "step": 50 }, { "epoch": 0.59, "learning_rate": 0.0003976, "loss": 7.1352, "step": 60 }, { "epoch": 0.69, "learning_rate": 0.0003972, "loss": 7.0625, "step": 70 }, { "epoch": 0.78, "learning_rate": 0.0003968, "loss": 6.911, "step": 80 }, { "epoch": 0.88, "learning_rate": 0.00039640000000000004, "loss": 6.7717, "step": 90 }, { "epoch": 0.98, "learning_rate": 0.00039600000000000003, "loss": 6.6484, "step": 100 }, { "epoch": 1.08, "learning_rate": 0.0003956, "loss": 6.5492, "step": 110 }, { "epoch": 1.18, "learning_rate": 0.0003952, "loss": 6.4417, "step": 120 }, { "epoch": 1.27, "learning_rate": 0.0003948, "loss": 6.3311, "step": 130 }, { "epoch": 1.37, "learning_rate": 0.0003944, "loss": 6.2206, "step": 140 }, { "epoch": 1.47, "learning_rate": 0.00039400000000000004, "loss": 6.1179, "step": 150 }, { "epoch": 1.57, "learning_rate": 0.0003936, "loss": 6.0151, "step": 160 }, { "epoch": 1.67, "learning_rate": 0.0003932, "loss": 5.9035, "step": 170 }, { "epoch": 1.76, "learning_rate": 0.0003928, "loss": 5.8222, "step": 180 }, { "epoch": 1.86, "learning_rate": 0.0003924, "loss": 5.7042, "step": 190 }, { "epoch": 1.96, "learning_rate": 0.000392, "loss": 5.6265, "step": 200 }, { "epoch": 2.06, "learning_rate": 0.00039160000000000003, "loss": 5.5338, "step": 210 }, { "epoch": 2.16, "learning_rate": 0.0003912, "loss": 5.4521, "step": 220 }, { "epoch": 2.25, "learning_rate": 0.0003908, "loss": 5.3552, "step": 230 }, { "epoch": 2.35, "learning_rate": 0.0003904, "loss": 5.2771, "step": 240 }, { "epoch": 2.45, "learning_rate": 0.00039, "loss": 5.1587, "step": 250 }, { "epoch": 2.55, "learning_rate": 0.0003896, "loss": 5.0899, "step": 260 }, { "epoch": 2.65, "learning_rate": 0.00038920000000000003, "loss": 5.0191, "step": 270 }, { "epoch": 2.75, "learning_rate": 0.0003888, "loss": 4.9602, "step": 280 }, { "epoch": 2.84, "learning_rate": 0.0003884, "loss": 4.8366, "step": 290 }, { "epoch": 2.94, "learning_rate": 0.000388, "loss": 4.7848, "step": 300 }, { "epoch": 3.04, "learning_rate": 0.0003876, "loss": 4.7199, "step": 310 }, { "epoch": 3.14, "learning_rate": 0.00038720000000000003, "loss": 4.6306, "step": 320 }, { "epoch": 3.24, "learning_rate": 0.0003868, "loss": 4.5337, "step": 330 }, { "epoch": 3.33, "learning_rate": 0.0003864, "loss": 4.4796, "step": 340 }, { "epoch": 3.43, "learning_rate": 0.000386, "loss": 4.3881, "step": 350 }, { "epoch": 3.53, "learning_rate": 0.0003856, "loss": 4.2989, "step": 360 }, { "epoch": 3.63, "learning_rate": 0.0003852, "loss": 4.2533, "step": 370 }, { "epoch": 3.73, "learning_rate": 0.00038480000000000003, "loss": 4.2379, "step": 380 }, { "epoch": 3.82, "learning_rate": 0.0003844, "loss": 4.142, "step": 390 }, { "epoch": 3.92, "learning_rate": 0.000384, "loss": 4.0793, "step": 400 }, { "epoch": 4.02, "learning_rate": 0.0003836, "loss": 4.0005, "step": 410 }, { "epoch": 4.12, "learning_rate": 0.0003832, "loss": 3.9619, "step": 420 }, { "epoch": 4.22, "learning_rate": 0.0003828, "loss": 3.8687, "step": 430 }, { "epoch": 4.31, "learning_rate": 0.0003824, "loss": 3.8486, "step": 440 }, { "epoch": 4.41, "learning_rate": 0.000382, "loss": 3.7684, "step": 450 }, { "epoch": 4.51, "learning_rate": 0.0003816, "loss": 3.7013, "step": 460 }, { "epoch": 4.61, "learning_rate": 0.0003812, "loss": 3.6851, "step": 470 }, { "epoch": 4.71, "learning_rate": 0.0003808, "loss": 3.6585, "step": 480 }, { "epoch": 4.8, "learning_rate": 0.0003804, "loss": 3.6172, "step": 490 }, { "epoch": 4.9, "learning_rate": 0.00038, "loss": 3.5557, "step": 500 }, { "epoch": 5.0, "learning_rate": 0.0003796, "loss": 3.4746, "step": 510 }, { "epoch": 5.1, "learning_rate": 0.0003792, "loss": 3.4473, "step": 520 }, { "epoch": 5.2, "learning_rate": 0.0003788, "loss": 3.3828, "step": 530 }, { "epoch": 5.29, "learning_rate": 0.0003784, "loss": 3.3868, "step": 540 }, { "epoch": 5.39, "learning_rate": 0.00037799999999999997, "loss": 3.31, "step": 550 }, { "epoch": 5.49, "learning_rate": 0.0003776, "loss": 3.2628, "step": 560 }, { "epoch": 5.59, "learning_rate": 0.0003772, "loss": 3.2541, "step": 570 }, { "epoch": 5.69, "learning_rate": 0.0003768, "loss": 3.223, "step": 580 }, { "epoch": 5.78, "learning_rate": 0.0003764, "loss": 3.2028, "step": 590 }, { "epoch": 5.88, "learning_rate": 0.000376, "loss": 3.1659, "step": 600 }, { "epoch": 5.98, "learning_rate": 0.0003756, "loss": 3.0847, "step": 610 }, { "epoch": 6.08, "learning_rate": 0.0003752, "loss": 3.0215, "step": 620 }, { "epoch": 6.18, "learning_rate": 0.00037480000000000006, "loss": 3.0149, "step": 630 }, { "epoch": 6.27, "learning_rate": 0.00037440000000000005, "loss": 3.0177, "step": 640 }, { "epoch": 6.37, "learning_rate": 0.00037400000000000004, "loss": 2.9474, "step": 650 }, { "epoch": 6.47, "learning_rate": 0.00037360000000000003, "loss": 2.9245, "step": 660 }, { "epoch": 6.57, "learning_rate": 0.0003732, "loss": 2.9218, "step": 670 }, { "epoch": 6.67, "learning_rate": 0.00037280000000000006, "loss": 2.8666, "step": 680 }, { "epoch": 6.76, "learning_rate": 0.00037240000000000005, "loss": 2.8821, "step": 690 }, { "epoch": 6.86, "learning_rate": 0.00037200000000000004, "loss": 2.8243, "step": 700 }, { "epoch": 6.96, "learning_rate": 0.00037160000000000003, "loss": 2.7753, "step": 710 }, { "epoch": 7.06, "learning_rate": 0.0003712, "loss": 2.7086, "step": 720 }, { "epoch": 7.16, "learning_rate": 0.0003708, "loss": 2.7104, "step": 730 }, { "epoch": 7.25, "learning_rate": 0.00037040000000000006, "loss": 2.7103, "step": 740 }, { "epoch": 7.35, "learning_rate": 0.00037000000000000005, "loss": 2.6707, "step": 750 }, { "epoch": 7.45, "learning_rate": 0.00036960000000000004, "loss": 2.6413, "step": 760 }, { "epoch": 7.55, "learning_rate": 0.00036920000000000003, "loss": 2.6359, "step": 770 }, { "epoch": 7.65, "learning_rate": 0.0003688, "loss": 2.5838, "step": 780 }, { "epoch": 7.75, "learning_rate": 0.0003684, "loss": 2.6212, "step": 790 }, { "epoch": 7.84, "learning_rate": 0.00036800000000000005, "loss": 2.5718, "step": 800 }, { "epoch": 7.94, "learning_rate": 0.00036760000000000004, "loss": 2.5348, "step": 810 }, { "epoch": 8.04, "learning_rate": 0.00036720000000000004, "loss": 2.4195, "step": 820 }, { "epoch": 8.14, "learning_rate": 0.0003668, "loss": 2.4938, "step": 830 }, { "epoch": 8.24, "learning_rate": 0.0003664, "loss": 2.4372, "step": 840 }, { "epoch": 8.33, "learning_rate": 0.000366, "loss": 2.4567, "step": 850 }, { "epoch": 8.43, "learning_rate": 0.00036560000000000005, "loss": 2.4108, "step": 860 }, { "epoch": 8.53, "learning_rate": 0.00036520000000000004, "loss": 2.3993, "step": 870 }, { "epoch": 8.63, "learning_rate": 0.00036480000000000003, "loss": 2.3739, "step": 880 }, { "epoch": 8.73, "learning_rate": 0.0003644, "loss": 2.391, "step": 890 }, { "epoch": 8.82, "learning_rate": 0.000364, "loss": 2.38, "step": 900 }, { "epoch": 8.92, "learning_rate": 0.00036360000000000006, "loss": 2.3257, "step": 910 }, { "epoch": 9.02, "learning_rate": 0.00036320000000000005, "loss": 2.1956, "step": 920 }, { "epoch": 9.12, "learning_rate": 0.00036280000000000004, "loss": 2.2944, "step": 930 }, { "epoch": 9.22, "learning_rate": 0.0003624, "loss": 2.2163, "step": 940 }, { "epoch": 9.31, "learning_rate": 0.000362, "loss": 2.2641, "step": 950 }, { "epoch": 9.41, "learning_rate": 0.0003616, "loss": 2.2035, "step": 960 }, { "epoch": 9.51, "learning_rate": 0.00036120000000000005, "loss": 2.188, "step": 970 }, { "epoch": 9.61, "learning_rate": 0.00036080000000000004, "loss": 2.197, "step": 980 }, { "epoch": 9.71, "learning_rate": 0.00036040000000000003, "loss": 2.18, "step": 990 }, { "epoch": 9.8, "learning_rate": 0.00036, "loss": 2.1994, "step": 1000 }, { "epoch": 9.9, "learning_rate": 0.0003596, "loss": 2.1538, "step": 1010 }, { "epoch": 10.0, "learning_rate": 0.0003592, "loss": 2.0495, "step": 1020 }, { "epoch": 10.1, "learning_rate": 0.00035880000000000005, "loss": 2.0501, "step": 1030 }, { "epoch": 10.2, "learning_rate": 0.00035840000000000004, "loss": 2.0302, "step": 1040 }, { "epoch": 10.29, "learning_rate": 0.00035800000000000003, "loss": 2.0996, "step": 1050 }, { "epoch": 10.39, "learning_rate": 0.0003576, "loss": 2.0369, "step": 1060 }, { "epoch": 10.49, "learning_rate": 0.0003572, "loss": 1.9996, "step": 1070 }, { "epoch": 10.59, "learning_rate": 0.0003568, "loss": 2.0254, "step": 1080 }, { "epoch": 10.69, "learning_rate": 0.00035640000000000004, "loss": 2.0004, "step": 1090 }, { "epoch": 10.78, "learning_rate": 0.00035600000000000003, "loss": 2.0154, "step": 1100 }, { "epoch": 10.88, "learning_rate": 0.0003556, "loss": 2.0014, "step": 1110 }, { "epoch": 10.98, "learning_rate": 0.00035524, "loss": 1.9377, "step": 1120 }, { "epoch": 11.08, "learning_rate": 0.00035484000000000004, "loss": 1.8296, "step": 1130 }, { "epoch": 11.18, "learning_rate": 0.00035444000000000003, "loss": 1.9002, "step": 1140 }, { "epoch": 11.27, "learning_rate": 0.00035404, "loss": 1.9336, "step": 1150 }, { "epoch": 11.37, "learning_rate": 0.00035364, "loss": 1.9041, "step": 1160 }, { "epoch": 11.47, "learning_rate": 0.00035324, "loss": 1.8324, "step": 1170 }, { "epoch": 11.57, "learning_rate": 0.00035284, "loss": 1.8663, "step": 1180 }, { "epoch": 11.67, "learning_rate": 0.00035244000000000003, "loss": 1.8358, "step": 1190 }, { "epoch": 11.76, "learning_rate": 0.00035204, "loss": 1.8586, "step": 1200 }, { "epoch": 11.86, "learning_rate": 0.00035164, "loss": 1.8452, "step": 1210 }, { "epoch": 11.96, "learning_rate": 0.00035124, "loss": 1.7957, "step": 1220 }, { "epoch": 12.06, "learning_rate": 0.00035084, "loss": 1.6603, "step": 1230 }, { "epoch": 12.16, "learning_rate": 0.00035044000000000004, "loss": 1.7903, "step": 1240 }, { "epoch": 12.25, "learning_rate": 0.00035004000000000003, "loss": 1.7574, "step": 1250 }, { "epoch": 12.35, "learning_rate": 0.00034964, "loss": 1.7823, "step": 1260 }, { "epoch": 12.45, "learning_rate": 0.00034924, "loss": 1.7038, "step": 1270 }, { "epoch": 12.55, "learning_rate": 0.00034884, "loss": 1.7186, "step": 1280 }, { "epoch": 12.65, "learning_rate": 0.00034844, "loss": 1.6948, "step": 1290 }, { "epoch": 12.75, "learning_rate": 0.00034804000000000004, "loss": 1.7347, "step": 1300 }, { "epoch": 12.84, "learning_rate": 0.00034764, "loss": 1.7304, "step": 1310 }, { "epoch": 12.94, "learning_rate": 0.00034724, "loss": 1.6862, "step": 1320 }, { "epoch": 13.04, "learning_rate": 0.00034684, "loss": 1.4857, "step": 1330 }, { "epoch": 13.14, "learning_rate": 0.00034644, "loss": 1.6981, "step": 1340 }, { "epoch": 13.24, "learning_rate": 0.00034604, "loss": 1.5926, "step": 1350 }, { "epoch": 13.33, "learning_rate": 0.00034564000000000003, "loss": 1.6545, "step": 1360 }, { "epoch": 13.43, "learning_rate": 0.00034524, "loss": 1.5918, "step": 1370 }, { "epoch": 13.53, "learning_rate": 0.00034484, "loss": 1.5743, "step": 1380 }, { "epoch": 13.63, "learning_rate": 0.00034444, "loss": 1.5867, "step": 1390 }, { "epoch": 13.73, "learning_rate": 0.00034404, "loss": 1.6042, "step": 1400 }, { "epoch": 13.82, "learning_rate": 0.00034364, "loss": 1.6308, "step": 1410 }, { "epoch": 13.92, "learning_rate": 0.00034324000000000003, "loss": 1.5813, "step": 1420 }, { "epoch": 14.02, "learning_rate": 0.00034284, "loss": 1.364, "step": 1430 }, { "epoch": 14.12, "learning_rate": 0.00034244, "loss": 1.5849, "step": 1440 }, { "epoch": 14.22, "learning_rate": 0.00034204, "loss": 1.4957, "step": 1450 }, { "epoch": 14.31, "learning_rate": 0.00034164, "loss": 1.5312, "step": 1460 }, { "epoch": 14.41, "learning_rate": 0.00034124, "loss": 1.4995, "step": 1470 }, { "epoch": 14.51, "learning_rate": 0.00034084, "loss": 1.4501, "step": 1480 }, { "epoch": 14.61, "learning_rate": 0.00034044, "loss": 1.4721, "step": 1490 }, { "epoch": 14.71, "learning_rate": 0.00034004, "loss": 1.4792, "step": 1500 }, { "epoch": 14.8, "learning_rate": 0.00033964, "loss": 1.513, "step": 1510 }, { "epoch": 14.9, "learning_rate": 0.00033924, "loss": 1.5028, "step": 1520 }, { "epoch": 15.0, "learning_rate": 0.00033884000000000003, "loss": 1.3511, "step": 1530 }, { "epoch": 15.1, "learning_rate": 0.00033844, "loss": 1.4152, "step": 1540 }, { "epoch": 15.2, "learning_rate": 0.00033804, "loss": 1.4216, "step": 1550 }, { "epoch": 15.29, "learning_rate": 0.00033764, "loss": 1.4177, "step": 1560 }, { "epoch": 15.39, "learning_rate": 0.00033724, "loss": 1.4074, "step": 1570 }, { "epoch": 15.49, "learning_rate": 0.00033684, "loss": 1.3532, "step": 1580 }, { "epoch": 15.59, "learning_rate": 0.00033644, "loss": 1.3525, "step": 1590 }, { "epoch": 15.69, "learning_rate": 0.00033604, "loss": 1.3755, "step": 1600 }, { "epoch": 15.78, "learning_rate": 0.00033564, "loss": 1.3989, "step": 1610 }, { "epoch": 15.88, "learning_rate": 0.00033524, "loss": 1.4012, "step": 1620 }, { "epoch": 15.98, "learning_rate": 0.00033484, "loss": 1.3386, "step": 1630 }, { "epoch": 16.08, "learning_rate": 0.00033444, "loss": 1.2476, "step": 1640 }, { "epoch": 16.18, "learning_rate": 0.00033404, "loss": 1.3435, "step": 1650 }, { "epoch": 16.27, "learning_rate": 0.00033364, "loss": 1.3131, "step": 1660 }, { "epoch": 16.37, "learning_rate": 0.00033324, "loss": 1.3099, "step": 1670 }, { "epoch": 16.47, "learning_rate": 0.00033284, "loss": 1.2693, "step": 1680 }, { "epoch": 16.57, "learning_rate": 0.00033244, "loss": 1.2649, "step": 1690 }, { "epoch": 16.67, "learning_rate": 0.00033203999999999997, "loss": 1.2777, "step": 1700 }, { "epoch": 16.76, "learning_rate": 0.00033164, "loss": 1.2889, "step": 1710 }, { "epoch": 16.86, "learning_rate": 0.00033124, "loss": 1.2729, "step": 1720 }, { "epoch": 16.96, "learning_rate": 0.00033084, "loss": 1.2529, "step": 1730 }, { "epoch": 17.06, "learning_rate": 0.00033044, "loss": 1.1566, "step": 1740 }, { "epoch": 17.16, "learning_rate": 0.00033004, "loss": 1.2668, "step": 1750 }, { "epoch": 17.25, "learning_rate": 0.00032964, "loss": 1.2052, "step": 1760 }, { "epoch": 17.35, "learning_rate": 0.00032924, "loss": 1.2159, "step": 1770 }, { "epoch": 17.45, "learning_rate": 0.00032884000000000006, "loss": 1.2003, "step": 1780 }, { "epoch": 17.55, "learning_rate": 0.00032844000000000005, "loss": 1.1836, "step": 1790 }, { "epoch": 17.65, "learning_rate": 0.00032804000000000004, "loss": 1.1857, "step": 1800 }, { "epoch": 17.75, "learning_rate": 0.00032764000000000003, "loss": 1.2012, "step": 1810 }, { "epoch": 17.84, "learning_rate": 0.00032724, "loss": 1.1907, "step": 1820 }, { "epoch": 17.94, "learning_rate": 0.00032684, "loss": 1.1723, "step": 1830 }, { "epoch": 18.04, "learning_rate": 0.00032644000000000005, "loss": 1.0299, "step": 1840 }, { "epoch": 18.14, "learning_rate": 0.00032604000000000004, "loss": 1.1996, "step": 1850 }, { "epoch": 18.24, "learning_rate": 0.00032564000000000003, "loss": 1.0919, "step": 1860 }, { "epoch": 18.33, "learning_rate": 0.00032524, "loss": 1.1291, "step": 1870 }, { "epoch": 18.43, "learning_rate": 0.00032484, "loss": 1.1343, "step": 1880 }, { "epoch": 18.53, "learning_rate": 0.00032444000000000006, "loss": 1.0965, "step": 1890 }, { "epoch": 18.63, "learning_rate": 0.00032404000000000005, "loss": 1.1149, "step": 1900 }, { "epoch": 18.73, "learning_rate": 0.00032364000000000004, "loss": 1.132, "step": 1910 }, { "epoch": 18.82, "learning_rate": 0.00032324000000000003, "loss": 1.1218, "step": 1920 }, { "epoch": 18.92, "learning_rate": 0.00032284, "loss": 1.0928, "step": 1930 }, { "epoch": 19.02, "learning_rate": 0.00032244, "loss": 0.9377, "step": 1940 }, { "epoch": 19.12, "learning_rate": 0.00032204000000000005, "loss": 1.1132, "step": 1950 }, { "epoch": 19.22, "learning_rate": 0.00032164000000000004, "loss": 1.0275, "step": 1960 }, { "epoch": 19.31, "learning_rate": 0.00032124000000000003, "loss": 1.0312, "step": 1970 }, { "epoch": 19.41, "learning_rate": 0.00032084, "loss": 1.0248, "step": 1980 }, { "epoch": 19.51, "learning_rate": 0.00032044, "loss": 1.0063, "step": 1990 }, { "epoch": 19.61, "learning_rate": 0.00032004, "loss": 1.0438, "step": 2000 }, { "epoch": 19.71, "learning_rate": 0.00031964000000000005, "loss": 1.0608, "step": 2010 }, { "epoch": 19.8, "learning_rate": 0.00031924000000000004, "loss": 1.0503, "step": 2020 }, { "epoch": 19.9, "learning_rate": 0.00031884000000000003, "loss": 1.0422, "step": 2030 }, { "epoch": 20.0, "learning_rate": 0.00031844, "loss": 0.9331, "step": 2040 }, { "epoch": 20.1, "learning_rate": 0.00031804, "loss": 0.9744, "step": 2050 }, { "epoch": 20.2, "learning_rate": 0.00031768000000000003, "loss": 0.9977, "step": 2060 }, { "epoch": 20.29, "learning_rate": 0.00031728, "loss": 0.9579, "step": 2070 }, { "epoch": 20.39, "learning_rate": 0.00031688, "loss": 0.908, "step": 2080 }, { "epoch": 20.49, "learning_rate": 0.00031648, "loss": 0.9323, "step": 2090 }, { "epoch": 20.59, "learning_rate": 0.00031608, "loss": 0.9597, "step": 2100 }, { "epoch": 20.69, "learning_rate": 0.00031568000000000004, "loss": 0.975, "step": 2110 }, { "epoch": 20.78, "learning_rate": 0.00031528000000000003, "loss": 0.9676, "step": 2120 }, { "epoch": 20.88, "learning_rate": 0.00031488, "loss": 0.979, "step": 2130 }, { "epoch": 20.98, "learning_rate": 0.00031448, "loss": 0.9263, "step": 2140 }, { "epoch": 21.08, "learning_rate": 0.00031408, "loss": 0.8463, "step": 2150 }, { "epoch": 21.18, "learning_rate": 0.00031368, "loss": 0.9642, "step": 2160 }, { "epoch": 21.27, "learning_rate": 0.00031328000000000004, "loss": 0.8091, "step": 2170 }, { "epoch": 21.37, "learning_rate": 0.00031288, "loss": 0.7911, "step": 2180 }, { "epoch": 21.47, "learning_rate": 0.00031248, "loss": 0.8679, "step": 2190 }, { "epoch": 21.57, "learning_rate": 0.00031208, "loss": 0.893, "step": 2200 }, { "epoch": 21.67, "learning_rate": 0.00031168, "loss": 0.9048, "step": 2210 }, { "epoch": 21.76, "learning_rate": 0.00031128000000000004, "loss": 0.9283, "step": 2220 }, { "epoch": 21.86, "learning_rate": 0.00031088000000000003, "loss": 0.9238, "step": 2230 }, { "epoch": 21.96, "learning_rate": 0.00031048, "loss": 0.8698, "step": 2240 }, { "epoch": 22.06, "learning_rate": 0.00031008, "loss": 0.794, "step": 2250 }, { "epoch": 22.16, "learning_rate": 0.00030968, "loss": 0.8646, "step": 2260 }, { "epoch": 22.25, "learning_rate": 0.00030928, "loss": 0.7995, "step": 2270 }, { "epoch": 22.35, "learning_rate": 0.00030888000000000004, "loss": 0.8045, "step": 2280 }, { "epoch": 22.45, "learning_rate": 0.00030848000000000003, "loss": 0.8186, "step": 2290 }, { "epoch": 22.55, "learning_rate": 0.00030808, "loss": 0.8085, "step": 2300 }, { "epoch": 22.65, "learning_rate": 0.00030768, "loss": 0.8098, "step": 2310 }, { "epoch": 22.75, "learning_rate": 0.00030728, "loss": 0.8741, "step": 2320 }, { "epoch": 22.84, "learning_rate": 0.00030688, "loss": 0.8602, "step": 2330 }, { "epoch": 22.94, "learning_rate": 0.00030648000000000003, "loss": 0.7975, "step": 2340 }, { "epoch": 23.04, "learning_rate": 0.00030608, "loss": 0.6979, "step": 2350 }, { "epoch": 23.14, "learning_rate": 0.00030568, "loss": 0.8089, "step": 2360 }, { "epoch": 23.24, "learning_rate": 0.00030528, "loss": 0.7418, "step": 2370 }, { "epoch": 23.33, "learning_rate": 0.00030488, "loss": 0.752, "step": 2380 }, { "epoch": 23.43, "learning_rate": 0.00030448, "loss": 0.7302, "step": 2390 }, { "epoch": 23.53, "learning_rate": 0.00030408000000000003, "loss": 0.7431, "step": 2400 }, { "epoch": 23.63, "learning_rate": 0.00030368, "loss": 0.7443, "step": 2410 }, { "epoch": 23.73, "learning_rate": 0.00030328, "loss": 0.7954, "step": 2420 }, { "epoch": 23.82, "learning_rate": 0.00030288, "loss": 0.8135, "step": 2430 }, { "epoch": 23.92, "learning_rate": 0.00030248, "loss": 0.7517, "step": 2440 }, { "epoch": 24.02, "learning_rate": 0.00030208, "loss": 0.6313, "step": 2450 }, { "epoch": 24.12, "learning_rate": 0.00030168, "loss": 0.7276, "step": 2460 }, { "epoch": 24.22, "learning_rate": 0.00030128, "loss": 0.6859, "step": 2470 }, { "epoch": 24.31, "learning_rate": 0.00030088, "loss": 0.6701, "step": 2480 }, { "epoch": 24.41, "learning_rate": 0.00030048, "loss": 0.6492, "step": 2490 }, { "epoch": 24.51, "learning_rate": 0.00030008, "loss": 0.7006, "step": 2500 }, { "epoch": 24.61, "learning_rate": 0.00029968000000000003, "loss": 0.7105, "step": 2510 }, { "epoch": 24.71, "learning_rate": 0.00029928, "loss": 0.7175, "step": 2520 }, { "epoch": 24.8, "learning_rate": 0.00029888, "loss": 0.7297, "step": 2530 }, { "epoch": 24.9, "learning_rate": 0.00029848, "loss": 0.7295, "step": 2540 }, { "epoch": 25.0, "learning_rate": 0.00029808, "loss": 0.6337, "step": 2550 }, { "epoch": 25.1, "learning_rate": 0.00029768, "loss": 0.6452, "step": 2560 }, { "epoch": 25.2, "learning_rate": 0.00029728, "loss": 0.6282, "step": 2570 }, { "epoch": 25.29, "learning_rate": 0.00029688, "loss": 0.6218, "step": 2580 }, { "epoch": 25.39, "learning_rate": 0.00029648, "loss": 0.5982, "step": 2590 }, { "epoch": 25.49, "learning_rate": 0.00029608, "loss": 0.6601, "step": 2600 }, { "epoch": 25.59, "learning_rate": 0.00029568, "loss": 0.6761, "step": 2610 }, { "epoch": 25.69, "learning_rate": 0.00029528, "loss": 0.6685, "step": 2620 }, { "epoch": 25.78, "learning_rate": 0.00029488, "loss": 0.6506, "step": 2630 }, { "epoch": 25.88, "learning_rate": 0.00029448, "loss": 0.6804, "step": 2640 }, { "epoch": 25.98, "learning_rate": 0.00029408, "loss": 0.6387, "step": 2650 }, { "epoch": 26.08, "learning_rate": 0.00029368, "loss": 0.5516, "step": 2660 }, { "epoch": 26.18, "learning_rate": 0.00029328, "loss": 0.5891, "step": 2670 }, { "epoch": 26.27, "learning_rate": 0.00029288, "loss": 0.5907, "step": 2680 }, { "epoch": 26.37, "learning_rate": 0.00029248, "loss": 0.5559, "step": 2690 }, { "epoch": 26.47, "learning_rate": 0.00029208, "loss": 0.5942, "step": 2700 }, { "epoch": 26.57, "learning_rate": 0.00029168, "loss": 0.625, "step": 2710 }, { "epoch": 26.67, "learning_rate": 0.00029128, "loss": 0.621, "step": 2720 }, { "epoch": 26.76, "learning_rate": 0.00029088, "loss": 0.5987, "step": 2730 }, { "epoch": 26.86, "learning_rate": 0.00029047999999999997, "loss": 0.6137, "step": 2740 }, { "epoch": 26.96, "learning_rate": 0.00029008, "loss": 0.5928, "step": 2750 }, { "epoch": 27.06, "learning_rate": 0.00028968, "loss": 0.5423, "step": 2760 }, { "epoch": 27.16, "learning_rate": 0.00028928, "loss": 0.5484, "step": 2770 }, { "epoch": 27.25, "learning_rate": 0.00028888, "loss": 0.563, "step": 2780 }, { "epoch": 27.35, "learning_rate": 0.00028848, "loss": 0.5161, "step": 2790 }, { "epoch": 27.45, "learning_rate": 0.00028808, "loss": 0.5432, "step": 2800 }, { "epoch": 27.55, "learning_rate": 0.00028768, "loss": 0.5607, "step": 2810 }, { "epoch": 27.65, "learning_rate": 0.00028728, "loss": 0.557, "step": 2820 }, { "epoch": 27.75, "learning_rate": 0.00028688, "loss": 0.5698, "step": 2830 }, { "epoch": 27.84, "learning_rate": 0.00028648, "loss": 0.5571, "step": 2840 }, { "epoch": 27.94, "learning_rate": 0.00028607999999999997, "loss": 0.5525, "step": 2850 }, { "epoch": 28.04, "learning_rate": 0.00028568, "loss": 0.5017, "step": 2860 }, { "epoch": 28.14, "learning_rate": 0.00028528, "loss": 0.5348, "step": 2870 }, { "epoch": 28.24, "learning_rate": 0.00028488000000000005, "loss": 0.4948, "step": 2880 }, { "epoch": 28.33, "learning_rate": 0.00028448000000000004, "loss": 0.4791, "step": 2890 }, { "epoch": 28.43, "learning_rate": 0.00028408000000000003, "loss": 0.4908, "step": 2900 }, { "epoch": 28.53, "learning_rate": 0.00028368, "loss": 0.4943, "step": 2910 }, { "epoch": 28.63, "learning_rate": 0.00028328, "loss": 0.4995, "step": 2920 }, { "epoch": 28.73, "learning_rate": 0.00028288000000000006, "loss": 0.5252, "step": 2930 }, { "epoch": 28.82, "learning_rate": 0.00028248000000000005, "loss": 0.5126, "step": 2940 }, { "epoch": 28.92, "learning_rate": 0.00028208000000000004, "loss": 0.5077, "step": 2950 }, { "epoch": 29.02, "learning_rate": 0.00028168000000000003, "loss": 0.4462, "step": 2960 }, { "epoch": 29.12, "learning_rate": 0.00028128, "loss": 0.4924, "step": 2970 }, { "epoch": 29.22, "learning_rate": 0.00028088, "loss": 0.4502, "step": 2980 }, { "epoch": 29.31, "learning_rate": 0.00028048000000000005, "loss": 0.4494, "step": 2990 }, { "epoch": 29.41, "learning_rate": 0.00028008000000000004, "loss": 0.4389, "step": 3000 }, { "epoch": 29.51, "learning_rate": 0.00027968000000000003, "loss": 0.4506, "step": 3010 }, { "epoch": 29.61, "learning_rate": 0.00027928, "loss": 0.4621, "step": 3020 }, { "epoch": 29.71, "learning_rate": 0.00027888, "loss": 0.4678, "step": 3030 }, { "epoch": 29.8, "learning_rate": 0.00027848, "loss": 0.4587, "step": 3040 }, { "epoch": 29.9, "learning_rate": 0.00027808000000000005, "loss": 0.4719, "step": 3050 }, { "epoch": 30.0, "learning_rate": 0.00027768000000000004, "loss": 0.4271, "step": 3060 }, { "epoch": 30.1, "learning_rate": 0.00027728000000000003, "loss": 0.4325, "step": 3070 }, { "epoch": 30.2, "learning_rate": 0.00027688, "loss": 0.4136, "step": 3080 }, { "epoch": 30.29, "learning_rate": 0.00027648, "loss": 0.4362, "step": 3090 }, { "epoch": 30.39, "learning_rate": 0.00027608000000000005, "loss": 0.4099, "step": 3100 }, { "epoch": 30.49, "learning_rate": 0.00027568000000000004, "loss": 0.4238, "step": 3110 }, { "epoch": 30.59, "learning_rate": 0.00027528000000000003, "loss": 0.4295, "step": 3120 }, { "epoch": 30.69, "learning_rate": 0.00027488, "loss": 0.4264, "step": 3130 }, { "epoch": 30.78, "learning_rate": 0.00027448, "loss": 0.417, "step": 3140 }, { "epoch": 30.88, "learning_rate": 0.00027408, "loss": 0.4304, "step": 3150 }, { "epoch": 30.98, "learning_rate": 0.00027368000000000005, "loss": 0.4272, "step": 3160 }, { "epoch": 31.08, "learning_rate": 0.00027328000000000004, "loss": 0.3643, "step": 3170 }, { "epoch": 31.18, "learning_rate": 0.00027288000000000003, "loss": 0.3797, "step": 3180 }, { "epoch": 31.27, "learning_rate": 0.00027248, "loss": 0.4051, "step": 3190 }, { "epoch": 31.37, "learning_rate": 0.00027208, "loss": 0.3869, "step": 3200 }, { "epoch": 31.47, "learning_rate": 0.00027168, "loss": 0.3997, "step": 3210 }, { "epoch": 31.57, "learning_rate": 0.00027128000000000005, "loss": 0.4037, "step": 3220 }, { "epoch": 31.67, "learning_rate": 0.00027088000000000004, "loss": 0.385, "step": 3230 }, { "epoch": 31.76, "learning_rate": 0.00027048, "loss": 0.3919, "step": 3240 }, { "epoch": 31.86, "learning_rate": 0.00027008, "loss": 0.3823, "step": 3250 }, { "epoch": 31.96, "learning_rate": 0.00026968, "loss": 0.382, "step": 3260 }, { "epoch": 32.06, "learning_rate": 0.00026928, "loss": 0.349, "step": 3270 }, { "epoch": 32.16, "learning_rate": 0.00026888000000000004, "loss": 0.3416, "step": 3280 }, { "epoch": 32.25, "learning_rate": 0.00026848000000000003, "loss": 0.3733, "step": 3290 }, { "epoch": 32.35, "learning_rate": 0.00026808, "loss": 0.3589, "step": 3300 }, { "epoch": 32.45, "learning_rate": 0.00026768, "loss": 0.3697, "step": 3310 }, { "epoch": 32.55, "learning_rate": 0.00026728, "loss": 0.3787, "step": 3320 }, { "epoch": 32.65, "learning_rate": 0.00026688, "loss": 0.3564, "step": 3330 }, { "epoch": 32.75, "learning_rate": 0.00026648000000000004, "loss": 0.3636, "step": 3340 }, { "epoch": 32.84, "learning_rate": 0.00026608000000000003, "loss": 0.3519, "step": 3350 }, { "epoch": 32.94, "learning_rate": 0.00026568, "loss": 0.3526, "step": 3360 }, { "epoch": 33.04, "learning_rate": 0.00026528, "loss": 0.3287, "step": 3370 }, { "epoch": 33.14, "learning_rate": 0.00026488, "loss": 0.3138, "step": 3380 }, { "epoch": 33.24, "learning_rate": 0.00026448000000000004, "loss": 0.3274, "step": 3390 }, { "epoch": 33.33, "learning_rate": 0.00026408000000000003, "loss": 0.3302, "step": 3400 }, { "epoch": 33.43, "learning_rate": 0.00026368, "loss": 0.3327, "step": 3410 }, { "epoch": 33.53, "learning_rate": 0.00026328, "loss": 0.3383, "step": 3420 }, { "epoch": 33.63, "learning_rate": 0.00026288, "loss": 0.3343, "step": 3430 }, { "epoch": 33.73, "learning_rate": 0.00026248, "loss": 0.3375, "step": 3440 }, { "epoch": 33.82, "learning_rate": 0.00026208000000000004, "loss": 0.3378, "step": 3450 }, { "epoch": 33.92, "learning_rate": 0.00026168000000000003, "loss": 0.3298, "step": 3460 }, { "epoch": 34.02, "learning_rate": 0.00026128, "loss": 0.2956, "step": 3470 }, { "epoch": 34.12, "learning_rate": 0.00026088, "loss": 0.295, "step": 3480 }, { "epoch": 34.22, "learning_rate": 0.00026048, "loss": 0.2847, "step": 3490 }, { "epoch": 34.31, "learning_rate": 0.00026008, "loss": 0.2993, "step": 3500 }, { "epoch": 34.41, "learning_rate": 0.00025968000000000003, "loss": 0.2974, "step": 3510 }, { "epoch": 34.51, "learning_rate": 0.00025928, "loss": 0.3076, "step": 3520 }, { "epoch": 34.61, "learning_rate": 0.00025888, "loss": 0.3146, "step": 3530 }, { "epoch": 34.71, "learning_rate": 0.00025848, "loss": 0.3124, "step": 3540 }, { "epoch": 34.8, "learning_rate": 0.00025808, "loss": 0.3072, "step": 3550 }, { "epoch": 34.9, "learning_rate": 0.00025768, "loss": 0.3078, "step": 3560 }, { "epoch": 35.0, "learning_rate": 0.00025728000000000003, "loss": 0.2738, "step": 3570 }, { "epoch": 35.1, "learning_rate": 0.00025688, "loss": 0.2736, "step": 3580 }, { "epoch": 35.2, "learning_rate": 0.00025648, "loss": 0.257, "step": 3590 }, { "epoch": 35.29, "learning_rate": 0.00025608, "loss": 0.2671, "step": 3600 }, { "epoch": 35.39, "learning_rate": 0.00025568, "loss": 0.2638, "step": 3610 }, { "epoch": 35.49, "learning_rate": 0.00025528, "loss": 0.2707, "step": 3620 }, { "epoch": 35.59, "learning_rate": 0.00025488, "loss": 0.2795, "step": 3630 }, { "epoch": 35.69, "learning_rate": 0.00025448, "loss": 0.2765, "step": 3640 }, { "epoch": 35.78, "learning_rate": 0.00025408, "loss": 0.2757, "step": 3650 }, { "epoch": 35.88, "learning_rate": 0.00025368, "loss": 0.2779, "step": 3660 }, { "epoch": 35.98, "learning_rate": 0.00025328, "loss": 0.2628, "step": 3670 }, { "epoch": 36.08, "learning_rate": 0.00025288000000000003, "loss": 0.2403, "step": 3680 }, { "epoch": 36.18, "learning_rate": 0.00025248, "loss": 0.2361, "step": 3690 }, { "epoch": 36.27, "learning_rate": 0.00025208, "loss": 0.2394, "step": 3700 }, { "epoch": 36.37, "learning_rate": 0.00025168, "loss": 0.2309, "step": 3710 }, { "epoch": 36.47, "learning_rate": 0.00025128, "loss": 0.233, "step": 3720 }, { "epoch": 36.57, "learning_rate": 0.00025088, "loss": 0.2446, "step": 3730 }, { "epoch": 36.67, "learning_rate": 0.00025048000000000003, "loss": 0.2385, "step": 3740 }, { "epoch": 36.76, "learning_rate": 0.00025008, "loss": 0.239, "step": 3750 }, { "epoch": 36.86, "learning_rate": 0.00024968, "loss": 0.2392, "step": 3760 }, { "epoch": 36.96, "learning_rate": 0.00024928, "loss": 0.2318, "step": 3770 }, { "epoch": 37.06, "learning_rate": 0.00024888, "loss": 0.218, "step": 3780 }, { "epoch": 37.16, "learning_rate": 0.00024848, "loss": 0.2059, "step": 3790 }, { "epoch": 37.25, "learning_rate": 0.00024808, "loss": 0.2082, "step": 3800 }, { "epoch": 37.35, "learning_rate": 0.00024768, "loss": 0.2058, "step": 3810 }, { "epoch": 37.45, "learning_rate": 0.00024728, "loss": 0.2037, "step": 3820 }, { "epoch": 37.55, "learning_rate": 0.00024688, "loss": 0.2083, "step": 3830 }, { "epoch": 37.65, "learning_rate": 0.00024648, "loss": 0.2007, "step": 3840 }, { "epoch": 37.75, "learning_rate": 0.00024608, "loss": 0.2012, "step": 3850 }, { "epoch": 37.84, "learning_rate": 0.00024568, "loss": 0.1972, "step": 3860 }, { "epoch": 37.94, "learning_rate": 0.00024528, "loss": 0.1996, "step": 3870 }, { "epoch": 38.04, "learning_rate": 0.00024488, "loss": 0.1859, "step": 3880 }, { "epoch": 38.14, "learning_rate": 0.00024448, "loss": 0.1801, "step": 3890 }, { "epoch": 38.24, "learning_rate": 0.00024408, "loss": 0.1708, "step": 3900 }, { "epoch": 38.33, "learning_rate": 0.00024368, "loss": 0.1709, "step": 3910 }, { "epoch": 38.43, "learning_rate": 0.00024328, "loss": 0.1736, "step": 3920 }, { "epoch": 38.53, "learning_rate": 0.00024288, "loss": 0.1726, "step": 3930 }, { "epoch": 38.63, "learning_rate": 0.00024248, "loss": 0.1681, "step": 3940 }, { "epoch": 38.73, "learning_rate": 0.00024207999999999999, "loss": 0.1674, "step": 3950 }, { "epoch": 38.82, "learning_rate": 0.00024168, "loss": 0.1601, "step": 3960 }, { "epoch": 38.92, "learning_rate": 0.00024128, "loss": 0.1552, "step": 3970 }, { "epoch": 39.02, "learning_rate": 0.00024087999999999998, "loss": 0.1414, "step": 3980 }, { "epoch": 39.12, "learning_rate": 0.00024048, "loss": 0.151, "step": 3990 }, { "epoch": 39.22, "learning_rate": 0.00024008, "loss": 0.1418, "step": 4000 }, { "epoch": 39.31, "learning_rate": 0.00023967999999999998, "loss": 0.1372, "step": 4010 }, { "epoch": 39.41, "learning_rate": 0.00023928, "loss": 0.1334, "step": 4020 }, { "epoch": 39.51, "learning_rate": 0.00023888, "loss": 0.1339, "step": 4030 }, { "epoch": 39.61, "learning_rate": 0.00023847999999999998, "loss": 0.137, "step": 4040 }, { "epoch": 39.71, "learning_rate": 0.00023808, "loss": 0.1325, "step": 4050 }, { "epoch": 39.8, "learning_rate": 0.00023768, "loss": 0.1322, "step": 4060 }, { "epoch": 39.9, "learning_rate": 0.00023727999999999998, "loss": 0.1262, "step": 4070 }, { "epoch": 40.0, "learning_rate": 0.00023688, "loss": 0.1111, "step": 4080 }, { "epoch": 40.1, "learning_rate": 0.00023647999999999999, "loss": 0.1125, "step": 4090 }, { "epoch": 40.2, "learning_rate": 0.00023608, "loss": 0.1128, "step": 4100 }, { "epoch": 40.29, "learning_rate": 0.00023568, "loss": 0.1098, "step": 4110 }, { "epoch": 40.39, "learning_rate": 0.00023527999999999998, "loss": 0.1057, "step": 4120 }, { "epoch": 40.49, "learning_rate": 0.00023488000000000003, "loss": 0.1019, "step": 4130 }, { "epoch": 40.59, "learning_rate": 0.00023448000000000005, "loss": 0.1024, "step": 4140 }, { "epoch": 40.69, "learning_rate": 0.00023408000000000004, "loss": 0.1014, "step": 4150 }, { "epoch": 40.78, "learning_rate": 0.00023368000000000003, "loss": 0.1038, "step": 4160 }, { "epoch": 40.88, "learning_rate": 0.00023328000000000004, "loss": 0.096, "step": 4170 }, { "epoch": 40.98, "learning_rate": 0.00023288000000000003, "loss": 0.0943, "step": 4180 }, { "epoch": 41.08, "learning_rate": 0.00023248000000000002, "loss": 0.0821, "step": 4190 }, { "epoch": 41.18, "learning_rate": 0.00023208000000000004, "loss": 0.0861, "step": 4200 }, { "epoch": 41.27, "learning_rate": 0.00023168000000000003, "loss": 0.0854, "step": 4210 }, { "epoch": 41.37, "learning_rate": 0.00023128000000000002, "loss": 0.081, "step": 4220 }, { "epoch": 41.47, "learning_rate": 0.00023088000000000004, "loss": 0.0787, "step": 4230 }, { "epoch": 41.57, "learning_rate": 0.00023048000000000003, "loss": 0.0775, "step": 4240 }, { "epoch": 41.67, "learning_rate": 0.00023008000000000002, "loss": 0.0742, "step": 4250 }, { "epoch": 41.76, "learning_rate": 0.00022968000000000004, "loss": 0.0727, "step": 4260 }, { "epoch": 41.86, "learning_rate": 0.00022928000000000003, "loss": 0.0716, "step": 4270 }, { "epoch": 41.96, "learning_rate": 0.00022888000000000002, "loss": 0.0701, "step": 4280 }, { "epoch": 42.06, "learning_rate": 0.00022848000000000004, "loss": 0.0639, "step": 4290 }, { "epoch": 42.16, "learning_rate": 0.00022808000000000003, "loss": 0.0636, "step": 4300 }, { "epoch": 42.25, "learning_rate": 0.00022768000000000002, "loss": 0.0625, "step": 4310 }, { "epoch": 42.35, "learning_rate": 0.00022728000000000003, "loss": 0.0607, "step": 4320 }, { "epoch": 42.45, "learning_rate": 0.00022688000000000002, "loss": 0.058, "step": 4330 }, { "epoch": 42.55, "learning_rate": 0.00022648000000000001, "loss": 0.0567, "step": 4340 }, { "epoch": 42.65, "learning_rate": 0.00022608000000000003, "loss": 0.0538, "step": 4350 }, { "epoch": 42.75, "learning_rate": 0.00022568000000000002, "loss": 0.0528, "step": 4360 }, { "epoch": 42.84, "learning_rate": 0.00022528, "loss": 0.0507, "step": 4370 }, { "epoch": 42.94, "learning_rate": 0.00022488000000000003, "loss": 0.0481, "step": 4380 }, { "epoch": 43.04, "learning_rate": 0.00022448000000000002, "loss": 0.0434, "step": 4390 }, { "epoch": 43.14, "learning_rate": 0.00022408000000000004, "loss": 0.0451, "step": 4400 }, { "epoch": 43.24, "learning_rate": 0.00022368000000000003, "loss": 0.0433, "step": 4410 }, { "epoch": 43.33, "learning_rate": 0.00022328000000000002, "loss": 0.0419, "step": 4420 }, { "epoch": 43.43, "learning_rate": 0.00022288000000000003, "loss": 0.0408, "step": 4430 }, { "epoch": 43.53, "learning_rate": 0.00022248000000000002, "loss": 0.0386, "step": 4440 }, { "epoch": 43.63, "learning_rate": 0.00022208000000000002, "loss": 0.038, "step": 4450 }, { "epoch": 43.73, "learning_rate": 0.00022168000000000003, "loss": 0.0367, "step": 4460 }, { "epoch": 43.82, "learning_rate": 0.00022128000000000002, "loss": 0.0362, "step": 4470 }, { "epoch": 43.92, "learning_rate": 0.00022088, "loss": 0.0338, "step": 4480 }, { "epoch": 44.02, "learning_rate": 0.00022048000000000003, "loss": 0.0303, "step": 4490 }, { "epoch": 44.12, "learning_rate": 0.00022008000000000002, "loss": 0.0318, "step": 4500 }, { "epoch": 44.22, "learning_rate": 0.00021968, "loss": 0.031, "step": 4510 }, { "epoch": 44.31, "learning_rate": 0.00021928000000000003, "loss": 0.0295, "step": 4520 }, { "epoch": 44.41, "learning_rate": 0.00021888000000000002, "loss": 0.029, "step": 4530 }, { "epoch": 44.51, "learning_rate": 0.00021848, "loss": 0.0275, "step": 4540 }, { "epoch": 44.61, "learning_rate": 0.00021808000000000003, "loss": 0.0267, "step": 4550 }, { "epoch": 44.71, "learning_rate": 0.00021768000000000002, "loss": 0.0269, "step": 4560 }, { "epoch": 44.8, "learning_rate": 0.00021728, "loss": 0.0262, "step": 4570 }, { "epoch": 44.9, "learning_rate": 0.00021688000000000002, "loss": 0.0259, "step": 4580 }, { "epoch": 45.0, "learning_rate": 0.00021648000000000001, "loss": 0.0235, "step": 4590 }, { "epoch": 45.1, "learning_rate": 0.00021608, "loss": 0.0225, "step": 4600 }, { "epoch": 45.2, "learning_rate": 0.00021568000000000002, "loss": 0.0238, "step": 4610 }, { "epoch": 45.29, "learning_rate": 0.00021528, "loss": 0.0233, "step": 4620 }, { "epoch": 45.39, "learning_rate": 0.00021488, "loss": 0.0222, "step": 4630 }, { "epoch": 45.49, "learning_rate": 0.00021448000000000002, "loss": 0.0218, "step": 4640 }, { "epoch": 45.59, "learning_rate": 0.00021408, "loss": 0.0207, "step": 4650 }, { "epoch": 45.69, "learning_rate": 0.00021368, "loss": 0.0207, "step": 4660 }, { "epoch": 45.78, "learning_rate": 0.00021328000000000002, "loss": 0.0214, "step": 4670 }, { "epoch": 45.88, "learning_rate": 0.00021288, "loss": 0.021, "step": 4680 }, { "epoch": 45.98, "learning_rate": 0.00021248000000000003, "loss": 0.0205, "step": 4690 }, { "epoch": 46.08, "learning_rate": 0.00021208000000000002, "loss": 0.0179, "step": 4700 }, { "epoch": 46.18, "learning_rate": 0.00021168, "loss": 0.019, "step": 4710 }, { "epoch": 46.27, "learning_rate": 0.00021128000000000002, "loss": 0.0192, "step": 4720 }, { "epoch": 46.37, "learning_rate": 0.00021088000000000001, "loss": 0.0187, "step": 4730 }, { "epoch": 46.47, "learning_rate": 0.00021048, "loss": 0.0183, "step": 4740 }, { "epoch": 46.57, "learning_rate": 0.00021008000000000002, "loss": 0.0176, "step": 4750 }, { "epoch": 46.67, "learning_rate": 0.00020968, "loss": 0.0169, "step": 4760 }, { "epoch": 46.76, "learning_rate": 0.00020928, "loss": 0.0171, "step": 4770 }, { "epoch": 46.86, "learning_rate": 0.00020888000000000002, "loss": 0.0172, "step": 4780 }, { "epoch": 46.96, "learning_rate": 0.00020848, "loss": 0.0175, "step": 4790 }, { "epoch": 47.06, "learning_rate": 0.00020808, "loss": 0.0153, "step": 4800 }, { "epoch": 47.16, "learning_rate": 0.00020768000000000002, "loss": 0.0162, "step": 4810 }, { "epoch": 47.25, "learning_rate": 0.00020728, "loss": 0.0161, "step": 4820 }, { "epoch": 47.35, "learning_rate": 0.00020688, "loss": 0.0161, "step": 4830 }, { "epoch": 47.45, "learning_rate": 0.00020648000000000002, "loss": 0.0157, "step": 4840 }, { "epoch": 47.55, "learning_rate": 0.00020608, "loss": 0.0149, "step": 4850 }, { "epoch": 47.65, "learning_rate": 0.00020568, "loss": 0.0143, "step": 4860 }, { "epoch": 47.75, "learning_rate": 0.00020528, "loss": 0.0138, "step": 4870 }, { "epoch": 47.84, "learning_rate": 0.00020488, "loss": 0.0136, "step": 4880 }, { "epoch": 47.94, "learning_rate": 0.00020448, "loss": 0.014, "step": 4890 }, { "epoch": 48.04, "learning_rate": 0.00020408, "loss": 0.0131, "step": 4900 }, { "epoch": 48.14, "learning_rate": 0.00020368, "loss": 0.0141, "step": 4910 }, { "epoch": 48.24, "learning_rate": 0.00020328, "loss": 0.0134, "step": 4920 }, { "epoch": 48.33, "learning_rate": 0.00020288, "loss": 0.013, "step": 4930 }, { "epoch": 48.43, "learning_rate": 0.00020248, "loss": 0.0128, "step": 4940 }, { "epoch": 48.53, "learning_rate": 0.00020208, "loss": 0.0127, "step": 4950 }, { "epoch": 48.63, "learning_rate": 0.00020168, "loss": 0.012, "step": 4960 }, { "epoch": 48.73, "learning_rate": 0.00020128, "loss": 0.0114, "step": 4970 }, { "epoch": 48.82, "learning_rate": 0.00020088000000000001, "loss": 0.0113, "step": 4980 }, { "epoch": 48.92, "learning_rate": 0.00020048, "loss": 0.0118, "step": 4990 }, { "epoch": 49.02, "learning_rate": 0.00020008, "loss": 0.011, "step": 5000 } ], "logging_steps": 10, "max_steps": 10000, "num_train_epochs": 99, "save_steps": 1000, "total_flos": 3.3405215440896e+17, "trial_name": null, "trial_params": null }