{ "best_metric": 0.6965588331222534, "best_model_checkpoint": "./vit-base-hate-meme/checkpoint-532", "epoch": 8.0, "eval_steps": 500, "global_step": 4256, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02, "grad_norm": 6.126795768737793, "learning_rate": 4.000000000000001e-06, "loss": 0.7608, "step": 10 }, { "epoch": 0.04, "grad_norm": 5.251371383666992, "learning_rate": 8.000000000000001e-06, "loss": 0.7016, "step": 20 }, { "epoch": 0.06, "grad_norm": 5.532265663146973, "learning_rate": 1.2e-05, "loss": 0.6429, "step": 30 }, { "epoch": 0.08, "grad_norm": 5.4727888107299805, "learning_rate": 1.6000000000000003e-05, "loss": 0.7306, "step": 40 }, { "epoch": 0.09, "grad_norm": 5.2745771408081055, "learning_rate": 2e-05, "loss": 0.7079, "step": 50 }, { "epoch": 0.11, "grad_norm": 4.616114616394043, "learning_rate": 2.36e-05, "loss": 0.671, "step": 60 }, { "epoch": 0.13, "grad_norm": 4.487009048461914, "learning_rate": 2.7600000000000003e-05, "loss": 0.6954, "step": 70 }, { "epoch": 0.15, "grad_norm": 6.081264495849609, "learning_rate": 3.16e-05, "loss": 0.6968, "step": 80 }, { "epoch": 0.17, "grad_norm": 4.151589393615723, "learning_rate": 3.56e-05, "loss": 0.6523, "step": 90 }, { "epoch": 0.19, "grad_norm": 4.241616249084473, "learning_rate": 3.960000000000001e-05, "loss": 0.6785, "step": 100 }, { "epoch": 0.21, "grad_norm": 4.035406112670898, "learning_rate": 4.36e-05, "loss": 0.6779, "step": 110 }, { "epoch": 0.23, "grad_norm": 4.655460357666016, "learning_rate": 4.76e-05, "loss": 0.6418, "step": 120 }, { "epoch": 0.24, "grad_norm": 3.683321952819824, "learning_rate": 5.16e-05, "loss": 0.6415, "step": 130 }, { "epoch": 0.26, "grad_norm": 5.034379959106445, "learning_rate": 5.560000000000001e-05, "loss": 0.7168, "step": 140 }, { "epoch": 0.28, "grad_norm": 7.606465816497803, "learning_rate": 5.96e-05, "loss": 0.7146, "step": 150 }, { "epoch": 0.3, "grad_norm": 3.5599539279937744, "learning_rate": 6.36e-05, "loss": 0.6495, "step": 160 }, { "epoch": 0.32, "grad_norm": 8.060068130493164, "learning_rate": 6.76e-05, "loss": 0.6523, "step": 170 }, { "epoch": 0.34, "grad_norm": 2.229842185974121, "learning_rate": 7.16e-05, "loss": 0.6252, "step": 180 }, { "epoch": 0.36, "grad_norm": 2.0872910022735596, "learning_rate": 7.560000000000001e-05, "loss": 0.6729, "step": 190 }, { "epoch": 0.38, "grad_norm": 5.176510334014893, "learning_rate": 7.960000000000001e-05, "loss": 0.6706, "step": 200 }, { "epoch": 0.39, "grad_norm": 3.0015692710876465, "learning_rate": 8.36e-05, "loss": 0.6279, "step": 210 }, { "epoch": 0.41, "grad_norm": 3.563908576965332, "learning_rate": 8.76e-05, "loss": 0.5934, "step": 220 }, { "epoch": 0.43, "grad_norm": 4.134763240814209, "learning_rate": 9.16e-05, "loss": 0.6436, "step": 230 }, { "epoch": 0.45, "grad_norm": 6.163305759429932, "learning_rate": 9.56e-05, "loss": 0.6519, "step": 240 }, { "epoch": 0.47, "grad_norm": 4.374207019805908, "learning_rate": 9.960000000000001e-05, "loss": 0.6637, "step": 250 }, { "epoch": 0.49, "grad_norm": 4.880618572235107, "learning_rate": 0.00010360000000000001, "loss": 0.6486, "step": 260 }, { "epoch": 0.51, "grad_norm": 1.537030577659607, "learning_rate": 0.00010760000000000001, "loss": 0.622, "step": 270 }, { "epoch": 0.53, "grad_norm": 3.5295188426971436, "learning_rate": 0.00011160000000000002, "loss": 0.6234, "step": 280 }, { "epoch": 0.55, "grad_norm": 1.5816266536712646, "learning_rate": 0.00011559999999999999, "loss": 0.6321, "step": 290 }, { "epoch": 0.56, "grad_norm": 1.2675076723098755, "learning_rate": 0.0001192, "loss": 0.7066, "step": 300 }, { "epoch": 0.58, "grad_norm": 2.379913568496704, "learning_rate": 0.0001232, "loss": 0.6625, "step": 310 }, { "epoch": 0.6, "grad_norm": 2.8097121715545654, "learning_rate": 0.0001272, "loss": 0.6018, "step": 320 }, { "epoch": 0.62, "grad_norm": 6.209559917449951, "learning_rate": 0.00013120000000000002, "loss": 0.6946, "step": 330 }, { "epoch": 0.64, "grad_norm": 4.973850727081299, "learning_rate": 0.0001352, "loss": 0.6724, "step": 340 }, { "epoch": 0.66, "grad_norm": 1.5640188455581665, "learning_rate": 0.0001392, "loss": 0.6227, "step": 350 }, { "epoch": 0.68, "grad_norm": 3.425607442855835, "learning_rate": 0.0001432, "loss": 0.6515, "step": 360 }, { "epoch": 0.7, "grad_norm": 2.7312209606170654, "learning_rate": 0.0001472, "loss": 0.6477, "step": 370 }, { "epoch": 0.71, "grad_norm": 2.8867738246917725, "learning_rate": 0.00015120000000000002, "loss": 0.6713, "step": 380 }, { "epoch": 0.73, "grad_norm": 0.647611141204834, "learning_rate": 0.0001552, "loss": 0.6554, "step": 390 }, { "epoch": 0.75, "grad_norm": 1.5883524417877197, "learning_rate": 0.00015920000000000002, "loss": 0.6394, "step": 400 }, { "epoch": 0.77, "grad_norm": 3.988485097885132, "learning_rate": 0.0001632, "loss": 0.6706, "step": 410 }, { "epoch": 0.79, "grad_norm": 7.253389358520508, "learning_rate": 0.0001672, "loss": 0.6658, "step": 420 }, { "epoch": 0.81, "grad_norm": 0.7725067138671875, "learning_rate": 0.00017120000000000001, "loss": 0.7349, "step": 430 }, { "epoch": 0.83, "grad_norm": 3.858313798904419, "learning_rate": 0.0001752, "loss": 0.5985, "step": 440 }, { "epoch": 0.85, "grad_norm": 3.0623714923858643, "learning_rate": 0.00017920000000000002, "loss": 0.5857, "step": 450 }, { "epoch": 0.86, "grad_norm": 2.89859938621521, "learning_rate": 0.0001832, "loss": 0.7285, "step": 460 }, { "epoch": 0.88, "grad_norm": 5.508333683013916, "learning_rate": 0.00018720000000000002, "loss": 0.643, "step": 470 }, { "epoch": 0.9, "grad_norm": 1.188262939453125, "learning_rate": 0.0001912, "loss": 0.664, "step": 480 }, { "epoch": 0.92, "grad_norm": 2.729386329650879, "learning_rate": 0.0001952, "loss": 0.622, "step": 490 }, { "epoch": 0.94, "grad_norm": 1.0896118879318237, "learning_rate": 0.00019920000000000002, "loss": 0.5781, "step": 500 }, { "epoch": 0.96, "grad_norm": 1.856227159500122, "learning_rate": 0.0001995740149094782, "loss": 0.6579, "step": 510 }, { "epoch": 0.98, "grad_norm": 0.9613101482391357, "learning_rate": 0.0001990415335463259, "loss": 0.6003, "step": 520 }, { "epoch": 1.0, "grad_norm": 2.3655054569244385, "learning_rate": 0.0001985090521831736, "loss": 0.6407, "step": 530 }, { "epoch": 1.0, "eval_loss": 0.6965588331222534, "eval_runtime": 5.5301, "eval_samples_per_second": 90.415, "eval_steps_per_second": 2.893, "step": 532 }, { "epoch": 1.02, "grad_norm": 4.0246710777282715, "learning_rate": 0.0001979765708200213, "loss": 0.6144, "step": 540 }, { "epoch": 1.03, "grad_norm": 0.397265762090683, "learning_rate": 0.000197444089456869, "loss": 0.7252, "step": 550 }, { "epoch": 1.05, "grad_norm": 6.538491249084473, "learning_rate": 0.00019691160809371674, "loss": 0.6467, "step": 560 }, { "epoch": 1.07, "grad_norm": 5.099808692932129, "learning_rate": 0.00019637912673056444, "loss": 0.6668, "step": 570 }, { "epoch": 1.09, "grad_norm": 2.9582018852233887, "learning_rate": 0.00019584664536741216, "loss": 0.6395, "step": 580 }, { "epoch": 1.11, "grad_norm": 6.390412330627441, "learning_rate": 0.00019531416400425986, "loss": 0.6392, "step": 590 }, { "epoch": 1.13, "grad_norm": 0.7637470960617065, "learning_rate": 0.00019478168264110756, "loss": 0.6227, "step": 600 }, { "epoch": 1.15, "grad_norm": 3.5703885555267334, "learning_rate": 0.00019424920127795528, "loss": 0.6981, "step": 610 }, { "epoch": 1.17, "grad_norm": 0.8668873310089111, "learning_rate": 0.000193716719914803, "loss": 0.6175, "step": 620 }, { "epoch": 1.18, "grad_norm": 2.3757476806640625, "learning_rate": 0.0001931842385516507, "loss": 0.6463, "step": 630 }, { "epoch": 1.2, "grad_norm": 0.4950419068336487, "learning_rate": 0.0001926517571884984, "loss": 0.6656, "step": 640 }, { "epoch": 1.22, "grad_norm": 0.4995131492614746, "learning_rate": 0.00019211927582534613, "loss": 0.6087, "step": 650 }, { "epoch": 1.24, "grad_norm": 4.226231098175049, "learning_rate": 0.00019158679446219383, "loss": 0.6768, "step": 660 }, { "epoch": 1.26, "grad_norm": 4.077276706695557, "learning_rate": 0.00019105431309904153, "loss": 0.6551, "step": 670 }, { "epoch": 1.28, "grad_norm": 0.6081713438034058, "learning_rate": 0.00019052183173588926, "loss": 0.6334, "step": 680 }, { "epoch": 1.3, "grad_norm": 1.841720461845398, "learning_rate": 0.00018998935037273698, "loss": 0.6541, "step": 690 }, { "epoch": 1.32, "grad_norm": 0.833160400390625, "learning_rate": 0.00018945686900958468, "loss": 0.5732, "step": 700 }, { "epoch": 1.33, "grad_norm": 1.2434287071228027, "learning_rate": 0.00018892438764643238, "loss": 0.7005, "step": 710 }, { "epoch": 1.35, "grad_norm": 1.5411890745162964, "learning_rate": 0.0001883919062832801, "loss": 0.604, "step": 720 }, { "epoch": 1.37, "grad_norm": 0.5615927577018738, "learning_rate": 0.0001878594249201278, "loss": 0.6982, "step": 730 }, { "epoch": 1.39, "grad_norm": 1.1369315385818481, "learning_rate": 0.0001873269435569755, "loss": 0.6451, "step": 740 }, { "epoch": 1.41, "grad_norm": 0.7013240456581116, "learning_rate": 0.00018679446219382323, "loss": 0.6697, "step": 750 }, { "epoch": 1.43, "grad_norm": 0.7162796854972839, "learning_rate": 0.00018626198083067095, "loss": 0.6793, "step": 760 }, { "epoch": 1.45, "grad_norm": 0.7037357091903687, "learning_rate": 0.00018572949946751865, "loss": 0.6546, "step": 770 }, { "epoch": 1.47, "grad_norm": 3.4446752071380615, "learning_rate": 0.00018519701810436635, "loss": 0.5589, "step": 780 }, { "epoch": 1.48, "grad_norm": 2.896850109100342, "learning_rate": 0.00018466453674121408, "loss": 0.6368, "step": 790 }, { "epoch": 1.5, "grad_norm": 0.8148677945137024, "learning_rate": 0.00018413205537806177, "loss": 0.6817, "step": 800 }, { "epoch": 1.52, "grad_norm": 1.7753351926803589, "learning_rate": 0.00018359957401490947, "loss": 0.6491, "step": 810 }, { "epoch": 1.54, "grad_norm": 1.8540890216827393, "learning_rate": 0.0001830670926517572, "loss": 0.6578, "step": 820 }, { "epoch": 1.56, "grad_norm": 1.7155886888504028, "learning_rate": 0.00018253461128860492, "loss": 0.6473, "step": 830 }, { "epoch": 1.58, "grad_norm": 0.892581582069397, "learning_rate": 0.00018200212992545262, "loss": 0.6021, "step": 840 }, { "epoch": 1.6, "grad_norm": 0.7859013080596924, "learning_rate": 0.00018146964856230032, "loss": 0.6064, "step": 850 }, { "epoch": 1.62, "grad_norm": 6.423700332641602, "learning_rate": 0.00018093716719914805, "loss": 0.612, "step": 860 }, { "epoch": 1.64, "grad_norm": 1.622660517692566, "learning_rate": 0.00018040468583599575, "loss": 0.6688, "step": 870 }, { "epoch": 1.65, "grad_norm": 1.050557255744934, "learning_rate": 0.00017987220447284344, "loss": 0.6549, "step": 880 }, { "epoch": 1.67, "grad_norm": 3.1418724060058594, "learning_rate": 0.00017933972310969117, "loss": 0.6104, "step": 890 }, { "epoch": 1.69, "grad_norm": 3.0087502002716064, "learning_rate": 0.0001788072417465389, "loss": 0.6451, "step": 900 }, { "epoch": 1.71, "grad_norm": 3.000000476837158, "learning_rate": 0.0001782747603833866, "loss": 0.6247, "step": 910 }, { "epoch": 1.73, "grad_norm": 1.215047001838684, "learning_rate": 0.0001777422790202343, "loss": 0.6418, "step": 920 }, { "epoch": 1.75, "grad_norm": 2.694948434829712, "learning_rate": 0.00017720979765708202, "loss": 0.657, "step": 930 }, { "epoch": 1.77, "grad_norm": 3.017652988433838, "learning_rate": 0.00017667731629392972, "loss": 0.6534, "step": 940 }, { "epoch": 1.79, "grad_norm": 3.342782497406006, "learning_rate": 0.00017614483493077742, "loss": 0.66, "step": 950 }, { "epoch": 1.8, "grad_norm": 3.4681971073150635, "learning_rate": 0.00017561235356762514, "loss": 0.6681, "step": 960 }, { "epoch": 1.82, "grad_norm": 0.638871431350708, "learning_rate": 0.00017507987220447287, "loss": 0.6062, "step": 970 }, { "epoch": 1.84, "grad_norm": 0.615951657295227, "learning_rate": 0.00017454739084132057, "loss": 0.5745, "step": 980 }, { "epoch": 1.86, "grad_norm": 2.9999630451202393, "learning_rate": 0.00017401490947816826, "loss": 0.6137, "step": 990 }, { "epoch": 1.88, "grad_norm": 3.212564468383789, "learning_rate": 0.000173482428115016, "loss": 0.5914, "step": 1000 }, { "epoch": 1.9, "grad_norm": 2.063244104385376, "learning_rate": 0.0001729499467518637, "loss": 0.5904, "step": 1010 }, { "epoch": 1.92, "grad_norm": 4.356505393981934, "learning_rate": 0.00017241746538871139, "loss": 0.6904, "step": 1020 }, { "epoch": 1.94, "grad_norm": 2.85433292388916, "learning_rate": 0.0001718849840255591, "loss": 0.6621, "step": 1030 }, { "epoch": 1.95, "grad_norm": 0.8630966544151306, "learning_rate": 0.00017135250266240684, "loss": 0.6591, "step": 1040 }, { "epoch": 1.97, "grad_norm": 7.148132801055908, "learning_rate": 0.00017082002129925454, "loss": 0.6562, "step": 1050 }, { "epoch": 1.99, "grad_norm": 0.8249239921569824, "learning_rate": 0.00017028753993610223, "loss": 0.5847, "step": 1060 }, { "epoch": 2.0, "eval_loss": 0.7979968190193176, "eval_runtime": 4.8835, "eval_samples_per_second": 102.387, "eval_steps_per_second": 3.276, "step": 1064 }, { "epoch": 2.01, "grad_norm": 0.5970304012298584, "learning_rate": 0.00016975505857294996, "loss": 0.6398, "step": 1070 }, { "epoch": 2.03, "grad_norm": 2.641237258911133, "learning_rate": 0.00016922257720979766, "loss": 0.631, "step": 1080 }, { "epoch": 2.05, "grad_norm": 0.5415599942207336, "learning_rate": 0.00016869009584664536, "loss": 0.6337, "step": 1090 }, { "epoch": 2.07, "grad_norm": 5.68765115737915, "learning_rate": 0.00016815761448349308, "loss": 0.6154, "step": 1100 }, { "epoch": 2.09, "grad_norm": 3.6129837036132812, "learning_rate": 0.0001676251331203408, "loss": 0.68, "step": 1110 }, { "epoch": 2.11, "grad_norm": 0.3377436399459839, "learning_rate": 0.0001670926517571885, "loss": 0.6461, "step": 1120 }, { "epoch": 2.12, "grad_norm": 5.473901271820068, "learning_rate": 0.0001665601703940362, "loss": 0.6727, "step": 1130 }, { "epoch": 2.14, "grad_norm": 0.664064884185791, "learning_rate": 0.00016602768903088393, "loss": 0.5921, "step": 1140 }, { "epoch": 2.16, "grad_norm": 1.5953819751739502, "learning_rate": 0.00016549520766773163, "loss": 0.6052, "step": 1150 }, { "epoch": 2.18, "grad_norm": 2.9201242923736572, "learning_rate": 0.00016496272630457933, "loss": 0.6278, "step": 1160 }, { "epoch": 2.2, "grad_norm": 2.1254889965057373, "learning_rate": 0.00016443024494142705, "loss": 0.5732, "step": 1170 }, { "epoch": 2.22, "grad_norm": 4.827600002288818, "learning_rate": 0.00016389776357827478, "loss": 0.5687, "step": 1180 }, { "epoch": 2.24, "grad_norm": 4.18930721282959, "learning_rate": 0.00016336528221512248, "loss": 0.5923, "step": 1190 }, { "epoch": 2.26, "grad_norm": 3.059593439102173, "learning_rate": 0.00016283280085197018, "loss": 0.6697, "step": 1200 }, { "epoch": 2.27, "grad_norm": 0.7229830622673035, "learning_rate": 0.0001623003194888179, "loss": 0.5956, "step": 1210 }, { "epoch": 2.29, "grad_norm": 0.41750237345695496, "learning_rate": 0.0001617678381256656, "loss": 0.6543, "step": 1220 }, { "epoch": 2.31, "grad_norm": 1.0170073509216309, "learning_rate": 0.0001612353567625133, "loss": 0.6222, "step": 1230 }, { "epoch": 2.33, "grad_norm": 1.94419527053833, "learning_rate": 0.00016070287539936103, "loss": 0.6379, "step": 1240 }, { "epoch": 2.35, "grad_norm": 2.2258355617523193, "learning_rate": 0.00016017039403620875, "loss": 0.5887, "step": 1250 }, { "epoch": 2.37, "grad_norm": 1.861475944519043, "learning_rate": 0.00015963791267305645, "loss": 0.5848, "step": 1260 }, { "epoch": 2.39, "grad_norm": 1.9490504264831543, "learning_rate": 0.00015910543130990418, "loss": 0.6283, "step": 1270 }, { "epoch": 2.41, "grad_norm": 0.9739388823509216, "learning_rate": 0.00015857294994675187, "loss": 0.6027, "step": 1280 }, { "epoch": 2.42, "grad_norm": 1.012650966644287, "learning_rate": 0.00015804046858359957, "loss": 0.5685, "step": 1290 }, { "epoch": 2.44, "grad_norm": 0.5264813303947449, "learning_rate": 0.00015750798722044727, "loss": 0.619, "step": 1300 }, { "epoch": 2.46, "grad_norm": 1.5455741882324219, "learning_rate": 0.00015697550585729502, "loss": 0.6441, "step": 1310 }, { "epoch": 2.48, "grad_norm": 1.331289529800415, "learning_rate": 0.00015644302449414272, "loss": 0.6583, "step": 1320 }, { "epoch": 2.5, "grad_norm": 1.2682970762252808, "learning_rate": 0.00015591054313099042, "loss": 0.6573, "step": 1330 }, { "epoch": 2.52, "grad_norm": 2.107792377471924, "learning_rate": 0.00015537806176783815, "loss": 0.6044, "step": 1340 }, { "epoch": 2.54, "grad_norm": 0.5763486623764038, "learning_rate": 0.00015484558040468585, "loss": 0.5277, "step": 1350 }, { "epoch": 2.56, "grad_norm": 5.478113651275635, "learning_rate": 0.00015431309904153354, "loss": 0.6821, "step": 1360 }, { "epoch": 2.58, "grad_norm": 3.25028395652771, "learning_rate": 0.00015378061767838124, "loss": 0.6411, "step": 1370 }, { "epoch": 2.59, "grad_norm": 1.4305381774902344, "learning_rate": 0.000153248136315229, "loss": 0.6273, "step": 1380 }, { "epoch": 2.61, "grad_norm": 1.6666404008865356, "learning_rate": 0.0001527156549520767, "loss": 0.6636, "step": 1390 }, { "epoch": 2.63, "grad_norm": 4.375661373138428, "learning_rate": 0.0001521831735889244, "loss": 0.5989, "step": 1400 }, { "epoch": 2.65, "grad_norm": 3.827533483505249, "learning_rate": 0.00015165069222577212, "loss": 0.5814, "step": 1410 }, { "epoch": 2.67, "grad_norm": 2.938608169555664, "learning_rate": 0.00015111821086261982, "loss": 0.5575, "step": 1420 }, { "epoch": 2.69, "grad_norm": 2.966648817062378, "learning_rate": 0.00015058572949946752, "loss": 0.623, "step": 1430 }, { "epoch": 2.71, "grad_norm": 3.7637674808502197, "learning_rate": 0.00015005324813631521, "loss": 0.5792, "step": 1440 }, { "epoch": 2.73, "grad_norm": 2.175443172454834, "learning_rate": 0.00014952076677316297, "loss": 0.6161, "step": 1450 }, { "epoch": 2.74, "grad_norm": 0.6871302723884583, "learning_rate": 0.00014898828541001067, "loss": 0.6145, "step": 1460 }, { "epoch": 2.76, "grad_norm": 2.2254390716552734, "learning_rate": 0.00014845580404685836, "loss": 0.5978, "step": 1470 }, { "epoch": 2.78, "grad_norm": 2.5566136837005615, "learning_rate": 0.0001479233226837061, "loss": 0.6676, "step": 1480 }, { "epoch": 2.8, "grad_norm": 1.8439075946807861, "learning_rate": 0.0001473908413205538, "loss": 0.6386, "step": 1490 }, { "epoch": 2.82, "grad_norm": 2.6792383193969727, "learning_rate": 0.0001468583599574015, "loss": 0.6562, "step": 1500 }, { "epoch": 2.84, "grad_norm": 0.6229275465011597, "learning_rate": 0.0001463258785942492, "loss": 0.5914, "step": 1510 }, { "epoch": 2.86, "grad_norm": 5.081364631652832, "learning_rate": 0.00014579339723109694, "loss": 0.5784, "step": 1520 }, { "epoch": 2.88, "grad_norm": 4.535241603851318, "learning_rate": 0.00014526091586794464, "loss": 0.5823, "step": 1530 }, { "epoch": 2.89, "grad_norm": 0.9447025060653687, "learning_rate": 0.00014472843450479234, "loss": 0.6573, "step": 1540 }, { "epoch": 2.91, "grad_norm": 1.4323358535766602, "learning_rate": 0.00014419595314164006, "loss": 0.5481, "step": 1550 }, { "epoch": 2.93, "grad_norm": 2.6219279766082764, "learning_rate": 0.00014366347177848776, "loss": 0.6331, "step": 1560 }, { "epoch": 2.95, "grad_norm": 0.7964478135108948, "learning_rate": 0.00014313099041533546, "loss": 0.6071, "step": 1570 }, { "epoch": 2.97, "grad_norm": 1.5989696979522705, "learning_rate": 0.00014259850905218318, "loss": 0.621, "step": 1580 }, { "epoch": 2.99, "grad_norm": 2.2981741428375244, "learning_rate": 0.0001420660276890309, "loss": 0.6144, "step": 1590 }, { "epoch": 3.0, "eval_loss": 0.7764111161231995, "eval_runtime": 4.9574, "eval_samples_per_second": 100.859, "eval_steps_per_second": 3.227, "step": 1596 }, { "epoch": 3.01, "grad_norm": 1.12291419506073, "learning_rate": 0.0001415335463258786, "loss": 0.6111, "step": 1600 }, { "epoch": 3.03, "grad_norm": 3.5445947647094727, "learning_rate": 0.0001410010649627263, "loss": 0.5686, "step": 1610 }, { "epoch": 3.05, "grad_norm": 5.601056098937988, "learning_rate": 0.00014046858359957403, "loss": 0.5344, "step": 1620 }, { "epoch": 3.06, "grad_norm": 2.273033380508423, "learning_rate": 0.00013993610223642173, "loss": 0.5662, "step": 1630 }, { "epoch": 3.08, "grad_norm": 0.932608962059021, "learning_rate": 0.00013940362087326943, "loss": 0.5788, "step": 1640 }, { "epoch": 3.1, "grad_norm": 1.4313303232192993, "learning_rate": 0.00013887113951011716, "loss": 0.6199, "step": 1650 }, { "epoch": 3.12, "grad_norm": 1.161855936050415, "learning_rate": 0.00013833865814696488, "loss": 0.5824, "step": 1660 }, { "epoch": 3.14, "grad_norm": 3.128478527069092, "learning_rate": 0.00013780617678381258, "loss": 0.5774, "step": 1670 }, { "epoch": 3.16, "grad_norm": 2.3241376876831055, "learning_rate": 0.00013727369542066028, "loss": 0.6157, "step": 1680 }, { "epoch": 3.18, "grad_norm": 3.1155593395233154, "learning_rate": 0.000136741214057508, "loss": 0.6216, "step": 1690 }, { "epoch": 3.2, "grad_norm": 1.408307433128357, "learning_rate": 0.0001362087326943557, "loss": 0.6149, "step": 1700 }, { "epoch": 3.21, "grad_norm": 1.0534541606903076, "learning_rate": 0.0001356762513312034, "loss": 0.579, "step": 1710 }, { "epoch": 3.23, "grad_norm": 1.5528515577316284, "learning_rate": 0.00013514376996805113, "loss": 0.5386, "step": 1720 }, { "epoch": 3.25, "grad_norm": 2.1062605381011963, "learning_rate": 0.00013461128860489885, "loss": 0.6308, "step": 1730 }, { "epoch": 3.27, "grad_norm": 2.3048410415649414, "learning_rate": 0.00013407880724174655, "loss": 0.6255, "step": 1740 }, { "epoch": 3.29, "grad_norm": 3.4321556091308594, "learning_rate": 0.00013354632587859425, "loss": 0.6106, "step": 1750 }, { "epoch": 3.31, "grad_norm": 1.6664164066314697, "learning_rate": 0.00013301384451544198, "loss": 0.5444, "step": 1760 }, { "epoch": 3.33, "grad_norm": 2.9253435134887695, "learning_rate": 0.00013248136315228967, "loss": 0.5808, "step": 1770 }, { "epoch": 3.35, "grad_norm": 2.220623731613159, "learning_rate": 0.00013194888178913737, "loss": 0.5715, "step": 1780 }, { "epoch": 3.36, "grad_norm": 1.8884316682815552, "learning_rate": 0.0001314164004259851, "loss": 0.5269, "step": 1790 }, { "epoch": 3.38, "grad_norm": 1.592698335647583, "learning_rate": 0.00013088391906283282, "loss": 0.5806, "step": 1800 }, { "epoch": 3.4, "grad_norm": 1.9080153703689575, "learning_rate": 0.00013035143769968052, "loss": 0.6371, "step": 1810 }, { "epoch": 3.42, "grad_norm": 2.2922542095184326, "learning_rate": 0.00012981895633652822, "loss": 0.6221, "step": 1820 }, { "epoch": 3.44, "grad_norm": 1.9848778247833252, "learning_rate": 0.00012928647497337595, "loss": 0.5616, "step": 1830 }, { "epoch": 3.46, "grad_norm": 3.399679660797119, "learning_rate": 0.00012875399361022365, "loss": 0.6284, "step": 1840 }, { "epoch": 3.48, "grad_norm": 1.6504648923873901, "learning_rate": 0.00012822151224707134, "loss": 0.5542, "step": 1850 }, { "epoch": 3.5, "grad_norm": 3.0508906841278076, "learning_rate": 0.00012768903088391907, "loss": 0.5928, "step": 1860 }, { "epoch": 3.52, "grad_norm": 3.353853464126587, "learning_rate": 0.0001271565495207668, "loss": 0.5858, "step": 1870 }, { "epoch": 3.53, "grad_norm": 3.4700353145599365, "learning_rate": 0.0001266240681576145, "loss": 0.6719, "step": 1880 }, { "epoch": 3.55, "grad_norm": 0.9324420094490051, "learning_rate": 0.0001260915867944622, "loss": 0.5396, "step": 1890 }, { "epoch": 3.57, "grad_norm": 1.1515653133392334, "learning_rate": 0.00012555910543130992, "loss": 0.6344, "step": 1900 }, { "epoch": 3.59, "grad_norm": 1.1183910369873047, "learning_rate": 0.00012502662406815762, "loss": 0.5346, "step": 1910 }, { "epoch": 3.61, "grad_norm": 0.988416850566864, "learning_rate": 0.00012449414270500531, "loss": 0.611, "step": 1920 }, { "epoch": 3.63, "grad_norm": 2.163034439086914, "learning_rate": 0.00012396166134185304, "loss": 0.5806, "step": 1930 }, { "epoch": 3.65, "grad_norm": 0.785860538482666, "learning_rate": 0.00012342917997870077, "loss": 0.5344, "step": 1940 }, { "epoch": 3.67, "grad_norm": 1.1014069318771362, "learning_rate": 0.00012289669861554846, "loss": 0.5961, "step": 1950 }, { "epoch": 3.68, "grad_norm": 1.5329209566116333, "learning_rate": 0.00012236421725239616, "loss": 0.5685, "step": 1960 }, { "epoch": 3.7, "grad_norm": 2.057729721069336, "learning_rate": 0.00012183173588924388, "loss": 0.6254, "step": 1970 }, { "epoch": 3.72, "grad_norm": 1.8107123374938965, "learning_rate": 0.00012129925452609159, "loss": 0.5715, "step": 1980 }, { "epoch": 3.74, "grad_norm": 2.0086817741394043, "learning_rate": 0.0001207667731629393, "loss": 0.5606, "step": 1990 }, { "epoch": 3.76, "grad_norm": 3.7626357078552246, "learning_rate": 0.000120234291799787, "loss": 0.5334, "step": 2000 }, { "epoch": 3.78, "grad_norm": 3.1028711795806885, "learning_rate": 0.00011970181043663474, "loss": 0.6453, "step": 2010 }, { "epoch": 3.8, "grad_norm": 1.0698882341384888, "learning_rate": 0.00011916932907348244, "loss": 0.5851, "step": 2020 }, { "epoch": 3.82, "grad_norm": 2.4677796363830566, "learning_rate": 0.00011863684771033015, "loss": 0.6008, "step": 2030 }, { "epoch": 3.83, "grad_norm": 1.6755846738815308, "learning_rate": 0.00011810436634717785, "loss": 0.5903, "step": 2040 }, { "epoch": 3.85, "grad_norm": 1.8782827854156494, "learning_rate": 0.00011757188498402556, "loss": 0.6018, "step": 2050 }, { "epoch": 3.87, "grad_norm": 1.1513298749923706, "learning_rate": 0.00011703940362087327, "loss": 0.5819, "step": 2060 }, { "epoch": 3.89, "grad_norm": 2.3282463550567627, "learning_rate": 0.000116506922257721, "loss": 0.5476, "step": 2070 }, { "epoch": 3.91, "grad_norm": 2.159614324569702, "learning_rate": 0.00011597444089456871, "loss": 0.5934, "step": 2080 }, { "epoch": 3.93, "grad_norm": 1.6344852447509766, "learning_rate": 0.00011544195953141641, "loss": 0.4688, "step": 2090 }, { "epoch": 3.95, "grad_norm": 1.4373761415481567, "learning_rate": 0.00011490947816826412, "loss": 0.5604, "step": 2100 }, { "epoch": 3.97, "grad_norm": 2.425197124481201, "learning_rate": 0.00011437699680511182, "loss": 0.4934, "step": 2110 }, { "epoch": 3.98, "grad_norm": 3.7428033351898193, "learning_rate": 0.00011384451544195953, "loss": 0.5307, "step": 2120 }, { "epoch": 4.0, "eval_loss": 0.9912688136100769, "eval_runtime": 5.4585, "eval_samples_per_second": 91.6, "eval_steps_per_second": 2.931, "step": 2128 }, { "epoch": 4.0, "grad_norm": 1.4787501096725464, "learning_rate": 0.00011331203407880724, "loss": 0.5672, "step": 2130 }, { "epoch": 4.02, "grad_norm": 3.0657687187194824, "learning_rate": 0.00011277955271565497, "loss": 0.4967, "step": 2140 }, { "epoch": 4.04, "grad_norm": 1.8531616926193237, "learning_rate": 0.00011224707135250268, "loss": 0.4264, "step": 2150 }, { "epoch": 4.06, "grad_norm": 2.5244009494781494, "learning_rate": 0.00011171458998935038, "loss": 0.5213, "step": 2160 }, { "epoch": 4.08, "grad_norm": 1.9924266338348389, "learning_rate": 0.00011118210862619809, "loss": 0.5376, "step": 2170 }, { "epoch": 4.1, "grad_norm": 1.3312138319015503, "learning_rate": 0.0001106496272630458, "loss": 0.5506, "step": 2180 }, { "epoch": 4.12, "grad_norm": 0.9579734206199646, "learning_rate": 0.0001101171458998935, "loss": 0.4378, "step": 2190 }, { "epoch": 4.14, "grad_norm": 1.8804982900619507, "learning_rate": 0.00010958466453674121, "loss": 0.4032, "step": 2200 }, { "epoch": 4.15, "grad_norm": 1.5957534313201904, "learning_rate": 0.00010905218317358894, "loss": 0.5157, "step": 2210 }, { "epoch": 4.17, "grad_norm": 3.0028767585754395, "learning_rate": 0.00010851970181043665, "loss": 0.5172, "step": 2220 }, { "epoch": 4.19, "grad_norm": 2.0464072227478027, "learning_rate": 0.00010798722044728435, "loss": 0.5099, "step": 2230 }, { "epoch": 4.21, "grad_norm": 2.517566204071045, "learning_rate": 0.00010745473908413206, "loss": 0.5021, "step": 2240 }, { "epoch": 4.23, "grad_norm": 1.1620187759399414, "learning_rate": 0.00010692225772097977, "loss": 0.4711, "step": 2250 }, { "epoch": 4.25, "grad_norm": 1.9912981986999512, "learning_rate": 0.00010638977635782747, "loss": 0.5725, "step": 2260 }, { "epoch": 4.27, "grad_norm": 0.9746074080467224, "learning_rate": 0.00010585729499467518, "loss": 0.5196, "step": 2270 }, { "epoch": 4.29, "grad_norm": 3.125993490219116, "learning_rate": 0.00010532481363152291, "loss": 0.55, "step": 2280 }, { "epoch": 4.3, "grad_norm": 1.6319783926010132, "learning_rate": 0.00010479233226837062, "loss": 0.5553, "step": 2290 }, { "epoch": 4.32, "grad_norm": 2.2027761936187744, "learning_rate": 0.00010425985090521832, "loss": 0.5072, "step": 2300 }, { "epoch": 4.34, "grad_norm": 3.756495475769043, "learning_rate": 0.00010372736954206603, "loss": 0.5979, "step": 2310 }, { "epoch": 4.36, "grad_norm": 0.9569351673126221, "learning_rate": 0.00010319488817891375, "loss": 0.5269, "step": 2320 }, { "epoch": 4.38, "grad_norm": 1.8240845203399658, "learning_rate": 0.00010266240681576144, "loss": 0.5554, "step": 2330 }, { "epoch": 4.4, "grad_norm": 1.4558720588684082, "learning_rate": 0.00010212992545260916, "loss": 0.4523, "step": 2340 }, { "epoch": 4.42, "grad_norm": 1.966886281967163, "learning_rate": 0.00010159744408945688, "loss": 0.5397, "step": 2350 }, { "epoch": 4.44, "grad_norm": 2.034513473510742, "learning_rate": 0.0001010649627263046, "loss": 0.4826, "step": 2360 }, { "epoch": 4.45, "grad_norm": 2.524667263031006, "learning_rate": 0.00010053248136315229, "loss": 0.4602, "step": 2370 }, { "epoch": 4.47, "grad_norm": 1.973412036895752, "learning_rate": 0.00010005324813631522, "loss": 0.5769, "step": 2380 }, { "epoch": 4.49, "grad_norm": 2.0014846324920654, "learning_rate": 9.952076677316294e-05, "loss": 0.464, "step": 2390 }, { "epoch": 4.51, "grad_norm": 1.190698266029358, "learning_rate": 9.898828541001066e-05, "loss": 0.4745, "step": 2400 }, { "epoch": 4.53, "grad_norm": 1.8334747552871704, "learning_rate": 9.845580404685837e-05, "loss": 0.4763, "step": 2410 }, { "epoch": 4.55, "grad_norm": 1.5026592016220093, "learning_rate": 9.792332268370608e-05, "loss": 0.4714, "step": 2420 }, { "epoch": 4.57, "grad_norm": 3.0638644695281982, "learning_rate": 9.739084132055378e-05, "loss": 0.3997, "step": 2430 }, { "epoch": 4.59, "grad_norm": 3.050144672393799, "learning_rate": 9.68583599574015e-05, "loss": 0.4471, "step": 2440 }, { "epoch": 4.61, "grad_norm": 2.4229140281677246, "learning_rate": 9.63258785942492e-05, "loss": 0.4742, "step": 2450 }, { "epoch": 4.62, "grad_norm": 2.2217772006988525, "learning_rate": 9.579339723109692e-05, "loss": 0.5514, "step": 2460 }, { "epoch": 4.64, "grad_norm": 2.5604331493377686, "learning_rate": 9.526091586794463e-05, "loss": 0.538, "step": 2470 }, { "epoch": 4.66, "grad_norm": 1.479228138923645, "learning_rate": 9.472843450479234e-05, "loss": 0.5559, "step": 2480 }, { "epoch": 4.68, "grad_norm": 1.5853182077407837, "learning_rate": 9.419595314164005e-05, "loss": 0.4786, "step": 2490 }, { "epoch": 4.7, "grad_norm": 4.546787261962891, "learning_rate": 9.366347177848775e-05, "loss": 0.5031, "step": 2500 }, { "epoch": 4.72, "grad_norm": 2.7307677268981934, "learning_rate": 9.313099041533548e-05, "loss": 0.4968, "step": 2510 }, { "epoch": 4.74, "grad_norm": 2.9583988189697266, "learning_rate": 9.259850905218317e-05, "loss": 0.5033, "step": 2520 }, { "epoch": 4.76, "grad_norm": 2.163015604019165, "learning_rate": 9.206602768903089e-05, "loss": 0.533, "step": 2530 }, { "epoch": 4.77, "grad_norm": 2.1034317016601562, "learning_rate": 9.15335463258786e-05, "loss": 0.4756, "step": 2540 }, { "epoch": 4.79, "grad_norm": 3.1241562366485596, "learning_rate": 9.100106496272631e-05, "loss": 0.5204, "step": 2550 }, { "epoch": 4.81, "grad_norm": 1.9084091186523438, "learning_rate": 9.046858359957402e-05, "loss": 0.4767, "step": 2560 }, { "epoch": 4.83, "grad_norm": 2.3724541664123535, "learning_rate": 8.993610223642172e-05, "loss": 0.4941, "step": 2570 }, { "epoch": 4.85, "grad_norm": 1.9276816844940186, "learning_rate": 8.940362087326945e-05, "loss": 0.5011, "step": 2580 }, { "epoch": 4.87, "grad_norm": 3.0274977684020996, "learning_rate": 8.887113951011715e-05, "loss": 0.4222, "step": 2590 }, { "epoch": 4.89, "grad_norm": 3.4712843894958496, "learning_rate": 8.833865814696486e-05, "loss": 0.4536, "step": 2600 }, { "epoch": 4.91, "grad_norm": 1.3534293174743652, "learning_rate": 8.780617678381257e-05, "loss": 0.4915, "step": 2610 }, { "epoch": 4.92, "grad_norm": 2.5444459915161133, "learning_rate": 8.727369542066028e-05, "loss": 0.5122, "step": 2620 }, { "epoch": 4.94, "grad_norm": 2.352358102798462, "learning_rate": 8.6741214057508e-05, "loss": 0.5088, "step": 2630 }, { "epoch": 4.96, "grad_norm": 2.807650327682495, "learning_rate": 8.620873269435569e-05, "loss": 0.5527, "step": 2640 }, { "epoch": 4.98, "grad_norm": 1.5027140378952026, "learning_rate": 8.567625133120342e-05, "loss": 0.5642, "step": 2650 }, { "epoch": 5.0, "grad_norm": 9.080309867858887, "learning_rate": 8.514376996805112e-05, "loss": 0.5545, "step": 2660 }, { "epoch": 5.0, "eval_loss": 0.9512230157852173, "eval_runtime": 5.1199, "eval_samples_per_second": 97.657, "eval_steps_per_second": 3.125, "step": 2660 }, { "epoch": 5.02, "grad_norm": 1.3162051439285278, "learning_rate": 8.461128860489883e-05, "loss": 0.405, "step": 2670 }, { "epoch": 5.04, "grad_norm": 2.260577440261841, "learning_rate": 8.407880724174654e-05, "loss": 0.3254, "step": 2680 }, { "epoch": 5.06, "grad_norm": 2.4929864406585693, "learning_rate": 8.354632587859425e-05, "loss": 0.4028, "step": 2690 }, { "epoch": 5.08, "grad_norm": 5.688662052154541, "learning_rate": 8.301384451544197e-05, "loss": 0.2906, "step": 2700 }, { "epoch": 5.09, "grad_norm": 1.3222452402114868, "learning_rate": 8.248136315228966e-05, "loss": 0.3261, "step": 2710 }, { "epoch": 5.11, "grad_norm": 1.3454262018203735, "learning_rate": 8.194888178913739e-05, "loss": 0.3013, "step": 2720 }, { "epoch": 5.13, "grad_norm": 3.683896064758301, "learning_rate": 8.141640042598509e-05, "loss": 0.3953, "step": 2730 }, { "epoch": 5.15, "grad_norm": 5.352532863616943, "learning_rate": 8.08839190628328e-05, "loss": 0.4251, "step": 2740 }, { "epoch": 5.17, "grad_norm": 2.004856586456299, "learning_rate": 8.035143769968051e-05, "loss": 0.3757, "step": 2750 }, { "epoch": 5.19, "grad_norm": 3.9098947048187256, "learning_rate": 7.981895633652823e-05, "loss": 0.3575, "step": 2760 }, { "epoch": 5.21, "grad_norm": 3.314652919769287, "learning_rate": 7.928647497337594e-05, "loss": 0.3759, "step": 2770 }, { "epoch": 5.23, "grad_norm": 5.735342979431152, "learning_rate": 7.875399361022364e-05, "loss": 0.346, "step": 2780 }, { "epoch": 5.24, "grad_norm": 1.5605205297470093, "learning_rate": 7.822151224707136e-05, "loss": 0.2185, "step": 2790 }, { "epoch": 5.26, "grad_norm": 4.125953197479248, "learning_rate": 7.768903088391907e-05, "loss": 0.4666, "step": 2800 }, { "epoch": 5.28, "grad_norm": 3.755160331726074, "learning_rate": 7.715654952076677e-05, "loss": 0.4115, "step": 2810 }, { "epoch": 5.3, "grad_norm": 3.822338104248047, "learning_rate": 7.66240681576145e-05, "loss": 0.3338, "step": 2820 }, { "epoch": 5.32, "grad_norm": 7.3369269371032715, "learning_rate": 7.60915867944622e-05, "loss": 0.4178, "step": 2830 }, { "epoch": 5.34, "grad_norm": 3.8311543464660645, "learning_rate": 7.555910543130991e-05, "loss": 0.3084, "step": 2840 }, { "epoch": 5.36, "grad_norm": 4.502466678619385, "learning_rate": 7.502662406815761e-05, "loss": 0.2962, "step": 2850 }, { "epoch": 5.38, "grad_norm": 3.489075183868408, "learning_rate": 7.449414270500533e-05, "loss": 0.3073, "step": 2860 }, { "epoch": 5.39, "grad_norm": 2.344810962677002, "learning_rate": 7.396166134185304e-05, "loss": 0.3059, "step": 2870 }, { "epoch": 5.41, "grad_norm": 4.37916374206543, "learning_rate": 7.342917997870074e-05, "loss": 0.3749, "step": 2880 }, { "epoch": 5.43, "grad_norm": 2.2266342639923096, "learning_rate": 7.289669861554847e-05, "loss": 0.2954, "step": 2890 }, { "epoch": 5.45, "grad_norm": 2.1551311016082764, "learning_rate": 7.236421725239617e-05, "loss": 0.3377, "step": 2900 }, { "epoch": 5.47, "grad_norm": 5.243711948394775, "learning_rate": 7.183173588924388e-05, "loss": 0.2986, "step": 2910 }, { "epoch": 5.49, "grad_norm": 3.0937271118164062, "learning_rate": 7.129925452609159e-05, "loss": 0.2355, "step": 2920 }, { "epoch": 5.51, "grad_norm": 1.6880041360855103, "learning_rate": 7.07667731629393e-05, "loss": 0.2808, "step": 2930 }, { "epoch": 5.53, "grad_norm": 2.438985586166382, "learning_rate": 7.023429179978702e-05, "loss": 0.4147, "step": 2940 }, { "epoch": 5.55, "grad_norm": 2.1790943145751953, "learning_rate": 6.970181043663471e-05, "loss": 0.3619, "step": 2950 }, { "epoch": 5.56, "grad_norm": 4.307677745819092, "learning_rate": 6.916932907348244e-05, "loss": 0.3657, "step": 2960 }, { "epoch": 5.58, "grad_norm": 4.3886566162109375, "learning_rate": 6.863684771033014e-05, "loss": 0.3333, "step": 2970 }, { "epoch": 5.6, "grad_norm": 2.579944610595703, "learning_rate": 6.810436634717785e-05, "loss": 0.4132, "step": 2980 }, { "epoch": 5.62, "grad_norm": 2.7568283081054688, "learning_rate": 6.757188498402556e-05, "loss": 0.2668, "step": 2990 }, { "epoch": 5.64, "grad_norm": 1.8786532878875732, "learning_rate": 6.703940362087328e-05, "loss": 0.3285, "step": 3000 }, { "epoch": 5.66, "grad_norm": 1.1324492692947388, "learning_rate": 6.650692225772099e-05, "loss": 0.3481, "step": 3010 }, { "epoch": 5.68, "grad_norm": 1.0097510814666748, "learning_rate": 6.597444089456869e-05, "loss": 0.2596, "step": 3020 }, { "epoch": 5.7, "grad_norm": 5.024123191833496, "learning_rate": 6.544195953141641e-05, "loss": 0.2748, "step": 3030 }, { "epoch": 5.71, "grad_norm": 4.416189193725586, "learning_rate": 6.490947816826411e-05, "loss": 0.3929, "step": 3040 }, { "epoch": 5.73, "grad_norm": 2.678957462310791, "learning_rate": 6.437699680511182e-05, "loss": 0.2433, "step": 3050 }, { "epoch": 5.75, "grad_norm": 2.553661346435547, "learning_rate": 6.384451544195953e-05, "loss": 0.2908, "step": 3060 }, { "epoch": 5.77, "grad_norm": 2.256429433822632, "learning_rate": 6.331203407880725e-05, "loss": 0.2666, "step": 3070 }, { "epoch": 5.79, "grad_norm": 5.489305019378662, "learning_rate": 6.277955271565496e-05, "loss": 0.3884, "step": 3080 }, { "epoch": 5.81, "grad_norm": 4.188615322113037, "learning_rate": 6.224707135250266e-05, "loss": 0.3637, "step": 3090 }, { "epoch": 5.83, "grad_norm": 3.3800997734069824, "learning_rate": 6.171458998935038e-05, "loss": 0.2935, "step": 3100 }, { "epoch": 5.85, "grad_norm": 1.9177793264389038, "learning_rate": 6.118210862619808e-05, "loss": 0.3361, "step": 3110 }, { "epoch": 5.86, "grad_norm": 1.3074955940246582, "learning_rate": 6.0649627263045794e-05, "loss": 0.2567, "step": 3120 }, { "epoch": 5.88, "grad_norm": 4.866466045379639, "learning_rate": 6.01171458998935e-05, "loss": 0.3597, "step": 3130 }, { "epoch": 5.9, "grad_norm": 2.4640395641326904, "learning_rate": 5.958466453674122e-05, "loss": 0.2919, "step": 3140 }, { "epoch": 5.92, "grad_norm": 4.659738540649414, "learning_rate": 5.9052183173588923e-05, "loss": 0.3253, "step": 3150 }, { "epoch": 5.94, "grad_norm": 3.3256399631500244, "learning_rate": 5.8519701810436636e-05, "loss": 0.3053, "step": 3160 }, { "epoch": 5.96, "grad_norm": 1.7987494468688965, "learning_rate": 5.7987220447284354e-05, "loss": 0.3421, "step": 3170 }, { "epoch": 5.98, "grad_norm": 3.481025218963623, "learning_rate": 5.745473908413206e-05, "loss": 0.2801, "step": 3180 }, { "epoch": 6.0, "grad_norm": 3.000406265258789, "learning_rate": 5.6922257720979765e-05, "loss": 0.3243, "step": 3190 }, { "epoch": 6.0, "eval_loss": 1.507144808769226, "eval_runtime": 4.9299, "eval_samples_per_second": 101.421, "eval_steps_per_second": 3.245, "step": 3192 }, { "epoch": 6.02, "grad_norm": 1.909064531326294, "learning_rate": 5.6389776357827484e-05, "loss": 0.195, "step": 3200 }, { "epoch": 6.03, "grad_norm": 1.058613896369934, "learning_rate": 5.585729499467519e-05, "loss": 0.1155, "step": 3210 }, { "epoch": 6.05, "grad_norm": 3.8664138317108154, "learning_rate": 5.53248136315229e-05, "loss": 0.2217, "step": 3220 }, { "epoch": 6.07, "grad_norm": 0.6019624471664429, "learning_rate": 5.479233226837061e-05, "loss": 0.1283, "step": 3230 }, { "epoch": 6.09, "grad_norm": 4.904506683349609, "learning_rate": 5.4259850905218326e-05, "loss": 0.1153, "step": 3240 }, { "epoch": 6.11, "grad_norm": 2.7440154552459717, "learning_rate": 5.372736954206603e-05, "loss": 0.0852, "step": 3250 }, { "epoch": 6.13, "grad_norm": 1.7608120441436768, "learning_rate": 5.3194888178913736e-05, "loss": 0.1631, "step": 3260 }, { "epoch": 6.15, "grad_norm": 0.817927360534668, "learning_rate": 5.2662406815761455e-05, "loss": 0.0899, "step": 3270 }, { "epoch": 6.17, "grad_norm": 0.4968046545982361, "learning_rate": 5.212992545260916e-05, "loss": 0.1774, "step": 3280 }, { "epoch": 6.18, "grad_norm": 3.789479970932007, "learning_rate": 5.159744408945687e-05, "loss": 0.1012, "step": 3290 }, { "epoch": 6.2, "grad_norm": 2.593512535095215, "learning_rate": 5.106496272630458e-05, "loss": 0.0886, "step": 3300 }, { "epoch": 6.22, "grad_norm": 1.9879395961761475, "learning_rate": 5.05324813631523e-05, "loss": 0.0865, "step": 3310 }, { "epoch": 6.24, "grad_norm": 0.262031614780426, "learning_rate": 5e-05, "loss": 0.1385, "step": 3320 }, { "epoch": 6.26, "grad_norm": 0.5335056781768799, "learning_rate": 4.9467518636847715e-05, "loss": 0.12, "step": 3330 }, { "epoch": 6.28, "grad_norm": 3.7105541229248047, "learning_rate": 4.893503727369542e-05, "loss": 0.2251, "step": 3340 }, { "epoch": 6.3, "grad_norm": 5.374883651733398, "learning_rate": 4.840255591054313e-05, "loss": 0.154, "step": 3350 }, { "epoch": 6.32, "grad_norm": 11.730541229248047, "learning_rate": 4.7870074547390844e-05, "loss": 0.2578, "step": 3360 }, { "epoch": 6.33, "grad_norm": 2.778977632522583, "learning_rate": 4.7337593184238556e-05, "loss": 0.1043, "step": 3370 }, { "epoch": 6.35, "grad_norm": 1.2093865871429443, "learning_rate": 4.680511182108626e-05, "loss": 0.1851, "step": 3380 }, { "epoch": 6.37, "grad_norm": 0.44236135482788086, "learning_rate": 4.6272630457933974e-05, "loss": 0.193, "step": 3390 }, { "epoch": 6.39, "grad_norm": 5.98400354385376, "learning_rate": 4.5740149094781686e-05, "loss": 0.1475, "step": 3400 }, { "epoch": 6.41, "grad_norm": 9.247739791870117, "learning_rate": 4.520766773162939e-05, "loss": 0.1972, "step": 3410 }, { "epoch": 6.43, "grad_norm": 0.7436373829841614, "learning_rate": 4.46751863684771e-05, "loss": 0.141, "step": 3420 }, { "epoch": 6.45, "grad_norm": 2.57186222076416, "learning_rate": 4.4142705005324815e-05, "loss": 0.1439, "step": 3430 }, { "epoch": 6.47, "grad_norm": 6.6432600021362305, "learning_rate": 4.361022364217253e-05, "loss": 0.1118, "step": 3440 }, { "epoch": 6.48, "grad_norm": 5.326605796813965, "learning_rate": 4.307774227902024e-05, "loss": 0.1989, "step": 3450 }, { "epoch": 6.5, "grad_norm": 4.456277370452881, "learning_rate": 4.2545260915867945e-05, "loss": 0.1541, "step": 3460 }, { "epoch": 6.52, "grad_norm": 0.17808414995670319, "learning_rate": 4.201277955271566e-05, "loss": 0.0874, "step": 3470 }, { "epoch": 6.54, "grad_norm": 14.033349990844727, "learning_rate": 4.148029818956336e-05, "loss": 0.1112, "step": 3480 }, { "epoch": 6.56, "grad_norm": 7.038208484649658, "learning_rate": 4.094781682641108e-05, "loss": 0.2453, "step": 3490 }, { "epoch": 6.58, "grad_norm": 1.385833501815796, "learning_rate": 4.041533546325879e-05, "loss": 0.134, "step": 3500 }, { "epoch": 6.6, "grad_norm": 3.095097303390503, "learning_rate": 3.98828541001065e-05, "loss": 0.1328, "step": 3510 }, { "epoch": 6.62, "grad_norm": 0.5865158438682556, "learning_rate": 3.935037273695421e-05, "loss": 0.1189, "step": 3520 }, { "epoch": 6.64, "grad_norm": 6.275442600250244, "learning_rate": 3.8817891373801916e-05, "loss": 0.2252, "step": 3530 }, { "epoch": 6.65, "grad_norm": 4.78378438949585, "learning_rate": 3.828541001064963e-05, "loss": 0.2046, "step": 3540 }, { "epoch": 6.67, "grad_norm": 0.8807236552238464, "learning_rate": 3.775292864749734e-05, "loss": 0.0828, "step": 3550 }, { "epoch": 6.69, "grad_norm": 0.18408401310443878, "learning_rate": 3.722044728434505e-05, "loss": 0.0516, "step": 3560 }, { "epoch": 6.71, "grad_norm": 8.150348663330078, "learning_rate": 3.668796592119276e-05, "loss": 0.2113, "step": 3570 }, { "epoch": 6.73, "grad_norm": 4.728875637054443, "learning_rate": 3.615548455804047e-05, "loss": 0.1763, "step": 3580 }, { "epoch": 6.75, "grad_norm": 8.548792839050293, "learning_rate": 3.562300319488818e-05, "loss": 0.1872, "step": 3590 }, { "epoch": 6.77, "grad_norm": 0.4689349830150604, "learning_rate": 3.509052183173589e-05, "loss": 0.0679, "step": 3600 }, { "epoch": 6.79, "grad_norm": 1.130393624305725, "learning_rate": 3.45580404685836e-05, "loss": 0.15, "step": 3610 }, { "epoch": 6.8, "grad_norm": 1.4335063695907593, "learning_rate": 3.402555910543131e-05, "loss": 0.1507, "step": 3620 }, { "epoch": 6.82, "grad_norm": 1.1329718828201294, "learning_rate": 3.3493077742279024e-05, "loss": 0.1569, "step": 3630 }, { "epoch": 6.84, "grad_norm": 0.7500413656234741, "learning_rate": 3.2960596379126736e-05, "loss": 0.1514, "step": 3640 }, { "epoch": 6.86, "grad_norm": 0.21835221350193024, "learning_rate": 3.242811501597444e-05, "loss": 0.1324, "step": 3650 }, { "epoch": 6.88, "grad_norm": 1.765376329421997, "learning_rate": 3.1895633652822154e-05, "loss": 0.0542, "step": 3660 }, { "epoch": 6.9, "grad_norm": 1.007416844367981, "learning_rate": 3.136315228966986e-05, "loss": 0.1132, "step": 3670 }, { "epoch": 6.92, "grad_norm": 0.4407300353050232, "learning_rate": 3.083067092651757e-05, "loss": 0.1133, "step": 3680 }, { "epoch": 6.94, "grad_norm": 3.9314990043640137, "learning_rate": 3.029818956336528e-05, "loss": 0.153, "step": 3690 }, { "epoch": 6.95, "grad_norm": 9.002704620361328, "learning_rate": 2.9765708200212995e-05, "loss": 0.1732, "step": 3700 }, { "epoch": 6.97, "grad_norm": 8.162276268005371, "learning_rate": 2.9233226837060707e-05, "loss": 0.1315, "step": 3710 }, { "epoch": 6.99, "grad_norm": 0.3330124318599701, "learning_rate": 2.8700745473908413e-05, "loss": 0.1371, "step": 3720 }, { "epoch": 7.0, "eval_loss": 2.0203089714050293, "eval_runtime": 4.9645, "eval_samples_per_second": 100.715, "eval_steps_per_second": 3.223, "step": 3724 }, { "epoch": 7.01, "grad_norm": 1.0700572729110718, "learning_rate": 2.8168264110756125e-05, "loss": 0.0909, "step": 3730 }, { "epoch": 7.03, "grad_norm": 0.4397372007369995, "learning_rate": 2.7635782747603834e-05, "loss": 0.0137, "step": 3740 }, { "epoch": 7.05, "grad_norm": 6.55562686920166, "learning_rate": 2.7103301384451546e-05, "loss": 0.0651, "step": 3750 }, { "epoch": 7.07, "grad_norm": 0.3805689215660095, "learning_rate": 2.6570820021299255e-05, "loss": 0.0863, "step": 3760 }, { "epoch": 7.09, "grad_norm": 3.0128049850463867, "learning_rate": 2.6038338658146967e-05, "loss": 0.0868, "step": 3770 }, { "epoch": 7.11, "grad_norm": 2.7534263134002686, "learning_rate": 2.550585729499468e-05, "loss": 0.0183, "step": 3780 }, { "epoch": 7.12, "grad_norm": 0.03968283161520958, "learning_rate": 2.4973375931842384e-05, "loss": 0.0045, "step": 3790 }, { "epoch": 7.14, "grad_norm": 11.932687759399414, "learning_rate": 2.44408945686901e-05, "loss": 0.0418, "step": 3800 }, { "epoch": 7.16, "grad_norm": 0.20918692648410797, "learning_rate": 2.390841320553781e-05, "loss": 0.0048, "step": 3810 }, { "epoch": 7.18, "grad_norm": 0.48621994256973267, "learning_rate": 2.3375931842385517e-05, "loss": 0.0285, "step": 3820 }, { "epoch": 7.2, "grad_norm": 0.008626734837889671, "learning_rate": 2.284345047923323e-05, "loss": 0.0382, "step": 3830 }, { "epoch": 7.22, "grad_norm": 12.129301071166992, "learning_rate": 2.2310969116080938e-05, "loss": 0.1381, "step": 3840 }, { "epoch": 7.24, "grad_norm": 3.3187191486358643, "learning_rate": 2.1778487752928647e-05, "loss": 0.034, "step": 3850 }, { "epoch": 7.26, "grad_norm": 19.58869171142578, "learning_rate": 2.124600638977636e-05, "loss": 0.078, "step": 3860 }, { "epoch": 7.27, "grad_norm": 0.011677253991365433, "learning_rate": 2.071352502662407e-05, "loss": 0.0278, "step": 3870 }, { "epoch": 7.29, "grad_norm": 0.0037781130522489548, "learning_rate": 2.018104366347178e-05, "loss": 0.019, "step": 3880 }, { "epoch": 7.31, "grad_norm": 0.012461444362998009, "learning_rate": 1.964856230031949e-05, "loss": 0.022, "step": 3890 }, { "epoch": 7.33, "grad_norm": 0.13421136140823364, "learning_rate": 1.91160809371672e-05, "loss": 0.0591, "step": 3900 }, { "epoch": 7.35, "grad_norm": 0.015879683196544647, "learning_rate": 1.858359957401491e-05, "loss": 0.0587, "step": 3910 }, { "epoch": 7.37, "grad_norm": 0.006961719132959843, "learning_rate": 1.805111821086262e-05, "loss": 0.0549, "step": 3920 }, { "epoch": 7.39, "grad_norm": 0.0046995761804282665, "learning_rate": 1.7518636847710333e-05, "loss": 0.0096, "step": 3930 }, { "epoch": 7.41, "grad_norm": 3.304826021194458, "learning_rate": 1.6986155484558042e-05, "loss": 0.083, "step": 3940 }, { "epoch": 7.42, "grad_norm": 0.3136584758758545, "learning_rate": 1.645367412140575e-05, "loss": 0.0191, "step": 3950 }, { "epoch": 7.44, "grad_norm": 2.098071575164795, "learning_rate": 1.5921192758253463e-05, "loss": 0.1322, "step": 3960 }, { "epoch": 7.46, "grad_norm": 0.0669000893831253, "learning_rate": 1.5388711395101172e-05, "loss": 0.0662, "step": 3970 }, { "epoch": 7.48, "grad_norm": 0.024517908692359924, "learning_rate": 1.485623003194888e-05, "loss": 0.0431, "step": 3980 }, { "epoch": 7.5, "grad_norm": 0.01538030430674553, "learning_rate": 1.4323748668796594e-05, "loss": 0.0201, "step": 3990 }, { "epoch": 7.52, "grad_norm": 0.019790129736065865, "learning_rate": 1.3791267305644303e-05, "loss": 0.0768, "step": 4000 }, { "epoch": 7.54, "grad_norm": 0.6820014119148254, "learning_rate": 1.3258785942492014e-05, "loss": 0.0222, "step": 4010 }, { "epoch": 7.56, "grad_norm": 0.027402225881814957, "learning_rate": 1.2726304579339724e-05, "loss": 0.0191, "step": 4020 }, { "epoch": 7.58, "grad_norm": 0.012008791789412498, "learning_rate": 1.2193823216187434e-05, "loss": 0.0137, "step": 4030 }, { "epoch": 7.59, "grad_norm": 10.289400100708008, "learning_rate": 1.1661341853035145e-05, "loss": 0.0738, "step": 4040 }, { "epoch": 7.61, "grad_norm": 0.019787069410085678, "learning_rate": 1.1128860489882854e-05, "loss": 0.0201, "step": 4050 }, { "epoch": 7.63, "grad_norm": 3.301687717437744, "learning_rate": 1.0596379126730564e-05, "loss": 0.0393, "step": 4060 }, { "epoch": 7.65, "grad_norm": 1.3089189529418945, "learning_rate": 1.0063897763578276e-05, "loss": 0.031, "step": 4070 }, { "epoch": 7.67, "grad_norm": 0.03497765213251114, "learning_rate": 9.531416400425985e-06, "loss": 0.159, "step": 4080 }, { "epoch": 7.69, "grad_norm": 0.23914293944835663, "learning_rate": 8.998935037273695e-06, "loss": 0.0757, "step": 4090 }, { "epoch": 7.71, "grad_norm": 0.011268123984336853, "learning_rate": 8.466453674121406e-06, "loss": 0.0051, "step": 4100 }, { "epoch": 7.73, "grad_norm": 0.021614540368318558, "learning_rate": 7.933972310969116e-06, "loss": 0.0606, "step": 4110 }, { "epoch": 7.74, "grad_norm": 2.5865726470947266, "learning_rate": 7.4014909478168266e-06, "loss": 0.0776, "step": 4120 }, { "epoch": 7.76, "grad_norm": 0.5956721305847168, "learning_rate": 6.869009584664538e-06, "loss": 0.0222, "step": 4130 }, { "epoch": 7.78, "grad_norm": 5.844602108001709, "learning_rate": 6.336528221512247e-06, "loss": 0.0293, "step": 4140 }, { "epoch": 7.8, "grad_norm": 0.03582174703478813, "learning_rate": 5.804046858359958e-06, "loss": 0.032, "step": 4150 }, { "epoch": 7.82, "grad_norm": 4.13277006149292, "learning_rate": 5.2715654952076674e-06, "loss": 0.0635, "step": 4160 }, { "epoch": 7.84, "grad_norm": 0.02260272204875946, "learning_rate": 4.739084132055379e-06, "loss": 0.0024, "step": 4170 }, { "epoch": 7.86, "grad_norm": 0.026146868243813515, "learning_rate": 4.206602768903089e-06, "loss": 0.0522, "step": 4180 }, { "epoch": 7.88, "grad_norm": 0.3275425434112549, "learning_rate": 3.6741214057507987e-06, "loss": 0.0492, "step": 4190 }, { "epoch": 7.89, "grad_norm": 0.07726357877254486, "learning_rate": 3.141640042598509e-06, "loss": 0.0205, "step": 4200 }, { "epoch": 7.91, "grad_norm": 0.6571511626243591, "learning_rate": 2.6091586794462196e-06, "loss": 0.0593, "step": 4210 }, { "epoch": 7.93, "grad_norm": 0.18334336578845978, "learning_rate": 2.0766773162939296e-06, "loss": 0.0052, "step": 4220 }, { "epoch": 7.95, "grad_norm": 0.056655995547771454, "learning_rate": 1.54419595314164e-06, "loss": 0.0362, "step": 4230 }, { "epoch": 7.97, "grad_norm": 0.3170097768306732, "learning_rate": 1.0117145899893504e-06, "loss": 0.0056, "step": 4240 }, { "epoch": 7.99, "grad_norm": 6.882564544677734, "learning_rate": 4.792332268370607e-07, "loss": 0.0571, "step": 4250 }, { "epoch": 8.0, "eval_loss": 2.9498751163482666, "eval_runtime": 5.0819, "eval_samples_per_second": 98.389, "eval_steps_per_second": 3.148, "step": 4256 }, { "epoch": 8.0, "step": 4256, "total_flos": 5.269455293792256e+18, "train_loss": 0.4405817036390921, "train_runtime": 1400.9279, "train_samples_per_second": 48.539, "train_steps_per_second": 3.038 } ], "logging_steps": 10, "max_steps": 4256, "num_input_tokens_seen": 0, "num_train_epochs": 8, "save_steps": 500, "total_flos": 5.269455293792256e+18, "train_batch_size": 16, "trial_name": null, "trial_params": null }