| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.18467220683287167, |
| "eval_steps": 500, |
| "global_step": 300, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0030778701138811943, |
| "grad_norm": 0.47482776641845703, |
| "learning_rate": 4.9999916942231176e-05, |
| "loss": 1.1992, |
| "num_input_tokens_seen": 42384, |
| "step": 5, |
| "train_runtime": 19.849, |
| "train_tokens_per_second": 2135.325 |
| }, |
| { |
| "epoch": 0.006155740227762389, |
| "grad_norm": 0.24747319519519806, |
| "learning_rate": 4.999957952099119e-05, |
| "loss": 1.0247, |
| "num_input_tokens_seen": 86000, |
| "step": 10, |
| "train_runtime": 39.0811, |
| "train_tokens_per_second": 2200.554 |
| }, |
| { |
| "epoch": 0.009233610341643583, |
| "grad_norm": 0.10441822558641434, |
| "learning_rate": 4.9998982548669973e-05, |
| "loss": 0.8631, |
| "num_input_tokens_seen": 124784, |
| "step": 15, |
| "train_runtime": 57.1682, |
| "train_tokens_per_second": 2182.751 |
| }, |
| { |
| "epoch": 0.012311480455524777, |
| "grad_norm": 0.11058320850133896, |
| "learning_rate": 4.999812603146542e-05, |
| "loss": 0.9173, |
| "num_input_tokens_seen": 165952, |
| "step": 20, |
| "train_runtime": 75.5733, |
| "train_tokens_per_second": 2195.909 |
| }, |
| { |
| "epoch": 0.015389350569405972, |
| "grad_norm": 0.09879708290100098, |
| "learning_rate": 4.9997009978270085e-05, |
| "loss": 0.8722, |
| "num_input_tokens_seen": 201584, |
| "step": 25, |
| "train_runtime": 92.5588, |
| "train_tokens_per_second": 2177.903 |
| }, |
| { |
| "epoch": 0.018467220683287166, |
| "grad_norm": 0.11391397565603256, |
| "learning_rate": 4.999563440067108e-05, |
| "loss": 0.866, |
| "num_input_tokens_seen": 240656, |
| "step": 30, |
| "train_runtime": 111.1373, |
| "train_tokens_per_second": 2165.393 |
| }, |
| { |
| "epoch": 0.02154509079716836, |
| "grad_norm": 0.09722574800252914, |
| "learning_rate": 4.9993999312949936e-05, |
| "loss": 0.853, |
| "num_input_tokens_seen": 287248, |
| "step": 35, |
| "train_runtime": 130.6684, |
| "train_tokens_per_second": 2198.297 |
| }, |
| { |
| "epoch": 0.024622960911049555, |
| "grad_norm": 0.12133017182350159, |
| "learning_rate": 4.99921047320825e-05, |
| "loss": 0.88, |
| "num_input_tokens_seen": 328448, |
| "step": 40, |
| "train_runtime": 149.193, |
| "train_tokens_per_second": 2201.498 |
| }, |
| { |
| "epoch": 0.027700831024930747, |
| "grad_norm": 0.12174953520298004, |
| "learning_rate": 4.9989950677738726e-05, |
| "loss": 0.8503, |
| "num_input_tokens_seen": 370512, |
| "step": 45, |
| "train_runtime": 168.2236, |
| "train_tokens_per_second": 2202.497 |
| }, |
| { |
| "epoch": 0.030778701138811943, |
| "grad_norm": 0.12440948933362961, |
| "learning_rate": 4.9987537172282466e-05, |
| "loss": 0.8707, |
| "num_input_tokens_seen": 409936, |
| "step": 50, |
| "train_runtime": 186.7223, |
| "train_tokens_per_second": 2195.432 |
| }, |
| { |
| "epoch": 0.03385657125269314, |
| "grad_norm": 0.12797962129116058, |
| "learning_rate": 4.998486424077127e-05, |
| "loss": 0.8475, |
| "num_input_tokens_seen": 450928, |
| "step": 55, |
| "train_runtime": 205.1107, |
| "train_tokens_per_second": 2198.462 |
| }, |
| { |
| "epoch": 0.03693444136657433, |
| "grad_norm": 0.09793120622634888, |
| "learning_rate": 4.9981931910956086e-05, |
| "loss": 0.8409, |
| "num_input_tokens_seen": 491152, |
| "step": 60, |
| "train_runtime": 223.3671, |
| "train_tokens_per_second": 2198.856 |
| }, |
| { |
| "epoch": 0.040012311480455524, |
| "grad_norm": 0.10957301408052444, |
| "learning_rate": 4.9978740213281005e-05, |
| "loss": 0.8524, |
| "num_input_tokens_seen": 532160, |
| "step": 65, |
| "train_runtime": 241.5682, |
| "train_tokens_per_second": 2202.939 |
| }, |
| { |
| "epoch": 0.04309018159433672, |
| "grad_norm": 0.12283050268888474, |
| "learning_rate": 4.9975289180882936e-05, |
| "loss": 0.8944, |
| "num_input_tokens_seen": 573392, |
| "step": 70, |
| "train_runtime": 259.8567, |
| "train_tokens_per_second": 2206.57 |
| }, |
| { |
| "epoch": 0.046168051708217916, |
| "grad_norm": 0.1338459849357605, |
| "learning_rate": 4.9971578849591225e-05, |
| "loss": 0.8478, |
| "num_input_tokens_seen": 613488, |
| "step": 75, |
| "train_runtime": 278.0456, |
| "train_tokens_per_second": 2206.429 |
| }, |
| { |
| "epoch": 0.04924592182209911, |
| "grad_norm": 0.10969562828540802, |
| "learning_rate": 4.996760925792737e-05, |
| "loss": 0.8331, |
| "num_input_tokens_seen": 656480, |
| "step": 80, |
| "train_runtime": 296.3508, |
| "train_tokens_per_second": 2215.212 |
| }, |
| { |
| "epoch": 0.0523237919359803, |
| "grad_norm": 0.1411936730146408, |
| "learning_rate": 4.996338044710452e-05, |
| "loss": 0.8306, |
| "num_input_tokens_seen": 698272, |
| "step": 85, |
| "train_runtime": 315.096, |
| "train_tokens_per_second": 2216.061 |
| }, |
| { |
| "epoch": 0.055401662049861494, |
| "grad_norm": 0.1435951143503189, |
| "learning_rate": 4.995889246102711e-05, |
| "loss": 0.8318, |
| "num_input_tokens_seen": 735376, |
| "step": 90, |
| "train_runtime": 332.4774, |
| "train_tokens_per_second": 2211.807 |
| }, |
| { |
| "epoch": 0.05847953216374269, |
| "grad_norm": 0.15133805572986603, |
| "learning_rate": 4.995414534629039e-05, |
| "loss": 0.8674, |
| "num_input_tokens_seen": 772544, |
| "step": 95, |
| "train_runtime": 350.9252, |
| "train_tokens_per_second": 2201.449 |
| }, |
| { |
| "epoch": 0.061557402277623886, |
| "grad_norm": 0.10615513473749161, |
| "learning_rate": 4.9949139152179944e-05, |
| "loss": 0.901, |
| "num_input_tokens_seen": 814928, |
| "step": 100, |
| "train_runtime": 369.5452, |
| "train_tokens_per_second": 2205.219 |
| }, |
| { |
| "epoch": 0.06463527239150507, |
| "grad_norm": 0.10441838204860687, |
| "learning_rate": 4.994387393067117e-05, |
| "loss": 0.8522, |
| "num_input_tokens_seen": 855104, |
| "step": 105, |
| "train_runtime": 388.7482, |
| "train_tokens_per_second": 2199.635 |
| }, |
| { |
| "epoch": 0.06771314250538628, |
| "grad_norm": 0.1405644714832306, |
| "learning_rate": 4.993834973642874e-05, |
| "loss": 0.8614, |
| "num_input_tokens_seen": 896272, |
| "step": 110, |
| "train_runtime": 406.8129, |
| "train_tokens_per_second": 2203.155 |
| }, |
| { |
| "epoch": 0.07079101261926747, |
| "grad_norm": 0.11736899614334106, |
| "learning_rate": 4.993256662680604e-05, |
| "loss": 0.8371, |
| "num_input_tokens_seen": 940864, |
| "step": 115, |
| "train_runtime": 425.836, |
| "train_tokens_per_second": 2209.452 |
| }, |
| { |
| "epoch": 0.07386888273314866, |
| "grad_norm": 0.1269819736480713, |
| "learning_rate": 4.992652466184458e-05, |
| "loss": 0.9, |
| "num_input_tokens_seen": 983360, |
| "step": 120, |
| "train_runtime": 444.5802, |
| "train_tokens_per_second": 2211.884 |
| }, |
| { |
| "epoch": 0.07694675284702986, |
| "grad_norm": 0.10944321751594543, |
| "learning_rate": 4.992022390427335e-05, |
| "loss": 0.8629, |
| "num_input_tokens_seen": 1025120, |
| "step": 125, |
| "train_runtime": 463.3805, |
| "train_tokens_per_second": 2212.264 |
| }, |
| { |
| "epoch": 0.08002462296091105, |
| "grad_norm": 0.1438760608434677, |
| "learning_rate": 4.99136644195082e-05, |
| "loss": 0.8797, |
| "num_input_tokens_seen": 1065360, |
| "step": 130, |
| "train_runtime": 480.9152, |
| "train_tokens_per_second": 2215.276 |
| }, |
| { |
| "epoch": 0.08310249307479224, |
| "grad_norm": 0.13859188556671143, |
| "learning_rate": 4.9906846275651125e-05, |
| "loss": 0.8317, |
| "num_input_tokens_seen": 1106112, |
| "step": 135, |
| "train_runtime": 498.8557, |
| "train_tokens_per_second": 2217.299 |
| }, |
| { |
| "epoch": 0.08618036318867343, |
| "grad_norm": 0.1412304937839508, |
| "learning_rate": 4.989976954348958e-05, |
| "loss": 0.8422, |
| "num_input_tokens_seen": 1149632, |
| "step": 140, |
| "train_runtime": 517.2481, |
| "train_tokens_per_second": 2222.593 |
| }, |
| { |
| "epoch": 0.08925823330255463, |
| "grad_norm": 0.1513959765434265, |
| "learning_rate": 4.989243429649573e-05, |
| "loss": 0.8403, |
| "num_input_tokens_seen": 1187824, |
| "step": 145, |
| "train_runtime": 535.2196, |
| "train_tokens_per_second": 2219.321 |
| }, |
| { |
| "epoch": 0.09233610341643583, |
| "grad_norm": 0.13685107231140137, |
| "learning_rate": 4.988484061082572e-05, |
| "loss": 0.8961, |
| "num_input_tokens_seen": 1229088, |
| "step": 150, |
| "train_runtime": 553.4777, |
| "train_tokens_per_second": 2220.664 |
| }, |
| { |
| "epoch": 0.09541397353031703, |
| "grad_norm": 0.12999333441257477, |
| "learning_rate": 4.987698856531884e-05, |
| "loss": 0.814, |
| "num_input_tokens_seen": 1273056, |
| "step": 155, |
| "train_runtime": 572.9621, |
| "train_tokens_per_second": 2221.885 |
| }, |
| { |
| "epoch": 0.09849184364419822, |
| "grad_norm": 0.1481194943189621, |
| "learning_rate": 4.986887824149674e-05, |
| "loss": 0.8542, |
| "num_input_tokens_seen": 1310288, |
| "step": 160, |
| "train_runtime": 589.8405, |
| "train_tokens_per_second": 2221.427 |
| }, |
| { |
| "epoch": 0.10156971375807941, |
| "grad_norm": 0.1396579146385193, |
| "learning_rate": 4.9860509723562573e-05, |
| "loss": 0.8722, |
| "num_input_tokens_seen": 1351312, |
| "step": 165, |
| "train_runtime": 607.8908, |
| "train_tokens_per_second": 2222.952 |
| }, |
| { |
| "epoch": 0.1046475838719606, |
| "grad_norm": 0.12688206136226654, |
| "learning_rate": 4.985188309840012e-05, |
| "loss": 0.8587, |
| "num_input_tokens_seen": 1396336, |
| "step": 170, |
| "train_runtime": 626.8902, |
| "train_tokens_per_second": 2227.401 |
| }, |
| { |
| "epoch": 0.1077254539858418, |
| "grad_norm": 0.13825534284114838, |
| "learning_rate": 4.984299845557287e-05, |
| "loss": 0.8014, |
| "num_input_tokens_seen": 1437984, |
| "step": 175, |
| "train_runtime": 645.2717, |
| "train_tokens_per_second": 2228.494 |
| }, |
| { |
| "epoch": 0.11080332409972299, |
| "grad_norm": 0.12385495752096176, |
| "learning_rate": 4.983385588732312e-05, |
| "loss": 0.8385, |
| "num_input_tokens_seen": 1475360, |
| "step": 180, |
| "train_runtime": 662.6866, |
| "train_tokens_per_second": 2226.332 |
| }, |
| { |
| "epoch": 0.11388119421360418, |
| "grad_norm": 0.12246419489383698, |
| "learning_rate": 4.982445548857102e-05, |
| "loss": 0.8372, |
| "num_input_tokens_seen": 1513152, |
| "step": 185, |
| "train_runtime": 680.3316, |
| "train_tokens_per_second": 2224.139 |
| }, |
| { |
| "epoch": 0.11695906432748537, |
| "grad_norm": 0.12281708419322968, |
| "learning_rate": 4.981479735691354e-05, |
| "loss": 0.8446, |
| "num_input_tokens_seen": 1556256, |
| "step": 190, |
| "train_runtime": 698.7652, |
| "train_tokens_per_second": 2227.152 |
| }, |
| { |
| "epoch": 0.12003693444136658, |
| "grad_norm": 0.13810615241527557, |
| "learning_rate": 4.980488159262353e-05, |
| "loss": 0.8141, |
| "num_input_tokens_seen": 1598832, |
| "step": 195, |
| "train_runtime": 717.4389, |
| "train_tokens_per_second": 2228.527 |
| }, |
| { |
| "epoch": 0.12311480455524777, |
| "grad_norm": 0.15435761213302612, |
| "learning_rate": 4.97947082986486e-05, |
| "loss": 0.8398, |
| "num_input_tokens_seen": 1641712, |
| "step": 200, |
| "train_runtime": 736.4476, |
| "train_tokens_per_second": 2229.231 |
| }, |
| { |
| "epoch": 0.12619267466912895, |
| "grad_norm": 0.15266934037208557, |
| "learning_rate": 4.978427758061009e-05, |
| "loss": 0.8714, |
| "num_input_tokens_seen": 1684544, |
| "step": 205, |
| "train_runtime": 756.7894, |
| "train_tokens_per_second": 2225.908 |
| }, |
| { |
| "epoch": 0.12927054478301014, |
| "grad_norm": 0.13250574469566345, |
| "learning_rate": 4.977358954680202e-05, |
| "loss": 0.8116, |
| "num_input_tokens_seen": 1723120, |
| "step": 210, |
| "train_runtime": 774.2541, |
| "train_tokens_per_second": 2225.523 |
| }, |
| { |
| "epoch": 0.13234841489689136, |
| "grad_norm": 0.11579591035842896, |
| "learning_rate": 4.9762644308189865e-05, |
| "loss": 0.8204, |
| "num_input_tokens_seen": 1768096, |
| "step": 215, |
| "train_runtime": 793.0535, |
| "train_tokens_per_second": 2229.479 |
| }, |
| { |
| "epoch": 0.13542628501077256, |
| "grad_norm": 0.12654677033424377, |
| "learning_rate": 4.9751441978409484e-05, |
| "loss": 0.8334, |
| "num_input_tokens_seen": 1808176, |
| "step": 220, |
| "train_runtime": 810.905, |
| "train_tokens_per_second": 2229.825 |
| }, |
| { |
| "epoch": 0.13850415512465375, |
| "grad_norm": 0.15607228875160217, |
| "learning_rate": 4.97399826737659e-05, |
| "loss": 0.8177, |
| "num_input_tokens_seen": 1848480, |
| "step": 225, |
| "train_runtime": 829.3126, |
| "train_tokens_per_second": 2228.93 |
| }, |
| { |
| "epoch": 0.14158202523853494, |
| "grad_norm": 0.1535269021987915, |
| "learning_rate": 4.972826651323211e-05, |
| "loss": 0.807, |
| "num_input_tokens_seen": 1889168, |
| "step": 230, |
| "train_runtime": 847.0449, |
| "train_tokens_per_second": 2230.304 |
| }, |
| { |
| "epoch": 0.14465989535241613, |
| "grad_norm": 0.13663041591644287, |
| "learning_rate": 4.971629361844785e-05, |
| "loss": 0.8557, |
| "num_input_tokens_seen": 1930784, |
| "step": 235, |
| "train_runtime": 865.4457, |
| "train_tokens_per_second": 2230.971 |
| }, |
| { |
| "epoch": 0.14773776546629733, |
| "grad_norm": 0.1719619184732437, |
| "learning_rate": 4.9704064113718326e-05, |
| "loss": 0.8436, |
| "num_input_tokens_seen": 1974176, |
| "step": 240, |
| "train_runtime": 883.9962, |
| "train_tokens_per_second": 2233.24 |
| }, |
| { |
| "epoch": 0.15081563558017852, |
| "grad_norm": 0.15830345451831818, |
| "learning_rate": 4.9691578126012905e-05, |
| "loss": 0.853, |
| "num_input_tokens_seen": 2013008, |
| "step": 245, |
| "train_runtime": 902.1681, |
| "train_tokens_per_second": 2231.3 |
| }, |
| { |
| "epoch": 0.1538935056940597, |
| "grad_norm": 0.14472956955432892, |
| "learning_rate": 4.967883578496385e-05, |
| "loss": 0.874, |
| "num_input_tokens_seen": 2053328, |
| "step": 250, |
| "train_runtime": 919.7822, |
| "train_tokens_per_second": 2232.407 |
| }, |
| { |
| "epoch": 0.1569713758079409, |
| "grad_norm": 0.13329900801181793, |
| "learning_rate": 4.966583722286491e-05, |
| "loss": 0.8333, |
| "num_input_tokens_seen": 2094992, |
| "step": 255, |
| "train_runtime": 937.9537, |
| "train_tokens_per_second": 2233.577 |
| }, |
| { |
| "epoch": 0.1600492459218221, |
| "grad_norm": 0.1551639437675476, |
| "learning_rate": 4.9652582574669995e-05, |
| "loss": 0.864, |
| "num_input_tokens_seen": 2135376, |
| "step": 260, |
| "train_runtime": 955.8955, |
| "train_tokens_per_second": 2233.901 |
| }, |
| { |
| "epoch": 0.1631271160357033, |
| "grad_norm": 0.12460612505674362, |
| "learning_rate": 4.963907197799174e-05, |
| "loss": 0.8321, |
| "num_input_tokens_seen": 2176624, |
| "step": 265, |
| "train_runtime": 974.3283, |
| "train_tokens_per_second": 2233.974 |
| }, |
| { |
| "epoch": 0.16620498614958448, |
| "grad_norm": 0.13981439173221588, |
| "learning_rate": 4.9625305573100115e-05, |
| "loss": 0.8349, |
| "num_input_tokens_seen": 2220496, |
| "step": 270, |
| "train_runtime": 992.5551, |
| "train_tokens_per_second": 2237.151 |
| }, |
| { |
| "epoch": 0.16928285626346568, |
| "grad_norm": 0.15424658358097076, |
| "learning_rate": 4.961128350292094e-05, |
| "loss": 0.8248, |
| "num_input_tokens_seen": 2263808, |
| "step": 275, |
| "train_runtime": 1010.9218, |
| "train_tokens_per_second": 2239.35 |
| }, |
| { |
| "epoch": 0.17236072637734687, |
| "grad_norm": 0.14288848638534546, |
| "learning_rate": 4.959700591303438e-05, |
| "loss": 0.8488, |
| "num_input_tokens_seen": 2311696, |
| "step": 280, |
| "train_runtime": 1030.8143, |
| "train_tokens_per_second": 2242.592 |
| }, |
| { |
| "epoch": 0.17543859649122806, |
| "grad_norm": 0.15639737248420715, |
| "learning_rate": 4.95824729516735e-05, |
| "loss": 0.8602, |
| "num_input_tokens_seen": 2353200, |
| "step": 285, |
| "train_runtime": 1049.2401, |
| "train_tokens_per_second": 2242.766 |
| }, |
| { |
| "epoch": 0.17851646660510925, |
| "grad_norm": 0.1268603354692459, |
| "learning_rate": 4.9567684769722664e-05, |
| "loss": 0.8987, |
| "num_input_tokens_seen": 2396000, |
| "step": 290, |
| "train_runtime": 1067.7324, |
| "train_tokens_per_second": 2244.008 |
| }, |
| { |
| "epoch": 0.18159433671899045, |
| "grad_norm": 0.14782211184501648, |
| "learning_rate": 4.9552641520716014e-05, |
| "loss": 0.7794, |
| "num_input_tokens_seen": 2438032, |
| "step": 295, |
| "train_runtime": 1085.9257, |
| "train_tokens_per_second": 2245.119 |
| }, |
| { |
| "epoch": 0.18467220683287167, |
| "grad_norm": 0.15259359776973724, |
| "learning_rate": 4.953734336083583e-05, |
| "loss": 0.8648, |
| "num_input_tokens_seen": 2478048, |
| "step": 300, |
| "train_runtime": 1103.5538, |
| "train_tokens_per_second": 2245.516 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 4875, |
| "num_input_tokens_seen": 2478048, |
| "num_train_epochs": 3, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 4.765645728959693e+17, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|