| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.21384084894817032, |
| "eval_steps": 500, |
| "global_step": 2000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0010692042447408515, |
| "grad_norm": 0.4522557854652405, |
| "learning_rate": 0.00018, |
| "loss": 2.4398, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.002138408489481703, |
| "grad_norm": 0.5223538875579834, |
| "learning_rate": 0.0001990954773869347, |
| "loss": 2.1227, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.003207612734222555, |
| "grad_norm": 0.47437742352485657, |
| "learning_rate": 0.00019809045226130653, |
| "loss": 1.9881, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.004276816978963406, |
| "grad_norm": 0.3975315988063812, |
| "learning_rate": 0.0001970854271356784, |
| "loss": 2.0059, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.0053460212237042585, |
| "grad_norm": 0.29887351393699646, |
| "learning_rate": 0.00019608040201005026, |
| "loss": 1.9811, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.00641522546844511, |
| "grad_norm": 0.3441367745399475, |
| "learning_rate": 0.00019507537688442212, |
| "loss": 1.9828, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.007484429713185962, |
| "grad_norm": 0.2983023226261139, |
| "learning_rate": 0.00019407035175879398, |
| "loss": 2.0323, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.008553633957926812, |
| "grad_norm": 0.30667832493782043, |
| "learning_rate": 0.00019306532663316582, |
| "loss": 2.0422, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.009622838202667664, |
| "grad_norm": 0.3337627947330475, |
| "learning_rate": 0.0001920603015075377, |
| "loss": 1.9896, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.010692042447408517, |
| "grad_norm": 0.3218510150909424, |
| "learning_rate": 0.00019105527638190955, |
| "loss": 1.976, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.011761246692149369, |
| "grad_norm": 0.3417668044567108, |
| "learning_rate": 0.00019005025125628142, |
| "loss": 1.9818, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.01283045093689022, |
| "grad_norm": 0.2790803015232086, |
| "learning_rate": 0.00018904522613065328, |
| "loss": 2.0031, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.013899655181631072, |
| "grad_norm": 0.4081987142562866, |
| "learning_rate": 0.00018804020100502512, |
| "loss": 1.9905, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.014968859426371923, |
| "grad_norm": 0.36898791790008545, |
| "learning_rate": 0.00018703517587939698, |
| "loss": 2.0049, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.016038063671112775, |
| "grad_norm": 0.29053226113319397, |
| "learning_rate": 0.00018603015075376885, |
| "loss": 1.9797, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.017107267915853624, |
| "grad_norm": 0.25266537070274353, |
| "learning_rate": 0.00018502512562814071, |
| "loss": 1.9509, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.018176472160594478, |
| "grad_norm": 0.2833998501300812, |
| "learning_rate": 0.00018402010050251258, |
| "loss": 1.9574, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.019245676405335328, |
| "grad_norm": 0.5657384395599365, |
| "learning_rate": 0.00018301507537688442, |
| "loss": 2.0581, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.02031488065007618, |
| "grad_norm": 0.29206541180610657, |
| "learning_rate": 0.00018201005025125628, |
| "loss": 1.8847, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.021384084894817034, |
| "grad_norm": 0.33557596802711487, |
| "learning_rate": 0.00018100502512562815, |
| "loss": 2.0308, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.022453289139557884, |
| "grad_norm": 0.3242787718772888, |
| "learning_rate": 0.00018, |
| "loss": 2.0498, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.023522493384298737, |
| "grad_norm": 0.37068045139312744, |
| "learning_rate": 0.00017899497487437188, |
| "loss": 1.9788, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.024591697629039587, |
| "grad_norm": 0.3749759793281555, |
| "learning_rate": 0.00017798994974874371, |
| "loss": 1.9289, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.02566090187378044, |
| "grad_norm": 0.3888619840145111, |
| "learning_rate": 0.00017698492462311558, |
| "loss": 1.9179, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.02673010611852129, |
| "grad_norm": 0.377247154712677, |
| "learning_rate": 0.00017597989949748744, |
| "loss": 2.0299, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.027799310363262143, |
| "grad_norm": 0.49286213517189026, |
| "learning_rate": 0.0001749748743718593, |
| "loss": 1.9895, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.028868514608002993, |
| "grad_norm": 0.30507227778434753, |
| "learning_rate": 0.00017396984924623117, |
| "loss": 1.9596, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.029937718852743846, |
| "grad_norm": 0.3705558180809021, |
| "learning_rate": 0.000172964824120603, |
| "loss": 2.0357, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.031006923097484696, |
| "grad_norm": 0.4542909562587738, |
| "learning_rate": 0.00017195979899497488, |
| "loss": 1.9323, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.03207612734222555, |
| "grad_norm": 0.3561468720436096, |
| "learning_rate": 0.00017095477386934674, |
| "loss": 1.9695, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.0331453315869664, |
| "grad_norm": 0.39809998869895935, |
| "learning_rate": 0.0001699497487437186, |
| "loss": 2.01, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.03421453583170725, |
| "grad_norm": 0.279765784740448, |
| "learning_rate": 0.00016894472361809047, |
| "loss": 1.9361, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.035283740076448106, |
| "grad_norm": 0.30650126934051514, |
| "learning_rate": 0.0001679396984924623, |
| "loss": 1.9411, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.036352944321188956, |
| "grad_norm": 0.33605650067329407, |
| "learning_rate": 0.00016693467336683417, |
| "loss": 1.9908, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.037422148565929805, |
| "grad_norm": 0.3010708689689636, |
| "learning_rate": 0.00016592964824120604, |
| "loss": 1.9326, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.038491352810670655, |
| "grad_norm": 0.33158665895462036, |
| "learning_rate": 0.0001649246231155779, |
| "loss": 1.9362, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.03956055705541151, |
| "grad_norm": 0.2630426585674286, |
| "learning_rate": 0.00016391959798994977, |
| "loss": 1.9345, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.04062976130015236, |
| "grad_norm": 0.2744496762752533, |
| "learning_rate": 0.0001629145728643216, |
| "loss": 1.916, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.04169896554489321, |
| "grad_norm": 0.33794867992401123, |
| "learning_rate": 0.00016190954773869347, |
| "loss": 1.9082, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.04276816978963407, |
| "grad_norm": 0.2850714921951294, |
| "learning_rate": 0.00016090452261306533, |
| "loss": 1.9752, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.04383737403437492, |
| "grad_norm": 0.34638258814811707, |
| "learning_rate": 0.0001598994974874372, |
| "loss": 1.8934, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.04490657827911577, |
| "grad_norm": 0.2859291434288025, |
| "learning_rate": 0.00015889447236180906, |
| "loss": 1.9893, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.04597578252385662, |
| "grad_norm": 0.31361544132232666, |
| "learning_rate": 0.0001578894472361809, |
| "loss": 1.9233, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.047044986768597474, |
| "grad_norm": 0.3515094816684723, |
| "learning_rate": 0.00015688442211055277, |
| "loss": 1.9668, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.048114191013338324, |
| "grad_norm": 0.3527173399925232, |
| "learning_rate": 0.00015587939698492463, |
| "loss": 1.9771, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.049183395258079174, |
| "grad_norm": 0.289266973733902, |
| "learning_rate": 0.0001548743718592965, |
| "loss": 2.0094, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.050252599502820024, |
| "grad_norm": 0.30293336510658264, |
| "learning_rate": 0.00015386934673366836, |
| "loss": 1.9822, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.05132180374756088, |
| "grad_norm": 0.31669914722442627, |
| "learning_rate": 0.0001528643216080402, |
| "loss": 1.9341, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.05239100799230173, |
| "grad_norm": 0.3566587567329407, |
| "learning_rate": 0.00015185929648241206, |
| "loss": 1.8382, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.05346021223704258, |
| "grad_norm": 0.2622707188129425, |
| "learning_rate": 0.00015085427135678393, |
| "loss": 1.9165, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.05452941648178343, |
| "grad_norm": 0.41455262899398804, |
| "learning_rate": 0.0001498492462311558, |
| "loss": 1.9656, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.05559862072652429, |
| "grad_norm": 0.37541866302490234, |
| "learning_rate": 0.00014884422110552766, |
| "loss": 1.9291, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.056667824971265136, |
| "grad_norm": 0.36062073707580566, |
| "learning_rate": 0.0001478391959798995, |
| "loss": 1.9518, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.057737029216005986, |
| "grad_norm": 0.30928629636764526, |
| "learning_rate": 0.00014683417085427136, |
| "loss": 1.8933, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.058806233460746836, |
| "grad_norm": 0.29138654470443726, |
| "learning_rate": 0.00014582914572864323, |
| "loss": 1.9848, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.05987543770548769, |
| "grad_norm": 0.259957492351532, |
| "learning_rate": 0.0001448241206030151, |
| "loss": 1.8802, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.06094464195022854, |
| "grad_norm": 0.40919119119644165, |
| "learning_rate": 0.00014381909547738696, |
| "loss": 1.9503, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.06201384619496939, |
| "grad_norm": 0.323599249124527, |
| "learning_rate": 0.0001428140703517588, |
| "loss": 1.9125, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.06308305043971024, |
| "grad_norm": 0.31641989946365356, |
| "learning_rate": 0.00014180904522613066, |
| "loss": 1.8805, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.0641522546844511, |
| "grad_norm": 0.4122176766395569, |
| "learning_rate": 0.00014080402010050252, |
| "loss": 1.8804, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.06522145892919196, |
| "grad_norm": 0.3238036334514618, |
| "learning_rate": 0.0001397989949748744, |
| "loss": 1.9645, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.0662906631739328, |
| "grad_norm": 0.3590157926082611, |
| "learning_rate": 0.00013879396984924625, |
| "loss": 1.9316, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.06735986741867366, |
| "grad_norm": 0.3035522997379303, |
| "learning_rate": 0.0001377889447236181, |
| "loss": 1.9626, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.0684290716634145, |
| "grad_norm": 0.3481757640838623, |
| "learning_rate": 0.00013678391959798996, |
| "loss": 1.9469, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.06949827590815535, |
| "grad_norm": 0.41229742765426636, |
| "learning_rate": 0.0001357788944723618, |
| "loss": 2.0198, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.07056748015289621, |
| "grad_norm": 0.5381227731704712, |
| "learning_rate": 0.00013477386934673368, |
| "loss": 1.9417, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.07163668439763705, |
| "grad_norm": 0.3313769996166229, |
| "learning_rate": 0.00013376884422110555, |
| "loss": 1.9537, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.07270588864237791, |
| "grad_norm": 0.3597472310066223, |
| "learning_rate": 0.0001327638190954774, |
| "loss": 1.9603, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.07377509288711877, |
| "grad_norm": 0.29230740666389465, |
| "learning_rate": 0.00013175879396984925, |
| "loss": 1.9189, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.07484429713185961, |
| "grad_norm": 0.31095099449157715, |
| "learning_rate": 0.0001307537688442211, |
| "loss": 1.8667, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.07591350137660047, |
| "grad_norm": 0.6215882897377014, |
| "learning_rate": 0.00012974874371859298, |
| "loss": 1.9415, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.07698270562134131, |
| "grad_norm": 0.3483082056045532, |
| "learning_rate": 0.00012874371859296485, |
| "loss": 1.9062, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.07805190986608217, |
| "grad_norm": 0.3564540147781372, |
| "learning_rate": 0.00012773869346733668, |
| "loss": 1.9511, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.07912111411082302, |
| "grad_norm": 0.2901301980018616, |
| "learning_rate": 0.00012673366834170855, |
| "loss": 1.9828, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.08019031835556387, |
| "grad_norm": 0.35523131489753723, |
| "learning_rate": 0.0001257286432160804, |
| "loss": 1.9864, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.08125952260030472, |
| "grad_norm": 0.3558770716190338, |
| "learning_rate": 0.00012472361809045228, |
| "loss": 1.8961, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.08232872684504558, |
| "grad_norm": 0.3421788811683655, |
| "learning_rate": 0.00012371859296482414, |
| "loss": 1.9287, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.08339793108978642, |
| "grad_norm": 0.2824752926826477, |
| "learning_rate": 0.00012271356783919598, |
| "loss": 1.8427, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.08446713533452728, |
| "grad_norm": 0.3791171908378601, |
| "learning_rate": 0.00012170854271356785, |
| "loss": 1.9233, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.08553633957926814, |
| "grad_norm": 0.3605143129825592, |
| "learning_rate": 0.0001207035175879397, |
| "loss": 1.9573, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.08660554382400898, |
| "grad_norm": 0.4379942715167999, |
| "learning_rate": 0.00011969849246231158, |
| "loss": 1.9844, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.08767474806874984, |
| "grad_norm": 0.2933856248855591, |
| "learning_rate": 0.00011869346733668343, |
| "loss": 1.8231, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.08874395231349068, |
| "grad_norm": 0.31451016664505005, |
| "learning_rate": 0.00011768844221105528, |
| "loss": 1.9715, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.08981315655823154, |
| "grad_norm": 0.3226332366466522, |
| "learning_rate": 0.00011668341708542714, |
| "loss": 1.9178, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.09088236080297239, |
| "grad_norm": 0.28982308506965637, |
| "learning_rate": 0.000115678391959799, |
| "loss": 2.0103, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.09195156504771324, |
| "grad_norm": 0.30275505781173706, |
| "learning_rate": 0.00011467336683417087, |
| "loss": 1.9218, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.09302076929245409, |
| "grad_norm": 0.3631080687046051, |
| "learning_rate": 0.00011366834170854272, |
| "loss": 1.9509, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.09408997353719495, |
| "grad_norm": 0.3824511468410492, |
| "learning_rate": 0.00011266331658291458, |
| "loss": 1.9624, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.09515917778193579, |
| "grad_norm": 0.24786897003650665, |
| "learning_rate": 0.00011165829145728644, |
| "loss": 1.8977, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.09622838202667665, |
| "grad_norm": 0.39904719591140747, |
| "learning_rate": 0.00011065326633165829, |
| "loss": 1.9995, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.09729758627141749, |
| "grad_norm": 0.3084559738636017, |
| "learning_rate": 0.00010964824120603017, |
| "loss": 1.9172, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.09836679051615835, |
| "grad_norm": 0.35854995250701904, |
| "learning_rate": 0.00010864321608040202, |
| "loss": 1.9198, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.0994359947608992, |
| "grad_norm": 0.29974690079689026, |
| "learning_rate": 0.00010763819095477387, |
| "loss": 1.9342, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.10050519900564005, |
| "grad_norm": 0.3411110043525696, |
| "learning_rate": 0.00010663316582914574, |
| "loss": 1.9233, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.1015744032503809, |
| "grad_norm": 0.3226073980331421, |
| "learning_rate": 0.00010562814070351759, |
| "loss": 1.8911, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.10264360749512176, |
| "grad_norm": 0.35765019059181213, |
| "learning_rate": 0.00010462311557788944, |
| "loss": 1.9626, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.1037128117398626, |
| "grad_norm": 0.3223406672477722, |
| "learning_rate": 0.00010361809045226132, |
| "loss": 1.8816, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.10478201598460346, |
| "grad_norm": 0.3049243986606598, |
| "learning_rate": 0.00010261306532663317, |
| "loss": 1.8941, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.10585122022934432, |
| "grad_norm": 0.2669491469860077, |
| "learning_rate": 0.00010160804020100503, |
| "loss": 1.8701, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.10692042447408516, |
| "grad_norm": 0.3321566581726074, |
| "learning_rate": 0.00010060301507537689, |
| "loss": 1.9848, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.10798962871882602, |
| "grad_norm": 0.3456732928752899, |
| "learning_rate": 9.959798994974875e-05, |
| "loss": 1.9494, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.10905883296356686, |
| "grad_norm": 0.32657039165496826, |
| "learning_rate": 9.85929648241206e-05, |
| "loss": 1.9775, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.11012803720830772, |
| "grad_norm": 0.3481104373931885, |
| "learning_rate": 9.758793969849247e-05, |
| "loss": 1.9383, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.11119724145304857, |
| "grad_norm": 0.3254976272583008, |
| "learning_rate": 9.658291457286432e-05, |
| "loss": 1.9553, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.11226644569778942, |
| "grad_norm": 0.365232914686203, |
| "learning_rate": 9.55778894472362e-05, |
| "loss": 1.935, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.11333564994253027, |
| "grad_norm": 0.3333910405635834, |
| "learning_rate": 9.457286432160805e-05, |
| "loss": 1.9184, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.11440485418727113, |
| "grad_norm": 0.3671784996986389, |
| "learning_rate": 9.35678391959799e-05, |
| "loss": 1.9303, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.11547405843201197, |
| "grad_norm": 0.2980143427848816, |
| "learning_rate": 9.256281407035176e-05, |
| "loss": 1.9606, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.11654326267675283, |
| "grad_norm": 0.29698073863983154, |
| "learning_rate": 9.155778894472362e-05, |
| "loss": 1.9324, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.11761246692149367, |
| "grad_norm": 0.35693129897117615, |
| "learning_rate": 9.055276381909548e-05, |
| "loss": 1.9567, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.11868167116623453, |
| "grad_norm": 0.39428821206092834, |
| "learning_rate": 8.954773869346734e-05, |
| "loss": 1.9712, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.11975087541097539, |
| "grad_norm": 0.4319991171360016, |
| "learning_rate": 8.85427135678392e-05, |
| "loss": 1.9229, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.12082007965571623, |
| "grad_norm": 0.271451860666275, |
| "learning_rate": 8.753768844221106e-05, |
| "loss": 1.9374, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.12188928390045708, |
| "grad_norm": 0.3350280523300171, |
| "learning_rate": 8.653266331658291e-05, |
| "loss": 1.97, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.12295848814519794, |
| "grad_norm": 0.23576690256595612, |
| "learning_rate": 8.552763819095478e-05, |
| "loss": 1.9165, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.12402769238993878, |
| "grad_norm": 0.3180292844772339, |
| "learning_rate": 8.452261306532664e-05, |
| "loss": 1.9629, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.12509689663467963, |
| "grad_norm": 0.35042503476142883, |
| "learning_rate": 8.351758793969849e-05, |
| "loss": 1.9137, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.12616610087942048, |
| "grad_norm": 0.33998608589172363, |
| "learning_rate": 8.251256281407036e-05, |
| "loss": 1.8604, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.12723530512416134, |
| "grad_norm": 0.3942660093307495, |
| "learning_rate": 8.150753768844221e-05, |
| "loss": 1.9758, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.1283045093689022, |
| "grad_norm": 0.34661826491355896, |
| "learning_rate": 8.050251256281407e-05, |
| "loss": 1.9638, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.12937371361364305, |
| "grad_norm": 0.31751224398612976, |
| "learning_rate": 7.949748743718594e-05, |
| "loss": 1.8879, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.1304429178583839, |
| "grad_norm": 0.3244868516921997, |
| "learning_rate": 7.849246231155779e-05, |
| "loss": 1.9034, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.13151212210312474, |
| "grad_norm": 0.28967052698135376, |
| "learning_rate": 7.748743718592966e-05, |
| "loss": 1.8447, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.1325813263478656, |
| "grad_norm": 0.23402956128120422, |
| "learning_rate": 7.64824120603015e-05, |
| "loss": 1.8901, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.13365053059260645, |
| "grad_norm": 0.29249387979507446, |
| "learning_rate": 7.547738693467337e-05, |
| "loss": 1.958, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.1347197348373473, |
| "grad_norm": 0.3257281184196472, |
| "learning_rate": 7.447236180904524e-05, |
| "loss": 1.9411, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.13578893908208817, |
| "grad_norm": 0.2677409052848816, |
| "learning_rate": 7.346733668341709e-05, |
| "loss": 1.9618, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.136858143326829, |
| "grad_norm": 0.5253011584281921, |
| "learning_rate": 7.246231155778895e-05, |
| "loss": 1.888, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.13792734757156985, |
| "grad_norm": 0.3806459307670593, |
| "learning_rate": 7.14572864321608e-05, |
| "loss": 1.9266, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.1389965518163107, |
| "grad_norm": 0.4092906415462494, |
| "learning_rate": 7.045226130653267e-05, |
| "loss": 1.9766, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.14006575606105157, |
| "grad_norm": 0.2884989082813263, |
| "learning_rate": 6.944723618090453e-05, |
| "loss": 1.946, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.14113496030579242, |
| "grad_norm": 0.3107132315635681, |
| "learning_rate": 6.844221105527638e-05, |
| "loss": 1.8267, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.14220416455053325, |
| "grad_norm": 0.29931432008743286, |
| "learning_rate": 6.743718592964824e-05, |
| "loss": 1.9226, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.1432733687952741, |
| "grad_norm": 0.32327061891555786, |
| "learning_rate": 6.64321608040201e-05, |
| "loss": 1.9389, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.14434257304001497, |
| "grad_norm": 0.3812110126018524, |
| "learning_rate": 6.542713567839197e-05, |
| "loss": 1.9137, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.14541177728475582, |
| "grad_norm": 0.3208372890949249, |
| "learning_rate": 6.442211055276383e-05, |
| "loss": 1.8635, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.14648098152949668, |
| "grad_norm": 0.3378104567527771, |
| "learning_rate": 6.341708542713568e-05, |
| "loss": 1.9773, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.14755018577423754, |
| "grad_norm": 0.32731178402900696, |
| "learning_rate": 6.241206030150753e-05, |
| "loss": 1.9294, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.14861939001897836, |
| "grad_norm": 0.30876997113227844, |
| "learning_rate": 6.14070351758794e-05, |
| "loss": 1.9457, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.14968859426371922, |
| "grad_norm": 0.2618562579154968, |
| "learning_rate": 6.0402010050251256e-05, |
| "loss": 1.9014, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.15075779850846008, |
| "grad_norm": 0.3791179656982422, |
| "learning_rate": 5.939698492462312e-05, |
| "loss": 1.897, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.15182700275320093, |
| "grad_norm": 0.31716200709342957, |
| "learning_rate": 5.839195979899498e-05, |
| "loss": 1.947, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.1528962069979418, |
| "grad_norm": 0.2968918979167938, |
| "learning_rate": 5.738693467336683e-05, |
| "loss": 1.9895, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.15396541124268262, |
| "grad_norm": 0.3310339152812958, |
| "learning_rate": 5.63819095477387e-05, |
| "loss": 1.9455, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.15503461548742348, |
| "grad_norm": 0.30880674719810486, |
| "learning_rate": 5.537688442211055e-05, |
| "loss": 1.9152, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.15610381973216433, |
| "grad_norm": 0.28629955649375916, |
| "learning_rate": 5.437185929648242e-05, |
| "loss": 1.8665, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.1571730239769052, |
| "grad_norm": 0.3241690695285797, |
| "learning_rate": 5.3366834170854276e-05, |
| "loss": 1.9451, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.15824222822164605, |
| "grad_norm": 0.26445063948631287, |
| "learning_rate": 5.236180904522613e-05, |
| "loss": 1.897, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.1593114324663869, |
| "grad_norm": 0.3061549961566925, |
| "learning_rate": 5.135678391959799e-05, |
| "loss": 1.8208, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.16038063671112773, |
| "grad_norm": 0.334157794713974, |
| "learning_rate": 5.035175879396985e-05, |
| "loss": 1.854, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.1614498409558686, |
| "grad_norm": 0.2809840142726898, |
| "learning_rate": 4.934673366834171e-05, |
| "loss": 1.9399, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.16251904520060945, |
| "grad_norm": 0.331537663936615, |
| "learning_rate": 4.834170854271357e-05, |
| "loss": 1.9757, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.1635882494453503, |
| "grad_norm": 0.30003657937049866, |
| "learning_rate": 4.733668341708543e-05, |
| "loss": 1.9077, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.16465745369009116, |
| "grad_norm": 0.29070165753364563, |
| "learning_rate": 4.633165829145729e-05, |
| "loss": 1.9441, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.165726657934832, |
| "grad_norm": 0.3907763659954071, |
| "learning_rate": 4.532663316582915e-05, |
| "loss": 1.8893, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.16679586217957285, |
| "grad_norm": 0.3012675940990448, |
| "learning_rate": 4.4321608040201005e-05, |
| "loss": 1.9539, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.1678650664243137, |
| "grad_norm": 0.28272444009780884, |
| "learning_rate": 4.331658291457287e-05, |
| "loss": 1.9571, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.16893427066905456, |
| "grad_norm": 0.3389994502067566, |
| "learning_rate": 4.231155778894473e-05, |
| "loss": 1.8685, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.17000347491379542, |
| "grad_norm": 0.3380429744720459, |
| "learning_rate": 4.1306532663316586e-05, |
| "loss": 1.9519, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.17107267915853627, |
| "grad_norm": 0.3569534122943878, |
| "learning_rate": 4.0301507537688444e-05, |
| "loss": 1.8919, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.1721418834032771, |
| "grad_norm": 0.32496699690818787, |
| "learning_rate": 3.92964824120603e-05, |
| "loss": 1.8853, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.17321108764801796, |
| "grad_norm": 0.3131369650363922, |
| "learning_rate": 3.829145728643217e-05, |
| "loss": 1.8952, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.17428029189275882, |
| "grad_norm": 0.3050549626350403, |
| "learning_rate": 3.7286432160804025e-05, |
| "loss": 1.85, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.17534949613749967, |
| "grad_norm": 0.2770765423774719, |
| "learning_rate": 3.628140703517588e-05, |
| "loss": 1.9345, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.17641870038224053, |
| "grad_norm": 0.40778982639312744, |
| "learning_rate": 3.527638190954774e-05, |
| "loss": 1.885, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.17748790462698136, |
| "grad_norm": 0.305587500333786, |
| "learning_rate": 3.42713567839196e-05, |
| "loss": 1.9261, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.17855710887172221, |
| "grad_norm": 0.2813640534877777, |
| "learning_rate": 3.3266331658291464e-05, |
| "loss": 1.981, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.17962631311646307, |
| "grad_norm": 0.2698950469493866, |
| "learning_rate": 3.2261306532663315e-05, |
| "loss": 1.9081, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.18069551736120393, |
| "grad_norm": 0.3598436713218689, |
| "learning_rate": 3.125628140703517e-05, |
| "loss": 1.9129, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.18176472160594478, |
| "grad_norm": 0.25739961862564087, |
| "learning_rate": 3.0251256281407038e-05, |
| "loss": 1.9376, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.1828339258506856, |
| "grad_norm": 0.30413052439689636, |
| "learning_rate": 2.9246231155778896e-05, |
| "loss": 1.9186, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.18390313009542647, |
| "grad_norm": 0.2563576400279999, |
| "learning_rate": 2.8241206030150757e-05, |
| "loss": 1.847, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.18497233434016733, |
| "grad_norm": 0.3041762113571167, |
| "learning_rate": 2.7236180904522612e-05, |
| "loss": 1.9606, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.18604153858490818, |
| "grad_norm": 0.283974826335907, |
| "learning_rate": 2.6231155778894474e-05, |
| "loss": 1.9263, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.18711074282964904, |
| "grad_norm": 0.3330858051776886, |
| "learning_rate": 2.522613065326633e-05, |
| "loss": 1.8791, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.1881799470743899, |
| "grad_norm": 0.31441086530685425, |
| "learning_rate": 2.422110552763819e-05, |
| "loss": 1.9351, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.18924915131913073, |
| "grad_norm": 0.31107625365257263, |
| "learning_rate": 2.321608040201005e-05, |
| "loss": 1.9231, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.19031835556387158, |
| "grad_norm": 0.34114930033683777, |
| "learning_rate": 2.2211055276381913e-05, |
| "loss": 1.9559, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.19138755980861244, |
| "grad_norm": 0.23545467853546143, |
| "learning_rate": 2.120603015075377e-05, |
| "loss": 1.849, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.1924567640533533, |
| "grad_norm": 0.29872679710388184, |
| "learning_rate": 2.020100502512563e-05, |
| "loss": 1.9186, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.19352596829809415, |
| "grad_norm": 0.34200844168663025, |
| "learning_rate": 1.9195979899497487e-05, |
| "loss": 1.9165, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.19459517254283498, |
| "grad_norm": 0.3605504035949707, |
| "learning_rate": 1.8190954773869348e-05, |
| "loss": 1.9204, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.19566437678757584, |
| "grad_norm": 0.3210300803184509, |
| "learning_rate": 1.7185929648241206e-05, |
| "loss": 1.9172, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.1967335810323167, |
| "grad_norm": 0.31484687328338623, |
| "learning_rate": 1.6180904522613068e-05, |
| "loss": 1.9356, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.19780278527705755, |
| "grad_norm": 0.335658460855484, |
| "learning_rate": 1.5175879396984927e-05, |
| "loss": 1.9379, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.1988719895217984, |
| "grad_norm": 0.36707040667533875, |
| "learning_rate": 1.4170854271356784e-05, |
| "loss": 1.9189, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.19994119376653927, |
| "grad_norm": 0.2470160871744156, |
| "learning_rate": 1.3165829145728645e-05, |
| "loss": 1.9095, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.2010103980112801, |
| "grad_norm": 0.24876108765602112, |
| "learning_rate": 1.2160804020100503e-05, |
| "loss": 1.8907, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.20207960225602095, |
| "grad_norm": 0.31706857681274414, |
| "learning_rate": 1.1155778894472363e-05, |
| "loss": 1.9588, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.2031488065007618, |
| "grad_norm": 0.27313029766082764, |
| "learning_rate": 1.0150753768844223e-05, |
| "loss": 1.936, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.20421801074550266, |
| "grad_norm": 0.2927481532096863, |
| "learning_rate": 9.14572864321608e-06, |
| "loss": 1.904, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.20528721499024352, |
| "grad_norm": 0.261416494846344, |
| "learning_rate": 8.14070351758794e-06, |
| "loss": 1.9233, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.20635641923498435, |
| "grad_norm": 0.2947026193141937, |
| "learning_rate": 7.1356783919597995e-06, |
| "loss": 1.8992, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.2074256234797252, |
| "grad_norm": 0.31104397773742676, |
| "learning_rate": 6.130653266331659e-06, |
| "loss": 1.9726, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.20849482772446606, |
| "grad_norm": 0.2600247263908386, |
| "learning_rate": 5.125628140703518e-06, |
| "loss": 2.02, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.20956403196920692, |
| "grad_norm": 0.30064964294433594, |
| "learning_rate": 4.120603015075377e-06, |
| "loss": 1.8599, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.21063323621394778, |
| "grad_norm": 0.2797416150569916, |
| "learning_rate": 3.1155778894472364e-06, |
| "loss": 1.9514, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.21170244045868863, |
| "grad_norm": 0.34935057163238525, |
| "learning_rate": 2.1105527638190953e-06, |
| "loss": 1.8716, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.21277164470342946, |
| "grad_norm": 0.2744124233722687, |
| "learning_rate": 1.1055276381909548e-06, |
| "loss": 1.9384, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.21384084894817032, |
| "grad_norm": 0.2645925283432007, |
| "learning_rate": 1.0050251256281409e-07, |
| "loss": 1.9418, |
| "step": 2000 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 2000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 50, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 6.9738373840896e+17, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|