| { |
| "best_metric": 0.02887474000453949, |
| "best_model_checkpoint": "saves/psy-course/MentaLLaMA-chat-7B/train/fold2/checkpoint-1900", |
| "epoch": 5.0, |
| "eval_steps": 50, |
| "global_step": 3225, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.015503875968992248, |
| "grad_norm": 1.8097485303878784, |
| "learning_rate": 3.0959752321981426e-06, |
| "loss": 1.678, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.031007751937984496, |
| "grad_norm": 1.9774603843688965, |
| "learning_rate": 6.191950464396285e-06, |
| "loss": 1.7145, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.046511627906976744, |
| "grad_norm": 2.036001443862915, |
| "learning_rate": 9.287925696594429e-06, |
| "loss": 1.5924, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.06201550387596899, |
| "grad_norm": 3.265268087387085, |
| "learning_rate": 1.238390092879257e-05, |
| "loss": 1.2457, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.07751937984496124, |
| "grad_norm": 1.1102505922317505, |
| "learning_rate": 1.5479876160990712e-05, |
| "loss": 0.7519, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.07751937984496124, |
| "eval_loss": 0.638915479183197, |
| "eval_runtime": 177.4395, |
| "eval_samples_per_second": 6.464, |
| "eval_steps_per_second": 6.464, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.09302325581395349, |
| "grad_norm": 0.6308065056800842, |
| "learning_rate": 1.8575851393188857e-05, |
| "loss": 0.6351, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.10852713178294573, |
| "grad_norm": 0.8012827634811401, |
| "learning_rate": 2.1671826625387e-05, |
| "loss": 0.4782, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.12403100775193798, |
| "grad_norm": 0.6199682354927063, |
| "learning_rate": 2.476780185758514e-05, |
| "loss": 0.2779, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.13953488372093023, |
| "grad_norm": 0.5094736218452454, |
| "learning_rate": 2.7863777089783283e-05, |
| "loss": 0.1896, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.15503875968992248, |
| "grad_norm": 0.5959056615829468, |
| "learning_rate": 3.0959752321981425e-05, |
| "loss": 0.1485, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.15503875968992248, |
| "eval_loss": 0.11244650185108185, |
| "eval_runtime": 177.3837, |
| "eval_samples_per_second": 6.466, |
| "eval_steps_per_second": 6.466, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.17054263565891473, |
| "grad_norm": 0.919588029384613, |
| "learning_rate": 3.4055727554179566e-05, |
| "loss": 0.1271, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.18604651162790697, |
| "grad_norm": 0.9113105535507202, |
| "learning_rate": 3.7151702786377715e-05, |
| "loss": 0.101, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.20155038759689922, |
| "grad_norm": 0.74349045753479, |
| "learning_rate": 4.024767801857585e-05, |
| "loss": 0.1009, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.21705426356589147, |
| "grad_norm": 0.6830365061759949, |
| "learning_rate": 4.3343653250774e-05, |
| "loss": 0.096, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.23255813953488372, |
| "grad_norm": 0.5473061203956604, |
| "learning_rate": 4.6439628482972134e-05, |
| "loss": 0.074, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.23255813953488372, |
| "eval_loss": 0.06496970355510712, |
| "eval_runtime": 177.3867, |
| "eval_samples_per_second": 6.466, |
| "eval_steps_per_second": 6.466, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.24806201550387597, |
| "grad_norm": 0.45949482917785645, |
| "learning_rate": 4.953560371517028e-05, |
| "loss": 0.076, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.26356589147286824, |
| "grad_norm": 0.6822295188903809, |
| "learning_rate": 5.2631578947368424e-05, |
| "loss": 0.0689, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.27906976744186046, |
| "grad_norm": 0.5395657420158386, |
| "learning_rate": 5.5727554179566566e-05, |
| "loss": 0.0634, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.29457364341085274, |
| "grad_norm": 1.2032052278518677, |
| "learning_rate": 5.882352941176471e-05, |
| "loss": 0.0687, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.31007751937984496, |
| "grad_norm": 0.8278073072433472, |
| "learning_rate": 6.191950464396285e-05, |
| "loss": 0.0655, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.31007751937984496, |
| "eval_loss": 0.06194576248526573, |
| "eval_runtime": 177.5104, |
| "eval_samples_per_second": 6.462, |
| "eval_steps_per_second": 6.462, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.32558139534883723, |
| "grad_norm": 0.48666349053382874, |
| "learning_rate": 6.501547987616098e-05, |
| "loss": 0.0591, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.34108527131782945, |
| "grad_norm": 0.6362942457199097, |
| "learning_rate": 6.811145510835913e-05, |
| "loss": 0.0557, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.35658914728682173, |
| "grad_norm": 0.44466376304626465, |
| "learning_rate": 7.120743034055728e-05, |
| "loss": 0.0531, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.37209302325581395, |
| "grad_norm": 0.473111093044281, |
| "learning_rate": 7.430340557275543e-05, |
| "loss": 0.0704, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.3875968992248062, |
| "grad_norm": 0.6141083836555481, |
| "learning_rate": 7.739938080495357e-05, |
| "loss": 0.0598, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.3875968992248062, |
| "eval_loss": 0.05116499215364456, |
| "eval_runtime": 177.5255, |
| "eval_samples_per_second": 6.461, |
| "eval_steps_per_second": 6.461, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.40310077519379844, |
| "grad_norm": 0.34960028529167175, |
| "learning_rate": 8.04953560371517e-05, |
| "loss": 0.0502, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.4186046511627907, |
| "grad_norm": 0.6346738338470459, |
| "learning_rate": 8.359133126934985e-05, |
| "loss": 0.0469, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.43410852713178294, |
| "grad_norm": 0.34611088037490845, |
| "learning_rate": 8.6687306501548e-05, |
| "loss": 0.0769, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.4496124031007752, |
| "grad_norm": 0.746246337890625, |
| "learning_rate": 8.978328173374613e-05, |
| "loss": 0.0547, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.46511627906976744, |
| "grad_norm": 0.3229033350944519, |
| "learning_rate": 9.287925696594427e-05, |
| "loss": 0.0414, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.46511627906976744, |
| "eval_loss": 0.04494742304086685, |
| "eval_runtime": 177.4386, |
| "eval_samples_per_second": 6.464, |
| "eval_steps_per_second": 6.464, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.4806201550387597, |
| "grad_norm": 0.7018037438392639, |
| "learning_rate": 9.597523219814242e-05, |
| "loss": 0.0469, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.49612403100775193, |
| "grad_norm": 0.3599799871444702, |
| "learning_rate": 9.907120743034056e-05, |
| "loss": 0.0377, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.5116279069767442, |
| "grad_norm": 0.3153678774833679, |
| "learning_rate": 9.999856438185238e-05, |
| "loss": 0.0441, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.5271317829457365, |
| "grad_norm": 0.2962203621864319, |
| "learning_rate": 9.999153298122152e-05, |
| "loss": 0.0522, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.5426356589147286, |
| "grad_norm": 0.1656709760427475, |
| "learning_rate": 9.997864293614054e-05, |
| "loss": 0.0427, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.5426356589147286, |
| "eval_loss": 0.04140669107437134, |
| "eval_runtime": 177.5437, |
| "eval_samples_per_second": 6.46, |
| "eval_steps_per_second": 6.46, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.5581395348837209, |
| "grad_norm": 0.2917194068431854, |
| "learning_rate": 9.995989575722902e-05, |
| "loss": 0.047, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.5736434108527132, |
| "grad_norm": 0.3477674424648285, |
| "learning_rate": 9.993529364152018e-05, |
| "loss": 0.0501, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.5891472868217055, |
| "grad_norm": 0.48425808548927307, |
| "learning_rate": 9.990483947220319e-05, |
| "loss": 0.0593, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.6046511627906976, |
| "grad_norm": 0.21694518625736237, |
| "learning_rate": 9.986853681828546e-05, |
| "loss": 0.0513, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.6201550387596899, |
| "grad_norm": 0.5227199792861938, |
| "learning_rate": 9.982638993417425e-05, |
| "loss": 0.0471, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.6201550387596899, |
| "eval_loss": 0.038663338869810104, |
| "eval_runtime": 177.5346, |
| "eval_samples_per_second": 6.461, |
| "eval_steps_per_second": 6.461, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.6356589147286822, |
| "grad_norm": 0.25696733593940735, |
| "learning_rate": 9.977840375917817e-05, |
| "loss": 0.0475, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.6511627906976745, |
| "grad_norm": 0.22936882078647614, |
| "learning_rate": 9.972458391692827e-05, |
| "loss": 0.0405, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.6666666666666666, |
| "grad_norm": 0.31321045756340027, |
| "learning_rate": 9.966493671471904e-05, |
| "loss": 0.0449, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.6821705426356589, |
| "grad_norm": 0.3151184916496277, |
| "learning_rate": 9.959946914276922e-05, |
| "loss": 0.056, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.6976744186046512, |
| "grad_norm": 0.50681471824646, |
| "learning_rate": 9.952818887340257e-05, |
| "loss": 0.0433, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.6976744186046512, |
| "eval_loss": 0.03619127348065376, |
| "eval_runtime": 177.5548, |
| "eval_samples_per_second": 6.46, |
| "eval_steps_per_second": 6.46, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.7131782945736435, |
| "grad_norm": 0.31854957342147827, |
| "learning_rate": 9.945110426014878e-05, |
| "loss": 0.0494, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.7286821705426356, |
| "grad_norm": 0.28269049525260925, |
| "learning_rate": 9.936822433676444e-05, |
| "loss": 0.0387, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.7441860465116279, |
| "grad_norm": 0.24601627886295319, |
| "learning_rate": 9.927955881617444e-05, |
| "loss": 0.0479, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.7596899224806202, |
| "grad_norm": 0.19982942938804626, |
| "learning_rate": 9.918511808933358e-05, |
| "loss": 0.0378, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.7751937984496124, |
| "grad_norm": 0.20261038839817047, |
| "learning_rate": 9.908491322400885e-05, |
| "loss": 0.0432, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.7751937984496124, |
| "eval_loss": 0.03534514829516411, |
| "eval_runtime": 177.4652, |
| "eval_samples_per_second": 6.463, |
| "eval_steps_per_second": 6.463, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.7906976744186046, |
| "grad_norm": 0.1283186674118042, |
| "learning_rate": 9.897895596348247e-05, |
| "loss": 0.0395, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.8062015503875969, |
| "grad_norm": 0.2949117124080658, |
| "learning_rate": 9.886725872517552e-05, |
| "loss": 0.0326, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.8217054263565892, |
| "grad_norm": 0.42478370666503906, |
| "learning_rate": 9.874983459919277e-05, |
| "loss": 0.0442, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.8372093023255814, |
| "grad_norm": 0.270944207906723, |
| "learning_rate": 9.862669734678867e-05, |
| "loss": 0.038, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.8527131782945736, |
| "grad_norm": 0.19820506870746613, |
| "learning_rate": 9.849786139875452e-05, |
| "loss": 0.0445, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.8527131782945736, |
| "eval_loss": 0.03533301129937172, |
| "eval_runtime": 177.5388, |
| "eval_samples_per_second": 6.461, |
| "eval_steps_per_second": 6.461, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.8682170542635659, |
| "grad_norm": 0.33879902958869934, |
| "learning_rate": 9.836334185372738e-05, |
| "loss": 0.0464, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.8837209302325582, |
| "grad_norm": 0.330963134765625, |
| "learning_rate": 9.822315447642056e-05, |
| "loss": 0.0368, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.8992248062015504, |
| "grad_norm": 0.22305184602737427, |
| "learning_rate": 9.807731569577615e-05, |
| "loss": 0.0296, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.9147286821705426, |
| "grad_norm": 0.28611594438552856, |
| "learning_rate": 9.792584260303964e-05, |
| "loss": 0.0364, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.9302325581395349, |
| "grad_norm": 0.24084076285362244, |
| "learning_rate": 9.776875294975698e-05, |
| "loss": 0.0529, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.9302325581395349, |
| "eval_loss": 0.03533976525068283, |
| "eval_runtime": 177.6164, |
| "eval_samples_per_second": 6.458, |
| "eval_steps_per_second": 6.458, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.9457364341085271, |
| "grad_norm": 0.28942936658859253, |
| "learning_rate": 9.760606514569416e-05, |
| "loss": 0.0381, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.9612403100775194, |
| "grad_norm": 0.2056342214345932, |
| "learning_rate": 9.743779825667984e-05, |
| "loss": 0.0337, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.9767441860465116, |
| "grad_norm": 0.14191323518753052, |
| "learning_rate": 9.726397200237085e-05, |
| "loss": 0.0368, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.9922480620155039, |
| "grad_norm": 0.33825528621673584, |
| "learning_rate": 9.708460675394126e-05, |
| "loss": 0.0465, |
| "step": 640 |
| }, |
| { |
| "epoch": 1.0077519379844961, |
| "grad_norm": 0.12557874619960785, |
| "learning_rate": 9.689972353169499e-05, |
| "loss": 0.0313, |
| "step": 650 |
| }, |
| { |
| "epoch": 1.0077519379844961, |
| "eval_loss": 0.03178540617227554, |
| "eval_runtime": 177.627, |
| "eval_samples_per_second": 6.457, |
| "eval_steps_per_second": 6.457, |
| "step": 650 |
| }, |
| { |
| "epoch": 1.0232558139534884, |
| "grad_norm": 0.35606396198272705, |
| "learning_rate": 9.67093440026024e-05, |
| "loss": 0.0303, |
| "step": 660 |
| }, |
| { |
| "epoch": 1.0387596899224807, |
| "grad_norm": 0.183203786611557, |
| "learning_rate": 9.651349047776112e-05, |
| "loss": 0.0386, |
| "step": 670 |
| }, |
| { |
| "epoch": 1.054263565891473, |
| "grad_norm": 0.23599869012832642, |
| "learning_rate": 9.631218590978126e-05, |
| "loss": 0.0222, |
| "step": 680 |
| }, |
| { |
| "epoch": 1.069767441860465, |
| "grad_norm": 0.13795703649520874, |
| "learning_rate": 9.610545389009562e-05, |
| "loss": 0.0307, |
| "step": 690 |
| }, |
| { |
| "epoch": 1.0852713178294573, |
| "grad_norm": 0.2756401002407074, |
| "learning_rate": 9.589331864619486e-05, |
| "loss": 0.0301, |
| "step": 700 |
| }, |
| { |
| "epoch": 1.0852713178294573, |
| "eval_loss": 0.03220707178115845, |
| "eval_runtime": 177.6556, |
| "eval_samples_per_second": 6.456, |
| "eval_steps_per_second": 6.456, |
| "step": 700 |
| }, |
| { |
| "epoch": 1.1007751937984496, |
| "grad_norm": 0.22982142865657806, |
| "learning_rate": 9.567580503878833e-05, |
| "loss": 0.025, |
| "step": 710 |
| }, |
| { |
| "epoch": 1.1162790697674418, |
| "grad_norm": 0.21597911417484283, |
| "learning_rate": 9.545293855889043e-05, |
| "loss": 0.0342, |
| "step": 720 |
| }, |
| { |
| "epoch": 1.1317829457364341, |
| "grad_norm": 0.16282838582992554, |
| "learning_rate": 9.522474532483337e-05, |
| "loss": 0.0275, |
| "step": 730 |
| }, |
| { |
| "epoch": 1.1472868217054264, |
| "grad_norm": 0.32020196318626404, |
| "learning_rate": 9.499125207920622e-05, |
| "loss": 0.0402, |
| "step": 740 |
| }, |
| { |
| "epoch": 1.1627906976744187, |
| "grad_norm": 0.12987101078033447, |
| "learning_rate": 9.475248618572096e-05, |
| "loss": 0.0289, |
| "step": 750 |
| }, |
| { |
| "epoch": 1.1627906976744187, |
| "eval_loss": 0.03383904695510864, |
| "eval_runtime": 177.667, |
| "eval_samples_per_second": 6.456, |
| "eval_steps_per_second": 6.456, |
| "step": 750 |
| }, |
| { |
| "epoch": 1.178294573643411, |
| "grad_norm": 0.37133800983428955, |
| "learning_rate": 9.45084756260055e-05, |
| "loss": 0.035, |
| "step": 760 |
| }, |
| { |
| "epoch": 1.193798449612403, |
| "grad_norm": 0.07524913549423218, |
| "learning_rate": 9.425924899632458e-05, |
| "loss": 0.0291, |
| "step": 770 |
| }, |
| { |
| "epoch": 1.2093023255813953, |
| "grad_norm": 0.14250116050243378, |
| "learning_rate": 9.400483550422845e-05, |
| "loss": 0.0221, |
| "step": 780 |
| }, |
| { |
| "epoch": 1.2248062015503876, |
| "grad_norm": 0.33380234241485596, |
| "learning_rate": 9.374526496512992e-05, |
| "loss": 0.0303, |
| "step": 790 |
| }, |
| { |
| "epoch": 1.2403100775193798, |
| "grad_norm": 0.6035688519477844, |
| "learning_rate": 9.348056779881025e-05, |
| "loss": 0.0267, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.2403100775193798, |
| "eval_loss": 0.03135337680578232, |
| "eval_runtime": 177.6172, |
| "eval_samples_per_second": 6.458, |
| "eval_steps_per_second": 6.458, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.255813953488372, |
| "grad_norm": 0.25495055317878723, |
| "learning_rate": 9.321077502585417e-05, |
| "loss": 0.0267, |
| "step": 810 |
| }, |
| { |
| "epoch": 1.2713178294573644, |
| "grad_norm": 0.352885365486145, |
| "learning_rate": 9.293591826401451e-05, |
| "loss": 0.0351, |
| "step": 820 |
| }, |
| { |
| "epoch": 1.2868217054263567, |
| "grad_norm": 0.25123533606529236, |
| "learning_rate": 9.265602972450673e-05, |
| "loss": 0.0328, |
| "step": 830 |
| }, |
| { |
| "epoch": 1.302325581395349, |
| "grad_norm": 0.47959160804748535, |
| "learning_rate": 9.237114220823413e-05, |
| "loss": 0.0326, |
| "step": 840 |
| }, |
| { |
| "epoch": 1.3178294573643412, |
| "grad_norm": 0.13225722312927246, |
| "learning_rate": 9.208128910194377e-05, |
| "loss": 0.0314, |
| "step": 850 |
| }, |
| { |
| "epoch": 1.3178294573643412, |
| "eval_loss": 0.03171215206384659, |
| "eval_runtime": 177.6409, |
| "eval_samples_per_second": 6.457, |
| "eval_steps_per_second": 6.457, |
| "step": 850 |
| }, |
| { |
| "epoch": 1.3333333333333333, |
| "grad_norm": 0.27190810441970825, |
| "learning_rate": 9.17865043743138e-05, |
| "loss": 0.031, |
| "step": 860 |
| }, |
| { |
| "epoch": 1.3488372093023255, |
| "grad_norm": 0.10395022481679916, |
| "learning_rate": 9.148682257197245e-05, |
| "loss": 0.0238, |
| "step": 870 |
| }, |
| { |
| "epoch": 1.3643410852713178, |
| "grad_norm": 0.33697041869163513, |
| "learning_rate": 9.118227881544967e-05, |
| "loss": 0.0397, |
| "step": 880 |
| }, |
| { |
| "epoch": 1.37984496124031, |
| "grad_norm": 0.2214643359184265, |
| "learning_rate": 9.087290879506104e-05, |
| "loss": 0.0281, |
| "step": 890 |
| }, |
| { |
| "epoch": 1.3953488372093024, |
| "grad_norm": 0.11761048436164856, |
| "learning_rate": 9.055874876672519e-05, |
| "loss": 0.0382, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.3953488372093024, |
| "eval_loss": 0.03273608908057213, |
| "eval_runtime": 177.7052, |
| "eval_samples_per_second": 6.455, |
| "eval_steps_per_second": 6.455, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.4108527131782946, |
| "grad_norm": 0.17550455033779144, |
| "learning_rate": 9.023983554771492e-05, |
| "loss": 0.0263, |
| "step": 910 |
| }, |
| { |
| "epoch": 1.4263565891472867, |
| "grad_norm": 0.21578268706798553, |
| "learning_rate": 8.991620651234242e-05, |
| "loss": 0.0261, |
| "step": 920 |
| }, |
| { |
| "epoch": 1.441860465116279, |
| "grad_norm": 0.21228443086147308, |
| "learning_rate": 8.958789958757928e-05, |
| "loss": 0.029, |
| "step": 930 |
| }, |
| { |
| "epoch": 1.4573643410852712, |
| "grad_norm": 0.2734227180480957, |
| "learning_rate": 8.925495324861177e-05, |
| "loss": 0.0286, |
| "step": 940 |
| }, |
| { |
| "epoch": 1.4728682170542635, |
| "grad_norm": 0.4174731373786926, |
| "learning_rate": 8.89174065143318e-05, |
| "loss": 0.0354, |
| "step": 950 |
| }, |
| { |
| "epoch": 1.4728682170542635, |
| "eval_loss": 0.031981565058231354, |
| "eval_runtime": 177.7226, |
| "eval_samples_per_second": 6.454, |
| "eval_steps_per_second": 6.454, |
| "step": 950 |
| }, |
| { |
| "epoch": 1.4883720930232558, |
| "grad_norm": 0.3433147370815277, |
| "learning_rate": 8.857529894276412e-05, |
| "loss": 0.0352, |
| "step": 960 |
| }, |
| { |
| "epoch": 1.503875968992248, |
| "grad_norm": 0.21089136600494385, |
| "learning_rate": 8.822867062643059e-05, |
| "loss": 0.0298, |
| "step": 970 |
| }, |
| { |
| "epoch": 1.5193798449612403, |
| "grad_norm": 0.1780281960964203, |
| "learning_rate": 8.78775621876514e-05, |
| "loss": 0.0319, |
| "step": 980 |
| }, |
| { |
| "epoch": 1.5348837209302326, |
| "grad_norm": 0.3001708686351776, |
| "learning_rate": 8.752201477378459e-05, |
| "loss": 0.0217, |
| "step": 990 |
| }, |
| { |
| "epoch": 1.550387596899225, |
| "grad_norm": 0.15130403637886047, |
| "learning_rate": 8.716207005240382e-05, |
| "loss": 0.0265, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.550387596899225, |
| "eval_loss": 0.03209313005208969, |
| "eval_runtime": 177.6849, |
| "eval_samples_per_second": 6.455, |
| "eval_steps_per_second": 6.455, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.5658914728682172, |
| "grad_norm": 0.28896772861480713, |
| "learning_rate": 8.679777020641525e-05, |
| "loss": 0.0345, |
| "step": 1010 |
| }, |
| { |
| "epoch": 1.5813953488372094, |
| "grad_norm": 0.4542350471019745, |
| "learning_rate": 8.6429157929114e-05, |
| "loss": 0.033, |
| "step": 1020 |
| }, |
| { |
| "epoch": 1.5968992248062015, |
| "grad_norm": 0.26365530490875244, |
| "learning_rate": 8.60562764191808e-05, |
| "loss": 0.0277, |
| "step": 1030 |
| }, |
| { |
| "epoch": 1.6124031007751938, |
| "grad_norm": 0.17537464201450348, |
| "learning_rate": 8.567916937561944e-05, |
| "loss": 0.0307, |
| "step": 1040 |
| }, |
| { |
| "epoch": 1.627906976744186, |
| "grad_norm": 0.39150458574295044, |
| "learning_rate": 8.529788099263554e-05, |
| "loss": 0.0301, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.627906976744186, |
| "eval_loss": 0.03326583281159401, |
| "eval_runtime": 177.6473, |
| "eval_samples_per_second": 6.457, |
| "eval_steps_per_second": 6.457, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.6434108527131783, |
| "grad_norm": 0.2252993881702423, |
| "learning_rate": 8.49124559544573e-05, |
| "loss": 0.0279, |
| "step": 1060 |
| }, |
| { |
| "epoch": 1.6589147286821704, |
| "grad_norm": 0.3254898488521576, |
| "learning_rate": 8.452293943009889e-05, |
| "loss": 0.0269, |
| "step": 1070 |
| }, |
| { |
| "epoch": 1.6744186046511627, |
| "grad_norm": 0.2006569355726242, |
| "learning_rate": 8.412937706806692e-05, |
| "loss": 0.0361, |
| "step": 1080 |
| }, |
| { |
| "epoch": 1.689922480620155, |
| "grad_norm": 0.2423427700996399, |
| "learning_rate": 8.373181499101077e-05, |
| "loss": 0.0301, |
| "step": 1090 |
| }, |
| { |
| "epoch": 1.7054263565891472, |
| "grad_norm": 0.36214545369148254, |
| "learning_rate": 8.333029979031737e-05, |
| "loss": 0.0262, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.7054263565891472, |
| "eval_loss": 0.031156621873378754, |
| "eval_runtime": 177.8049, |
| "eval_samples_per_second": 6.451, |
| "eval_steps_per_second": 6.451, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.7209302325581395, |
| "grad_norm": 0.17312179505825043, |
| "learning_rate": 8.292487852065104e-05, |
| "loss": 0.0392, |
| "step": 1110 |
| }, |
| { |
| "epoch": 1.7364341085271318, |
| "grad_norm": 0.275713711977005, |
| "learning_rate": 8.251559869443897e-05, |
| "loss": 0.0322, |
| "step": 1120 |
| }, |
| { |
| "epoch": 1.751937984496124, |
| "grad_norm": 0.33000582456588745, |
| "learning_rate": 8.210250827630313e-05, |
| "loss": 0.029, |
| "step": 1130 |
| }, |
| { |
| "epoch": 1.7674418604651163, |
| "grad_norm": 0.29563483595848083, |
| "learning_rate": 8.168565567743924e-05, |
| "loss": 0.0263, |
| "step": 1140 |
| }, |
| { |
| "epoch": 1.7829457364341086, |
| "grad_norm": 0.31916943192481995, |
| "learning_rate": 8.126508974994321e-05, |
| "loss": 0.0273, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.7829457364341086, |
| "eval_loss": 0.030605774372816086, |
| "eval_runtime": 177.725, |
| "eval_samples_per_second": 6.454, |
| "eval_steps_per_second": 6.454, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.7984496124031009, |
| "grad_norm": 0.1885177344083786, |
| "learning_rate": 8.084085978108611e-05, |
| "loss": 0.0169, |
| "step": 1160 |
| }, |
| { |
| "epoch": 1.8139534883720931, |
| "grad_norm": 0.15678612887859344, |
| "learning_rate": 8.041301548753804e-05, |
| "loss": 0.0284, |
| "step": 1170 |
| }, |
| { |
| "epoch": 1.8294573643410854, |
| "grad_norm": 0.24891909956932068, |
| "learning_rate": 7.998160700954165e-05, |
| "loss": 0.0244, |
| "step": 1180 |
| }, |
| { |
| "epoch": 1.8449612403100775, |
| "grad_norm": 0.27802011370658875, |
| "learning_rate": 7.954668490503621e-05, |
| "loss": 0.027, |
| "step": 1190 |
| }, |
| { |
| "epoch": 1.8604651162790697, |
| "grad_norm": 0.2340599000453949, |
| "learning_rate": 7.910830014373237e-05, |
| "loss": 0.0283, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.8604651162790697, |
| "eval_loss": 0.029726749286055565, |
| "eval_runtime": 177.8483, |
| "eval_samples_per_second": 6.449, |
| "eval_steps_per_second": 6.449, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.875968992248062, |
| "grad_norm": 0.13624481856822968, |
| "learning_rate": 7.866650410113907e-05, |
| "loss": 0.0243, |
| "step": 1210 |
| }, |
| { |
| "epoch": 1.8914728682170543, |
| "grad_norm": 0.30038079619407654, |
| "learning_rate": 7.822134855254263e-05, |
| "loss": 0.0296, |
| "step": 1220 |
| }, |
| { |
| "epoch": 1.9069767441860463, |
| "grad_norm": 0.23166057467460632, |
| "learning_rate": 7.777288566693905e-05, |
| "loss": 0.0273, |
| "step": 1230 |
| }, |
| { |
| "epoch": 1.9224806201550386, |
| "grad_norm": 0.26034536957740784, |
| "learning_rate": 7.732116800092018e-05, |
| "loss": 0.029, |
| "step": 1240 |
| }, |
| { |
| "epoch": 1.937984496124031, |
| "grad_norm": 0.2269134223461151, |
| "learning_rate": 7.686624849251453e-05, |
| "loss": 0.0381, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.937984496124031, |
| "eval_loss": 0.029867732897400856, |
| "eval_runtime": 177.8147, |
| "eval_samples_per_second": 6.451, |
| "eval_steps_per_second": 6.451, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.9534883720930232, |
| "grad_norm": 0.30028459429740906, |
| "learning_rate": 7.640818045498324e-05, |
| "loss": 0.0378, |
| "step": 1260 |
| }, |
| { |
| "epoch": 1.9689922480620154, |
| "grad_norm": 0.10794702172279358, |
| "learning_rate": 7.59470175705722e-05, |
| "loss": 0.0303, |
| "step": 1270 |
| }, |
| { |
| "epoch": 1.9844961240310077, |
| "grad_norm": 0.27119511365890503, |
| "learning_rate": 7.548281388422088e-05, |
| "loss": 0.0361, |
| "step": 1280 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.22147968411445618, |
| "learning_rate": 7.50156237972286e-05, |
| "loss": 0.0263, |
| "step": 1290 |
| }, |
| { |
| "epoch": 2.0155038759689923, |
| "grad_norm": 0.12922464311122894, |
| "learning_rate": 7.454550206087921e-05, |
| "loss": 0.0207, |
| "step": 1300 |
| }, |
| { |
| "epoch": 2.0155038759689923, |
| "eval_loss": 0.029372563585639, |
| "eval_runtime": 177.7077, |
| "eval_samples_per_second": 6.454, |
| "eval_steps_per_second": 6.454, |
| "step": 1300 |
| }, |
| { |
| "epoch": 2.0310077519379846, |
| "grad_norm": 0.20252908766269684, |
| "learning_rate": 7.407250377002451e-05, |
| "loss": 0.0184, |
| "step": 1310 |
| }, |
| { |
| "epoch": 2.046511627906977, |
| "grad_norm": 0.20628713071346283, |
| "learning_rate": 7.359668435662757e-05, |
| "loss": 0.0216, |
| "step": 1320 |
| }, |
| { |
| "epoch": 2.062015503875969, |
| "grad_norm": 0.16820959746837616, |
| "learning_rate": 7.311809958326659e-05, |
| "loss": 0.0153, |
| "step": 1330 |
| }, |
| { |
| "epoch": 2.0775193798449614, |
| "grad_norm": 0.28085780143737793, |
| "learning_rate": 7.263680553659983e-05, |
| "loss": 0.0172, |
| "step": 1340 |
| }, |
| { |
| "epoch": 2.0930232558139537, |
| "grad_norm": 0.4147201180458069, |
| "learning_rate": 7.21528586207927e-05, |
| "loss": 0.0163, |
| "step": 1350 |
| }, |
| { |
| "epoch": 2.0930232558139537, |
| "eval_loss": 0.03293036296963692, |
| "eval_runtime": 177.2709, |
| "eval_samples_per_second": 6.47, |
| "eval_steps_per_second": 6.47, |
| "step": 1350 |
| }, |
| { |
| "epoch": 2.108527131782946, |
| "grad_norm": 0.2516096830368042, |
| "learning_rate": 7.166631555090759e-05, |
| "loss": 0.025, |
| "step": 1360 |
| }, |
| { |
| "epoch": 2.124031007751938, |
| "grad_norm": 0.26919105648994446, |
| "learning_rate": 7.117723334625731e-05, |
| "loss": 0.024, |
| "step": 1370 |
| }, |
| { |
| "epoch": 2.13953488372093, |
| "grad_norm": 0.228290855884552, |
| "learning_rate": 7.068566932372278e-05, |
| "loss": 0.0183, |
| "step": 1380 |
| }, |
| { |
| "epoch": 2.1550387596899223, |
| "grad_norm": 0.28446343541145325, |
| "learning_rate": 7.019168109103599e-05, |
| "loss": 0.0162, |
| "step": 1390 |
| }, |
| { |
| "epoch": 2.1705426356589146, |
| "grad_norm": 0.19133096933364868, |
| "learning_rate": 6.96953265400287e-05, |
| "loss": 0.0236, |
| "step": 1400 |
| }, |
| { |
| "epoch": 2.1705426356589146, |
| "eval_loss": 0.031053410843014717, |
| "eval_runtime": 176.3297, |
| "eval_samples_per_second": 6.505, |
| "eval_steps_per_second": 6.505, |
| "step": 1400 |
| }, |
| { |
| "epoch": 2.186046511627907, |
| "grad_norm": 0.14185263216495514, |
| "learning_rate": 6.9196663839848e-05, |
| "loss": 0.0142, |
| "step": 1410 |
| }, |
| { |
| "epoch": 2.201550387596899, |
| "grad_norm": 0.2277035117149353, |
| "learning_rate": 6.869575143013934e-05, |
| "loss": 0.0166, |
| "step": 1420 |
| }, |
| { |
| "epoch": 2.2170542635658914, |
| "grad_norm": 0.24037382006645203, |
| "learning_rate": 6.819264801419771e-05, |
| "loss": 0.0274, |
| "step": 1430 |
| }, |
| { |
| "epoch": 2.2325581395348837, |
| "grad_norm": 0.10074973106384277, |
| "learning_rate": 6.768741255208813e-05, |
| "loss": 0.0154, |
| "step": 1440 |
| }, |
| { |
| "epoch": 2.248062015503876, |
| "grad_norm": 0.24921472370624542, |
| "learning_rate": 6.718010425373599e-05, |
| "loss": 0.0191, |
| "step": 1450 |
| }, |
| { |
| "epoch": 2.248062015503876, |
| "eval_loss": 0.030957119539380074, |
| "eval_runtime": 175.6681, |
| "eval_samples_per_second": 6.529, |
| "eval_steps_per_second": 6.529, |
| "step": 1450 |
| }, |
| { |
| "epoch": 2.2635658914728682, |
| "grad_norm": 0.22070886194705963, |
| "learning_rate": 6.667078257198796e-05, |
| "loss": 0.0184, |
| "step": 1460 |
| }, |
| { |
| "epoch": 2.2790697674418605, |
| "grad_norm": 0.39727696776390076, |
| "learning_rate": 6.615950719564466e-05, |
| "loss": 0.0142, |
| "step": 1470 |
| }, |
| { |
| "epoch": 2.294573643410853, |
| "grad_norm": 0.31478437781333923, |
| "learning_rate": 6.56463380424655e-05, |
| "loss": 0.0252, |
| "step": 1480 |
| }, |
| { |
| "epoch": 2.310077519379845, |
| "grad_norm": 0.1986107975244522, |
| "learning_rate": 6.51313352521468e-05, |
| "loss": 0.0151, |
| "step": 1490 |
| }, |
| { |
| "epoch": 2.3255813953488373, |
| "grad_norm": 0.3045531213283539, |
| "learning_rate": 6.46145591792738e-05, |
| "loss": 0.0243, |
| "step": 1500 |
| }, |
| { |
| "epoch": 2.3255813953488373, |
| "eval_loss": 0.03084620088338852, |
| "eval_runtime": 175.2668, |
| "eval_samples_per_second": 6.544, |
| "eval_steps_per_second": 6.544, |
| "step": 1500 |
| }, |
| { |
| "epoch": 2.3410852713178296, |
| "grad_norm": 0.08667182922363281, |
| "learning_rate": 6.409607038624759e-05, |
| "loss": 0.0142, |
| "step": 1510 |
| }, |
| { |
| "epoch": 2.356589147286822, |
| "grad_norm": 0.17333078384399414, |
| "learning_rate": 6.357592963618761e-05, |
| "loss": 0.0241, |
| "step": 1520 |
| }, |
| { |
| "epoch": 2.3720930232558137, |
| "grad_norm": 0.20276910066604614, |
| "learning_rate": 6.305419788581067e-05, |
| "loss": 0.0185, |
| "step": 1530 |
| }, |
| { |
| "epoch": 2.387596899224806, |
| "grad_norm": 0.287349134683609, |
| "learning_rate": 6.253093627828724e-05, |
| "loss": 0.0176, |
| "step": 1540 |
| }, |
| { |
| "epoch": 2.4031007751937983, |
| "grad_norm": 0.32019197940826416, |
| "learning_rate": 6.200620613607597e-05, |
| "loss": 0.0165, |
| "step": 1550 |
| }, |
| { |
| "epoch": 2.4031007751937983, |
| "eval_loss": 0.03274994716048241, |
| "eval_runtime": 174.5607, |
| "eval_samples_per_second": 6.571, |
| "eval_steps_per_second": 6.571, |
| "step": 1550 |
| }, |
| { |
| "epoch": 2.4186046511627906, |
| "grad_norm": 0.34726911783218384, |
| "learning_rate": 6.148006895373706e-05, |
| "loss": 0.0196, |
| "step": 1560 |
| }, |
| { |
| "epoch": 2.434108527131783, |
| "grad_norm": 0.15877018868923187, |
| "learning_rate": 6.0952586390725644e-05, |
| "loss": 0.0171, |
| "step": 1570 |
| }, |
| { |
| "epoch": 2.449612403100775, |
| "grad_norm": 0.21150444447994232, |
| "learning_rate": 6.042382026416563e-05, |
| "loss": 0.0164, |
| "step": 1580 |
| }, |
| { |
| "epoch": 2.4651162790697674, |
| "grad_norm": 0.19103744626045227, |
| "learning_rate": 5.989383254160529e-05, |
| "loss": 0.0171, |
| "step": 1590 |
| }, |
| { |
| "epoch": 2.4806201550387597, |
| "grad_norm": 0.44117987155914307, |
| "learning_rate": 5.936268533375506e-05, |
| "loss": 0.0224, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.4806201550387597, |
| "eval_loss": 0.03294115141034126, |
| "eval_runtime": 174.4995, |
| "eval_samples_per_second": 6.573, |
| "eval_steps_per_second": 6.573, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.496124031007752, |
| "grad_norm": 0.3606320321559906, |
| "learning_rate": 5.883044088720865e-05, |
| "loss": 0.0214, |
| "step": 1610 |
| }, |
| { |
| "epoch": 2.511627906976744, |
| "grad_norm": 0.13814851641654968, |
| "learning_rate": 5.829716157714814e-05, |
| "loss": 0.0177, |
| "step": 1620 |
| }, |
| { |
| "epoch": 2.5271317829457365, |
| "grad_norm": 0.23484356701374054, |
| "learning_rate": 5.77629099000341e-05, |
| "loss": 0.0263, |
| "step": 1630 |
| }, |
| { |
| "epoch": 2.5426356589147288, |
| "grad_norm": 0.21404676139354706, |
| "learning_rate": 5.7227748466281486e-05, |
| "loss": 0.0187, |
| "step": 1640 |
| }, |
| { |
| "epoch": 2.558139534883721, |
| "grad_norm": 0.3276318609714508, |
| "learning_rate": 5.669173999292208e-05, |
| "loss": 0.0289, |
| "step": 1650 |
| }, |
| { |
| "epoch": 2.558139534883721, |
| "eval_loss": 0.03190125152468681, |
| "eval_runtime": 174.4148, |
| "eval_samples_per_second": 6.576, |
| "eval_steps_per_second": 6.576, |
| "step": 1650 |
| }, |
| { |
| "epoch": 2.5736434108527133, |
| "grad_norm": 0.41611889004707336, |
| "learning_rate": 5.615494729625458e-05, |
| "loss": 0.0237, |
| "step": 1660 |
| }, |
| { |
| "epoch": 2.5891472868217056, |
| "grad_norm": 0.372302383184433, |
| "learning_rate": 5.561743328448296e-05, |
| "loss": 0.0246, |
| "step": 1670 |
| }, |
| { |
| "epoch": 2.604651162790698, |
| "grad_norm": 0.1654350608587265, |
| "learning_rate": 5.5079260950344035e-05, |
| "loss": 0.019, |
| "step": 1680 |
| }, |
| { |
| "epoch": 2.62015503875969, |
| "grad_norm": 0.09135215729475021, |
| "learning_rate": 5.454049336372531e-05, |
| "loss": 0.0165, |
| "step": 1690 |
| }, |
| { |
| "epoch": 2.6356589147286824, |
| "grad_norm": 0.24403300881385803, |
| "learning_rate": 5.4001193664273454e-05, |
| "loss": 0.014, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.6356589147286824, |
| "eval_loss": 0.03161098435521126, |
| "eval_runtime": 174.2648, |
| "eval_samples_per_second": 6.582, |
| "eval_steps_per_second": 6.582, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.6511627906976747, |
| "grad_norm": 0.19514554738998413, |
| "learning_rate": 5.346142505399495e-05, |
| "loss": 0.0198, |
| "step": 1710 |
| }, |
| { |
| "epoch": 2.6666666666666665, |
| "grad_norm": 0.13203318417072296, |
| "learning_rate": 5.292125078984925e-05, |
| "loss": 0.0181, |
| "step": 1720 |
| }, |
| { |
| "epoch": 2.682170542635659, |
| "grad_norm": 0.09618645161390305, |
| "learning_rate": 5.2380734176335425e-05, |
| "loss": 0.0171, |
| "step": 1730 |
| }, |
| { |
| "epoch": 2.697674418604651, |
| "grad_norm": 0.17840640246868134, |
| "learning_rate": 5.183993855807343e-05, |
| "loss": 0.0186, |
| "step": 1740 |
| }, |
| { |
| "epoch": 2.7131782945736433, |
| "grad_norm": 0.12565214931964874, |
| "learning_rate": 5.1298927312380586e-05, |
| "loss": 0.0182, |
| "step": 1750 |
| }, |
| { |
| "epoch": 2.7131782945736433, |
| "eval_loss": 0.03337998315691948, |
| "eval_runtime": 174.2061, |
| "eval_samples_per_second": 6.584, |
| "eval_steps_per_second": 6.584, |
| "step": 1750 |
| }, |
| { |
| "epoch": 2.7286821705426356, |
| "grad_norm": 0.1932082623243332, |
| "learning_rate": 5.075776384184411e-05, |
| "loss": 0.0275, |
| "step": 1760 |
| }, |
| { |
| "epoch": 2.744186046511628, |
| "grad_norm": 0.14624637365341187, |
| "learning_rate": 5.021651156689094e-05, |
| "loss": 0.0284, |
| "step": 1770 |
| }, |
| { |
| "epoch": 2.75968992248062, |
| "grad_norm": 0.20881560444831848, |
| "learning_rate": 4.967523391835521e-05, |
| "loss": 0.0217, |
| "step": 1780 |
| }, |
| { |
| "epoch": 2.7751937984496124, |
| "grad_norm": 0.2833938002586365, |
| "learning_rate": 4.9133994330044644e-05, |
| "loss": 0.0234, |
| "step": 1790 |
| }, |
| { |
| "epoch": 2.7906976744186047, |
| "grad_norm": 0.23201899230480194, |
| "learning_rate": 4.85928562313066e-05, |
| "loss": 0.0175, |
| "step": 1800 |
| }, |
| { |
| "epoch": 2.7906976744186047, |
| "eval_loss": 0.029777053743600845, |
| "eval_runtime": 174.1005, |
| "eval_samples_per_second": 6.588, |
| "eval_steps_per_second": 6.588, |
| "step": 1800 |
| }, |
| { |
| "epoch": 2.806201550387597, |
| "grad_norm": 0.2119215726852417, |
| "learning_rate": 4.8051883039594616e-05, |
| "loss": 0.0151, |
| "step": 1810 |
| }, |
| { |
| "epoch": 2.8217054263565893, |
| "grad_norm": 0.30587702989578247, |
| "learning_rate": 4.751113815303624e-05, |
| "loss": 0.0187, |
| "step": 1820 |
| }, |
| { |
| "epoch": 2.8372093023255816, |
| "grad_norm": 0.18018727004528046, |
| "learning_rate": 4.697068494300343e-05, |
| "loss": 0.0151, |
| "step": 1830 |
| }, |
| { |
| "epoch": 2.8527131782945734, |
| "grad_norm": 0.21248644590377808, |
| "learning_rate": 4.6430586746685724e-05, |
| "loss": 0.0222, |
| "step": 1840 |
| }, |
| { |
| "epoch": 2.8682170542635657, |
| "grad_norm": 0.18737761676311493, |
| "learning_rate": 4.589090685966758e-05, |
| "loss": 0.0218, |
| "step": 1850 |
| }, |
| { |
| "epoch": 2.8682170542635657, |
| "eval_loss": 0.029682571068406105, |
| "eval_runtime": 174.0789, |
| "eval_samples_per_second": 6.589, |
| "eval_steps_per_second": 6.589, |
| "step": 1850 |
| }, |
| { |
| "epoch": 2.883720930232558, |
| "grad_norm": 0.3078942894935608, |
| "learning_rate": 4.535170852851073e-05, |
| "loss": 0.0265, |
| "step": 1860 |
| }, |
| { |
| "epoch": 2.89922480620155, |
| "grad_norm": 0.2036602944135666, |
| "learning_rate": 4.481305494334201e-05, |
| "loss": 0.0167, |
| "step": 1870 |
| }, |
| { |
| "epoch": 2.9147286821705425, |
| "grad_norm": 0.08146923035383224, |
| "learning_rate": 4.427500923044801e-05, |
| "loss": 0.0199, |
| "step": 1880 |
| }, |
| { |
| "epoch": 2.9302325581395348, |
| "grad_norm": 0.12505419552326202, |
| "learning_rate": 4.373763444487705e-05, |
| "loss": 0.0179, |
| "step": 1890 |
| }, |
| { |
| "epoch": 2.945736434108527, |
| "grad_norm": 0.18068896234035492, |
| "learning_rate": 4.3200993563049725e-05, |
| "loss": 0.018, |
| "step": 1900 |
| }, |
| { |
| "epoch": 2.945736434108527, |
| "eval_loss": 0.02887474000453949, |
| "eval_runtime": 173.8738, |
| "eval_samples_per_second": 6.597, |
| "eval_steps_per_second": 6.597, |
| "step": 1900 |
| }, |
| { |
| "epoch": 2.9612403100775193, |
| "grad_norm": 0.19300121068954468, |
| "learning_rate": 4.266514947537839e-05, |
| "loss": 0.0157, |
| "step": 1910 |
| }, |
| { |
| "epoch": 2.9767441860465116, |
| "grad_norm": 0.1891685128211975, |
| "learning_rate": 4.2130164978896916e-05, |
| "loss": 0.0188, |
| "step": 1920 |
| }, |
| { |
| "epoch": 2.992248062015504, |
| "grad_norm": 0.027009131386876106, |
| "learning_rate": 4.159610276990137e-05, |
| "loss": 0.0118, |
| "step": 1930 |
| }, |
| { |
| "epoch": 3.007751937984496, |
| "grad_norm": 0.11215299367904663, |
| "learning_rate": 4.106302543660235e-05, |
| "loss": 0.0164, |
| "step": 1940 |
| }, |
| { |
| "epoch": 3.0232558139534884, |
| "grad_norm": 0.07163859158754349, |
| "learning_rate": 4.053099545179028e-05, |
| "loss": 0.01, |
| "step": 1950 |
| }, |
| { |
| "epoch": 3.0232558139534884, |
| "eval_loss": 0.030872860923409462, |
| "eval_runtime": 173.945, |
| "eval_samples_per_second": 6.594, |
| "eval_steps_per_second": 6.594, |
| "step": 1950 |
| }, |
| { |
| "epoch": 3.0387596899224807, |
| "grad_norm": 0.12158221006393433, |
| "learning_rate": 4.0000075165513845e-05, |
| "loss": 0.014, |
| "step": 1960 |
| }, |
| { |
| "epoch": 3.054263565891473, |
| "grad_norm": 0.17903171479701996, |
| "learning_rate": 3.9470326797773216e-05, |
| "loss": 0.0085, |
| "step": 1970 |
| }, |
| { |
| "epoch": 3.0697674418604652, |
| "grad_norm": 0.23486894369125366, |
| "learning_rate": 3.8941812431228166e-05, |
| "loss": 0.012, |
| "step": 1980 |
| }, |
| { |
| "epoch": 3.0852713178294575, |
| "grad_norm": 0.19141437113285065, |
| "learning_rate": 3.8414594003922515e-05, |
| "loss": 0.0141, |
| "step": 1990 |
| }, |
| { |
| "epoch": 3.10077519379845, |
| "grad_norm": 0.047276828438043594, |
| "learning_rate": 3.788873330202544e-05, |
| "loss": 0.0109, |
| "step": 2000 |
| }, |
| { |
| "epoch": 3.10077519379845, |
| "eval_loss": 0.033768460154533386, |
| "eval_runtime": 173.8687, |
| "eval_samples_per_second": 6.597, |
| "eval_steps_per_second": 6.597, |
| "step": 2000 |
| }, |
| { |
| "epoch": 3.116279069767442, |
| "grad_norm": 0.3079964518547058, |
| "learning_rate": 3.736429195259051e-05, |
| "loss": 0.0069, |
| "step": 2010 |
| }, |
| { |
| "epoch": 3.1317829457364343, |
| "grad_norm": 0.207097128033638, |
| "learning_rate": 3.684133141633358e-05, |
| "loss": 0.009, |
| "step": 2020 |
| }, |
| { |
| "epoch": 3.147286821705426, |
| "grad_norm": 0.16703131794929504, |
| "learning_rate": 3.6319912980429846e-05, |
| "loss": 0.0098, |
| "step": 2030 |
| }, |
| { |
| "epoch": 3.1627906976744184, |
| "grad_norm": 0.1511267125606537, |
| "learning_rate": 3.580009775133168e-05, |
| "loss": 0.0076, |
| "step": 2040 |
| }, |
| { |
| "epoch": 3.1782945736434107, |
| "grad_norm": 0.184535413980484, |
| "learning_rate": 3.528194664760714e-05, |
| "loss": 0.0076, |
| "step": 2050 |
| }, |
| { |
| "epoch": 3.1782945736434107, |
| "eval_loss": 0.03470822051167488, |
| "eval_runtime": 173.7828, |
| "eval_samples_per_second": 6.6, |
| "eval_steps_per_second": 6.6, |
| "step": 2050 |
| }, |
| { |
| "epoch": 3.193798449612403, |
| "grad_norm": 0.19592364132404327, |
| "learning_rate": 3.476552039280096e-05, |
| "loss": 0.0089, |
| "step": 2060 |
| }, |
| { |
| "epoch": 3.2093023255813953, |
| "grad_norm": 0.1496165543794632, |
| "learning_rate": 3.42508795083181e-05, |
| "loss": 0.0092, |
| "step": 2070 |
| }, |
| { |
| "epoch": 3.2248062015503876, |
| "grad_norm": 0.19381049275398254, |
| "learning_rate": 3.373808430633106e-05, |
| "loss": 0.0086, |
| "step": 2080 |
| }, |
| { |
| "epoch": 3.24031007751938, |
| "grad_norm": 0.30487295985221863, |
| "learning_rate": 3.32271948827118e-05, |
| "loss": 0.0099, |
| "step": 2090 |
| }, |
| { |
| "epoch": 3.255813953488372, |
| "grad_norm": 0.21724191308021545, |
| "learning_rate": 3.2718271109988863e-05, |
| "loss": 0.0087, |
| "step": 2100 |
| }, |
| { |
| "epoch": 3.255813953488372, |
| "eval_loss": 0.03577126935124397, |
| "eval_runtime": 173.7017, |
| "eval_samples_per_second": 6.603, |
| "eval_steps_per_second": 6.603, |
| "step": 2100 |
| }, |
| { |
| "epoch": 3.2713178294573644, |
| "grad_norm": 0.8536518812179565, |
| "learning_rate": 3.2211372630330835e-05, |
| "loss": 0.0094, |
| "step": 2110 |
| }, |
| { |
| "epoch": 3.2868217054263567, |
| "grad_norm": 0.23373013734817505, |
| "learning_rate": 3.170655884855661e-05, |
| "loss": 0.0121, |
| "step": 2120 |
| }, |
| { |
| "epoch": 3.302325581395349, |
| "grad_norm": 0.09981165081262589, |
| "learning_rate": 3.120388892517368e-05, |
| "loss": 0.0193, |
| "step": 2130 |
| }, |
| { |
| "epoch": 3.317829457364341, |
| "grad_norm": 0.2072608470916748, |
| "learning_rate": 3.070342176944494e-05, |
| "loss": 0.0135, |
| "step": 2140 |
| }, |
| { |
| "epoch": 3.3333333333333335, |
| "grad_norm": 0.07687430083751678, |
| "learning_rate": 3.0205216032484805e-05, |
| "loss": 0.0092, |
| "step": 2150 |
| }, |
| { |
| "epoch": 3.3333333333333335, |
| "eval_loss": 0.032335054129362106, |
| "eval_runtime": 173.7494, |
| "eval_samples_per_second": 6.601, |
| "eval_steps_per_second": 6.601, |
| "step": 2150 |
| }, |
| { |
| "epoch": 3.3488372093023258, |
| "grad_norm": 0.2054612785577774, |
| "learning_rate": 2.970933010038599e-05, |
| "loss": 0.013, |
| "step": 2160 |
| }, |
| { |
| "epoch": 3.3643410852713176, |
| "grad_norm": 0.17977559566497803, |
| "learning_rate": 2.921582208737681e-05, |
| "loss": 0.0088, |
| "step": 2170 |
| }, |
| { |
| "epoch": 3.37984496124031, |
| "grad_norm": 0.23377564549446106, |
| "learning_rate": 2.872474982901081e-05, |
| "loss": 0.0136, |
| "step": 2180 |
| }, |
| { |
| "epoch": 3.395348837209302, |
| "grad_norm": 0.16202200949192047, |
| "learning_rate": 2.8236170875388744e-05, |
| "loss": 0.01, |
| "step": 2190 |
| }, |
| { |
| "epoch": 3.4108527131782944, |
| "grad_norm": 0.2739998698234558, |
| "learning_rate": 2.775014248441422e-05, |
| "loss": 0.0078, |
| "step": 2200 |
| }, |
| { |
| "epoch": 3.4108527131782944, |
| "eval_loss": 0.03312241658568382, |
| "eval_runtime": 173.7097, |
| "eval_samples_per_second": 6.603, |
| "eval_steps_per_second": 6.603, |
| "step": 2200 |
| }, |
| { |
| "epoch": 3.4263565891472867, |
| "grad_norm": 0.10180757939815521, |
| "learning_rate": 2.726672161508341e-05, |
| "loss": 0.0098, |
| "step": 2210 |
| }, |
| { |
| "epoch": 3.441860465116279, |
| "grad_norm": 0.27830150723457336, |
| "learning_rate": 2.678596492080984e-05, |
| "loss": 0.0077, |
| "step": 2220 |
| }, |
| { |
| "epoch": 3.4573643410852712, |
| "grad_norm": 0.25859320163726807, |
| "learning_rate": 2.630792874278516e-05, |
| "loss": 0.0099, |
| "step": 2230 |
| }, |
| { |
| "epoch": 3.4728682170542635, |
| "grad_norm": 0.4165274500846863, |
| "learning_rate": 2.583266910337624e-05, |
| "loss": 0.0095, |
| "step": 2240 |
| }, |
| { |
| "epoch": 3.488372093023256, |
| "grad_norm": 0.5020641684532166, |
| "learning_rate": 2.5360241699559816e-05, |
| "loss": 0.0109, |
| "step": 2250 |
| }, |
| { |
| "epoch": 3.488372093023256, |
| "eval_loss": 0.035617418587207794, |
| "eval_runtime": 173.6138, |
| "eval_samples_per_second": 6.607, |
| "eval_steps_per_second": 6.607, |
| "step": 2250 |
| }, |
| { |
| "epoch": 3.503875968992248, |
| "grad_norm": 0.19337324798107147, |
| "learning_rate": 2.4890701896395146e-05, |
| "loss": 0.0104, |
| "step": 2260 |
| }, |
| { |
| "epoch": 3.5193798449612403, |
| "grad_norm": 0.1879683881998062, |
| "learning_rate": 2.4424104720535735e-05, |
| "loss": 0.0101, |
| "step": 2270 |
| }, |
| { |
| "epoch": 3.5348837209302326, |
| "grad_norm": 0.063876673579216, |
| "learning_rate": 2.3960504853780462e-05, |
| "loss": 0.0115, |
| "step": 2280 |
| }, |
| { |
| "epoch": 3.550387596899225, |
| "grad_norm": 0.11843185126781464, |
| "learning_rate": 2.349995662666547e-05, |
| "loss": 0.0085, |
| "step": 2290 |
| }, |
| { |
| "epoch": 3.565891472868217, |
| "grad_norm": 0.21422308683395386, |
| "learning_rate": 2.3042514012096843e-05, |
| "loss": 0.0137, |
| "step": 2300 |
| }, |
| { |
| "epoch": 3.565891472868217, |
| "eval_loss": 0.035960037261247635, |
| "eval_runtime": 173.6653, |
| "eval_samples_per_second": 6.605, |
| "eval_steps_per_second": 6.605, |
| "step": 2300 |
| }, |
| { |
| "epoch": 3.5813953488372094, |
| "grad_norm": 0.354839563369751, |
| "learning_rate": 2.2588230619025407e-05, |
| "loss": 0.0159, |
| "step": 2310 |
| }, |
| { |
| "epoch": 3.5968992248062017, |
| "grad_norm": 0.2568114995956421, |
| "learning_rate": 2.213715968616425e-05, |
| "loss": 0.0109, |
| "step": 2320 |
| }, |
| { |
| "epoch": 3.612403100775194, |
| "grad_norm": 0.06852281093597412, |
| "learning_rate": 2.16893540757494e-05, |
| "loss": 0.0091, |
| "step": 2330 |
| }, |
| { |
| "epoch": 3.6279069767441863, |
| "grad_norm": 0.194693461060524, |
| "learning_rate": 2.1244866267344866e-05, |
| "loss": 0.0084, |
| "step": 2340 |
| }, |
| { |
| "epoch": 3.6434108527131785, |
| "grad_norm": 0.3488699495792389, |
| "learning_rate": 2.080374835169235e-05, |
| "loss": 0.013, |
| "step": 2350 |
| }, |
| { |
| "epoch": 3.6434108527131785, |
| "eval_loss": 0.03496725112199783, |
| "eval_runtime": 173.6559, |
| "eval_samples_per_second": 6.605, |
| "eval_steps_per_second": 6.605, |
| "step": 2350 |
| }, |
| { |
| "epoch": 3.6589147286821704, |
| "grad_norm": 0.16774657368659973, |
| "learning_rate": 2.0366052024606612e-05, |
| "loss": 0.0082, |
| "step": 2360 |
| }, |
| { |
| "epoch": 3.6744186046511627, |
| "grad_norm": 0.2563646733760834, |
| "learning_rate": 1.9931828580917107e-05, |
| "loss": 0.0114, |
| "step": 2370 |
| }, |
| { |
| "epoch": 3.689922480620155, |
| "grad_norm": 0.254089891910553, |
| "learning_rate": 1.9501128908456523e-05, |
| "loss": 0.0087, |
| "step": 2380 |
| }, |
| { |
| "epoch": 3.705426356589147, |
| "grad_norm": 0.2673187255859375, |
| "learning_rate": 1.9074003482097258e-05, |
| "loss": 0.0093, |
| "step": 2390 |
| }, |
| { |
| "epoch": 3.7209302325581395, |
| "grad_norm": 0.47327151894569397, |
| "learning_rate": 1.8650502357835925e-05, |
| "loss": 0.0133, |
| "step": 2400 |
| }, |
| { |
| "epoch": 3.7209302325581395, |
| "eval_loss": 0.03533458709716797, |
| "eval_runtime": 173.6028, |
| "eval_samples_per_second": 6.607, |
| "eval_steps_per_second": 6.607, |
| "step": 2400 |
| }, |
| { |
| "epoch": 3.7364341085271318, |
| "grad_norm": 0.4366307556629181, |
| "learning_rate": 1.82306751669274e-05, |
| "loss": 0.0085, |
| "step": 2410 |
| }, |
| { |
| "epoch": 3.751937984496124, |
| "grad_norm": 0.39723125100135803, |
| "learning_rate": 1.7814571110068135e-05, |
| "loss": 0.0123, |
| "step": 2420 |
| }, |
| { |
| "epoch": 3.7674418604651163, |
| "grad_norm": 0.16450275480747223, |
| "learning_rate": 1.740223895163039e-05, |
| "loss": 0.0091, |
| "step": 2430 |
| }, |
| { |
| "epoch": 3.7829457364341086, |
| "grad_norm": 0.06844169646501541, |
| "learning_rate": 1.6993727013947336e-05, |
| "loss": 0.0109, |
| "step": 2440 |
| }, |
| { |
| "epoch": 3.798449612403101, |
| "grad_norm": 0.29116564989089966, |
| "learning_rate": 1.6589083171649977e-05, |
| "loss": 0.0068, |
| "step": 2450 |
| }, |
| { |
| "epoch": 3.798449612403101, |
| "eval_loss": 0.035682931542396545, |
| "eval_runtime": 173.5613, |
| "eval_samples_per_second": 6.609, |
| "eval_steps_per_second": 6.609, |
| "step": 2450 |
| }, |
| { |
| "epoch": 3.813953488372093, |
| "grad_norm": 0.1012692078948021, |
| "learning_rate": 1.6188354846056698e-05, |
| "loss": 0.008, |
| "step": 2460 |
| }, |
| { |
| "epoch": 3.8294573643410854, |
| "grad_norm": 0.1340951770544052, |
| "learning_rate": 1.579158899961575e-05, |
| "loss": 0.0054, |
| "step": 2470 |
| }, |
| { |
| "epoch": 3.8449612403100772, |
| "grad_norm": 0.22337476909160614, |
| "learning_rate": 1.5398832130401637e-05, |
| "loss": 0.0147, |
| "step": 2480 |
| }, |
| { |
| "epoch": 3.8604651162790695, |
| "grad_norm": 0.18346606194972992, |
| "learning_rate": 1.5010130266665807e-05, |
| "loss": 0.0108, |
| "step": 2490 |
| }, |
| { |
| "epoch": 3.875968992248062, |
| "grad_norm": 0.2582853436470032, |
| "learning_rate": 1.4625528961442591e-05, |
| "loss": 0.012, |
| "step": 2500 |
| }, |
| { |
| "epoch": 3.875968992248062, |
| "eval_loss": 0.034770578145980835, |
| "eval_runtime": 173.5234, |
| "eval_samples_per_second": 6.61, |
| "eval_steps_per_second": 6.61, |
| "step": 2500 |
| }, |
| { |
| "epoch": 3.891472868217054, |
| "grad_norm": 0.22837190330028534, |
| "learning_rate": 1.4245073287210608e-05, |
| "loss": 0.0098, |
| "step": 2510 |
| }, |
| { |
| "epoch": 3.9069767441860463, |
| "grad_norm": 0.49539613723754883, |
| "learning_rate": 1.3868807830610653e-05, |
| "loss": 0.016, |
| "step": 2520 |
| }, |
| { |
| "epoch": 3.9224806201550386, |
| "grad_norm": 0.17993128299713135, |
| "learning_rate": 1.3496776687220514e-05, |
| "loss": 0.0129, |
| "step": 2530 |
| }, |
| { |
| "epoch": 3.937984496124031, |
| "grad_norm": 0.25489673018455505, |
| "learning_rate": 1.3129023456387151e-05, |
| "loss": 0.0068, |
| "step": 2540 |
| }, |
| { |
| "epoch": 3.953488372093023, |
| "grad_norm": 0.06681143492460251, |
| "learning_rate": 1.2765591236117324e-05, |
| "loss": 0.0088, |
| "step": 2550 |
| }, |
| { |
| "epoch": 3.953488372093023, |
| "eval_loss": 0.034369926899671555, |
| "eval_runtime": 173.574, |
| "eval_samples_per_second": 6.608, |
| "eval_steps_per_second": 6.608, |
| "step": 2550 |
| }, |
| { |
| "epoch": 3.9689922480620154, |
| "grad_norm": 0.14447127282619476, |
| "learning_rate": 1.240652261802669e-05, |
| "loss": 0.0105, |
| "step": 2560 |
| }, |
| { |
| "epoch": 3.9844961240310077, |
| "grad_norm": 0.2608015239238739, |
| "learning_rate": 1.205185968234847e-05, |
| "loss": 0.0136, |
| "step": 2570 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 0.17687876522541046, |
| "learning_rate": 1.1701643993001916e-05, |
| "loss": 0.0093, |
| "step": 2580 |
| }, |
| { |
| "epoch": 4.015503875968992, |
| "grad_norm": 0.1070009246468544, |
| "learning_rate": 1.1355916592721316e-05, |
| "loss": 0.0045, |
| "step": 2590 |
| }, |
| { |
| "epoch": 4.0310077519379846, |
| "grad_norm": 0.11302992701530457, |
| "learning_rate": 1.1014717998246099e-05, |
| "loss": 0.0066, |
| "step": 2600 |
| }, |
| { |
| "epoch": 4.0310077519379846, |
| "eval_loss": 0.03459889069199562, |
| "eval_runtime": 173.5445, |
| "eval_samples_per_second": 6.609, |
| "eval_steps_per_second": 6.609, |
| "step": 2600 |
| }, |
| { |
| "epoch": 4.046511627906977, |
| "grad_norm": 0.107364721596241, |
| "learning_rate": 1.0678088195572517e-05, |
| "loss": 0.0053, |
| "step": 2610 |
| }, |
| { |
| "epoch": 4.062015503875969, |
| "grad_norm": 0.2821330428123474, |
| "learning_rate": 1.0346066635267676e-05, |
| "loss": 0.0069, |
| "step": 2620 |
| }, |
| { |
| "epoch": 4.077519379844961, |
| "grad_norm": 0.17317578196525574, |
| "learning_rate": 1.001869222784611e-05, |
| "loss": 0.0064, |
| "step": 2630 |
| }, |
| { |
| "epoch": 4.093023255813954, |
| "grad_norm": 0.08701343834400177, |
| "learning_rate": 9.696003339209819e-06, |
| "loss": 0.007, |
| "step": 2640 |
| }, |
| { |
| "epoch": 4.108527131782946, |
| "grad_norm": 0.23731686174869537, |
| "learning_rate": 9.378037786152055e-06, |
| "loss": 0.0052, |
| "step": 2650 |
| }, |
| { |
| "epoch": 4.108527131782946, |
| "eval_loss": 0.03613542765378952, |
| "eval_runtime": 173.8232, |
| "eval_samples_per_second": 6.599, |
| "eval_steps_per_second": 6.599, |
| "step": 2650 |
| }, |
| { |
| "epoch": 4.124031007751938, |
| "grad_norm": 0.13787880539894104, |
| "learning_rate": 9.064832831925441e-06, |
| "loss": 0.0054, |
| "step": 2660 |
| }, |
| { |
| "epoch": 4.1395348837209305, |
| "grad_norm": 0.1381884068250656, |
| "learning_rate": 8.756425181875028e-06, |
| "loss": 0.0035, |
| "step": 2670 |
| }, |
| { |
| "epoch": 4.155038759689923, |
| "grad_norm": 0.16911053657531738, |
| "learning_rate": 8.452850979136617e-06, |
| "loss": 0.0045, |
| "step": 2680 |
| }, |
| { |
| "epoch": 4.170542635658915, |
| "grad_norm": 0.04475264623761177, |
| "learning_rate": 8.15414580040117e-06, |
| "loss": 0.0032, |
| "step": 2690 |
| }, |
| { |
| "epoch": 4.186046511627907, |
| "grad_norm": 0.09773126989603043, |
| "learning_rate": 7.860344651745366e-06, |
| "loss": 0.008, |
| "step": 2700 |
| }, |
| { |
| "epoch": 4.186046511627907, |
| "eval_loss": 0.037369657307863235, |
| "eval_runtime": 173.9253, |
| "eval_samples_per_second": 6.595, |
| "eval_steps_per_second": 6.595, |
| "step": 2700 |
| }, |
| { |
| "epoch": 4.2015503875969, |
| "grad_norm": 0.1751956194639206, |
| "learning_rate": 7.571481964529226e-06, |
| "loss": 0.0057, |
| "step": 2710 |
| }, |
| { |
| "epoch": 4.217054263565892, |
| "grad_norm": 0.06625787168741226, |
| "learning_rate": 7.287591591360926e-06, |
| "loss": 0.0028, |
| "step": 2720 |
| }, |
| { |
| "epoch": 4.232558139534884, |
| "grad_norm": 0.1029408872127533, |
| "learning_rate": 7.008706802129606e-06, |
| "loss": 0.0069, |
| "step": 2730 |
| }, |
| { |
| "epoch": 4.248062015503876, |
| "grad_norm": 0.01715020090341568, |
| "learning_rate": 6.734860280106292e-06, |
| "loss": 0.0021, |
| "step": 2740 |
| }, |
| { |
| "epoch": 4.263565891472869, |
| "grad_norm": 0.5773080587387085, |
| "learning_rate": 6.466084118113735e-06, |
| "loss": 0.0062, |
| "step": 2750 |
| }, |
| { |
| "epoch": 4.263565891472869, |
| "eval_loss": 0.0383269228041172, |
| "eval_runtime": 173.7893, |
| "eval_samples_per_second": 6.6, |
| "eval_steps_per_second": 6.6, |
| "step": 2750 |
| }, |
| { |
| "epoch": 4.27906976744186, |
| "grad_norm": 0.09289313852787018, |
| "learning_rate": 6.202409814765325e-06, |
| "loss": 0.0043, |
| "step": 2760 |
| }, |
| { |
| "epoch": 4.294573643410852, |
| "grad_norm": 0.16086533665657043, |
| "learning_rate": 5.9438682707736725e-06, |
| "loss": 0.0064, |
| "step": 2770 |
| }, |
| { |
| "epoch": 4.310077519379845, |
| "grad_norm": 0.175432026386261, |
| "learning_rate": 5.690489785329301e-06, |
| "loss": 0.0052, |
| "step": 2780 |
| }, |
| { |
| "epoch": 4.325581395348837, |
| "grad_norm": 0.03287180885672569, |
| "learning_rate": 5.442304052549752e-06, |
| "loss": 0.0037, |
| "step": 2790 |
| }, |
| { |
| "epoch": 4.341085271317829, |
| "grad_norm": 0.09991113096475601, |
| "learning_rate": 5.199340157999733e-06, |
| "loss": 0.005, |
| "step": 2800 |
| }, |
| { |
| "epoch": 4.341085271317829, |
| "eval_loss": 0.038646504282951355, |
| "eval_runtime": 173.5908, |
| "eval_samples_per_second": 6.607, |
| "eval_steps_per_second": 6.607, |
| "step": 2800 |
| }, |
| { |
| "epoch": 4.3565891472868215, |
| "grad_norm": 0.07446504384279251, |
| "learning_rate": 4.961626575282396e-06, |
| "loss": 0.0028, |
| "step": 2810 |
| }, |
| { |
| "epoch": 4.372093023255814, |
| "grad_norm": 0.1563124656677246, |
| "learning_rate": 4.7291911627025235e-06, |
| "loss": 0.0045, |
| "step": 2820 |
| }, |
| { |
| "epoch": 4.387596899224806, |
| "grad_norm": 0.09954715520143509, |
| "learning_rate": 4.502061160001725e-06, |
| "loss": 0.0038, |
| "step": 2830 |
| }, |
| { |
| "epoch": 4.403100775193798, |
| "grad_norm": 0.03728143870830536, |
| "learning_rate": 4.280263185166078e-06, |
| "loss": 0.007, |
| "step": 2840 |
| }, |
| { |
| "epoch": 4.4186046511627906, |
| "grad_norm": 0.14879867434501648, |
| "learning_rate": 4.063823231306757e-06, |
| "loss": 0.004, |
| "step": 2850 |
| }, |
| { |
| "epoch": 4.4186046511627906, |
| "eval_loss": 0.03950056806206703, |
| "eval_runtime": 173.7546, |
| "eval_samples_per_second": 6.601, |
| "eval_steps_per_second": 6.601, |
| "step": 2850 |
| }, |
| { |
| "epoch": 4.434108527131783, |
| "grad_norm": 0.04011745750904083, |
| "learning_rate": 3.8527666636137885e-06, |
| "loss": 0.0039, |
| "step": 2860 |
| }, |
| { |
| "epoch": 4.449612403100775, |
| "grad_norm": 0.16904973983764648, |
| "learning_rate": 3.6471182163834825e-06, |
| "loss": 0.0075, |
| "step": 2870 |
| }, |
| { |
| "epoch": 4.465116279069767, |
| "grad_norm": 0.05882781371474266, |
| "learning_rate": 3.4469019901197054e-06, |
| "loss": 0.0027, |
| "step": 2880 |
| }, |
| { |
| "epoch": 4.48062015503876, |
| "grad_norm": 0.36288192868232727, |
| "learning_rate": 3.252141448709495e-06, |
| "loss": 0.0052, |
| "step": 2890 |
| }, |
| { |
| "epoch": 4.496124031007752, |
| "grad_norm": 0.25636863708496094, |
| "learning_rate": 3.0628594166732693e-06, |
| "loss": 0.0075, |
| "step": 2900 |
| }, |
| { |
| "epoch": 4.496124031007752, |
| "eval_loss": 0.040012069046497345, |
| "eval_runtime": 173.5684, |
| "eval_samples_per_second": 6.608, |
| "eval_steps_per_second": 6.608, |
| "step": 2900 |
| }, |
| { |
| "epoch": 4.511627906976744, |
| "grad_norm": 0.07002587616443634, |
| "learning_rate": 2.8790780764899384e-06, |
| "loss": 0.0053, |
| "step": 2910 |
| }, |
| { |
| "epoch": 4.5271317829457365, |
| "grad_norm": 0.18481026589870453, |
| "learning_rate": 2.700818965997315e-06, |
| "loss": 0.0064, |
| "step": 2920 |
| }, |
| { |
| "epoch": 4.542635658914729, |
| "grad_norm": 0.1123792976140976, |
| "learning_rate": 2.528102975867991e-06, |
| "loss": 0.0033, |
| "step": 2930 |
| }, |
| { |
| "epoch": 4.558139534883721, |
| "grad_norm": 0.039597250521183014, |
| "learning_rate": 2.3609503471611284e-06, |
| "loss": 0.0041, |
| "step": 2940 |
| }, |
| { |
| "epoch": 4.573643410852713, |
| "grad_norm": 0.05195772275328636, |
| "learning_rate": 2.1993806689503738e-06, |
| "loss": 0.003, |
| "step": 2950 |
| }, |
| { |
| "epoch": 4.573643410852713, |
| "eval_loss": 0.040207840502262115, |
| "eval_runtime": 173.5139, |
| "eval_samples_per_second": 6.61, |
| "eval_steps_per_second": 6.61, |
| "step": 2950 |
| }, |
| { |
| "epoch": 4.589147286821706, |
| "grad_norm": 0.06371759623289108, |
| "learning_rate": 2.0434128760281167e-06, |
| "loss": 0.0021, |
| "step": 2960 |
| }, |
| { |
| "epoch": 4.604651162790698, |
| "grad_norm": 0.0519726499915123, |
| "learning_rate": 1.893065246686504e-06, |
| "loss": 0.0029, |
| "step": 2970 |
| }, |
| { |
| "epoch": 4.62015503875969, |
| "grad_norm": 0.05901438370347023, |
| "learning_rate": 1.74835540057533e-06, |
| "loss": 0.0067, |
| "step": 2980 |
| }, |
| { |
| "epoch": 4.635658914728682, |
| "grad_norm": 0.16322703659534454, |
| "learning_rate": 1.6093002966371617e-06, |
| "loss": 0.0044, |
| "step": 2990 |
| }, |
| { |
| "epoch": 4.651162790697675, |
| "grad_norm": 0.25651177763938904, |
| "learning_rate": 1.4759162311198783e-06, |
| "loss": 0.0066, |
| "step": 3000 |
| }, |
| { |
| "epoch": 4.651162790697675, |
| "eval_loss": 0.0405086986720562, |
| "eval_runtime": 173.4642, |
| "eval_samples_per_second": 6.612, |
| "eval_steps_per_second": 6.612, |
| "step": 3000 |
| }, |
| { |
| "epoch": 4.666666666666667, |
| "grad_norm": 0.009503856301307678, |
| "learning_rate": 1.348218835666859e-06, |
| "loss": 0.0032, |
| "step": 3010 |
| }, |
| { |
| "epoch": 4.682170542635659, |
| "grad_norm": 0.12524276971817017, |
| "learning_rate": 1.2262230754850445e-06, |
| "loss": 0.0032, |
| "step": 3020 |
| }, |
| { |
| "epoch": 4.6976744186046515, |
| "grad_norm": 0.06170937791466713, |
| "learning_rate": 1.109943247591172e-06, |
| "loss": 0.0045, |
| "step": 3030 |
| }, |
| { |
| "epoch": 4.713178294573644, |
| "grad_norm": 0.12521466612815857, |
| "learning_rate": 9.993929791362323e-07, |
| "loss": 0.0052, |
| "step": 3040 |
| }, |
| { |
| "epoch": 4.728682170542635, |
| "grad_norm": 0.17083865404129028, |
| "learning_rate": 8.945852258084863e-07, |
| "loss": 0.005, |
| "step": 3050 |
| }, |
| { |
| "epoch": 4.728682170542635, |
| "eval_loss": 0.04061457887291908, |
| "eval_runtime": 173.5187, |
| "eval_samples_per_second": 6.61, |
| "eval_steps_per_second": 6.61, |
| "step": 3050 |
| }, |
| { |
| "epoch": 4.7441860465116275, |
| "grad_norm": 0.35159188508987427, |
| "learning_rate": 7.955322703151358e-07, |
| "loss": 0.0046, |
| "step": 3060 |
| }, |
| { |
| "epoch": 4.75968992248062, |
| "grad_norm": 0.2449195384979248, |
| "learning_rate": 7.022457209428901e-07, |
| "loss": 0.0044, |
| "step": 3070 |
| }, |
| { |
| "epoch": 4.775193798449612, |
| "grad_norm": 0.2619343101978302, |
| "learning_rate": 6.147365101975666e-07, |
| "loss": 0.0035, |
| "step": 3080 |
| }, |
| { |
| "epoch": 4.790697674418604, |
| "grad_norm": 0.09077528119087219, |
| "learning_rate": 5.33014893522854e-07, |
| "loss": 0.0032, |
| "step": 3090 |
| }, |
| { |
| "epoch": 4.8062015503875966, |
| "grad_norm": 0.14284414052963257, |
| "learning_rate": 4.57090448098485e-07, |
| "loss": 0.0067, |
| "step": 3100 |
| }, |
| { |
| "epoch": 4.8062015503875966, |
| "eval_loss": 0.04071947932243347, |
| "eval_runtime": 173.7232, |
| "eval_samples_per_second": 6.602, |
| "eval_steps_per_second": 6.602, |
| "step": 3100 |
| }, |
| { |
| "epoch": 4.821705426356589, |
| "grad_norm": 0.10482044517993927, |
| "learning_rate": 3.8697207171781714e-07, |
| "loss": 0.0037, |
| "step": 3110 |
| }, |
| { |
| "epoch": 4.837209302325581, |
| "grad_norm": 0.03253592178225517, |
| "learning_rate": 3.2266798174512837e-07, |
| "loss": 0.0045, |
| "step": 3120 |
| }, |
| { |
| "epoch": 4.852713178294573, |
| "grad_norm": 0.03622818365693092, |
| "learning_rate": 2.6418571415255387e-07, |
| "loss": 0.0093, |
| "step": 3130 |
| }, |
| { |
| "epoch": 4.868217054263566, |
| "grad_norm": 0.09220151603221893, |
| "learning_rate": 2.1153212263695378e-07, |
| "loss": 0.0046, |
| "step": 3140 |
| }, |
| { |
| "epoch": 4.883720930232558, |
| "grad_norm": 0.1942395716905594, |
| "learning_rate": 1.6471337781669982e-07, |
| "loss": 0.0067, |
| "step": 3150 |
| }, |
| { |
| "epoch": 4.883720930232558, |
| "eval_loss": 0.040707945823669434, |
| "eval_runtime": 173.7396, |
| "eval_samples_per_second": 6.602, |
| "eval_steps_per_second": 6.602, |
| "step": 3150 |
| }, |
| { |
| "epoch": 4.89922480620155, |
| "grad_norm": 0.09173876792192459, |
| "learning_rate": 1.237349665085097e-07, |
| "loss": 0.0049, |
| "step": 3160 |
| }, |
| { |
| "epoch": 4.9147286821705425, |
| "grad_norm": 0.045584529638290405, |
| "learning_rate": 8.86016910844667e-08, |
| "loss": 0.003, |
| "step": 3170 |
| }, |
| { |
| "epoch": 4.930232558139535, |
| "grad_norm": 0.46576347947120667, |
| "learning_rate": 5.9317668909192323e-08, |
| "loss": 0.0061, |
| "step": 3180 |
| }, |
| { |
| "epoch": 4.945736434108527, |
| "grad_norm": 0.059987977147102356, |
| "learning_rate": 3.588633185730994e-08, |
| "loss": 0.0062, |
| "step": 3190 |
| }, |
| { |
| "epoch": 4.961240310077519, |
| "grad_norm": 0.2531404197216034, |
| "learning_rate": 1.8310425911294283e-08, |
| "loss": 0.006, |
| "step": 3200 |
| }, |
| { |
| "epoch": 4.961240310077519, |
| "eval_loss": 0.04073048010468483, |
| "eval_runtime": 173.54, |
| "eval_samples_per_second": 6.609, |
| "eval_steps_per_second": 6.609, |
| "step": 3200 |
| }, |
| { |
| "epoch": 4.976744186046512, |
| "grad_norm": 0.24996767938137054, |
| "learning_rate": 6.592010839612251e-09, |
| "loss": 0.0076, |
| "step": 3210 |
| }, |
| { |
| "epoch": 4.992248062015504, |
| "grad_norm": 0.020084669813513756, |
| "learning_rate": 7.324599553770739e-10, |
| "loss": 0.0025, |
| "step": 3220 |
| }, |
| { |
| "epoch": 5.0, |
| "step": 3225, |
| "total_flos": 8.718478050646426e+17, |
| "train_loss": 0.049027663960821866, |
| "train_runtime": 36714.2196, |
| "train_samples_per_second": 1.405, |
| "train_steps_per_second": 0.088 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 3225, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 50, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 8.718478050646426e+17, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|