| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.7684295008623486, |
| "eval_steps": 10000.0, |
| "global_step": 90000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 8.538105565137207e-06, |
| "grad_norm": 6.5811052322387695, |
| "learning_rate": 1.7073587160662455e-08, |
| "loss": 0.9425575733184814, |
| "memory(GiB)": 4.21, |
| "step": 1, |
| "train_speed(iter/s)": 0.107952 |
| }, |
| { |
| "epoch": 0.004269052782568603, |
| "grad_norm": 0.775414764881134, |
| "learning_rate": 8.536793580331227e-06, |
| "loss": 0.2293025076030968, |
| "memory(GiB)": 5.71, |
| "step": 500, |
| "train_speed(iter/s)": 0.938128 |
| }, |
| { |
| "epoch": 0.008538105565137207, |
| "grad_norm": 0.5124035477638245, |
| "learning_rate": 1.7073587160662455e-05, |
| "loss": 0.05007815933227539, |
| "memory(GiB)": 5.71, |
| "step": 1000, |
| "train_speed(iter/s)": 0.946051 |
| }, |
| { |
| "epoch": 0.012807158347705811, |
| "grad_norm": 0.34229570627212524, |
| "learning_rate": 2.5610380740993682e-05, |
| "loss": 0.036965576171875, |
| "memory(GiB)": 5.71, |
| "step": 1500, |
| "train_speed(iter/s)": 0.980261 |
| }, |
| { |
| "epoch": 0.017076211130274414, |
| "grad_norm": 0.29022017121315, |
| "learning_rate": 3.414717432132491e-05, |
| "loss": 0.030718547821044922, |
| "memory(GiB)": 5.71, |
| "step": 2000, |
| "train_speed(iter/s)": 1.013563 |
| }, |
| { |
| "epoch": 0.02134526391284302, |
| "grad_norm": 0.2701531946659088, |
| "learning_rate": 4.268396790165614e-05, |
| "loss": 0.026950979232788087, |
| "memory(GiB)": 5.71, |
| "step": 2500, |
| "train_speed(iter/s)": 1.038614 |
| }, |
| { |
| "epoch": 0.025614316695411622, |
| "grad_norm": 0.1944938600063324, |
| "learning_rate": 5.1220761481987364e-05, |
| "loss": 0.025618917465209962, |
| "memory(GiB)": 5.71, |
| "step": 3000, |
| "train_speed(iter/s)": 1.081032 |
| }, |
| { |
| "epoch": 0.029883369477980225, |
| "grad_norm": 0.15646104514598846, |
| "learning_rate": 5.9757555062318595e-05, |
| "loss": 0.0239105281829834, |
| "memory(GiB)": 5.71, |
| "step": 3500, |
| "train_speed(iter/s)": 1.023247 |
| }, |
| { |
| "epoch": 0.03415242226054883, |
| "grad_norm": 0.12373720854520798, |
| "learning_rate": 6.829434864264982e-05, |
| "loss": 0.021082000732421877, |
| "memory(GiB)": 5.71, |
| "step": 4000, |
| "train_speed(iter/s)": 0.971265 |
| }, |
| { |
| "epoch": 0.03842147504311743, |
| "grad_norm": 0.11245039105415344, |
| "learning_rate": 7.683114222298106e-05, |
| "loss": 0.020011337280273437, |
| "memory(GiB)": 5.71, |
| "step": 4500, |
| "train_speed(iter/s)": 0.934224 |
| }, |
| { |
| "epoch": 0.04269052782568604, |
| "grad_norm": 0.12453507632017136, |
| "learning_rate": 8.536793580331228e-05, |
| "loss": 0.019226245880126953, |
| "memory(GiB)": 5.71, |
| "step": 5000, |
| "train_speed(iter/s)": 0.906199 |
| }, |
| { |
| "epoch": 0.04695958060825464, |
| "grad_norm": 0.08364573866128922, |
| "learning_rate": 9.39047293836435e-05, |
| "loss": 0.018570756912231444, |
| "memory(GiB)": 5.71, |
| "step": 5500, |
| "train_speed(iter/s)": 0.887629 |
| }, |
| { |
| "epoch": 0.051228633390823244, |
| "grad_norm": 0.08422534167766571, |
| "learning_rate": 9.999959243761552e-05, |
| "loss": 0.018008022308349608, |
| "memory(GiB)": 5.71, |
| "step": 6000, |
| "train_speed(iter/s)": 0.882382 |
| }, |
| { |
| "epoch": 0.05549768617339185, |
| "grad_norm": 0.0722428560256958, |
| "learning_rate": 9.999175989726746e-05, |
| "loss": 0.016920005798339844, |
| "memory(GiB)": 5.71, |
| "step": 6500, |
| "train_speed(iter/s)": 0.887128 |
| }, |
| { |
| "epoch": 0.05976673895596045, |
| "grad_norm": 0.0696873590350151, |
| "learning_rate": 9.997396381339795e-05, |
| "loss": 0.016275758743286132, |
| "memory(GiB)": 5.71, |
| "step": 7000, |
| "train_speed(iter/s)": 0.896874 |
| }, |
| { |
| "epoch": 0.06403579173852905, |
| "grad_norm": 0.0715385377407074, |
| "learning_rate": 9.994620773283261e-05, |
| "loss": 0.03211669921875, |
| "memory(GiB)": 5.71, |
| "step": 7500, |
| "train_speed(iter/s)": 0.906916 |
| }, |
| { |
| "epoch": 0.06830484452109765, |
| "grad_norm": 0.07314470410346985, |
| "learning_rate": 9.990849718746144e-05, |
| "loss": 0.014735573768615723, |
| "memory(GiB)": 5.71, |
| "step": 8000, |
| "train_speed(iter/s)": 0.915054 |
| }, |
| { |
| "epoch": 0.07257389730366626, |
| "grad_norm": 0.0624830424785614, |
| "learning_rate": 9.986083969313632e-05, |
| "loss": 0.014247347831726074, |
| "memory(GiB)": 5.71, |
| "step": 8500, |
| "train_speed(iter/s)": 0.923566 |
| }, |
| { |
| "epoch": 0.07684295008623486, |
| "grad_norm": 0.054315660148859024, |
| "learning_rate": 9.980324474817292e-05, |
| "loss": 0.014038642883300782, |
| "memory(GiB)": 5.71, |
| "step": 9000, |
| "train_speed(iter/s)": 0.931957 |
| }, |
| { |
| "epoch": 0.08111200286880348, |
| "grad_norm": 0.04966143146157265, |
| "learning_rate": 9.973572383145782e-05, |
| "loss": 0.013697422027587891, |
| "memory(GiB)": 5.71, |
| "step": 9500, |
| "train_speed(iter/s)": 0.939331 |
| }, |
| { |
| "epoch": 0.08538105565137208, |
| "grad_norm": 0.055357128381729126, |
| "learning_rate": 9.965829040016061e-05, |
| "loss": 0.013380534172058105, |
| "memory(GiB)": 5.71, |
| "step": 10000, |
| "train_speed(iter/s)": 0.946434 |
| }, |
| { |
| "epoch": 0.08965010843394068, |
| "grad_norm": 0.05234465003013611, |
| "learning_rate": 9.957095988705193e-05, |
| "loss": 0.013177488327026367, |
| "memory(GiB)": 5.71, |
| "step": 10500, |
| "train_speed(iter/s)": 0.960222 |
| }, |
| { |
| "epoch": 0.09391916121650928, |
| "grad_norm": 0.053024690598249435, |
| "learning_rate": 9.947374969742755e-05, |
| "loss": 0.013030742645263672, |
| "memory(GiB)": 5.71, |
| "step": 11000, |
| "train_speed(iter/s)": 0.973308 |
| }, |
| { |
| "epoch": 0.09818821399907789, |
| "grad_norm": 0.04746083542704582, |
| "learning_rate": 9.936667920563951e-05, |
| "loss": 0.012802671432495118, |
| "memory(GiB)": 5.71, |
| "step": 11500, |
| "train_speed(iter/s)": 0.985248 |
| }, |
| { |
| "epoch": 0.10245726678164649, |
| "grad_norm": 0.045355405658483505, |
| "learning_rate": 9.924976975123472e-05, |
| "loss": 0.012366548538208007, |
| "memory(GiB)": 5.71, |
| "step": 12000, |
| "train_speed(iter/s)": 0.996787 |
| }, |
| { |
| "epoch": 0.10672631956421509, |
| "grad_norm": 0.0430605448782444, |
| "learning_rate": 9.912304463470185e-05, |
| "loss": 0.018324718475341797, |
| "memory(GiB)": 5.71, |
| "step": 12500, |
| "train_speed(iter/s)": 1.007619 |
| }, |
| { |
| "epoch": 0.1109953723467837, |
| "grad_norm": 0.03760723024606705, |
| "learning_rate": 9.89865291128276e-05, |
| "loss": 0.011288459777832032, |
| "memory(GiB)": 5.71, |
| "step": 13000, |
| "train_speed(iter/s)": 1.017831 |
| }, |
| { |
| "epoch": 0.1152644251293523, |
| "grad_norm": 0.04123725742101669, |
| "learning_rate": 9.884025039366274e-05, |
| "loss": 0.011067386627197265, |
| "memory(GiB)": 5.71, |
| "step": 13500, |
| "train_speed(iter/s)": 1.027479 |
| }, |
| { |
| "epoch": 0.1195334779119209, |
| "grad_norm": 0.046554189175367355, |
| "learning_rate": 9.868423763109962e-05, |
| "loss": 0.010972289085388184, |
| "memory(GiB)": 5.71, |
| "step": 14000, |
| "train_speed(iter/s)": 1.0366 |
| }, |
| { |
| "epoch": 0.1238025306944895, |
| "grad_norm": 0.04529291018843651, |
| "learning_rate": 9.851852191906155e-05, |
| "loss": 0.010918002128601074, |
| "memory(GiB)": 5.71, |
| "step": 14500, |
| "train_speed(iter/s)": 1.045276 |
| }, |
| { |
| "epoch": 0.1280715834770581, |
| "grad_norm": 0.039391227066516876, |
| "learning_rate": 9.834313628530574e-05, |
| "loss": 0.010901000022888184, |
| "memory(GiB)": 5.71, |
| "step": 15000, |
| "train_speed(iter/s)": 1.053467 |
| }, |
| { |
| "epoch": 0.13234063625962672, |
| "grad_norm": 0.050082143396139145, |
| "learning_rate": 9.81581156848408e-05, |
| "loss": 0.010822629928588868, |
| "memory(GiB)": 5.71, |
| "step": 15500, |
| "train_speed(iter/s)": 1.061046 |
| }, |
| { |
| "epoch": 0.1366096890421953, |
| "grad_norm": 0.04181528836488724, |
| "learning_rate": 9.79634969929599e-05, |
| "loss": 0.01061639404296875, |
| "memory(GiB)": 5.71, |
| "step": 16000, |
| "train_speed(iter/s)": 1.068454 |
| }, |
| { |
| "epoch": 0.14087874182476393, |
| "grad_norm": 0.037736013531684875, |
| "learning_rate": 9.775931899789159e-05, |
| "loss": 0.01050139808654785, |
| "memory(GiB)": 5.71, |
| "step": 16500, |
| "train_speed(iter/s)": 1.075513 |
| }, |
| { |
| "epoch": 0.14514779460733251, |
| "grad_norm": 0.03446267917752266, |
| "learning_rate": 9.754562239306902e-05, |
| "loss": 0.010301560401916503, |
| "memory(GiB)": 5.71, |
| "step": 17000, |
| "train_speed(iter/s)": 1.08225 |
| }, |
| { |
| "epoch": 0.14941684738990113, |
| "grad_norm": 0.035218626260757446, |
| "learning_rate": 9.732244976901965e-05, |
| "loss": 0.010299022674560548, |
| "memory(GiB)": 5.71, |
| "step": 17500, |
| "train_speed(iter/s)": 1.088656 |
| }, |
| { |
| "epoch": 0.15368590017246972, |
| "grad_norm": 0.034140028059482574, |
| "learning_rate": 9.708984560487677e-05, |
| "loss": 0.010085094451904296, |
| "memory(GiB)": 5.71, |
| "step": 18000, |
| "train_speed(iter/s)": 1.094769 |
| }, |
| { |
| "epoch": 0.15795495295503834, |
| "grad_norm": 0.040227197110652924, |
| "learning_rate": 9.684785625951468e-05, |
| "loss": 0.009981593132019044, |
| "memory(GiB)": 5.71, |
| "step": 18500, |
| "train_speed(iter/s)": 1.100633 |
| }, |
| { |
| "epoch": 0.16222400573760695, |
| "grad_norm": 0.03376320004463196, |
| "learning_rate": 9.659652996230917e-05, |
| "loss": 0.009874713897705079, |
| "memory(GiB)": 5.71, |
| "step": 19000, |
| "train_speed(iter/s)": 1.106243 |
| }, |
| { |
| "epoch": 0.16649305852017554, |
| "grad_norm": 0.03337237238883972, |
| "learning_rate": 9.633591680352522e-05, |
| "loss": 0.009621439933776855, |
| "memory(GiB)": 5.71, |
| "step": 19500, |
| "train_speed(iter/s)": 1.111407 |
| }, |
| { |
| "epoch": 0.17076211130274416, |
| "grad_norm": 0.03156784921884537, |
| "learning_rate": 9.606606872433384e-05, |
| "loss": 0.01175856876373291, |
| "memory(GiB)": 5.73, |
| "step": 20000, |
| "train_speed(iter/s)": 1.116578 |
| }, |
| { |
| "epoch": 0.17503116408531275, |
| "grad_norm": 0.06789804250001907, |
| "learning_rate": 9.578703950645998e-05, |
| "loss": 0.008876850128173828, |
| "memory(GiB)": 5.73, |
| "step": 20500, |
| "train_speed(iter/s)": 1.121386 |
| }, |
| { |
| "epoch": 0.17930021686788136, |
| "grad_norm": 0.029559865593910217, |
| "learning_rate": 9.549888476146366e-05, |
| "loss": 0.008808825492858887, |
| "memory(GiB)": 5.73, |
| "step": 21000, |
| "train_speed(iter/s)": 1.126144 |
| }, |
| { |
| "epoch": 0.18356926965044995, |
| "grad_norm": 0.027149997651576996, |
| "learning_rate": 9.52016619196564e-05, |
| "loss": 0.008746042251586914, |
| "memory(GiB)": 5.73, |
| "step": 21500, |
| "train_speed(iter/s)": 1.130706 |
| }, |
| { |
| "epoch": 0.18783832243301857, |
| "grad_norm": 0.031270887702703476, |
| "learning_rate": 9.489543021865507e-05, |
| "loss": 0.008727970123291016, |
| "memory(GiB)": 5.73, |
| "step": 22000, |
| "train_speed(iter/s)": 1.135093 |
| }, |
| { |
| "epoch": 0.19210737521558716, |
| "grad_norm": 0.03134565427899361, |
| "learning_rate": 9.458025069157563e-05, |
| "loss": 0.008822738647460937, |
| "memory(GiB)": 5.73, |
| "step": 22500, |
| "train_speed(iter/s)": 1.139318 |
| }, |
| { |
| "epoch": 0.19637642799815577, |
| "grad_norm": 0.02970048598945141, |
| "learning_rate": 9.425618615486908e-05, |
| "loss": 0.008724775314331055, |
| "memory(GiB)": 5.73, |
| "step": 23000, |
| "train_speed(iter/s)": 1.143306 |
| }, |
| { |
| "epoch": 0.20064548078072436, |
| "grad_norm": 0.037413984537124634, |
| "learning_rate": 9.392330119580186e-05, |
| "loss": 0.008617961883544922, |
| "memory(GiB)": 5.73, |
| "step": 23500, |
| "train_speed(iter/s)": 1.147167 |
| }, |
| { |
| "epoch": 0.20491453356329298, |
| "grad_norm": 0.031085532158613205, |
| "learning_rate": 9.358166215958333e-05, |
| "loss": 0.008613507270812988, |
| "memory(GiB)": 5.73, |
| "step": 24000, |
| "train_speed(iter/s)": 1.150974 |
| }, |
| { |
| "epoch": 0.2091835863458616, |
| "grad_norm": 0.030068758875131607, |
| "learning_rate": 9.323133713614297e-05, |
| "loss": 0.008516620635986329, |
| "memory(GiB)": 5.73, |
| "step": 24500, |
| "train_speed(iter/s)": 1.154635 |
| }, |
| { |
| "epoch": 0.21345263912843018, |
| "grad_norm": 0.030049536377191544, |
| "learning_rate": 9.287239594655976e-05, |
| "loss": 0.00915114688873291, |
| "memory(GiB)": 5.73, |
| "step": 25000, |
| "train_speed(iter/s)": 1.158172 |
| }, |
| { |
| "epoch": 0.2177216919109988, |
| "grad_norm": 0.03236347809433937, |
| "learning_rate": 9.250491012914668e-05, |
| "loss": 0.008387946128845214, |
| "memory(GiB)": 5.73, |
| "step": 25500, |
| "train_speed(iter/s)": 1.161598 |
| }, |
| { |
| "epoch": 0.2219907446935674, |
| "grad_norm": 0.028227701783180237, |
| "learning_rate": 9.212895292519276e-05, |
| "loss": 0.008091423034667969, |
| "memory(GiB)": 5.73, |
| "step": 26000, |
| "train_speed(iter/s)": 1.164914 |
| }, |
| { |
| "epoch": 0.226259797476136, |
| "grad_norm": 0.026642831042408943, |
| "learning_rate": 9.17445992643658e-05, |
| "loss": 0.008073025703430176, |
| "memory(GiB)": 5.73, |
| "step": 26500, |
| "train_speed(iter/s)": 1.168129 |
| }, |
| { |
| "epoch": 0.2305288502587046, |
| "grad_norm": 0.029216019436717033, |
| "learning_rate": 9.135192574977873e-05, |
| "loss": 0.008088951110839843, |
| "memory(GiB)": 5.73, |
| "step": 27000, |
| "train_speed(iter/s)": 1.169593 |
| }, |
| { |
| "epoch": 0.2347979030412732, |
| "grad_norm": 0.02682262659072876, |
| "learning_rate": 9.09510106427222e-05, |
| "loss": 0.007971211433410645, |
| "memory(GiB)": 5.73, |
| "step": 27500, |
| "train_speed(iter/s)": 1.172618 |
| }, |
| { |
| "epoch": 0.2390669558238418, |
| "grad_norm": 0.027296727523207664, |
| "learning_rate": 9.054193384706688e-05, |
| "loss": 0.007928550243377686, |
| "memory(GiB)": 5.73, |
| "step": 28000, |
| "train_speed(iter/s)": 1.165921 |
| }, |
| { |
| "epoch": 0.24333600860641041, |
| "grad_norm": 0.03066374734044075, |
| "learning_rate": 9.012477689333834e-05, |
| "loss": 0.007805256366729736, |
| "memory(GiB)": 5.73, |
| "step": 28500, |
| "train_speed(iter/s)": 1.155767 |
| }, |
| { |
| "epoch": 0.247605061388979, |
| "grad_norm": 0.027467776089906693, |
| "learning_rate": 8.96996229224676e-05, |
| "loss": 0.007825798034667968, |
| "memory(GiB)": 5.73, |
| "step": 29000, |
| "train_speed(iter/s)": 1.147611 |
| }, |
| { |
| "epoch": 0.2518741141715476, |
| "grad_norm": 0.028940001502633095, |
| "learning_rate": 8.926655666922102e-05, |
| "loss": 0.007748476028442383, |
| "memory(GiB)": 5.73, |
| "step": 29500, |
| "train_speed(iter/s)": 1.139656 |
| }, |
| { |
| "epoch": 0.2561431669541162, |
| "grad_norm": 0.0281364805996418, |
| "learning_rate": 8.882566444531216e-05, |
| "loss": 0.007644564628601074, |
| "memory(GiB)": 5.73, |
| "step": 30000, |
| "train_speed(iter/s)": 1.131607 |
| }, |
| { |
| "epoch": 0.2604122197366848, |
| "grad_norm": 0.03272758424282074, |
| "learning_rate": 8.837703412219962e-05, |
| "loss": 0.007614383697509766, |
| "memory(GiB)": 5.73, |
| "step": 30500, |
| "train_speed(iter/s)": 1.123907 |
| }, |
| { |
| "epoch": 0.26468127251925344, |
| "grad_norm": 0.033637482672929764, |
| "learning_rate": 8.7920755113574e-05, |
| "loss": 0.007485725402832031, |
| "memory(GiB)": 5.73, |
| "step": 31000, |
| "train_speed(iter/s)": 1.118652 |
| }, |
| { |
| "epoch": 0.26895032530182206, |
| "grad_norm": 0.028677962720394135, |
| "learning_rate": 8.745691835753724e-05, |
| "loss": 0.007466458320617676, |
| "memory(GiB)": 5.73, |
| "step": 31500, |
| "train_speed(iter/s)": 1.114283 |
| }, |
| { |
| "epoch": 0.2732193780843906, |
| "grad_norm": 0.026501238346099854, |
| "learning_rate": 8.698561629847851e-05, |
| "loss": 0.00739455795288086, |
| "memory(GiB)": 5.73, |
| "step": 32000, |
| "train_speed(iter/s)": 1.110807 |
| }, |
| { |
| "epoch": 0.27748843086695923, |
| "grad_norm": 0.03185174614191055, |
| "learning_rate": 8.650694286864957e-05, |
| "loss": 0.007317279815673828, |
| "memory(GiB)": 5.73, |
| "step": 32500, |
| "train_speed(iter/s)": 1.109657 |
| }, |
| { |
| "epoch": 0.28175748364952785, |
| "grad_norm": 0.031619079411029816, |
| "learning_rate": 8.602099346944379e-05, |
| "loss": 0.007236574649810791, |
| "memory(GiB)": 5.73, |
| "step": 33000, |
| "train_speed(iter/s)": 1.109762 |
| }, |
| { |
| "epoch": 0.28602653643209647, |
| "grad_norm": 0.028590602800250053, |
| "learning_rate": 8.552786495238226e-05, |
| "loss": 0.00712824535369873, |
| "memory(GiB)": 5.73, |
| "step": 33500, |
| "train_speed(iter/s)": 1.10959 |
| }, |
| { |
| "epoch": 0.29029558921466503, |
| "grad_norm": 0.026914609596133232, |
| "learning_rate": 8.502765559981091e-05, |
| "loss": 0.007133237838745117, |
| "memory(GiB)": 5.73, |
| "step": 34000, |
| "train_speed(iter/s)": 1.108437 |
| }, |
| { |
| "epoch": 0.29456464199723364, |
| "grad_norm": 0.03186658397316933, |
| "learning_rate": 8.452046510531258e-05, |
| "loss": 0.00705194091796875, |
| "memory(GiB)": 5.73, |
| "step": 34500, |
| "train_speed(iter/s)": 1.111376 |
| }, |
| { |
| "epoch": 0.29883369477980226, |
| "grad_norm": 0.022866345942020416, |
| "learning_rate": 8.400639455383754e-05, |
| "loss": 0.006991560935974121, |
| "memory(GiB)": 5.73, |
| "step": 35000, |
| "train_speed(iter/s)": 1.103171 |
| }, |
| { |
| "epoch": 0.3031027475623709, |
| "grad_norm": 0.027075253427028656, |
| "learning_rate": 8.348554640155709e-05, |
| "loss": 0.006916217803955078, |
| "memory(GiB)": 5.73, |
| "step": 35500, |
| "train_speed(iter/s)": 1.094784 |
| }, |
| { |
| "epoch": 0.30737180034493944, |
| "grad_norm": 0.027598075568675995, |
| "learning_rate": 8.295802445544345e-05, |
| "loss": 0.0068712844848632815, |
| "memory(GiB)": 5.73, |
| "step": 36000, |
| "train_speed(iter/s)": 1.087037 |
| }, |
| { |
| "epoch": 0.31164085312750806, |
| "grad_norm": 0.022508256137371063, |
| "learning_rate": 8.242393385258083e-05, |
| "loss": 0.006878099918365479, |
| "memory(GiB)": 5.73, |
| "step": 36500, |
| "train_speed(iter/s)": 1.080097 |
| }, |
| { |
| "epoch": 0.31590990591007667, |
| "grad_norm": 0.02388549968600273, |
| "learning_rate": 8.188338103921109e-05, |
| "loss": 0.006974416732788086, |
| "memory(GiB)": 5.73, |
| "step": 37000, |
| "train_speed(iter/s)": 1.073446 |
| }, |
| { |
| "epoch": 0.3201789586926453, |
| "grad_norm": 0.025459734722971916, |
| "learning_rate": 8.13364737495187e-05, |
| "loss": 0.0067239184379577635, |
| "memory(GiB)": 5.73, |
| "step": 37500, |
| "train_speed(iter/s)": 1.067642 |
| }, |
| { |
| "epoch": 0.3244480114752139, |
| "grad_norm": 0.023768454790115356, |
| "learning_rate": 8.078332098415881e-05, |
| "loss": 0.006635515213012695, |
| "memory(GiB)": 5.73, |
| "step": 38000, |
| "train_speed(iter/s)": 1.067385 |
| }, |
| { |
| "epoch": 0.32871706425778247, |
| "grad_norm": 0.028179064393043518, |
| "learning_rate": 8.022403298853317e-05, |
| "loss": 0.00661515998840332, |
| "memory(GiB)": 5.73, |
| "step": 38500, |
| "train_speed(iter/s)": 1.067485 |
| }, |
| { |
| "epoch": 0.3329861170403511, |
| "grad_norm": 0.026124022901058197, |
| "learning_rate": 7.965872123081765e-05, |
| "loss": 0.006523737907409668, |
| "memory(GiB)": 5.73, |
| "step": 39000, |
| "train_speed(iter/s)": 1.067501 |
| }, |
| { |
| "epoch": 0.3372551698229197, |
| "grad_norm": 0.02668868564069271, |
| "learning_rate": 7.908749837974632e-05, |
| "loss": 0.006474626541137695, |
| "memory(GiB)": 5.73, |
| "step": 39500, |
| "train_speed(iter/s)": 1.067703 |
| }, |
| { |
| "epoch": 0.3415242226054883, |
| "grad_norm": 0.02441154234111309, |
| "learning_rate": 7.851047828215611e-05, |
| "loss": 0.006419078826904297, |
| "memory(GiB)": 5.73, |
| "step": 40000, |
| "train_speed(iter/s)": 1.06808 |
| }, |
| { |
| "epoch": 0.3457932753880569, |
| "grad_norm": 0.0254750307649374, |
| "learning_rate": 7.792777594029674e-05, |
| "loss": 0.006350691795349121, |
| "memory(GiB)": 5.73, |
| "step": 40500, |
| "train_speed(iter/s)": 1.068683 |
| }, |
| { |
| "epoch": 0.3500623281706255, |
| "grad_norm": 0.027933409437537193, |
| "learning_rate": 7.73395074889103e-05, |
| "loss": 0.006355803966522217, |
| "memory(GiB)": 5.73, |
| "step": 41000, |
| "train_speed(iter/s)": 1.064808 |
| }, |
| { |
| "epoch": 0.3543313809531941, |
| "grad_norm": 0.022553391754627228, |
| "learning_rate": 7.67457901720852e-05, |
| "loss": 0.006336944103240967, |
| "memory(GiB)": 5.73, |
| "step": 41500, |
| "train_speed(iter/s)": 1.060284 |
| }, |
| { |
| "epoch": 0.3586004337357627, |
| "grad_norm": 0.027581321075558662, |
| "learning_rate": 7.614674231988903e-05, |
| "loss": 0.00619974422454834, |
| "memory(GiB)": 5.73, |
| "step": 42000, |
| "train_speed(iter/s)": 1.055658 |
| }, |
| { |
| "epoch": 0.3628694865183313, |
| "grad_norm": 0.02141967974603176, |
| "learning_rate": 7.554248332478485e-05, |
| "loss": 0.006249521732330322, |
| "memory(GiB)": 5.73, |
| "step": 42500, |
| "train_speed(iter/s)": 1.051341 |
| }, |
| { |
| "epoch": 0.3671385393008999, |
| "grad_norm": 0.025843387469649315, |
| "learning_rate": 7.49331336178358e-05, |
| "loss": 0.006162589550018311, |
| "memory(GiB)": 5.73, |
| "step": 43000, |
| "train_speed(iter/s)": 1.046866 |
| }, |
| { |
| "epoch": 0.3714075920834685, |
| "grad_norm": 0.02431940846145153, |
| "learning_rate": 7.431881464470293e-05, |
| "loss": 0.0060729503631591795, |
| "memory(GiB)": 5.73, |
| "step": 43500, |
| "train_speed(iter/s)": 1.042554 |
| }, |
| { |
| "epoch": 0.37567664486603713, |
| "grad_norm": 0.0244905948638916, |
| "learning_rate": 7.369964884144047e-05, |
| "loss": 0.006033665180206299, |
| "memory(GiB)": 5.73, |
| "step": 44000, |
| "train_speed(iter/s)": 1.041578 |
| }, |
| { |
| "epoch": 0.37994569764860575, |
| "grad_norm": 0.02309691719710827, |
| "learning_rate": 7.307575961009385e-05, |
| "loss": 0.006005731582641602, |
| "memory(GiB)": 5.73, |
| "step": 44500, |
| "train_speed(iter/s)": 1.041875 |
| }, |
| { |
| "epoch": 0.3842147504311743, |
| "grad_norm": 0.023321352899074554, |
| "learning_rate": 7.24472712941053e-05, |
| "loss": 0.005931224346160889, |
| "memory(GiB)": 5.73, |
| "step": 45000, |
| "train_speed(iter/s)": 1.044453 |
| }, |
| { |
| "epoch": 0.38848380321374293, |
| "grad_norm": 0.024199847131967545, |
| "learning_rate": 7.181430915353171e-05, |
| "loss": 0.0059114408493041995, |
| "memory(GiB)": 5.73, |
| "step": 45500, |
| "train_speed(iter/s)": 1.047147 |
| }, |
| { |
| "epoch": 0.39275285599631155, |
| "grad_norm": 0.02660815231502056, |
| "learning_rate": 7.117699934007987e-05, |
| "loss": 0.005867915630340576, |
| "memory(GiB)": 5.73, |
| "step": 46000, |
| "train_speed(iter/s)": 1.049798 |
| }, |
| { |
| "epoch": 0.39702190877888016, |
| "grad_norm": 0.02538706362247467, |
| "learning_rate": 7.053546887196391e-05, |
| "loss": 0.005895719528198242, |
| "memory(GiB)": 5.73, |
| "step": 46500, |
| "train_speed(iter/s)": 1.052318 |
| }, |
| { |
| "epoch": 0.4012909615614487, |
| "grad_norm": 0.023992260918021202, |
| "learning_rate": 6.988984560859009e-05, |
| "loss": 0.005823767662048339, |
| "memory(GiB)": 5.73, |
| "step": 47000, |
| "train_speed(iter/s)": 1.054874 |
| }, |
| { |
| "epoch": 0.40556001434401734, |
| "grad_norm": 0.024961460381746292, |
| "learning_rate": 6.924025822507398e-05, |
| "loss": 0.005796549797058105, |
| "memory(GiB)": 5.73, |
| "step": 47500, |
| "train_speed(iter/s)": 1.057392 |
| }, |
| { |
| "epoch": 0.40982906712658596, |
| "grad_norm": 0.026839323341846466, |
| "learning_rate": 6.858683618659509e-05, |
| "loss": 0.0057229394912719726, |
| "memory(GiB)": 5.73, |
| "step": 48000, |
| "train_speed(iter/s)": 1.059871 |
| }, |
| { |
| "epoch": 0.41409811990915457, |
| "grad_norm": 0.026930488646030426, |
| "learning_rate": 6.792970972259381e-05, |
| "loss": 0.005688785552978515, |
| "memory(GiB)": 5.73, |
| "step": 48500, |
| "train_speed(iter/s)": 1.062309 |
| }, |
| { |
| "epoch": 0.4183671726917232, |
| "grad_norm": 0.024773526936769485, |
| "learning_rate": 6.726900980081639e-05, |
| "loss": 0.005612356185913086, |
| "memory(GiB)": 5.73, |
| "step": 49000, |
| "train_speed(iter/s)": 1.06471 |
| }, |
| { |
| "epoch": 0.42263622547429175, |
| "grad_norm": 0.025835830718278885, |
| "learning_rate": 6.660486810121244e-05, |
| "loss": 0.005570381164550781, |
| "memory(GiB)": 5.73, |
| "step": 49500, |
| "train_speed(iter/s)": 1.067072 |
| }, |
| { |
| "epoch": 0.42690527825686037, |
| "grad_norm": 0.028116557747125626, |
| "learning_rate": 6.593741698969073e-05, |
| "loss": 0.005553098201751709, |
| "memory(GiB)": 5.73, |
| "step": 50000, |
| "train_speed(iter/s)": 1.069395 |
| }, |
| { |
| "epoch": 0.431174331039429, |
| "grad_norm": 0.026658741757273674, |
| "learning_rate": 6.526678949173808e-05, |
| "loss": 0.005453477859497071, |
| "memory(GiB)": 5.73, |
| "step": 50500, |
| "train_speed(iter/s)": 1.065559 |
| }, |
| { |
| "epoch": 0.4354433838219976, |
| "grad_norm": 0.02522198110818863, |
| "learning_rate": 6.459311926590695e-05, |
| "loss": 0.005405562877655029, |
| "memory(GiB)": 5.73, |
| "step": 51000, |
| "train_speed(iter/s)": 1.061202 |
| }, |
| { |
| "epoch": 0.43971243660456616, |
| "grad_norm": 0.019938671961426735, |
| "learning_rate": 6.391654057717676e-05, |
| "loss": 0.005375346183776855, |
| "memory(GiB)": 5.73, |
| "step": 51500, |
| "train_speed(iter/s)": 1.05697 |
| }, |
| { |
| "epoch": 0.4439814893871348, |
| "grad_norm": 0.02449255809187889, |
| "learning_rate": 6.32371882701944e-05, |
| "loss": 0.00538975715637207, |
| "memory(GiB)": 5.73, |
| "step": 52000, |
| "train_speed(iter/s)": 1.053086 |
| }, |
| { |
| "epoch": 0.4482505421697034, |
| "grad_norm": 0.027349578216671944, |
| "learning_rate": 6.25551977423992e-05, |
| "loss": 0.005338613510131836, |
| "memory(GiB)": 5.73, |
| "step": 52500, |
| "train_speed(iter/s)": 1.049655 |
| }, |
| { |
| "epoch": 0.452519594952272, |
| "grad_norm": 0.02677008882164955, |
| "learning_rate": 6.187070491703767e-05, |
| "loss": 0.005392338752746582, |
| "memory(GiB)": 5.73, |
| "step": 53000, |
| "train_speed(iter/s)": 1.046434 |
| }, |
| { |
| "epoch": 0.45678864773484057, |
| "grad_norm": 0.021387379616498947, |
| "learning_rate": 6.118384621607356e-05, |
| "loss": 0.0052757196426391605, |
| "memory(GiB)": 5.73, |
| "step": 53500, |
| "train_speed(iter/s)": 1.043484 |
| }, |
| { |
| "epoch": 0.4610577005174092, |
| "grad_norm": 0.021920237690210342, |
| "learning_rate": 6.0494758532998397e-05, |
| "loss": 0.0052754092216491695, |
| "memory(GiB)": 5.73, |
| "step": 54000, |
| "train_speed(iter/s)": 1.040652 |
| }, |
| { |
| "epoch": 0.4653267532999778, |
| "grad_norm": 0.02255011908710003, |
| "learning_rate": 5.980357920554813e-05, |
| "loss": 0.005176177024841308, |
| "memory(GiB)": 5.73, |
| "step": 54500, |
| "train_speed(iter/s)": 1.03761 |
| }, |
| { |
| "epoch": 0.4695958060825464, |
| "grad_norm": 0.023933693766593933, |
| "learning_rate": 5.91104459883312e-05, |
| "loss": 0.00518220043182373, |
| "memory(GiB)": 5.73, |
| "step": 55000, |
| "train_speed(iter/s)": 1.03464 |
| }, |
| { |
| "epoch": 0.47386485886511504, |
| "grad_norm": 0.02640974149107933, |
| "learning_rate": 5.8415497025373545e-05, |
| "loss": 0.0051289405822753905, |
| "memory(GiB)": 5.73, |
| "step": 55500, |
| "train_speed(iter/s)": 1.032562 |
| }, |
| { |
| "epoch": 0.4781339116476836, |
| "grad_norm": 0.027417296543717384, |
| "learning_rate": 5.771887082258598e-05, |
| "loss": 0.005091516494750976, |
| "memory(GiB)": 5.73, |
| "step": 56000, |
| "train_speed(iter/s)": 1.031309 |
| }, |
| { |
| "epoch": 0.4824029644302522, |
| "grad_norm": 0.02626318484544754, |
| "learning_rate": 5.7020706220159446e-05, |
| "loss": 0.005014698505401611, |
| "memory(GiB)": 5.73, |
| "step": 56500, |
| "train_speed(iter/s)": 1.030475 |
| }, |
| { |
| "epoch": 0.48667201721282083, |
| "grad_norm": 0.022486470639705658, |
| "learning_rate": 5.6321142364893655e-05, |
| "loss": 0.00502289867401123, |
| "memory(GiB)": 5.73, |
| "step": 57000, |
| "train_speed(iter/s)": 1.029867 |
| }, |
| { |
| "epoch": 0.49094106999538945, |
| "grad_norm": 0.024762239307165146, |
| "learning_rate": 5.562031868246459e-05, |
| "loss": 0.004976710319519043, |
| "memory(GiB)": 5.73, |
| "step": 57500, |
| "train_speed(iter/s)": 1.029133 |
| }, |
| { |
| "epoch": 0.495210122777958, |
| "grad_norm": 0.02197747305035591, |
| "learning_rate": 5.49183748496365e-05, |
| "loss": 0.004930309295654297, |
| "memory(GiB)": 5.73, |
| "step": 58000, |
| "train_speed(iter/s)": 1.03126 |
| }, |
| { |
| "epoch": 0.4994791755605266, |
| "grad_norm": 0.017993444576859474, |
| "learning_rate": 5.421545076642376e-05, |
| "loss": 0.004885564804077149, |
| "memory(GiB)": 5.73, |
| "step": 58500, |
| "train_speed(iter/s)": 1.033407 |
| }, |
| { |
| "epoch": 0.5037482283430952, |
| "grad_norm": 0.023290056735277176, |
| "learning_rate": 5.351168652820825e-05, |
| "loss": 0.004815481662750244, |
| "memory(GiB)": 5.73, |
| "step": 59000, |
| "train_speed(iter/s)": 1.035534 |
| }, |
| { |
| "epoch": 0.5080172811256638, |
| "grad_norm": 0.02278745174407959, |
| "learning_rate": 5.2807222397817946e-05, |
| "loss": 0.0048018951416015625, |
| "memory(GiB)": 5.73, |
| "step": 59500, |
| "train_speed(iter/s)": 1.037635 |
| }, |
| { |
| "epoch": 0.5122863339082324, |
| "grad_norm": 0.01709616929292679, |
| "learning_rate": 5.210219877757185e-05, |
| "loss": 0.004790943622589111, |
| "memory(GiB)": 5.73, |
| "step": 60000, |
| "train_speed(iter/s)": 1.039708 |
| }, |
| { |
| "epoch": 0.516555386690801, |
| "grad_norm": 0.024141253903508186, |
| "learning_rate": 5.139675618129741e-05, |
| "loss": 0.0047971105575561526, |
| "memory(GiB)": 5.73, |
| "step": 60500, |
| "train_speed(iter/s)": 1.041714 |
| }, |
| { |
| "epoch": 0.5208244394733696, |
| "grad_norm": 0.025403697043657303, |
| "learning_rate": 5.069103520632558e-05, |
| "loss": 0.0046922645568847655, |
| "memory(GiB)": 5.73, |
| "step": 61000, |
| "train_speed(iter/s)": 1.043735 |
| }, |
| { |
| "epoch": 0.5250934922559383, |
| "grad_norm": 0.023161958903074265, |
| "learning_rate": 4.998517650546916e-05, |
| "loss": 0.0046929998397827145, |
| "memory(GiB)": 5.73, |
| "step": 61500, |
| "train_speed(iter/s)": 1.045718 |
| }, |
| { |
| "epoch": 0.5293625450385069, |
| "grad_norm": 0.022052627056837082, |
| "learning_rate": 4.927932075899032e-05, |
| "loss": 0.004638696193695068, |
| "memory(GiB)": 5.73, |
| "step": 62000, |
| "train_speed(iter/s)": 1.043334 |
| }, |
| { |
| "epoch": 0.5336315978210755, |
| "grad_norm": 0.023017114028334618, |
| "learning_rate": 4.857360864656229e-05, |
| "loss": 0.004680471420288086, |
| "memory(GiB)": 5.73, |
| "step": 62500, |
| "train_speed(iter/s)": 1.041039 |
| }, |
| { |
| "epoch": 0.5379006506036441, |
| "grad_norm": 0.023162037134170532, |
| "learning_rate": 4.7868180819231614e-05, |
| "loss": 0.004635006904602051, |
| "memory(GiB)": 5.73, |
| "step": 63000, |
| "train_speed(iter/s)": 1.039508 |
| }, |
| { |
| "epoch": 0.5421697033862126, |
| "grad_norm": 0.02154356613755226, |
| "learning_rate": 4.7163177871385713e-05, |
| "loss": 0.004594725131988525, |
| "memory(GiB)": 5.73, |
| "step": 63500, |
| "train_speed(iter/s)": 1.037869 |
| }, |
| { |
| "epoch": 0.5464387561687812, |
| "grad_norm": 0.024489399045705795, |
| "learning_rate": 4.6458740312731915e-05, |
| "loss": 0.004505970001220703, |
| "memory(GiB)": 5.73, |
| "step": 64000, |
| "train_speed(iter/s)": 1.036001 |
| }, |
| { |
| "epoch": 0.5507078089513499, |
| "grad_norm": 0.02230563387274742, |
| "learning_rate": 4.575500854029343e-05, |
| "loss": 0.004512208938598633, |
| "memory(GiB)": 5.73, |
| "step": 64500, |
| "train_speed(iter/s)": 1.034042 |
| }, |
| { |
| "epoch": 0.5549768617339185, |
| "grad_norm": 0.024222563952207565, |
| "learning_rate": 4.5052122810427655e-05, |
| "loss": 0.004453976154327393, |
| "memory(GiB)": 5.73, |
| "step": 65000, |
| "train_speed(iter/s)": 1.033383 |
| }, |
| { |
| "epoch": 0.5592459145164871, |
| "grad_norm": 0.024108612909913063, |
| "learning_rate": 4.435022321087251e-05, |
| "loss": 0.004433969497680664, |
| "memory(GiB)": 5.73, |
| "step": 65500, |
| "train_speed(iter/s)": 1.031504 |
| }, |
| { |
| "epoch": 0.5635149672990557, |
| "grad_norm": 0.02404128573834896, |
| "learning_rate": 4.3649449632826524e-05, |
| "loss": 0.004369840621948242, |
| "memory(GiB)": 5.73, |
| "step": 66000, |
| "train_speed(iter/s)": 1.029515 |
| }, |
| { |
| "epoch": 0.5677840200816243, |
| "grad_norm": 0.02122694067656994, |
| "learning_rate": 4.294994174306796e-05, |
| "loss": 0.00436569881439209, |
| "memory(GiB)": 5.73, |
| "step": 66500, |
| "train_speed(iter/s)": 1.027863 |
| }, |
| { |
| "epoch": 0.5720530728641929, |
| "grad_norm": 0.02231895923614502, |
| "learning_rate": 4.2251838956118646e-05, |
| "loss": 0.004324491500854492, |
| "memory(GiB)": 5.73, |
| "step": 67000, |
| "train_speed(iter/s)": 1.026447 |
| }, |
| { |
| "epoch": 0.5763221256467616, |
| "grad_norm": 0.01995609700679779, |
| "learning_rate": 4.1555280406458243e-05, |
| "loss": 0.004273086071014404, |
| "memory(GiB)": 5.73, |
| "step": 67500, |
| "train_speed(iter/s)": 1.025095 |
| }, |
| { |
| "epoch": 0.5805911784293301, |
| "grad_norm": 0.023028602823615074, |
| "learning_rate": 4.086040492079418e-05, |
| "loss": 0.004247576713562012, |
| "memory(GiB)": 5.73, |
| "step": 68000, |
| "train_speed(iter/s)": 1.024105 |
| }, |
| { |
| "epoch": 0.5848602312118987, |
| "grad_norm": 0.02473682351410389, |
| "learning_rate": 4.016735099039299e-05, |
| "loss": 0.004212839603424072, |
| "memory(GiB)": 5.73, |
| "step": 68500, |
| "train_speed(iter/s)": 1.025833 |
| }, |
| { |
| "epoch": 0.5891292839944673, |
| "grad_norm": 0.02591153420507908, |
| "learning_rate": 3.947625674347842e-05, |
| "loss": 0.004188227653503418, |
| "memory(GiB)": 5.73, |
| "step": 69000, |
| "train_speed(iter/s)": 1.027676 |
| }, |
| { |
| "epoch": 0.5933983367770359, |
| "grad_norm": 0.022372225299477577, |
| "learning_rate": 3.878725991770206e-05, |
| "loss": 0.00420154619216919, |
| "memory(GiB)": 5.73, |
| "step": 69500, |
| "train_speed(iter/s)": 1.029494 |
| }, |
| { |
| "epoch": 0.5976673895596045, |
| "grad_norm": 0.020848704501986504, |
| "learning_rate": 3.810049783269169e-05, |
| "loss": 0.004149648189544677, |
| "memory(GiB)": 5.73, |
| "step": 70000, |
| "train_speed(iter/s)": 1.031265 |
| }, |
| { |
| "epoch": 0.6019364423421731, |
| "grad_norm": 0.020646043121814728, |
| "learning_rate": 3.7416107362682874e-05, |
| "loss": 0.004120903968811035, |
| "memory(GiB)": 5.73, |
| "step": 70500, |
| "train_speed(iter/s)": 1.03288 |
| }, |
| { |
| "epoch": 0.6062054951247418, |
| "grad_norm": 0.02318960428237915, |
| "learning_rate": 3.673422490923957e-05, |
| "loss": 0.004070096492767334, |
| "memory(GiB)": 5.73, |
| "step": 71000, |
| "train_speed(iter/s)": 1.031156 |
| }, |
| { |
| "epoch": 0.6104745479073104, |
| "grad_norm": 0.01929691806435585, |
| "learning_rate": 3.605498637406871e-05, |
| "loss": 0.0040385212898254395, |
| "memory(GiB)": 5.73, |
| "step": 71500, |
| "train_speed(iter/s)": 1.029013 |
| }, |
| { |
| "epoch": 0.6147436006898789, |
| "grad_norm": 0.0221713837236166, |
| "learning_rate": 3.5378527131934415e-05, |
| "loss": 0.004040939807891846, |
| "memory(GiB)": 5.73, |
| "step": 72000, |
| "train_speed(iter/s)": 1.027147 |
| }, |
| { |
| "epoch": 0.6190126534724475, |
| "grad_norm": 0.026295281946659088, |
| "learning_rate": 3.470498200367745e-05, |
| "loss": 0.003968184471130371, |
| "memory(GiB)": 5.73, |
| "step": 72500, |
| "train_speed(iter/s)": 1.025599 |
| }, |
| { |
| "epoch": 0.6232817062550161, |
| "grad_norm": 0.022878218442201614, |
| "learning_rate": 3.403448522934484e-05, |
| "loss": 0.00394676160812378, |
| "memory(GiB)": 5.73, |
| "step": 73000, |
| "train_speed(iter/s)": 1.024666 |
| }, |
| { |
| "epoch": 0.6275507590375847, |
| "grad_norm": 0.017653649672865868, |
| "learning_rate": 3.3367170441435326e-05, |
| "loss": 0.003906076669692993, |
| "memory(GiB)": 5.73, |
| "step": 73500, |
| "train_speed(iter/s)": 1.023811 |
| }, |
| { |
| "epoch": 0.6318198118201533, |
| "grad_norm": 0.021848097443580627, |
| "learning_rate": 3.270317063826594e-05, |
| "loss": 0.0038814377784729005, |
| "memory(GiB)": 5.73, |
| "step": 74000, |
| "train_speed(iter/s)": 1.023855 |
| }, |
| { |
| "epoch": 0.636088864602722, |
| "grad_norm": 0.022029753774404526, |
| "learning_rate": 3.204261815746496e-05, |
| "loss": 0.003879170894622803, |
| "memory(GiB)": 5.73, |
| "step": 74500, |
| "train_speed(iter/s)": 1.024885 |
| }, |
| { |
| "epoch": 0.6403579173852906, |
| "grad_norm": 0.02403407171368599, |
| "learning_rate": 3.1385644649596445e-05, |
| "loss": 0.003841569900512695, |
| "memory(GiB)": 5.73, |
| "step": 75000, |
| "train_speed(iter/s)": 1.026583 |
| }, |
| { |
| "epoch": 0.6446269701678592, |
| "grad_norm": 0.023609979078173637, |
| "learning_rate": 3.073238105192191e-05, |
| "loss": 0.0038005766868591307, |
| "memory(GiB)": 5.73, |
| "step": 75500, |
| "train_speed(iter/s)": 1.028269 |
| }, |
| { |
| "epoch": 0.6488960229504278, |
| "grad_norm": 0.01760837249457836, |
| "learning_rate": 3.008295756230397e-05, |
| "loss": 0.0037522752285003664, |
| "memory(GiB)": 5.73, |
| "step": 76000, |
| "train_speed(iter/s)": 1.029938 |
| }, |
| { |
| "epoch": 0.6531650757329963, |
| "grad_norm": 0.021126747131347656, |
| "learning_rate": 2.943750361325739e-05, |
| "loss": 0.003741382837295532, |
| "memory(GiB)": 5.73, |
| "step": 76500, |
| "train_speed(iter/s)": 1.03159 |
| }, |
| { |
| "epoch": 0.6574341285155649, |
| "grad_norm": 0.02278253622353077, |
| "learning_rate": 2.879614784615281e-05, |
| "loss": 0.0037315216064453126, |
| "memory(GiB)": 5.73, |
| "step": 77000, |
| "train_speed(iter/s)": 1.033224 |
| }, |
| { |
| "epoch": 0.6617031812981335, |
| "grad_norm": 0.023599898442626, |
| "learning_rate": 2.8159018085577936e-05, |
| "loss": 0.0037167372703552247, |
| "memory(GiB)": 5.73, |
| "step": 77500, |
| "train_speed(iter/s)": 1.034809 |
| }, |
| { |
| "epoch": 0.6659722340807022, |
| "grad_norm": 0.02341049537062645, |
| "learning_rate": 2.752624131386169e-05, |
| "loss": 0.0036745924949645997, |
| "memory(GiB)": 5.73, |
| "step": 78000, |
| "train_speed(iter/s)": 1.036412 |
| }, |
| { |
| "epoch": 0.6702412868632708, |
| "grad_norm": 0.021224385127425194, |
| "learning_rate": 2.68979436457661e-05, |
| "loss": 0.0036270735263824465, |
| "memory(GiB)": 5.73, |
| "step": 78500, |
| "train_speed(iter/s)": 1.035171 |
| }, |
| { |
| "epoch": 0.6745103396458394, |
| "grad_norm": 0.02113701030611992, |
| "learning_rate": 2.6274250303351277e-05, |
| "loss": 0.003653192758560181, |
| "memory(GiB)": 5.73, |
| "step": 79000, |
| "train_speed(iter/s)": 1.033322 |
| }, |
| { |
| "epoch": 0.678779392428408, |
| "grad_norm": 0.02273395285010338, |
| "learning_rate": 2.5655285591018053e-05, |
| "loss": 0.003600950241088867, |
| "memory(GiB)": 5.73, |
| "step": 79500, |
| "train_speed(iter/s)": 1.03175 |
| }, |
| { |
| "epoch": 0.6830484452109766, |
| "grad_norm": 0.01843477226793766, |
| "learning_rate": 2.5041172870733688e-05, |
| "loss": 0.003576310634613037, |
| "memory(GiB)": 5.73, |
| "step": 80000, |
| "train_speed(iter/s)": 1.030732 |
| }, |
| { |
| "epoch": 0.6873174979935452, |
| "grad_norm": 0.021858269348740578, |
| "learning_rate": 2.4432034537445504e-05, |
| "loss": 0.0035532989501953125, |
| "memory(GiB)": 5.73, |
| "step": 80500, |
| "train_speed(iter/s)": 1.030234 |
| }, |
| { |
| "epoch": 0.6915865507761138, |
| "grad_norm": 0.02257091924548149, |
| "learning_rate": 2.3827991994686855e-05, |
| "loss": 0.0034713072776794435, |
| "memory(GiB)": 5.73, |
| "step": 81000, |
| "train_speed(iter/s)": 1.029758 |
| }, |
| { |
| "epoch": 0.6958556035586824, |
| "grad_norm": 0.02086802013218403, |
| "learning_rate": 2.3229165630381254e-05, |
| "loss": 0.0035013933181762694, |
| "memory(GiB)": 5.73, |
| "step": 81500, |
| "train_speed(iter/s)": 1.029855 |
| }, |
| { |
| "epoch": 0.700124656341251, |
| "grad_norm": 0.02159390039741993, |
| "learning_rate": 2.263567479284836e-05, |
| "loss": 0.0034512946605682374, |
| "memory(GiB)": 5.73, |
| "step": 82000, |
| "train_speed(iter/s)": 1.031396 |
| }, |
| { |
| "epoch": 0.7043937091238196, |
| "grad_norm": 0.02238837257027626, |
| "learning_rate": 2.2047637767017594e-05, |
| "loss": 0.0034342200756073, |
| "memory(GiB)": 5.73, |
| "step": 82500, |
| "train_speed(iter/s)": 1.029892 |
| }, |
| { |
| "epoch": 0.7086627619063882, |
| "grad_norm": 0.022957606241106987, |
| "learning_rate": 2.1465171750853386e-05, |
| "loss": 0.003412749528884888, |
| "memory(GiB)": 5.73, |
| "step": 83000, |
| "train_speed(iter/s)": 1.02831 |
| }, |
| { |
| "epoch": 0.7129318146889568, |
| "grad_norm": 0.019689923152327538, |
| "learning_rate": 2.0888392831997238e-05, |
| "loss": 0.00341141414642334, |
| "memory(GiB)": 5.73, |
| "step": 83500, |
| "train_speed(iter/s)": 1.026956 |
| }, |
| { |
| "epoch": 0.7172008674715254, |
| "grad_norm": 0.023503178730607033, |
| "learning_rate": 2.03174159646311e-05, |
| "loss": 0.0033840060234069822, |
| "memory(GiB)": 5.73, |
| "step": 84000, |
| "train_speed(iter/s)": 1.025661 |
| }, |
| { |
| "epoch": 0.7214699202540941, |
| "grad_norm": 0.02037668041884899, |
| "learning_rate": 1.9752354946566354e-05, |
| "loss": 0.0033505113124847412, |
| "memory(GiB)": 5.73, |
| "step": 84500, |
| "train_speed(iter/s)": 1.02442 |
| }, |
| { |
| "epoch": 0.7257389730366626, |
| "grad_norm": 0.025251047685742378, |
| "learning_rate": 1.9193322396563785e-05, |
| "loss": 0.0033303892612457277, |
| "memory(GiB)": 5.73, |
| "step": 85000, |
| "train_speed(iter/s)": 1.023539 |
| }, |
| { |
| "epoch": 0.7300080258192312, |
| "grad_norm": 0.01954658329486847, |
| "learning_rate": 1.8640429731887998e-05, |
| "loss": 0.003283708333969116, |
| "memory(GiB)": 5.73, |
| "step": 85500, |
| "train_speed(iter/s)": 1.023297 |
| }, |
| { |
| "epoch": 0.7342770786017998, |
| "grad_norm": 0.0188963171094656, |
| "learning_rate": 1.809378714610167e-05, |
| "loss": 0.003271867275238037, |
| "memory(GiB)": 5.73, |
| "step": 86000, |
| "train_speed(iter/s)": 1.023267 |
| }, |
| { |
| "epoch": 0.7385461313843684, |
| "grad_norm": 0.02359418198466301, |
| "learning_rate": 1.7553503587103505e-05, |
| "loss": 0.0032482266426086424, |
| "memory(GiB)": 5.73, |
| "step": 86500, |
| "train_speed(iter/s)": 1.023221 |
| }, |
| { |
| "epoch": 0.742815184166937, |
| "grad_norm": 0.020105060189962387, |
| "learning_rate": 1.701968673541458e-05, |
| "loss": 0.003266146183013916, |
| "memory(GiB)": 5.73, |
| "step": 87000, |
| "train_speed(iter/s)": 1.023208 |
| }, |
| { |
| "epoch": 0.7470842369495057, |
| "grad_norm": 0.018053073436021805, |
| "learning_rate": 1.649244298271714e-05, |
| "loss": 0.003204747676849365, |
| "memory(GiB)": 5.73, |
| "step": 87500, |
| "train_speed(iter/s)": 1.023206 |
| }, |
| { |
| "epoch": 0.7513532897320743, |
| "grad_norm": 0.021992964670062065, |
| "learning_rate": 1.5971877410650354e-05, |
| "loss": 0.0031999170780181883, |
| "memory(GiB)": 5.73, |
| "step": 88000, |
| "train_speed(iter/s)": 1.023363 |
| }, |
| { |
| "epoch": 0.7556223425146429, |
| "grad_norm": 0.022893013432621956, |
| "learning_rate": 1.545809376986727e-05, |
| "loss": 0.0031597645282745363, |
| "memory(GiB)": 5.73, |
| "step": 88500, |
| "train_speed(iter/s)": 1.023915 |
| }, |
| { |
| "epoch": 0.7598913952972115, |
| "grad_norm": 0.021105078980326653, |
| "learning_rate": 1.4951194459356693e-05, |
| "loss": 0.003171279191970825, |
| "memory(GiB)": 5.73, |
| "step": 89000, |
| "train_speed(iter/s)": 1.024754 |
| }, |
| { |
| "epoch": 0.76416044807978, |
| "grad_norm": 0.01919909007847309, |
| "learning_rate": 1.445128050603493e-05, |
| "loss": 0.0031237168312072752, |
| "memory(GiB)": 5.73, |
| "step": 89500, |
| "train_speed(iter/s)": 1.026165 |
| }, |
| { |
| "epoch": 0.7684295008623486, |
| "grad_norm": 0.02016974799335003, |
| "learning_rate": 1.39584515446106e-05, |
| "loss": 0.003100724697113037, |
| "memory(GiB)": 5.73, |
| "step": 90000, |
| "train_speed(iter/s)": 1.027578 |
| } |
| ], |
| "logging_steps": 500, |
| "max_steps": 117122, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 10000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3.069758781932123e+19, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|