{ "best_metric": 0.90963293, "best_model_checkpoint": "/volume/output/run2/checkpoint-252560", "epoch": 20.0, "eval_steps": 500, "global_step": 252560, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.05004751346214761, "grad_norm": 1.831141710281372, "learning_rate": 9.974976243268926e-05, "loss": 1.5344, "step": 632 }, { "epoch": 0.10009502692429521, "grad_norm": 3.449512004852295, "learning_rate": 9.949952486537853e-05, "loss": 0.3225, "step": 1264 }, { "epoch": 0.1501425403864428, "grad_norm": 1.8513766527175903, "learning_rate": 9.92492872980678e-05, "loss": 0.2466, "step": 1896 }, { "epoch": 0.20019005384859043, "grad_norm": 1.0473196506500244, "learning_rate": 9.899904973075705e-05, "loss": 0.2176, "step": 2528 }, { "epoch": 0.25023756731073804, "grad_norm": 1.3224372863769531, "learning_rate": 9.874881216344632e-05, "loss": 0.1852, "step": 3160 }, { "epoch": 0.3002850807728856, "grad_norm": 1.1163212060928345, "learning_rate": 9.849857459613557e-05, "loss": 0.1725, "step": 3792 }, { "epoch": 0.35033259423503327, "grad_norm": 0.2766059339046478, "learning_rate": 9.824833702882485e-05, "loss": 0.1613, "step": 4424 }, { "epoch": 0.40038010769718085, "grad_norm": 1.0185670852661133, "learning_rate": 9.79980994615141e-05, "loss": 0.1345, "step": 5056 }, { "epoch": 0.4504276211593285, "grad_norm": 1.164562702178955, "learning_rate": 9.774786189420336e-05, "loss": 0.1364, "step": 5688 }, { "epoch": 0.5004751346214761, "grad_norm": 1.1677042245864868, "learning_rate": 9.749762432689263e-05, "loss": 0.1277, "step": 6320 }, { "epoch": 0.5505226480836237, "grad_norm": 3.836665153503418, "learning_rate": 9.724738675958188e-05, "loss": 0.1254, "step": 6952 }, { "epoch": 0.6005701615457713, "grad_norm": 2.3849477767944336, "learning_rate": 9.699714919227115e-05, "loss": 0.1188, "step": 7584 }, { "epoch": 0.6506176750079189, "grad_norm": 3.4638051986694336, "learning_rate": 9.674691162496041e-05, "loss": 0.1159, "step": 8216 }, { "epoch": 0.7006651884700665, "grad_norm": 0.2078377604484558, "learning_rate": 9.649667405764967e-05, "loss": 0.1097, "step": 8848 }, { "epoch": 0.7507127019322142, "grad_norm": 2.513908624649048, "learning_rate": 9.624643649033894e-05, "loss": 0.1073, "step": 9480 }, { "epoch": 0.8007602153943617, "grad_norm": 1.9936473369598389, "learning_rate": 9.599619892302819e-05, "loss": 0.1006, "step": 10112 }, { "epoch": 0.8508077288565093, "grad_norm": 1.9626480340957642, "learning_rate": 9.574596135571746e-05, "loss": 0.1019, "step": 10744 }, { "epoch": 0.900855242318657, "grad_norm": 1.7610087394714355, "learning_rate": 9.549572378840672e-05, "loss": 0.0898, "step": 11376 }, { "epoch": 0.9509027557808045, "grad_norm": 1.6743286848068237, "learning_rate": 9.524548622109598e-05, "loss": 0.0912, "step": 12008 }, { "epoch": 1.0, "eval_gen_len": 12.34557138, "eval_loss": 0.06976257264614105, "eval_runtime": 538.4523, "eval_samples_per_second": 98.843, "eval_steps_per_second": 1.545, "eval_wordacc": 0.8983691, "eval_wordacc_oov": 0.84212982, "step": 12628 }, { "epoch": 1.0009502692429522, "grad_norm": 0.813258707523346, "learning_rate": 9.499524865378525e-05, "loss": 0.0908, "step": 12640 }, { "epoch": 1.0509977827050998, "grad_norm": 1.4410091638565063, "learning_rate": 9.474501108647451e-05, "loss": 0.0864, "step": 13272 }, { "epoch": 1.1010452961672474, "grad_norm": 0.1239413172006607, "learning_rate": 9.449477351916377e-05, "loss": 0.0819, "step": 13904 }, { "epoch": 1.151092809629395, "grad_norm": 0.41668111085891724, "learning_rate": 9.424453595185304e-05, "loss": 0.0872, "step": 14536 }, { "epoch": 1.2011403230915425, "grad_norm": 0.2707739472389221, "learning_rate": 9.399429838454229e-05, "loss": 0.082, "step": 15168 }, { "epoch": 1.2511878365536901, "grad_norm": 0.5811319351196289, "learning_rate": 9.374406081723154e-05, "loss": 0.0847, "step": 15800 }, { "epoch": 1.3012353500158378, "grad_norm": 1.7169886827468872, "learning_rate": 9.349382324992082e-05, "loss": 0.0817, "step": 16432 }, { "epoch": 1.3512828634779854, "grad_norm": 0.11871356517076492, "learning_rate": 9.324358568261008e-05, "loss": 0.0884, "step": 17064 }, { "epoch": 1.401330376940133, "grad_norm": 1.614180326461792, "learning_rate": 9.299334811529935e-05, "loss": 0.0781, "step": 17696 }, { "epoch": 1.4513778904022807, "grad_norm": 0.5827309489250183, "learning_rate": 9.27431105479886e-05, "loss": 0.0722, "step": 18328 }, { "epoch": 1.5014254038644284, "grad_norm": 0.09628592431545258, "learning_rate": 9.249287298067785e-05, "loss": 0.0815, "step": 18960 }, { "epoch": 1.551472917326576, "grad_norm": 1.4028997421264648, "learning_rate": 9.224263541336713e-05, "loss": 0.0758, "step": 19592 }, { "epoch": 1.6015204307887236, "grad_norm": 0.37902089953422546, "learning_rate": 9.199239784605639e-05, "loss": 0.0772, "step": 20224 }, { "epoch": 1.651567944250871, "grad_norm": 0.5193475484848022, "learning_rate": 9.174216027874564e-05, "loss": 0.0731, "step": 20856 }, { "epoch": 1.7016154577130187, "grad_norm": 1.5854244232177734, "learning_rate": 9.149192271143491e-05, "loss": 0.0704, "step": 21488 }, { "epoch": 1.7516629711751663, "grad_norm": 2.4236505031585693, "learning_rate": 9.124168514412418e-05, "loss": 0.0733, "step": 22120 }, { "epoch": 1.8017104846373138, "grad_norm": 0.610543429851532, "learning_rate": 9.099144757681343e-05, "loss": 0.0712, "step": 22752 }, { "epoch": 1.8517579980994614, "grad_norm": 0.8024447560310364, "learning_rate": 9.07412100095027e-05, "loss": 0.0758, "step": 23384 }, { "epoch": 1.901805511561609, "grad_norm": 0.1649412214756012, "learning_rate": 9.049097244219195e-05, "loss": 0.0736, "step": 24016 }, { "epoch": 1.9518530250237567, "grad_norm": 0.15900301933288574, "learning_rate": 9.024073487488122e-05, "loss": 0.0746, "step": 24648 }, { "epoch": 2.0, "eval_gen_len": 12.34418098, "eval_loss": 0.057007092982530594, "eval_runtime": 547.57, "eval_samples_per_second": 97.197, "eval_steps_per_second": 1.519, "eval_wordacc": 0.91236707, "eval_wordacc_oov": 0.85835753, "step": 25256 }, { "epoch": 2.0019005384859043, "grad_norm": 1.2868945598602295, "learning_rate": 8.999049730757049e-05, "loss": 0.0676, "step": 25280 }, { "epoch": 2.051948051948052, "grad_norm": 0.5984334945678711, "learning_rate": 8.974025974025974e-05, "loss": 0.0627, "step": 25912 }, { "epoch": 2.1019955654101996, "grad_norm": 0.043390534818172455, "learning_rate": 8.949002217294901e-05, "loss": 0.0583, "step": 26544 }, { "epoch": 2.1520430788723472, "grad_norm": 0.544601321220398, "learning_rate": 8.923978460563826e-05, "loss": 0.0614, "step": 27176 }, { "epoch": 2.202090592334495, "grad_norm": 0.22555088996887207, "learning_rate": 8.898954703832753e-05, "loss": 0.0713, "step": 27808 }, { "epoch": 2.2521381057966425, "grad_norm": 2.1508424282073975, "learning_rate": 8.87393094710168e-05, "loss": 0.071, "step": 28440 }, { "epoch": 2.30218561925879, "grad_norm": 0.2992984354496002, "learning_rate": 8.848907190370605e-05, "loss": 0.0655, "step": 29072 }, { "epoch": 2.3522331327209374, "grad_norm": 1.4744491577148438, "learning_rate": 8.823883433639532e-05, "loss": 0.0697, "step": 29704 }, { "epoch": 2.402280646183085, "grad_norm": 0.6834865808486938, "learning_rate": 8.798859676908457e-05, "loss": 0.0661, "step": 30336 }, { "epoch": 2.4523281596452327, "grad_norm": 0.06805714964866638, "learning_rate": 8.773835920177384e-05, "loss": 0.0672, "step": 30968 }, { "epoch": 2.5023756731073803, "grad_norm": 1.880346655845642, "learning_rate": 8.748812163446311e-05, "loss": 0.063, "step": 31600 }, { "epoch": 2.552423186569528, "grad_norm": 0.5247331857681274, "learning_rate": 8.723788406715236e-05, "loss": 0.0621, "step": 32232 }, { "epoch": 2.6024707000316756, "grad_norm": 0.15831807255744934, "learning_rate": 8.698764649984163e-05, "loss": 0.0653, "step": 32864 }, { "epoch": 2.652518213493823, "grad_norm": 0.04121825844049454, "learning_rate": 8.673740893253088e-05, "loss": 0.0642, "step": 33496 }, { "epoch": 2.702565726955971, "grad_norm": 0.3549499809741974, "learning_rate": 8.648717136522015e-05, "loss": 0.0679, "step": 34128 }, { "epoch": 2.7526132404181185, "grad_norm": 0.16000640392303467, "learning_rate": 8.623693379790942e-05, "loss": 0.0597, "step": 34760 }, { "epoch": 2.802660753880266, "grad_norm": 0.6360165476799011, "learning_rate": 8.598669623059867e-05, "loss": 0.0666, "step": 35392 }, { "epoch": 2.852708267342414, "grad_norm": 0.4808698296546936, "learning_rate": 8.573645866328793e-05, "loss": 0.0654, "step": 36024 }, { "epoch": 2.9027557808045614, "grad_norm": 1.7070688009262085, "learning_rate": 8.54862210959772e-05, "loss": 0.0645, "step": 36656 }, { "epoch": 2.952803294266709, "grad_norm": 0.3017909824848175, "learning_rate": 8.523598352866646e-05, "loss": 0.0622, "step": 37288 }, { "epoch": 3.0, "eval_gen_len": 12.35115178, "eval_loss": 0.04925922676920891, "eval_runtime": 519.5392, "eval_samples_per_second": 102.441, "eval_steps_per_second": 1.601, "eval_wordacc": 0.9194506, "eval_wordacc_oov": 0.87168187, "step": 37884 }, { "epoch": 3.0028508077288567, "grad_norm": 0.5821070671081543, "learning_rate": 8.498574596135573e-05, "loss": 0.0661, "step": 37920 }, { "epoch": 3.052898321191004, "grad_norm": 0.4475654363632202, "learning_rate": 8.473550839404498e-05, "loss": 0.0574, "step": 38552 }, { "epoch": 3.1029458346531515, "grad_norm": 1.4051363468170166, "learning_rate": 8.448527082673424e-05, "loss": 0.0614, "step": 39184 }, { "epoch": 3.152993348115299, "grad_norm": 0.5270406603813171, "learning_rate": 8.423503325942352e-05, "loss": 0.059, "step": 39816 }, { "epoch": 3.203040861577447, "grad_norm": 0.2516399621963501, "learning_rate": 8.398479569211277e-05, "loss": 0.0568, "step": 40448 }, { "epoch": 3.2530883750395945, "grad_norm": 2.4771833419799805, "learning_rate": 8.373455812480203e-05, "loss": 0.055, "step": 41080 }, { "epoch": 3.303135888501742, "grad_norm": 0.39709779620170593, "learning_rate": 8.348432055749129e-05, "loss": 0.0591, "step": 41712 }, { "epoch": 3.3531834019638898, "grad_norm": 0.2029147893190384, "learning_rate": 8.323408299018055e-05, "loss": 0.0572, "step": 42344 }, { "epoch": 3.4032309154260374, "grad_norm": 0.1706971675157547, "learning_rate": 8.298384542286983e-05, "loss": 0.0618, "step": 42976 }, { "epoch": 3.453278428888185, "grad_norm": 0.14079004526138306, "learning_rate": 8.273360785555908e-05, "loss": 0.0583, "step": 43608 }, { "epoch": 3.5033259423503327, "grad_norm": 0.8446473479270935, "learning_rate": 8.248337028824834e-05, "loss": 0.0582, "step": 44240 }, { "epoch": 3.5533734558124803, "grad_norm": 1.3202866315841675, "learning_rate": 8.22331327209376e-05, "loss": 0.0594, "step": 44872 }, { "epoch": 3.603420969274628, "grad_norm": 0.1289588212966919, "learning_rate": 8.198289515362686e-05, "loss": 0.0668, "step": 45504 }, { "epoch": 3.653468482736775, "grad_norm": 0.7715129852294922, "learning_rate": 8.173265758631612e-05, "loss": 0.056, "step": 46136 }, { "epoch": 3.703515996198923, "grad_norm": 0.8882943391799927, "learning_rate": 8.148242001900539e-05, "loss": 0.0585, "step": 46768 }, { "epoch": 3.7535635096610704, "grad_norm": 0.5238478183746338, "learning_rate": 8.123218245169465e-05, "loss": 0.0523, "step": 47400 }, { "epoch": 3.803611023123218, "grad_norm": 0.7119426727294922, "learning_rate": 8.098194488438391e-05, "loss": 0.0566, "step": 48032 }, { "epoch": 3.8536585365853657, "grad_norm": 1.7843942642211914, "learning_rate": 8.073170731707318e-05, "loss": 0.0577, "step": 48664 }, { "epoch": 3.9037060500475134, "grad_norm": 0.09263037890195847, "learning_rate": 8.048146974976244e-05, "loss": 0.0602, "step": 49296 }, { "epoch": 3.953753563509661, "grad_norm": 0.8220856785774231, "learning_rate": 8.02312321824517e-05, "loss": 0.0584, "step": 49928 }, { "epoch": 4.0, "eval_gen_len": 12.34404945, "eval_loss": 0.04651999473571777, "eval_runtime": 525.6595, "eval_samples_per_second": 101.248, "eval_steps_per_second": 1.583, "eval_wordacc": 0.92211867, "eval_wordacc_oov": 0.87494815, "step": 50512 }, { "epoch": 4.003801076971809, "grad_norm": 0.09420084208250046, "learning_rate": 7.998099461514096e-05, "loss": 0.0559, "step": 50560 }, { "epoch": 4.053848590433956, "grad_norm": 0.6216241121292114, "learning_rate": 7.973075704783022e-05, "loss": 0.0507, "step": 51192 }, { "epoch": 4.103896103896104, "grad_norm": 0.6072413921356201, "learning_rate": 7.948051948051949e-05, "loss": 0.0508, "step": 51824 }, { "epoch": 4.153943617358252, "grad_norm": 0.3261624872684479, "learning_rate": 7.923028191320875e-05, "loss": 0.0584, "step": 52456 }, { "epoch": 4.203991130820399, "grad_norm": 3.996793508529663, "learning_rate": 7.898004434589801e-05, "loss": 0.0542, "step": 53088 }, { "epoch": 4.254038644282547, "grad_norm": 0.09758679568767548, "learning_rate": 7.872980677858727e-05, "loss": 0.0543, "step": 53720 }, { "epoch": 4.3040861577446945, "grad_norm": 0.8341479897499084, "learning_rate": 7.847956921127652e-05, "loss": 0.06, "step": 54352 }, { "epoch": 4.354133671206842, "grad_norm": 0.3256041407585144, "learning_rate": 7.82293316439658e-05, "loss": 0.0547, "step": 54984 }, { "epoch": 4.40418118466899, "grad_norm": 0.8251773715019226, "learning_rate": 7.797909407665506e-05, "loss": 0.0563, "step": 55616 }, { "epoch": 4.454228698131137, "grad_norm": 0.23588858544826508, "learning_rate": 7.772885650934432e-05, "loss": 0.0566, "step": 56248 }, { "epoch": 4.504276211593285, "grad_norm": 0.8622148633003235, "learning_rate": 7.747861894203358e-05, "loss": 0.0511, "step": 56880 }, { "epoch": 4.554323725055433, "grad_norm": 0.29321447014808655, "learning_rate": 7.722838137472284e-05, "loss": 0.0502, "step": 57512 }, { "epoch": 4.60437123851758, "grad_norm": 1.1889938116073608, "learning_rate": 7.697814380741211e-05, "loss": 0.0525, "step": 58144 }, { "epoch": 4.654418751979728, "grad_norm": 0.5421351790428162, "learning_rate": 7.672790624010137e-05, "loss": 0.0509, "step": 58776 }, { "epoch": 4.704466265441875, "grad_norm": 0.4371638000011444, "learning_rate": 7.647766867279062e-05, "loss": 0.0548, "step": 59408 }, { "epoch": 4.754513778904023, "grad_norm": 0.06439998745918274, "learning_rate": 7.622743110547989e-05, "loss": 0.0553, "step": 60040 }, { "epoch": 4.80456129236617, "grad_norm": 0.8337986469268799, "learning_rate": 7.597719353816916e-05, "loss": 0.0539, "step": 60672 }, { "epoch": 4.854608805828318, "grad_norm": 0.5551128387451172, "learning_rate": 7.572695597085841e-05, "loss": 0.0531, "step": 61304 }, { "epoch": 4.904656319290465, "grad_norm": 0.842311680316925, "learning_rate": 7.547671840354768e-05, "loss": 0.053, "step": 61936 }, { "epoch": 4.954703832752613, "grad_norm": 0.6434153318405151, "learning_rate": 7.522648083623693e-05, "loss": 0.0497, "step": 62568 }, { "epoch": 5.0, "eval_gen_len": 12.35519146, "eval_loss": 0.04361514747142792, "eval_runtime": 525.1326, "eval_samples_per_second": 101.35, "eval_steps_per_second": 1.584, "eval_wordacc": 0.92741723, "eval_wordacc_oov": 0.88205102, "step": 63140 }, { "epoch": 5.004751346214761, "grad_norm": 0.11094748228788376, "learning_rate": 7.49762432689262e-05, "loss": 0.0509, "step": 63200 }, { "epoch": 5.054798859676908, "grad_norm": 0.46878868341445923, "learning_rate": 7.472600570161547e-05, "loss": 0.0445, "step": 63832 }, { "epoch": 5.104846373139056, "grad_norm": 1.7899694442749023, "learning_rate": 7.447576813430472e-05, "loss": 0.045, "step": 64464 }, { "epoch": 5.1548938866012035, "grad_norm": 0.03743477538228035, "learning_rate": 7.422553056699399e-05, "loss": 0.0538, "step": 65096 }, { "epoch": 5.204941400063351, "grad_norm": 0.9636221528053284, "learning_rate": 7.397529299968324e-05, "loss": 0.0444, "step": 65728 }, { "epoch": 5.254988913525499, "grad_norm": 0.1024821326136589, "learning_rate": 7.372505543237251e-05, "loss": 0.0543, "step": 66360 }, { "epoch": 5.305036426987646, "grad_norm": 0.6220707297325134, "learning_rate": 7.347481786506178e-05, "loss": 0.0537, "step": 66992 }, { "epoch": 5.355083940449794, "grad_norm": 0.13747639954090118, "learning_rate": 7.322458029775103e-05, "loss": 0.0443, "step": 67624 }, { "epoch": 5.405131453911942, "grad_norm": 0.25481894612312317, "learning_rate": 7.29743427304403e-05, "loss": 0.0544, "step": 68256 }, { "epoch": 5.455178967374089, "grad_norm": 0.40640395879745483, "learning_rate": 7.272410516312955e-05, "loss": 0.0501, "step": 68888 }, { "epoch": 5.505226480836237, "grad_norm": 0.017994888126850128, "learning_rate": 7.247386759581882e-05, "loss": 0.0497, "step": 69520 }, { "epoch": 5.555273994298385, "grad_norm": 0.7399475574493408, "learning_rate": 7.222363002850809e-05, "loss": 0.0503, "step": 70152 }, { "epoch": 5.605321507760532, "grad_norm": 0.1965421885251999, "learning_rate": 7.197339246119734e-05, "loss": 0.0494, "step": 70784 }, { "epoch": 5.65536902122268, "grad_norm": 0.603735625743866, "learning_rate": 7.172315489388661e-05, "loss": 0.0492, "step": 71416 }, { "epoch": 5.705416534684828, "grad_norm": 0.4098168909549713, "learning_rate": 7.147291732657586e-05, "loss": 0.0496, "step": 72048 }, { "epoch": 5.755464048146975, "grad_norm": 0.026117555797100067, "learning_rate": 7.122267975926513e-05, "loss": 0.05, "step": 72680 }, { "epoch": 5.805511561609123, "grad_norm": 0.03871222585439682, "learning_rate": 7.09724421919544e-05, "loss": 0.0508, "step": 73312 }, { "epoch": 5.8555590750712705, "grad_norm": 1.100329041481018, "learning_rate": 7.072220462464365e-05, "loss": 0.0503, "step": 73944 }, { "epoch": 5.905606588533418, "grad_norm": 0.8121901154518127, "learning_rate": 7.04719670573329e-05, "loss": 0.0468, "step": 74576 }, { "epoch": 5.955654101995566, "grad_norm": 0.6457042694091797, "learning_rate": 7.022172949002219e-05, "loss": 0.0502, "step": 75208 }, { "epoch": 6.0, "eval_gen_len": 12.35192214, "eval_loss": 0.041068777441978455, "eval_runtime": 522.1947, "eval_samples_per_second": 101.92, "eval_steps_per_second": 1.593, "eval_wordacc": 0.93108113, "eval_wordacc_oov": 0.88583575, "step": 75768 }, { "epoch": 6.005701615457713, "grad_norm": 0.20237529277801514, "learning_rate": 6.997149192271144e-05, "loss": 0.0496, "step": 75840 }, { "epoch": 6.055749128919861, "grad_norm": 0.10558341443538666, "learning_rate": 6.97212543554007e-05, "loss": 0.0422, "step": 76472 }, { "epoch": 6.105796642382008, "grad_norm": 0.38815170526504517, "learning_rate": 6.947101678808996e-05, "loss": 0.0468, "step": 77104 }, { "epoch": 6.1558441558441555, "grad_norm": 0.042554233223199844, "learning_rate": 6.922077922077921e-05, "loss": 0.0446, "step": 77736 }, { "epoch": 6.205891669306303, "grad_norm": 1.2894216775894165, "learning_rate": 6.89705416534685e-05, "loss": 0.0462, "step": 78368 }, { "epoch": 6.255939182768451, "grad_norm": 0.7259889841079712, "learning_rate": 6.872030408615775e-05, "loss": 0.045, "step": 79000 }, { "epoch": 6.305986696230598, "grad_norm": 0.5286532044410706, "learning_rate": 6.8470066518847e-05, "loss": 0.0446, "step": 79632 }, { "epoch": 6.356034209692746, "grad_norm": 0.20671215653419495, "learning_rate": 6.821982895153627e-05, "loss": 0.0442, "step": 80264 }, { "epoch": 6.406081723154894, "grad_norm": 0.2522045969963074, "learning_rate": 6.796959138422552e-05, "loss": 0.0455, "step": 80896 }, { "epoch": 6.456129236617041, "grad_norm": 0.09235669672489166, "learning_rate": 6.771935381691479e-05, "loss": 0.0463, "step": 81528 }, { "epoch": 6.506176750079189, "grad_norm": 1.3369249105453491, "learning_rate": 6.746911624960406e-05, "loss": 0.0467, "step": 82160 }, { "epoch": 6.556224263541337, "grad_norm": 0.1361207365989685, "learning_rate": 6.721887868229331e-05, "loss": 0.0464, "step": 82792 }, { "epoch": 6.606271777003484, "grad_norm": 0.057785116136074066, "learning_rate": 6.696864111498258e-05, "loss": 0.0471, "step": 83424 }, { "epoch": 6.656319290465632, "grad_norm": 0.5493625998497009, "learning_rate": 6.671840354767185e-05, "loss": 0.0493, "step": 84056 }, { "epoch": 6.7063668039277795, "grad_norm": 0.08289259672164917, "learning_rate": 6.64681659803611e-05, "loss": 0.0491, "step": 84688 }, { "epoch": 6.756414317389927, "grad_norm": 0.08197712898254395, "learning_rate": 6.621792841305037e-05, "loss": 0.0487, "step": 85320 }, { "epoch": 6.806461830852075, "grad_norm": 0.1574297547340393, "learning_rate": 6.596769084573962e-05, "loss": 0.042, "step": 85952 }, { "epoch": 6.856509344314222, "grad_norm": 0.07021531462669373, "learning_rate": 6.571745327842889e-05, "loss": 0.0469, "step": 86584 }, { "epoch": 6.90655685777637, "grad_norm": 0.5342025756835938, "learning_rate": 6.546721571111816e-05, "loss": 0.0458, "step": 87216 }, { "epoch": 6.956604371238518, "grad_norm": 0.754435658454895, "learning_rate": 6.521697814380741e-05, "loss": 0.0428, "step": 87848 }, { "epoch": 7.0, "eval_gen_len": 12.34435008, "eval_loss": 0.039582036435604095, "eval_runtime": 513.2393, "eval_samples_per_second": 103.698, "eval_steps_per_second": 1.621, "eval_wordacc": 0.93356131, "eval_wordacc_oov": 0.88780589, "step": 88396 }, { "epoch": 7.006651884700665, "grad_norm": 1.7553069591522217, "learning_rate": 6.496674057649668e-05, "loss": 0.0429, "step": 88480 }, { "epoch": 7.056699398162813, "grad_norm": 0.15994039177894592, "learning_rate": 6.471650300918593e-05, "loss": 0.0451, "step": 89112 }, { "epoch": 7.106746911624961, "grad_norm": 0.21786805987358093, "learning_rate": 6.44662654418752e-05, "loss": 0.0415, "step": 89744 }, { "epoch": 7.156794425087108, "grad_norm": 0.07405902445316315, "learning_rate": 6.421602787456447e-05, "loss": 0.0408, "step": 90376 }, { "epoch": 7.206841938549256, "grad_norm": 0.1853848397731781, "learning_rate": 6.396579030725372e-05, "loss": 0.0415, "step": 91008 }, { "epoch": 7.256889452011404, "grad_norm": 0.41366642713546753, "learning_rate": 6.371555273994299e-05, "loss": 0.046, "step": 91640 }, { "epoch": 7.306936965473551, "grad_norm": 0.2615118622779846, "learning_rate": 6.346531517263224e-05, "loss": 0.0413, "step": 92272 }, { "epoch": 7.356984478935699, "grad_norm": 0.06805741786956787, "learning_rate": 6.321507760532151e-05, "loss": 0.0436, "step": 92904 }, { "epoch": 7.407031992397846, "grad_norm": 1.3070762157440186, "learning_rate": 6.296484003801078e-05, "loss": 0.0464, "step": 93536 }, { "epoch": 7.457079505859994, "grad_norm": 0.11481507122516632, "learning_rate": 6.271460247070003e-05, "loss": 0.0442, "step": 94168 }, { "epoch": 7.507127019322141, "grad_norm": 0.9575181603431702, "learning_rate": 6.246436490338929e-05, "loss": 0.043, "step": 94800 }, { "epoch": 7.5571745327842885, "grad_norm": 0.2559140920639038, "learning_rate": 6.221412733607856e-05, "loss": 0.0375, "step": 95432 }, { "epoch": 7.607222046246436, "grad_norm": 0.043967317789793015, "learning_rate": 6.196388976876782e-05, "loss": 0.0412, "step": 96064 }, { "epoch": 7.657269559708584, "grad_norm": 0.5306654572486877, "learning_rate": 6.171365220145709e-05, "loss": 0.0449, "step": 96696 }, { "epoch": 7.7073170731707314, "grad_norm": 0.03713352233171463, "learning_rate": 6.146341463414634e-05, "loss": 0.0443, "step": 97328 }, { "epoch": 7.757364586632879, "grad_norm": 0.6375567317008972, "learning_rate": 6.12131770668356e-05, "loss": 0.0441, "step": 97960 }, { "epoch": 7.807412100095027, "grad_norm": 0.9398515224456787, "learning_rate": 6.096293949952487e-05, "loss": 0.0421, "step": 98592 }, { "epoch": 7.857459613557174, "grad_norm": 0.6615290641784668, "learning_rate": 6.071270193221413e-05, "loss": 0.0468, "step": 99224 }, { "epoch": 7.907507127019322, "grad_norm": 0.48468518257141113, "learning_rate": 6.0462464364903394e-05, "loss": 0.0427, "step": 99856 }, { "epoch": 7.95755464048147, "grad_norm": 0.967910647392273, "learning_rate": 6.0212226797592654e-05, "loss": 0.0416, "step": 100488 }, { "epoch": 8.0, "eval_gen_len": 12.34714967, "eval_loss": 0.0372321754693985, "eval_runtime": 517.1292, "eval_samples_per_second": 102.918, "eval_steps_per_second": 1.609, "eval_wordacc": 0.93393709, "eval_wordacc_oov": 0.88868727, "step": 101024 }, { "epoch": 8.007602153943617, "grad_norm": 0.5545419454574585, "learning_rate": 5.9961989230281915e-05, "loss": 0.0447, "step": 101120 }, { "epoch": 8.057649667405766, "grad_norm": 1.7096141576766968, "learning_rate": 5.971175166297118e-05, "loss": 0.038, "step": 101752 }, { "epoch": 8.107697180867913, "grad_norm": 0.06642602384090424, "learning_rate": 5.946151409566044e-05, "loss": 0.0395, "step": 102384 }, { "epoch": 8.15774469433006, "grad_norm": 0.15396763384342194, "learning_rate": 5.9211276528349704e-05, "loss": 0.0411, "step": 103016 }, { "epoch": 8.207792207792208, "grad_norm": 1.0655204057693481, "learning_rate": 5.8961038961038965e-05, "loss": 0.0416, "step": 103648 }, { "epoch": 8.257839721254355, "grad_norm": 0.42243492603302, "learning_rate": 5.871080139372822e-05, "loss": 0.0445, "step": 104280 }, { "epoch": 8.307887234716503, "grad_norm": 0.05212310701608658, "learning_rate": 5.846056382641749e-05, "loss": 0.0392, "step": 104912 }, { "epoch": 8.35793474817865, "grad_norm": 0.6314841508865356, "learning_rate": 5.8210326259106754e-05, "loss": 0.0392, "step": 105544 }, { "epoch": 8.407982261640798, "grad_norm": 0.4014628827571869, "learning_rate": 5.796008869179601e-05, "loss": 0.0387, "step": 106176 }, { "epoch": 8.458029775102945, "grad_norm": 0.13107645511627197, "learning_rate": 5.770985112448527e-05, "loss": 0.0425, "step": 106808 }, { "epoch": 8.508077288565094, "grad_norm": 0.1390117108821869, "learning_rate": 5.745961355717454e-05, "loss": 0.038, "step": 107440 }, { "epoch": 8.55812480202724, "grad_norm": 0.1780831664800644, "learning_rate": 5.72093759898638e-05, "loss": 0.0397, "step": 108072 }, { "epoch": 8.608172315489389, "grad_norm": 0.8216772675514221, "learning_rate": 5.695913842255306e-05, "loss": 0.0377, "step": 108704 }, { "epoch": 8.658219828951536, "grad_norm": 0.040023334324359894, "learning_rate": 5.670890085524232e-05, "loss": 0.0395, "step": 109336 }, { "epoch": 8.708267342413684, "grad_norm": 0.7334257364273071, "learning_rate": 5.645866328793158e-05, "loss": 0.0401, "step": 109968 }, { "epoch": 8.758314855875831, "grad_norm": 0.09213205426931381, "learning_rate": 5.620842572062085e-05, "loss": 0.038, "step": 110600 }, { "epoch": 8.80836236933798, "grad_norm": 0.8264344930648804, "learning_rate": 5.595818815331011e-05, "loss": 0.0404, "step": 111232 }, { "epoch": 8.858409882800126, "grad_norm": 0.4497428238391876, "learning_rate": 5.570795058599937e-05, "loss": 0.0417, "step": 111864 }, { "epoch": 8.908457396262275, "grad_norm": 0.02390374056994915, "learning_rate": 5.545771301868863e-05, "loss": 0.0351, "step": 112496 }, { "epoch": 8.958504909724422, "grad_norm": 0.1745648831129074, "learning_rate": 5.520747545137789e-05, "loss": 0.042, "step": 113128 }, { "epoch": 9.0, "eval_gen_len": 12.3485025, "eval_loss": 0.03648155927658081, "eval_runtime": 514.9505, "eval_samples_per_second": 103.354, "eval_steps_per_second": 1.616, "eval_wordacc": 0.93964902, "eval_wordacc_oov": 0.89444214, "step": 113652 }, { "epoch": 9.00855242318657, "grad_norm": 0.7463224530220032, "learning_rate": 5.4957237884067156e-05, "loss": 0.0393, "step": 113760 }, { "epoch": 9.058599936648717, "grad_norm": 0.2374447137117386, "learning_rate": 5.470700031675642e-05, "loss": 0.0349, "step": 114392 }, { "epoch": 9.108647450110865, "grad_norm": 1.1859426498413086, "learning_rate": 5.445676274944568e-05, "loss": 0.039, "step": 115024 }, { "epoch": 9.158694963573012, "grad_norm": 0.03181586042046547, "learning_rate": 5.420652518213494e-05, "loss": 0.0354, "step": 115656 }, { "epoch": 9.20874247703516, "grad_norm": 0.0614316463470459, "learning_rate": 5.3956287614824206e-05, "loss": 0.0337, "step": 116288 }, { "epoch": 9.258789990497307, "grad_norm": 0.031755685806274414, "learning_rate": 5.370605004751347e-05, "loss": 0.0346, "step": 116920 }, { "epoch": 9.308837503959456, "grad_norm": 0.4263891875743866, "learning_rate": 5.345581248020273e-05, "loss": 0.0365, "step": 117552 }, { "epoch": 9.358885017421603, "grad_norm": 0.7516904473304749, "learning_rate": 5.320557491289199e-05, "loss": 0.0394, "step": 118184 }, { "epoch": 9.408932530883751, "grad_norm": 0.547282338142395, "learning_rate": 5.295533734558125e-05, "loss": 0.037, "step": 118816 }, { "epoch": 9.458980044345898, "grad_norm": 0.048598043620586395, "learning_rate": 5.2705099778270516e-05, "loss": 0.0396, "step": 119448 }, { "epoch": 9.509027557808047, "grad_norm": 0.019015343859791756, "learning_rate": 5.245486221095978e-05, "loss": 0.0377, "step": 120080 }, { "epoch": 9.559075071270193, "grad_norm": 0.42539820075035095, "learning_rate": 5.220462464364904e-05, "loss": 0.0361, "step": 120712 }, { "epoch": 9.60912258473234, "grad_norm": 0.5751402974128723, "learning_rate": 5.19543870763383e-05, "loss": 0.0407, "step": 121344 }, { "epoch": 9.659170098194489, "grad_norm": 1.380823016166687, "learning_rate": 5.170414950902755e-05, "loss": 0.0388, "step": 121976 }, { "epoch": 9.709217611656635, "grad_norm": 0.6849233508110046, "learning_rate": 5.145391194171683e-05, "loss": 0.0387, "step": 122608 }, { "epoch": 9.759265125118784, "grad_norm": 1.4156357049942017, "learning_rate": 5.120367437440609e-05, "loss": 0.0373, "step": 123240 }, { "epoch": 9.80931263858093, "grad_norm": 0.18062171339988708, "learning_rate": 5.095343680709535e-05, "loss": 0.0333, "step": 123872 }, { "epoch": 9.85936015204308, "grad_norm": 0.6566870212554932, "learning_rate": 5.07031992397846e-05, "loss": 0.0367, "step": 124504 }, { "epoch": 9.909407665505226, "grad_norm": 0.2741030156612396, "learning_rate": 5.0452961672473876e-05, "loss": 0.0426, "step": 125136 }, { "epoch": 9.959455178967374, "grad_norm": 0.7864658236503601, "learning_rate": 5.020272410516314e-05, "loss": 0.0376, "step": 125768 }, { "epoch": 10.0, "eval_gen_len": 12.34852129, "eval_loss": 0.03532838076353073, "eval_runtime": 515.5211, "eval_samples_per_second": 103.239, "eval_steps_per_second": 1.614, "eval_wordacc": 0.94117094, "eval_wordacc_oov": 0.89620489, "step": 126280 }, { "epoch": 10.009502692429521, "grad_norm": 0.08346331119537354, "learning_rate": 4.995248653785239e-05, "loss": 0.0358, "step": 126400 }, { "epoch": 10.05955020589167, "grad_norm": 0.1552925556898117, "learning_rate": 4.970224897054165e-05, "loss": 0.0286, "step": 127032 }, { "epoch": 10.109597719353816, "grad_norm": 0.017832357436418533, "learning_rate": 4.945201140323092e-05, "loss": 0.0342, "step": 127664 }, { "epoch": 10.159645232815965, "grad_norm": 0.7822960019111633, "learning_rate": 4.920177383592018e-05, "loss": 0.0342, "step": 128296 }, { "epoch": 10.209692746278112, "grad_norm": 0.8483818173408508, "learning_rate": 4.895153626860944e-05, "loss": 0.036, "step": 128928 }, { "epoch": 10.25974025974026, "grad_norm": 0.12484422326087952, "learning_rate": 4.87012987012987e-05, "loss": 0.0353, "step": 129560 }, { "epoch": 10.309787773202407, "grad_norm": 1.0866436958312988, "learning_rate": 4.845106113398797e-05, "loss": 0.0336, "step": 130192 }, { "epoch": 10.359835286664556, "grad_norm": 2.065387487411499, "learning_rate": 4.820082356667723e-05, "loss": 0.036, "step": 130824 }, { "epoch": 10.409882800126702, "grad_norm": 0.13544411957263947, "learning_rate": 4.795058599936649e-05, "loss": 0.0353, "step": 131456 }, { "epoch": 10.45993031358885, "grad_norm": 0.8644410967826843, "learning_rate": 4.770034843205575e-05, "loss": 0.0366, "step": 132088 }, { "epoch": 10.509977827050998, "grad_norm": 0.044758204370737076, "learning_rate": 4.745011086474501e-05, "loss": 0.0337, "step": 132720 }, { "epoch": 10.560025340513146, "grad_norm": 0.07954395562410355, "learning_rate": 4.719987329743428e-05, "loss": 0.038, "step": 133352 }, { "epoch": 10.610072853975293, "grad_norm": 0.15263523161411285, "learning_rate": 4.694963573012354e-05, "loss": 0.037, "step": 133984 }, { "epoch": 10.660120367437441, "grad_norm": 0.4090266227722168, "learning_rate": 4.66993981628128e-05, "loss": 0.0341, "step": 134616 }, { "epoch": 10.710167880899588, "grad_norm": 0.3766542077064514, "learning_rate": 4.644916059550206e-05, "loss": 0.0395, "step": 135248 }, { "epoch": 10.760215394361737, "grad_norm": 0.044228482991456985, "learning_rate": 4.619892302819132e-05, "loss": 0.0354, "step": 135880 }, { "epoch": 10.810262907823883, "grad_norm": 3.4171650409698486, "learning_rate": 4.594868546088059e-05, "loss": 0.0344, "step": 136512 }, { "epoch": 10.86031042128603, "grad_norm": 0.1112111434340477, "learning_rate": 4.569844789356984e-05, "loss": 0.0361, "step": 137144 }, { "epoch": 10.910357934748179, "grad_norm": 0.09063247591257095, "learning_rate": 4.544821032625911e-05, "loss": 0.0325, "step": 137776 }, { "epoch": 10.960405448210325, "grad_norm": 0.2144654095172882, "learning_rate": 4.519797275894837e-05, "loss": 0.031, "step": 138408 }, { "epoch": 11.0, "eval_gen_len": 12.35186577, "eval_loss": 0.03388630226254463, "eval_runtime": 519.1861, "eval_samples_per_second": 102.51, "eval_steps_per_second": 1.603, "eval_wordacc": 0.94387659, "eval_wordacc_oov": 0.90081916, "step": 138908 }, { "epoch": 11.010452961672474, "grad_norm": 0.9230628609657288, "learning_rate": 4.494773519163763e-05, "loss": 0.0336, "step": 139040 }, { "epoch": 11.06050047513462, "grad_norm": 0.04931863397359848, "learning_rate": 4.469749762432689e-05, "loss": 0.0332, "step": 139672 }, { "epoch": 11.11054798859677, "grad_norm": 0.43372172117233276, "learning_rate": 4.4447260057016154e-05, "loss": 0.0303, "step": 140304 }, { "epoch": 11.160595502058916, "grad_norm": 1.8216339349746704, "learning_rate": 4.419702248970542e-05, "loss": 0.0327, "step": 140936 }, { "epoch": 11.210643015521065, "grad_norm": 0.4168206453323364, "learning_rate": 4.394678492239468e-05, "loss": 0.0299, "step": 141568 }, { "epoch": 11.260690528983211, "grad_norm": 0.2748374938964844, "learning_rate": 4.369654735508394e-05, "loss": 0.031, "step": 142200 }, { "epoch": 11.31073804244536, "grad_norm": 1.0721484422683716, "learning_rate": 4.34463097877732e-05, "loss": 0.0338, "step": 142832 }, { "epoch": 11.360785555907507, "grad_norm": 0.6880238056182861, "learning_rate": 4.319607222046247e-05, "loss": 0.0306, "step": 143464 }, { "epoch": 11.410833069369655, "grad_norm": 0.6363319158554077, "learning_rate": 4.294583465315173e-05, "loss": 0.0324, "step": 144096 }, { "epoch": 11.460880582831802, "grad_norm": 0.9793805480003357, "learning_rate": 4.2695597085840985e-05, "loss": 0.0352, "step": 144728 }, { "epoch": 11.51092809629395, "grad_norm": 0.05823446065187454, "learning_rate": 4.244535951853025e-05, "loss": 0.0302, "step": 145360 }, { "epoch": 11.560975609756097, "grad_norm": 0.44829338788986206, "learning_rate": 4.2195121951219514e-05, "loss": 0.0317, "step": 145992 }, { "epoch": 11.611023123218246, "grad_norm": 0.21787405014038086, "learning_rate": 4.194488438390878e-05, "loss": 0.0327, "step": 146624 }, { "epoch": 11.661070636680392, "grad_norm": 0.09327682852745056, "learning_rate": 4.1694646816598035e-05, "loss": 0.0343, "step": 147256 }, { "epoch": 11.711118150142541, "grad_norm": 0.4916613698005676, "learning_rate": 4.14444092492873e-05, "loss": 0.0358, "step": 147888 }, { "epoch": 11.761165663604688, "grad_norm": 0.04805804416537285, "learning_rate": 4.119417168197656e-05, "loss": 0.0343, "step": 148520 }, { "epoch": 11.811213177066836, "grad_norm": 0.019468722864985466, "learning_rate": 4.0943934114665824e-05, "loss": 0.0298, "step": 149152 }, { "epoch": 11.861260690528983, "grad_norm": 0.09931553900241852, "learning_rate": 4.0693696547355085e-05, "loss": 0.0304, "step": 149784 }, { "epoch": 11.911308203991132, "grad_norm": 0.03299326449632645, "learning_rate": 4.0443458980044345e-05, "loss": 0.0355, "step": 150416 }, { "epoch": 11.961355717453278, "grad_norm": 0.11701735109090805, "learning_rate": 4.019322141273361e-05, "loss": 0.0298, "step": 151048 }, { "epoch": 12.0, "eval_gen_len": 12.34792003, "eval_loss": 0.03365420550107956, "eval_runtime": 512.0738, "eval_samples_per_second": 103.934, "eval_steps_per_second": 1.625, "eval_wordacc": 0.94539852, "eval_wordacc_oov": 0.90128577, "step": 151536 }, { "epoch": 12.011403230915427, "grad_norm": 0.14153118431568146, "learning_rate": 3.9942983845422874e-05, "loss": 0.0294, "step": 151680 }, { "epoch": 12.061450744377574, "grad_norm": 0.8496013283729553, "learning_rate": 3.9692746278112134e-05, "loss": 0.0302, "step": 152312 }, { "epoch": 12.111498257839722, "grad_norm": 0.3620108664035797, "learning_rate": 3.9442508710801395e-05, "loss": 0.0293, "step": 152944 }, { "epoch": 12.161545771301869, "grad_norm": 0.8358076810836792, "learning_rate": 3.9192271143490656e-05, "loss": 0.0279, "step": 153576 }, { "epoch": 12.211593284764016, "grad_norm": 2.183274745941162, "learning_rate": 3.894203357617992e-05, "loss": 0.0298, "step": 154208 }, { "epoch": 12.261640798226164, "grad_norm": 0.11177387088537216, "learning_rate": 3.869179600886918e-05, "loss": 0.0279, "step": 154840 }, { "epoch": 12.311688311688311, "grad_norm": 0.16532179713249207, "learning_rate": 3.8441558441558445e-05, "loss": 0.0283, "step": 155472 }, { "epoch": 12.36173582515046, "grad_norm": 0.2667466104030609, "learning_rate": 3.8191320874247705e-05, "loss": 0.0324, "step": 156104 }, { "epoch": 12.411783338612606, "grad_norm": 1.4449559450149536, "learning_rate": 3.794108330693697e-05, "loss": 0.0327, "step": 156736 }, { "epoch": 12.461830852074755, "grad_norm": 1.9251255989074707, "learning_rate": 3.769084573962623e-05, "loss": 0.0293, "step": 157368 }, { "epoch": 12.511878365536901, "grad_norm": 1.981763243675232, "learning_rate": 3.744060817231549e-05, "loss": 0.0296, "step": 158000 }, { "epoch": 12.56192587899905, "grad_norm": 0.5154635906219482, "learning_rate": 3.7190370605004755e-05, "loss": 0.0326, "step": 158632 }, { "epoch": 12.611973392461197, "grad_norm": 0.7047850489616394, "learning_rate": 3.6940133037694016e-05, "loss": 0.0313, "step": 159264 }, { "epoch": 12.662020905923345, "grad_norm": 0.05169570446014404, "learning_rate": 3.6689895470383276e-05, "loss": 0.0311, "step": 159896 }, { "epoch": 12.712068419385492, "grad_norm": 0.3161393105983734, "learning_rate": 3.643965790307254e-05, "loss": 0.0295, "step": 160528 }, { "epoch": 12.76211593284764, "grad_norm": 0.4407559633255005, "learning_rate": 3.6189420335761805e-05, "loss": 0.0307, "step": 161160 }, { "epoch": 12.812163446309787, "grad_norm": 2.090928792953491, "learning_rate": 3.5939182768451065e-05, "loss": 0.0288, "step": 161792 }, { "epoch": 12.862210959771936, "grad_norm": 0.039744604378938675, "learning_rate": 3.568894520114032e-05, "loss": 0.0285, "step": 162424 }, { "epoch": 12.912258473234083, "grad_norm": 0.48494115471839905, "learning_rate": 3.543870763382959e-05, "loss": 0.0302, "step": 163056 }, { "epoch": 12.962305986696231, "grad_norm": 0.38117051124572754, "learning_rate": 3.518847006651885e-05, "loss": 0.0302, "step": 163688 }, { "epoch": 13.0, "eval_gen_len": 12.34829582, "eval_loss": 0.032208241522312164, "eval_runtime": 516.3051, "eval_samples_per_second": 103.082, "eval_steps_per_second": 1.611, "eval_wordacc": 0.94701439, "eval_wordacc_oov": 0.90429282, "step": 164164 }, { "epoch": 13.012353500158378, "grad_norm": 0.5338648557662964, "learning_rate": 3.4938232499208115e-05, "loss": 0.0273, "step": 164320 }, { "epoch": 13.062401013620526, "grad_norm": 0.5787109136581421, "learning_rate": 3.468799493189737e-05, "loss": 0.0278, "step": 164952 }, { "epoch": 13.112448527082673, "grad_norm": 3.402815818786621, "learning_rate": 3.4437757364586636e-05, "loss": 0.0309, "step": 165584 }, { "epoch": 13.162496040544822, "grad_norm": 0.3387065827846527, "learning_rate": 3.41875197972759e-05, "loss": 0.0266, "step": 166216 }, { "epoch": 13.212543554006968, "grad_norm": 0.6726759076118469, "learning_rate": 3.393728222996516e-05, "loss": 0.0257, "step": 166848 }, { "epoch": 13.262591067469117, "grad_norm": 0.039511628448963165, "learning_rate": 3.368704466265442e-05, "loss": 0.0285, "step": 167480 }, { "epoch": 13.312638580931264, "grad_norm": 0.16470862925052643, "learning_rate": 3.343680709534368e-05, "loss": 0.0265, "step": 168112 }, { "epoch": 13.362686094393412, "grad_norm": 1.1072640419006348, "learning_rate": 3.318656952803295e-05, "loss": 0.0289, "step": 168744 }, { "epoch": 13.412733607855559, "grad_norm": 0.4822203516960144, "learning_rate": 3.293633196072221e-05, "loss": 0.0282, "step": 169376 }, { "epoch": 13.462781121317708, "grad_norm": 0.29173898696899414, "learning_rate": 3.268609439341147e-05, "loss": 0.025, "step": 170008 }, { "epoch": 13.512828634779854, "grad_norm": 0.12304585427045822, "learning_rate": 3.243585682610073e-05, "loss": 0.0286, "step": 170640 }, { "epoch": 13.562876148242001, "grad_norm": 0.02250618301331997, "learning_rate": 3.218561925878999e-05, "loss": 0.0261, "step": 171272 }, { "epoch": 13.61292366170415, "grad_norm": 1.214078664779663, "learning_rate": 3.193538169147926e-05, "loss": 0.0265, "step": 171904 }, { "epoch": 13.662971175166296, "grad_norm": 1.080644965171814, "learning_rate": 3.168514412416851e-05, "loss": 0.0264, "step": 172536 }, { "epoch": 13.713018688628445, "grad_norm": 0.47105857729911804, "learning_rate": 3.143490655685778e-05, "loss": 0.03, "step": 173168 }, { "epoch": 13.763066202090592, "grad_norm": 0.8579492568969727, "learning_rate": 3.118466898954704e-05, "loss": 0.03, "step": 173800 }, { "epoch": 13.81311371555274, "grad_norm": 1.1287840604782104, "learning_rate": 3.0934431422236307e-05, "loss": 0.0274, "step": 174432 }, { "epoch": 13.863161229014887, "grad_norm": 1.2223260402679443, "learning_rate": 3.068419385492556e-05, "loss": 0.0275, "step": 175064 }, { "epoch": 13.913208742477035, "grad_norm": 0.6263837814331055, "learning_rate": 3.0433956287614825e-05, "loss": 0.03, "step": 175696 }, { "epoch": 13.963256255939182, "grad_norm": 2.3187966346740723, "learning_rate": 3.018371872030409e-05, "loss": 0.0277, "step": 176328 }, { "epoch": 14.0, "eval_gen_len": 12.35064447, "eval_loss": 0.031636711210012436, "eval_runtime": 514.9724, "eval_samples_per_second": 103.349, "eval_steps_per_second": 1.616, "eval_wordacc": 0.94785991, "eval_wordacc_oov": 0.9040336, "step": 176792 }, { "epoch": 14.01330376940133, "grad_norm": 0.09013538807630539, "learning_rate": 2.993348115299335e-05, "loss": 0.0291, "step": 176960 }, { "epoch": 14.063351282863477, "grad_norm": 0.7201142907142639, "learning_rate": 2.9683243585682614e-05, "loss": 0.0216, "step": 177592 }, { "epoch": 14.113398796325626, "grad_norm": 0.2142077535390854, "learning_rate": 2.943300601837187e-05, "loss": 0.0264, "step": 178224 }, { "epoch": 14.163446309787773, "grad_norm": 0.618773341178894, "learning_rate": 2.918276845106114e-05, "loss": 0.0248, "step": 178856 }, { "epoch": 14.213493823249921, "grad_norm": 1.0433090925216675, "learning_rate": 2.8932530883750396e-05, "loss": 0.0286, "step": 179488 }, { "epoch": 14.263541336712068, "grad_norm": 0.11434757709503174, "learning_rate": 2.8682293316439656e-05, "loss": 0.0253, "step": 180120 }, { "epoch": 14.313588850174217, "grad_norm": 1.1828645467758179, "learning_rate": 2.843205574912892e-05, "loss": 0.0245, "step": 180752 }, { "epoch": 14.363636363636363, "grad_norm": 0.3703102171421051, "learning_rate": 2.818181818181818e-05, "loss": 0.0261, "step": 181384 }, { "epoch": 14.413683877098512, "grad_norm": 0.20109039545059204, "learning_rate": 2.7931580614507445e-05, "loss": 0.0255, "step": 182016 }, { "epoch": 14.463731390560659, "grad_norm": 0.627756655216217, "learning_rate": 2.7681343047196706e-05, "loss": 0.0228, "step": 182648 }, { "epoch": 14.513778904022807, "grad_norm": 0.04185875132679939, "learning_rate": 2.743110547988597e-05, "loss": 0.0237, "step": 183280 }, { "epoch": 14.563826417484954, "grad_norm": 0.10042019188404083, "learning_rate": 2.718086791257523e-05, "loss": 0.0272, "step": 183912 }, { "epoch": 14.613873930947102, "grad_norm": 1.0942548513412476, "learning_rate": 2.693063034526449e-05, "loss": 0.0242, "step": 184544 }, { "epoch": 14.66392144440925, "grad_norm": 0.7228376269340515, "learning_rate": 2.6680392777953756e-05, "loss": 0.0252, "step": 185176 }, { "epoch": 14.713968957871398, "grad_norm": 0.6470568776130676, "learning_rate": 2.6430155210643016e-05, "loss": 0.0269, "step": 185808 }, { "epoch": 14.764016471333544, "grad_norm": 0.4635055363178253, "learning_rate": 2.617991764333228e-05, "loss": 0.0243, "step": 186440 }, { "epoch": 14.814063984795691, "grad_norm": 0.43463265895843506, "learning_rate": 2.592968007602154e-05, "loss": 0.0265, "step": 187072 }, { "epoch": 14.86411149825784, "grad_norm": 0.7284203767776489, "learning_rate": 2.5679442508710805e-05, "loss": 0.0269, "step": 187704 }, { "epoch": 14.914159011719988, "grad_norm": 1.462714433670044, "learning_rate": 2.5429204941400066e-05, "loss": 0.0236, "step": 188336 }, { "epoch": 14.964206525182135, "grad_norm": 0.029578888788819313, "learning_rate": 2.5178967374089323e-05, "loss": 0.0277, "step": 188968 }, { "epoch": 15.0, "eval_gen_len": 12.35139604, "eval_loss": 0.032311730086803436, "eval_runtime": 516.3173, "eval_samples_per_second": 103.08, "eval_steps_per_second": 1.611, "eval_wordacc": 0.94876179, "eval_wordacc_oov": 0.90299668, "step": 189420 }, { "epoch": 15.014254038644282, "grad_norm": 0.3032616376876831, "learning_rate": 2.4928729806778587e-05, "loss": 0.0275, "step": 189600 }, { "epoch": 15.06430155210643, "grad_norm": 0.08022745698690414, "learning_rate": 2.467849223946785e-05, "loss": 0.0212, "step": 190232 }, { "epoch": 15.114349065568577, "grad_norm": 0.06174493208527565, "learning_rate": 2.4428254672157112e-05, "loss": 0.0233, "step": 190864 }, { "epoch": 15.164396579030726, "grad_norm": 0.7364438772201538, "learning_rate": 2.4178017104846373e-05, "loss": 0.0214, "step": 191496 }, { "epoch": 15.214444092492872, "grad_norm": 0.022581493481993675, "learning_rate": 2.3927779537535637e-05, "loss": 0.0236, "step": 192128 }, { "epoch": 15.26449160595502, "grad_norm": 0.03313857316970825, "learning_rate": 2.3677541970224898e-05, "loss": 0.0246, "step": 192760 }, { "epoch": 15.314539119417168, "grad_norm": 0.7156698703765869, "learning_rate": 2.3427304402914162e-05, "loss": 0.0263, "step": 193392 }, { "epoch": 15.364586632879316, "grad_norm": 0.626669704914093, "learning_rate": 2.3177066835603423e-05, "loss": 0.0219, "step": 194024 }, { "epoch": 15.414634146341463, "grad_norm": 1.3695429563522339, "learning_rate": 2.2926829268292687e-05, "loss": 0.0244, "step": 194656 }, { "epoch": 15.464681659803611, "grad_norm": 0.03239896893501282, "learning_rate": 2.2676591700981947e-05, "loss": 0.0217, "step": 195288 }, { "epoch": 15.514729173265758, "grad_norm": 0.4434269666671753, "learning_rate": 2.2426354133671208e-05, "loss": 0.026, "step": 195920 }, { "epoch": 15.564776686727907, "grad_norm": 0.11541499197483063, "learning_rate": 2.217611656636047e-05, "loss": 0.0251, "step": 196552 }, { "epoch": 15.614824200190053, "grad_norm": 0.11521150171756744, "learning_rate": 2.1925878999049733e-05, "loss": 0.0234, "step": 197184 }, { "epoch": 15.664871713652202, "grad_norm": 0.10318135470151901, "learning_rate": 2.1675641431738994e-05, "loss": 0.0239, "step": 197816 }, { "epoch": 15.714919227114349, "grad_norm": 0.181732639670372, "learning_rate": 2.1425403864428258e-05, "loss": 0.0231, "step": 198448 }, { "epoch": 15.764966740576497, "grad_norm": 0.43891534209251404, "learning_rate": 2.117516629711752e-05, "loss": 0.0258, "step": 199080 }, { "epoch": 15.815014254038644, "grad_norm": 0.2669106125831604, "learning_rate": 2.0924928729806782e-05, "loss": 0.0258, "step": 199712 }, { "epoch": 15.865061767500793, "grad_norm": 0.787148118019104, "learning_rate": 2.067469116249604e-05, "loss": 0.023, "step": 200344 }, { "epoch": 15.91510928096294, "grad_norm": 0.09489845484495163, "learning_rate": 2.0424453595185304e-05, "loss": 0.0233, "step": 200976 }, { "epoch": 15.965156794425088, "grad_norm": 0.07817026227712631, "learning_rate": 2.0174216027874565e-05, "loss": 0.0245, "step": 201608 }, { "epoch": 16.0, "eval_gen_len": 12.35013716, "eval_loss": 0.03139541670680046, "eval_runtime": 516.248, "eval_samples_per_second": 103.094, "eval_steps_per_second": 1.612, "eval_wordacc": 0.95127955, "eval_wordacc_oov": 0.90724803, "step": 202048 }, { "epoch": 16.015204307887235, "grad_norm": 0.026963738724589348, "learning_rate": 1.992397846056383e-05, "loss": 0.0209, "step": 202240 }, { "epoch": 16.06525182134938, "grad_norm": 0.031288400292396545, "learning_rate": 1.967374089325309e-05, "loss": 0.0211, "step": 202872 }, { "epoch": 16.11529933481153, "grad_norm": 0.04831352084875107, "learning_rate": 1.9423503325942354e-05, "loss": 0.0216, "step": 203504 }, { "epoch": 16.16534684827368, "grad_norm": 0.3481796383857727, "learning_rate": 1.9173265758631614e-05, "loss": 0.0215, "step": 204136 }, { "epoch": 16.215394361735825, "grad_norm": 0.21324285864830017, "learning_rate": 1.8923028191320875e-05, "loss": 0.0218, "step": 204768 }, { "epoch": 16.265441875197972, "grad_norm": 1.6428892612457275, "learning_rate": 1.8672790624010136e-05, "loss": 0.021, "step": 205400 }, { "epoch": 16.31548938866012, "grad_norm": 0.2430795133113861, "learning_rate": 1.84225530566994e-05, "loss": 0.02, "step": 206032 }, { "epoch": 16.36553690212227, "grad_norm": 0.5367847681045532, "learning_rate": 1.817231548938866e-05, "loss": 0.0209, "step": 206664 }, { "epoch": 16.415584415584416, "grad_norm": 0.08580495417118073, "learning_rate": 1.7922077922077925e-05, "loss": 0.0204, "step": 207296 }, { "epoch": 16.465631929046562, "grad_norm": 0.6704065203666687, "learning_rate": 1.7671840354767185e-05, "loss": 0.0218, "step": 207928 }, { "epoch": 16.51567944250871, "grad_norm": 1.014145016670227, "learning_rate": 1.742160278745645e-05, "loss": 0.0216, "step": 208560 }, { "epoch": 16.56572695597086, "grad_norm": 0.040132321417331696, "learning_rate": 1.7171365220145707e-05, "loss": 0.0249, "step": 209192 }, { "epoch": 16.615774469433006, "grad_norm": 1.3288291692733765, "learning_rate": 1.692112765283497e-05, "loss": 0.0216, "step": 209824 }, { "epoch": 16.665821982895153, "grad_norm": 0.6535305976867676, "learning_rate": 1.667089008552423e-05, "loss": 0.0221, "step": 210456 }, { "epoch": 16.7158694963573, "grad_norm": 0.516942024230957, "learning_rate": 1.6420652518213496e-05, "loss": 0.0209, "step": 211088 }, { "epoch": 16.76591700981945, "grad_norm": 0.6040933728218079, "learning_rate": 1.6170414950902756e-05, "loss": 0.0231, "step": 211720 }, { "epoch": 16.815964523281597, "grad_norm": 0.07999061793088913, "learning_rate": 1.592017738359202e-05, "loss": 0.023, "step": 212352 }, { "epoch": 16.866012036743744, "grad_norm": 2.897904872894287, "learning_rate": 1.566993981628128e-05, "loss": 0.0227, "step": 212984 }, { "epoch": 16.91605955020589, "grad_norm": 0.2583966553211212, "learning_rate": 1.5419702248970542e-05, "loss": 0.0218, "step": 213616 }, { "epoch": 16.96610706366804, "grad_norm": 1.3506840467453003, "learning_rate": 1.5169464681659804e-05, "loss": 0.0235, "step": 214248 }, { "epoch": 17.0, "eval_gen_len": 12.35109541, "eval_loss": 0.03128550201654434, "eval_runtime": 515.496, "eval_samples_per_second": 103.244, "eval_steps_per_second": 1.614, "eval_wordacc": 0.95197475, "eval_wordacc_oov": 0.90714434, "step": 214676 }, { "epoch": 17.016154577130187, "grad_norm": 0.28379184007644653, "learning_rate": 1.4919227114349067e-05, "loss": 0.0196, "step": 214880 }, { "epoch": 17.066202090592334, "grad_norm": 0.2609073221683502, "learning_rate": 1.4668989547038327e-05, "loss": 0.0199, "step": 215512 }, { "epoch": 17.11624960405448, "grad_norm": 0.22402538359165192, "learning_rate": 1.441875197972759e-05, "loss": 0.0203, "step": 216144 }, { "epoch": 17.16629711751663, "grad_norm": 0.020612630993127823, "learning_rate": 1.4168514412416852e-05, "loss": 0.0196, "step": 216776 }, { "epoch": 17.216344630978778, "grad_norm": 0.036708053201436996, "learning_rate": 1.3918276845106115e-05, "loss": 0.0164, "step": 217408 }, { "epoch": 17.266392144440925, "grad_norm": 1.0852469205856323, "learning_rate": 1.3668039277795375e-05, "loss": 0.0197, "step": 218040 }, { "epoch": 17.31643965790307, "grad_norm": 0.08419329673051834, "learning_rate": 1.3417801710484638e-05, "loss": 0.0223, "step": 218672 }, { "epoch": 17.366487171365222, "grad_norm": 0.8586103320121765, "learning_rate": 1.31675641431739e-05, "loss": 0.0217, "step": 219304 }, { "epoch": 17.41653468482737, "grad_norm": 0.41822636127471924, "learning_rate": 1.2917326575863162e-05, "loss": 0.0203, "step": 219936 }, { "epoch": 17.466582198289515, "grad_norm": 0.16271276772022247, "learning_rate": 1.2667089008552425e-05, "loss": 0.0211, "step": 220568 }, { "epoch": 17.516629711751662, "grad_norm": 0.7023501992225647, "learning_rate": 1.2416851441241686e-05, "loss": 0.0197, "step": 221200 }, { "epoch": 17.56667722521381, "grad_norm": 0.038701217621564865, "learning_rate": 1.2166613873930948e-05, "loss": 0.022, "step": 221832 }, { "epoch": 17.61672473867596, "grad_norm": 0.03481602668762207, "learning_rate": 1.191637630662021e-05, "loss": 0.0178, "step": 222464 }, { "epoch": 17.666772252138106, "grad_norm": 0.3818272352218628, "learning_rate": 1.1666138739309473e-05, "loss": 0.0204, "step": 223096 }, { "epoch": 17.716819765600253, "grad_norm": 1.1725687980651855, "learning_rate": 1.1415901171998734e-05, "loss": 0.0214, "step": 223728 }, { "epoch": 17.7668672790624, "grad_norm": 0.5618239641189575, "learning_rate": 1.1165663604687996e-05, "loss": 0.0198, "step": 224360 }, { "epoch": 17.81691479252455, "grad_norm": 0.11917304992675781, "learning_rate": 1.0915426037377258e-05, "loss": 0.0203, "step": 224992 }, { "epoch": 17.866962305986696, "grad_norm": 0.1154344379901886, "learning_rate": 1.066518847006652e-05, "loss": 0.0217, "step": 225624 }, { "epoch": 17.917009819448843, "grad_norm": 0.12502028048038483, "learning_rate": 1.0414950902755781e-05, "loss": 0.0217, "step": 226256 }, { "epoch": 17.96705733291099, "grad_norm": 0.27321675419807434, "learning_rate": 1.0164713335445044e-05, "loss": 0.0206, "step": 226888 }, { "epoch": 18.0, "eval_gen_len": 12.35015595, "eval_loss": 0.030980365350842476, "eval_runtime": 547.6731, "eval_samples_per_second": 97.178, "eval_steps_per_second": 1.519, "eval_wordacc": 0.95308331, "eval_wordacc_oov": 0.90844048, "step": 227304 }, { "epoch": 18.01710484637314, "grad_norm": 0.15305259823799133, "learning_rate": 9.914475768134306e-06, "loss": 0.018, "step": 227520 }, { "epoch": 18.067152359835287, "grad_norm": 2.5299997329711914, "learning_rate": 9.664238200823567e-06, "loss": 0.0183, "step": 228152 }, { "epoch": 18.117199873297434, "grad_norm": 0.2780356705188751, "learning_rate": 9.41400063351283e-06, "loss": 0.0179, "step": 228784 }, { "epoch": 18.16724738675958, "grad_norm": 0.06621097773313522, "learning_rate": 9.163763066202092e-06, "loss": 0.0196, "step": 229416 }, { "epoch": 18.21729490022173, "grad_norm": 0.05437783896923065, "learning_rate": 8.913525498891354e-06, "loss": 0.0186, "step": 230048 }, { "epoch": 18.267342413683878, "grad_norm": 0.62905353307724, "learning_rate": 8.663287931580615e-06, "loss": 0.0202, "step": 230680 }, { "epoch": 18.317389927146024, "grad_norm": 0.03628231957554817, "learning_rate": 8.413050364269877e-06, "loss": 0.0186, "step": 231312 }, { "epoch": 18.36743744060817, "grad_norm": 0.1532977819442749, "learning_rate": 8.16281279695914e-06, "loss": 0.0197, "step": 231944 }, { "epoch": 18.41748495407032, "grad_norm": 0.15254627168178558, "learning_rate": 7.9125752296484e-06, "loss": 0.0177, "step": 232576 }, { "epoch": 18.467532467532468, "grad_norm": 1.0562399625778198, "learning_rate": 7.662337662337663e-06, "loss": 0.0172, "step": 233208 }, { "epoch": 18.517579980994615, "grad_norm": 0.4453679323196411, "learning_rate": 7.412100095026925e-06, "loss": 0.0197, "step": 233840 }, { "epoch": 18.56762749445676, "grad_norm": 0.26340222358703613, "learning_rate": 7.161862527716187e-06, "loss": 0.018, "step": 234472 }, { "epoch": 18.617675007918912, "grad_norm": 0.052670177072286606, "learning_rate": 6.911624960405448e-06, "loss": 0.0184, "step": 235104 }, { "epoch": 18.66772252138106, "grad_norm": 0.026885686442255974, "learning_rate": 6.661387393094711e-06, "loss": 0.0183, "step": 235736 }, { "epoch": 18.717770034843205, "grad_norm": 0.34494689106941223, "learning_rate": 6.411149825783973e-06, "loss": 0.0192, "step": 236368 }, { "epoch": 18.767817548305352, "grad_norm": 0.0610111765563488, "learning_rate": 6.160912258473235e-06, "loss": 0.0179, "step": 237000 }, { "epoch": 18.817865061767503, "grad_norm": 0.02205130271613598, "learning_rate": 5.910674691162496e-06, "loss": 0.0218, "step": 237632 }, { "epoch": 18.86791257522965, "grad_norm": 0.046364884823560715, "learning_rate": 5.660437123851759e-06, "loss": 0.0167, "step": 238264 }, { "epoch": 18.917960088691796, "grad_norm": 0.8253034949302673, "learning_rate": 5.41019955654102e-06, "loss": 0.0206, "step": 238896 }, { "epoch": 18.968007602153943, "grad_norm": 0.07394740730524063, "learning_rate": 5.159961989230283e-06, "loss": 0.0178, "step": 239528 }, { "epoch": 19.0, "eval_gen_len": 12.35068205, "eval_loss": 0.03073255531489849, "eval_runtime": 534.9855, "eval_samples_per_second": 99.483, "eval_steps_per_second": 1.555, "eval_wordacc": 0.95447371, "eval_wordacc_oov": 0.9093737, "step": 239932 }, { "epoch": 19.01805511561609, "grad_norm": 0.05381210148334503, "learning_rate": 4.909724421919544e-06, "loss": 0.017, "step": 240160 }, { "epoch": 19.06810262907824, "grad_norm": 0.05161185935139656, "learning_rate": 4.659486854608807e-06, "loss": 0.0184, "step": 240792 }, { "epoch": 19.118150142540387, "grad_norm": 0.054974284023046494, "learning_rate": 4.409249287298068e-06, "loss": 0.0165, "step": 241424 }, { "epoch": 19.168197656002533, "grad_norm": 0.4919557571411133, "learning_rate": 4.15901171998733e-06, "loss": 0.0171, "step": 242056 }, { "epoch": 19.21824516946468, "grad_norm": 0.8104053735733032, "learning_rate": 3.908774152676592e-06, "loss": 0.0161, "step": 242688 }, { "epoch": 19.26829268292683, "grad_norm": 0.04452453926205635, "learning_rate": 3.6585365853658537e-06, "loss": 0.017, "step": 243320 }, { "epoch": 19.318340196388977, "grad_norm": 2.039726972579956, "learning_rate": 3.408299018055116e-06, "loss": 0.0188, "step": 243952 }, { "epoch": 19.368387709851124, "grad_norm": 0.05839849263429642, "learning_rate": 3.1580614507443777e-06, "loss": 0.0174, "step": 244584 }, { "epoch": 19.41843522331327, "grad_norm": 0.06209117919206619, "learning_rate": 2.9078238834336396e-06, "loss": 0.0177, "step": 245216 }, { "epoch": 19.46848273677542, "grad_norm": 0.023634545505046844, "learning_rate": 2.6575863161229016e-06, "loss": 0.0187, "step": 245848 }, { "epoch": 19.518530250237568, "grad_norm": 0.4597359001636505, "learning_rate": 2.4073487488121636e-06, "loss": 0.0157, "step": 246480 }, { "epoch": 19.568577763699714, "grad_norm": 0.03822094574570656, "learning_rate": 2.1571111815014256e-06, "loss": 0.0191, "step": 247112 }, { "epoch": 19.61862527716186, "grad_norm": 0.028475474566221237, "learning_rate": 1.9068736141906876e-06, "loss": 0.0184, "step": 247744 }, { "epoch": 19.66867279062401, "grad_norm": 2.0368409156799316, "learning_rate": 1.6566360468799495e-06, "loss": 0.016, "step": 248376 }, { "epoch": 19.71872030408616, "grad_norm": 0.6395711898803711, "learning_rate": 1.4063984795692113e-06, "loss": 0.0182, "step": 249008 }, { "epoch": 19.768767817548305, "grad_norm": 0.11478591710329056, "learning_rate": 1.1561609122584733e-06, "loss": 0.0172, "step": 249640 }, { "epoch": 19.818815331010452, "grad_norm": 2.0897409915924072, "learning_rate": 9.059233449477353e-07, "loss": 0.0178, "step": 250272 }, { "epoch": 19.868862844472602, "grad_norm": 1.6272854804992676, "learning_rate": 6.556857776369973e-07, "loss": 0.0165, "step": 250904 }, { "epoch": 19.91891035793475, "grad_norm": 0.018388045951724052, "learning_rate": 4.0544821032625913e-07, "loss": 0.0153, "step": 251536 }, { "epoch": 19.968957871396896, "grad_norm": 0.097270667552948, "learning_rate": 1.5521064301552106e-07, "loss": 0.016, "step": 252168 }, { "epoch": 20.0, "eval_gen_len": 12.35156514, "eval_loss": 0.030767865478992462, "eval_runtime": 535.8673, "eval_samples_per_second": 99.319, "eval_steps_per_second": 1.553, "eval_wordacc": 0.95458645, "eval_wordacc_oov": 0.90963293, "step": 252560 }, { "epoch": 20.0, "step": 252560, "total_flos": 1486519029399552.0, "train_loss": 0.04684994914012881, "train_runtime": 19459.2196, "train_samples_per_second": 103.832, "train_steps_per_second": 12.979 } ], "logging_steps": 632, "max_steps": 252560, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1486519029399552.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }