|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9981298423724285, |
|
"eval_steps": 200, |
|
"global_step": 467, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0021373230029388193, |
|
"grad_norm": 0.0, |
|
"learning_rate": 0.0, |
|
"logits": -2.7276527881622314, |
|
"logps": -123.19757843017578, |
|
"loss": -0.7819, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.010686615014694095, |
|
"grad_norm": 0.0, |
|
"learning_rate": 0.0, |
|
"logits": -2.8715224266052246, |
|
"logps": -234.59034729003906, |
|
"loss": -0.7834, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.02137323002938819, |
|
"grad_norm": 62.58979166531292, |
|
"learning_rate": 3.191489361702127e-08, |
|
"logits": -2.846045732498169, |
|
"logps": -248.165771484375, |
|
"loss": -0.7829, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03205984504408229, |
|
"grad_norm": 165.21106318043798, |
|
"learning_rate": 7.446808510638298e-08, |
|
"logits": -2.777956485748291, |
|
"logps": -229.2180938720703, |
|
"loss": -0.7878, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.04274646005877638, |
|
"grad_norm": 73.29436509866687, |
|
"learning_rate": 1.2765957446808508e-07, |
|
"logits": -2.769435405731201, |
|
"logps": -204.04058837890625, |
|
"loss": -0.86, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.053433075073470476, |
|
"grad_norm": 43.89267095591359, |
|
"learning_rate": 1.8085106382978725e-07, |
|
"logits": -2.9436452388763428, |
|
"logps": -289.7768859863281, |
|
"loss": -0.939, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.06411969008816458, |
|
"grad_norm": 29.65124187980817, |
|
"learning_rate": 2.3404255319148937e-07, |
|
"logits": -2.9300923347473145, |
|
"logps": -277.5953674316406, |
|
"loss": -0.9585, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07480630510285867, |
|
"grad_norm": 27.17018222606885, |
|
"learning_rate": 2.872340425531915e-07, |
|
"logits": -2.9261233806610107, |
|
"logps": -234.77285766601562, |
|
"loss": -0.9666, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.08549292011755276, |
|
"grad_norm": 23.213967474767983, |
|
"learning_rate": 3.404255319148936e-07, |
|
"logits": -2.960068464279175, |
|
"logps": -236.64358520507812, |
|
"loss": -0.9697, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09617953513224686, |
|
"grad_norm": 21.820617838068017, |
|
"learning_rate": 3.9361702127659574e-07, |
|
"logits": -2.910217046737671, |
|
"logps": -254.3408203125, |
|
"loss": -0.9706, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.10686615014694095, |
|
"grad_norm": 19.42383634052037, |
|
"learning_rate": 4.4680851063829783e-07, |
|
"logits": -2.80342960357666, |
|
"logps": -266.91448974609375, |
|
"loss": -0.9712, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.11755276516163506, |
|
"grad_norm": 19.073870329982725, |
|
"learning_rate": 5e-07, |
|
"logits": -3.0574581623077393, |
|
"logps": -278.97833251953125, |
|
"loss": -0.9752, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.12823938017632916, |
|
"grad_norm": 22.140537506897275, |
|
"learning_rate": 4.998251761970996e-07, |
|
"logits": -2.982924699783325, |
|
"logps": -246.08999633789062, |
|
"loss": -0.9711, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.13892599519102325, |
|
"grad_norm": 22.027187117031247, |
|
"learning_rate": 4.993009492952949e-07, |
|
"logits": -3.066633939743042, |
|
"logps": -287.9792175292969, |
|
"loss": -0.9739, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.14961261020571734, |
|
"grad_norm": 28.360950734285428, |
|
"learning_rate": 4.984280524733107e-07, |
|
"logits": -2.9325830936431885, |
|
"logps": -254.9385528564453, |
|
"loss": -0.969, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.16029922522041143, |
|
"grad_norm": 26.470079510381215, |
|
"learning_rate": 4.972077065562821e-07, |
|
"logits": -2.8176350593566895, |
|
"logps": -247.518798828125, |
|
"loss": -0.9746, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.17098584023510552, |
|
"grad_norm": 19.869439451969985, |
|
"learning_rate": 4.956416183083221e-07, |
|
"logits": -3.0621449947357178, |
|
"logps": -289.036376953125, |
|
"loss": -0.974, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.18167245524979964, |
|
"grad_norm": 18.092489458888704, |
|
"learning_rate": 4.937319780454559e-07, |
|
"logits": -2.8688838481903076, |
|
"logps": -279.94183349609375, |
|
"loss": -0.9776, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.19235907026449373, |
|
"grad_norm": 23.563928671841666, |
|
"learning_rate": 4.91481456572267e-07, |
|
"logits": -3.0502922534942627, |
|
"logps": -219.67733764648438, |
|
"loss": -0.9768, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.20304568527918782, |
|
"grad_norm": 18.6682864992676, |
|
"learning_rate": 4.888932014465352e-07, |
|
"logits": -2.9103236198425293, |
|
"logps": -270.4436340332031, |
|
"loss": -0.9753, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.2137323002938819, |
|
"grad_norm": 18.181568458747662, |
|
"learning_rate": 4.859708325770919e-07, |
|
"logits": -2.9558568000793457, |
|
"logps": -222.9886932373047, |
|
"loss": -0.9758, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.224418915308576, |
|
"grad_norm": 22.1137260269499, |
|
"learning_rate": 4.82718437161051e-07, |
|
"logits": -3.025911808013916, |
|
"logps": -265.393310546875, |
|
"loss": -0.9784, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.2351055303232701, |
|
"grad_norm": 18.34089290810698, |
|
"learning_rate": 4.79140563967494e-07, |
|
"logits": -3.0452187061309814, |
|
"logps": -291.10003662109375, |
|
"loss": -0.977, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.2457921453379642, |
|
"grad_norm": 24.000868485282172, |
|
"learning_rate": 4.752422169756047e-07, |
|
"logits": -2.9980249404907227, |
|
"logps": -270.8134765625, |
|
"loss": -0.9769, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.2564787603526583, |
|
"grad_norm": 22.432824006478324, |
|
"learning_rate": 4.710288483761524e-07, |
|
"logits": -3.0276284217834473, |
|
"logps": -242.854248046875, |
|
"loss": -0.9781, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.2671653753673524, |
|
"grad_norm": 19.623943572874033, |
|
"learning_rate": 4.6650635094610966e-07, |
|
"logits": -2.9869930744171143, |
|
"logps": -264.6936340332031, |
|
"loss": -0.9784, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2778519903820465, |
|
"grad_norm": 16.514800564439714, |
|
"learning_rate": 4.6168104980707103e-07, |
|
"logits": -3.074154853820801, |
|
"logps": -254.39559936523438, |
|
"loss": -0.9789, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.2885386053967406, |
|
"grad_norm": 17.940806695880376, |
|
"learning_rate": 4.565596935789987e-07, |
|
"logits": -3.064635753631592, |
|
"logps": -253.6359405517578, |
|
"loss": -0.9772, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.2992252204114347, |
|
"grad_norm": 15.99587855831518, |
|
"learning_rate": 4.511494449416671e-07, |
|
"logits": -3.021132707595825, |
|
"logps": -291.09625244140625, |
|
"loss": -0.9776, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.30991183542612877, |
|
"grad_norm": 15.5792510590759, |
|
"learning_rate": 4.4545787061700746e-07, |
|
"logits": -2.853203296661377, |
|
"logps": -281.72210693359375, |
|
"loss": -0.9771, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.32059845044082286, |
|
"grad_norm": 17.480990084091566, |
|
"learning_rate": 4.394929307863632e-07, |
|
"logits": -2.993375539779663, |
|
"logps": -265.2592468261719, |
|
"loss": -0.9753, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.33128506545551695, |
|
"grad_norm": 17.70526730878467, |
|
"learning_rate": 4.332629679574565e-07, |
|
"logits": -3.0605502128601074, |
|
"logps": -305.0889587402344, |
|
"loss": -0.9766, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.34197168047021104, |
|
"grad_norm": 18.40262893823541, |
|
"learning_rate": 4.2677669529663686e-07, |
|
"logits": -2.9697351455688477, |
|
"logps": -306.7701721191406, |
|
"loss": -0.9792, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.3526582954849052, |
|
"grad_norm": 20.54973812877713, |
|
"learning_rate": 4.200431844427298e-07, |
|
"logits": -3.0495619773864746, |
|
"logps": -295.7186279296875, |
|
"loss": -0.9785, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.36334491049959927, |
|
"grad_norm": 23.008419291973887, |
|
"learning_rate": 4.130718528195303e-07, |
|
"logits": -2.997997283935547, |
|
"logps": -275.9290466308594, |
|
"loss": -0.9772, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.37403152551429336, |
|
"grad_norm": 22.007125412303857, |
|
"learning_rate": 4.058724504646834e-07, |
|
"logits": -2.998748779296875, |
|
"logps": -276.2112731933594, |
|
"loss": -0.9773, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.38471814052898745, |
|
"grad_norm": 14.558828760504346, |
|
"learning_rate": 3.9845504639337535e-07, |
|
"logits": -3.035895586013794, |
|
"logps": -280.32708740234375, |
|
"loss": -0.9778, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.39540475554368154, |
|
"grad_norm": 15.094528036756223, |
|
"learning_rate": 3.908300145159055e-07, |
|
"logits": -2.99044132232666, |
|
"logps": -231.31869506835938, |
|
"loss": -0.9782, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.40609137055837563, |
|
"grad_norm": 19.931751129133623, |
|
"learning_rate": 3.8300801912883414e-07, |
|
"logits": -3.1248080730438232, |
|
"logps": -272.81658935546875, |
|
"loss": -0.9794, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.4167779855730697, |
|
"grad_norm": 19.163508181353045, |
|
"learning_rate": 3.75e-07, |
|
"logits": -3.047405242919922, |
|
"logps": -332.17578125, |
|
"loss": -0.979, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.4274646005877638, |
|
"grad_norm": 16.407530767721205, |
|
"learning_rate": 3.668171570682655e-07, |
|
"logits": -3.0323641300201416, |
|
"logps": -251.59506225585938, |
|
"loss": -0.9788, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4274646005877638, |
|
"eval_logits": -3.0493266582489014, |
|
"eval_logps": -289.3082275390625, |
|
"eval_loss": -0.9775694608688354, |
|
"eval_runtime": 586.0302, |
|
"eval_samples_per_second": 3.358, |
|
"eval_steps_per_second": 0.21, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4381512156024579, |
|
"grad_norm": 18.72596941199976, |
|
"learning_rate": 3.584709347793895e-07, |
|
"logits": -2.9899675846099854, |
|
"logps": -264.5972595214844, |
|
"loss": -0.9774, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.448837830617152, |
|
"grad_norm": 14.684631204607477, |
|
"learning_rate": 3.499730060799352e-07, |
|
"logits": -2.9389395713806152, |
|
"logps": -270.917236328125, |
|
"loss": -0.977, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.45952444563184613, |
|
"grad_norm": 19.60342980482045, |
|
"learning_rate": 3.413352560915988e-07, |
|
"logits": -2.9131505489349365, |
|
"logps": -250.50747680664062, |
|
"loss": -0.9779, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.4702110606465402, |
|
"grad_norm": 12.44502702657658, |
|
"learning_rate": 3.325697654887918e-07, |
|
"logits": -2.8331167697906494, |
|
"logps": -209.47265625, |
|
"loss": -0.9759, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.4808976756612343, |
|
"grad_norm": 19.223731552060446, |
|
"learning_rate": 3.2368879360272606e-07, |
|
"logits": -3.003699779510498, |
|
"logps": -293.6314392089844, |
|
"loss": -0.9776, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.4915842906759284, |
|
"grad_norm": 18.20205804125661, |
|
"learning_rate": 3.147047612756302e-07, |
|
"logits": -3.0232555866241455, |
|
"logps": -305.18316650390625, |
|
"loss": -0.9783, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.5022709056906225, |
|
"grad_norm": 16.46859278231364, |
|
"learning_rate": 3.056302334890786e-07, |
|
"logits": -2.9345574378967285, |
|
"logps": -237.68637084960938, |
|
"loss": -0.9797, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.5129575207053166, |
|
"grad_norm": 16.465984487681027, |
|
"learning_rate": 2.964779017907287e-07, |
|
"logits": -2.9307010173797607, |
|
"logps": -286.076171875, |
|
"loss": -0.9805, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.5236441357200107, |
|
"grad_norm": 17.54623610211809, |
|
"learning_rate": 2.872605665440436e-07, |
|
"logits": -2.974175453186035, |
|
"logps": -262.3833312988281, |
|
"loss": -0.9787, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.5343307507347048, |
|
"grad_norm": 18.88275216616494, |
|
"learning_rate": 2.7799111902582693e-07, |
|
"logits": -3.046766996383667, |
|
"logps": -251.1654052734375, |
|
"loss": -0.9818, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.5450173657493989, |
|
"grad_norm": 16.60487185722242, |
|
"learning_rate": 2.6868252339660607e-07, |
|
"logits": -2.889235496520996, |
|
"logps": -254.0398712158203, |
|
"loss": -0.9779, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.555703980764093, |
|
"grad_norm": 19.312080394342164, |
|
"learning_rate": 2.593477985690815e-07, |
|
"logits": -2.904877185821533, |
|
"logps": -260.0421142578125, |
|
"loss": -0.9792, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.566390595778787, |
|
"grad_norm": 17.837317168454984, |
|
"learning_rate": 2.5e-07, |
|
"logits": -3.0580921173095703, |
|
"logps": -253.33425903320312, |
|
"loss": -0.9783, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.5770772107934812, |
|
"grad_norm": 19.59229912598117, |
|
"learning_rate": 2.406522014309186e-07, |
|
"logits": -2.8213086128234863, |
|
"logps": -248.05899047851562, |
|
"loss": -0.9789, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.5877638258081752, |
|
"grad_norm": 19.31749257546076, |
|
"learning_rate": 2.3131747660339394e-07, |
|
"logits": -3.022545576095581, |
|
"logps": -261.65911865234375, |
|
"loss": -0.9795, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.5984504408228694, |
|
"grad_norm": 18.893405028741554, |
|
"learning_rate": 2.2200888097417302e-07, |
|
"logits": -2.8628716468811035, |
|
"logps": -276.71771240234375, |
|
"loss": -0.9787, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.6091370558375635, |
|
"grad_norm": 12.031725724871942, |
|
"learning_rate": 2.1273943345595635e-07, |
|
"logits": -2.8501031398773193, |
|
"logps": -225.07083129882812, |
|
"loss": -0.9802, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.6198236708522575, |
|
"grad_norm": 17.440618788817403, |
|
"learning_rate": 2.0352209820927135e-07, |
|
"logits": -2.9809391498565674, |
|
"logps": -279.31085205078125, |
|
"loss": -0.98, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.6305102858669517, |
|
"grad_norm": 17.533312297864594, |
|
"learning_rate": 1.9436976651092142e-07, |
|
"logits": -2.8259799480438232, |
|
"logps": -233.43637084960938, |
|
"loss": -0.9791, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.6411969008816457, |
|
"grad_norm": 19.09114013567347, |
|
"learning_rate": 1.8529523872436977e-07, |
|
"logits": -2.877838134765625, |
|
"logps": -250.38919067382812, |
|
"loss": -0.979, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.6518835158963399, |
|
"grad_norm": 14.469325828491328, |
|
"learning_rate": 1.763112063972739e-07, |
|
"logits": -2.9694724082946777, |
|
"logps": -265.880615234375, |
|
"loss": -0.9795, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.6625701309110339, |
|
"grad_norm": 13.892022212321942, |
|
"learning_rate": 1.674302345112083e-07, |
|
"logits": -2.8365535736083984, |
|
"logps": -249.2300567626953, |
|
"loss": -0.9815, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.673256745925728, |
|
"grad_norm": 16.882467097549785, |
|
"learning_rate": 1.5866474390840124e-07, |
|
"logits": -3.010043144226074, |
|
"logps": -297.1403503417969, |
|
"loss": -0.9803, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.6839433609404221, |
|
"grad_norm": 16.561775281344463, |
|
"learning_rate": 1.500269939200648e-07, |
|
"logits": -2.910022735595703, |
|
"logps": -245.9589080810547, |
|
"loss": -0.981, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.6946299759551162, |
|
"grad_norm": 11.956009192905249, |
|
"learning_rate": 1.4152906522061047e-07, |
|
"logits": -3.090505599975586, |
|
"logps": -297.404052734375, |
|
"loss": -0.9801, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.7053165909698104, |
|
"grad_norm": 20.558342213046632, |
|
"learning_rate": 1.3318284293173449e-07, |
|
"logits": -3.017181396484375, |
|
"logps": -282.6482238769531, |
|
"loss": -0.9807, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.7160032059845044, |
|
"grad_norm": 13.386951582952191, |
|
"learning_rate": 1.2500000000000005e-07, |
|
"logits": -2.884382963180542, |
|
"logps": -307.8055419921875, |
|
"loss": -0.982, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.7266898209991985, |
|
"grad_norm": 20.886776560670956, |
|
"learning_rate": 1.1699198087116588e-07, |
|
"logits": -2.919635057449341, |
|
"logps": -249.56100463867188, |
|
"loss": -0.9798, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.7373764360138926, |
|
"grad_norm": 18.641185325431465, |
|
"learning_rate": 1.0916998548409447e-07, |
|
"logits": -2.8742291927337646, |
|
"logps": -281.08062744140625, |
|
"loss": -0.9797, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.7480630510285867, |
|
"grad_norm": 12.516321612741994, |
|
"learning_rate": 1.0154495360662463e-07, |
|
"logits": -2.859714984893799, |
|
"logps": -246.27450561523438, |
|
"loss": -0.9812, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.7587496660432808, |
|
"grad_norm": 15.783450182742786, |
|
"learning_rate": 9.412754953531663e-08, |
|
"logits": -2.9711270332336426, |
|
"logps": -252.8061981201172, |
|
"loss": -0.9806, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.7694362810579749, |
|
"grad_norm": 16.203610890848484, |
|
"learning_rate": 8.692814718046978e-08, |
|
"logits": -2.942884683609009, |
|
"logps": -284.9074401855469, |
|
"loss": -0.9813, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.7801228960726689, |
|
"grad_norm": 15.344868690260943, |
|
"learning_rate": 7.99568155572701e-08, |
|
"logits": -2.9541680812835693, |
|
"logps": -254.73123168945312, |
|
"loss": -0.9793, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.7908095110873631, |
|
"grad_norm": 15.263546884716034, |
|
"learning_rate": 7.322330470336313e-08, |
|
"logits": -2.816589832305908, |
|
"logps": -258.56964111328125, |
|
"loss": -0.9813, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.8014961261020572, |
|
"grad_norm": 15.150826166868379, |
|
"learning_rate": 6.673703204254347e-08, |
|
"logits": -2.874589681625366, |
|
"logps": -264.58929443359375, |
|
"loss": -0.9818, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.8121827411167513, |
|
"grad_norm": 17.746820632954226, |
|
"learning_rate": 6.050706921363672e-08, |
|
"logits": -2.811462879180908, |
|
"logps": -274.1175842285156, |
|
"loss": -0.9821, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.8228693561314454, |
|
"grad_norm": 17.866909167465558, |
|
"learning_rate": 5.454212938299255e-08, |
|
"logits": -2.751802444458008, |
|
"logps": -265.1951904296875, |
|
"loss": -0.981, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.8335559711461394, |
|
"grad_norm": 19.146717420675806, |
|
"learning_rate": 4.885055505833291e-08, |
|
"logits": -2.872286319732666, |
|
"logps": -271.44451904296875, |
|
"loss": -0.9816, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.8442425861608336, |
|
"grad_norm": 16.977738077995557, |
|
"learning_rate": 4.3440306421001324e-08, |
|
"logits": -2.9000906944274902, |
|
"logps": -246.639892578125, |
|
"loss": -0.9818, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.8549292011755276, |
|
"grad_norm": 15.148125130733664, |
|
"learning_rate": 3.831895019292897e-08, |
|
"logits": -2.962857723236084, |
|
"logps": -276.59356689453125, |
|
"loss": -0.9812, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.8549292011755276, |
|
"eval_logits": -2.9494612216949463, |
|
"eval_logps": -291.1517028808594, |
|
"eval_loss": -0.9810261130332947, |
|
"eval_runtime": 560.2713, |
|
"eval_samples_per_second": 3.513, |
|
"eval_steps_per_second": 0.22, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.8656158161902218, |
|
"grad_norm": 17.51200256618434, |
|
"learning_rate": 3.349364905389032e-08, |
|
"logits": -2.8892321586608887, |
|
"logps": -291.7529296875, |
|
"loss": -0.9807, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.8763024312049158, |
|
"grad_norm": 22.832692473414674, |
|
"learning_rate": 2.8971151623847584e-08, |
|
"logits": -2.803345203399658, |
|
"logps": -268.4388122558594, |
|
"loss": -0.9811, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.88698904621961, |
|
"grad_norm": 17.595837977158578, |
|
"learning_rate": 2.475778302439524e-08, |
|
"logits": -2.8722121715545654, |
|
"logps": -251.7541961669922, |
|
"loss": -0.9823, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.897675661234304, |
|
"grad_norm": 18.312171813693368, |
|
"learning_rate": 2.085943603250595e-08, |
|
"logits": -2.9985263347625732, |
|
"logps": -270.0942077636719, |
|
"loss": -0.9802, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.9083622762489981, |
|
"grad_norm": 16.517443174199403, |
|
"learning_rate": 1.7281562838948966e-08, |
|
"logits": -2.8859758377075195, |
|
"logps": -265.142578125, |
|
"loss": -0.9812, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.9190488912636923, |
|
"grad_norm": 19.65350536281161, |
|
"learning_rate": 1.4029167422908105e-08, |
|
"logits": -2.95444393157959, |
|
"logps": -290.2696228027344, |
|
"loss": -0.9818, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.9297355062783863, |
|
"grad_norm": 15.855089650199325, |
|
"learning_rate": 1.1106798553464802e-08, |
|
"logits": -2.9999821186065674, |
|
"logps": -269.6681213378906, |
|
"loss": -0.9815, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.9404221212930804, |
|
"grad_norm": 17.34381390762727, |
|
"learning_rate": 8.518543427732949e-09, |
|
"logits": -2.989065647125244, |
|
"logps": -246.51412963867188, |
|
"loss": -0.9808, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.9511087363077745, |
|
"grad_norm": 16.16955244150207, |
|
"learning_rate": 6.268021954544095e-09, |
|
"logits": -2.859179973602295, |
|
"logps": -305.86053466796875, |
|
"loss": -0.9795, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.9617953513224686, |
|
"grad_norm": 14.813287522652269, |
|
"learning_rate": 4.358381691677931e-09, |
|
"logits": -2.835251808166504, |
|
"logps": -273.0709533691406, |
|
"loss": -0.9813, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.9724819663371627, |
|
"grad_norm": 18.12603626613458, |
|
"learning_rate": 2.7922934437178692e-09, |
|
"logits": -2.8190817832946777, |
|
"logps": -264.83544921875, |
|
"loss": -0.981, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.9831685813518568, |
|
"grad_norm": 17.624270425975272, |
|
"learning_rate": 1.5719475266893489e-09, |
|
"logits": -2.793806552886963, |
|
"logps": -224.26779174804688, |
|
"loss": -0.9821, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.9938551963665508, |
|
"grad_norm": 17.165545325823846, |
|
"learning_rate": 6.990507047049676e-10, |
|
"logits": -2.9348580837249756, |
|
"logps": -265.0174255371094, |
|
"loss": -0.9799, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.9981298423724285, |
|
"step": 467, |
|
"total_flos": 0.0, |
|
"train_loss": -0.9702747563733789, |
|
"train_runtime": 45631.769, |
|
"train_samples_per_second": 1.312, |
|
"train_steps_per_second": 0.01 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 467, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 250, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|