|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"global_step": 71754, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.9651587367951616e-05, |
|
"loss": 2.6727, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.930317473590323e-05, |
|
"loss": 2.2524, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.895476210385484e-05, |
|
"loss": 2.1243, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.860634947180645e-05, |
|
"loss": 2.0282, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.825793683975807e-05, |
|
"loss": 1.9537, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.7909524207709675e-05, |
|
"loss": 1.9412, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.756111157566129e-05, |
|
"loss": 1.8912, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.72126989436129e-05, |
|
"loss": 1.8241, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.6864286311564515e-05, |
|
"loss": 1.7936, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.651587367951613e-05, |
|
"loss": 1.7729, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.616746104746774e-05, |
|
"loss": 1.7618, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.5819048415419355e-05, |
|
"loss": 1.7382, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.547063578337096e-05, |
|
"loss": 1.7277, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.512222315132258e-05, |
|
"loss": 1.6827, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.477381051927419e-05, |
|
"loss": 1.6632, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.44253978872258e-05, |
|
"loss": 1.651, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.4076985255177414e-05, |
|
"loss": 1.6343, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.372857262312903e-05, |
|
"loss": 1.6402, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.338015999108064e-05, |
|
"loss": 1.6284, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.303174735903225e-05, |
|
"loss": 1.6198, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.268333472698387e-05, |
|
"loss": 1.5991, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.2334922094935473e-05, |
|
"loss": 1.5921, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.198650946288709e-05, |
|
"loss": 1.5648, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.16380968308387e-05, |
|
"loss": 1.5915, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.128968419879031e-05, |
|
"loss": 1.5482, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.0941271566741926e-05, |
|
"loss": 1.5351, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.059285893469354e-05, |
|
"loss": 1.5637, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.024444630264515e-05, |
|
"loss": 1.5312, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.989603367059676e-05, |
|
"loss": 1.5249, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.954762103854838e-05, |
|
"loss": 1.5168, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.9199208406499986e-05, |
|
"loss": 1.4931, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.88507957744516e-05, |
|
"loss": 1.5159, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.850238314240321e-05, |
|
"loss": 1.4926, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.8153970510354826e-05, |
|
"loss": 1.5232, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.780555787830644e-05, |
|
"loss": 1.4861, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.745714524625805e-05, |
|
"loss": 1.4828, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.7108732614209665e-05, |
|
"loss": 1.4798, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.676031998216127e-05, |
|
"loss": 1.4484, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.641190735011289e-05, |
|
"loss": 1.4321, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.60634947180645e-05, |
|
"loss": 1.4593, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.571508208601611e-05, |
|
"loss": 1.4524, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.5366669453967725e-05, |
|
"loss": 1.4315, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.501825682191934e-05, |
|
"loss": 1.432, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.466984418987095e-05, |
|
"loss": 1.4086, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 3.4321431557822564e-05, |
|
"loss": 1.4256, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.397301892577418e-05, |
|
"loss": 1.4312, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.3624606293725784e-05, |
|
"loss": 1.429, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.3276193661677404e-05, |
|
"loss": 1.3781, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 3.292778102962901e-05, |
|
"loss": 1.2389, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 3.2579368397580624e-05, |
|
"loss": 1.2274, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 3.223095576553224e-05, |
|
"loss": 1.248, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 3.188254313348385e-05, |
|
"loss": 1.2383, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 3.1534130501435464e-05, |
|
"loss": 1.2561, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.118571786938708e-05, |
|
"loss": 1.2113, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 3.083730523733869e-05, |
|
"loss": 1.2452, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.0488892605290297e-05, |
|
"loss": 1.2477, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.014047997324191e-05, |
|
"loss": 1.2327, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 2.9792067341193523e-05, |
|
"loss": 1.2308, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.9443654709145136e-05, |
|
"loss": 1.2349, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 2.9095242077096746e-05, |
|
"loss": 1.2426, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 2.8746829445048363e-05, |
|
"loss": 1.2198, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 2.8398416812999973e-05, |
|
"loss": 1.2329, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 2.8050004180951583e-05, |
|
"loss": 1.2437, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 2.77015915489032e-05, |
|
"loss": 1.24, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 2.735317891685481e-05, |
|
"loss": 1.2085, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 2.7004766284806422e-05, |
|
"loss": 1.2179, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 2.6656353652758035e-05, |
|
"loss": 1.2073, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 2.630794102070965e-05, |
|
"loss": 1.2168, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 2.595952838866126e-05, |
|
"loss": 1.2198, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 2.5611115756612875e-05, |
|
"loss": 1.2073, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 2.5262703124564485e-05, |
|
"loss": 1.2348, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 2.49142904925161e-05, |
|
"loss": 1.2014, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 2.4565877860467708e-05, |
|
"loss": 1.2108, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 2.421746522841932e-05, |
|
"loss": 1.195, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 2.3869052596370935e-05, |
|
"loss": 1.1997, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 2.3520639964322548e-05, |
|
"loss": 1.1957, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 2.317222733227416e-05, |
|
"loss": 1.1704, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 2.2823814700225774e-05, |
|
"loss": 1.1826, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 2.2475402068177384e-05, |
|
"loss": 1.1957, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 2.2126989436128997e-05, |
|
"loss": 1.1773, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 2.177857680408061e-05, |
|
"loss": 1.1977, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.143016417203222e-05, |
|
"loss": 1.1744, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.1081751539983834e-05, |
|
"loss": 1.1794, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.0733338907935447e-05, |
|
"loss": 1.1686, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.038492627588706e-05, |
|
"loss": 1.1729, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.0036513643838674e-05, |
|
"loss": 1.1693, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.9688101011790283e-05, |
|
"loss": 1.1593, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.9339688379741897e-05, |
|
"loss": 1.1629, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.899127574769351e-05, |
|
"loss": 1.1769, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.864286311564512e-05, |
|
"loss": 1.142, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.8294450483596733e-05, |
|
"loss": 1.1707, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.7946037851548346e-05, |
|
"loss": 1.1564, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.759762521949996e-05, |
|
"loss": 1.1419, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.7249212587451573e-05, |
|
"loss": 1.1432, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.6900799955403186e-05, |
|
"loss": 1.1542, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 1.6552387323354796e-05, |
|
"loss": 1.0809, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.620397469130641e-05, |
|
"loss": 1.001, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 1.5855562059258022e-05, |
|
"loss": 0.9794, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.5507149427209632e-05, |
|
"loss": 0.9856, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.5158736795161247e-05, |
|
"loss": 0.9815, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.481032416311286e-05, |
|
"loss": 0.9833, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.446191153106447e-05, |
|
"loss": 0.9746, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.4113498899016083e-05, |
|
"loss": 0.9953, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.3765086266967697e-05, |
|
"loss": 0.9918, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.3416673634919308e-05, |
|
"loss": 0.9825, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.3068261002870921e-05, |
|
"loss": 0.9986, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.2719848370822531e-05, |
|
"loss": 0.9579, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.2371435738774145e-05, |
|
"loss": 0.9845, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.202302310672576e-05, |
|
"loss": 0.9711, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.1674610474677371e-05, |
|
"loss": 0.9944, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.1326197842628983e-05, |
|
"loss": 0.9784, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.0977785210580594e-05, |
|
"loss": 0.9882, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.0629372578532207e-05, |
|
"loss": 0.9813, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.028095994648382e-05, |
|
"loss": 0.9667, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 9.932547314435432e-06, |
|
"loss": 0.9585, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.584134682387045e-06, |
|
"loss": 0.9685, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 9.235722050338657e-06, |
|
"loss": 0.9943, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 8.88730941829027e-06, |
|
"loss": 0.962, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 8.538896786241882e-06, |
|
"loss": 0.9667, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 8.190484154193495e-06, |
|
"loss": 0.9758, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.842071522145107e-06, |
|
"loss": 0.9683, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.493658890096719e-06, |
|
"loss": 0.9813, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.145246258048332e-06, |
|
"loss": 0.9577, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 6.796833625999945e-06, |
|
"loss": 0.9577, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 6.448420993951557e-06, |
|
"loss": 0.9791, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 6.100008361903169e-06, |
|
"loss": 0.9675, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 5.751595729854782e-06, |
|
"loss": 0.9594, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 5.403183097806394e-06, |
|
"loss": 0.968, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 5.054770465758007e-06, |
|
"loss": 0.978, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.70635783370962e-06, |
|
"loss": 0.9623, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.357945201661231e-06, |
|
"loss": 0.9446, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 4.009532569612845e-06, |
|
"loss": 0.9673, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.661119937564456e-06, |
|
"loss": 0.9528, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.312707305516069e-06, |
|
"loss": 0.9548, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 2.9642946734676814e-06, |
|
"loss": 0.9608, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.615882041419294e-06, |
|
"loss": 0.951, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.267469409370906e-06, |
|
"loss": 0.9476, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.919056777322519e-06, |
|
"loss": 0.9765, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.5706441452741312e-06, |
|
"loss": 0.9551, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.2222315132257436e-06, |
|
"loss": 0.954, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 8.73818881177356e-07, |
|
"loss": 0.9611, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 5.254062491289685e-07, |
|
"loss": 0.9548, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.7699361708058088e-07, |
|
"loss": 0.9542, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 71754, |
|
"total_flos": 9.831206834203853e+16, |
|
"train_loss": 1.271412817926867, |
|
"train_runtime": 18699.5043, |
|
"train_samples_per_second": 38.372, |
|
"train_steps_per_second": 3.837 |
|
} |
|
], |
|
"max_steps": 71754, |
|
"num_train_epochs": 3, |
|
"total_flos": 9.831206834203853e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|