|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9992627698788836, |
|
"eval_steps": 10000000000000, |
|
"global_step": 1186, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": "8.3333e-06", |
|
"loss": 0.2161, |
|
"slid_loss": 0.2161, |
|
"step": 1, |
|
"time": 63.57 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": "1.6667e-05", |
|
"loss": 0.2083, |
|
"slid_loss": 0.2122, |
|
"step": 2, |
|
"time": 44.72 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": "2.5000e-05", |
|
"loss": 0.2154, |
|
"slid_loss": 0.2133, |
|
"step": 3, |
|
"time": 46.86 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": "3.3333e-05", |
|
"loss": 0.1934, |
|
"slid_loss": 0.2083, |
|
"step": 4, |
|
"time": 46.36 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": "4.1667e-05", |
|
"loss": 0.1994, |
|
"slid_loss": 0.2065, |
|
"step": 5, |
|
"time": 45.98 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "5.0000e-05", |
|
"loss": 0.187, |
|
"slid_loss": 0.2033, |
|
"step": 6, |
|
"time": 50.17 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "5.8333e-05", |
|
"loss": 0.1959, |
|
"slid_loss": 0.2022, |
|
"step": 7, |
|
"time": 49.06 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "6.6667e-05", |
|
"loss": 0.1983, |
|
"slid_loss": 0.2017, |
|
"step": 8, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "7.5000e-05", |
|
"loss": 0.1768, |
|
"slid_loss": 0.1989, |
|
"step": 9, |
|
"time": 46.75 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "8.3333e-05", |
|
"loss": 0.1891, |
|
"slid_loss": 0.198, |
|
"step": 10, |
|
"time": 47.51 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "9.1667e-05", |
|
"loss": 0.176, |
|
"slid_loss": 0.196, |
|
"step": 11, |
|
"time": 47.4 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "1.0000e-04", |
|
"loss": 0.172, |
|
"slid_loss": 0.194, |
|
"step": 12, |
|
"time": 48.28 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "1.0000e-04", |
|
"loss": 0.1811, |
|
"slid_loss": 0.193, |
|
"step": 13, |
|
"time": 45.98 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "9.9999e-05", |
|
"loss": 0.1826, |
|
"slid_loss": 0.1922, |
|
"step": 14, |
|
"time": 49.9 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "9.9999e-05", |
|
"loss": 0.1658, |
|
"slid_loss": 0.1905, |
|
"step": 15, |
|
"time": 48.77 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "9.9998e-05", |
|
"loss": 0.1667, |
|
"slid_loss": 0.189, |
|
"step": 16, |
|
"time": 44.99 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "9.9996e-05", |
|
"loss": 0.1641, |
|
"slid_loss": 0.1875, |
|
"step": 17, |
|
"time": 46.92 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9995e-05", |
|
"loss": 0.1717, |
|
"slid_loss": 0.1867, |
|
"step": 18, |
|
"time": 47.76 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9993e-05", |
|
"loss": 0.1579, |
|
"slid_loss": 0.1851, |
|
"step": 19, |
|
"time": 46.11 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9991e-05", |
|
"loss": 0.1678, |
|
"slid_loss": 0.1843, |
|
"step": 20, |
|
"time": 47.59 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9988e-05", |
|
"loss": 0.1666, |
|
"slid_loss": 0.1834, |
|
"step": 21, |
|
"time": 46.51 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9986e-05", |
|
"loss": 0.1689, |
|
"slid_loss": 0.1828, |
|
"step": 22, |
|
"time": 47.13 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9983e-05", |
|
"loss": 0.1634, |
|
"slid_loss": 0.1819, |
|
"step": 23, |
|
"time": 46.48 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9979e-05", |
|
"loss": 0.1587, |
|
"slid_loss": 0.181, |
|
"step": 24, |
|
"time": 50.81 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9976e-05", |
|
"loss": 0.1598, |
|
"slid_loss": 0.1801, |
|
"step": 25, |
|
"time": 49.95 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9972e-05", |
|
"loss": 0.1741, |
|
"slid_loss": 0.1799, |
|
"step": 26, |
|
"time": 48.44 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9968e-05", |
|
"loss": 0.1595, |
|
"slid_loss": 0.1791, |
|
"step": 27, |
|
"time": 45.33 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9963e-05", |
|
"loss": 0.1633, |
|
"slid_loss": 0.1786, |
|
"step": 28, |
|
"time": 48.15 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "9.9959e-05", |
|
"loss": 0.1643, |
|
"slid_loss": 0.1781, |
|
"step": 29, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9954e-05", |
|
"loss": 0.1596, |
|
"slid_loss": 0.1775, |
|
"step": 30, |
|
"time": 47.56 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9948e-05", |
|
"loss": 0.1653, |
|
"slid_loss": 0.1771, |
|
"step": 31, |
|
"time": 47.08 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9943e-05", |
|
"loss": 0.1521, |
|
"slid_loss": 0.1763, |
|
"step": 32, |
|
"time": 45.78 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9937e-05", |
|
"loss": 0.1595, |
|
"slid_loss": 0.1758, |
|
"step": 33, |
|
"time": 49.14 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9931e-05", |
|
"loss": 0.1543, |
|
"slid_loss": 0.1751, |
|
"step": 34, |
|
"time": 46.07 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9924e-05", |
|
"loss": 0.1569, |
|
"slid_loss": 0.1746, |
|
"step": 35, |
|
"time": 46.77 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9918e-05", |
|
"loss": 0.139, |
|
"slid_loss": 0.1736, |
|
"step": 36, |
|
"time": 45.64 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9911e-05", |
|
"loss": 0.1632, |
|
"slid_loss": 0.1733, |
|
"step": 37, |
|
"time": 47.39 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9903e-05", |
|
"loss": 0.1501, |
|
"slid_loss": 0.1727, |
|
"step": 38, |
|
"time": 47.17 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9896e-05", |
|
"loss": 0.1521, |
|
"slid_loss": 0.1722, |
|
"step": 39, |
|
"time": 45.96 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9888e-05", |
|
"loss": 0.149, |
|
"slid_loss": 0.1716, |
|
"step": 40, |
|
"time": 47.63 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "9.9880e-05", |
|
"loss": 0.1515, |
|
"slid_loss": 0.1711, |
|
"step": 41, |
|
"time": 45.83 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9871e-05", |
|
"loss": 0.1471, |
|
"slid_loss": 0.1706, |
|
"step": 42, |
|
"time": 49.04 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9862e-05", |
|
"loss": 0.1536, |
|
"slid_loss": 0.1702, |
|
"step": 43, |
|
"time": 46.63 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9853e-05", |
|
"loss": 0.1485, |
|
"slid_loss": 0.1697, |
|
"step": 44, |
|
"time": 48.86 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9844e-05", |
|
"loss": 0.1496, |
|
"slid_loss": 0.1692, |
|
"step": 45, |
|
"time": 47.03 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9835e-05", |
|
"loss": 0.1519, |
|
"slid_loss": 0.1689, |
|
"step": 46, |
|
"time": 49.21 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9825e-05", |
|
"loss": 0.146, |
|
"slid_loss": 0.1684, |
|
"step": 47, |
|
"time": 45.37 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9815e-05", |
|
"loss": 0.1516, |
|
"slid_loss": 0.168, |
|
"step": 48, |
|
"time": 46.19 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9804e-05", |
|
"loss": 0.1475, |
|
"slid_loss": 0.1676, |
|
"step": 49, |
|
"time": 45.62 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9793e-05", |
|
"loss": 0.1509, |
|
"slid_loss": 0.1673, |
|
"step": 50, |
|
"time": 46.11 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9782e-05", |
|
"loss": 0.1513, |
|
"slid_loss": 0.167, |
|
"step": 51, |
|
"time": 44.94 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9771e-05", |
|
"loss": 0.1425, |
|
"slid_loss": 0.1665, |
|
"step": 52, |
|
"time": 46.8 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "9.9759e-05", |
|
"loss": 0.1461, |
|
"slid_loss": 0.1661, |
|
"step": 53, |
|
"time": 47.33 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9748e-05", |
|
"loss": 0.1457, |
|
"slid_loss": 0.1657, |
|
"step": 54, |
|
"time": 45.88 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9735e-05", |
|
"loss": 0.1503, |
|
"slid_loss": 0.1654, |
|
"step": 55, |
|
"time": 47.22 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9723e-05", |
|
"loss": 0.1525, |
|
"slid_loss": 0.1652, |
|
"step": 56, |
|
"time": 46.7 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9710e-05", |
|
"loss": 0.1466, |
|
"slid_loss": 0.1649, |
|
"step": 57, |
|
"time": 46.63 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9697e-05", |
|
"loss": 0.1464, |
|
"slid_loss": 0.1646, |
|
"step": 58, |
|
"time": 46.62 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9684e-05", |
|
"loss": 0.1434, |
|
"slid_loss": 0.1642, |
|
"step": 59, |
|
"time": 46.47 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9670e-05", |
|
"loss": 0.1478, |
|
"slid_loss": 0.1639, |
|
"step": 60, |
|
"time": 49.29 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9657e-05", |
|
"loss": 0.1542, |
|
"slid_loss": 0.1638, |
|
"step": 61, |
|
"time": 47.25 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9642e-05", |
|
"loss": 0.1569, |
|
"slid_loss": 0.1637, |
|
"step": 62, |
|
"time": 46.6 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9628e-05", |
|
"loss": 0.1486, |
|
"slid_loss": 0.1634, |
|
"step": 63, |
|
"time": 47.97 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9613e-05", |
|
"loss": 0.1481, |
|
"slid_loss": 0.1632, |
|
"step": 64, |
|
"time": 46.52 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.9598e-05", |
|
"loss": 0.1501, |
|
"slid_loss": 0.163, |
|
"step": 65, |
|
"time": 46.88 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9583e-05", |
|
"loss": 0.1522, |
|
"slid_loss": 0.1628, |
|
"step": 66, |
|
"time": 46.17 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9568e-05", |
|
"loss": 0.1539, |
|
"slid_loss": 0.1627, |
|
"step": 67, |
|
"time": 46.06 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9552e-05", |
|
"loss": 0.1496, |
|
"slid_loss": 0.1625, |
|
"step": 68, |
|
"time": 47.13 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9536e-05", |
|
"loss": 0.1535, |
|
"slid_loss": 0.1624, |
|
"step": 69, |
|
"time": 45.6 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9519e-05", |
|
"loss": 0.1493, |
|
"slid_loss": 0.1622, |
|
"step": 70, |
|
"time": 47.53 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9502e-05", |
|
"loss": 0.146, |
|
"slid_loss": 0.1619, |
|
"step": 71, |
|
"time": 47.76 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9486e-05", |
|
"loss": 0.1372, |
|
"slid_loss": 0.1616, |
|
"step": 72, |
|
"time": 47.97 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9468e-05", |
|
"loss": 0.1411, |
|
"slid_loss": 0.1613, |
|
"step": 73, |
|
"time": 46.46 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9451e-05", |
|
"loss": 0.1403, |
|
"slid_loss": 0.161, |
|
"step": 74, |
|
"time": 45.97 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9433e-05", |
|
"loss": 0.1385, |
|
"slid_loss": 0.1607, |
|
"step": 75, |
|
"time": 47.65 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9415e-05", |
|
"loss": 0.1463, |
|
"slid_loss": 0.1605, |
|
"step": 76, |
|
"time": 45.63 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9396e-05", |
|
"loss": 0.1549, |
|
"slid_loss": 0.1605, |
|
"step": 77, |
|
"time": 47.19 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9378e-05", |
|
"loss": 0.1484, |
|
"slid_loss": 0.1603, |
|
"step": 78, |
|
"time": 48.92 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9359e-05", |
|
"loss": 0.1446, |
|
"slid_loss": 0.1601, |
|
"step": 79, |
|
"time": 46.14 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9340e-05", |
|
"loss": 0.1403, |
|
"slid_loss": 0.1599, |
|
"step": 80, |
|
"time": 45.0 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9320e-05", |
|
"loss": 0.1444, |
|
"slid_loss": 0.1597, |
|
"step": 81, |
|
"time": 46.98 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9300e-05", |
|
"loss": 0.1492, |
|
"slid_loss": 0.1596, |
|
"step": 82, |
|
"time": 45.89 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9280e-05", |
|
"loss": 0.14, |
|
"slid_loss": 0.1593, |
|
"step": 83, |
|
"time": 46.72 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9260e-05", |
|
"loss": 0.1398, |
|
"slid_loss": 0.1591, |
|
"step": 84, |
|
"time": 44.65 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9239e-05", |
|
"loss": 0.1457, |
|
"slid_loss": 0.1589, |
|
"step": 85, |
|
"time": 45.24 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9218e-05", |
|
"loss": 0.1442, |
|
"slid_loss": 0.1588, |
|
"step": 86, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9197e-05", |
|
"loss": 0.147, |
|
"slid_loss": 0.1586, |
|
"step": 87, |
|
"time": 47.13 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9176e-05", |
|
"loss": 0.1385, |
|
"slid_loss": 0.1584, |
|
"step": 88, |
|
"time": 45.66 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9154e-05", |
|
"loss": 0.1359, |
|
"slid_loss": 0.1581, |
|
"step": 89, |
|
"time": 46.29 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9132e-05", |
|
"loss": 0.1498, |
|
"slid_loss": 0.158, |
|
"step": 90, |
|
"time": 45.65 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9110e-05", |
|
"loss": 0.145, |
|
"slid_loss": 0.1579, |
|
"step": 91, |
|
"time": 46.64 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9087e-05", |
|
"loss": 0.1426, |
|
"slid_loss": 0.1577, |
|
"step": 92, |
|
"time": 47.69 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9064e-05", |
|
"loss": 0.143, |
|
"slid_loss": 0.1576, |
|
"step": 93, |
|
"time": 50.06 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9041e-05", |
|
"loss": 0.1417, |
|
"slid_loss": 0.1574, |
|
"step": 94, |
|
"time": 47.81 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9017e-05", |
|
"loss": 0.1503, |
|
"slid_loss": 0.1573, |
|
"step": 95, |
|
"time": 46.11 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.8994e-05", |
|
"loss": 0.1413, |
|
"slid_loss": 0.1572, |
|
"step": 96, |
|
"time": 45.96 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.8970e-05", |
|
"loss": 0.1367, |
|
"slid_loss": 0.157, |
|
"step": 97, |
|
"time": 47.16 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.8945e-05", |
|
"loss": 0.1431, |
|
"slid_loss": 0.1568, |
|
"step": 98, |
|
"time": 47.64 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.8921e-05", |
|
"loss": 0.1434, |
|
"slid_loss": 0.1567, |
|
"step": 99, |
|
"time": 45.85 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.8896e-05", |
|
"loss": 0.1463, |
|
"slid_loss": 0.1566, |
|
"step": 100, |
|
"time": 46.13 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8871e-05", |
|
"loss": 0.1357, |
|
"slid_loss": 0.1558, |
|
"step": 101, |
|
"time": 46.45 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8846e-05", |
|
"loss": 0.1451, |
|
"slid_loss": 0.1551, |
|
"step": 102, |
|
"time": 47.18 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8820e-05", |
|
"loss": 0.1489, |
|
"slid_loss": 0.1545, |
|
"step": 103, |
|
"time": 48.45 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8794e-05", |
|
"loss": 0.1416, |
|
"slid_loss": 0.154, |
|
"step": 104, |
|
"time": 47.46 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8768e-05", |
|
"loss": 0.1428, |
|
"slid_loss": 0.1534, |
|
"step": 105, |
|
"time": 47.11 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8741e-05", |
|
"loss": 0.1397, |
|
"slid_loss": 0.1529, |
|
"step": 106, |
|
"time": 46.69 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8714e-05", |
|
"loss": 0.1459, |
|
"slid_loss": 0.1524, |
|
"step": 107, |
|
"time": 45.91 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8687e-05", |
|
"loss": 0.1504, |
|
"slid_loss": 0.1519, |
|
"step": 108, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8660e-05", |
|
"loss": 0.147, |
|
"slid_loss": 0.1516, |
|
"step": 109, |
|
"time": 46.61 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8632e-05", |
|
"loss": 0.1346, |
|
"slid_loss": 0.1511, |
|
"step": 110, |
|
"time": 46.59 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8605e-05", |
|
"loss": 0.1442, |
|
"slid_loss": 0.1508, |
|
"step": 111, |
|
"time": 45.14 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.8576e-05", |
|
"loss": 0.1427, |
|
"slid_loss": 0.1505, |
|
"step": 112, |
|
"time": 48.58 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8548e-05", |
|
"loss": 0.1339, |
|
"slid_loss": 0.15, |
|
"step": 113, |
|
"time": 46.61 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8519e-05", |
|
"loss": 0.1427, |
|
"slid_loss": 0.1496, |
|
"step": 114, |
|
"time": 45.24 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8490e-05", |
|
"loss": 0.1372, |
|
"slid_loss": 0.1493, |
|
"step": 115, |
|
"time": 46.51 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8461e-05", |
|
"loss": 0.147, |
|
"slid_loss": 0.1491, |
|
"step": 116, |
|
"time": 48.66 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8431e-05", |
|
"loss": 0.1475, |
|
"slid_loss": 0.149, |
|
"step": 117, |
|
"time": 45.56 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8402e-05", |
|
"loss": 0.14, |
|
"slid_loss": 0.1486, |
|
"step": 118, |
|
"time": 45.94 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8371e-05", |
|
"loss": 0.1438, |
|
"slid_loss": 0.1485, |
|
"step": 119, |
|
"time": 45.77 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8341e-05", |
|
"loss": 0.1394, |
|
"slid_loss": 0.1482, |
|
"step": 120, |
|
"time": 45.5 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8310e-05", |
|
"loss": 0.1455, |
|
"slid_loss": 0.148, |
|
"step": 121, |
|
"time": 48.74 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8280e-05", |
|
"loss": 0.1388, |
|
"slid_loss": 0.1477, |
|
"step": 122, |
|
"time": 46.07 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8248e-05", |
|
"loss": 0.1398, |
|
"slid_loss": 0.1475, |
|
"step": 123, |
|
"time": 45.6 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.8217e-05", |
|
"loss": 0.1472, |
|
"slid_loss": 0.1474, |
|
"step": 124, |
|
"time": 47.77 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.8185e-05", |
|
"loss": 0.1488, |
|
"slid_loss": 0.1472, |
|
"step": 125, |
|
"time": 45.48 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.8153e-05", |
|
"loss": 0.143, |
|
"slid_loss": 0.1469, |
|
"step": 126, |
|
"time": 48.73 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.8121e-05", |
|
"loss": 0.1421, |
|
"slid_loss": 0.1468, |
|
"step": 127, |
|
"time": 45.63 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.8088e-05", |
|
"loss": 0.1357, |
|
"slid_loss": 0.1465, |
|
"step": 128, |
|
"time": 47.22 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.8055e-05", |
|
"loss": 0.1416, |
|
"slid_loss": 0.1463, |
|
"step": 129, |
|
"time": 46.94 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.8022e-05", |
|
"loss": 0.1332, |
|
"slid_loss": 0.146, |
|
"step": 130, |
|
"time": 47.46 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.7989e-05", |
|
"loss": 0.1464, |
|
"slid_loss": 0.1458, |
|
"step": 131, |
|
"time": 46.41 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.7955e-05", |
|
"loss": 0.1465, |
|
"slid_loss": 0.1458, |
|
"step": 132, |
|
"time": 46.8 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.7921e-05", |
|
"loss": 0.1473, |
|
"slid_loss": 0.1456, |
|
"step": 133, |
|
"time": 48.26 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.7887e-05", |
|
"loss": 0.1481, |
|
"slid_loss": 0.1456, |
|
"step": 134, |
|
"time": 49.79 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.7853e-05", |
|
"loss": 0.1391, |
|
"slid_loss": 0.1454, |
|
"step": 135, |
|
"time": 46.83 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.7818e-05", |
|
"loss": 0.1474, |
|
"slid_loss": 0.1455, |
|
"step": 136, |
|
"time": 46.18 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7783e-05", |
|
"loss": 0.1435, |
|
"slid_loss": 0.1453, |
|
"step": 137, |
|
"time": 45.84 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7748e-05", |
|
"loss": 0.1387, |
|
"slid_loss": 0.1452, |
|
"step": 138, |
|
"time": 44.86 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7712e-05", |
|
"loss": 0.1458, |
|
"slid_loss": 0.1451, |
|
"step": 139, |
|
"time": 45.62 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7676e-05", |
|
"loss": 0.1474, |
|
"slid_loss": 0.1451, |
|
"step": 140, |
|
"time": 46.06 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7640e-05", |
|
"loss": 0.1358, |
|
"slid_loss": 0.1449, |
|
"step": 141, |
|
"time": 45.46 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7604e-05", |
|
"loss": 0.1416, |
|
"slid_loss": 0.1449, |
|
"step": 142, |
|
"time": 47.35 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7567e-05", |
|
"loss": 0.1276, |
|
"slid_loss": 0.1446, |
|
"step": 143, |
|
"time": 47.16 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7530e-05", |
|
"loss": 0.1344, |
|
"slid_loss": 0.1445, |
|
"step": 144, |
|
"time": 45.11 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7493e-05", |
|
"loss": 0.1369, |
|
"slid_loss": 0.1443, |
|
"step": 145, |
|
"time": 46.5 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7456e-05", |
|
"loss": 0.1415, |
|
"slid_loss": 0.1442, |
|
"step": 146, |
|
"time": 45.43 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7418e-05", |
|
"loss": 0.1359, |
|
"slid_loss": 0.1441, |
|
"step": 147, |
|
"time": 48.09 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.7380e-05", |
|
"loss": 0.1359, |
|
"slid_loss": 0.144, |
|
"step": 148, |
|
"time": 47.34 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.7342e-05", |
|
"loss": 0.1315, |
|
"slid_loss": 0.1438, |
|
"step": 149, |
|
"time": 49.79 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.7303e-05", |
|
"loss": 0.1404, |
|
"slid_loss": 0.1437, |
|
"step": 150, |
|
"time": 46.45 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.7265e-05", |
|
"loss": 0.1402, |
|
"slid_loss": 0.1436, |
|
"step": 151, |
|
"time": 46.96 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.7226e-05", |
|
"loss": 0.1397, |
|
"slid_loss": 0.1436, |
|
"step": 152, |
|
"time": 46.84 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.7186e-05", |
|
"loss": 0.1522, |
|
"slid_loss": 0.1436, |
|
"step": 153, |
|
"time": 46.07 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.7147e-05", |
|
"loss": 0.1413, |
|
"slid_loss": 0.1436, |
|
"step": 154, |
|
"time": 45.52 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.7107e-05", |
|
"loss": 0.1399, |
|
"slid_loss": 0.1435, |
|
"step": 155, |
|
"time": 46.48 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.7067e-05", |
|
"loss": 0.1346, |
|
"slid_loss": 0.1433, |
|
"step": 156, |
|
"time": 47.82 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.7026e-05", |
|
"loss": 0.1414, |
|
"slid_loss": 0.1433, |
|
"step": 157, |
|
"time": 44.85 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.6986e-05", |
|
"loss": 0.1399, |
|
"slid_loss": 0.1432, |
|
"step": 158, |
|
"time": 46.29 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.6945e-05", |
|
"loss": 0.1374, |
|
"slid_loss": 0.1431, |
|
"step": 159, |
|
"time": 47.2 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.6904e-05", |
|
"loss": 0.1385, |
|
"slid_loss": 0.143, |
|
"step": 160, |
|
"time": 47.41 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6862e-05", |
|
"loss": 0.1434, |
|
"slid_loss": 0.1429, |
|
"step": 161, |
|
"time": 46.89 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6821e-05", |
|
"loss": 0.1402, |
|
"slid_loss": 0.1428, |
|
"step": 162, |
|
"time": 45.75 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6779e-05", |
|
"loss": 0.1412, |
|
"slid_loss": 0.1427, |
|
"step": 163, |
|
"time": 45.58 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6736e-05", |
|
"loss": 0.1435, |
|
"slid_loss": 0.1426, |
|
"step": 164, |
|
"time": 48.24 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6694e-05", |
|
"loss": 0.1362, |
|
"slid_loss": 0.1425, |
|
"step": 165, |
|
"time": 47.07 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6651e-05", |
|
"loss": 0.1415, |
|
"slid_loss": 0.1424, |
|
"step": 166, |
|
"time": 46.34 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6608e-05", |
|
"loss": 0.1412, |
|
"slid_loss": 0.1423, |
|
"step": 167, |
|
"time": 47.21 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6565e-05", |
|
"loss": 0.1383, |
|
"slid_loss": 0.1422, |
|
"step": 168, |
|
"time": 45.14 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6521e-05", |
|
"loss": 0.1257, |
|
"slid_loss": 0.1419, |
|
"step": 169, |
|
"time": 47.43 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6478e-05", |
|
"loss": 0.1301, |
|
"slid_loss": 0.1417, |
|
"step": 170, |
|
"time": 47.65 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6434e-05", |
|
"loss": 0.1454, |
|
"slid_loss": 0.1417, |
|
"step": 171, |
|
"time": 47.82 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.6389e-05", |
|
"loss": 0.1501, |
|
"slid_loss": 0.1418, |
|
"step": 172, |
|
"time": 45.79 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.6345e-05", |
|
"loss": 0.1367, |
|
"slid_loss": 0.1418, |
|
"step": 173, |
|
"time": 47.27 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.6300e-05", |
|
"loss": 0.1427, |
|
"slid_loss": 0.1418, |
|
"step": 174, |
|
"time": 48.48 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.6255e-05", |
|
"loss": 0.1379, |
|
"slid_loss": 0.1418, |
|
"step": 175, |
|
"time": 44.65 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.6209e-05", |
|
"loss": 0.1382, |
|
"slid_loss": 0.1417, |
|
"step": 176, |
|
"time": 45.96 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.6164e-05", |
|
"loss": 0.1371, |
|
"slid_loss": 0.1415, |
|
"step": 177, |
|
"time": 48.49 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.6118e-05", |
|
"loss": 0.1429, |
|
"slid_loss": 0.1415, |
|
"step": 178, |
|
"time": 46.0 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.6072e-05", |
|
"loss": 0.1495, |
|
"slid_loss": 0.1415, |
|
"step": 179, |
|
"time": 46.39 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.6025e-05", |
|
"loss": 0.1334, |
|
"slid_loss": 0.1415, |
|
"step": 180, |
|
"time": 46.35 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.5979e-05", |
|
"loss": 0.1394, |
|
"slid_loss": 0.1414, |
|
"step": 181, |
|
"time": 46.39 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.5932e-05", |
|
"loss": 0.1337, |
|
"slid_loss": 0.1413, |
|
"step": 182, |
|
"time": 45.25 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.5885e-05", |
|
"loss": 0.1342, |
|
"slid_loss": 0.1412, |
|
"step": 183, |
|
"time": 47.43 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5837e-05", |
|
"loss": 0.1356, |
|
"slid_loss": 0.1412, |
|
"step": 184, |
|
"time": 47.52 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5790e-05", |
|
"loss": 0.1373, |
|
"slid_loss": 0.1411, |
|
"step": 185, |
|
"time": 45.52 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5742e-05", |
|
"loss": 0.1377, |
|
"slid_loss": 0.141, |
|
"step": 186, |
|
"time": 45.8 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5694e-05", |
|
"loss": 0.1407, |
|
"slid_loss": 0.1409, |
|
"step": 187, |
|
"time": 49.12 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5645e-05", |
|
"loss": 0.1379, |
|
"slid_loss": 0.1409, |
|
"step": 188, |
|
"time": 46.77 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5596e-05", |
|
"loss": 0.1448, |
|
"slid_loss": 0.141, |
|
"step": 189, |
|
"time": 46.49 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5547e-05", |
|
"loss": 0.139, |
|
"slid_loss": 0.1409, |
|
"step": 190, |
|
"time": 45.94 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5498e-05", |
|
"loss": 0.1451, |
|
"slid_loss": 0.1409, |
|
"step": 191, |
|
"time": 46.17 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5449e-05", |
|
"loss": 0.1369, |
|
"slid_loss": 0.1409, |
|
"step": 192, |
|
"time": 46.11 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5399e-05", |
|
"loss": 0.1339, |
|
"slid_loss": 0.1408, |
|
"step": 193, |
|
"time": 46.96 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5349e-05", |
|
"loss": 0.1345, |
|
"slid_loss": 0.1407, |
|
"step": 194, |
|
"time": 47.56 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.5299e-05", |
|
"loss": 0.1346, |
|
"slid_loss": 0.1405, |
|
"step": 195, |
|
"time": 46.84 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.5248e-05", |
|
"loss": 0.1351, |
|
"slid_loss": 0.1405, |
|
"step": 196, |
|
"time": 49.93 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.5198e-05", |
|
"loss": 0.1386, |
|
"slid_loss": 0.1405, |
|
"step": 197, |
|
"time": 45.32 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.5147e-05", |
|
"loss": 0.1417, |
|
"slid_loss": 0.1405, |
|
"step": 198, |
|
"time": 47.06 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.5095e-05", |
|
"loss": 0.1379, |
|
"slid_loss": 0.1404, |
|
"step": 199, |
|
"time": 45.81 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.5044e-05", |
|
"loss": 0.1324, |
|
"slid_loss": 0.1403, |
|
"step": 200, |
|
"time": 45.83 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.4992e-05", |
|
"loss": 0.1402, |
|
"slid_loss": 0.1403, |
|
"step": 201, |
|
"time": 46.5 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.4940e-05", |
|
"loss": 0.1349, |
|
"slid_loss": 0.1402, |
|
"step": 202, |
|
"time": 49.6 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.4888e-05", |
|
"loss": 0.1409, |
|
"slid_loss": 0.1402, |
|
"step": 203, |
|
"time": 49.01 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.4836e-05", |
|
"loss": 0.1312, |
|
"slid_loss": 0.14, |
|
"step": 204, |
|
"time": 47.51 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.4783e-05", |
|
"loss": 0.14, |
|
"slid_loss": 0.14, |
|
"step": 205, |
|
"time": 45.38 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.4730e-05", |
|
"loss": 0.1367, |
|
"slid_loss": 0.14, |
|
"step": 206, |
|
"time": 46.64 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.4677e-05", |
|
"loss": 0.1363, |
|
"slid_loss": 0.1399, |
|
"step": 207, |
|
"time": 46.32 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4623e-05", |
|
"loss": 0.1275, |
|
"slid_loss": 0.1397, |
|
"step": 208, |
|
"time": 46.34 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4569e-05", |
|
"loss": 0.1346, |
|
"slid_loss": 0.1395, |
|
"step": 209, |
|
"time": 48.32 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4515e-05", |
|
"loss": 0.1368, |
|
"slid_loss": 0.1396, |
|
"step": 210, |
|
"time": 45.5 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4461e-05", |
|
"loss": 0.1401, |
|
"slid_loss": 0.1395, |
|
"step": 211, |
|
"time": 46.83 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4407e-05", |
|
"loss": 0.136, |
|
"slid_loss": 0.1395, |
|
"step": 212, |
|
"time": 47.35 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4352e-05", |
|
"loss": 0.1433, |
|
"slid_loss": 0.1395, |
|
"step": 213, |
|
"time": 46.19 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4297e-05", |
|
"loss": 0.1332, |
|
"slid_loss": 0.1395, |
|
"step": 214, |
|
"time": 45.95 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4242e-05", |
|
"loss": 0.1427, |
|
"slid_loss": 0.1395, |
|
"step": 215, |
|
"time": 48.28 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4186e-05", |
|
"loss": 0.1444, |
|
"slid_loss": 0.1395, |
|
"step": 216, |
|
"time": 47.17 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4131e-05", |
|
"loss": 0.1356, |
|
"slid_loss": 0.1394, |
|
"step": 217, |
|
"time": 47.52 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4075e-05", |
|
"loss": 0.1422, |
|
"slid_loss": 0.1394, |
|
"step": 218, |
|
"time": 45.98 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.4019e-05", |
|
"loss": 0.1369, |
|
"slid_loss": 0.1393, |
|
"step": 219, |
|
"time": 45.78 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3962e-05", |
|
"loss": 0.1389, |
|
"slid_loss": 0.1393, |
|
"step": 220, |
|
"time": 45.72 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3906e-05", |
|
"loss": 0.1319, |
|
"slid_loss": 0.1392, |
|
"step": 221, |
|
"time": 45.04 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3849e-05", |
|
"loss": 0.1309, |
|
"slid_loss": 0.1391, |
|
"step": 222, |
|
"time": 46.05 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3791e-05", |
|
"loss": 0.1342, |
|
"slid_loss": 0.139, |
|
"step": 223, |
|
"time": 46.89 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3734e-05", |
|
"loss": 0.1373, |
|
"slid_loss": 0.1389, |
|
"step": 224, |
|
"time": 45.79 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3676e-05", |
|
"loss": 0.1387, |
|
"slid_loss": 0.1388, |
|
"step": 225, |
|
"time": 46.5 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3619e-05", |
|
"loss": 0.1317, |
|
"slid_loss": 0.1387, |
|
"step": 226, |
|
"time": 48.07 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3560e-05", |
|
"loss": 0.1432, |
|
"slid_loss": 0.1387, |
|
"step": 227, |
|
"time": 46.22 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3502e-05", |
|
"loss": 0.1358, |
|
"slid_loss": 0.1387, |
|
"step": 228, |
|
"time": 47.6 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3443e-05", |
|
"loss": 0.1341, |
|
"slid_loss": 0.1387, |
|
"step": 229, |
|
"time": 46.8 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3385e-05", |
|
"loss": 0.1334, |
|
"slid_loss": 0.1387, |
|
"step": 230, |
|
"time": 49.29 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.3326e-05", |
|
"loss": 0.1339, |
|
"slid_loss": 0.1385, |
|
"step": 231, |
|
"time": 46.45 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.3266e-05", |
|
"loss": 0.1376, |
|
"slid_loss": 0.1385, |
|
"step": 232, |
|
"time": 46.14 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.3207e-05", |
|
"loss": 0.1363, |
|
"slid_loss": 0.1383, |
|
"step": 233, |
|
"time": 45.01 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.3147e-05", |
|
"loss": 0.1382, |
|
"slid_loss": 0.1382, |
|
"step": 234, |
|
"time": 47.27 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.3087e-05", |
|
"loss": 0.1356, |
|
"slid_loss": 0.1382, |
|
"step": 235, |
|
"time": 44.83 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.3027e-05", |
|
"loss": 0.1396, |
|
"slid_loss": 0.1381, |
|
"step": 236, |
|
"time": 45.61 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.2966e-05", |
|
"loss": 0.1398, |
|
"slid_loss": 0.1381, |
|
"step": 237, |
|
"time": 45.87 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.2905e-05", |
|
"loss": 0.1305, |
|
"slid_loss": 0.138, |
|
"step": 238, |
|
"time": 48.21 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.2844e-05", |
|
"loss": 0.1344, |
|
"slid_loss": 0.1379, |
|
"step": 239, |
|
"time": 45.7 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.2783e-05", |
|
"loss": 0.1443, |
|
"slid_loss": 0.1379, |
|
"step": 240, |
|
"time": 47.16 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.2722e-05", |
|
"loss": 0.1361, |
|
"slid_loss": 0.1379, |
|
"step": 241, |
|
"time": 48.03 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.2660e-05", |
|
"loss": 0.1289, |
|
"slid_loss": 0.1377, |
|
"step": 242, |
|
"time": 48.08 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.2598e-05", |
|
"loss": 0.1469, |
|
"slid_loss": 0.1379, |
|
"step": 243, |
|
"time": 47.05 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.2536e-05", |
|
"loss": 0.1327, |
|
"slid_loss": 0.1379, |
|
"step": 244, |
|
"time": 44.99 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.2474e-05", |
|
"loss": 0.1368, |
|
"slid_loss": 0.1379, |
|
"step": 245, |
|
"time": 46.44 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.2411e-05", |
|
"loss": 0.1385, |
|
"slid_loss": 0.1379, |
|
"step": 246, |
|
"time": 47.77 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.2348e-05", |
|
"loss": 0.1367, |
|
"slid_loss": 0.1379, |
|
"step": 247, |
|
"time": 45.9 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.2285e-05", |
|
"loss": 0.13, |
|
"slid_loss": 0.1378, |
|
"step": 248, |
|
"time": 45.19 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.2222e-05", |
|
"loss": 0.1264, |
|
"slid_loss": 0.1378, |
|
"step": 249, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.2158e-05", |
|
"loss": 0.1413, |
|
"slid_loss": 0.1378, |
|
"step": 250, |
|
"time": 47.56 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.2094e-05", |
|
"loss": 0.1404, |
|
"slid_loss": 0.1378, |
|
"step": 251, |
|
"time": 45.91 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.2030e-05", |
|
"loss": 0.1368, |
|
"slid_loss": 0.1378, |
|
"step": 252, |
|
"time": 44.73 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.1966e-05", |
|
"loss": 0.1397, |
|
"slid_loss": 0.1376, |
|
"step": 253, |
|
"time": 46.38 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.1902e-05", |
|
"loss": 0.1376, |
|
"slid_loss": 0.1376, |
|
"step": 254, |
|
"time": 45.53 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.1837e-05", |
|
"loss": 0.1328, |
|
"slid_loss": 0.1375, |
|
"step": 255, |
|
"time": 46.59 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1772e-05", |
|
"loss": 0.1337, |
|
"slid_loss": 0.1375, |
|
"step": 256, |
|
"time": 46.71 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1707e-05", |
|
"loss": 0.1374, |
|
"slid_loss": 0.1375, |
|
"step": 257, |
|
"time": 46.27 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1642e-05", |
|
"loss": 0.1326, |
|
"slid_loss": 0.1374, |
|
"step": 258, |
|
"time": 47.03 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1576e-05", |
|
"loss": 0.1279, |
|
"slid_loss": 0.1373, |
|
"step": 259, |
|
"time": 44.97 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1510e-05", |
|
"loss": 0.1303, |
|
"slid_loss": 0.1372, |
|
"step": 260, |
|
"time": 47.27 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1444e-05", |
|
"loss": 0.134, |
|
"slid_loss": 0.1371, |
|
"step": 261, |
|
"time": 47.77 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1378e-05", |
|
"loss": 0.131, |
|
"slid_loss": 0.1371, |
|
"step": 262, |
|
"time": 47.75 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1311e-05", |
|
"loss": 0.1343, |
|
"slid_loss": 0.137, |
|
"step": 263, |
|
"time": 49.38 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1245e-05", |
|
"loss": 0.1419, |
|
"slid_loss": 0.137, |
|
"step": 264, |
|
"time": 48.7 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1178e-05", |
|
"loss": 0.139, |
|
"slid_loss": 0.137, |
|
"step": 265, |
|
"time": 44.41 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1111e-05", |
|
"loss": 0.1325, |
|
"slid_loss": 0.1369, |
|
"step": 266, |
|
"time": 47.23 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.1043e-05", |
|
"loss": 0.1396, |
|
"slid_loss": 0.1369, |
|
"step": 267, |
|
"time": 45.76 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0976e-05", |
|
"loss": 0.1381, |
|
"slid_loss": 0.1369, |
|
"step": 268, |
|
"time": 45.71 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0908e-05", |
|
"loss": 0.135, |
|
"slid_loss": 0.137, |
|
"step": 269, |
|
"time": 47.32 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0840e-05", |
|
"loss": 0.1312, |
|
"slid_loss": 0.137, |
|
"step": 270, |
|
"time": 45.63 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0771e-05", |
|
"loss": 0.1276, |
|
"slid_loss": 0.1368, |
|
"step": 271, |
|
"time": 45.77 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0703e-05", |
|
"loss": 0.1407, |
|
"slid_loss": 0.1367, |
|
"step": 272, |
|
"time": 46.35 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0634e-05", |
|
"loss": 0.1302, |
|
"slid_loss": 0.1367, |
|
"step": 273, |
|
"time": 47.35 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0565e-05", |
|
"loss": 0.143, |
|
"slid_loss": 0.1367, |
|
"step": 274, |
|
"time": 45.74 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0496e-05", |
|
"loss": 0.1337, |
|
"slid_loss": 0.1366, |
|
"step": 275, |
|
"time": 45.74 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0427e-05", |
|
"loss": 0.1318, |
|
"slid_loss": 0.1366, |
|
"step": 276, |
|
"time": 47.6 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0357e-05", |
|
"loss": 0.133, |
|
"slid_loss": 0.1365, |
|
"step": 277, |
|
"time": 45.32 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.0287e-05", |
|
"loss": 0.1359, |
|
"slid_loss": 0.1364, |
|
"step": 278, |
|
"time": 48.58 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.0217e-05", |
|
"loss": 0.1359, |
|
"slid_loss": 0.1363, |
|
"step": 279, |
|
"time": 45.08 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.0147e-05", |
|
"loss": 0.1377, |
|
"slid_loss": 0.1363, |
|
"step": 280, |
|
"time": 46.46 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.0077e-05", |
|
"loss": 0.1306, |
|
"slid_loss": 0.1363, |
|
"step": 281, |
|
"time": 46.85 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.0006e-05", |
|
"loss": 0.1301, |
|
"slid_loss": 0.1362, |
|
"step": 282, |
|
"time": 45.49 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "8.9935e-05", |
|
"loss": 0.1345, |
|
"slid_loss": 0.1362, |
|
"step": 283, |
|
"time": 48.25 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "8.9864e-05", |
|
"loss": 0.1297, |
|
"slid_loss": 0.1362, |
|
"step": 284, |
|
"time": 45.85 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "8.9793e-05", |
|
"loss": 0.1373, |
|
"slid_loss": 0.1362, |
|
"step": 285, |
|
"time": 46.44 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "8.9721e-05", |
|
"loss": 0.1368, |
|
"slid_loss": 0.1362, |
|
"step": 286, |
|
"time": 46.42 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "8.9649e-05", |
|
"loss": 0.138, |
|
"slid_loss": 0.1361, |
|
"step": 287, |
|
"time": 46.31 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "8.9577e-05", |
|
"loss": 0.1345, |
|
"slid_loss": 0.1361, |
|
"step": 288, |
|
"time": 45.31 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "8.9505e-05", |
|
"loss": 0.1374, |
|
"slid_loss": 0.136, |
|
"step": 289, |
|
"time": 47.19 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "8.9433e-05", |
|
"loss": 0.133, |
|
"slid_loss": 0.136, |
|
"step": 290, |
|
"time": 46.14 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.9360e-05", |
|
"loss": 0.1334, |
|
"slid_loss": 0.1358, |
|
"step": 291, |
|
"time": 45.89 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.9287e-05", |
|
"loss": 0.1331, |
|
"slid_loss": 0.1358, |
|
"step": 292, |
|
"time": 48.08 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.9214e-05", |
|
"loss": 0.1314, |
|
"slid_loss": 0.1358, |
|
"step": 293, |
|
"time": 46.56 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.9141e-05", |
|
"loss": 0.1275, |
|
"slid_loss": 0.1357, |
|
"step": 294, |
|
"time": 48.63 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.9068e-05", |
|
"loss": 0.1387, |
|
"slid_loss": 0.1358, |
|
"step": 295, |
|
"time": 48.59 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.8994e-05", |
|
"loss": 0.1335, |
|
"slid_loss": 0.1357, |
|
"step": 296, |
|
"time": 47.66 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.8920e-05", |
|
"loss": 0.1363, |
|
"slid_loss": 0.1357, |
|
"step": 297, |
|
"time": 46.77 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.8846e-05", |
|
"loss": 0.1322, |
|
"slid_loss": 0.1356, |
|
"step": 298, |
|
"time": 48.14 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.8772e-05", |
|
"loss": 0.1324, |
|
"slid_loss": 0.1356, |
|
"step": 299, |
|
"time": 48.01 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.8697e-05", |
|
"loss": 0.1272, |
|
"slid_loss": 0.1355, |
|
"step": 300, |
|
"time": 47.08 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.8623e-05", |
|
"loss": 0.1399, |
|
"slid_loss": 0.1355, |
|
"step": 301, |
|
"time": 47.34 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "8.8548e-05", |
|
"loss": 0.1364, |
|
"slid_loss": 0.1355, |
|
"step": 302, |
|
"time": 45.45 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.8473e-05", |
|
"loss": 0.1307, |
|
"slid_loss": 0.1354, |
|
"step": 303, |
|
"time": 45.43 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.8398e-05", |
|
"loss": 0.1303, |
|
"slid_loss": 0.1354, |
|
"step": 304, |
|
"time": 45.13 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.8322e-05", |
|
"loss": 0.1412, |
|
"slid_loss": 0.1354, |
|
"step": 305, |
|
"time": 47.48 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.8246e-05", |
|
"loss": 0.1257, |
|
"slid_loss": 0.1353, |
|
"step": 306, |
|
"time": 47.46 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.8171e-05", |
|
"loss": 0.1307, |
|
"slid_loss": 0.1353, |
|
"step": 307, |
|
"time": 46.46 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.8094e-05", |
|
"loss": 0.1294, |
|
"slid_loss": 0.1353, |
|
"step": 308, |
|
"time": 47.94 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.8018e-05", |
|
"loss": 0.1318, |
|
"slid_loss": 0.1352, |
|
"step": 309, |
|
"time": 47.16 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.7942e-05", |
|
"loss": 0.1346, |
|
"slid_loss": 0.1352, |
|
"step": 310, |
|
"time": 45.65 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.7865e-05", |
|
"loss": 0.1362, |
|
"slid_loss": 0.1352, |
|
"step": 311, |
|
"time": 48.16 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.7788e-05", |
|
"loss": 0.1388, |
|
"slid_loss": 0.1352, |
|
"step": 312, |
|
"time": 51.78 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.7711e-05", |
|
"loss": 0.1291, |
|
"slid_loss": 0.1351, |
|
"step": 313, |
|
"time": 46.35 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "8.7634e-05", |
|
"loss": 0.1372, |
|
"slid_loss": 0.1351, |
|
"step": 314, |
|
"time": 48.89 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.7556e-05", |
|
"loss": 0.1288, |
|
"slid_loss": 0.135, |
|
"step": 315, |
|
"time": 45.88 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.7478e-05", |
|
"loss": 0.1332, |
|
"slid_loss": 0.1349, |
|
"step": 316, |
|
"time": 47.52 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.7401e-05", |
|
"loss": 0.1337, |
|
"slid_loss": 0.1348, |
|
"step": 317, |
|
"time": 45.04 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.7323e-05", |
|
"loss": 0.131, |
|
"slid_loss": 0.1347, |
|
"step": 318, |
|
"time": 45.84 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.7244e-05", |
|
"loss": 0.1293, |
|
"slid_loss": 0.1347, |
|
"step": 319, |
|
"time": 48.9 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.7166e-05", |
|
"loss": 0.124, |
|
"slid_loss": 0.1345, |
|
"step": 320, |
|
"time": 46.7 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.7087e-05", |
|
"loss": 0.1322, |
|
"slid_loss": 0.1345, |
|
"step": 321, |
|
"time": 46.28 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.7008e-05", |
|
"loss": 0.1285, |
|
"slid_loss": 0.1345, |
|
"step": 322, |
|
"time": 46.66 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.6929e-05", |
|
"loss": 0.1279, |
|
"slid_loss": 0.1344, |
|
"step": 323, |
|
"time": 46.51 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.6850e-05", |
|
"loss": 0.1397, |
|
"slid_loss": 0.1344, |
|
"step": 324, |
|
"time": 45.32 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.6771e-05", |
|
"loss": 0.1308, |
|
"slid_loss": 0.1344, |
|
"step": 325, |
|
"time": 46.31 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "8.6691e-05", |
|
"loss": 0.1334, |
|
"slid_loss": 0.1344, |
|
"step": 326, |
|
"time": 46.36 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.6611e-05", |
|
"loss": 0.1366, |
|
"slid_loss": 0.1343, |
|
"step": 327, |
|
"time": 45.64 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.6531e-05", |
|
"loss": 0.1333, |
|
"slid_loss": 0.1343, |
|
"step": 328, |
|
"time": 45.06 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.6451e-05", |
|
"loss": 0.1258, |
|
"slid_loss": 0.1342, |
|
"step": 329, |
|
"time": 47.7 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.6370e-05", |
|
"loss": 0.1377, |
|
"slid_loss": 0.1343, |
|
"step": 330, |
|
"time": 47.65 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.6290e-05", |
|
"loss": 0.1372, |
|
"slid_loss": 0.1343, |
|
"step": 331, |
|
"time": 46.1 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.6209e-05", |
|
"loss": 0.1288, |
|
"slid_loss": 0.1342, |
|
"step": 332, |
|
"time": 46.35 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.6128e-05", |
|
"loss": 0.1328, |
|
"slid_loss": 0.1342, |
|
"step": 333, |
|
"time": 45.03 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.6047e-05", |
|
"loss": 0.1307, |
|
"slid_loss": 0.1341, |
|
"step": 334, |
|
"time": 45.68 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.5966e-05", |
|
"loss": 0.1349, |
|
"slid_loss": 0.1341, |
|
"step": 335, |
|
"time": 49.23 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.5884e-05", |
|
"loss": 0.1324, |
|
"slid_loss": 0.134, |
|
"step": 336, |
|
"time": 48.05 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.5802e-05", |
|
"loss": 0.1393, |
|
"slid_loss": 0.134, |
|
"step": 337, |
|
"time": 45.29 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "8.5721e-05", |
|
"loss": 0.1318, |
|
"slid_loss": 0.134, |
|
"step": 338, |
|
"time": 47.05 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.5639e-05", |
|
"loss": 0.1376, |
|
"slid_loss": 0.1341, |
|
"step": 339, |
|
"time": 47.72 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.5556e-05", |
|
"loss": 0.1318, |
|
"slid_loss": 0.1339, |
|
"step": 340, |
|
"time": 46.61 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.5474e-05", |
|
"loss": 0.1319, |
|
"slid_loss": 0.1339, |
|
"step": 341, |
|
"time": 46.36 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.5391e-05", |
|
"loss": 0.1356, |
|
"slid_loss": 0.134, |
|
"step": 342, |
|
"time": 47.32 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.5308e-05", |
|
"loss": 0.1245, |
|
"slid_loss": 0.1337, |
|
"step": 343, |
|
"time": 48.38 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.5225e-05", |
|
"loss": 0.1311, |
|
"slid_loss": 0.1337, |
|
"step": 344, |
|
"time": 47.25 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.5142e-05", |
|
"loss": 0.1378, |
|
"slid_loss": 0.1337, |
|
"step": 345, |
|
"time": 48.14 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.5059e-05", |
|
"loss": 0.1413, |
|
"slid_loss": 0.1337, |
|
"step": 346, |
|
"time": 46.51 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.4975e-05", |
|
"loss": 0.1356, |
|
"slid_loss": 0.1337, |
|
"step": 347, |
|
"time": 45.1 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.4892e-05", |
|
"loss": 0.1295, |
|
"slid_loss": 0.1337, |
|
"step": 348, |
|
"time": 46.81 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.4808e-05", |
|
"loss": 0.1277, |
|
"slid_loss": 0.1337, |
|
"step": 349, |
|
"time": 47.47 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "8.4724e-05", |
|
"loss": 0.1292, |
|
"slid_loss": 0.1336, |
|
"step": 350, |
|
"time": 47.44 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.4640e-05", |
|
"loss": 0.1283, |
|
"slid_loss": 0.1335, |
|
"step": 351, |
|
"time": 50.1 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.4555e-05", |
|
"loss": 0.1251, |
|
"slid_loss": 0.1334, |
|
"step": 352, |
|
"time": 46.83 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.4471e-05", |
|
"loss": 0.1405, |
|
"slid_loss": 0.1334, |
|
"step": 353, |
|
"time": 48.32 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.4386e-05", |
|
"loss": 0.1254, |
|
"slid_loss": 0.1333, |
|
"step": 354, |
|
"time": 47.68 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.4301e-05", |
|
"loss": 0.1326, |
|
"slid_loss": 0.1333, |
|
"step": 355, |
|
"time": 47.78 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.4216e-05", |
|
"loss": 0.1436, |
|
"slid_loss": 0.1334, |
|
"step": 356, |
|
"time": 47.13 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.4131e-05", |
|
"loss": 0.13, |
|
"slid_loss": 0.1333, |
|
"step": 357, |
|
"time": 46.48 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.4045e-05", |
|
"loss": 0.1323, |
|
"slid_loss": 0.1333, |
|
"step": 358, |
|
"time": 47.54 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.3959e-05", |
|
"loss": 0.1331, |
|
"slid_loss": 0.1333, |
|
"step": 359, |
|
"time": 46.67 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.3874e-05", |
|
"loss": 0.1387, |
|
"slid_loss": 0.1334, |
|
"step": 360, |
|
"time": 48.05 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "8.3788e-05", |
|
"loss": 0.1309, |
|
"slid_loss": 0.1334, |
|
"step": 361, |
|
"time": 46.49 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.3702e-05", |
|
"loss": 0.131, |
|
"slid_loss": 0.1334, |
|
"step": 362, |
|
"time": 47.59 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.3615e-05", |
|
"loss": 0.1396, |
|
"slid_loss": 0.1335, |
|
"step": 363, |
|
"time": 47.87 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.3529e-05", |
|
"loss": 0.1297, |
|
"slid_loss": 0.1333, |
|
"step": 364, |
|
"time": 45.82 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.3442e-05", |
|
"loss": 0.1253, |
|
"slid_loss": 0.1332, |
|
"step": 365, |
|
"time": 47.38 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.3355e-05", |
|
"loss": 0.1288, |
|
"slid_loss": 0.1332, |
|
"step": 366, |
|
"time": 45.19 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.3268e-05", |
|
"loss": 0.1376, |
|
"slid_loss": 0.1331, |
|
"step": 367, |
|
"time": 46.03 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.3181e-05", |
|
"loss": 0.1346, |
|
"slid_loss": 0.1331, |
|
"step": 368, |
|
"time": 46.73 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.3094e-05", |
|
"loss": 0.1412, |
|
"slid_loss": 0.1332, |
|
"step": 369, |
|
"time": 47.2 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.3006e-05", |
|
"loss": 0.131, |
|
"slid_loss": 0.1332, |
|
"step": 370, |
|
"time": 46.76 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.2919e-05", |
|
"loss": 0.1436, |
|
"slid_loss": 0.1333, |
|
"step": 371, |
|
"time": 47.27 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.2831e-05", |
|
"loss": 0.1296, |
|
"slid_loss": 0.1332, |
|
"step": 372, |
|
"time": 46.89 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "8.2743e-05", |
|
"loss": 0.1287, |
|
"slid_loss": 0.1332, |
|
"step": 373, |
|
"time": 44.38 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.2655e-05", |
|
"loss": 0.1384, |
|
"slid_loss": 0.1331, |
|
"step": 374, |
|
"time": 47.52 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.2566e-05", |
|
"loss": 0.1357, |
|
"slid_loss": 0.1332, |
|
"step": 375, |
|
"time": 46.74 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.2478e-05", |
|
"loss": 0.1257, |
|
"slid_loss": 0.1331, |
|
"step": 376, |
|
"time": 45.27 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.2389e-05", |
|
"loss": 0.1301, |
|
"slid_loss": 0.1331, |
|
"step": 377, |
|
"time": 46.68 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.2301e-05", |
|
"loss": 0.1363, |
|
"slid_loss": 0.1331, |
|
"step": 378, |
|
"time": 47.75 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.2212e-05", |
|
"loss": 0.1263, |
|
"slid_loss": 0.133, |
|
"step": 379, |
|
"time": 44.94 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.2123e-05", |
|
"loss": 0.1325, |
|
"slid_loss": 0.1329, |
|
"step": 380, |
|
"time": 47.44 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.2033e-05", |
|
"loss": 0.1373, |
|
"slid_loss": 0.133, |
|
"step": 381, |
|
"time": 46.08 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.1944e-05", |
|
"loss": 0.139, |
|
"slid_loss": 0.1331, |
|
"step": 382, |
|
"time": 50.22 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.1854e-05", |
|
"loss": 0.1341, |
|
"slid_loss": 0.1331, |
|
"step": 383, |
|
"time": 46.37 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.1765e-05", |
|
"loss": 0.128, |
|
"slid_loss": 0.1331, |
|
"step": 384, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "8.1675e-05", |
|
"loss": 0.125, |
|
"slid_loss": 0.1329, |
|
"step": 385, |
|
"time": 46.41 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.1585e-05", |
|
"loss": 0.1328, |
|
"slid_loss": 0.1329, |
|
"step": 386, |
|
"time": 45.24 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.1494e-05", |
|
"loss": 0.1305, |
|
"slid_loss": 0.1328, |
|
"step": 387, |
|
"time": 46.78 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.1404e-05", |
|
"loss": 0.1334, |
|
"slid_loss": 0.1328, |
|
"step": 388, |
|
"time": 47.67 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.1314e-05", |
|
"loss": 0.1325, |
|
"slid_loss": 0.1328, |
|
"step": 389, |
|
"time": 47.71 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.1223e-05", |
|
"loss": 0.1226, |
|
"slid_loss": 0.1327, |
|
"step": 390, |
|
"time": 46.58 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.1132e-05", |
|
"loss": 0.1348, |
|
"slid_loss": 0.1327, |
|
"step": 391, |
|
"time": 47.42 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.1041e-05", |
|
"loss": 0.1377, |
|
"slid_loss": 0.1327, |
|
"step": 392, |
|
"time": 47.11 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.0950e-05", |
|
"loss": 0.1367, |
|
"slid_loss": 0.1328, |
|
"step": 393, |
|
"time": 46.24 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.0859e-05", |
|
"loss": 0.1331, |
|
"slid_loss": 0.1328, |
|
"step": 394, |
|
"time": 46.44 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.0767e-05", |
|
"loss": 0.1334, |
|
"slid_loss": 0.1328, |
|
"step": 395, |
|
"time": 44.03 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.0676e-05", |
|
"loss": 0.1288, |
|
"slid_loss": 0.1327, |
|
"step": 396, |
|
"time": 46.18 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "8.0584e-05", |
|
"loss": 0.1351, |
|
"slid_loss": 0.1327, |
|
"step": 397, |
|
"time": 47.62 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "8.0492e-05", |
|
"loss": 0.1269, |
|
"slid_loss": 0.1327, |
|
"step": 398, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "8.0400e-05", |
|
"loss": 0.1278, |
|
"slid_loss": 0.1326, |
|
"step": 399, |
|
"time": 45.92 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "8.0308e-05", |
|
"loss": 0.1385, |
|
"slid_loss": 0.1327, |
|
"step": 400, |
|
"time": 45.8 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "8.0216e-05", |
|
"loss": 0.1272, |
|
"slid_loss": 0.1326, |
|
"step": 401, |
|
"time": 48.02 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "8.0123e-05", |
|
"loss": 0.1162, |
|
"slid_loss": 0.1324, |
|
"step": 402, |
|
"time": 48.85 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "8.0031e-05", |
|
"loss": 0.1308, |
|
"slid_loss": 0.1324, |
|
"step": 403, |
|
"time": 45.57 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "7.9938e-05", |
|
"loss": 0.1266, |
|
"slid_loss": 0.1324, |
|
"step": 404, |
|
"time": 45.77 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "7.9845e-05", |
|
"loss": 0.1349, |
|
"slid_loss": 0.1323, |
|
"step": 405, |
|
"time": 46.6 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "7.9752e-05", |
|
"loss": 0.1197, |
|
"slid_loss": 0.1323, |
|
"step": 406, |
|
"time": 44.61 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "7.9659e-05", |
|
"loss": 0.1363, |
|
"slid_loss": 0.1323, |
|
"step": 407, |
|
"time": 47.86 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "7.9566e-05", |
|
"loss": 0.1336, |
|
"slid_loss": 0.1323, |
|
"step": 408, |
|
"time": 47.58 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "7.9472e-05", |
|
"loss": 0.1332, |
|
"slid_loss": 0.1324, |
|
"step": 409, |
|
"time": 44.94 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.9379e-05", |
|
"loss": 0.1252, |
|
"slid_loss": 0.1323, |
|
"step": 410, |
|
"time": 46.27 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.9285e-05", |
|
"loss": 0.129, |
|
"slid_loss": 0.1322, |
|
"step": 411, |
|
"time": 46.03 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.9191e-05", |
|
"loss": 0.1257, |
|
"slid_loss": 0.1321, |
|
"step": 412, |
|
"time": 46.49 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.9097e-05", |
|
"loss": 0.14, |
|
"slid_loss": 0.1322, |
|
"step": 413, |
|
"time": 45.87 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.9003e-05", |
|
"loss": 0.1209, |
|
"slid_loss": 0.132, |
|
"step": 414, |
|
"time": 46.87 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.8909e-05", |
|
"loss": 0.1216, |
|
"slid_loss": 0.1319, |
|
"step": 415, |
|
"time": 46.89 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.8815e-05", |
|
"loss": 0.1263, |
|
"slid_loss": 0.1319, |
|
"step": 416, |
|
"time": 45.85 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.8720e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1317, |
|
"step": 417, |
|
"time": 45.86 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.8625e-05", |
|
"loss": 0.1332, |
|
"slid_loss": 0.1317, |
|
"step": 418, |
|
"time": 46.26 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.8531e-05", |
|
"loss": 0.1247, |
|
"slid_loss": 0.1317, |
|
"step": 419, |
|
"time": 46.52 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.8436e-05", |
|
"loss": 0.1429, |
|
"slid_loss": 0.1319, |
|
"step": 420, |
|
"time": 44.08 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "7.8341e-05", |
|
"loss": 0.1346, |
|
"slid_loss": 0.1319, |
|
"step": 421, |
|
"time": 45.56 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.8245e-05", |
|
"loss": 0.1236, |
|
"slid_loss": 0.1319, |
|
"step": 422, |
|
"time": 46.09 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.8150e-05", |
|
"loss": 0.1339, |
|
"slid_loss": 0.1319, |
|
"step": 423, |
|
"time": 46.39 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.8055e-05", |
|
"loss": 0.1274, |
|
"slid_loss": 0.1318, |
|
"step": 424, |
|
"time": 47.06 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.7959e-05", |
|
"loss": 0.1316, |
|
"slid_loss": 0.1318, |
|
"step": 425, |
|
"time": 48.23 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.7863e-05", |
|
"loss": 0.1291, |
|
"slid_loss": 0.1318, |
|
"step": 426, |
|
"time": 45.25 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.7768e-05", |
|
"loss": 0.1232, |
|
"slid_loss": 0.1316, |
|
"step": 427, |
|
"time": 45.73 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.7672e-05", |
|
"loss": 0.1252, |
|
"slid_loss": 0.1315, |
|
"step": 428, |
|
"time": 46.88 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.7575e-05", |
|
"loss": 0.1267, |
|
"slid_loss": 0.1315, |
|
"step": 429, |
|
"time": 46.09 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.7479e-05", |
|
"loss": 0.1245, |
|
"slid_loss": 0.1314, |
|
"step": 430, |
|
"time": 44.91 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.7383e-05", |
|
"loss": 0.1279, |
|
"slid_loss": 0.1313, |
|
"step": 431, |
|
"time": 46.5 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.7286e-05", |
|
"loss": 0.13, |
|
"slid_loss": 0.1313, |
|
"step": 432, |
|
"time": 48.81 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "7.7190e-05", |
|
"loss": 0.1209, |
|
"slid_loss": 0.1312, |
|
"step": 433, |
|
"time": 47.13 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.7093e-05", |
|
"loss": 0.1358, |
|
"slid_loss": 0.1313, |
|
"step": 434, |
|
"time": 45.47 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6996e-05", |
|
"loss": 0.1366, |
|
"slid_loss": 0.1313, |
|
"step": 435, |
|
"time": 45.98 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6899e-05", |
|
"loss": 0.1315, |
|
"slid_loss": 0.1313, |
|
"step": 436, |
|
"time": 45.82 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6802e-05", |
|
"loss": 0.1305, |
|
"slid_loss": 0.1312, |
|
"step": 437, |
|
"time": 46.56 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6705e-05", |
|
"loss": 0.1326, |
|
"slid_loss": 0.1312, |
|
"step": 438, |
|
"time": 46.4 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6608e-05", |
|
"loss": 0.1221, |
|
"slid_loss": 0.131, |
|
"step": 439, |
|
"time": 47.54 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6510e-05", |
|
"loss": 0.1307, |
|
"slid_loss": 0.131, |
|
"step": 440, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6413e-05", |
|
"loss": 0.1247, |
|
"slid_loss": 0.131, |
|
"step": 441, |
|
"time": 47.53 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6315e-05", |
|
"loss": 0.1246, |
|
"slid_loss": 0.1308, |
|
"step": 442, |
|
"time": 47.38 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6217e-05", |
|
"loss": 0.1271, |
|
"slid_loss": 0.1309, |
|
"step": 443, |
|
"time": 46.51 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6119e-05", |
|
"loss": 0.1265, |
|
"slid_loss": 0.1308, |
|
"step": 444, |
|
"time": 46.68 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "7.6021e-05", |
|
"loss": 0.1167, |
|
"slid_loss": 0.1306, |
|
"step": 445, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5923e-05", |
|
"loss": 0.1242, |
|
"slid_loss": 0.1304, |
|
"step": 446, |
|
"time": 45.1 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5825e-05", |
|
"loss": 0.1241, |
|
"slid_loss": 0.1303, |
|
"step": 447, |
|
"time": 47.81 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5727e-05", |
|
"loss": 0.1274, |
|
"slid_loss": 0.1303, |
|
"step": 448, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5628e-05", |
|
"loss": 0.1281, |
|
"slid_loss": 0.1303, |
|
"step": 449, |
|
"time": 47.53 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5530e-05", |
|
"loss": 0.13, |
|
"slid_loss": 0.1303, |
|
"step": 450, |
|
"time": 45.53 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5431e-05", |
|
"loss": 0.114, |
|
"slid_loss": 0.1302, |
|
"step": 451, |
|
"time": 46.52 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5332e-05", |
|
"loss": 0.1267, |
|
"slid_loss": 0.1302, |
|
"step": 452, |
|
"time": 47.67 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5233e-05", |
|
"loss": 0.1176, |
|
"slid_loss": 0.13, |
|
"step": 453, |
|
"time": 44.83 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5134e-05", |
|
"loss": 0.1332, |
|
"slid_loss": 0.13, |
|
"step": 454, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.5035e-05", |
|
"loss": 0.1275, |
|
"slid_loss": 0.13, |
|
"step": 455, |
|
"time": 47.44 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "7.4936e-05", |
|
"loss": 0.1231, |
|
"slid_loss": 0.1298, |
|
"step": 456, |
|
"time": 45.97 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.4836e-05", |
|
"loss": 0.1282, |
|
"slid_loss": 0.1298, |
|
"step": 457, |
|
"time": 46.83 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.4737e-05", |
|
"loss": 0.1312, |
|
"slid_loss": 0.1298, |
|
"step": 458, |
|
"time": 45.03 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.4637e-05", |
|
"loss": 0.1236, |
|
"slid_loss": 0.1297, |
|
"step": 459, |
|
"time": 46.63 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.4538e-05", |
|
"loss": 0.1302, |
|
"slid_loss": 0.1296, |
|
"step": 460, |
|
"time": 47.17 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.4438e-05", |
|
"loss": 0.1287, |
|
"slid_loss": 0.1296, |
|
"step": 461, |
|
"time": 48.56 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.4338e-05", |
|
"loss": 0.1271, |
|
"slid_loss": 0.1295, |
|
"step": 462, |
|
"time": 45.87 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.4238e-05", |
|
"loss": 0.1279, |
|
"slid_loss": 0.1294, |
|
"step": 463, |
|
"time": 46.72 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.4138e-05", |
|
"loss": 0.1247, |
|
"slid_loss": 0.1293, |
|
"step": 464, |
|
"time": 46.61 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.4038e-05", |
|
"loss": 0.1261, |
|
"slid_loss": 0.1294, |
|
"step": 465, |
|
"time": 45.42 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.3938e-05", |
|
"loss": 0.1252, |
|
"slid_loss": 0.1293, |
|
"step": 466, |
|
"time": 45.49 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.3837e-05", |
|
"loss": 0.1267, |
|
"slid_loss": 0.1292, |
|
"step": 467, |
|
"time": 46.25 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "7.3737e-05", |
|
"loss": 0.1329, |
|
"slid_loss": 0.1292, |
|
"step": 468, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.3636e-05", |
|
"loss": 0.1196, |
|
"slid_loss": 0.129, |
|
"step": 469, |
|
"time": 46.99 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.3535e-05", |
|
"loss": 0.122, |
|
"slid_loss": 0.1289, |
|
"step": 470, |
|
"time": 47.19 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.3435e-05", |
|
"loss": 0.1296, |
|
"slid_loss": 0.1287, |
|
"step": 471, |
|
"time": 45.5 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.3334e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1286, |
|
"step": 472, |
|
"time": 45.99 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.3233e-05", |
|
"loss": 0.1258, |
|
"slid_loss": 0.1286, |
|
"step": 473, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.3132e-05", |
|
"loss": 0.1372, |
|
"slid_loss": 0.1286, |
|
"step": 474, |
|
"time": 44.81 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.3031e-05", |
|
"loss": 0.1176, |
|
"slid_loss": 0.1284, |
|
"step": 475, |
|
"time": 47.61 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.2929e-05", |
|
"loss": 0.1225, |
|
"slid_loss": 0.1284, |
|
"step": 476, |
|
"time": 44.58 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.2828e-05", |
|
"loss": 0.1284, |
|
"slid_loss": 0.1284, |
|
"step": 477, |
|
"time": 46.35 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.2727e-05", |
|
"loss": 0.1286, |
|
"slid_loss": 0.1283, |
|
"step": 478, |
|
"time": 46.28 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.2625e-05", |
|
"loss": 0.1276, |
|
"slid_loss": 0.1283, |
|
"step": 479, |
|
"time": 45.77 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "7.2523e-05", |
|
"loss": 0.1316, |
|
"slid_loss": 0.1283, |
|
"step": 480, |
|
"time": 45.24 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.2422e-05", |
|
"loss": 0.1221, |
|
"slid_loss": 0.1281, |
|
"step": 481, |
|
"time": 45.57 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.2320e-05", |
|
"loss": 0.1216, |
|
"slid_loss": 0.128, |
|
"step": 482, |
|
"time": 46.64 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.2218e-05", |
|
"loss": 0.1202, |
|
"slid_loss": 0.1278, |
|
"step": 483, |
|
"time": 45.27 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.2116e-05", |
|
"loss": 0.129, |
|
"slid_loss": 0.1278, |
|
"step": 484, |
|
"time": 46.2 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.2014e-05", |
|
"loss": 0.1314, |
|
"slid_loss": 0.1279, |
|
"step": 485, |
|
"time": 49.14 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.1912e-05", |
|
"loss": 0.1242, |
|
"slid_loss": 0.1278, |
|
"step": 486, |
|
"time": 47.23 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.1810e-05", |
|
"loss": 0.1282, |
|
"slid_loss": 0.1278, |
|
"step": 487, |
|
"time": 45.96 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.1707e-05", |
|
"loss": 0.1317, |
|
"slid_loss": 0.1278, |
|
"step": 488, |
|
"time": 47.52 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.1605e-05", |
|
"loss": 0.127, |
|
"slid_loss": 0.1277, |
|
"step": 489, |
|
"time": 45.46 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.1503e-05", |
|
"loss": 0.1296, |
|
"slid_loss": 0.1278, |
|
"step": 490, |
|
"time": 46.13 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.1400e-05", |
|
"loss": 0.1215, |
|
"slid_loss": 0.1277, |
|
"step": 491, |
|
"time": 46.36 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "7.1297e-05", |
|
"loss": 0.1211, |
|
"slid_loss": 0.1275, |
|
"step": 492, |
|
"time": 46.4 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.1195e-05", |
|
"loss": 0.1284, |
|
"slid_loss": 0.1274, |
|
"step": 493, |
|
"time": 46.68 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.1092e-05", |
|
"loss": 0.1205, |
|
"slid_loss": 0.1273, |
|
"step": 494, |
|
"time": 46.9 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0989e-05", |
|
"loss": 0.1231, |
|
"slid_loss": 0.1272, |
|
"step": 495, |
|
"time": 46.14 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0886e-05", |
|
"loss": 0.1299, |
|
"slid_loss": 0.1272, |
|
"step": 496, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0783e-05", |
|
"loss": 0.1286, |
|
"slid_loss": 0.1271, |
|
"step": 497, |
|
"time": 46.68 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0680e-05", |
|
"loss": 0.1201, |
|
"slid_loss": 0.1271, |
|
"step": 498, |
|
"time": 48.29 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0577e-05", |
|
"loss": 0.1255, |
|
"slid_loss": 0.127, |
|
"step": 499, |
|
"time": 49.17 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0473e-05", |
|
"loss": 0.124, |
|
"slid_loss": 0.1269, |
|
"step": 500, |
|
"time": 47.87 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0370e-05", |
|
"loss": 0.1234, |
|
"slid_loss": 0.1269, |
|
"step": 501, |
|
"time": 45.78 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0267e-05", |
|
"loss": 0.1197, |
|
"slid_loss": 0.1269, |
|
"step": 502, |
|
"time": 47.5 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0163e-05", |
|
"loss": 0.1296, |
|
"slid_loss": 0.1269, |
|
"step": 503, |
|
"time": 45.32 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "7.0060e-05", |
|
"loss": 0.1307, |
|
"slid_loss": 0.1269, |
|
"step": 504, |
|
"time": 45.23 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9956e-05", |
|
"loss": 0.1251, |
|
"slid_loss": 0.1268, |
|
"step": 505, |
|
"time": 47.33 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9852e-05", |
|
"loss": 0.1176, |
|
"slid_loss": 0.1268, |
|
"step": 506, |
|
"time": 47.45 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9748e-05", |
|
"loss": 0.1299, |
|
"slid_loss": 0.1267, |
|
"step": 507, |
|
"time": 48.54 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9645e-05", |
|
"loss": 0.132, |
|
"slid_loss": 0.1267, |
|
"step": 508, |
|
"time": 46.86 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9541e-05", |
|
"loss": 0.1166, |
|
"slid_loss": 0.1266, |
|
"step": 509, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9437e-05", |
|
"loss": 0.1306, |
|
"slid_loss": 0.1266, |
|
"step": 510, |
|
"time": 47.96 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9333e-05", |
|
"loss": 0.1281, |
|
"slid_loss": 0.1266, |
|
"step": 511, |
|
"time": 47.94 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9228e-05", |
|
"loss": 0.1199, |
|
"slid_loss": 0.1265, |
|
"step": 512, |
|
"time": 48.0 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9124e-05", |
|
"loss": 0.1222, |
|
"slid_loss": 0.1264, |
|
"step": 513, |
|
"time": 46.22 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.9020e-05", |
|
"loss": 0.1249, |
|
"slid_loss": 0.1264, |
|
"step": 514, |
|
"time": 46.94 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.8916e-05", |
|
"loss": 0.1193, |
|
"slid_loss": 0.1264, |
|
"step": 515, |
|
"time": 45.78 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "6.8811e-05", |
|
"loss": 0.1265, |
|
"slid_loss": 0.1264, |
|
"step": 516, |
|
"time": 45.7 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.8707e-05", |
|
"loss": 0.1284, |
|
"slid_loss": 0.1265, |
|
"step": 517, |
|
"time": 48.55 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.8602e-05", |
|
"loss": 0.123, |
|
"slid_loss": 0.1264, |
|
"step": 518, |
|
"time": 44.82 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.8498e-05", |
|
"loss": 0.1274, |
|
"slid_loss": 0.1264, |
|
"step": 519, |
|
"time": 48.3 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.8393e-05", |
|
"loss": 0.1292, |
|
"slid_loss": 0.1263, |
|
"step": 520, |
|
"time": 46.15 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.8289e-05", |
|
"loss": 0.1263, |
|
"slid_loss": 0.1262, |
|
"step": 521, |
|
"time": 46.29 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.8184e-05", |
|
"loss": 0.1273, |
|
"slid_loss": 0.1262, |
|
"step": 522, |
|
"time": 48.24 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.8079e-05", |
|
"loss": 0.1266, |
|
"slid_loss": 0.1262, |
|
"step": 523, |
|
"time": 46.75 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.7974e-05", |
|
"loss": 0.1346, |
|
"slid_loss": 0.1262, |
|
"step": 524, |
|
"time": 47.09 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.7869e-05", |
|
"loss": 0.1289, |
|
"slid_loss": 0.1262, |
|
"step": 525, |
|
"time": 49.67 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.7764e-05", |
|
"loss": 0.1255, |
|
"slid_loss": 0.1262, |
|
"step": 526, |
|
"time": 47.09 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.7659e-05", |
|
"loss": 0.127, |
|
"slid_loss": 0.1262, |
|
"step": 527, |
|
"time": 46.68 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "6.7554e-05", |
|
"loss": 0.1217, |
|
"slid_loss": 0.1262, |
|
"step": 528, |
|
"time": 47.36 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.7449e-05", |
|
"loss": 0.1274, |
|
"slid_loss": 0.1262, |
|
"step": 529, |
|
"time": 46.03 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.7344e-05", |
|
"loss": 0.1226, |
|
"slid_loss": 0.1262, |
|
"step": 530, |
|
"time": 48.25 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.7239e-05", |
|
"loss": 0.125, |
|
"slid_loss": 0.1261, |
|
"step": 531, |
|
"time": 48.33 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.7133e-05", |
|
"loss": 0.1205, |
|
"slid_loss": 0.126, |
|
"step": 532, |
|
"time": 47.15 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.7028e-05", |
|
"loss": 0.1273, |
|
"slid_loss": 0.1261, |
|
"step": 533, |
|
"time": 45.32 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.6923e-05", |
|
"loss": 0.1239, |
|
"slid_loss": 0.126, |
|
"step": 534, |
|
"time": 47.65 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.6817e-05", |
|
"loss": 0.1339, |
|
"slid_loss": 0.126, |
|
"step": 535, |
|
"time": 47.39 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.6712e-05", |
|
"loss": 0.1238, |
|
"slid_loss": 0.1259, |
|
"step": 536, |
|
"time": 46.21 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.6606e-05", |
|
"loss": 0.1249, |
|
"slid_loss": 0.1258, |
|
"step": 537, |
|
"time": 46.77 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.6500e-05", |
|
"loss": 0.1272, |
|
"slid_loss": 0.1258, |
|
"step": 538, |
|
"time": 47.54 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.6395e-05", |
|
"loss": 0.1276, |
|
"slid_loss": 0.1258, |
|
"step": 539, |
|
"time": 47.36 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "6.6289e-05", |
|
"loss": 0.1329, |
|
"slid_loss": 0.1258, |
|
"step": 540, |
|
"time": 46.69 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.6183e-05", |
|
"loss": 0.1278, |
|
"slid_loss": 0.1259, |
|
"step": 541, |
|
"time": 45.83 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.6078e-05", |
|
"loss": 0.1298, |
|
"slid_loss": 0.1259, |
|
"step": 542, |
|
"time": 46.98 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.5972e-05", |
|
"loss": 0.1241, |
|
"slid_loss": 0.1259, |
|
"step": 543, |
|
"time": 47.74 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.5866e-05", |
|
"loss": 0.1241, |
|
"slid_loss": 0.1259, |
|
"step": 544, |
|
"time": 46.3 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.5760e-05", |
|
"loss": 0.1269, |
|
"slid_loss": 0.126, |
|
"step": 545, |
|
"time": 45.15 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.5654e-05", |
|
"loss": 0.1277, |
|
"slid_loss": 0.126, |
|
"step": 546, |
|
"time": 47.36 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.5548e-05", |
|
"loss": 0.1294, |
|
"slid_loss": 0.1261, |
|
"step": 547, |
|
"time": 45.29 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.5442e-05", |
|
"loss": 0.1275, |
|
"slid_loss": 0.1261, |
|
"step": 548, |
|
"time": 45.9 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.5336e-05", |
|
"loss": 0.1264, |
|
"slid_loss": 0.126, |
|
"step": 549, |
|
"time": 46.54 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.5230e-05", |
|
"loss": 0.1259, |
|
"slid_loss": 0.126, |
|
"step": 550, |
|
"time": 45.18 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "6.5124e-05", |
|
"loss": 0.1221, |
|
"slid_loss": 0.1261, |
|
"step": 551, |
|
"time": 45.83 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.5018e-05", |
|
"loss": 0.1215, |
|
"slid_loss": 0.126, |
|
"step": 552, |
|
"time": 47.8 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.4911e-05", |
|
"loss": 0.1319, |
|
"slid_loss": 0.1262, |
|
"step": 553, |
|
"time": 45.78 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.4805e-05", |
|
"loss": 0.1266, |
|
"slid_loss": 0.1261, |
|
"step": 554, |
|
"time": 47.0 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.4699e-05", |
|
"loss": 0.1228, |
|
"slid_loss": 0.1261, |
|
"step": 555, |
|
"time": 48.07 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.4593e-05", |
|
"loss": 0.128, |
|
"slid_loss": 0.1261, |
|
"step": 556, |
|
"time": 48.73 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.4486e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.126, |
|
"step": 557, |
|
"time": 47.06 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.4380e-05", |
|
"loss": 0.1273, |
|
"slid_loss": 0.126, |
|
"step": 558, |
|
"time": 48.46 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.4273e-05", |
|
"loss": 0.1264, |
|
"slid_loss": 0.126, |
|
"step": 559, |
|
"time": 50.44 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.4167e-05", |
|
"loss": 0.1247, |
|
"slid_loss": 0.126, |
|
"step": 560, |
|
"time": 46.43 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.4060e-05", |
|
"loss": 0.1224, |
|
"slid_loss": 0.1259, |
|
"step": 561, |
|
"time": 46.38 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.3954e-05", |
|
"loss": 0.1213, |
|
"slid_loss": 0.1258, |
|
"step": 562, |
|
"time": 48.2 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "6.3847e-05", |
|
"loss": 0.1308, |
|
"slid_loss": 0.1259, |
|
"step": 563, |
|
"time": 46.5 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.3741e-05", |
|
"loss": 0.1247, |
|
"slid_loss": 0.1259, |
|
"step": 564, |
|
"time": 45.24 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.3634e-05", |
|
"loss": 0.1248, |
|
"slid_loss": 0.1259, |
|
"step": 565, |
|
"time": 45.23 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.3528e-05", |
|
"loss": 0.1262, |
|
"slid_loss": 0.1259, |
|
"step": 566, |
|
"time": 49.45 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.3421e-05", |
|
"loss": 0.1274, |
|
"slid_loss": 0.1259, |
|
"step": 567, |
|
"time": 48.02 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.3314e-05", |
|
"loss": 0.1203, |
|
"slid_loss": 0.1257, |
|
"step": 568, |
|
"time": 46.67 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.3208e-05", |
|
"loss": 0.1215, |
|
"slid_loss": 0.1258, |
|
"step": 569, |
|
"time": 46.92 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.3101e-05", |
|
"loss": 0.1261, |
|
"slid_loss": 0.1258, |
|
"step": 570, |
|
"time": 49.22 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.2994e-05", |
|
"loss": 0.1187, |
|
"slid_loss": 0.1257, |
|
"step": 571, |
|
"time": 44.64 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.2888e-05", |
|
"loss": 0.136, |
|
"slid_loss": 0.1259, |
|
"step": 572, |
|
"time": 48.45 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.2781e-05", |
|
"loss": 0.1301, |
|
"slid_loss": 0.1259, |
|
"step": 573, |
|
"time": 47.53 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.2674e-05", |
|
"loss": 0.1217, |
|
"slid_loss": 0.1258, |
|
"step": 574, |
|
"time": 47.97 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "6.2567e-05", |
|
"loss": 0.1187, |
|
"slid_loss": 0.1258, |
|
"step": 575, |
|
"time": 45.52 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.2460e-05", |
|
"loss": 0.1265, |
|
"slid_loss": 0.1258, |
|
"step": 576, |
|
"time": 46.06 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.2353e-05", |
|
"loss": 0.1167, |
|
"slid_loss": 0.1257, |
|
"step": 577, |
|
"time": 47.47 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.2247e-05", |
|
"loss": 0.1217, |
|
"slid_loss": 0.1256, |
|
"step": 578, |
|
"time": 46.4 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.2140e-05", |
|
"loss": 0.1251, |
|
"slid_loss": 0.1256, |
|
"step": 579, |
|
"time": 46.21 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.2033e-05", |
|
"loss": 0.1236, |
|
"slid_loss": 0.1255, |
|
"step": 580, |
|
"time": 46.81 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.1926e-05", |
|
"loss": 0.1212, |
|
"slid_loss": 0.1255, |
|
"step": 581, |
|
"time": 47.14 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.1819e-05", |
|
"loss": 0.126, |
|
"slid_loss": 0.1255, |
|
"step": 582, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.1712e-05", |
|
"loss": 0.1253, |
|
"slid_loss": 0.1256, |
|
"step": 583, |
|
"time": 47.18 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.1605e-05", |
|
"loss": 0.1245, |
|
"slid_loss": 0.1256, |
|
"step": 584, |
|
"time": 46.46 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.1498e-05", |
|
"loss": 0.1212, |
|
"slid_loss": 0.1255, |
|
"step": 585, |
|
"time": 46.98 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.1391e-05", |
|
"loss": 0.1174, |
|
"slid_loss": 0.1254, |
|
"step": 586, |
|
"time": 46.05 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "6.1284e-05", |
|
"loss": 0.1298, |
|
"slid_loss": 0.1254, |
|
"step": 587, |
|
"time": 48.71 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.1177e-05", |
|
"loss": 0.128, |
|
"slid_loss": 0.1254, |
|
"step": 588, |
|
"time": 45.56 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.1070e-05", |
|
"loss": 0.1196, |
|
"slid_loss": 0.1253, |
|
"step": 589, |
|
"time": 47.79 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0963e-05", |
|
"loss": 0.1289, |
|
"slid_loss": 0.1253, |
|
"step": 590, |
|
"time": 48.66 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0856e-05", |
|
"loss": 0.1251, |
|
"slid_loss": 0.1253, |
|
"step": 591, |
|
"time": 45.29 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0749e-05", |
|
"loss": 0.121, |
|
"slid_loss": 0.1253, |
|
"step": 592, |
|
"time": 47.72 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0642e-05", |
|
"loss": 0.122, |
|
"slid_loss": 0.1253, |
|
"step": 593, |
|
"time": 45.48 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0535e-05", |
|
"loss": 0.1303, |
|
"slid_loss": 0.1253, |
|
"step": 594, |
|
"time": 45.98 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0428e-05", |
|
"loss": 0.113, |
|
"slid_loss": 0.1252, |
|
"step": 595, |
|
"time": 45.64 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0321e-05", |
|
"loss": 0.125, |
|
"slid_loss": 0.1252, |
|
"step": 596, |
|
"time": 47.26 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0214e-05", |
|
"loss": 0.1247, |
|
"slid_loss": 0.1252, |
|
"step": 597, |
|
"time": 49.22 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0107e-05", |
|
"loss": 0.1232, |
|
"slid_loss": 0.1252, |
|
"step": 598, |
|
"time": 46.48 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "6.0000e-05", |
|
"loss": 0.1211, |
|
"slid_loss": 0.1251, |
|
"step": 599, |
|
"time": 47.23 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.9893e-05", |
|
"loss": 0.1148, |
|
"slid_loss": 0.1251, |
|
"step": 600, |
|
"time": 48.26 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.9786e-05", |
|
"loss": 0.1208, |
|
"slid_loss": 0.125, |
|
"step": 601, |
|
"time": 149.89 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.9679e-05", |
|
"loss": 0.123, |
|
"slid_loss": 0.1251, |
|
"step": 602, |
|
"time": 47.96 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.9572e-05", |
|
"loss": 0.1151, |
|
"slid_loss": 0.1249, |
|
"step": 603, |
|
"time": 46.82 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.9465e-05", |
|
"loss": 0.1222, |
|
"slid_loss": 0.1248, |
|
"step": 604, |
|
"time": 46.66 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.9358e-05", |
|
"loss": 0.121, |
|
"slid_loss": 0.1248, |
|
"step": 605, |
|
"time": 46.25 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.9251e-05", |
|
"loss": 0.1235, |
|
"slid_loss": 0.1248, |
|
"step": 606, |
|
"time": 48.74 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.9144e-05", |
|
"loss": 0.1177, |
|
"slid_loss": 0.1247, |
|
"step": 607, |
|
"time": 48.81 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.9037e-05", |
|
"loss": 0.1214, |
|
"slid_loss": 0.1246, |
|
"step": 608, |
|
"time": 46.74 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.8930e-05", |
|
"loss": 0.122, |
|
"slid_loss": 0.1247, |
|
"step": 609, |
|
"time": 44.76 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.8823e-05", |
|
"loss": 0.1243, |
|
"slid_loss": 0.1246, |
|
"step": 610, |
|
"time": 45.94 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "5.8716e-05", |
|
"loss": 0.1242, |
|
"slid_loss": 0.1246, |
|
"step": 611, |
|
"time": 45.59 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.8609e-05", |
|
"loss": 0.125, |
|
"slid_loss": 0.1246, |
|
"step": 612, |
|
"time": 45.63 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.8502e-05", |
|
"loss": 0.1277, |
|
"slid_loss": 0.1247, |
|
"step": 613, |
|
"time": 47.56 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.8395e-05", |
|
"loss": 0.1183, |
|
"slid_loss": 0.1246, |
|
"step": 614, |
|
"time": 44.87 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.8288e-05", |
|
"loss": 0.121, |
|
"slid_loss": 0.1246, |
|
"step": 615, |
|
"time": 46.38 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.8181e-05", |
|
"loss": 0.1222, |
|
"slid_loss": 0.1246, |
|
"step": 616, |
|
"time": 46.97 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.8074e-05", |
|
"loss": 0.1291, |
|
"slid_loss": 0.1246, |
|
"step": 617, |
|
"time": 45.82 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.7967e-05", |
|
"loss": 0.1268, |
|
"slid_loss": 0.1246, |
|
"step": 618, |
|
"time": 46.08 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.7860e-05", |
|
"loss": 0.1165, |
|
"slid_loss": 0.1245, |
|
"step": 619, |
|
"time": 47.16 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.7753e-05", |
|
"loss": 0.1251, |
|
"slid_loss": 0.1245, |
|
"step": 620, |
|
"time": 46.75 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.7647e-05", |
|
"loss": 0.1315, |
|
"slid_loss": 0.1245, |
|
"step": 621, |
|
"time": 49.53 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.7540e-05", |
|
"loss": 0.1285, |
|
"slid_loss": 0.1245, |
|
"step": 622, |
|
"time": 46.15 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "5.7433e-05", |
|
"loss": 0.1199, |
|
"slid_loss": 0.1245, |
|
"step": 623, |
|
"time": 47.39 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.7326e-05", |
|
"loss": 0.1154, |
|
"slid_loss": 0.1243, |
|
"step": 624, |
|
"time": 45.78 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.7219e-05", |
|
"loss": 0.1209, |
|
"slid_loss": 0.1242, |
|
"step": 625, |
|
"time": 44.79 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.7112e-05", |
|
"loss": 0.1168, |
|
"slid_loss": 0.1241, |
|
"step": 626, |
|
"time": 46.43 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.7006e-05", |
|
"loss": 0.1247, |
|
"slid_loss": 0.1241, |
|
"step": 627, |
|
"time": 47.42 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.6899e-05", |
|
"loss": 0.1314, |
|
"slid_loss": 0.1242, |
|
"step": 628, |
|
"time": 46.0 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.6792e-05", |
|
"loss": 0.1234, |
|
"slid_loss": 0.1242, |
|
"step": 629, |
|
"time": 45.57 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.6686e-05", |
|
"loss": 0.1255, |
|
"slid_loss": 0.1242, |
|
"step": 630, |
|
"time": 44.54 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.6579e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1241, |
|
"step": 631, |
|
"time": 47.12 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.6472e-05", |
|
"loss": 0.1203, |
|
"slid_loss": 0.1241, |
|
"step": 632, |
|
"time": 47.42 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.6366e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1241, |
|
"step": 633, |
|
"time": 47.23 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "5.6259e-05", |
|
"loss": 0.1206, |
|
"slid_loss": 0.124, |
|
"step": 634, |
|
"time": 46.73 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.6153e-05", |
|
"loss": 0.1246, |
|
"slid_loss": 0.1239, |
|
"step": 635, |
|
"time": 47.57 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.6046e-05", |
|
"loss": 0.1186, |
|
"slid_loss": 0.1239, |
|
"step": 636, |
|
"time": 48.39 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.5940e-05", |
|
"loss": 0.1215, |
|
"slid_loss": 0.1238, |
|
"step": 637, |
|
"time": 46.95 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.5833e-05", |
|
"loss": 0.1255, |
|
"slid_loss": 0.1238, |
|
"step": 638, |
|
"time": 47.03 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.5727e-05", |
|
"loss": 0.1156, |
|
"slid_loss": 0.1237, |
|
"step": 639, |
|
"time": 47.87 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.5620e-05", |
|
"loss": 0.1206, |
|
"slid_loss": 0.1236, |
|
"step": 640, |
|
"time": 46.27 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.5514e-05", |
|
"loss": 0.1207, |
|
"slid_loss": 0.1235, |
|
"step": 641, |
|
"time": 45.77 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.5407e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1234, |
|
"step": 642, |
|
"time": 51.08 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.5301e-05", |
|
"loss": 0.1256, |
|
"slid_loss": 0.1234, |
|
"step": 643, |
|
"time": 46.26 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.5195e-05", |
|
"loss": 0.1149, |
|
"slid_loss": 0.1233, |
|
"step": 644, |
|
"time": 45.48 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.5089e-05", |
|
"loss": 0.1259, |
|
"slid_loss": 0.1233, |
|
"step": 645, |
|
"time": 45.26 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "5.4982e-05", |
|
"loss": 0.1189, |
|
"slid_loss": 0.1232, |
|
"step": 646, |
|
"time": 45.44 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.4876e-05", |
|
"loss": 0.1154, |
|
"slid_loss": 0.1231, |
|
"step": 647, |
|
"time": 46.8 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.4770e-05", |
|
"loss": 0.1219, |
|
"slid_loss": 0.123, |
|
"step": 648, |
|
"time": 46.87 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.4664e-05", |
|
"loss": 0.1133, |
|
"slid_loss": 0.1229, |
|
"step": 649, |
|
"time": 47.42 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.4558e-05", |
|
"loss": 0.1176, |
|
"slid_loss": 0.1228, |
|
"step": 650, |
|
"time": 47.24 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.4452e-05", |
|
"loss": 0.1164, |
|
"slid_loss": 0.1228, |
|
"step": 651, |
|
"time": 49.44 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.4346e-05", |
|
"loss": 0.1243, |
|
"slid_loss": 0.1228, |
|
"step": 652, |
|
"time": 45.62 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.4240e-05", |
|
"loss": 0.1215, |
|
"slid_loss": 0.1227, |
|
"step": 653, |
|
"time": 46.94 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.4134e-05", |
|
"loss": 0.1268, |
|
"slid_loss": 0.1227, |
|
"step": 654, |
|
"time": 45.39 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.4028e-05", |
|
"loss": 0.1211, |
|
"slid_loss": 0.1227, |
|
"step": 655, |
|
"time": 46.99 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.3922e-05", |
|
"loss": 0.1201, |
|
"slid_loss": 0.1226, |
|
"step": 656, |
|
"time": 48.14 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.3817e-05", |
|
"loss": 0.1246, |
|
"slid_loss": 0.1227, |
|
"step": 657, |
|
"time": 47.39 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "5.3711e-05", |
|
"loss": 0.123, |
|
"slid_loss": 0.1226, |
|
"step": 658, |
|
"time": 46.64 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.3605e-05", |
|
"loss": 0.1141, |
|
"slid_loss": 0.1225, |
|
"step": 659, |
|
"time": 46.64 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.3500e-05", |
|
"loss": 0.1209, |
|
"slid_loss": 0.1225, |
|
"step": 660, |
|
"time": 45.91 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.3394e-05", |
|
"loss": 0.1208, |
|
"slid_loss": 0.1224, |
|
"step": 661, |
|
"time": 46.43 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.3288e-05", |
|
"loss": 0.1346, |
|
"slid_loss": 0.1226, |
|
"step": 662, |
|
"time": 46.66 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.3183e-05", |
|
"loss": 0.1216, |
|
"slid_loss": 0.1225, |
|
"step": 663, |
|
"time": 46.06 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.3077e-05", |
|
"loss": 0.117, |
|
"slid_loss": 0.1224, |
|
"step": 664, |
|
"time": 46.37 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.2972e-05", |
|
"loss": 0.1258, |
|
"slid_loss": 0.1224, |
|
"step": 665, |
|
"time": 46.81 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.2867e-05", |
|
"loss": 0.1208, |
|
"slid_loss": 0.1224, |
|
"step": 666, |
|
"time": 46.32 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.2761e-05", |
|
"loss": 0.1225, |
|
"slid_loss": 0.1223, |
|
"step": 667, |
|
"time": 46.81 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.2656e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1223, |
|
"step": 668, |
|
"time": 45.87 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.2551e-05", |
|
"loss": 0.1213, |
|
"slid_loss": 0.1223, |
|
"step": 669, |
|
"time": 46.03 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "5.2446e-05", |
|
"loss": 0.1239, |
|
"slid_loss": 0.1223, |
|
"step": 670, |
|
"time": 45.12 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.2341e-05", |
|
"loss": 0.125, |
|
"slid_loss": 0.1223, |
|
"step": 671, |
|
"time": 44.65 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.2236e-05", |
|
"loss": 0.1196, |
|
"slid_loss": 0.1222, |
|
"step": 672, |
|
"time": 44.51 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.2131e-05", |
|
"loss": 0.1246, |
|
"slid_loss": 0.1221, |
|
"step": 673, |
|
"time": 46.49 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.2026e-05", |
|
"loss": 0.117, |
|
"slid_loss": 0.1221, |
|
"step": 674, |
|
"time": 45.56 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.1921e-05", |
|
"loss": 0.1153, |
|
"slid_loss": 0.122, |
|
"step": 675, |
|
"time": 45.52 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.1816e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.122, |
|
"step": 676, |
|
"time": 47.11 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.1711e-05", |
|
"loss": 0.1246, |
|
"slid_loss": 0.1221, |
|
"step": 677, |
|
"time": 45.87 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.1607e-05", |
|
"loss": 0.1212, |
|
"slid_loss": 0.122, |
|
"step": 678, |
|
"time": 44.99 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.1502e-05", |
|
"loss": 0.1171, |
|
"slid_loss": 0.122, |
|
"step": 679, |
|
"time": 46.24 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.1398e-05", |
|
"loss": 0.1215, |
|
"slid_loss": 0.1219, |
|
"step": 680, |
|
"time": 47.35 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.1293e-05", |
|
"loss": 0.1118, |
|
"slid_loss": 0.1219, |
|
"step": 681, |
|
"time": 48.56 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "5.1189e-05", |
|
"loss": 0.1268, |
|
"slid_loss": 0.1219, |
|
"step": 682, |
|
"time": 46.27 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.1084e-05", |
|
"loss": 0.1274, |
|
"slid_loss": 0.1219, |
|
"step": 683, |
|
"time": 49.49 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0980e-05", |
|
"loss": 0.1222, |
|
"slid_loss": 0.1219, |
|
"step": 684, |
|
"time": 45.26 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0876e-05", |
|
"loss": 0.1213, |
|
"slid_loss": 0.1219, |
|
"step": 685, |
|
"time": 47.56 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0772e-05", |
|
"loss": 0.1239, |
|
"slid_loss": 0.1219, |
|
"step": 686, |
|
"time": 47.89 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0667e-05", |
|
"loss": 0.1151, |
|
"slid_loss": 0.1218, |
|
"step": 687, |
|
"time": 45.36 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0563e-05", |
|
"loss": 0.1102, |
|
"slid_loss": 0.1216, |
|
"step": 688, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0459e-05", |
|
"loss": 0.1269, |
|
"slid_loss": 0.1217, |
|
"step": 689, |
|
"time": 49.43 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0355e-05", |
|
"loss": 0.1225, |
|
"slid_loss": 0.1216, |
|
"step": 690, |
|
"time": 46.27 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0252e-05", |
|
"loss": 0.121, |
|
"slid_loss": 0.1216, |
|
"step": 691, |
|
"time": 47.47 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0148e-05", |
|
"loss": 0.1188, |
|
"slid_loss": 0.1215, |
|
"step": 692, |
|
"time": 47.74 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "5.0044e-05", |
|
"loss": 0.1165, |
|
"slid_loss": 0.1215, |
|
"step": 693, |
|
"time": 46.68 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "4.9940e-05", |
|
"loss": 0.117, |
|
"slid_loss": 0.1214, |
|
"step": 694, |
|
"time": 48.16 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.9837e-05", |
|
"loss": 0.1207, |
|
"slid_loss": 0.1214, |
|
"step": 695, |
|
"time": 45.3 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.9733e-05", |
|
"loss": 0.1185, |
|
"slid_loss": 0.1214, |
|
"step": 696, |
|
"time": 45.82 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.9630e-05", |
|
"loss": 0.125, |
|
"slid_loss": 0.1214, |
|
"step": 697, |
|
"time": 47.56 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.9527e-05", |
|
"loss": 0.1246, |
|
"slid_loss": 0.1214, |
|
"step": 698, |
|
"time": 48.19 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.9423e-05", |
|
"loss": 0.1222, |
|
"slid_loss": 0.1214, |
|
"step": 699, |
|
"time": 47.65 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.9320e-05", |
|
"loss": 0.1267, |
|
"slid_loss": 0.1215, |
|
"step": 700, |
|
"time": 48.47 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.9217e-05", |
|
"loss": 0.1222, |
|
"slid_loss": 0.1215, |
|
"step": 701, |
|
"time": 46.84 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.9114e-05", |
|
"loss": 0.118, |
|
"slid_loss": 0.1215, |
|
"step": 702, |
|
"time": 45.36 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.9011e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1215, |
|
"step": 703, |
|
"time": 46.8 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.8908e-05", |
|
"loss": 0.1215, |
|
"slid_loss": 0.1215, |
|
"step": 704, |
|
"time": 47.75 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.8805e-05", |
|
"loss": 0.1272, |
|
"slid_loss": 0.1216, |
|
"step": 705, |
|
"time": 47.34 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "4.8703e-05", |
|
"loss": 0.1251, |
|
"slid_loss": 0.1216, |
|
"step": 706, |
|
"time": 46.05 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.8600e-05", |
|
"loss": 0.123, |
|
"slid_loss": 0.1216, |
|
"step": 707, |
|
"time": 45.38 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.8497e-05", |
|
"loss": 0.1223, |
|
"slid_loss": 0.1216, |
|
"step": 708, |
|
"time": 44.58 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.8395e-05", |
|
"loss": 0.1112, |
|
"slid_loss": 0.1215, |
|
"step": 709, |
|
"time": 46.83 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.8293e-05", |
|
"loss": 0.1168, |
|
"slid_loss": 0.1215, |
|
"step": 710, |
|
"time": 45.9 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.8190e-05", |
|
"loss": 0.1194, |
|
"slid_loss": 0.1214, |
|
"step": 711, |
|
"time": 47.86 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.8088e-05", |
|
"loss": 0.1167, |
|
"slid_loss": 0.1213, |
|
"step": 712, |
|
"time": 45.71 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.7986e-05", |
|
"loss": 0.1279, |
|
"slid_loss": 0.1213, |
|
"step": 713, |
|
"time": 48.04 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.7884e-05", |
|
"loss": 0.1299, |
|
"slid_loss": 0.1214, |
|
"step": 714, |
|
"time": 46.67 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.7782e-05", |
|
"loss": 0.1237, |
|
"slid_loss": 0.1215, |
|
"step": 715, |
|
"time": 46.35 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.7680e-05", |
|
"loss": 0.1284, |
|
"slid_loss": 0.1215, |
|
"step": 716, |
|
"time": 47.04 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.7578e-05", |
|
"loss": 0.117, |
|
"slid_loss": 0.1214, |
|
"step": 717, |
|
"time": 46.39 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "4.7477e-05", |
|
"loss": 0.1185, |
|
"slid_loss": 0.1213, |
|
"step": 718, |
|
"time": 45.73 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.7375e-05", |
|
"loss": 0.1186, |
|
"slid_loss": 0.1213, |
|
"step": 719, |
|
"time": 45.73 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.7273e-05", |
|
"loss": 0.1237, |
|
"slid_loss": 0.1213, |
|
"step": 720, |
|
"time": 46.84 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.7172e-05", |
|
"loss": 0.1163, |
|
"slid_loss": 0.1212, |
|
"step": 721, |
|
"time": 46.18 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.7071e-05", |
|
"loss": 0.1197, |
|
"slid_loss": 0.1211, |
|
"step": 722, |
|
"time": 44.48 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.6969e-05", |
|
"loss": 0.1177, |
|
"slid_loss": 0.1211, |
|
"step": 723, |
|
"time": 46.24 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.6868e-05", |
|
"loss": 0.1255, |
|
"slid_loss": 0.1212, |
|
"step": 724, |
|
"time": 47.96 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.6767e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1212, |
|
"step": 725, |
|
"time": 46.05 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.6666e-05", |
|
"loss": 0.1178, |
|
"slid_loss": 0.1212, |
|
"step": 726, |
|
"time": 45.39 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.6565e-05", |
|
"loss": 0.1126, |
|
"slid_loss": 0.121, |
|
"step": 727, |
|
"time": 46.66 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.6465e-05", |
|
"loss": 0.1126, |
|
"slid_loss": 0.1209, |
|
"step": 728, |
|
"time": 45.38 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "4.6364e-05", |
|
"loss": 0.125, |
|
"slid_loss": 0.1209, |
|
"step": 729, |
|
"time": 48.38 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.6263e-05", |
|
"loss": 0.116, |
|
"slid_loss": 0.1208, |
|
"step": 730, |
|
"time": 46.89 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.6163e-05", |
|
"loss": 0.1305, |
|
"slid_loss": 0.1209, |
|
"step": 731, |
|
"time": 44.79 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.6062e-05", |
|
"loss": 0.1116, |
|
"slid_loss": 0.1208, |
|
"step": 732, |
|
"time": 45.82 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.5962e-05", |
|
"loss": 0.1205, |
|
"slid_loss": 0.1208, |
|
"step": 733, |
|
"time": 46.2 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.5862e-05", |
|
"loss": 0.1163, |
|
"slid_loss": 0.1208, |
|
"step": 734, |
|
"time": 46.33 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.5762e-05", |
|
"loss": 0.1203, |
|
"slid_loss": 0.1207, |
|
"step": 735, |
|
"time": 45.73 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.5662e-05", |
|
"loss": 0.1288, |
|
"slid_loss": 0.1208, |
|
"step": 736, |
|
"time": 48.51 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.5562e-05", |
|
"loss": 0.1177, |
|
"slid_loss": 0.1208, |
|
"step": 737, |
|
"time": 47.25 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.5462e-05", |
|
"loss": 0.1298, |
|
"slid_loss": 0.1208, |
|
"step": 738, |
|
"time": 47.49 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.5363e-05", |
|
"loss": 0.1146, |
|
"slid_loss": 0.1208, |
|
"step": 739, |
|
"time": 45.81 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.5263e-05", |
|
"loss": 0.1247, |
|
"slid_loss": 0.1209, |
|
"step": 740, |
|
"time": 47.21 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "4.5164e-05", |
|
"loss": 0.1176, |
|
"slid_loss": 0.1208, |
|
"step": 741, |
|
"time": 46.84 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.5064e-05", |
|
"loss": 0.115, |
|
"slid_loss": 0.1208, |
|
"step": 742, |
|
"time": 48.34 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4965e-05", |
|
"loss": 0.1216, |
|
"slid_loss": 0.1207, |
|
"step": 743, |
|
"time": 45.92 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4866e-05", |
|
"loss": 0.1132, |
|
"slid_loss": 0.1207, |
|
"step": 744, |
|
"time": 46.59 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4767e-05", |
|
"loss": 0.1124, |
|
"slid_loss": 0.1206, |
|
"step": 745, |
|
"time": 46.28 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4668e-05", |
|
"loss": 0.1197, |
|
"slid_loss": 0.1206, |
|
"step": 746, |
|
"time": 48.43 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4569e-05", |
|
"loss": 0.122, |
|
"slid_loss": 0.1207, |
|
"step": 747, |
|
"time": 47.84 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4470e-05", |
|
"loss": 0.124, |
|
"slid_loss": 0.1207, |
|
"step": 748, |
|
"time": 45.65 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4372e-05", |
|
"loss": 0.1226, |
|
"slid_loss": 0.1208, |
|
"step": 749, |
|
"time": 46.86 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4273e-05", |
|
"loss": 0.1188, |
|
"slid_loss": 0.1208, |
|
"step": 750, |
|
"time": 46.47 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4175e-05", |
|
"loss": 0.1252, |
|
"slid_loss": 0.1209, |
|
"step": 751, |
|
"time": 44.67 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.4077e-05", |
|
"loss": 0.1216, |
|
"slid_loss": 0.1209, |
|
"step": 752, |
|
"time": 46.69 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "4.3979e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1208, |
|
"step": 753, |
|
"time": 47.98 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3881e-05", |
|
"loss": 0.1198, |
|
"slid_loss": 0.1208, |
|
"step": 754, |
|
"time": 46.17 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3783e-05", |
|
"loss": 0.1252, |
|
"slid_loss": 0.1208, |
|
"step": 755, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3685e-05", |
|
"loss": 0.1213, |
|
"slid_loss": 0.1208, |
|
"step": 756, |
|
"time": 47.16 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3587e-05", |
|
"loss": 0.1183, |
|
"slid_loss": 0.1207, |
|
"step": 757, |
|
"time": 46.91 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3490e-05", |
|
"loss": 0.1112, |
|
"slid_loss": 0.1206, |
|
"step": 758, |
|
"time": 46.41 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3392e-05", |
|
"loss": 0.1228, |
|
"slid_loss": 0.1207, |
|
"step": 759, |
|
"time": 48.0 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3295e-05", |
|
"loss": 0.1201, |
|
"slid_loss": 0.1207, |
|
"step": 760, |
|
"time": 47.42 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3198e-05", |
|
"loss": 0.1221, |
|
"slid_loss": 0.1207, |
|
"step": 761, |
|
"time": 46.62 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3101e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1206, |
|
"step": 762, |
|
"time": 50.4 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.3004e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1205, |
|
"step": 763, |
|
"time": 47.21 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.2907e-05", |
|
"loss": 0.118, |
|
"slid_loss": 0.1205, |
|
"step": 764, |
|
"time": 46.38 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "4.2810e-05", |
|
"loss": 0.1144, |
|
"slid_loss": 0.1204, |
|
"step": 765, |
|
"time": 47.06 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.2714e-05", |
|
"loss": 0.1279, |
|
"slid_loss": 0.1205, |
|
"step": 766, |
|
"time": 44.95 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.2617e-05", |
|
"loss": 0.1109, |
|
"slid_loss": 0.1204, |
|
"step": 767, |
|
"time": 47.12 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.2521e-05", |
|
"loss": 0.1142, |
|
"slid_loss": 0.1203, |
|
"step": 768, |
|
"time": 49.81 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.2425e-05", |
|
"loss": 0.117, |
|
"slid_loss": 0.1203, |
|
"step": 769, |
|
"time": 47.69 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.2328e-05", |
|
"loss": 0.1216, |
|
"slid_loss": 0.1203, |
|
"step": 770, |
|
"time": 46.84 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.2232e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1202, |
|
"step": 771, |
|
"time": 47.39 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.2137e-05", |
|
"loss": 0.1248, |
|
"slid_loss": 0.1203, |
|
"step": 772, |
|
"time": 47.73 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.2041e-05", |
|
"loss": 0.1276, |
|
"slid_loss": 0.1203, |
|
"step": 773, |
|
"time": 45.08 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.1945e-05", |
|
"loss": 0.1153, |
|
"slid_loss": 0.1203, |
|
"step": 774, |
|
"time": 46.15 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.1850e-05", |
|
"loss": 0.1159, |
|
"slid_loss": 0.1203, |
|
"step": 775, |
|
"time": 46.06 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.1755e-05", |
|
"loss": 0.1206, |
|
"slid_loss": 0.1203, |
|
"step": 776, |
|
"time": 45.24 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "4.1659e-05", |
|
"loss": 0.1253, |
|
"slid_loss": 0.1203, |
|
"step": 777, |
|
"time": 45.69 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.1564e-05", |
|
"loss": 0.1191, |
|
"slid_loss": 0.1203, |
|
"step": 778, |
|
"time": 46.7 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.1469e-05", |
|
"loss": 0.1154, |
|
"slid_loss": 0.1203, |
|
"step": 779, |
|
"time": 45.64 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.1375e-05", |
|
"loss": 0.1267, |
|
"slid_loss": 0.1203, |
|
"step": 780, |
|
"time": 46.72 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.1280e-05", |
|
"loss": 0.1186, |
|
"slid_loss": 0.1204, |
|
"step": 781, |
|
"time": 46.05 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.1185e-05", |
|
"loss": 0.1162, |
|
"slid_loss": 0.1203, |
|
"step": 782, |
|
"time": 46.63 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.1091e-05", |
|
"loss": 0.1171, |
|
"slid_loss": 0.1202, |
|
"step": 783, |
|
"time": 46.76 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.0997e-05", |
|
"loss": 0.1177, |
|
"slid_loss": 0.1201, |
|
"step": 784, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.0903e-05", |
|
"loss": 0.1198, |
|
"slid_loss": 0.1201, |
|
"step": 785, |
|
"time": 47.51 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.0809e-05", |
|
"loss": 0.1217, |
|
"slid_loss": 0.1201, |
|
"step": 786, |
|
"time": 47.21 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.0715e-05", |
|
"loss": 0.1238, |
|
"slid_loss": 0.1202, |
|
"step": 787, |
|
"time": 45.95 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.0621e-05", |
|
"loss": 0.1216, |
|
"slid_loss": 0.1203, |
|
"step": 788, |
|
"time": 45.33 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "4.0528e-05", |
|
"loss": 0.114, |
|
"slid_loss": 0.1202, |
|
"step": 789, |
|
"time": 46.59 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "4.0434e-05", |
|
"loss": 0.1195, |
|
"slid_loss": 0.1201, |
|
"step": 790, |
|
"time": 47.54 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "4.0341e-05", |
|
"loss": 0.1215, |
|
"slid_loss": 0.1201, |
|
"step": 791, |
|
"time": 50.13 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "4.0248e-05", |
|
"loss": 0.1212, |
|
"slid_loss": 0.1202, |
|
"step": 792, |
|
"time": 44.09 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "4.0155e-05", |
|
"loss": 0.1137, |
|
"slid_loss": 0.1201, |
|
"step": 793, |
|
"time": 46.41 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "4.0062e-05", |
|
"loss": 0.1187, |
|
"slid_loss": 0.1202, |
|
"step": 794, |
|
"time": 46.21 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "3.9969e-05", |
|
"loss": 0.1217, |
|
"slid_loss": 0.1202, |
|
"step": 795, |
|
"time": 45.73 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "3.9877e-05", |
|
"loss": 0.1254, |
|
"slid_loss": 0.1202, |
|
"step": 796, |
|
"time": 44.63 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "3.9784e-05", |
|
"loss": 0.1267, |
|
"slid_loss": 0.1203, |
|
"step": 797, |
|
"time": 45.66 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "3.9692e-05", |
|
"loss": 0.1231, |
|
"slid_loss": 0.1202, |
|
"step": 798, |
|
"time": 47.52 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "3.9600e-05", |
|
"loss": 0.1205, |
|
"slid_loss": 0.1202, |
|
"step": 799, |
|
"time": 47.11 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "3.9508e-05", |
|
"loss": 0.1149, |
|
"slid_loss": 0.1201, |
|
"step": 800, |
|
"time": 46.8 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "3.9416e-05", |
|
"loss": 0.118, |
|
"slid_loss": 0.1201, |
|
"step": 801, |
|
"time": 47.67 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.9324e-05", |
|
"loss": 0.1202, |
|
"slid_loss": 0.1201, |
|
"step": 802, |
|
"time": 46.56 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.9233e-05", |
|
"loss": 0.1128, |
|
"slid_loss": 0.12, |
|
"step": 803, |
|
"time": 47.23 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.9141e-05", |
|
"loss": 0.1214, |
|
"slid_loss": 0.12, |
|
"step": 804, |
|
"time": 46.44 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.9050e-05", |
|
"loss": 0.1283, |
|
"slid_loss": 0.12, |
|
"step": 805, |
|
"time": 47.95 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.8959e-05", |
|
"loss": 0.1264, |
|
"slid_loss": 0.1201, |
|
"step": 806, |
|
"time": 46.54 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.8868e-05", |
|
"loss": 0.1185, |
|
"slid_loss": 0.12, |
|
"step": 807, |
|
"time": 46.57 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.8777e-05", |
|
"loss": 0.1118, |
|
"slid_loss": 0.1199, |
|
"step": 808, |
|
"time": 45.26 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.8686e-05", |
|
"loss": 0.1137, |
|
"slid_loss": 0.1199, |
|
"step": 809, |
|
"time": 46.26 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.8596e-05", |
|
"loss": 0.1226, |
|
"slid_loss": 0.12, |
|
"step": 810, |
|
"time": 48.42 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.8506e-05", |
|
"loss": 0.1186, |
|
"slid_loss": 0.12, |
|
"step": 811, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.8415e-05", |
|
"loss": 0.1229, |
|
"slid_loss": 0.12, |
|
"step": 812, |
|
"time": 46.66 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "3.8325e-05", |
|
"loss": 0.1102, |
|
"slid_loss": 0.1199, |
|
"step": 813, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.8235e-05", |
|
"loss": 0.1159, |
|
"slid_loss": 0.1197, |
|
"step": 814, |
|
"time": 47.39 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.8146e-05", |
|
"loss": 0.1204, |
|
"slid_loss": 0.1197, |
|
"step": 815, |
|
"time": 45.79 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.8056e-05", |
|
"loss": 0.1153, |
|
"slid_loss": 0.1196, |
|
"step": 816, |
|
"time": 46.22 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.7967e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1196, |
|
"step": 817, |
|
"time": 47.15 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.7877e-05", |
|
"loss": 0.1149, |
|
"slid_loss": 0.1196, |
|
"step": 818, |
|
"time": 46.56 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.7788e-05", |
|
"loss": 0.1203, |
|
"slid_loss": 0.1196, |
|
"step": 819, |
|
"time": 47.96 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.7699e-05", |
|
"loss": 0.1165, |
|
"slid_loss": 0.1195, |
|
"step": 820, |
|
"time": 44.85 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.7611e-05", |
|
"loss": 0.1162, |
|
"slid_loss": 0.1195, |
|
"step": 821, |
|
"time": 48.31 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.7522e-05", |
|
"loss": 0.1112, |
|
"slid_loss": 0.1194, |
|
"step": 822, |
|
"time": 46.11 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.7434e-05", |
|
"loss": 0.1156, |
|
"slid_loss": 0.1194, |
|
"step": 823, |
|
"time": 46.89 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "3.7345e-05", |
|
"loss": 0.1138, |
|
"slid_loss": 0.1193, |
|
"step": 824, |
|
"time": 46.08 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.7257e-05", |
|
"loss": 0.116, |
|
"slid_loss": 0.1192, |
|
"step": 825, |
|
"time": 51.03 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.7169e-05", |
|
"loss": 0.1181, |
|
"slid_loss": 0.1192, |
|
"step": 826, |
|
"time": 47.67 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.7081e-05", |
|
"loss": 0.1187, |
|
"slid_loss": 0.1193, |
|
"step": 827, |
|
"time": 47.69 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.6994e-05", |
|
"loss": 0.1124, |
|
"slid_loss": 0.1193, |
|
"step": 828, |
|
"time": 47.49 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.6906e-05", |
|
"loss": 0.1199, |
|
"slid_loss": 0.1192, |
|
"step": 829, |
|
"time": 45.32 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.6819e-05", |
|
"loss": 0.1088, |
|
"slid_loss": 0.1192, |
|
"step": 830, |
|
"time": 46.17 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.6732e-05", |
|
"loss": 0.1143, |
|
"slid_loss": 0.119, |
|
"step": 831, |
|
"time": 47.72 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.6645e-05", |
|
"loss": 0.1212, |
|
"slid_loss": 0.1191, |
|
"step": 832, |
|
"time": 47.58 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.6558e-05", |
|
"loss": 0.1204, |
|
"slid_loss": 0.1191, |
|
"step": 833, |
|
"time": 44.6 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.6471e-05", |
|
"loss": 0.1142, |
|
"slid_loss": 0.1191, |
|
"step": 834, |
|
"time": 47.64 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.6385e-05", |
|
"loss": 0.119, |
|
"slid_loss": 0.1191, |
|
"step": 835, |
|
"time": 45.21 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "3.6298e-05", |
|
"loss": 0.1112, |
|
"slid_loss": 0.1189, |
|
"step": 836, |
|
"time": 46.24 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.6212e-05", |
|
"loss": 0.1157, |
|
"slid_loss": 0.1189, |
|
"step": 837, |
|
"time": 49.35 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.6126e-05", |
|
"loss": 0.1156, |
|
"slid_loss": 0.1187, |
|
"step": 838, |
|
"time": 47.95 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.6041e-05", |
|
"loss": 0.1113, |
|
"slid_loss": 0.1187, |
|
"step": 839, |
|
"time": 46.78 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.5955e-05", |
|
"loss": 0.1195, |
|
"slid_loss": 0.1187, |
|
"step": 840, |
|
"time": 45.81 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.5869e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1187, |
|
"step": 841, |
|
"time": 46.43 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.5784e-05", |
|
"loss": 0.1096, |
|
"slid_loss": 0.1186, |
|
"step": 842, |
|
"time": 47.27 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.5699e-05", |
|
"loss": 0.1169, |
|
"slid_loss": 0.1186, |
|
"step": 843, |
|
"time": 45.92 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.5614e-05", |
|
"loss": 0.1204, |
|
"slid_loss": 0.1186, |
|
"step": 844, |
|
"time": 48.48 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.5529e-05", |
|
"loss": 0.1219, |
|
"slid_loss": 0.1187, |
|
"step": 845, |
|
"time": 47.8 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.5445e-05", |
|
"loss": 0.124, |
|
"slid_loss": 0.1188, |
|
"step": 846, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.5360e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1187, |
|
"step": 847, |
|
"time": 46.46 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "3.5276e-05", |
|
"loss": 0.1132, |
|
"slid_loss": 0.1186, |
|
"step": 848, |
|
"time": 48.07 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.5192e-05", |
|
"loss": 0.1206, |
|
"slid_loss": 0.1186, |
|
"step": 849, |
|
"time": 48.53 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.5108e-05", |
|
"loss": 0.1161, |
|
"slid_loss": 0.1186, |
|
"step": 850, |
|
"time": 46.09 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.5025e-05", |
|
"loss": 0.1146, |
|
"slid_loss": 0.1185, |
|
"step": 851, |
|
"time": 48.48 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.4941e-05", |
|
"loss": 0.1196, |
|
"slid_loss": 0.1184, |
|
"step": 852, |
|
"time": 47.18 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.4858e-05", |
|
"loss": 0.1151, |
|
"slid_loss": 0.1184, |
|
"step": 853, |
|
"time": 46.16 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.4775e-05", |
|
"loss": 0.1268, |
|
"slid_loss": 0.1185, |
|
"step": 854, |
|
"time": 46.71 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.4692e-05", |
|
"loss": 0.1176, |
|
"slid_loss": 0.1184, |
|
"step": 855, |
|
"time": 45.81 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.4609e-05", |
|
"loss": 0.1152, |
|
"slid_loss": 0.1183, |
|
"step": 856, |
|
"time": 47.43 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.4526e-05", |
|
"loss": 0.1259, |
|
"slid_loss": 0.1184, |
|
"step": 857, |
|
"time": 46.53 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.4444e-05", |
|
"loss": 0.1076, |
|
"slid_loss": 0.1184, |
|
"step": 858, |
|
"time": 47.91 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.4361e-05", |
|
"loss": 0.1193, |
|
"slid_loss": 0.1183, |
|
"step": 859, |
|
"time": 47.58 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "3.4279e-05", |
|
"loss": 0.1154, |
|
"slid_loss": 0.1183, |
|
"step": 860, |
|
"time": 46.0 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.4198e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1183, |
|
"step": 861, |
|
"time": 46.49 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.4116e-05", |
|
"loss": 0.1222, |
|
"slid_loss": 0.1183, |
|
"step": 862, |
|
"time": 46.74 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.4034e-05", |
|
"loss": 0.1129, |
|
"slid_loss": 0.1182, |
|
"step": 863, |
|
"time": 46.19 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.3953e-05", |
|
"loss": 0.1159, |
|
"slid_loss": 0.1182, |
|
"step": 864, |
|
"time": 48.54 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.3872e-05", |
|
"loss": 0.1183, |
|
"slid_loss": 0.1182, |
|
"step": 865, |
|
"time": 47.87 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.3791e-05", |
|
"loss": 0.1165, |
|
"slid_loss": 0.1181, |
|
"step": 866, |
|
"time": 45.96 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.3710e-05", |
|
"loss": 0.1114, |
|
"slid_loss": 0.1181, |
|
"step": 867, |
|
"time": 47.74 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.3630e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1182, |
|
"step": 868, |
|
"time": 47.14 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.3549e-05", |
|
"loss": 0.1181, |
|
"slid_loss": 0.1182, |
|
"step": 869, |
|
"time": 46.14 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.3469e-05", |
|
"loss": 0.1178, |
|
"slid_loss": 0.1182, |
|
"step": 870, |
|
"time": 45.85 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.3389e-05", |
|
"loss": 0.1191, |
|
"slid_loss": 0.1182, |
|
"step": 871, |
|
"time": 46.84 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "3.3309e-05", |
|
"loss": 0.121, |
|
"slid_loss": 0.1181, |
|
"step": 872, |
|
"time": 48.94 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.3229e-05", |
|
"loss": 0.1175, |
|
"slid_loss": 0.118, |
|
"step": 873, |
|
"time": 47.76 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.3150e-05", |
|
"loss": 0.1235, |
|
"slid_loss": 0.1181, |
|
"step": 874, |
|
"time": 47.11 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.3071e-05", |
|
"loss": 0.1157, |
|
"slid_loss": 0.1181, |
|
"step": 875, |
|
"time": 47.86 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.2992e-05", |
|
"loss": 0.1211, |
|
"slid_loss": 0.1181, |
|
"step": 876, |
|
"time": 49.33 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.2913e-05", |
|
"loss": 0.1093, |
|
"slid_loss": 0.1179, |
|
"step": 877, |
|
"time": 46.76 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.2834e-05", |
|
"loss": 0.1207, |
|
"slid_loss": 0.118, |
|
"step": 878, |
|
"time": 47.07 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.2756e-05", |
|
"loss": 0.1247, |
|
"slid_loss": 0.1181, |
|
"step": 879, |
|
"time": 45.94 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.2677e-05", |
|
"loss": 0.1186, |
|
"slid_loss": 0.118, |
|
"step": 880, |
|
"time": 46.12 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.2599e-05", |
|
"loss": 0.1146, |
|
"slid_loss": 0.1179, |
|
"step": 881, |
|
"time": 48.24 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.2522e-05", |
|
"loss": 0.1171, |
|
"slid_loss": 0.1179, |
|
"step": 882, |
|
"time": 46.17 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.2444e-05", |
|
"loss": 0.1285, |
|
"slid_loss": 0.1181, |
|
"step": 883, |
|
"time": 47.01 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "3.2366e-05", |
|
"loss": 0.1132, |
|
"slid_loss": 0.118, |
|
"step": 884, |
|
"time": 45.85 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.2289e-05", |
|
"loss": 0.1168, |
|
"slid_loss": 0.118, |
|
"step": 885, |
|
"time": 45.91 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.2212e-05", |
|
"loss": 0.1176, |
|
"slid_loss": 0.1179, |
|
"step": 886, |
|
"time": 45.82 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.2135e-05", |
|
"loss": 0.1133, |
|
"slid_loss": 0.1178, |
|
"step": 887, |
|
"time": 46.58 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.2058e-05", |
|
"loss": 0.1185, |
|
"slid_loss": 0.1178, |
|
"step": 888, |
|
"time": 45.12 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.1982e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1179, |
|
"step": 889, |
|
"time": 47.64 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.1906e-05", |
|
"loss": 0.1083, |
|
"slid_loss": 0.1177, |
|
"step": 890, |
|
"time": 47.17 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.1829e-05", |
|
"loss": 0.111, |
|
"slid_loss": 0.1176, |
|
"step": 891, |
|
"time": 48.31 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.1754e-05", |
|
"loss": 0.11, |
|
"slid_loss": 0.1175, |
|
"step": 892, |
|
"time": 47.13 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.1678e-05", |
|
"loss": 0.119, |
|
"slid_loss": 0.1176, |
|
"step": 893, |
|
"time": 45.72 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.1602e-05", |
|
"loss": 0.1122, |
|
"slid_loss": 0.1175, |
|
"step": 894, |
|
"time": 46.13 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.1527e-05", |
|
"loss": 0.113, |
|
"slid_loss": 0.1174, |
|
"step": 895, |
|
"time": 48.87 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "3.1452e-05", |
|
"loss": 0.1094, |
|
"slid_loss": 0.1173, |
|
"step": 896, |
|
"time": 49.0 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.1377e-05", |
|
"loss": 0.1165, |
|
"slid_loss": 0.1172, |
|
"step": 897, |
|
"time": 45.51 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.1303e-05", |
|
"loss": 0.1152, |
|
"slid_loss": 0.1171, |
|
"step": 898, |
|
"time": 45.83 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.1228e-05", |
|
"loss": 0.1186, |
|
"slid_loss": 0.1171, |
|
"step": 899, |
|
"time": 47.17 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.1154e-05", |
|
"loss": 0.117, |
|
"slid_loss": 0.1171, |
|
"step": 900, |
|
"time": 46.53 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.1080e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1171, |
|
"step": 901, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.1006e-05", |
|
"loss": 0.1152, |
|
"slid_loss": 0.1171, |
|
"step": 902, |
|
"time": 48.7 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.0932e-05", |
|
"loss": 0.117, |
|
"slid_loss": 0.1171, |
|
"step": 903, |
|
"time": 45.46 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.0859e-05", |
|
"loss": 0.1143, |
|
"slid_loss": 0.117, |
|
"step": 904, |
|
"time": 46.95 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.0786e-05", |
|
"loss": 0.1189, |
|
"slid_loss": 0.1169, |
|
"step": 905, |
|
"time": 46.39 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.0713e-05", |
|
"loss": 0.1161, |
|
"slid_loss": 0.1168, |
|
"step": 906, |
|
"time": 47.25 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "3.0640e-05", |
|
"loss": 0.1148, |
|
"slid_loss": 0.1168, |
|
"step": 907, |
|
"time": 46.94 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "3.0567e-05", |
|
"loss": 0.1112, |
|
"slid_loss": 0.1168, |
|
"step": 908, |
|
"time": 46.12 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "3.0495e-05", |
|
"loss": 0.1151, |
|
"slid_loss": 0.1168, |
|
"step": 909, |
|
"time": 46.94 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "3.0423e-05", |
|
"loss": 0.1129, |
|
"slid_loss": 0.1167, |
|
"step": 910, |
|
"time": 47.02 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "3.0351e-05", |
|
"loss": 0.1109, |
|
"slid_loss": 0.1166, |
|
"step": 911, |
|
"time": 46.06 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "3.0279e-05", |
|
"loss": 0.1222, |
|
"slid_loss": 0.1166, |
|
"step": 912, |
|
"time": 46.7 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "3.0207e-05", |
|
"loss": 0.1184, |
|
"slid_loss": 0.1167, |
|
"step": 913, |
|
"time": 49.91 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "3.0136e-05", |
|
"loss": 0.1135, |
|
"slid_loss": 0.1167, |
|
"step": 914, |
|
"time": 47.68 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "3.0065e-05", |
|
"loss": 0.1132, |
|
"slid_loss": 0.1166, |
|
"step": 915, |
|
"time": 45.99 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "2.9994e-05", |
|
"loss": 0.1175, |
|
"slid_loss": 0.1166, |
|
"step": 916, |
|
"time": 47.32 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "2.9923e-05", |
|
"loss": 0.11, |
|
"slid_loss": 0.1165, |
|
"step": 917, |
|
"time": 47.84 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "2.9853e-05", |
|
"loss": 0.1085, |
|
"slid_loss": 0.1165, |
|
"step": 918, |
|
"time": 45.78 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "2.9783e-05", |
|
"loss": 0.1136, |
|
"slid_loss": 0.1164, |
|
"step": 919, |
|
"time": 45.71 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9713e-05", |
|
"loss": 0.1123, |
|
"slid_loss": 0.1163, |
|
"step": 920, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9643e-05", |
|
"loss": 0.1094, |
|
"slid_loss": 0.1163, |
|
"step": 921, |
|
"time": 46.4 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9573e-05", |
|
"loss": 0.1165, |
|
"slid_loss": 0.1163, |
|
"step": 922, |
|
"time": 45.21 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9504e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1164, |
|
"step": 923, |
|
"time": 45.09 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9435e-05", |
|
"loss": 0.1105, |
|
"slid_loss": 0.1163, |
|
"step": 924, |
|
"time": 49.4 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9366e-05", |
|
"loss": 0.1157, |
|
"slid_loss": 0.1163, |
|
"step": 925, |
|
"time": 45.97 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9297e-05", |
|
"loss": 0.1114, |
|
"slid_loss": 0.1163, |
|
"step": 926, |
|
"time": 47.66 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9229e-05", |
|
"loss": 0.119, |
|
"slid_loss": 0.1163, |
|
"step": 927, |
|
"time": 46.03 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9160e-05", |
|
"loss": 0.1205, |
|
"slid_loss": 0.1163, |
|
"step": 928, |
|
"time": 46.26 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9092e-05", |
|
"loss": 0.1196, |
|
"slid_loss": 0.1163, |
|
"step": 929, |
|
"time": 47.92 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.9024e-05", |
|
"loss": 0.109, |
|
"slid_loss": 0.1163, |
|
"step": 930, |
|
"time": 46.43 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "2.8957e-05", |
|
"loss": 0.1186, |
|
"slid_loss": 0.1164, |
|
"step": 931, |
|
"time": 48.1 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8889e-05", |
|
"loss": 0.1201, |
|
"slid_loss": 0.1164, |
|
"step": 932, |
|
"time": 46.14 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8822e-05", |
|
"loss": 0.1181, |
|
"slid_loss": 0.1163, |
|
"step": 933, |
|
"time": 46.97 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8755e-05", |
|
"loss": 0.1159, |
|
"slid_loss": 0.1164, |
|
"step": 934, |
|
"time": 45.61 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8689e-05", |
|
"loss": 0.1097, |
|
"slid_loss": 0.1163, |
|
"step": 935, |
|
"time": 47.74 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8622e-05", |
|
"loss": 0.1231, |
|
"slid_loss": 0.1164, |
|
"step": 936, |
|
"time": 45.26 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8556e-05", |
|
"loss": 0.1208, |
|
"slid_loss": 0.1164, |
|
"step": 937, |
|
"time": 48.38 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8490e-05", |
|
"loss": 0.1139, |
|
"slid_loss": 0.1164, |
|
"step": 938, |
|
"time": 46.4 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8424e-05", |
|
"loss": 0.1126, |
|
"slid_loss": 0.1164, |
|
"step": 939, |
|
"time": 45.93 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8358e-05", |
|
"loss": 0.1088, |
|
"slid_loss": 0.1163, |
|
"step": 940, |
|
"time": 50.99 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8293e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1163, |
|
"step": 941, |
|
"time": 47.68 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8228e-05", |
|
"loss": 0.1127, |
|
"slid_loss": 0.1164, |
|
"step": 942, |
|
"time": 46.44 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "2.8163e-05", |
|
"loss": 0.1268, |
|
"slid_loss": 0.1165, |
|
"step": 943, |
|
"time": 45.53 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.8098e-05", |
|
"loss": 0.1257, |
|
"slid_loss": 0.1165, |
|
"step": 944, |
|
"time": 46.87 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.8034e-05", |
|
"loss": 0.1135, |
|
"slid_loss": 0.1164, |
|
"step": 945, |
|
"time": 48.51 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7970e-05", |
|
"loss": 0.1159, |
|
"slid_loss": 0.1164, |
|
"step": 946, |
|
"time": 45.8 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7906e-05", |
|
"loss": 0.1171, |
|
"slid_loss": 0.1164, |
|
"step": 947, |
|
"time": 47.06 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7842e-05", |
|
"loss": 0.1099, |
|
"slid_loss": 0.1163, |
|
"step": 948, |
|
"time": 45.61 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7778e-05", |
|
"loss": 0.1158, |
|
"slid_loss": 0.1163, |
|
"step": 949, |
|
"time": 48.64 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7715e-05", |
|
"loss": 0.1108, |
|
"slid_loss": 0.1162, |
|
"step": 950, |
|
"time": 46.89 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7652e-05", |
|
"loss": 0.1181, |
|
"slid_loss": 0.1163, |
|
"step": 951, |
|
"time": 45.91 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7589e-05", |
|
"loss": 0.1097, |
|
"slid_loss": 0.1162, |
|
"step": 952, |
|
"time": 46.57 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7526e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1162, |
|
"step": 953, |
|
"time": 46.58 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7464e-05", |
|
"loss": 0.1069, |
|
"slid_loss": 0.116, |
|
"step": 954, |
|
"time": 44.75 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "2.7402e-05", |
|
"loss": 0.116, |
|
"slid_loss": 0.116, |
|
"step": 955, |
|
"time": 46.85 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.7340e-05", |
|
"loss": 0.1187, |
|
"slid_loss": 0.116, |
|
"step": 956, |
|
"time": 46.7 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.7278e-05", |
|
"loss": 0.1139, |
|
"slid_loss": 0.1159, |
|
"step": 957, |
|
"time": 45.91 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.7217e-05", |
|
"loss": 0.1053, |
|
"slid_loss": 0.1159, |
|
"step": 958, |
|
"time": 46.81 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.7156e-05", |
|
"loss": 0.1125, |
|
"slid_loss": 0.1158, |
|
"step": 959, |
|
"time": 46.81 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.7095e-05", |
|
"loss": 0.1191, |
|
"slid_loss": 0.1158, |
|
"step": 960, |
|
"time": 44.53 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.7034e-05", |
|
"loss": 0.1089, |
|
"slid_loss": 0.1158, |
|
"step": 961, |
|
"time": 45.3 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.6973e-05", |
|
"loss": 0.1067, |
|
"slid_loss": 0.1156, |
|
"step": 962, |
|
"time": 46.3 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.6913e-05", |
|
"loss": 0.108, |
|
"slid_loss": 0.1155, |
|
"step": 963, |
|
"time": 49.07 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.6853e-05", |
|
"loss": 0.1148, |
|
"slid_loss": 0.1155, |
|
"step": 964, |
|
"time": 45.9 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.6793e-05", |
|
"loss": 0.1077, |
|
"slid_loss": 0.1154, |
|
"step": 965, |
|
"time": 45.71 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.6734e-05", |
|
"loss": 0.1205, |
|
"slid_loss": 0.1155, |
|
"step": 966, |
|
"time": 46.22 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "2.6674e-05", |
|
"loss": 0.1153, |
|
"slid_loss": 0.1155, |
|
"step": 967, |
|
"time": 45.77 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6615e-05", |
|
"loss": 0.1231, |
|
"slid_loss": 0.1155, |
|
"step": 968, |
|
"time": 47.15 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6557e-05", |
|
"loss": 0.1187, |
|
"slid_loss": 0.1155, |
|
"step": 969, |
|
"time": 47.57 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6498e-05", |
|
"loss": 0.1167, |
|
"slid_loss": 0.1155, |
|
"step": 970, |
|
"time": 45.24 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6440e-05", |
|
"loss": 0.1168, |
|
"slid_loss": 0.1155, |
|
"step": 971, |
|
"time": 48.19 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6381e-05", |
|
"loss": 0.1163, |
|
"slid_loss": 0.1155, |
|
"step": 972, |
|
"time": 50.11 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6324e-05", |
|
"loss": 0.1194, |
|
"slid_loss": 0.1155, |
|
"step": 973, |
|
"time": 46.14 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6266e-05", |
|
"loss": 0.1256, |
|
"slid_loss": 0.1155, |
|
"step": 974, |
|
"time": 45.84 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6209e-05", |
|
"loss": 0.1228, |
|
"slid_loss": 0.1156, |
|
"step": 975, |
|
"time": 45.73 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6151e-05", |
|
"loss": 0.1255, |
|
"slid_loss": 0.1156, |
|
"step": 976, |
|
"time": 46.78 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6094e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1157, |
|
"step": 977, |
|
"time": 50.52 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.6038e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1157, |
|
"step": 978, |
|
"time": 46.46 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "2.5981e-05", |
|
"loss": 0.1104, |
|
"slid_loss": 0.1156, |
|
"step": 979, |
|
"time": 47.8 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5925e-05", |
|
"loss": 0.1197, |
|
"slid_loss": 0.1156, |
|
"step": 980, |
|
"time": 46.11 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5869e-05", |
|
"loss": 0.1053, |
|
"slid_loss": 0.1155, |
|
"step": 981, |
|
"time": 48.43 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5814e-05", |
|
"loss": 0.1098, |
|
"slid_loss": 0.1154, |
|
"step": 982, |
|
"time": 47.15 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5758e-05", |
|
"loss": 0.1153, |
|
"slid_loss": 0.1153, |
|
"step": 983, |
|
"time": 47.75 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5703e-05", |
|
"loss": 0.1203, |
|
"slid_loss": 0.1154, |
|
"step": 984, |
|
"time": 45.52 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5648e-05", |
|
"loss": 0.1182, |
|
"slid_loss": 0.1154, |
|
"step": 985, |
|
"time": 46.57 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5593e-05", |
|
"loss": 0.1094, |
|
"slid_loss": 0.1153, |
|
"step": 986, |
|
"time": 47.1 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5539e-05", |
|
"loss": 0.1139, |
|
"slid_loss": 0.1153, |
|
"step": 987, |
|
"time": 47.68 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5485e-05", |
|
"loss": 0.1116, |
|
"slid_loss": 0.1152, |
|
"step": 988, |
|
"time": 44.93 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5431e-05", |
|
"loss": 0.1165, |
|
"slid_loss": 0.1152, |
|
"step": 989, |
|
"time": 47.18 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5377e-05", |
|
"loss": 0.116, |
|
"slid_loss": 0.1153, |
|
"step": 990, |
|
"time": 48.04 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "2.5323e-05", |
|
"loss": 0.1196, |
|
"slid_loss": 0.1154, |
|
"step": 991, |
|
"time": 46.07 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.5270e-05", |
|
"loss": 0.1067, |
|
"slid_loss": 0.1153, |
|
"step": 992, |
|
"time": 46.95 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.5217e-05", |
|
"loss": 0.1148, |
|
"slid_loss": 0.1153, |
|
"step": 993, |
|
"time": 46.44 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.5164e-05", |
|
"loss": 0.1128, |
|
"slid_loss": 0.1153, |
|
"step": 994, |
|
"time": 47.29 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.5112e-05", |
|
"loss": 0.1214, |
|
"slid_loss": 0.1154, |
|
"step": 995, |
|
"time": 45.88 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.5060e-05", |
|
"loss": 0.1116, |
|
"slid_loss": 0.1154, |
|
"step": 996, |
|
"time": 48.57 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.5008e-05", |
|
"loss": 0.106, |
|
"slid_loss": 0.1153, |
|
"step": 997, |
|
"time": 46.91 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.4956e-05", |
|
"loss": 0.1137, |
|
"slid_loss": 0.1153, |
|
"step": 998, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.4905e-05", |
|
"loss": 0.1185, |
|
"slid_loss": 0.1153, |
|
"step": 999, |
|
"time": 48.06 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.4853e-05", |
|
"loss": 0.1116, |
|
"slid_loss": 0.1152, |
|
"step": 1000, |
|
"time": 47.62 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.4802e-05", |
|
"loss": 0.1087, |
|
"slid_loss": 0.1151, |
|
"step": 1001, |
|
"time": 47.27 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "2.4752e-05", |
|
"loss": 0.1241, |
|
"slid_loss": 0.1152, |
|
"step": 1002, |
|
"time": 46.49 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4701e-05", |
|
"loss": 0.1145, |
|
"slid_loss": 0.1152, |
|
"step": 1003, |
|
"time": 46.74 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4651e-05", |
|
"loss": 0.1114, |
|
"slid_loss": 0.1151, |
|
"step": 1004, |
|
"time": 48.33 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4601e-05", |
|
"loss": 0.1096, |
|
"slid_loss": 0.1151, |
|
"step": 1005, |
|
"time": 46.77 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4551e-05", |
|
"loss": 0.1115, |
|
"slid_loss": 0.115, |
|
"step": 1006, |
|
"time": 46.68 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4502e-05", |
|
"loss": 0.1132, |
|
"slid_loss": 0.115, |
|
"step": 1007, |
|
"time": 48.08 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4453e-05", |
|
"loss": 0.1218, |
|
"slid_loss": 0.1151, |
|
"step": 1008, |
|
"time": 47.73 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4404e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1151, |
|
"step": 1009, |
|
"time": 47.98 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4355e-05", |
|
"loss": 0.1096, |
|
"slid_loss": 0.1151, |
|
"step": 1010, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4306e-05", |
|
"loss": 0.1202, |
|
"slid_loss": 0.1152, |
|
"step": 1011, |
|
"time": 45.19 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4258e-05", |
|
"loss": 0.1154, |
|
"slid_loss": 0.1151, |
|
"step": 1012, |
|
"time": 46.47 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4210e-05", |
|
"loss": 0.1181, |
|
"slid_loss": 0.1151, |
|
"step": 1013, |
|
"time": 47.2 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "2.4163e-05", |
|
"loss": 0.1154, |
|
"slid_loss": 0.1151, |
|
"step": 1014, |
|
"time": 46.84 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.4115e-05", |
|
"loss": 0.1129, |
|
"slid_loss": 0.1151, |
|
"step": 1015, |
|
"time": 45.4 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.4068e-05", |
|
"loss": 0.1159, |
|
"slid_loss": 0.1151, |
|
"step": 1016, |
|
"time": 49.58 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.4021e-05", |
|
"loss": 0.1113, |
|
"slid_loss": 0.1151, |
|
"step": 1017, |
|
"time": 48.75 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.3975e-05", |
|
"loss": 0.1106, |
|
"slid_loss": 0.1152, |
|
"step": 1018, |
|
"time": 46.42 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.3928e-05", |
|
"loss": 0.1189, |
|
"slid_loss": 0.1152, |
|
"step": 1019, |
|
"time": 45.68 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.3882e-05", |
|
"loss": 0.1173, |
|
"slid_loss": 0.1153, |
|
"step": 1020, |
|
"time": 47.61 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.3836e-05", |
|
"loss": 0.1178, |
|
"slid_loss": 0.1153, |
|
"step": 1021, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.3791e-05", |
|
"loss": 0.1111, |
|
"slid_loss": 0.1153, |
|
"step": 1022, |
|
"time": 47.24 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.3745e-05", |
|
"loss": 0.1213, |
|
"slid_loss": 0.1153, |
|
"step": 1023, |
|
"time": 46.95 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.3700e-05", |
|
"loss": 0.1255, |
|
"slid_loss": 0.1155, |
|
"step": 1024, |
|
"time": 46.45 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.3655e-05", |
|
"loss": 0.121, |
|
"slid_loss": 0.1155, |
|
"step": 1025, |
|
"time": 46.76 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "2.3611e-05", |
|
"loss": 0.1161, |
|
"slid_loss": 0.1156, |
|
"step": 1026, |
|
"time": 46.05 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3566e-05", |
|
"loss": 0.1122, |
|
"slid_loss": 0.1155, |
|
"step": 1027, |
|
"time": 48.79 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3522e-05", |
|
"loss": 0.1145, |
|
"slid_loss": 0.1154, |
|
"step": 1028, |
|
"time": 47.81 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3479e-05", |
|
"loss": 0.1181, |
|
"slid_loss": 0.1154, |
|
"step": 1029, |
|
"time": 47.66 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3435e-05", |
|
"loss": 0.1125, |
|
"slid_loss": 0.1155, |
|
"step": 1030, |
|
"time": 47.31 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3392e-05", |
|
"loss": 0.1113, |
|
"slid_loss": 0.1154, |
|
"step": 1031, |
|
"time": 46.28 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3349e-05", |
|
"loss": 0.1224, |
|
"slid_loss": 0.1154, |
|
"step": 1032, |
|
"time": 45.2 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3306e-05", |
|
"loss": 0.1088, |
|
"slid_loss": 0.1153, |
|
"step": 1033, |
|
"time": 48.14 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3264e-05", |
|
"loss": 0.1111, |
|
"slid_loss": 0.1153, |
|
"step": 1034, |
|
"time": 45.16 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3221e-05", |
|
"loss": 0.1147, |
|
"slid_loss": 0.1153, |
|
"step": 1035, |
|
"time": 46.36 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3179e-05", |
|
"loss": 0.1126, |
|
"slid_loss": 0.1152, |
|
"step": 1036, |
|
"time": 46.2 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3138e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1152, |
|
"step": 1037, |
|
"time": 47.44 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "2.3096e-05", |
|
"loss": 0.1141, |
|
"slid_loss": 0.1152, |
|
"step": 1038, |
|
"time": 49.26 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.3055e-05", |
|
"loss": 0.1162, |
|
"slid_loss": 0.1152, |
|
"step": 1039, |
|
"time": 48.24 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.3014e-05", |
|
"loss": 0.1146, |
|
"slid_loss": 0.1153, |
|
"step": 1040, |
|
"time": 45.57 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2974e-05", |
|
"loss": 0.1096, |
|
"slid_loss": 0.1152, |
|
"step": 1041, |
|
"time": 46.74 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2933e-05", |
|
"loss": 0.1071, |
|
"slid_loss": 0.1151, |
|
"step": 1042, |
|
"time": 46.73 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2893e-05", |
|
"loss": 0.123, |
|
"slid_loss": 0.1151, |
|
"step": 1043, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2853e-05", |
|
"loss": 0.11, |
|
"slid_loss": 0.115, |
|
"step": 1044, |
|
"time": 48.14 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2814e-05", |
|
"loss": 0.1118, |
|
"slid_loss": 0.1149, |
|
"step": 1045, |
|
"time": 44.74 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2774e-05", |
|
"loss": 0.1162, |
|
"slid_loss": 0.1149, |
|
"step": 1046, |
|
"time": 46.83 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2735e-05", |
|
"loss": 0.1075, |
|
"slid_loss": 0.1148, |
|
"step": 1047, |
|
"time": 49.54 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2697e-05", |
|
"loss": 0.1115, |
|
"slid_loss": 0.1149, |
|
"step": 1048, |
|
"time": 46.73 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2658e-05", |
|
"loss": 0.1076, |
|
"slid_loss": 0.1148, |
|
"step": 1049, |
|
"time": 46.04 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "2.2620e-05", |
|
"loss": 0.1045, |
|
"slid_loss": 0.1147, |
|
"step": 1050, |
|
"time": 46.85 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2582e-05", |
|
"loss": 0.1134, |
|
"slid_loss": 0.1147, |
|
"step": 1051, |
|
"time": 48.16 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2544e-05", |
|
"loss": 0.1146, |
|
"slid_loss": 0.1147, |
|
"step": 1052, |
|
"time": 44.94 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2507e-05", |
|
"loss": 0.1182, |
|
"slid_loss": 0.1147, |
|
"step": 1053, |
|
"time": 48.19 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2470e-05", |
|
"loss": 0.1172, |
|
"slid_loss": 0.1148, |
|
"step": 1054, |
|
"time": 48.89 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2433e-05", |
|
"loss": 0.1189, |
|
"slid_loss": 0.1148, |
|
"step": 1055, |
|
"time": 47.75 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2396e-05", |
|
"loss": 0.1113, |
|
"slid_loss": 0.1148, |
|
"step": 1056, |
|
"time": 47.32 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2360e-05", |
|
"loss": 0.1136, |
|
"slid_loss": 0.1148, |
|
"step": 1057, |
|
"time": 45.24 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2324e-05", |
|
"loss": 0.1161, |
|
"slid_loss": 0.1149, |
|
"step": 1058, |
|
"time": 47.15 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2288e-05", |
|
"loss": 0.1152, |
|
"slid_loss": 0.1149, |
|
"step": 1059, |
|
"time": 46.34 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2252e-05", |
|
"loss": 0.1129, |
|
"slid_loss": 0.1148, |
|
"step": 1060, |
|
"time": 47.62 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2217e-05", |
|
"loss": 0.1196, |
|
"slid_loss": 0.1149, |
|
"step": 1061, |
|
"time": 46.57 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "2.2182e-05", |
|
"loss": 0.1145, |
|
"slid_loss": 0.115, |
|
"step": 1062, |
|
"time": 48.06 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.2147e-05", |
|
"loss": 0.1085, |
|
"slid_loss": 0.115, |
|
"step": 1063, |
|
"time": 46.09 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.2113e-05", |
|
"loss": 0.1171, |
|
"slid_loss": 0.115, |
|
"step": 1064, |
|
"time": 46.02 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.2079e-05", |
|
"loss": 0.1189, |
|
"slid_loss": 0.1152, |
|
"step": 1065, |
|
"time": 47.42 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.2045e-05", |
|
"loss": 0.112, |
|
"slid_loss": 0.1151, |
|
"step": 1066, |
|
"time": 47.54 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.2011e-05", |
|
"loss": 0.1182, |
|
"slid_loss": 0.1151, |
|
"step": 1067, |
|
"time": 47.85 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.1978e-05", |
|
"loss": 0.116, |
|
"slid_loss": 0.115, |
|
"step": 1068, |
|
"time": 47.25 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.1945e-05", |
|
"loss": 0.1126, |
|
"slid_loss": 0.115, |
|
"step": 1069, |
|
"time": 46.86 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.1912e-05", |
|
"loss": 0.112, |
|
"slid_loss": 0.1149, |
|
"step": 1070, |
|
"time": 49.15 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.1879e-05", |
|
"loss": 0.1167, |
|
"slid_loss": 0.1149, |
|
"step": 1071, |
|
"time": 46.98 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.1847e-05", |
|
"loss": 0.1098, |
|
"slid_loss": 0.1149, |
|
"step": 1072, |
|
"time": 45.64 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.1815e-05", |
|
"loss": 0.1188, |
|
"slid_loss": 0.1148, |
|
"step": 1073, |
|
"time": 46.94 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "2.1783e-05", |
|
"loss": 0.1207, |
|
"slid_loss": 0.1148, |
|
"step": 1074, |
|
"time": 45.65 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1752e-05", |
|
"loss": 0.1143, |
|
"slid_loss": 0.1147, |
|
"step": 1075, |
|
"time": 45.88 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1720e-05", |
|
"loss": 0.1115, |
|
"slid_loss": 0.1146, |
|
"step": 1076, |
|
"time": 46.22 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1690e-05", |
|
"loss": 0.1112, |
|
"slid_loss": 0.1145, |
|
"step": 1077, |
|
"time": 45.05 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1659e-05", |
|
"loss": 0.1088, |
|
"slid_loss": 0.1144, |
|
"step": 1078, |
|
"time": 47.14 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1629e-05", |
|
"loss": 0.1157, |
|
"slid_loss": 0.1144, |
|
"step": 1079, |
|
"time": 46.14 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1598e-05", |
|
"loss": 0.1133, |
|
"slid_loss": 0.1144, |
|
"step": 1080, |
|
"time": 47.65 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1569e-05", |
|
"loss": 0.1097, |
|
"slid_loss": 0.1144, |
|
"step": 1081, |
|
"time": 46.47 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1539e-05", |
|
"loss": 0.1146, |
|
"slid_loss": 0.1145, |
|
"step": 1082, |
|
"time": 47.5 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1510e-05", |
|
"loss": 0.1162, |
|
"slid_loss": 0.1145, |
|
"step": 1083, |
|
"time": 45.48 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1481e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1145, |
|
"step": 1084, |
|
"time": 47.38 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "2.1452e-05", |
|
"loss": 0.1112, |
|
"slid_loss": 0.1144, |
|
"step": 1085, |
|
"time": 46.7 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1424e-05", |
|
"loss": 0.1093, |
|
"slid_loss": 0.1144, |
|
"step": 1086, |
|
"time": 49.39 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1395e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1144, |
|
"step": 1087, |
|
"time": 45.56 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1368e-05", |
|
"loss": 0.1082, |
|
"slid_loss": 0.1144, |
|
"step": 1088, |
|
"time": 47.11 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1340e-05", |
|
"loss": 0.1225, |
|
"slid_loss": 0.1144, |
|
"step": 1089, |
|
"time": 47.24 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1313e-05", |
|
"loss": 0.1201, |
|
"slid_loss": 0.1145, |
|
"step": 1090, |
|
"time": 47.99 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1286e-05", |
|
"loss": 0.1151, |
|
"slid_loss": 0.1144, |
|
"step": 1091, |
|
"time": 45.78 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1259e-05", |
|
"loss": 0.1146, |
|
"slid_loss": 0.1145, |
|
"step": 1092, |
|
"time": 46.39 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1232e-05", |
|
"loss": 0.1114, |
|
"slid_loss": 0.1145, |
|
"step": 1093, |
|
"time": 49.06 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1206e-05", |
|
"loss": 0.1182, |
|
"slid_loss": 0.1145, |
|
"step": 1094, |
|
"time": 47.98 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1180e-05", |
|
"loss": 0.1196, |
|
"slid_loss": 0.1145, |
|
"step": 1095, |
|
"time": 45.44 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1154e-05", |
|
"loss": 0.1152, |
|
"slid_loss": 0.1146, |
|
"step": 1096, |
|
"time": 46.09 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": "2.1129e-05", |
|
"loss": 0.1139, |
|
"slid_loss": 0.1146, |
|
"step": 1097, |
|
"time": 46.07 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.1104e-05", |
|
"loss": 0.1157, |
|
"slid_loss": 0.1147, |
|
"step": 1098, |
|
"time": 47.39 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.1079e-05", |
|
"loss": 0.1163, |
|
"slid_loss": 0.1146, |
|
"step": 1099, |
|
"time": 45.75 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.1055e-05", |
|
"loss": 0.113, |
|
"slid_loss": 0.1146, |
|
"step": 1100, |
|
"time": 47.66 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.1030e-05", |
|
"loss": 0.1101, |
|
"slid_loss": 0.1147, |
|
"step": 1101, |
|
"time": 45.75 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.1006e-05", |
|
"loss": 0.1153, |
|
"slid_loss": 0.1146, |
|
"step": 1102, |
|
"time": 46.29 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.0983e-05", |
|
"loss": 0.1164, |
|
"slid_loss": 0.1146, |
|
"step": 1103, |
|
"time": 45.14 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.0959e-05", |
|
"loss": 0.1127, |
|
"slid_loss": 0.1146, |
|
"step": 1104, |
|
"time": 48.06 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.0936e-05", |
|
"loss": 0.1137, |
|
"slid_loss": 0.1146, |
|
"step": 1105, |
|
"time": 46.65 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.0913e-05", |
|
"loss": 0.1248, |
|
"slid_loss": 0.1148, |
|
"step": 1106, |
|
"time": 46.96 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.0890e-05", |
|
"loss": 0.1099, |
|
"slid_loss": 0.1147, |
|
"step": 1107, |
|
"time": 48.73 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.0868e-05", |
|
"loss": 0.1062, |
|
"slid_loss": 0.1146, |
|
"step": 1108, |
|
"time": 46.86 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": "2.0846e-05", |
|
"loss": 0.1138, |
|
"slid_loss": 0.1146, |
|
"step": 1109, |
|
"time": 46.17 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0824e-05", |
|
"loss": 0.1239, |
|
"slid_loss": 0.1147, |
|
"step": 1110, |
|
"time": 46.76 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0803e-05", |
|
"loss": 0.1187, |
|
"slid_loss": 0.1147, |
|
"step": 1111, |
|
"time": 47.0 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0782e-05", |
|
"loss": 0.1144, |
|
"slid_loss": 0.1147, |
|
"step": 1112, |
|
"time": 46.01 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0761e-05", |
|
"loss": 0.1121, |
|
"slid_loss": 0.1146, |
|
"step": 1113, |
|
"time": 46.16 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0740e-05", |
|
"loss": 0.1023, |
|
"slid_loss": 0.1145, |
|
"step": 1114, |
|
"time": 46.6 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0720e-05", |
|
"loss": 0.1192, |
|
"slid_loss": 0.1145, |
|
"step": 1115, |
|
"time": 46.4 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0700e-05", |
|
"loss": 0.1215, |
|
"slid_loss": 0.1146, |
|
"step": 1116, |
|
"time": 45.24 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0680e-05", |
|
"loss": 0.1172, |
|
"slid_loss": 0.1147, |
|
"step": 1117, |
|
"time": 46.4 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0660e-05", |
|
"loss": 0.1107, |
|
"slid_loss": 0.1147, |
|
"step": 1118, |
|
"time": 45.15 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0641e-05", |
|
"loss": 0.1127, |
|
"slid_loss": 0.1146, |
|
"step": 1119, |
|
"time": 48.25 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0622e-05", |
|
"loss": 0.1221, |
|
"slid_loss": 0.1146, |
|
"step": 1120, |
|
"time": 45.73 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": "2.0604e-05", |
|
"loss": 0.1115, |
|
"slid_loss": 0.1146, |
|
"step": 1121, |
|
"time": 47.47 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0585e-05", |
|
"loss": 0.1149, |
|
"slid_loss": 0.1146, |
|
"step": 1122, |
|
"time": 46.26 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0567e-05", |
|
"loss": 0.12, |
|
"slid_loss": 0.1146, |
|
"step": 1123, |
|
"time": 44.85 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0549e-05", |
|
"loss": 0.1147, |
|
"slid_loss": 0.1145, |
|
"step": 1124, |
|
"time": 48.0 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0532e-05", |
|
"loss": 0.1107, |
|
"slid_loss": 0.1144, |
|
"step": 1125, |
|
"time": 46.44 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0514e-05", |
|
"loss": 0.1048, |
|
"slid_loss": 0.1143, |
|
"step": 1126, |
|
"time": 47.71 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0498e-05", |
|
"loss": 0.1152, |
|
"slid_loss": 0.1143, |
|
"step": 1127, |
|
"time": 47.88 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0481e-05", |
|
"loss": 0.1091, |
|
"slid_loss": 0.1143, |
|
"step": 1128, |
|
"time": 47.35 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0464e-05", |
|
"loss": 0.1254, |
|
"slid_loss": 0.1143, |
|
"step": 1129, |
|
"time": 47.7 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0448e-05", |
|
"loss": 0.1193, |
|
"slid_loss": 0.1144, |
|
"step": 1130, |
|
"time": 45.63 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0432e-05", |
|
"loss": 0.1174, |
|
"slid_loss": 0.1145, |
|
"step": 1131, |
|
"time": 46.13 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0417e-05", |
|
"loss": 0.1174, |
|
"slid_loss": 0.1144, |
|
"step": 1132, |
|
"time": 46.23 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": "2.0402e-05", |
|
"loss": 0.1123, |
|
"slid_loss": 0.1144, |
|
"step": 1133, |
|
"time": 47.42 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0387e-05", |
|
"loss": 0.1105, |
|
"slid_loss": 0.1144, |
|
"step": 1134, |
|
"time": 47.95 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0372e-05", |
|
"loss": 0.1154, |
|
"slid_loss": 0.1144, |
|
"step": 1135, |
|
"time": 45.59 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0358e-05", |
|
"loss": 0.1167, |
|
"slid_loss": 0.1145, |
|
"step": 1136, |
|
"time": 47.05 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0343e-05", |
|
"loss": 0.1145, |
|
"slid_loss": 0.1144, |
|
"step": 1137, |
|
"time": 47.51 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0330e-05", |
|
"loss": 0.1123, |
|
"slid_loss": 0.1144, |
|
"step": 1138, |
|
"time": 47.18 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0316e-05", |
|
"loss": 0.1078, |
|
"slid_loss": 0.1143, |
|
"step": 1139, |
|
"time": 48.45 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0303e-05", |
|
"loss": 0.1095, |
|
"slid_loss": 0.1143, |
|
"step": 1140, |
|
"time": 46.43 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0290e-05", |
|
"loss": 0.1101, |
|
"slid_loss": 0.1143, |
|
"step": 1141, |
|
"time": 46.39 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0277e-05", |
|
"loss": 0.1234, |
|
"slid_loss": 0.1144, |
|
"step": 1142, |
|
"time": 46.22 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0265e-05", |
|
"loss": 0.1055, |
|
"slid_loss": 0.1143, |
|
"step": 1143, |
|
"time": 48.46 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0252e-05", |
|
"loss": 0.1198, |
|
"slid_loss": 0.1144, |
|
"step": 1144, |
|
"time": 46.7 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": "2.0241e-05", |
|
"loss": 0.1056, |
|
"slid_loss": 0.1143, |
|
"step": 1145, |
|
"time": 45.29 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0229e-05", |
|
"loss": 0.116, |
|
"slid_loss": 0.1143, |
|
"step": 1146, |
|
"time": 46.78 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0218e-05", |
|
"loss": 0.1062, |
|
"slid_loss": 0.1143, |
|
"step": 1147, |
|
"time": 47.94 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0207e-05", |
|
"loss": 0.1157, |
|
"slid_loss": 0.1143, |
|
"step": 1148, |
|
"time": 46.77 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0196e-05", |
|
"loss": 0.1127, |
|
"slid_loss": 0.1144, |
|
"step": 1149, |
|
"time": 45.72 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0185e-05", |
|
"loss": 0.1199, |
|
"slid_loss": 0.1145, |
|
"step": 1150, |
|
"time": 47.32 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0175e-05", |
|
"loss": 0.1127, |
|
"slid_loss": 0.1145, |
|
"step": 1151, |
|
"time": 44.74 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0165e-05", |
|
"loss": 0.1115, |
|
"slid_loss": 0.1145, |
|
"step": 1152, |
|
"time": 45.66 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0156e-05", |
|
"loss": 0.1042, |
|
"slid_loss": 0.1144, |
|
"step": 1153, |
|
"time": 45.44 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0147e-05", |
|
"loss": 0.1144, |
|
"slid_loss": 0.1143, |
|
"step": 1154, |
|
"time": 47.6 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0138e-05", |
|
"loss": 0.1159, |
|
"slid_loss": 0.1143, |
|
"step": 1155, |
|
"time": 46.33 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0129e-05", |
|
"loss": 0.1123, |
|
"slid_loss": 0.1143, |
|
"step": 1156, |
|
"time": 47.49 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": "2.0120e-05", |
|
"loss": 0.1144, |
|
"slid_loss": 0.1143, |
|
"step": 1157, |
|
"time": 47.53 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0112e-05", |
|
"loss": 0.1138, |
|
"slid_loss": 0.1143, |
|
"step": 1158, |
|
"time": 47.61 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0104e-05", |
|
"loss": 0.1161, |
|
"slid_loss": 0.1143, |
|
"step": 1159, |
|
"time": 47.37 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0097e-05", |
|
"loss": 0.1179, |
|
"slid_loss": 0.1144, |
|
"step": 1160, |
|
"time": 46.91 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0089e-05", |
|
"loss": 0.119, |
|
"slid_loss": 0.1144, |
|
"step": 1161, |
|
"time": 45.47 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0082e-05", |
|
"loss": 0.1113, |
|
"slid_loss": 0.1143, |
|
"step": 1162, |
|
"time": 47.74 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0076e-05", |
|
"loss": 0.1136, |
|
"slid_loss": 0.1144, |
|
"step": 1163, |
|
"time": 45.57 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0069e-05", |
|
"loss": 0.1156, |
|
"slid_loss": 0.1144, |
|
"step": 1164, |
|
"time": 47.75 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0063e-05", |
|
"loss": 0.118, |
|
"slid_loss": 0.1143, |
|
"step": 1165, |
|
"time": 46.14 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0057e-05", |
|
"loss": 0.1091, |
|
"slid_loss": 0.1143, |
|
"step": 1166, |
|
"time": 45.85 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0052e-05", |
|
"loss": 0.1072, |
|
"slid_loss": 0.1142, |
|
"step": 1167, |
|
"time": 46.69 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0046e-05", |
|
"loss": 0.1128, |
|
"slid_loss": 0.1142, |
|
"step": 1168, |
|
"time": 47.05 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": "2.0041e-05", |
|
"loss": 0.1165, |
|
"slid_loss": 0.1142, |
|
"step": 1169, |
|
"time": 48.52 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0037e-05", |
|
"loss": 0.1176, |
|
"slid_loss": 0.1143, |
|
"step": 1170, |
|
"time": 45.19 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0032e-05", |
|
"loss": 0.1156, |
|
"slid_loss": 0.1143, |
|
"step": 1171, |
|
"time": 48.84 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0028e-05", |
|
"loss": 0.1182, |
|
"slid_loss": 0.1143, |
|
"step": 1172, |
|
"time": 46.05 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0024e-05", |
|
"loss": 0.1156, |
|
"slid_loss": 0.1143, |
|
"step": 1173, |
|
"time": 45.28 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0021e-05", |
|
"loss": 0.1119, |
|
"slid_loss": 0.1142, |
|
"step": 1174, |
|
"time": 47.03 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0017e-05", |
|
"loss": 0.1131, |
|
"slid_loss": 0.1142, |
|
"step": 1175, |
|
"time": 46.41 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0014e-05", |
|
"loss": 0.1171, |
|
"slid_loss": 0.1143, |
|
"step": 1176, |
|
"time": 47.64 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0012e-05", |
|
"loss": 0.1137, |
|
"slid_loss": 0.1143, |
|
"step": 1177, |
|
"time": 47.6 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0009e-05", |
|
"loss": 0.1142, |
|
"slid_loss": 0.1143, |
|
"step": 1178, |
|
"time": 45.89 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0007e-05", |
|
"loss": 0.1107, |
|
"slid_loss": 0.1143, |
|
"step": 1179, |
|
"time": 47.36 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": "2.0005e-05", |
|
"loss": 0.1177, |
|
"slid_loss": 0.1143, |
|
"step": 1180, |
|
"time": 46.62 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": "2.0004e-05", |
|
"loss": 0.1089, |
|
"slid_loss": 0.1143, |
|
"step": 1181, |
|
"time": 47.25 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": "2.0002e-05", |
|
"loss": 0.1056, |
|
"slid_loss": 0.1142, |
|
"step": 1182, |
|
"time": 47.04 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": "2.0001e-05", |
|
"loss": 0.1204, |
|
"slid_loss": 0.1143, |
|
"step": 1183, |
|
"time": 47.78 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": "2.0001e-05", |
|
"loss": 0.1193, |
|
"slid_loss": 0.1143, |
|
"step": 1184, |
|
"time": 46.95 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": "2.0000e-05", |
|
"loss": 0.114, |
|
"slid_loss": 0.1143, |
|
"step": 1185, |
|
"time": 45.8 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": "2.0000e-05", |
|
"loss": 0.1138, |
|
"slid_loss": 0.1144, |
|
"step": 1186, |
|
"time": 47.2 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 1186, |
|
"time": 1.43, |
|
"total_flos": 0.0, |
|
"train_loss": 0.12679563476554861, |
|
"train_runtime": 55610.9881, |
|
"train_samples_per_second": 5.464, |
|
"train_steps_per_second": 0.021 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 1186, |
|
"num_train_epochs": 1, |
|
"save_steps": 600, |
|
"total_flos": 0.0, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|