Training in progress, step 325, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 838904832
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:58bfc760e91b0d21d65d7923896fc95c4f25a4f944fedcb622d03de278369d4c
|
3 |
size 838904832
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 421458386
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a938569403d232d9b7ce7418e5911974858cee98b87ac1d1130486b3b546ff4
|
3 |
size 421458386
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ef8593826016a9c5e91074c43813975f98ccc4aaa6700967895ddedc6022efdd
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d0166e84a24bb5e8fbd3eec4a559ab0d784d0de1c9f1ce37bbb473bd77a0781
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
-
"epoch": 1.
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -1807,13 +1807,163 @@
|
|
1807 |
"learning_rate": 0.0002,
|
1808 |
"loss": 0.3419,
|
1809 |
"step": 300
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1810 |
}
|
1811 |
],
|
1812 |
"logging_steps": 1,
|
1813 |
"max_steps": 500,
|
1814 |
"num_train_epochs": 2,
|
1815 |
"save_steps": 25,
|
1816 |
-
"total_flos": 3.
|
1817 |
"trial_name": null,
|
1818 |
"trial_params": null
|
1819 |
}
|
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
+
"epoch": 1.3,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 325,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
1807 |
"learning_rate": 0.0002,
|
1808 |
"loss": 0.3419,
|
1809 |
"step": 300
|
1810 |
+
},
|
1811 |
+
{
|
1812 |
+
"epoch": 1.2,
|
1813 |
+
"learning_rate": 0.0002,
|
1814 |
+
"loss": 0.6055,
|
1815 |
+
"step": 301
|
1816 |
+
},
|
1817 |
+
{
|
1818 |
+
"epoch": 1.21,
|
1819 |
+
"learning_rate": 0.0002,
|
1820 |
+
"loss": 0.6022,
|
1821 |
+
"step": 302
|
1822 |
+
},
|
1823 |
+
{
|
1824 |
+
"epoch": 1.21,
|
1825 |
+
"learning_rate": 0.0002,
|
1826 |
+
"loss": 0.5849,
|
1827 |
+
"step": 303
|
1828 |
+
},
|
1829 |
+
{
|
1830 |
+
"epoch": 1.22,
|
1831 |
+
"learning_rate": 0.0002,
|
1832 |
+
"loss": 0.4591,
|
1833 |
+
"step": 304
|
1834 |
+
},
|
1835 |
+
{
|
1836 |
+
"epoch": 1.22,
|
1837 |
+
"learning_rate": 0.0002,
|
1838 |
+
"loss": 0.5527,
|
1839 |
+
"step": 305
|
1840 |
+
},
|
1841 |
+
{
|
1842 |
+
"epoch": 1.22,
|
1843 |
+
"learning_rate": 0.0002,
|
1844 |
+
"loss": 0.4893,
|
1845 |
+
"step": 306
|
1846 |
+
},
|
1847 |
+
{
|
1848 |
+
"epoch": 1.23,
|
1849 |
+
"learning_rate": 0.0002,
|
1850 |
+
"loss": 0.4901,
|
1851 |
+
"step": 307
|
1852 |
+
},
|
1853 |
+
{
|
1854 |
+
"epoch": 1.23,
|
1855 |
+
"learning_rate": 0.0002,
|
1856 |
+
"loss": 0.5406,
|
1857 |
+
"step": 308
|
1858 |
+
},
|
1859 |
+
{
|
1860 |
+
"epoch": 1.24,
|
1861 |
+
"learning_rate": 0.0002,
|
1862 |
+
"loss": 0.5152,
|
1863 |
+
"step": 309
|
1864 |
+
},
|
1865 |
+
{
|
1866 |
+
"epoch": 1.24,
|
1867 |
+
"learning_rate": 0.0002,
|
1868 |
+
"loss": 0.5013,
|
1869 |
+
"step": 310
|
1870 |
+
},
|
1871 |
+
{
|
1872 |
+
"epoch": 1.24,
|
1873 |
+
"learning_rate": 0.0002,
|
1874 |
+
"loss": 0.5074,
|
1875 |
+
"step": 311
|
1876 |
+
},
|
1877 |
+
{
|
1878 |
+
"epoch": 1.25,
|
1879 |
+
"learning_rate": 0.0002,
|
1880 |
+
"loss": 0.4718,
|
1881 |
+
"step": 312
|
1882 |
+
},
|
1883 |
+
{
|
1884 |
+
"epoch": 1.25,
|
1885 |
+
"learning_rate": 0.0002,
|
1886 |
+
"loss": 0.4602,
|
1887 |
+
"step": 313
|
1888 |
+
},
|
1889 |
+
{
|
1890 |
+
"epoch": 1.26,
|
1891 |
+
"learning_rate": 0.0002,
|
1892 |
+
"loss": 0.4424,
|
1893 |
+
"step": 314
|
1894 |
+
},
|
1895 |
+
{
|
1896 |
+
"epoch": 1.26,
|
1897 |
+
"learning_rate": 0.0002,
|
1898 |
+
"loss": 0.4844,
|
1899 |
+
"step": 315
|
1900 |
+
},
|
1901 |
+
{
|
1902 |
+
"epoch": 1.26,
|
1903 |
+
"learning_rate": 0.0002,
|
1904 |
+
"loss": 0.5499,
|
1905 |
+
"step": 316
|
1906 |
+
},
|
1907 |
+
{
|
1908 |
+
"epoch": 1.27,
|
1909 |
+
"learning_rate": 0.0002,
|
1910 |
+
"loss": 0.5265,
|
1911 |
+
"step": 317
|
1912 |
+
},
|
1913 |
+
{
|
1914 |
+
"epoch": 1.27,
|
1915 |
+
"learning_rate": 0.0002,
|
1916 |
+
"loss": 0.4419,
|
1917 |
+
"step": 318
|
1918 |
+
},
|
1919 |
+
{
|
1920 |
+
"epoch": 1.28,
|
1921 |
+
"learning_rate": 0.0002,
|
1922 |
+
"loss": 0.4753,
|
1923 |
+
"step": 319
|
1924 |
+
},
|
1925 |
+
{
|
1926 |
+
"epoch": 1.28,
|
1927 |
+
"learning_rate": 0.0002,
|
1928 |
+
"loss": 0.493,
|
1929 |
+
"step": 320
|
1930 |
+
},
|
1931 |
+
{
|
1932 |
+
"epoch": 1.28,
|
1933 |
+
"learning_rate": 0.0002,
|
1934 |
+
"loss": 0.4885,
|
1935 |
+
"step": 321
|
1936 |
+
},
|
1937 |
+
{
|
1938 |
+
"epoch": 1.29,
|
1939 |
+
"learning_rate": 0.0002,
|
1940 |
+
"loss": 0.464,
|
1941 |
+
"step": 322
|
1942 |
+
},
|
1943 |
+
{
|
1944 |
+
"epoch": 1.29,
|
1945 |
+
"learning_rate": 0.0002,
|
1946 |
+
"loss": 0.4767,
|
1947 |
+
"step": 323
|
1948 |
+
},
|
1949 |
+
{
|
1950 |
+
"epoch": 1.3,
|
1951 |
+
"learning_rate": 0.0002,
|
1952 |
+
"loss": 0.4766,
|
1953 |
+
"step": 324
|
1954 |
+
},
|
1955 |
+
{
|
1956 |
+
"epoch": 1.3,
|
1957 |
+
"learning_rate": 0.0002,
|
1958 |
+
"loss": 0.4449,
|
1959 |
+
"step": 325
|
1960 |
}
|
1961 |
],
|
1962 |
"logging_steps": 1,
|
1963 |
"max_steps": 500,
|
1964 |
"num_train_epochs": 2,
|
1965 |
"save_steps": 25,
|
1966 |
+
"total_flos": 3.921061058162688e+16,
|
1967 |
"trial_name": null,
|
1968 |
"trial_params": null
|
1969 |
}
|