Federic commited on
Commit
039cc31
1 Parent(s): 034fa04

Training in progress, step 325, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c7fede2ff68f9e472a2c2b349a9cbde0a4d12580727361822f038b204a98f85
3
  size 838904832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58bfc760e91b0d21d65d7923896fc95c4f25a4f944fedcb622d03de278369d4c
3
  size 838904832
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63f7fde2e75f42c7ea00494aae5de88398db12c3a2b9a16648ef1ab7106b12c3
3
  size 421458386
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a938569403d232d9b7ce7418e5911974858cee98b87ac1d1130486b3b546ff4
3
  size 421458386
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f3cd2689612d275f444201619598c0d2ff3e0c02732d05a46fb456bcfae2c9b
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef8593826016a9c5e91074c43813975f98ccc4aaa6700967895ddedc6022efdd
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c79c66b60a0abe3dc1f1792ced2b6c99f10b3ada4ba94ee60000ba5931c603a9
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d0166e84a24bb5e8fbd3eec4a559ab0d784d0de1c9f1ce37bbb473bd77a0781
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.2,
5
  "eval_steps": 500,
6
- "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1807,13 +1807,163 @@
1807
  "learning_rate": 0.0002,
1808
  "loss": 0.3419,
1809
  "step": 300
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1810
  }
1811
  ],
1812
  "logging_steps": 1,
1813
  "max_steps": 500,
1814
  "num_train_epochs": 2,
1815
  "save_steps": 25,
1816
- "total_flos": 3.589773546369024e+16,
1817
  "trial_name": null,
1818
  "trial_params": null
1819
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 1.3,
5
  "eval_steps": 500,
6
+ "global_step": 325,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1807
  "learning_rate": 0.0002,
1808
  "loss": 0.3419,
1809
  "step": 300
1810
+ },
1811
+ {
1812
+ "epoch": 1.2,
1813
+ "learning_rate": 0.0002,
1814
+ "loss": 0.6055,
1815
+ "step": 301
1816
+ },
1817
+ {
1818
+ "epoch": 1.21,
1819
+ "learning_rate": 0.0002,
1820
+ "loss": 0.6022,
1821
+ "step": 302
1822
+ },
1823
+ {
1824
+ "epoch": 1.21,
1825
+ "learning_rate": 0.0002,
1826
+ "loss": 0.5849,
1827
+ "step": 303
1828
+ },
1829
+ {
1830
+ "epoch": 1.22,
1831
+ "learning_rate": 0.0002,
1832
+ "loss": 0.4591,
1833
+ "step": 304
1834
+ },
1835
+ {
1836
+ "epoch": 1.22,
1837
+ "learning_rate": 0.0002,
1838
+ "loss": 0.5527,
1839
+ "step": 305
1840
+ },
1841
+ {
1842
+ "epoch": 1.22,
1843
+ "learning_rate": 0.0002,
1844
+ "loss": 0.4893,
1845
+ "step": 306
1846
+ },
1847
+ {
1848
+ "epoch": 1.23,
1849
+ "learning_rate": 0.0002,
1850
+ "loss": 0.4901,
1851
+ "step": 307
1852
+ },
1853
+ {
1854
+ "epoch": 1.23,
1855
+ "learning_rate": 0.0002,
1856
+ "loss": 0.5406,
1857
+ "step": 308
1858
+ },
1859
+ {
1860
+ "epoch": 1.24,
1861
+ "learning_rate": 0.0002,
1862
+ "loss": 0.5152,
1863
+ "step": 309
1864
+ },
1865
+ {
1866
+ "epoch": 1.24,
1867
+ "learning_rate": 0.0002,
1868
+ "loss": 0.5013,
1869
+ "step": 310
1870
+ },
1871
+ {
1872
+ "epoch": 1.24,
1873
+ "learning_rate": 0.0002,
1874
+ "loss": 0.5074,
1875
+ "step": 311
1876
+ },
1877
+ {
1878
+ "epoch": 1.25,
1879
+ "learning_rate": 0.0002,
1880
+ "loss": 0.4718,
1881
+ "step": 312
1882
+ },
1883
+ {
1884
+ "epoch": 1.25,
1885
+ "learning_rate": 0.0002,
1886
+ "loss": 0.4602,
1887
+ "step": 313
1888
+ },
1889
+ {
1890
+ "epoch": 1.26,
1891
+ "learning_rate": 0.0002,
1892
+ "loss": 0.4424,
1893
+ "step": 314
1894
+ },
1895
+ {
1896
+ "epoch": 1.26,
1897
+ "learning_rate": 0.0002,
1898
+ "loss": 0.4844,
1899
+ "step": 315
1900
+ },
1901
+ {
1902
+ "epoch": 1.26,
1903
+ "learning_rate": 0.0002,
1904
+ "loss": 0.5499,
1905
+ "step": 316
1906
+ },
1907
+ {
1908
+ "epoch": 1.27,
1909
+ "learning_rate": 0.0002,
1910
+ "loss": 0.5265,
1911
+ "step": 317
1912
+ },
1913
+ {
1914
+ "epoch": 1.27,
1915
+ "learning_rate": 0.0002,
1916
+ "loss": 0.4419,
1917
+ "step": 318
1918
+ },
1919
+ {
1920
+ "epoch": 1.28,
1921
+ "learning_rate": 0.0002,
1922
+ "loss": 0.4753,
1923
+ "step": 319
1924
+ },
1925
+ {
1926
+ "epoch": 1.28,
1927
+ "learning_rate": 0.0002,
1928
+ "loss": 0.493,
1929
+ "step": 320
1930
+ },
1931
+ {
1932
+ "epoch": 1.28,
1933
+ "learning_rate": 0.0002,
1934
+ "loss": 0.4885,
1935
+ "step": 321
1936
+ },
1937
+ {
1938
+ "epoch": 1.29,
1939
+ "learning_rate": 0.0002,
1940
+ "loss": 0.464,
1941
+ "step": 322
1942
+ },
1943
+ {
1944
+ "epoch": 1.29,
1945
+ "learning_rate": 0.0002,
1946
+ "loss": 0.4767,
1947
+ "step": 323
1948
+ },
1949
+ {
1950
+ "epoch": 1.3,
1951
+ "learning_rate": 0.0002,
1952
+ "loss": 0.4766,
1953
+ "step": 324
1954
+ },
1955
+ {
1956
+ "epoch": 1.3,
1957
+ "learning_rate": 0.0002,
1958
+ "loss": 0.4449,
1959
+ "step": 325
1960
  }
1961
  ],
1962
  "logging_steps": 1,
1963
  "max_steps": 500,
1964
  "num_train_epochs": 2,
1965
  "save_steps": 25,
1966
+ "total_flos": 3.921061058162688e+16,
1967
  "trial_name": null,
1968
  "trial_params": null
1969
  }