rootxhacker commited on
Commit
d6c884b
·
verified ·
1 Parent(s): 6b42b1d

Training in progress, step 11000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4324a3a5915156493f27435981bc5e067e17329d248df8e005af28872089db08
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0355dc1903f621175efb8489eb47a7c37c5bff9cb7109326e9981d55521c5a79
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:586643082c797c6471cec4d99035ab6336a6eec2692a30ba5ac6aa1ea78f1ad3
3
  size 1544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6da8cff90cb20d49e633651c8eacef51daa69bb2f0ab5dff6d6115d2c7b51d28
3
  size 1544
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:435157f2e32de040f237ef44fcb2f7d2ce0740bd523e00f2682e862fe6e5b77b
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9d988b8500be5654d081e2939fa8df78c5f27a1a3b79fa26f57611a14b3851f
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:371a8ee92df296caadab094d1243e6f6a8e33531055d940b9b2beb58c3296d37
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d7167878bddb7c0a88c66ad0b0be73885dac80e1980348d5e68d660c4a16ce9
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5b29e642bf24d79ebcf705888b6c986c0f26a1dd38cc2f292d82caae0977305
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ba0e34e1ad6da2db5da67e847d92cda308dc91647db7b3e13d90fd53261cd77
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d497e82650965830e0a6f7b87b60635133c5dcd48d2ca0d74cebf5c5827dfab0
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1aa74d2c9c42f12a34e14e605ae41a06b5b6cec868a4a2206366586ebba8210c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 9500,
3
- "best_metric": 4.357193470001221,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-9500",
5
- "epoch": 0.8076301822936697,
6
  "eval_steps": 250,
7
- "global_step": 10500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1814,6 +1814,92 @@
1814
  "eval_samples_per_second": 54.063,
1815
  "eval_steps_per_second": 13.516,
1816
  "step": 10500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1817
  }
1818
  ],
1819
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 11000,
3
+ "best_metric": 4.352676868438721,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-11000",
5
+ "epoch": 0.8460887624028921,
6
  "eval_steps": 250,
7
+ "global_step": 11000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1814
  "eval_samples_per_second": 54.063,
1815
  "eval_steps_per_second": 13.516,
1816
  "step": 10500
1817
+ },
1818
+ {
1819
+ "epoch": 0.811476040304592,
1820
+ "grad_norm": 4.379317760467529,
1821
+ "learning_rate": 0.00014784821961925045,
1822
+ "loss": 4.6715,
1823
+ "step": 10550
1824
+ },
1825
+ {
1826
+ "epoch": 0.8153218983155142,
1827
+ "grad_norm": 3.034796714782715,
1828
+ "learning_rate": 0.00014758849959743396,
1829
+ "loss": 4.4511,
1830
+ "step": 10600
1831
+ },
1832
+ {
1833
+ "epoch": 0.8191677563264365,
1834
+ "grad_norm": 3.8016927242279053,
1835
+ "learning_rate": 0.00014732877957561748,
1836
+ "loss": 4.2966,
1837
+ "step": 10650
1838
+ },
1839
+ {
1840
+ "epoch": 0.8230136143373586,
1841
+ "grad_norm": 4.330080509185791,
1842
+ "learning_rate": 0.00014706905955380102,
1843
+ "loss": 4.6496,
1844
+ "step": 10700
1845
+ },
1846
+ {
1847
+ "epoch": 0.8268594723482809,
1848
+ "grad_norm": 8.032389640808105,
1849
+ "learning_rate": 0.00014680933953198453,
1850
+ "loss": 4.2758,
1851
+ "step": 10750
1852
+ },
1853
+ {
1854
+ "epoch": 0.8268594723482809,
1855
+ "eval_loss": 4.388455867767334,
1856
+ "eval_runtime": 18.481,
1857
+ "eval_samples_per_second": 54.11,
1858
+ "eval_steps_per_second": 13.527,
1859
+ "step": 10750
1860
+ },
1861
+ {
1862
+ "epoch": 0.8307053303592031,
1863
+ "grad_norm": 3.1724319458007812,
1864
+ "learning_rate": 0.00014654961951016807,
1865
+ "loss": 4.405,
1866
+ "step": 10800
1867
+ },
1868
+ {
1869
+ "epoch": 0.8345511883701254,
1870
+ "grad_norm": 3.529196262359619,
1871
+ "learning_rate": 0.00014628989948835155,
1872
+ "loss": 4.4433,
1873
+ "step": 10850
1874
+ },
1875
+ {
1876
+ "epoch": 0.8383970463810476,
1877
+ "grad_norm": 4.758362293243408,
1878
+ "learning_rate": 0.00014603017946653507,
1879
+ "loss": 4.5747,
1880
+ "step": 10900
1881
+ },
1882
+ {
1883
+ "epoch": 0.8422429043919698,
1884
+ "grad_norm": 3.524068832397461,
1885
+ "learning_rate": 0.0001457704594447186,
1886
+ "loss": 4.3469,
1887
+ "step": 10950
1888
+ },
1889
+ {
1890
+ "epoch": 0.8460887624028921,
1891
+ "grad_norm": 4.452401161193848,
1892
+ "learning_rate": 0.00014551073942290212,
1893
+ "loss": 4.3767,
1894
+ "step": 11000
1895
+ },
1896
+ {
1897
+ "epoch": 0.8460887624028921,
1898
+ "eval_loss": 4.352676868438721,
1899
+ "eval_runtime": 18.5817,
1900
+ "eval_samples_per_second": 53.816,
1901
+ "eval_steps_per_second": 13.454,
1902
+ "step": 11000
1903
  }
1904
  ],
1905
  "logging_steps": 50,