ygaci commited on
Commit
9e20d93
·
verified ·
1 Parent(s): fd49d5e

Training in progress, epoch 19, checkpoint

Browse files
last-checkpoint/adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dacbb11ee26d39a11b078b4e41b186858595a541a46afa40d1d56f8d3f9d09cb
3
  size 778185130
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ae3a43f4efec245b701802773fa3672d33222c0901aefd3d8720e17deb1370c
3
  size 778185130
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc3648b7441555f2f1de64686d5391ee6f025ef43d67eb1f21801a5f2feeaa26
3
  size 395561780
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ed2ba4f7a29e297ae552e10f040d19b71f57c648e3efd9104e8940c29e0d6b1
3
  size 395561780
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff2a98b7b58dd272a86869334fd0edf641ab47ceb102b634b242c3ff92151a26
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:774612f2d4811ca7da639cec8d3b509c1f87b5ffc57761546af6466447a0619a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91dab35b307108e4bae50d4534d800057f9a6bba8495c6b805b7d97ab7a4077d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dfbf64ae704679e934ba0ab136667c219c97c5f82e2ab9b372076828c19fea3
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.4693966805934906,
3
  "best_model_checkpoint": "/tmp/tmp6i73fvu2/checkpoint-5596",
4
- "epoch": 18.0,
5
  "eval_steps": 500,
6
- "global_step": 25182,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1908,6 +1908,112 @@
1908
  "eval_samples_per_second": 3.578,
1909
  "eval_steps_per_second": 3.578,
1910
  "step": 25182
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1911
  }
1912
  ],
1913
  "logging_steps": 100,
@@ -1927,7 +2033,7 @@
1927
  "attributes": {}
1928
  }
1929
  },
1930
- "total_flos": 1.685860856763515e+17,
1931
  "train_batch_size": 1,
1932
  "trial_name": null,
1933
  "trial_params": null
 
1
  {
2
  "best_metric": 0.4693966805934906,
3
  "best_model_checkpoint": "/tmp/tmp6i73fvu2/checkpoint-5596",
4
+ "epoch": 19.0,
5
  "eval_steps": 500,
6
+ "global_step": 26581,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1908
  "eval_samples_per_second": 3.578,
1909
  "eval_steps_per_second": 3.578,
1910
  "step": 25182
1911
+ },
1912
+ {
1913
+ "epoch": 18.01286633309507,
1914
+ "grad_norm": 0.08682116866111755,
1915
+ "learning_rate": 4.989258861439313e-06,
1916
+ "loss": 0.3537,
1917
+ "step": 25200
1918
+ },
1919
+ {
1920
+ "epoch": 18.084345961401002,
1921
+ "grad_norm": 0.16758058965206146,
1922
+ "learning_rate": 4.810239885427856e-06,
1923
+ "loss": 0.3516,
1924
+ "step": 25300
1925
+ },
1926
+ {
1927
+ "epoch": 18.155825589706932,
1928
+ "grad_norm": 0.12864187359809875,
1929
+ "learning_rate": 4.631220909416398e-06,
1930
+ "loss": 0.345,
1931
+ "step": 25400
1932
+ },
1933
+ {
1934
+ "epoch": 18.227305218012866,
1935
+ "grad_norm": 0.12607234716415405,
1936
+ "learning_rate": 4.4522019334049415e-06,
1937
+ "loss": 0.3448,
1938
+ "step": 25500
1939
+ },
1940
+ {
1941
+ "epoch": 18.2987848463188,
1942
+ "grad_norm": 0.17426550388336182,
1943
+ "learning_rate": 4.273182957393484e-06,
1944
+ "loss": 0.3385,
1945
+ "step": 25600
1946
+ },
1947
+ {
1948
+ "epoch": 18.370264474624733,
1949
+ "grad_norm": 0.16788093745708466,
1950
+ "learning_rate": 4.0941639813820265e-06,
1951
+ "loss": 0.353,
1952
+ "step": 25700
1953
+ },
1954
+ {
1955
+ "epoch": 18.441744102930663,
1956
+ "grad_norm": 0.09532229602336884,
1957
+ "learning_rate": 3.915145005370569e-06,
1958
+ "loss": 0.3564,
1959
+ "step": 25800
1960
+ },
1961
+ {
1962
+ "epoch": 18.513223731236597,
1963
+ "grad_norm": 0.1609189808368683,
1964
+ "learning_rate": 3.7361260293591127e-06,
1965
+ "loss": 0.3567,
1966
+ "step": 25900
1967
+ },
1968
+ {
1969
+ "epoch": 18.58470335954253,
1970
+ "grad_norm": 0.16414226591587067,
1971
+ "learning_rate": 3.5571070533476552e-06,
1972
+ "loss": 0.3606,
1973
+ "step": 26000
1974
+ },
1975
+ {
1976
+ "epoch": 18.656182987848464,
1977
+ "grad_norm": 0.17029887437820435,
1978
+ "learning_rate": 3.378088077336198e-06,
1979
+ "loss": 0.3595,
1980
+ "step": 26100
1981
+ },
1982
+ {
1983
+ "epoch": 18.727662616154397,
1984
+ "grad_norm": 0.1379202902317047,
1985
+ "learning_rate": 3.1990691013247406e-06,
1986
+ "loss": 0.3475,
1987
+ "step": 26200
1988
+ },
1989
+ {
1990
+ "epoch": 18.799142244460327,
1991
+ "grad_norm": 0.12663400173187256,
1992
+ "learning_rate": 3.0200501253132835e-06,
1993
+ "loss": 0.3564,
1994
+ "step": 26300
1995
+ },
1996
+ {
1997
+ "epoch": 18.87062187276626,
1998
+ "grad_norm": 0.12856705486774445,
1999
+ "learning_rate": 2.841031149301826e-06,
2000
+ "loss": 0.3558,
2001
+ "step": 26400
2002
+ },
2003
+ {
2004
+ "epoch": 18.942101501072194,
2005
+ "grad_norm": 0.1797301322221756,
2006
+ "learning_rate": 2.662012173290369e-06,
2007
+ "loss": 0.3583,
2008
+ "step": 26500
2009
+ },
2010
+ {
2011
+ "epoch": 19.0,
2012
+ "eval_loss": 0.573534369468689,
2013
+ "eval_runtime": 195.7446,
2014
+ "eval_samples_per_second": 3.576,
2015
+ "eval_steps_per_second": 3.576,
2016
+ "step": 26581
2017
  }
2018
  ],
2019
  "logging_steps": 100,
 
2033
  "attributes": {}
2034
  }
2035
  },
2036
+ "total_flos": 1.7795197925064704e+17,
2037
  "train_batch_size": 1,
2038
  "trial_name": null,
2039
  "trial_params": null