update to 400k steps
Browse files- global_step400007/mp_rank_00_model_states.pt +3 -0
- global_step400007/zero_pp_rank_0_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_10_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_11_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_12_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_13_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_14_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_15_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_1_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_2_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_3_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_4_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_5_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_6_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_7_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_8_mp_rank_00optim_states.pt +3 -0
- global_step400007/zero_pp_rank_9_mp_rank_00optim_states.pt +3 -0
- latest +1 -1
- pytorch_model.bin +1 -1
- trainer_state.json +2 -2
global_step400007/mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cfc42cb7857ecd7615de3807a59c4c3a08e11fcae9d0f00aed152e97d55a3463
|
3 |
+
size 734884708
|
global_step400007/zero_pp_rank_0_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e8ccad8ba76226c5bb49a7a43415b5e45c2036dcd0b888dd7d5324e1027a4ec
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_10_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:93403ae6ea45276a14a7f95d2c017918b1b6fc046a09d096fcfc670f895d7b13
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_11_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5cf42aa02f4fe46440a653158ffc3af5ec64d4e9e61994445b7e5775bc326e3f
|
3 |
+
size 266119228
|
global_step400007/zero_pp_rank_12_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:acf01e75a1152b33b6f75339050be4f2d20d32597276e9cdb03fbfa176585766
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_13_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:094b206901db1a184f300740a69435082aea9d449d4f55d57744b46251599730
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_14_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5f849525c6a58dd2c1bd47d0c5f103bb188b037ce993ebbdc7a17b7c0fc8d2fd
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_15_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:97474f3209a2c6caf303e3d03c8da200b8ea496e2e1c92936fb688e1c343721e
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_1_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:901885a130f6521cd68381a5695026d0af499dc0c0e85e38d70b3ca154847654
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_2_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c47e01651f91ecdcb40b4ebc3b9ffa8d3cff31c094ea30d89491b06fd002b277
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_3_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8340769a235d21c7c12b7cfeead26f338cb10d987b0845e061459950daadda89
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_4_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4717797466dfedf59dc5045d1caf656063791209bfad48fe2858aab354bd6391
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_5_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d8c9bbd80a31adc37b266b5f2f11d2f3dc938dc7b4ceb5b196d0fb059bf004f
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_6_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0a3528f1933c499292ef82abad84d5095e740240c11d11f831fc37ea91faf074
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_7_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7fdc5659842102990cb80a81af5fa175c487cf3d3ce7d290f93a747128aacca2
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_8_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3230c2a1c09a81f526965c7d878e0289c0f0bb4eb836ce955e85333c2d6778ab
|
3 |
+
size 266119230
|
global_step400007/zero_pp_rank_9_mp_rank_00optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:25e695ce6ed08e1e1c3fbd372869c9b559dada2cf03dce1fc9df57b9e921d6f8
|
3 |
+
size 266119230
|
latest
CHANGED
@@ -1 +1 @@
|
|
1 |
-
|
1 |
+
global_step400007
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 734884388
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8c57d9af77c25ceb39ad6d831beae49e2d60e34a7810f5c277c86a6422470626
|
3 |
size 734884388
|
trainer_state.json
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:011f48bc721d68de45596f4e8f6316a6911c94620f644687185b5ef7b26f9cea
|
3 |
+
size 27984612
|