Outer Step 27. Inner Step 2. Batch Size 480
Browse files- config.json +7 -7
- model.safetensors +1 -1
config.json
CHANGED
@@ -28,19 +28,19 @@
|
|
28 |
"AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
|
29 |
},
|
30 |
"block_list": [
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
],
|
37 |
"block_size": 1024,
|
38 |
"bos_token_id": 50256,
|
39 |
"embd_pdrop": 0.1,
|
40 |
"eos_token_id": 50256,
|
41 |
"initializer_range": 0.02,
|
42 |
-
"inner_step":
|
43 |
-
"last_allreduce_block":
|
44 |
"layer_norm_epsilon": 1e-05,
|
45 |
"model_type": "gpt_optimized",
|
46 |
"n_embd": 1280,
|
|
|
28 |
"AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
|
29 |
},
|
30 |
"block_list": [
|
31 |
+
4019772,
|
32 |
+
4019777,
|
33 |
+
4019781,
|
34 |
+
4019786,
|
35 |
+
4019790
|
36 |
],
|
37 |
"block_size": 1024,
|
38 |
"bos_token_id": 50256,
|
39 |
"embd_pdrop": 0.1,
|
40 |
"eos_token_id": 50256,
|
41 |
"initializer_range": 0.02,
|
42 |
+
"inner_step": 2,
|
43 |
+
"last_allreduce_block": 4019541,
|
44 |
"layer_norm_epsilon": 1e-05,
|
45 |
"model_type": "gpt_optimized",
|
46 |
"n_embd": 1280,
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4040701744
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a5bb9c54e7dfcf15dee838799051169d44fe759b7db8441f751db256ac5a878
|
3 |
size 4040701744
|