kmfoda commited on
Commit
78eb49b
·
verified ·
1 Parent(s): 794c847

Outer Step 27. Inner Step 2. Batch Size 480

Browse files
Files changed (2) hide show
  1. config.json +7 -7
  2. model.safetensors +1 -1
config.json CHANGED
@@ -28,19 +28,19 @@
28
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
29
  },
30
  "block_list": [
31
- 4019687,
32
- 4019692,
33
- 4019696,
34
- 4019701,
35
- 4019705
36
  ],
37
  "block_size": 1024,
38
  "bos_token_id": 50256,
39
  "embd_pdrop": 0.1,
40
  "eos_token_id": 50256,
41
  "initializer_range": 0.02,
42
- "inner_step": 8,
43
- "last_allreduce_block": 4019290,
44
  "layer_norm_epsilon": 1e-05,
45
  "model_type": "gpt_optimized",
46
  "n_embd": 1280,
 
28
  "AutoModelForCausalLM": "distributed/optimized-gpt2-500m--modeling_gpt_optimized.GPTOptim"
29
  },
30
  "block_list": [
31
+ 4019772,
32
+ 4019777,
33
+ 4019781,
34
+ 4019786,
35
+ 4019790
36
  ],
37
  "block_size": 1024,
38
  "bos_token_id": 50256,
39
  "embd_pdrop": 0.1,
40
  "eos_token_id": 50256,
41
  "initializer_range": 0.02,
42
+ "inner_step": 2,
43
+ "last_allreduce_block": 4019541,
44
  "layer_norm_epsilon": 1e-05,
45
  "model_type": "gpt_optimized",
46
  "n_embd": 1280,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de8f4a65cfcc48dad69038eb1fae7daf1c42a18490856354b53ca16765625587
3
  size 4040701744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a5bb9c54e7dfcf15dee838799051169d44fe759b7db8441f751db256ac5a878
3
  size 4040701744