Yishi Xu commited on
Commit
b3d4aee
1 Parent(s): 4849c5d

add CerebrasGPT-2.7B_SP

Browse files
CerebrasGPT-2.7B_SP/checkpoint_10000_to_hf.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f8ba86cf7d8a26a76be42148da3fc0088a29b8f2b5f2f3c6a7fffffea897d52
3
+ size 10740579615
CerebrasGPT-2.7B_SP/checkpoint_20000_to_hf.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f7d17c9e697794347ad4442066be07496708a10379d03a960fee246508eff00
3
+ size 10740579615
CerebrasGPT-2.7B_SP/checkpoint_30000_to_hf.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74b698cff427f2e2074e0032ab4ad245a3fac7d7bf484de6c48796ff1ec04f72
3
+ size 10740579615
CerebrasGPT-2.7B_SP/checkpoint_40000_to_hf.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d451ae3b5c0f3481e50ea88aa26eb55328d9ea1e5670a5f582227a14dff24646
3
+ size 10740579615
CerebrasGPT-2.7B_SP/params_train_to_hf.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "attn_pdrop": 0.0,
3
+ "scale_attn_weights": true,
4
+ "resid_pdrop": 0.0,
5
+ "n_inner": 10240,
6
+ "n_embd": 2560,
7
+ "layer_norm_epsilon": 1e-05,
8
+ "n_positions": 2048,
9
+ "activation_function": "gelu",
10
+ "n_head": 32,
11
+ "n_layer": 32,
12
+ "tie_word_embeddings": true,
13
+ "vocab_size": 50257,
14
+ "embd_pdrop": 0.0
15
+ }