Simran Arora commited on
Commit
4a9a435
1 Parent(s): 1c5f865
Files changed (2) hide show
  1. config.json +2 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+
2
+ {"d_model": 1024, "n_layer": 46, "vocab_size": 50277, "ssm_cfg": {}, "rms_norm": true, "residual_in_fp32": true, "fused_add_norm": true, "pad_vocab_size_multiple": 16, "reorder_and_upcast_attn": false, "scale_attn_by_inverse_layer_idx": true, "n_positions": 2048, "n_embd": 1024, "n_head": 16, "use_flash_attn": true, "fused_dropout_add_ln": true, "fused_mlp": true, "fused_bias_fc": true, "use_fast_path": true}
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:842fe2cd777e4387901b42b1283028cf6d0c6cda9f65ab48c35581590b5e092d
3
+ size 1432942417