sabalanya commited on
Commit
84123e5
1 Parent(s): 73eafd8

Push model using huggingface_hub.

Browse files
Files changed (2) hide show
  1. config.json +18 -18
  2. model.safetensors +1 -1
config.json CHANGED
@@ -1,28 +1,28 @@
1
  {
2
- "attn_eq": false,
3
- "attn_type": "trittention",
4
- "autopad": true,
5
- "causal_attn": true,
6
- "d_head": 64,
7
- "d_mlp": 2048,
8
  "d_model": 768,
9
- "d_vocab": 50257,
10
  "debug": true,
11
- "dropout": 0.1,
12
- "dt_head": 64,
13
- "freqs_cis": null,
14
- "has_mlp": true,
15
- "init_range": 0.02,
16
- "is_gated": false,
17
  "layer_norm_eps": 1e-05,
18
- "look_backward": 1,
19
- "mlp_type": "all",
20
  "n_ctx": 128,
 
 
 
 
 
21
  "n_heads": 12,
22
- "n_layers": 1,
23
  "nt_heads": 2,
 
 
 
 
 
 
24
  "order_attn": true,
25
- "pad_value": 0,
26
  "window_size": 16,
27
- "with_ln": true
 
 
 
28
  }
 
1
  {
 
 
 
 
 
 
2
  "d_model": 768,
 
3
  "debug": true,
 
 
 
 
 
 
4
  "layer_norm_eps": 1e-05,
5
+ "d_vocab": 50257,
6
+ "init_range": 0.02,
7
  "n_ctx": 128,
8
+ "d_head": 64,
9
+ "dt_head": 64,
10
+ "d_mlp": 2048,
11
+ "causal_attn": true,
12
+ "attn_type": "trittention",
13
  "n_heads": 12,
 
14
  "nt_heads": 2,
15
+ "n_layers": 1,
16
+ "dropout": 0.1,
17
+ "mlp_type": "all",
18
+ "with_ln": true,
19
+ "is_gated": false,
20
+ "has_mlp": true,
21
  "order_attn": true,
22
+ "attn_eq": false,
23
  "window_size": 16,
24
+ "look_backward": 1,
25
+ "pad_value": 0,
26
+ "autopad": true,
27
+ "freqs_cis": null
28
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb9a89ead2f235a8979de4ad855b32e6dad1ce6d124a3ce1a87da582e7173f94
3
  size 336159192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6012ebc6498ee14a83e8a4cb09f17462774948018f3d3a9e792786bfd0ba220
3
  size 336159192