danieldk-explosion commited on
Commit
e50c460
1 Parent(s): d53e01a

Try to fixup causal lm

Browse files
Files changed (3) hide show
  1. config.json +5 -2
  2. generation_config.json +6 -0
  3. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alibi": false,
3
  "architectures": [
4
- "FalconModel"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bias": false,
@@ -11,6 +11,7 @@
11
  "hidden_size": 256,
12
  "initializer_range": 0.02,
13
  "layer_norm_epsilon": 1e-05,
 
14
  "model_type": "falcon",
15
  "multi_query": true,
16
  "new_decoder_architecture": true,
@@ -18,8 +19,10 @@
18
  "num_hidden_layers": 5,
19
  "num_kv_heads": 2,
20
  "parallel_attn": true,
 
 
21
  "torch_dtype": "float32",
22
- "transformers_version": "4.32.0.dev0",
23
  "use_cache": true,
24
  "vocab_size": 1024
25
  }
 
1
  {
2
  "alibi": false,
3
  "architectures": [
4
+ "FalconForCausalLM"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bias": false,
 
11
  "hidden_size": 256,
12
  "initializer_range": 0.02,
13
  "layer_norm_epsilon": 1e-05,
14
+ "max_position_embeddings": 2048,
15
  "model_type": "falcon",
16
  "multi_query": true,
17
  "new_decoder_architecture": true,
 
19
  "num_hidden_layers": 5,
20
  "num_kv_heads": 2,
21
  "parallel_attn": true,
22
+ "rope_scaling": null,
23
+ "rope_theta": 10000.0,
24
  "torch_dtype": "float32",
25
+ "transformers_version": "4.34.0.dev0",
26
  "use_cache": true,
27
  "vocab_size": 1024
28
  }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 11,
4
+ "eos_token_id": 11,
5
+ "transformers_version": "4.34.0.dev0"
6
+ }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0401b0bba6dd4f646231630ebb2a000afc4ca2d77b4ff96c841834ac906753f8
3
- size 15501938
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1248a031fea1cea953b8992e6ec7e595657e8bbbc9c39f2796d788a1be000bc1
3
+ size 15502514