Transformers
English
falcon
custom_code
text-generation-inference
erfanzar commited on
Commit
51f881f
1 Parent(s): 3364d40

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +9 -9
config.json CHANGED
@@ -4,22 +4,22 @@
4
  "FlaxAutoModel":"model.FlaxFalconModel",
5
  "FlaxAutoModelForCausalLM":"model.FlaxFalconForCausalLM"
6
  },
7
- "alibi": true,
8
  "apply_residual_connection_post_layernorm": false,
9
  "attention_dropout": 0.0,
10
- "bias": true,
11
- "bos_token_id": 1,
12
- "eos_token_id": 2,
13
  "hidden_dropout": 0.0,
14
- "hidden_size": 4096,
15
  "initializer_range": 0.02,
16
  "layer_norm_epsilon": 1e-05,
17
  "max_seq_len": 2048,
18
  "model_type": "falcon",
19
- "multi_query": false,
20
- "n_head": 64,
21
- "n_layer": 36,
22
- "parallel_attn": false,
23
  "transformers_version": "4.29.2",
24
  "use_cache": false,
25
  "vocab_size": 65024
 
4
  "FlaxAutoModel":"model.FlaxFalconModel",
5
  "FlaxAutoModelForCausalLM":"model.FlaxFalconForCausalLM"
6
  },
7
+ "alibi": false,
8
  "apply_residual_connection_post_layernorm": false,
9
  "attention_dropout": 0.0,
10
+ "bias": false,
11
+ "bos_token_id": 11,
12
+ "eos_token_id": 11,
13
  "hidden_dropout": 0.0,
14
+ "hidden_size": 4544,
15
  "initializer_range": 0.02,
16
  "layer_norm_epsilon": 1e-05,
17
  "max_seq_len": 2048,
18
  "model_type": "falcon",
19
+ "multi_query": true,
20
+ "n_head": 71,
21
+ "n_layer": 32,
22
+ "parallel_attn": true,
23
  "transformers_version": "4.29.2",
24
  "use_cache": false,
25
  "vocab_size": 65024