BAAI
/

ldwang commited on
Commit
356733c
1 Parent(s): 2e69b5e

Update config for test...

Browse files
Files changed (1) hide show
  1. config.json +2 -6
config.json CHANGED
@@ -1,11 +1,7 @@
1
  {
2
  "architectures": [
3
- "AquilaForCausalLM"
4
  ],
5
- "auto_map": {
6
- "AutoConfig": "configuration_aquila.AquilaConfig",
7
- "AutoModelForCausalLM": "modeling_aquila.AquilaForCausalLM"
8
- },
9
  "bos_token_id": 100006,
10
  "eos_token_id": 100007,
11
  "hidden_act": "silu",
@@ -13,7 +9,7 @@
13
  "initializer_range": 0.02,
14
  "intermediate_size": 24576,
15
  "max_position_embeddings": 4096,
16
- "model_type": "aquila",
17
  "num_attention_heads": 48,
18
  "num_hidden_layers": 60,
19
  "num_key_value_heads": 8,
 
1
  {
2
  "architectures": [
3
+ "LlamaForCausalLM"
4
  ],
 
 
 
 
5
  "bos_token_id": 100006,
6
  "eos_token_id": 100007,
7
  "hidden_act": "silu",
 
9
  "initializer_range": 0.02,
10
  "intermediate_size": 24576,
11
  "max_position_embeddings": 4096,
12
+ "model_type": "llama",
13
  "num_attention_heads": 48,
14
  "num_hidden_layers": 60,
15
  "num_key_value_heads": 8,