Commit
·
409299c
1
Parent(s):
4d2b40f
修改t5
Browse files
falcon-7b-instruct/bert4torch_config.json
CHANGED
@@ -17,7 +17,7 @@
|
|
17 |
"multi_query": true,
|
18 |
"num_attention_heads": 71,
|
19 |
"num_hidden_layers": 32,
|
20 |
-
"
|
21 |
"torch_dtype": "bfloat16",
|
22 |
"num_key_value_heads": 1,
|
23 |
"vocab_size": 65024,
|
|
|
17 |
"multi_query": true,
|
18 |
"num_attention_heads": 71,
|
19 |
"num_hidden_layers": 32,
|
20 |
+
"layer_type": "FalconParallelAttnLayer",
|
21 |
"torch_dtype": "bfloat16",
|
22 |
"num_key_value_heads": 1,
|
23 |
"vocab_size": 65024,
|
falcon-7b/bert4torch_config.json
CHANGED
@@ -17,7 +17,7 @@
|
|
17 |
"multi_query": true,
|
18 |
"num_attention_heads": 71,
|
19 |
"num_hidden_layers": 32,
|
20 |
-
"
|
21 |
"torch_dtype": "bfloat16",
|
22 |
"num_key_value_heads": 1,
|
23 |
"vocab_size": 65024,
|
|
|
17 |
"multi_query": true,
|
18 |
"num_attention_heads": 71,
|
19 |
"num_hidden_layers": 32,
|
20 |
+
"layer_type": "FalconParallelAttnLayer",
|
21 |
"torch_dtype": "bfloat16",
|
22 |
"num_key_value_heads": 1,
|
23 |
"vocab_size": 65024,
|
falcon-rw-1b/bert4torch_config.json
CHANGED
@@ -16,7 +16,6 @@
|
|
16 |
"multi_query": false,
|
17 |
"num_attention_heads": 32,
|
18 |
"num_hidden_layers": 24,
|
19 |
-
"parallel_attn": false,
|
20 |
"torch_dtype": "bfloat16",
|
21 |
"vocab_size": 50304,
|
22 |
"skip_init": true,
|
|
|
16 |
"multi_query": false,
|
17 |
"num_attention_heads": 32,
|
18 |
"num_hidden_layers": 24,
|
|
|
19 |
"torch_dtype": "bfloat16",
|
20 |
"vocab_size": 50304,
|
21 |
"skip_init": true,
|