Sengxian commited on
Commit
ad75f89
1 Parent(s): c3b3141

Update config

Browse files
Files changed (2) hide show
  1. config.json +0 -2
  2. configuration_chatglm.py +0 -3
config.json CHANGED
@@ -19,7 +19,6 @@
19
  "fp32_residual_connection": false,
20
  "hidden_dropout": 0.0,
21
  "hidden_size": 4096,
22
- "interleaved_qkv": false,
23
  "kv_channels": 128,
24
  "layernorm_epsilon": 1e-05,
25
  "multi_query_attention": true,
@@ -30,7 +29,6 @@
30
  "padded_vocab_size": 65024,
31
  "post_layer_norm": true,
32
  "rmsnorm": true,
33
- "rotary_percent": 0.5,
34
  "seq_length": 32768,
35
  "use_cache": true,
36
  "torch_dtype": "float16",
 
19
  "fp32_residual_connection": false,
20
  "hidden_dropout": 0.0,
21
  "hidden_size": 4096,
 
22
  "kv_channels": 128,
23
  "layernorm_epsilon": 1e-05,
24
  "multi_query_attention": true,
 
29
  "padded_vocab_size": 65024,
30
  "post_layer_norm": true,
31
  "rmsnorm": true,
 
32
  "seq_length": 32768,
33
  "use_cache": true,
34
  "torch_dtype": "float16",
configuration_chatglm.py CHANGED
@@ -21,7 +21,6 @@ class ChatGLMConfig(PretrainedConfig):
21
  add_qkv_bias=False,
22
  interleaved_qkv=False,
23
  bias_dropout_fusion=True,
24
- rotary_percent=1.0,
25
  multi_query_attention=False,
26
  multi_query_group_num=1,
27
  apply_query_key_layer_scaling=True,
@@ -45,9 +44,7 @@ class ChatGLMConfig(PretrainedConfig):
45
  self.post_layer_norm = post_layer_norm
46
  self.add_bias_linear = add_bias_linear
47
  self.add_qkv_bias = add_qkv_bias
48
- self.interleaved_qkv = interleaved_qkv
49
  self.bias_dropout_fusion = bias_dropout_fusion
50
- self.rotary_percent = rotary_percent
51
  self.multi_query_attention = multi_query_attention
52
  self.multi_query_group_num = multi_query_group_num
53
  self.apply_query_key_layer_scaling = apply_query_key_layer_scaling
 
21
  add_qkv_bias=False,
22
  interleaved_qkv=False,
23
  bias_dropout_fusion=True,
 
24
  multi_query_attention=False,
25
  multi_query_group_num=1,
26
  apply_query_key_layer_scaling=True,
 
44
  self.post_layer_norm = post_layer_norm
45
  self.add_bias_linear = add_bias_linear
46
  self.add_qkv_bias = add_qkv_bias
 
47
  self.bias_dropout_fusion = bias_dropout_fusion
 
48
  self.multi_query_attention = multi_query_attention
49
  self.multi_query_group_num = multi_query_group_num
50
  self.apply_query_key_layer_scaling = apply_query_key_layer_scaling