Crystalcareai commited on
Commit
082cf66
1 Parent(s): 95f3d4b

Update configuration_gemmoe.py

Browse files
Files changed (1) hide show
  1. configuration_gemmoe.py +1 -3
configuration_gemmoe.py CHANGED
@@ -118,7 +118,7 @@ class GemmoeConfig(PretrainedConfig):
118
  num_attention_heads=16,
119
  num_key_value_heads=16,
120
  head_dim=256,
121
- hidden_act="gelu",
122
  max_position_embeddings=8192,
123
  initializer_range=0.02,
124
  rms_norm_eps=1e-6,
@@ -126,7 +126,6 @@ class GemmoeConfig(PretrainedConfig):
126
  pad_token_id=0,
127
  eos_token_id=1,
128
  bos_token_id=2,
129
- hidden_act="gelu_pytorch_tanh",
130
  hidden_activation=None,
131
  tie_word_embeddings=True,
132
  rope_theta=10000.0,
@@ -148,7 +147,6 @@ class GemmoeConfig(PretrainedConfig):
148
  self.hidden_act = hidden_act
149
  self.hidden_activation = hidden_activation
150
  self.num_key_value_heads = num_key_value_heads
151
- self.hidden_act = hidden_act
152
  self.initializer_range = initializer_range
153
  self.rms_norm_eps = rms_norm_eps
154
  self.use_cache = use_cache
 
118
  num_attention_heads=16,
119
  num_key_value_heads=16,
120
  head_dim=256,
121
+ hidden_act="gelu_pytorch_tanh",
122
  max_position_embeddings=8192,
123
  initializer_range=0.02,
124
  rms_norm_eps=1e-6,
 
126
  pad_token_id=0,
127
  eos_token_id=1,
128
  bos_token_id=2,
 
129
  hidden_activation=None,
130
  tie_word_embeddings=True,
131
  rope_theta=10000.0,
 
147
  self.hidden_act = hidden_act
148
  self.hidden_activation = hidden_activation
149
  self.num_key_value_heads = num_key_value_heads
 
150
  self.initializer_range = initializer_range
151
  self.rms_norm_eps = rms_norm_eps
152
  self.use_cache = use_cache