Markus28 commited on
Commit
5b58f09
1 Parent(s): 5e7b835

Added additional config options

Browse files
Files changed (1) hide show
  1. configuration_bert.py +14 -1
configuration_bert.py CHANGED
@@ -75,8 +75,16 @@ class JinaBertConfig(PretrainedConfig):
75
  pad_token_id=0,
76
  window_size=(-1, -1),
77
  dense_seq_output=False,
 
 
 
 
 
 
78
  **kwargs,
79
  ):
 
 
80
  super().__init__(pad_token_id=pad_token_id, **kwargs)
81
 
82
  self.vocab_size = vocab_size
@@ -92,4 +100,9 @@ class JinaBertConfig(PretrainedConfig):
92
  self.layer_norm_eps = layer_norm_eps
93
  self.window_size = window_size
94
  self.dense_seq_output = dense_seq_output
95
-
 
 
 
 
 
 
75
  pad_token_id=0,
76
  window_size=(-1, -1),
77
  dense_seq_output=False,
78
+ fused_mlp=False,
79
+ mlp_checkpoint_lvl=0,
80
+ last_layer_subset=False,
81
+ fused_dropout_add_ln=False,
82
+ fused_bias_fc=False,
83
+ pad_vocab_size_multiple=1,
84
  **kwargs,
85
  ):
86
+ assert 'position_embedding_type' not in kwargs
87
+ assert 'max_position_embeddings' not in kwargs
88
  super().__init__(pad_token_id=pad_token_id, **kwargs)
89
 
90
  self.vocab_size = vocab_size
 
100
  self.layer_norm_eps = layer_norm_eps
101
  self.window_size = window_size
102
  self.dense_seq_output = dense_seq_output
103
+ self.fused_mlp = fused_mlp
104
+ self.mlp_checkpoint_lvl = mlp_checkpoint_lvl
105
+ self.last_layer_subset = last_layer_subset
106
+ self.fused_dropout_add_ln = fused_dropout_add_ln
107
+ self.fused_bias_fc = fused_bias_fc
108
+ self.pad_vocab_size_multiple = pad_vocab_size_multiple