yangwang825 commited on
Commit
866f15a
1 Parent(s): 2b1f795

Upload config

Browse files
Files changed (2) hide show
  1. config.json +2 -6
  2. configuration_bert.py +3 -1
config.json CHANGED
@@ -1,11 +1,8 @@
1
  {
2
- "architectures": [
3
- "BertForSequenceClassification"
4
- ],
5
  "attention_probs_dropout_prob": 0.1,
6
  "auto_map": {
7
- "AutoConfig": "configuration_bert.BertConfig",
8
- "AutoModelForSequenceClassification": "modeling_bert.BertForSequenceClassification"
9
  },
10
  "classifier_dropout": null,
11
  "hidden_act": "gelu",
@@ -20,7 +17,6 @@
20
  "num_hidden_layers": 12,
21
  "pad_token_id": 0,
22
  "position_embedding_type": "absolute",
23
- "torch_dtype": "float32",
24
  "transformers_version": "4.33.3",
25
  "type_vocab_size": 2,
26
  "use_cache": true,
 
1
  {
2
+ "affine": true,
 
 
3
  "attention_probs_dropout_prob": 0.1,
4
  "auto_map": {
5
+ "AutoConfig": "configuration_bert.BertConfig"
 
6
  },
7
  "classifier_dropout": null,
8
  "hidden_act": "gelu",
 
17
  "num_hidden_layers": 12,
18
  "pad_token_id": 0,
19
  "position_embedding_type": "absolute",
 
20
  "transformers_version": "4.33.3",
21
  "type_vocab_size": 2,
22
  "use_cache": true,
configuration_bert.py CHANGED
@@ -23,6 +23,7 @@ class BertConfig(PretrainedConfig):
23
  position_embedding_type="absolute",
24
  use_cache=True,
25
  classifier_dropout=None,
 
26
  **kwargs,
27
  ):
28
  super().__init__(pad_token_id=pad_token_id, **kwargs)
@@ -41,4 +42,5 @@ class BertConfig(PretrainedConfig):
41
  self.layer_norm_eps = layer_norm_eps
42
  self.position_embedding_type = position_embedding_type
43
  self.use_cache = use_cache
44
- self.classifier_dropout = classifier_dropout
 
 
23
  position_embedding_type="absolute",
24
  use_cache=True,
25
  classifier_dropout=None,
26
+ affine=True,
27
  **kwargs,
28
  ):
29
  super().__init__(pad_token_id=pad_token_id, **kwargs)
 
42
  self.layer_norm_eps = layer_norm_eps
43
  self.position_embedding_type = position_embedding_type
44
  self.use_cache = use_cache
45
+ self.classifier_dropout = classifier_dropout
46
+ self.affine = affine