WENGSYX commited on
Commit
462b598
1 Parent(s): ad0ae3a

Upload config.json

Browse files
Files changed (1) hide show
  1. config.json +20 -14
config.json CHANGED
@@ -1,22 +1,28 @@
1
  {
2
- "architectures": [
3
- "CoNNModel"
4
- ],
5
- "attention_probs_dropout_prob": 0.0,
 
 
6
  "hidden_act": "relu",
 
7
  "hidden_dropout_prob": 0.0,
8
- "hidden_size": 132,
 
9
  "initializer_range": 0.02,
10
- "intermediate_size": 42,
11
  "layer_norm": false,
12
  "layer_norm_eps": 0.02,
13
- "max_position_embeddings": 41,
14
  "mlp_hidden_size": 1959,
15
- "model_type": "conn",
16
- "num_attention_heads": 1,
17
- "num_hidden_layers": 4,
18
- "pad_token_id": 3,
19
- "torch_dtype": "float32",
20
- "transformers_version": "4.26.1",
21
- "vocab_size": 4
 
 
 
 
22
  }
 
1
  {
2
+ "pad_token_id": 3,
3
+ "transformers_version": "4.26.1",
4
+ "vocab_size": 4,
5
+ "hidden_size": 132,
6
+ "num_hidden_layers": 4,
7
+ "num_attention_heads": 1,
8
  "hidden_act": "relu",
9
+ "intermediate_size": 42,
10
  "hidden_dropout_prob": 0.0,
11
+ "attention_probs_dropout_prob": 0.0,
12
+ "max_position_embeddings": 41,
13
  "initializer_range": 0.02,
 
14
  "layer_norm": false,
15
  "layer_norm_eps": 0.02,
 
16
  "mlp_hidden_size": 1959,
17
+ "input_encoding_map": {
18
+ "0": 0,
19
+ "1": 1,
20
+ "bos": 2,
21
+ "pad": 3
22
+ },
23
+ "output_encoding_map": {
24
+ "0": 0,
25
+ "1": 1
26
+ },
27
+ "model_type": "conn"
28
  }