nielsr HF staff commited on
Commit
9f7d169
1 Parent(s): 760b167
Files changed (2) hide show
  1. config.json +9 -9
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -4,12 +4,18 @@
4
  ],
5
  "attention_probs_dropout_prob": 0.0,
6
  "auxiliary_loss_weight": 0.4,
7
- "channels": 256,
 
 
 
 
 
 
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.0,
10
  "hidden_size": 1024,
11
  "image_size": 384,
12
- "in_index": -1,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 4096,
15
  "layer_norm_eps": 1e-12,
@@ -23,12 +29,6 @@
23
  "num_attention_heads": 16,
24
  "num_channels": 3,
25
  "num_hidden_layers": 24,
26
- "out_indices": [
27
- 5,
28
- 11,
29
- 17,
30
- 23
31
- ],
32
  "patch_size": 16,
33
  "qkv_bias": true,
34
  "readout_type": "project",
@@ -43,5 +43,5 @@
43
  "torch_dtype": "float32",
44
  "transformers_version": "4.18.0.dev0",
45
  "use_auxiliary_head": true,
46
- "use_batch_norm": false
47
  }
 
4
  ],
5
  "attention_probs_dropout_prob": 0.0,
6
  "auxiliary_loss_weight": 0.4,
7
+ "backbone_out_indices": [
8
+ 5,
9
+ 11,
10
+ 17,
11
+ 23
12
+ ],
13
+ "fusion_hidden_size": 256,
14
+ "head_in_index": -1,
15
  "hidden_act": "gelu",
16
  "hidden_dropout_prob": 0.0,
17
  "hidden_size": 1024,
18
  "image_size": 384,
 
19
  "initializer_range": 0.02,
20
  "intermediate_size": 4096,
21
  "layer_norm_eps": 1e-12,
 
29
  "num_attention_heads": 16,
30
  "num_channels": 3,
31
  "num_hidden_layers": 24,
 
 
 
 
 
 
32
  "patch_size": 16,
33
  "qkv_bias": true,
34
  "readout_type": "project",
 
43
  "torch_dtype": "float32",
44
  "transformers_version": "4.18.0.dev0",
45
  "use_auxiliary_head": true,
46
+ "use_batch_norm_in_fusion_residual": false
47
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:434f2e147724ab260492e3f4d33ae6629da181957f05f2d503e7f4a4e0bf34f0
3
- size 1367580589
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71150941604c39c1c770a72a7b2f56487669f0e3a4d99c8759e233fb1be24080
3
+ size 1367581165