ybelkada commited on
Commit
99ac5c2
1 Parent(s): bf4ebd0

Upload SamForMaskGeneration

Browse files
Files changed (2) hide show
  1. config.json +9 -6
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_commit_hash": null,
3
  "architectures": [
4
- "SamForImageSegmentation"
5
  ],
6
  "initializer_factor": 1.0,
7
  "initializer_range": 0.02,
@@ -39,6 +39,7 @@
39
  "LABEL_0": 0,
40
  "LABEL_1": 1
41
  },
 
42
  "length_penalty": 1.0,
43
  "max_length": 20,
44
  "min_length": 0,
@@ -74,7 +75,7 @@
74
  "top_p": 1.0,
75
  "torch_dtype": null,
76
  "torchscript": false,
77
- "transformers_version": "4.28.0.dev0",
78
  "typical_p": 1.0,
79
  "use_bfloat16": false
80
  },
@@ -146,11 +147,11 @@
146
  "top_p": 1.0,
147
  "torch_dtype": null,
148
  "torchscript": false,
149
- "transformers_version": "4.28.0.dev0",
150
  "typical_p": 1.0,
151
  "use_bfloat16": false
152
  },
153
- "torch_dtype": "float16",
154
  "transformers_version": null,
155
  "vision_config": {
156
  "_name_or_path": "",
@@ -195,10 +196,11 @@
195
  "LABEL_0": 0,
196
  "LABEL_1": 1
197
  },
198
- "layer_norm_eps": 1e-05,
199
  "length_penalty": 1.0,
200
  "max_length": 20,
201
  "min_length": 0,
 
202
  "mlp_ratio": 4.0,
203
  "model_type": "sam_vision_model",
204
  "no_repeat_ngram_size": 0,
@@ -207,6 +209,7 @@
207
  "num_beams": 1,
208
  "num_channels": 3,
209
  "num_hidden_layers": 32,
 
210
  "num_return_sequences": 1,
211
  "output_attentions": false,
212
  "output_channels": 256,
@@ -236,7 +239,7 @@
236
  "top_p": 1.0,
237
  "torch_dtype": null,
238
  "torchscript": false,
239
- "transformers_version": "4.28.0.dev0",
240
  "typical_p": 1.0,
241
  "use_abs_pos": true,
242
  "use_bfloat16": false,
 
1
  {
2
  "_commit_hash": null,
3
  "architectures": [
4
+ "SamForMaskGeneration"
5
  ],
6
  "initializer_factor": 1.0,
7
  "initializer_range": 0.02,
 
39
  "LABEL_0": 0,
40
  "LABEL_1": 1
41
  },
42
+ "layer_norm_eps": 1e-06,
43
  "length_penalty": 1.0,
44
  "max_length": 20,
45
  "min_length": 0,
 
75
  "top_p": 1.0,
76
  "torch_dtype": null,
77
  "torchscript": false,
78
+ "transformers_version": "4.29.0.dev0",
79
  "typical_p": 1.0,
80
  "use_bfloat16": false
81
  },
 
147
  "top_p": 1.0,
148
  "torch_dtype": null,
149
  "torchscript": false,
150
+ "transformers_version": "4.29.0.dev0",
151
  "typical_p": 1.0,
152
  "use_bfloat16": false
153
  },
154
+ "torch_dtype": "float32",
155
  "transformers_version": null,
156
  "vision_config": {
157
  "_name_or_path": "",
 
196
  "LABEL_0": 0,
197
  "LABEL_1": 1
198
  },
199
+ "layer_norm_eps": 1e-06,
200
  "length_penalty": 1.0,
201
  "max_length": 20,
202
  "min_length": 0,
203
+ "mlp_dim": 5120,
204
  "mlp_ratio": 4.0,
205
  "model_type": "sam_vision_model",
206
  "no_repeat_ngram_size": 0,
 
209
  "num_beams": 1,
210
  "num_channels": 3,
211
  "num_hidden_layers": 32,
212
+ "num_pos_feats": 128,
213
  "num_return_sequences": 1,
214
  "output_attentions": false,
215
  "output_channels": 256,
 
239
  "top_p": 1.0,
240
  "torch_dtype": null,
241
  "torchscript": false,
242
+ "transformers_version": "4.29.0.dev0",
243
  "typical_p": 1.0,
244
  "use_abs_pos": true,
245
  "use_bfloat16": false,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d9ced09901119ebcbf97b53ca9a3bb5c62ee746c8fbf206b947214427182ac0
3
- size 1282379669
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02edcfeab5801c77aa03df04c95b3c919c6c8052c979a61f59ab499266aad5a3
3
+ size 2564565013