aaa12963337 commited on
Commit
96b6b03
1 Parent(s): b8f1130

Training in progress, epoch 0

Browse files
config.json CHANGED
@@ -1,13 +1,20 @@
1
  {
2
- "_name_or_path": "google/vit-base-patch16-224-in21k",
3
  "architectures": [
4
- "ViTForImageClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
- "encoder_stride": 16,
 
 
 
 
 
 
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.0,
10
- "hidden_size": 768,
11
  "id2label": {
12
  "0": "apple_pie",
13
  "1": "baby_back_ribs",
@@ -111,9 +118,8 @@
111
  "98": "tiramisu",
112
  "99": "tuna_tartare"
113
  },
114
- "image_size": 224,
115
  "initializer_range": 0.02,
116
- "intermediate_size": 3072,
117
  "label2id": {
118
  "apple_pie": "0",
119
  "baby_back_ribs": "1",
@@ -217,14 +223,35 @@
217
  "tuna_tartare": "99",
218
  "waffles": "100"
219
  },
220
- "layer_norm_eps": 1e-12,
221
- "model_type": "vit",
222
- "num_attention_heads": 12,
 
223
  "num_channels": 3,
224
- "num_hidden_layers": 12,
225
- "patch_size": 16,
 
 
 
 
 
 
 
 
 
 
 
 
 
226
  "problem_type": "single_label_classification",
227
  "qkv_bias": true,
 
 
 
 
 
 
 
228
  "torch_dtype": "float32",
229
  "transformers_version": "4.35.2"
230
  }
 
1
  {
2
+ "_name_or_path": "shi-labs/nat-mini-in1k-224",
3
  "architectures": [
4
+ "NatForImageClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
+ "depths": [
8
+ 3,
9
+ 4,
10
+ 6,
11
+ 5
12
+ ],
13
+ "drop_path_rate": 0.2,
14
+ "embed_dim": 64,
15
  "hidden_act": "gelu",
16
  "hidden_dropout_prob": 0.0,
17
+ "hidden_size": 512,
18
  "id2label": {
19
  "0": "apple_pie",
20
  "1": "baby_back_ribs",
 
118
  "98": "tiramisu",
119
  "99": "tuna_tartare"
120
  },
 
121
  "initializer_range": 0.02,
122
+ "kernel_size": 7,
123
  "label2id": {
124
  "apple_pie": "0",
125
  "baby_back_ribs": "1",
 
223
  "tuna_tartare": "99",
224
  "waffles": "100"
225
  },
226
+ "layer_norm_eps": 1e-05,
227
+ "layer_scale_init_value": 0.0,
228
+ "mlp_ratio": 3.0,
229
+ "model_type": "nat",
230
  "num_channels": 3,
231
+ "num_heads": [
232
+ 2,
233
+ 4,
234
+ 8,
235
+ 16
236
+ ],
237
+ "num_layers": 4,
238
+ "out_features": [
239
+ "stage4"
240
+ ],
241
+ "out_indices": [
242
+ 4
243
+ ],
244
+ "patch_size": 4,
245
+ "path_norm": true,
246
  "problem_type": "single_label_classification",
247
  "qkv_bias": true,
248
+ "stage_names": [
249
+ "stem",
250
+ "stage1",
251
+ "stage2",
252
+ "stage3",
253
+ "stage4"
254
+ ],
255
  "torch_dtype": "float32",
256
  "transformers_version": "4.35.2"
257
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dac0325cc49bda023ab0368bd01058903ec6135648f0d3c05d3ba7349da1ba49
3
- size 343528508
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2421f3149ca7c94e720fa8193b928373aa6bc9e856713e21e5cd1a798fbaff4
3
+ size 78131212
preprocessor_config.json CHANGED
@@ -3,17 +3,17 @@
3
  "do_rescale": true,
4
  "do_resize": true,
5
  "image_mean": [
6
- 0.5,
7
- 0.5,
8
- 0.5
9
  ],
10
  "image_processor_type": "ViTImageProcessor",
11
  "image_std": [
12
- 0.5,
13
- 0.5,
14
- 0.5
15
  ],
16
- "resample": 2,
17
  "rescale_factor": 0.00392156862745098,
18
  "size": {
19
  "height": 224,
 
3
  "do_rescale": true,
4
  "do_resize": true,
5
  "image_mean": [
6
+ 0.485,
7
+ 0.456,
8
+ 0.406
9
  ],
10
  "image_processor_type": "ViTImageProcessor",
11
  "image_std": [
12
+ 0.229,
13
+ 0.224,
14
+ 0.225
15
  ],
16
+ "resample": 3,
17
  "rescale_factor": 0.00392156862745098,
18
  "size": {
19
  "height": 224,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8df2f963698bca357cedcfcb68cdea4925d5b131032bd50d9987486dc9151830
3
  size 4155
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6422a2bcfcd68bcb3ec0785875ef3b3f62fe47d5134196d3db38a278361218e
3
  size 4155