Felixvi commited on
Commit
96df2c6
1 Parent(s): d8ef1d0

Training in progress, epoch 0

Browse files
config.json CHANGED
@@ -1,10 +1,18 @@
1
  {
2
- "_name_or_path": "google/vit-base-patch16-224-in21k",
3
  "architectures": [
4
- "ViTForImageClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
- "encoder_stride": 16,
 
 
 
 
 
 
 
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.0,
10
  "hidden_size": 768,
@@ -28,9 +36,8 @@
28
  "8": "\u6708\u4eae",
29
  "9": "\u72d7"
30
  },
31
- "image_size": 224,
32
  "initializer_range": 0.02,
33
- "intermediate_size": 3072,
34
  "label2id": {
35
  "\u54ed": "0",
36
  "\u5750": "1",
@@ -51,14 +58,29 @@
51
  "\u81ea\u884c\u8f66": "16",
52
  "\u829d\u9ebb": "17"
53
  },
54
- "layer_norm_eps": 1e-12,
55
- "model_type": "vit",
56
- "num_attention_heads": 12,
57
  "num_channels": 3,
58
- "num_hidden_layers": 12,
59
- "patch_size": 16,
 
 
 
 
 
 
 
 
 
 
 
 
 
60
  "problem_type": "single_label_classification",
61
  "qkv_bias": true,
62
  "torch_dtype": "float32",
63
- "transformers_version": "4.28.1"
 
 
64
  }
 
1
  {
2
+ "_name_or_path": "microsoft/swinv2-tiny-patch4-window8-256",
3
  "architectures": [
4
+ "Swinv2ForImageClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
+ "depths": [
8
+ 2,
9
+ 2,
10
+ 6,
11
+ 2
12
+ ],
13
+ "drop_path_rate": 0.1,
14
+ "embed_dim": 96,
15
+ "encoder_stride": 32,
16
  "hidden_act": "gelu",
17
  "hidden_dropout_prob": 0.0,
18
  "hidden_size": 768,
 
36
  "8": "\u6708\u4eae",
37
  "9": "\u72d7"
38
  },
39
+ "image_size": 256,
40
  "initializer_range": 0.02,
 
41
  "label2id": {
42
  "\u54ed": "0",
43
  "\u5750": "1",
 
58
  "\u81ea\u884c\u8f66": "16",
59
  "\u829d\u9ebb": "17"
60
  },
61
+ "layer_norm_eps": 1e-05,
62
+ "mlp_ratio": 4.0,
63
+ "model_type": "swinv2",
64
  "num_channels": 3,
65
+ "num_heads": [
66
+ 3,
67
+ 6,
68
+ 12,
69
+ 24
70
+ ],
71
+ "num_layers": 4,
72
+ "patch_size": 4,
73
+ "path_norm": true,
74
+ "pretrained_window_sizes": [
75
+ 0,
76
+ 0,
77
+ 0,
78
+ 0
79
+ ],
80
  "problem_type": "single_label_classification",
81
  "qkv_bias": true,
82
  "torch_dtype": "float32",
83
+ "transformers_version": "4.28.1",
84
+ "use_absolute_embeddings": false,
85
+ "window_size": 8
86
  }
preprocessor_config.json CHANGED
@@ -3,20 +3,20 @@
3
  "do_rescale": true,
4
  "do_resize": true,
5
  "image_mean": [
6
- 0.5,
7
- 0.5,
8
- 0.5
9
  ],
10
  "image_processor_type": "ViTImageProcessor",
11
  "image_std": [
12
- 0.5,
13
- 0.5,
14
- 0.5
15
  ],
16
- "resample": 2,
17
  "rescale_factor": 0.00392156862745098,
18
  "size": {
19
- "height": 224,
20
- "width": 224
21
  }
22
  }
 
3
  "do_rescale": true,
4
  "do_resize": true,
5
  "image_mean": [
6
+ 0.485,
7
+ 0.456,
8
+ 0.406
9
  ],
10
  "image_processor_type": "ViTImageProcessor",
11
  "image_std": [
12
+ 0.229,
13
+ 0.224,
14
+ 0.225
15
  ],
16
+ "resample": 3,
17
  "rescale_factor": 0.00392156862745098,
18
  "size": {
19
+ "height": 256,
20
+ "width": 256
21
  }
22
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0d8bc144fc95187b59b7d31852469f4a793533b85e829284957bace40863d11
3
- size 343317933
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:910535e82e2466731b4881cb923824ff7e71dd1864fcfb6be7205a6303573e5d
3
+ size 110454369
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08818a71e86674b571cc4e6edfa120472f9470579211e6659dc9ebe613c5f5d7
3
  size 3579
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a66eefcbd71bec5aec9d17e31e6da58c63b88728cf97b70418760c53fcae3e00
3
  size 3579