Felix Marty commited on
Commit
df4084d
1 Parent(s): fa16ec1
all_results.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
  "epoch": 6.0,
3
- "eval_accuracy": 0.518796992481203,
4
- "eval_loss": 0.9825071096420288,
5
- "eval_runtime": 0.5938,
6
- "eval_samples_per_second": 223.97,
7
- "eval_steps_per_second": 28.628,
8
- "train_loss": 0.9973401926984691,
9
- "train_runtime": 45.8453,
10
- "train_samples_per_second": 135.325,
11
- "train_steps_per_second": 4.319
12
  }
1
  {
2
  "epoch": 6.0,
3
+ "eval_accuracy": 0.6390977443609023,
4
+ "eval_loss": 0.7639745473861694,
5
+ "eval_runtime": 0.7192,
6
+ "eval_samples_per_second": 184.925,
7
+ "eval_steps_per_second": 23.637,
8
+ "train_loss": 0.8484223491013653,
9
+ "train_runtime": 32.6308,
10
+ "train_samples_per_second": 190.127,
11
+ "train_steps_per_second": 6.068
12
  }
config.json CHANGED
@@ -26,7 +26,7 @@
26
  },
27
  "layer_type": "basic",
28
  "model_type": "resnet",
29
- "num_channels": 1,
30
  "problem_type": "single_label_classification",
31
  "torch_dtype": "float32",
32
  "transformers_version": "4.21.0.dev0"
26
  },
27
  "layer_type": "basic",
28
  "model_type": "resnet",
29
+ "num_channels": 3,
30
  "problem_type": "single_label_classification",
31
  "torch_dtype": "float32",
32
  "transformers_version": "4.21.0.dev0"
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 6.0,
3
- "eval_accuracy": 0.518796992481203,
4
- "eval_loss": 0.9825071096420288,
5
- "eval_runtime": 0.5938,
6
- "eval_samples_per_second": 223.97,
7
- "eval_steps_per_second": 28.628
8
  }
1
  {
2
  "epoch": 6.0,
3
+ "eval_accuracy": 0.6390977443609023,
4
+ "eval_loss": 0.7639745473861694,
5
+ "eval_runtime": 0.7192,
6
+ "eval_samples_per_second": 184.925,
7
+ "eval_steps_per_second": 23.637
8
  }
preprocessor_config.json CHANGED
@@ -4,9 +4,13 @@
4
  "do_resize": true,
5
  "feature_extractor_type": "ConvNextFeatureExtractor",
6
  "image_mean": [
 
 
7
  0.45
8
  ],
9
  "image_std": [
 
 
10
  0.22
11
  ],
12
  "resample": 3,
4
  "do_resize": true,
5
  "feature_extractor_type": "ConvNextFeatureExtractor",
6
  "image_mean": [
7
+ 0.45,
8
+ 0.45,
9
  0.45
10
  ],
11
  "image_std": [
12
+ 0.22,
13
+ 0.22,
14
  0.22
15
  ],
16
  "resample": 3,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f75ed21005e99f438f2ed571a7b8db6b3dd13e53225752be4b204a6efb71388d
3
- size 761689
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64d031cd790aedfb724b3db0bccf4be88ab3b13655bab905ec329dfa39a8c0d6
3
+ size 786777
train.py CHANGED
@@ -7,7 +7,7 @@ import datasets
7
  import torch
8
  import transformers
9
  from torchinfo import summary
10
- from torchvision.transforms import Compose, Normalize, ToTensor
11
  from transformers import (
12
  ConvNextFeatureExtractor,
13
  HfArgumentParser,
@@ -103,7 +103,7 @@ def main():
103
  dataset["validation"] = split["test"]
104
 
105
  feature_extractor = ConvNextFeatureExtractor(
106
- do_resize=True, do_normalize=True, image_mean=[0.45], image_std=[0.22]
107
  )
108
 
109
  # Prepare label mappings.
@@ -115,7 +115,7 @@ def main():
115
  id2label[str(i)] = label
116
 
117
  config = ResNetConfig(
118
- num_channels=1,
119
  layer_type="basic",
120
  depths=[2, 2],
121
  hidden_sizes=[32, 64],
@@ -129,12 +129,17 @@ def main():
129
 
130
  # Define torchvision transforms to be applied to each image.
131
  normalize = Normalize(mean=feature_extractor.image_mean, std=feature_extractor.image_std)
132
- _transforms = Compose([ToTensor(), normalize])
 
 
 
 
 
133
 
134
  def transforms(example_batch):
135
  """Apply _train_transforms across a batch."""
136
  # black and white
137
- example_batch["pixel_values"] = [_transforms(pil_img.convert("L")) for pil_img in example_batch["image"]]
138
  return example_batch
139
 
140
  # Load the accuracy metric from the datasets package
7
  import torch
8
  import transformers
9
  from torchinfo import summary
10
+ from torchvision.transforms import Compose, Normalize, ToTensor, Resize, CenterCrop
11
  from transformers import (
12
  ConvNextFeatureExtractor,
13
  HfArgumentParser,
103
  dataset["validation"] = split["test"]
104
 
105
  feature_extractor = ConvNextFeatureExtractor(
106
+ do_resize=True, do_normalize=True, image_mean=[0.45, 0.45, 0.45], image_std=[0.22, 0.22, 0.22]
107
  )
108
 
109
  # Prepare label mappings.
115
  id2label[str(i)] = label
116
 
117
  config = ResNetConfig(
118
+ num_channels=3,
119
  layer_type="basic",
120
  depths=[2, 2],
121
  hidden_sizes=[32, 64],
129
 
130
  # Define torchvision transforms to be applied to each image.
131
  normalize = Normalize(mean=feature_extractor.image_mean, std=feature_extractor.image_std)
132
+ _transforms = Compose([
133
+ Resize(feature_extractor.size),
134
+ CenterCrop(feature_extractor.size),
135
+ ToTensor(),
136
+ normalize]
137
+ )
138
 
139
  def transforms(example_batch):
140
  """Apply _train_transforms across a batch."""
141
  # black and white
142
+ example_batch["pixel_values"] = [_transforms(pil_img.convert("RGB")) for pil_img in example_batch["image"]]
143
  return example_batch
144
 
145
  # Load the accuracy metric from the datasets package
train_results.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "epoch": 6.0,
3
- "train_loss": 0.9973401926984691,
4
- "train_runtime": 45.8453,
5
- "train_samples_per_second": 135.325,
6
- "train_steps_per_second": 4.319
7
  }
1
  {
2
  "epoch": 6.0,
3
+ "train_loss": 0.8484223491013653,
4
+ "train_runtime": 32.6308,
5
+ "train_samples_per_second": 190.127,
6
+ "train_steps_per_second": 6.068
7
  }
trainer_state.json CHANGED
@@ -10,16 +10,16 @@
10
  {
11
  "epoch": 6.0,
12
  "step": 198,
13
- "total_flos": 1708758414000000.0,
14
- "train_loss": 0.9973401926984691,
15
- "train_runtime": 45.8453,
16
- "train_samples_per_second": 135.325,
17
- "train_steps_per_second": 4.319
18
  }
19
  ],
20
  "max_steps": 198,
21
  "num_train_epochs": 6,
22
- "total_flos": 1708758414000000.0,
23
  "trial_name": null,
24
  "trial_params": null
25
  }
10
  {
11
  "epoch": 6.0,
12
  "step": 198,
13
+ "total_flos": 1064007556964352.0,
14
+ "train_loss": 0.8484223491013653,
15
+ "train_runtime": 32.6308,
16
+ "train_samples_per_second": 190.127,
17
+ "train_steps_per_second": 6.068
18
  }
19
  ],
20
  "max_steps": 198,
21
  "num_train_epochs": 6,
22
+ "total_flos": 1064007556964352.0,
23
  "trial_name": null,
24
  "trial_params": null
25
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a971e6637bccdbec3fdbb73349771760dd99e3c2a66a65ad63f1ab48570fa252
3
  size 3247
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5c30f97a34f52c729b4381ca5704c7c50d318011f10167ec3bfaa9db9be0777
3
  size 3247