dima806 commited on
Commit
c83dca3
1 Parent(s): 10cac41

Upload folder using huggingface_hub

Browse files
checkpoint-928/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "Model_Y",
13
+ "1": "Model_E",
14
+ "2": "Model_S",
15
+ "3": "Model_X"
16
+ },
17
+ "image_size": 224,
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 3072,
20
+ "label2id": {
21
+ "Model_E": 1,
22
+ "Model_S": 2,
23
+ "Model_X": 3,
24
+ "Model_Y": 0
25
+ },
26
+ "layer_norm_eps": 1e-12,
27
+ "model_type": "vit",
28
+ "num_attention_heads": 12,
29
+ "num_channels": 3,
30
+ "num_hidden_layers": 12,
31
+ "patch_size": 16,
32
+ "problem_type": "single_label_classification",
33
+ "qkv_bias": true,
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.33.2"
36
+ }
checkpoint-928/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6e9aee7a22ab5122279988e8d33dd90c4df6f19a5270777a4b21bd876044d4c
3
+ size 686580741
checkpoint-928/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
checkpoint-928/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:105cc66d36da826d44f11ae7883b910ffa990dd1e763feab9bb604797da7ba72
3
+ size 343274861
checkpoint-928/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51b0b2cdee095db66754707408eb4336bcbf358bc39cdf10132909fd9e39ecc2
3
+ size 14575
checkpoint-928/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1916fcd1c1d1cf8ae89b1c5c2b699fbf3669da08733f4311d3eeeeb9ac858f0d
3
+ size 627
checkpoint-928/trainer_state.json ADDED
@@ -0,0 +1,286 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8595060706138611,
3
+ "best_model_checkpoint": "tesla_car_model_image_detection/checkpoint-928",
4
+ "epoch": 29.0,
5
+ "eval_steps": 500,
6
+ "global_step": 928,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_accuracy": 0.3333333333333333,
14
+ "eval_loss": 1.356665015220642,
15
+ "eval_runtime": 4.0643,
16
+ "eval_samples_per_second": 62.003,
17
+ "eval_steps_per_second": 7.873,
18
+ "step": 32
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_accuracy": 0.3968253968253968,
23
+ "eval_loss": 1.3106689453125,
24
+ "eval_runtime": 3.6465,
25
+ "eval_samples_per_second": 69.107,
26
+ "eval_steps_per_second": 8.775,
27
+ "step": 64
28
+ },
29
+ {
30
+ "epoch": 3.0,
31
+ "eval_accuracy": 0.3968253968253968,
32
+ "eval_loss": 1.274211049079895,
33
+ "eval_runtime": 4.1634,
34
+ "eval_samples_per_second": 60.527,
35
+ "eval_steps_per_second": 7.686,
36
+ "step": 96
37
+ },
38
+ {
39
+ "epoch": 4.0,
40
+ "eval_accuracy": 0.4007936507936508,
41
+ "eval_loss": 1.2433098554611206,
42
+ "eval_runtime": 4.3624,
43
+ "eval_samples_per_second": 57.767,
44
+ "eval_steps_per_second": 7.335,
45
+ "step": 128
46
+ },
47
+ {
48
+ "epoch": 5.0,
49
+ "eval_accuracy": 0.44841269841269843,
50
+ "eval_loss": 1.212998628616333,
51
+ "eval_runtime": 3.6842,
52
+ "eval_samples_per_second": 68.399,
53
+ "eval_steps_per_second": 8.686,
54
+ "step": 160
55
+ },
56
+ {
57
+ "epoch": 6.0,
58
+ "eval_accuracy": 0.5238095238095238,
59
+ "eval_loss": 1.175584316253662,
60
+ "eval_runtime": 3.648,
61
+ "eval_samples_per_second": 69.079,
62
+ "eval_steps_per_second": 8.772,
63
+ "step": 192
64
+ },
65
+ {
66
+ "epoch": 7.0,
67
+ "eval_accuracy": 0.5515873015873016,
68
+ "eval_loss": 1.138845443725586,
69
+ "eval_runtime": 3.9253,
70
+ "eval_samples_per_second": 64.199,
71
+ "eval_steps_per_second": 8.152,
72
+ "step": 224
73
+ },
74
+ {
75
+ "epoch": 8.0,
76
+ "eval_accuracy": 0.6071428571428571,
77
+ "eval_loss": 1.1078771352767944,
78
+ "eval_runtime": 3.7402,
79
+ "eval_samples_per_second": 67.377,
80
+ "eval_steps_per_second": 8.556,
81
+ "step": 256
82
+ },
83
+ {
84
+ "epoch": 9.0,
85
+ "eval_accuracy": 0.5912698412698413,
86
+ "eval_loss": 1.0750828981399536,
87
+ "eval_runtime": 3.7683,
88
+ "eval_samples_per_second": 66.874,
89
+ "eval_steps_per_second": 8.492,
90
+ "step": 288
91
+ },
92
+ {
93
+ "epoch": 10.0,
94
+ "eval_accuracy": 0.6349206349206349,
95
+ "eval_loss": 1.049232840538025,
96
+ "eval_runtime": 3.6552,
97
+ "eval_samples_per_second": 68.942,
98
+ "eval_steps_per_second": 8.755,
99
+ "step": 320
100
+ },
101
+ {
102
+ "epoch": 11.0,
103
+ "eval_accuracy": 0.6428571428571429,
104
+ "eval_loss": 1.0186697244644165,
105
+ "eval_runtime": 3.6783,
106
+ "eval_samples_per_second": 68.51,
107
+ "eval_steps_per_second": 8.7,
108
+ "step": 352
109
+ },
110
+ {
111
+ "epoch": 12.0,
112
+ "eval_accuracy": 0.6428571428571429,
113
+ "eval_loss": 1.0005364418029785,
114
+ "eval_runtime": 3.6267,
115
+ "eval_samples_per_second": 69.485,
116
+ "eval_steps_per_second": 8.823,
117
+ "step": 384
118
+ },
119
+ {
120
+ "epoch": 13.0,
121
+ "eval_accuracy": 0.6666666666666666,
122
+ "eval_loss": 0.9789639115333557,
123
+ "eval_runtime": 3.6735,
124
+ "eval_samples_per_second": 68.6,
125
+ "eval_steps_per_second": 8.711,
126
+ "step": 416
127
+ },
128
+ {
129
+ "epoch": 14.0,
130
+ "eval_accuracy": 0.6587301587301587,
131
+ "eval_loss": 0.964360237121582,
132
+ "eval_runtime": 3.9861,
133
+ "eval_samples_per_second": 63.219,
134
+ "eval_steps_per_second": 8.028,
135
+ "step": 448
136
+ },
137
+ {
138
+ "epoch": 15.0,
139
+ "eval_accuracy": 0.6507936507936508,
140
+ "eval_loss": 0.9523670077323914,
141
+ "eval_runtime": 3.6602,
142
+ "eval_samples_per_second": 68.849,
143
+ "eval_steps_per_second": 8.743,
144
+ "step": 480
145
+ },
146
+ {
147
+ "epoch": 15.62,
148
+ "learning_rate": 5.054945054945055e-06,
149
+ "loss": 1.0015,
150
+ "step": 500
151
+ },
152
+ {
153
+ "epoch": 16.0,
154
+ "eval_accuracy": 0.6746031746031746,
155
+ "eval_loss": 0.935644805431366,
156
+ "eval_runtime": 3.6799,
157
+ "eval_samples_per_second": 68.481,
158
+ "eval_steps_per_second": 8.696,
159
+ "step": 512
160
+ },
161
+ {
162
+ "epoch": 17.0,
163
+ "eval_accuracy": 0.6706349206349206,
164
+ "eval_loss": 0.9163999557495117,
165
+ "eval_runtime": 3.6391,
166
+ "eval_samples_per_second": 69.247,
167
+ "eval_steps_per_second": 8.793,
168
+ "step": 544
169
+ },
170
+ {
171
+ "epoch": 18.0,
172
+ "eval_accuracy": 0.6587301587301587,
173
+ "eval_loss": 0.9083905816078186,
174
+ "eval_runtime": 3.7568,
175
+ "eval_samples_per_second": 67.079,
176
+ "eval_steps_per_second": 8.518,
177
+ "step": 576
178
+ },
179
+ {
180
+ "epoch": 19.0,
181
+ "eval_accuracy": 0.6785714285714286,
182
+ "eval_loss": 0.8992986679077148,
183
+ "eval_runtime": 3.6703,
184
+ "eval_samples_per_second": 68.659,
185
+ "eval_steps_per_second": 8.719,
186
+ "step": 608
187
+ },
188
+ {
189
+ "epoch": 20.0,
190
+ "eval_accuracy": 0.6746031746031746,
191
+ "eval_loss": 0.892761766910553,
192
+ "eval_runtime": 3.6781,
193
+ "eval_samples_per_second": 68.513,
194
+ "eval_steps_per_second": 8.7,
195
+ "step": 640
196
+ },
197
+ {
198
+ "epoch": 21.0,
199
+ "eval_accuracy": 0.6785714285714286,
200
+ "eval_loss": 0.8884325623512268,
201
+ "eval_runtime": 3.9427,
202
+ "eval_samples_per_second": 63.915,
203
+ "eval_steps_per_second": 8.116,
204
+ "step": 672
205
+ },
206
+ {
207
+ "epoch": 22.0,
208
+ "eval_accuracy": 0.6785714285714286,
209
+ "eval_loss": 0.8790464401245117,
210
+ "eval_runtime": 3.6773,
211
+ "eval_samples_per_second": 68.528,
212
+ "eval_steps_per_second": 8.702,
213
+ "step": 704
214
+ },
215
+ {
216
+ "epoch": 23.0,
217
+ "eval_accuracy": 0.6746031746031746,
218
+ "eval_loss": 0.8778495788574219,
219
+ "eval_runtime": 3.6291,
220
+ "eval_samples_per_second": 69.438,
221
+ "eval_steps_per_second": 8.818,
222
+ "step": 736
223
+ },
224
+ {
225
+ "epoch": 24.0,
226
+ "eval_accuracy": 0.6666666666666666,
227
+ "eval_loss": 0.8726167678833008,
228
+ "eval_runtime": 3.6835,
229
+ "eval_samples_per_second": 68.413,
230
+ "eval_steps_per_second": 8.687,
231
+ "step": 768
232
+ },
233
+ {
234
+ "epoch": 25.0,
235
+ "eval_accuracy": 0.6825396825396826,
236
+ "eval_loss": 0.8694801926612854,
237
+ "eval_runtime": 3.6959,
238
+ "eval_samples_per_second": 68.183,
239
+ "eval_steps_per_second": 8.658,
240
+ "step": 800
241
+ },
242
+ {
243
+ "epoch": 26.0,
244
+ "eval_accuracy": 0.6746031746031746,
245
+ "eval_loss": 0.8666487336158752,
246
+ "eval_runtime": 4.3527,
247
+ "eval_samples_per_second": 57.894,
248
+ "eval_steps_per_second": 7.352,
249
+ "step": 832
250
+ },
251
+ {
252
+ "epoch": 27.0,
253
+ "eval_accuracy": 0.6746031746031746,
254
+ "eval_loss": 0.8620003461837769,
255
+ "eval_runtime": 3.6904,
256
+ "eval_samples_per_second": 68.285,
257
+ "eval_steps_per_second": 8.671,
258
+ "step": 864
259
+ },
260
+ {
261
+ "epoch": 28.0,
262
+ "eval_accuracy": 0.6746031746031746,
263
+ "eval_loss": 0.8617379069328308,
264
+ "eval_runtime": 3.9652,
265
+ "eval_samples_per_second": 63.552,
266
+ "eval_steps_per_second": 8.07,
267
+ "step": 896
268
+ },
269
+ {
270
+ "epoch": 29.0,
271
+ "eval_accuracy": 0.6746031746031746,
272
+ "eval_loss": 0.8595060706138611,
273
+ "eval_runtime": 3.6428,
274
+ "eval_samples_per_second": 69.178,
275
+ "eval_steps_per_second": 8.785,
276
+ "step": 928
277
+ }
278
+ ],
279
+ "logging_steps": 500,
280
+ "max_steps": 960,
281
+ "num_train_epochs": 30,
282
+ "save_steps": 500,
283
+ "total_flos": 2.2652864457776824e+18,
284
+ "trial_name": null,
285
+ "trial_params": null
286
+ }
checkpoint-928/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f07d94c39451e68c671152dcbb778985d7d6827869cfdfd160fa96f751a0361
3
+ size 4027
config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "Model_Y",
13
+ "1": "Model_E",
14
+ "2": "Model_S",
15
+ "3": "Model_X"
16
+ },
17
+ "image_size": 224,
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 3072,
20
+ "label2id": {
21
+ "Model_E": 1,
22
+ "Model_S": 2,
23
+ "Model_X": 3,
24
+ "Model_Y": 0
25
+ },
26
+ "layer_norm_eps": 1e-12,
27
+ "model_type": "vit",
28
+ "num_attention_heads": 12,
29
+ "num_channels": 3,
30
+ "num_hidden_layers": 12,
31
+ "patch_size": 16,
32
+ "problem_type": "single_label_classification",
33
+ "qkv_bias": true,
34
+ "torch_dtype": "float32",
35
+ "transformers_version": "4.33.2"
36
+ }
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:105cc66d36da826d44f11ae7883b910ffa990dd1e763feab9bb604797da7ba72
3
+ size 343274861
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f07d94c39451e68c671152dcbb778985d7d6827869cfdfd160fa96f751a0361
3
+ size 4027