GuntramG commited on
Commit
aac3138
1 Parent(s): f3284b1

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - image-classification
6
+ widget:
7
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
8
+ example_title: Tiger
9
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
10
+ example_title: Teapot
11
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
12
+ example_title: Palace
13
+ ---
14
+
15
+ # Model Trained Using AutoTrain
16
+
17
+ - Problem type: Image Classification
18
+
19
+ ## Validation Metrics
20
+ loss: 0.18285594880580902
21
+
22
+ f1_macro: 0.9148614413559308
23
+
24
+ f1_micro: 0.9365456396335257
25
+
26
+ f1_weighted: 0.9364564915178187
27
+
28
+ precision_macro: 0.9159613318061458
29
+
30
+ precision_micro: 0.9365456396335257
31
+
32
+ precision_weighted: 0.9365041505044936
33
+
34
+ recall_macro: 0.9139276800740521
35
+
36
+ recall_micro: 0.9365456396335257
37
+
38
+ recall_weighted: 0.9365456396335257
39
+
40
+ accuracy: 0.9365456396335257
checkpoint-1107/config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/convnextv2-tiny-1k-224",
3
+ "_num_labels": 5,
4
+ "architectures": [
5
+ "ConvNextV2ForImageClassification"
6
+ ],
7
+ "depths": [
8
+ 3,
9
+ 3,
10
+ 9,
11
+ 3
12
+ ],
13
+ "drop_path_rate": 0.0,
14
+ "hidden_act": "gelu",
15
+ "hidden_sizes": [
16
+ 96,
17
+ 192,
18
+ 384,
19
+ 768
20
+ ],
21
+ "id2label": {
22
+ "0": "food",
23
+ "1": "inside",
24
+ "2": "menu",
25
+ "3": "outside",
26
+ "4": "people"
27
+ },
28
+ "image_size": 224,
29
+ "initializer_range": 0.02,
30
+ "label2id": {
31
+ "food": 0,
32
+ "inside": 1,
33
+ "menu": 2,
34
+ "outside": 3,
35
+ "people": 4
36
+ },
37
+ "layer_norm_eps": 1e-12,
38
+ "model_type": "convnextv2",
39
+ "num_channels": 3,
40
+ "num_stages": 4,
41
+ "out_features": [
42
+ "stage4"
43
+ ],
44
+ "out_indices": [
45
+ 4
46
+ ],
47
+ "patch_size": 4,
48
+ "problem_type": "single_label_classification",
49
+ "stage_names": [
50
+ "stem",
51
+ "stage1",
52
+ "stage2",
53
+ "stage3",
54
+ "stage4"
55
+ ],
56
+ "torch_dtype": "float32",
57
+ "transformers_version": "4.40.1"
58
+ }
checkpoint-1107/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f93c3ea4f6d582a9705f2c1a6a5023205bca5072bf88f23d270c97a488e79042
3
+ size 111505052
checkpoint-1107/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14d2504ff065847f1c6a41614c02a7bf0bec5c5afed5fc3ff00af6d3a7407585
3
+ size 223131066
checkpoint-1107/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48dd83a0ab96fee86ff887f31c13ac68c17490b563b62a8e0613f627910b8caf
3
+ size 14244
checkpoint-1107/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:894759274558b67a202f7aa023d4fc13faa30b70b87ff7c79fe0b865b4803a06
3
+ size 1064
checkpoint-1107/trainer_state.json ADDED
@@ -0,0 +1,383 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.18285594880580902,
3
+ "best_model_checkpoint": "autotrain-6doma-5m8vf/checkpoint-1107",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1107,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.06775067750677506,
13
+ "grad_norm": 16.566085815429688,
14
+ "learning_rate": 9.90990990990991e-06,
15
+ "loss": 1.5667,
16
+ "step": 25
17
+ },
18
+ {
19
+ "epoch": 0.13550135501355012,
20
+ "grad_norm": 25.27834129333496,
21
+ "learning_rate": 2.117117117117117e-05,
22
+ "loss": 1.2958,
23
+ "step": 50
24
+ },
25
+ {
26
+ "epoch": 0.2032520325203252,
27
+ "grad_norm": 14.882951736450195,
28
+ "learning_rate": 3.2432432432432436e-05,
29
+ "loss": 1.0021,
30
+ "step": 75
31
+ },
32
+ {
33
+ "epoch": 0.27100271002710025,
34
+ "grad_norm": 12.05569839477539,
35
+ "learning_rate": 4.369369369369369e-05,
36
+ "loss": 0.7635,
37
+ "step": 100
38
+ },
39
+ {
40
+ "epoch": 0.33875338753387535,
41
+ "grad_norm": 15.701233863830566,
42
+ "learning_rate": 4.944779116465864e-05,
43
+ "loss": 0.572,
44
+ "step": 125
45
+ },
46
+ {
47
+ "epoch": 0.4065040650406504,
48
+ "grad_norm": 45.54197692871094,
49
+ "learning_rate": 4.8242971887550205e-05,
50
+ "loss": 0.4778,
51
+ "step": 150
52
+ },
53
+ {
54
+ "epoch": 0.4742547425474255,
55
+ "grad_norm": 20.10284996032715,
56
+ "learning_rate": 4.698795180722892e-05,
57
+ "loss": 0.4213,
58
+ "step": 175
59
+ },
60
+ {
61
+ "epoch": 0.5420054200542005,
62
+ "grad_norm": 52.795291900634766,
63
+ "learning_rate": 4.573293172690764e-05,
64
+ "loss": 0.4171,
65
+ "step": 200
66
+ },
67
+ {
68
+ "epoch": 0.6097560975609756,
69
+ "grad_norm": 32.24135208129883,
70
+ "learning_rate": 4.447791164658635e-05,
71
+ "loss": 0.4031,
72
+ "step": 225
73
+ },
74
+ {
75
+ "epoch": 0.6775067750677507,
76
+ "grad_norm": 48.17521286010742,
77
+ "learning_rate": 4.3222891566265064e-05,
78
+ "loss": 0.3206,
79
+ "step": 250
80
+ },
81
+ {
82
+ "epoch": 0.7452574525745257,
83
+ "grad_norm": 27.5257511138916,
84
+ "learning_rate": 4.196787148594378e-05,
85
+ "loss": 0.3616,
86
+ "step": 275
87
+ },
88
+ {
89
+ "epoch": 0.8130081300813008,
90
+ "grad_norm": 15.912370681762695,
91
+ "learning_rate": 4.071285140562249e-05,
92
+ "loss": 0.508,
93
+ "step": 300
94
+ },
95
+ {
96
+ "epoch": 0.8807588075880759,
97
+ "grad_norm": 12.763589859008789,
98
+ "learning_rate": 3.9508032128514064e-05,
99
+ "loss": 0.3685,
100
+ "step": 325
101
+ },
102
+ {
103
+ "epoch": 0.948509485094851,
104
+ "grad_norm": 20.36044692993164,
105
+ "learning_rate": 3.8253012048192774e-05,
106
+ "loss": 0.393,
107
+ "step": 350
108
+ },
109
+ {
110
+ "epoch": 1.0,
111
+ "eval_accuracy": 0.9189005768578216,
112
+ "eval_f1_macro": 0.8913236764060113,
113
+ "eval_f1_micro": 0.9189005768578216,
114
+ "eval_f1_weighted": 0.9195807716070247,
115
+ "eval_loss": 0.2322985827922821,
116
+ "eval_precision_macro": 0.8992556342366311,
117
+ "eval_precision_micro": 0.9189005768578216,
118
+ "eval_precision_weighted": 0.9235420876186199,
119
+ "eval_recall_macro": 0.8887207219589304,
120
+ "eval_recall_micro": 0.9189005768578216,
121
+ "eval_recall_weighted": 0.9189005768578216,
122
+ "eval_runtime": 19.8492,
123
+ "eval_samples_per_second": 148.469,
124
+ "eval_steps_per_second": 9.32,
125
+ "step": 369
126
+ },
127
+ {
128
+ "epoch": 1.016260162601626,
129
+ "grad_norm": 14.446638107299805,
130
+ "learning_rate": 3.699799196787149e-05,
131
+ "loss": 0.2934,
132
+ "step": 375
133
+ },
134
+ {
135
+ "epoch": 1.084010840108401,
136
+ "grad_norm": 29.589651107788086,
137
+ "learning_rate": 3.57429718875502e-05,
138
+ "loss": 0.3801,
139
+ "step": 400
140
+ },
141
+ {
142
+ "epoch": 1.151761517615176,
143
+ "grad_norm": 23.76276397705078,
144
+ "learning_rate": 3.4487951807228916e-05,
145
+ "loss": 0.2276,
146
+ "step": 425
147
+ },
148
+ {
149
+ "epoch": 1.2195121951219512,
150
+ "grad_norm": 33.06072998046875,
151
+ "learning_rate": 3.323293172690763e-05,
152
+ "loss": 0.2623,
153
+ "step": 450
154
+ },
155
+ {
156
+ "epoch": 1.2872628726287263,
157
+ "grad_norm": 31.562694549560547,
158
+ "learning_rate": 3.197791164658634e-05,
159
+ "loss": 0.3324,
160
+ "step": 475
161
+ },
162
+ {
163
+ "epoch": 1.3550135501355014,
164
+ "grad_norm": 25.050046920776367,
165
+ "learning_rate": 3.072289156626506e-05,
166
+ "loss": 0.3613,
167
+ "step": 500
168
+ },
169
+ {
170
+ "epoch": 1.4227642276422765,
171
+ "grad_norm": 5.65738582611084,
172
+ "learning_rate": 2.9467871485943778e-05,
173
+ "loss": 0.3689,
174
+ "step": 525
175
+ },
176
+ {
177
+ "epoch": 1.4905149051490514,
178
+ "grad_norm": 30.50360870361328,
179
+ "learning_rate": 2.821285140562249e-05,
180
+ "loss": 0.2128,
181
+ "step": 550
182
+ },
183
+ {
184
+ "epoch": 1.5582655826558267,
185
+ "grad_norm": 31.306838989257812,
186
+ "learning_rate": 2.6957831325301207e-05,
187
+ "loss": 0.3329,
188
+ "step": 575
189
+ },
190
+ {
191
+ "epoch": 1.6260162601626016,
192
+ "grad_norm": 5.569540023803711,
193
+ "learning_rate": 2.570281124497992e-05,
194
+ "loss": 0.3934,
195
+ "step": 600
196
+ },
197
+ {
198
+ "epoch": 1.6937669376693767,
199
+ "grad_norm": 79.83793640136719,
200
+ "learning_rate": 2.4447791164658633e-05,
201
+ "loss": 0.3329,
202
+ "step": 625
203
+ },
204
+ {
205
+ "epoch": 1.7615176151761518,
206
+ "grad_norm": 11.711432456970215,
207
+ "learning_rate": 2.319277108433735e-05,
208
+ "loss": 0.3065,
209
+ "step": 650
210
+ },
211
+ {
212
+ "epoch": 1.8292682926829267,
213
+ "grad_norm": 27.71021842956543,
214
+ "learning_rate": 2.1937751004016066e-05,
215
+ "loss": 0.3361,
216
+ "step": 675
217
+ },
218
+ {
219
+ "epoch": 1.897018970189702,
220
+ "grad_norm": 24.346481323242188,
221
+ "learning_rate": 2.068273092369478e-05,
222
+ "loss": 0.3967,
223
+ "step": 700
224
+ },
225
+ {
226
+ "epoch": 1.9647696476964769,
227
+ "grad_norm": 7.5306549072265625,
228
+ "learning_rate": 1.9427710843373495e-05,
229
+ "loss": 0.3304,
230
+ "step": 725
231
+ },
232
+ {
233
+ "epoch": 2.0,
234
+ "eval_accuracy": 0.9284017645062775,
235
+ "eval_f1_macro": 0.903605865288441,
236
+ "eval_f1_micro": 0.9284017645062775,
237
+ "eval_f1_weighted": 0.9277100982185731,
238
+ "eval_loss": 0.2046061009168625,
239
+ "eval_precision_macro": 0.9095877174004062,
240
+ "eval_precision_micro": 0.9284017645062775,
241
+ "eval_precision_weighted": 0.9281331331487362,
242
+ "eval_recall_macro": 0.8989112570392443,
243
+ "eval_recall_micro": 0.9284017645062775,
244
+ "eval_recall_weighted": 0.9284017645062775,
245
+ "eval_runtime": 19.9432,
246
+ "eval_samples_per_second": 147.769,
247
+ "eval_steps_per_second": 9.276,
248
+ "step": 738
249
+ },
250
+ {
251
+ "epoch": 2.032520325203252,
252
+ "grad_norm": 28.1395206451416,
253
+ "learning_rate": 1.822289156626506e-05,
254
+ "loss": 0.332,
255
+ "step": 750
256
+ },
257
+ {
258
+ "epoch": 2.100271002710027,
259
+ "grad_norm": 7.682183265686035,
260
+ "learning_rate": 1.6967871485943776e-05,
261
+ "loss": 0.2995,
262
+ "step": 775
263
+ },
264
+ {
265
+ "epoch": 2.168021680216802,
266
+ "grad_norm": 23.640390396118164,
267
+ "learning_rate": 1.5712851405622492e-05,
268
+ "loss": 0.3089,
269
+ "step": 800
270
+ },
271
+ {
272
+ "epoch": 2.2357723577235773,
273
+ "grad_norm": 3.6244945526123047,
274
+ "learning_rate": 1.4457831325301205e-05,
275
+ "loss": 0.2557,
276
+ "step": 825
277
+ },
278
+ {
279
+ "epoch": 2.303523035230352,
280
+ "grad_norm": 109.68293762207031,
281
+ "learning_rate": 1.3202811244979921e-05,
282
+ "loss": 0.3734,
283
+ "step": 850
284
+ },
285
+ {
286
+ "epoch": 2.3712737127371275,
287
+ "grad_norm": 28.19609832763672,
288
+ "learning_rate": 1.1947791164658636e-05,
289
+ "loss": 0.3144,
290
+ "step": 875
291
+ },
292
+ {
293
+ "epoch": 2.4390243902439024,
294
+ "grad_norm": 17.588850021362305,
295
+ "learning_rate": 1.069277108433735e-05,
296
+ "loss": 0.4061,
297
+ "step": 900
298
+ },
299
+ {
300
+ "epoch": 2.5067750677506773,
301
+ "grad_norm": 21.384654998779297,
302
+ "learning_rate": 9.437751004016063e-06,
303
+ "loss": 0.2626,
304
+ "step": 925
305
+ },
306
+ {
307
+ "epoch": 2.5745257452574526,
308
+ "grad_norm": 0.35269397497177124,
309
+ "learning_rate": 8.18273092369478e-06,
310
+ "loss": 0.2822,
311
+ "step": 950
312
+ },
313
+ {
314
+ "epoch": 2.642276422764228,
315
+ "grad_norm": 21.37306785583496,
316
+ "learning_rate": 6.927710843373494e-06,
317
+ "loss": 0.4436,
318
+ "step": 975
319
+ },
320
+ {
321
+ "epoch": 2.710027100271003,
322
+ "grad_norm": 17.97796630859375,
323
+ "learning_rate": 5.672690763052209e-06,
324
+ "loss": 0.2517,
325
+ "step": 1000
326
+ },
327
+ {
328
+ "epoch": 2.7777777777777777,
329
+ "grad_norm": 9.19117259979248,
330
+ "learning_rate": 4.417670682730924e-06,
331
+ "loss": 0.2395,
332
+ "step": 1025
333
+ },
334
+ {
335
+ "epoch": 2.845528455284553,
336
+ "grad_norm": 64.22978210449219,
337
+ "learning_rate": 3.1626506024096387e-06,
338
+ "loss": 0.3387,
339
+ "step": 1050
340
+ },
341
+ {
342
+ "epoch": 2.913279132791328,
343
+ "grad_norm": 1.4875394105911255,
344
+ "learning_rate": 1.9076305220883537e-06,
345
+ "loss": 0.2559,
346
+ "step": 1075
347
+ },
348
+ {
349
+ "epoch": 2.9810298102981028,
350
+ "grad_norm": 0.5281310677528381,
351
+ "learning_rate": 6.526104417670682e-07,
352
+ "loss": 0.3605,
353
+ "step": 1100
354
+ },
355
+ {
356
+ "epoch": 3.0,
357
+ "eval_accuracy": 0.9365456396335257,
358
+ "eval_f1_macro": 0.9148614413559308,
359
+ "eval_f1_micro": 0.9365456396335257,
360
+ "eval_f1_weighted": 0.9364564915178187,
361
+ "eval_loss": 0.18285594880580902,
362
+ "eval_precision_macro": 0.9159613318061458,
363
+ "eval_precision_micro": 0.9365456396335257,
364
+ "eval_precision_weighted": 0.9365041505044936,
365
+ "eval_recall_macro": 0.9139276800740521,
366
+ "eval_recall_micro": 0.9365456396335257,
367
+ "eval_recall_weighted": 0.9365456396335257,
368
+ "eval_runtime": 19.7723,
369
+ "eval_samples_per_second": 149.047,
370
+ "eval_steps_per_second": 9.357,
371
+ "step": 1107
372
+ }
373
+ ],
374
+ "logging_steps": 25,
375
+ "max_steps": 1107,
376
+ "num_input_tokens_seen": 0,
377
+ "num_train_epochs": 3,
378
+ "save_steps": 500,
379
+ "total_flos": 2.225421168402862e+17,
380
+ "train_batch_size": 8,
381
+ "trial_name": null,
382
+ "trial_params": null
383
+ }
checkpoint-1107/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4a3f118e06589730c1ee59afd753a5192fc193e94fe295ebe821bdf09a4fab1
3
+ size 5048
config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/convnextv2-tiny-1k-224",
3
+ "_num_labels": 5,
4
+ "architectures": [
5
+ "ConvNextV2ForImageClassification"
6
+ ],
7
+ "depths": [
8
+ 3,
9
+ 3,
10
+ 9,
11
+ 3
12
+ ],
13
+ "drop_path_rate": 0.0,
14
+ "hidden_act": "gelu",
15
+ "hidden_sizes": [
16
+ 96,
17
+ 192,
18
+ 384,
19
+ 768
20
+ ],
21
+ "id2label": {
22
+ "0": "food",
23
+ "1": "inside",
24
+ "2": "menu",
25
+ "3": "outside",
26
+ "4": "people"
27
+ },
28
+ "image_size": 224,
29
+ "initializer_range": 0.02,
30
+ "label2id": {
31
+ "food": 0,
32
+ "inside": 1,
33
+ "menu": 2,
34
+ "outside": 3,
35
+ "people": 4
36
+ },
37
+ "layer_norm_eps": 1e-12,
38
+ "model_type": "convnextv2",
39
+ "num_channels": 3,
40
+ "num_stages": 4,
41
+ "out_features": [
42
+ "stage4"
43
+ ],
44
+ "out_indices": [
45
+ 4
46
+ ],
47
+ "patch_size": 4,
48
+ "problem_type": "single_label_classification",
49
+ "stage_names": [
50
+ "stem",
51
+ "stage1",
52
+ "stage2",
53
+ "stage3",
54
+ "stage4"
55
+ ],
56
+ "torch_dtype": "float32",
57
+ "transformers_version": "4.40.1"
58
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f93c3ea4f6d582a9705f2c1a6a5023205bca5072bf88f23d270c97a488e79042
3
+ size 111505052
preprocessor_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "crop_pct",
7
+ "resample",
8
+ "do_rescale",
9
+ "rescale_factor",
10
+ "do_normalize",
11
+ "image_mean",
12
+ "image_std",
13
+ "return_tensors",
14
+ "data_format",
15
+ "input_data_format"
16
+ ],
17
+ "crop_pct": 0.875,
18
+ "do_normalize": true,
19
+ "do_rescale": true,
20
+ "do_resize": true,
21
+ "image_mean": [
22
+ 0.485,
23
+ 0.456,
24
+ 0.406
25
+ ],
26
+ "image_processor_type": "ConvNextImageProcessor",
27
+ "image_std": [
28
+ 0.229,
29
+ 0.224,
30
+ 0.225
31
+ ],
32
+ "resample": 3,
33
+ "rescale_factor": 0.00392156862745098,
34
+ "size": {
35
+ "shortest_edge": 224
36
+ }
37
+ }
runs/May21_08-46-50_r-guntramg-autotrain-vfkldfv8-6db73-z9zi5/events.out.tfevents.1716281211.r-guntramg-autotrain-vfkldfv8-6db73-z9zi5.128.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d26bc278a8e510fa609dc72ee04707fd22ebfe8a4ac0e3da6396d5af941f485d
3
- size 4967
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f736de61f182d2db7f4199ce52104e45f333cf137f907aa40d06f64b0ec91fb
3
+ size 17084
runs/May21_08-46-50_r-guntramg-autotrain-vfkldfv8-6db73-z9zi5/events.out.tfevents.1716281454.r-guntramg-autotrain-vfkldfv8-6db73-z9zi5.128.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f665f28bf369698d97e14f8a85f7bf7b1b8399999633624a192ed5fdaf7bc85a
3
+ size 921
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4a3f118e06589730c1ee59afd753a5192fc193e94fe295ebe821bdf09a4fab1
3
+ size 5048
training_params.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "autotrain-6doma-5m8vf/autotrain-data",
3
+ "model": "facebook/convnextv2-tiny-1k-224",
4
+ "username": "GuntramG",
5
+ "lr": 5e-05,
6
+ "epochs": 3,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "logging_steps": -1,
18
+ "project_name": "autotrain-6doma-5m8vf",
19
+ "auto_find_batch_size": false,
20
+ "mixed_precision": "fp16",
21
+ "save_total_limit": 1,
22
+ "push_to_hub": true,
23
+ "evaluation_strategy": "epoch",
24
+ "image_column": "autotrain_image",
25
+ "target_column": "autotrain_label",
26
+ "log": "tensorboard",
27
+ "early_stopping_patience": 5,
28
+ "early_stopping_threshold": 0.01
29
+ }