howdyaendra commited on
Commit
61f26bb
1 Parent(s): 94bf55f

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - image-classification
6
+ widget:
7
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
8
+ example_title: Tiger
9
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
10
+ example_title: Teapot
11
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
12
+ example_title: Palace
13
+ datasets:
14
+ - xblock-large-patch3-224/autotrain-data
15
+ ---
16
+
17
+ # Model Trained Using AutoTrain
18
+
19
+ - Problem type: Image Classification
20
+
21
+ ## Validation Metrics
22
+ loss: 0.37855324149131775
23
+
24
+ f1_macro: 0.86094260720702
25
+
26
+ f1_micro: 0.8886217948717948
27
+
28
+ f1_weighted: 0.883196165156119
29
+
30
+ precision_macro: 0.8961444617693151
31
+
32
+ precision_micro: 0.8886217948717948
33
+
34
+ precision_weighted: 0.8922651559280282
35
+
36
+ recall_macro: 0.8524486181675118
37
+
38
+ recall_micro: 0.8886217948717948
39
+
40
+ recall_weighted: 0.8886217948717948
41
+
42
+ accuracy: 0.8886217948717948
checkpoint-1872/config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "howdyaendra/xblock-large-patch2-224",
3
+ "_num_labels": 11,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 1024,
12
+ "id2label": {
13
+ "0": "altright",
14
+ "1": "bluesky",
15
+ "2": "facebook",
16
+ "3": "fediverse",
17
+ "4": "instagram",
18
+ "5": "irrelevant",
19
+ "6": "ngl",
20
+ "7": "reddit",
21
+ "8": "threads",
22
+ "9": "tumblr",
23
+ "10": "twitter"
24
+ },
25
+ "image_size": 224,
26
+ "initializer_range": 0.02,
27
+ "intermediate_size": 4096,
28
+ "label2id": {
29
+ "altright": 0,
30
+ "bluesky": 1,
31
+ "facebook": 2,
32
+ "fediverse": 3,
33
+ "instagram": 4,
34
+ "irrelevant": 5,
35
+ "ngl": 6,
36
+ "reddit": 7,
37
+ "threads": 8,
38
+ "tumblr": 9,
39
+ "twitter": 10
40
+ },
41
+ "layer_norm_eps": 1e-12,
42
+ "model_type": "vit",
43
+ "num_attention_heads": 16,
44
+ "num_channels": 3,
45
+ "num_hidden_layers": 24,
46
+ "patch_size": 16,
47
+ "problem_type": "single_label_classification",
48
+ "qkv_bias": true,
49
+ "torch_dtype": "float32",
50
+ "transformers_version": "4.40.1"
51
+ }
checkpoint-1872/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33a192b221f88e6401d7a41ae47b4bb2306996c5d083d101b46ad946f5a993c5
3
+ size 1213298172
checkpoint-1872/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb6d57b1b76a62d92f9b1cfe9e85b82b7569fbc6fd771dd1eb65ee7953928e79
3
+ size 2426831274
checkpoint-1872/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e9fc52a27790ad5feb6ce8b92f10a90e40be76d441ba0254932062185798702
3
+ size 14244
checkpoint-1872/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64a708ef34e310c26770b3a84bf92a29c21d221438c4c23893c2307f36c83637
3
+ size 1064
checkpoint-1872/trainer_state.json ADDED
@@ -0,0 +1,593 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.37855324149131775,
3
+ "best_model_checkpoint": "xblock-large-patch3-224/checkpoint-1872",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1872,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.04006410256410257,
13
+ "grad_norm": 8.251535415649414,
14
+ "learning_rate": 6.3829787234042555e-06,
15
+ "loss": 1.3524,
16
+ "step": 25
17
+ },
18
+ {
19
+ "epoch": 0.08012820512820513,
20
+ "grad_norm": 12.843944549560547,
21
+ "learning_rate": 1.3031914893617023e-05,
22
+ "loss": 0.8737,
23
+ "step": 50
24
+ },
25
+ {
26
+ "epoch": 0.1201923076923077,
27
+ "grad_norm": 15.241576194763184,
28
+ "learning_rate": 1.9414893617021276e-05,
29
+ "loss": 0.6916,
30
+ "step": 75
31
+ },
32
+ {
33
+ "epoch": 0.16025641025641027,
34
+ "grad_norm": 6.287885665893555,
35
+ "learning_rate": 2.6063829787234046e-05,
36
+ "loss": 0.8028,
37
+ "step": 100
38
+ },
39
+ {
40
+ "epoch": 0.20032051282051283,
41
+ "grad_norm": 4.597654819488525,
42
+ "learning_rate": 3.271276595744681e-05,
43
+ "loss": 0.7592,
44
+ "step": 125
45
+ },
46
+ {
47
+ "epoch": 0.2403846153846154,
48
+ "grad_norm": 3.4197440147399902,
49
+ "learning_rate": 3.936170212765958e-05,
50
+ "loss": 0.8431,
51
+ "step": 150
52
+ },
53
+ {
54
+ "epoch": 0.28044871794871795,
55
+ "grad_norm": 7.936722755432129,
56
+ "learning_rate": 4.601063829787234e-05,
57
+ "loss": 0.6737,
58
+ "step": 175
59
+ },
60
+ {
61
+ "epoch": 0.32051282051282054,
62
+ "grad_norm": 27.45574188232422,
63
+ "learning_rate": 4.970308788598575e-05,
64
+ "loss": 0.9851,
65
+ "step": 200
66
+ },
67
+ {
68
+ "epoch": 0.3605769230769231,
69
+ "grad_norm": 7.184787273406982,
70
+ "learning_rate": 4.896080760095012e-05,
71
+ "loss": 0.9612,
72
+ "step": 225
73
+ },
74
+ {
75
+ "epoch": 0.40064102564102566,
76
+ "grad_norm": 5.246083736419678,
77
+ "learning_rate": 4.821852731591449e-05,
78
+ "loss": 0.7614,
79
+ "step": 250
80
+ },
81
+ {
82
+ "epoch": 0.4407051282051282,
83
+ "grad_norm": 14.747861862182617,
84
+ "learning_rate": 4.7476247030878864e-05,
85
+ "loss": 0.9273,
86
+ "step": 275
87
+ },
88
+ {
89
+ "epoch": 0.4807692307692308,
90
+ "grad_norm": 10.44289779663086,
91
+ "learning_rate": 4.673396674584323e-05,
92
+ "loss": 0.8679,
93
+ "step": 300
94
+ },
95
+ {
96
+ "epoch": 0.5208333333333334,
97
+ "grad_norm": 11.92969799041748,
98
+ "learning_rate": 4.5991686460807604e-05,
99
+ "loss": 0.8816,
100
+ "step": 325
101
+ },
102
+ {
103
+ "epoch": 0.5608974358974359,
104
+ "grad_norm": 8.216315269470215,
105
+ "learning_rate": 4.524940617577197e-05,
106
+ "loss": 0.8063,
107
+ "step": 350
108
+ },
109
+ {
110
+ "epoch": 0.6009615384615384,
111
+ "grad_norm": 9.245816230773926,
112
+ "learning_rate": 4.4507125890736345e-05,
113
+ "loss": 0.8215,
114
+ "step": 375
115
+ },
116
+ {
117
+ "epoch": 0.6410256410256411,
118
+ "grad_norm": 8.622091293334961,
119
+ "learning_rate": 4.376484560570072e-05,
120
+ "loss": 0.9864,
121
+ "step": 400
122
+ },
123
+ {
124
+ "epoch": 0.6810897435897436,
125
+ "grad_norm": 4.5337419509887695,
126
+ "learning_rate": 4.3022565320665086e-05,
127
+ "loss": 0.6767,
128
+ "step": 425
129
+ },
130
+ {
131
+ "epoch": 0.7211538461538461,
132
+ "grad_norm": 6.817728519439697,
133
+ "learning_rate": 4.228028503562946e-05,
134
+ "loss": 0.8387,
135
+ "step": 450
136
+ },
137
+ {
138
+ "epoch": 0.7612179487179487,
139
+ "grad_norm": 10.498202323913574,
140
+ "learning_rate": 4.153800475059383e-05,
141
+ "loss": 0.7565,
142
+ "step": 475
143
+ },
144
+ {
145
+ "epoch": 0.8012820512820513,
146
+ "grad_norm": 13.513296127319336,
147
+ "learning_rate": 4.07957244655582e-05,
148
+ "loss": 0.6909,
149
+ "step": 500
150
+ },
151
+ {
152
+ "epoch": 0.8413461538461539,
153
+ "grad_norm": 13.574014663696289,
154
+ "learning_rate": 4.005344418052257e-05,
155
+ "loss": 0.757,
156
+ "step": 525
157
+ },
158
+ {
159
+ "epoch": 0.8814102564102564,
160
+ "grad_norm": 6.092533588409424,
161
+ "learning_rate": 3.9311163895486934e-05,
162
+ "loss": 0.8626,
163
+ "step": 550
164
+ },
165
+ {
166
+ "epoch": 0.9214743589743589,
167
+ "grad_norm": 14.241544723510742,
168
+ "learning_rate": 3.856888361045131e-05,
169
+ "loss": 0.6826,
170
+ "step": 575
171
+ },
172
+ {
173
+ "epoch": 0.9615384615384616,
174
+ "grad_norm": 4.413307189941406,
175
+ "learning_rate": 3.7826603325415675e-05,
176
+ "loss": 0.7986,
177
+ "step": 600
178
+ },
179
+ {
180
+ "epoch": 1.0,
181
+ "eval_accuracy": 0.844551282051282,
182
+ "eval_f1_macro": 0.8160973181841822,
183
+ "eval_f1_micro": 0.844551282051282,
184
+ "eval_f1_weighted": 0.8402659980370851,
185
+ "eval_loss": 0.44926634430885315,
186
+ "eval_precision_macro": 0.8683442897703473,
187
+ "eval_precision_micro": 0.844551282051282,
188
+ "eval_precision_weighted": 0.8585860519825896,
189
+ "eval_recall_macro": 0.7986028134342184,
190
+ "eval_recall_micro": 0.844551282051282,
191
+ "eval_recall_weighted": 0.844551282051282,
192
+ "eval_runtime": 1324.3117,
193
+ "eval_samples_per_second": 0.942,
194
+ "eval_steps_per_second": 0.059,
195
+ "step": 624
196
+ },
197
+ {
198
+ "epoch": 1.001602564102564,
199
+ "grad_norm": 4.381113052368164,
200
+ "learning_rate": 3.708432304038005e-05,
201
+ "loss": 0.861,
202
+ "step": 625
203
+ },
204
+ {
205
+ "epoch": 1.0416666666666667,
206
+ "grad_norm": 5.680240631103516,
207
+ "learning_rate": 3.6342042755344416e-05,
208
+ "loss": 0.6535,
209
+ "step": 650
210
+ },
211
+ {
212
+ "epoch": 1.0817307692307692,
213
+ "grad_norm": 5.435142993927002,
214
+ "learning_rate": 3.559976247030879e-05,
215
+ "loss": 0.7551,
216
+ "step": 675
217
+ },
218
+ {
219
+ "epoch": 1.1217948717948718,
220
+ "grad_norm": 18.080663681030273,
221
+ "learning_rate": 3.4857482185273164e-05,
222
+ "loss": 0.7618,
223
+ "step": 700
224
+ },
225
+ {
226
+ "epoch": 1.1618589743589745,
227
+ "grad_norm": 12.246950149536133,
228
+ "learning_rate": 3.411520190023753e-05,
229
+ "loss": 0.7952,
230
+ "step": 725
231
+ },
232
+ {
233
+ "epoch": 1.2019230769230769,
234
+ "grad_norm": 9.545283317565918,
235
+ "learning_rate": 3.3372921615201904e-05,
236
+ "loss": 0.6002,
237
+ "step": 750
238
+ },
239
+ {
240
+ "epoch": 1.2419871794871795,
241
+ "grad_norm": 8.819951057434082,
242
+ "learning_rate": 3.263064133016627e-05,
243
+ "loss": 0.7556,
244
+ "step": 775
245
+ },
246
+ {
247
+ "epoch": 1.282051282051282,
248
+ "grad_norm": 10.624255180358887,
249
+ "learning_rate": 3.1888361045130645e-05,
250
+ "loss": 0.7373,
251
+ "step": 800
252
+ },
253
+ {
254
+ "epoch": 1.3221153846153846,
255
+ "grad_norm": 7.088049411773682,
256
+ "learning_rate": 3.114608076009501e-05,
257
+ "loss": 0.6166,
258
+ "step": 825
259
+ },
260
+ {
261
+ "epoch": 1.3621794871794872,
262
+ "grad_norm": 7.568330764770508,
263
+ "learning_rate": 3.0403800475059386e-05,
264
+ "loss": 0.6451,
265
+ "step": 850
266
+ },
267
+ {
268
+ "epoch": 1.4022435897435899,
269
+ "grad_norm": 6.381913185119629,
270
+ "learning_rate": 2.9661520190023756e-05,
271
+ "loss": 0.6397,
272
+ "step": 875
273
+ },
274
+ {
275
+ "epoch": 1.4423076923076923,
276
+ "grad_norm": 9.472349166870117,
277
+ "learning_rate": 2.8919239904988127e-05,
278
+ "loss": 0.6227,
279
+ "step": 900
280
+ },
281
+ {
282
+ "epoch": 1.482371794871795,
283
+ "grad_norm": 4.770060062408447,
284
+ "learning_rate": 2.8176959619952497e-05,
285
+ "loss": 0.6374,
286
+ "step": 925
287
+ },
288
+ {
289
+ "epoch": 1.5224358974358974,
290
+ "grad_norm": 6.0947771072387695,
291
+ "learning_rate": 2.7434679334916867e-05,
292
+ "loss": 0.5955,
293
+ "step": 950
294
+ },
295
+ {
296
+ "epoch": 1.5625,
297
+ "grad_norm": 5.420393466949463,
298
+ "learning_rate": 2.6692399049881234e-05,
299
+ "loss": 0.7621,
300
+ "step": 975
301
+ },
302
+ {
303
+ "epoch": 1.6025641025641026,
304
+ "grad_norm": 8.125214576721191,
305
+ "learning_rate": 2.5950118764845605e-05,
306
+ "loss": 0.624,
307
+ "step": 1000
308
+ },
309
+ {
310
+ "epoch": 1.6426282051282053,
311
+ "grad_norm": 11.498839378356934,
312
+ "learning_rate": 2.5207838479809975e-05,
313
+ "loss": 0.6889,
314
+ "step": 1025
315
+ },
316
+ {
317
+ "epoch": 1.6826923076923077,
318
+ "grad_norm": 6.599166393280029,
319
+ "learning_rate": 2.446555819477435e-05,
320
+ "loss": 0.5743,
321
+ "step": 1050
322
+ },
323
+ {
324
+ "epoch": 1.7227564102564101,
325
+ "grad_norm": 18.760013580322266,
326
+ "learning_rate": 2.372327790973872e-05,
327
+ "loss": 0.6392,
328
+ "step": 1075
329
+ },
330
+ {
331
+ "epoch": 1.7628205128205128,
332
+ "grad_norm": 10.053047180175781,
333
+ "learning_rate": 2.2980997624703086e-05,
334
+ "loss": 0.7208,
335
+ "step": 1100
336
+ },
337
+ {
338
+ "epoch": 1.8028846153846154,
339
+ "grad_norm": 6.171570777893066,
340
+ "learning_rate": 2.223871733966746e-05,
341
+ "loss": 0.5817,
342
+ "step": 1125
343
+ },
344
+ {
345
+ "epoch": 1.842948717948718,
346
+ "grad_norm": 5.438927173614502,
347
+ "learning_rate": 2.149643705463183e-05,
348
+ "loss": 0.4576,
349
+ "step": 1150
350
+ },
351
+ {
352
+ "epoch": 1.8830128205128205,
353
+ "grad_norm": 8.946006774902344,
354
+ "learning_rate": 2.07541567695962e-05,
355
+ "loss": 0.695,
356
+ "step": 1175
357
+ },
358
+ {
359
+ "epoch": 1.9230769230769231,
360
+ "grad_norm": 7.02325439453125,
361
+ "learning_rate": 2.001187648456057e-05,
362
+ "loss": 0.6768,
363
+ "step": 1200
364
+ },
365
+ {
366
+ "epoch": 1.9631410256410255,
367
+ "grad_norm": 4.542496204376221,
368
+ "learning_rate": 1.926959619952494e-05,
369
+ "loss": 0.6592,
370
+ "step": 1225
371
+ },
372
+ {
373
+ "epoch": 2.0,
374
+ "eval_accuracy": 0.8717948717948718,
375
+ "eval_f1_macro": 0.8420194226887088,
376
+ "eval_f1_micro": 0.8717948717948718,
377
+ "eval_f1_weighted": 0.8664623444122613,
378
+ "eval_loss": 0.39590924978256226,
379
+ "eval_precision_macro": 0.8724657298009468,
380
+ "eval_precision_micro": 0.8717948717948718,
381
+ "eval_precision_weighted": 0.8825043036411662,
382
+ "eval_recall_macro": 0.8395777006090396,
383
+ "eval_recall_micro": 0.8717948717948718,
384
+ "eval_recall_weighted": 0.8717948717948718,
385
+ "eval_runtime": 1293.5897,
386
+ "eval_samples_per_second": 0.965,
387
+ "eval_steps_per_second": 0.06,
388
+ "step": 1248
389
+ },
390
+ {
391
+ "epoch": 2.003205128205128,
392
+ "grad_norm": 7.841921329498291,
393
+ "learning_rate": 1.8527315914489312e-05,
394
+ "loss": 0.6025,
395
+ "step": 1250
396
+ },
397
+ {
398
+ "epoch": 2.043269230769231,
399
+ "grad_norm": 9.250267028808594,
400
+ "learning_rate": 1.7785035629453682e-05,
401
+ "loss": 0.4597,
402
+ "step": 1275
403
+ },
404
+ {
405
+ "epoch": 2.0833333333333335,
406
+ "grad_norm": 3.1655519008636475,
407
+ "learning_rate": 1.7042755344418056e-05,
408
+ "loss": 0.3962,
409
+ "step": 1300
410
+ },
411
+ {
412
+ "epoch": 2.123397435897436,
413
+ "grad_norm": 5.661818981170654,
414
+ "learning_rate": 1.6300475059382423e-05,
415
+ "loss": 0.6639,
416
+ "step": 1325
417
+ },
418
+ {
419
+ "epoch": 2.1634615384615383,
420
+ "grad_norm": 3.282247304916382,
421
+ "learning_rate": 1.5558194774346793e-05,
422
+ "loss": 0.464,
423
+ "step": 1350
424
+ },
425
+ {
426
+ "epoch": 2.203525641025641,
427
+ "grad_norm": 3.829854965209961,
428
+ "learning_rate": 1.4815914489311164e-05,
429
+ "loss": 0.4943,
430
+ "step": 1375
431
+ },
432
+ {
433
+ "epoch": 2.2435897435897436,
434
+ "grad_norm": 5.670718193054199,
435
+ "learning_rate": 1.4073634204275534e-05,
436
+ "loss": 0.6288,
437
+ "step": 1400
438
+ },
439
+ {
440
+ "epoch": 2.2836538461538463,
441
+ "grad_norm": 10.35056209564209,
442
+ "learning_rate": 1.3331353919239906e-05,
443
+ "loss": 0.4846,
444
+ "step": 1425
445
+ },
446
+ {
447
+ "epoch": 2.323717948717949,
448
+ "grad_norm": 0.6631863117218018,
449
+ "learning_rate": 1.2589073634204277e-05,
450
+ "loss": 0.3622,
451
+ "step": 1450
452
+ },
453
+ {
454
+ "epoch": 2.363782051282051,
455
+ "grad_norm": 6.619789123535156,
456
+ "learning_rate": 1.1846793349168647e-05,
457
+ "loss": 0.5013,
458
+ "step": 1475
459
+ },
460
+ {
461
+ "epoch": 2.4038461538461537,
462
+ "grad_norm": 6.15729284286499,
463
+ "learning_rate": 1.1104513064133017e-05,
464
+ "loss": 0.4955,
465
+ "step": 1500
466
+ },
467
+ {
468
+ "epoch": 2.4439102564102564,
469
+ "grad_norm": 0.05698273330926895,
470
+ "learning_rate": 1.0362232779097388e-05,
471
+ "loss": 0.4693,
472
+ "step": 1525
473
+ },
474
+ {
475
+ "epoch": 2.483974358974359,
476
+ "grad_norm": 3.433866262435913,
477
+ "learning_rate": 9.619952494061758e-06,
478
+ "loss": 0.4221,
479
+ "step": 1550
480
+ },
481
+ {
482
+ "epoch": 2.5240384615384617,
483
+ "grad_norm": 6.545916557312012,
484
+ "learning_rate": 8.877672209026128e-06,
485
+ "loss": 0.6492,
486
+ "step": 1575
487
+ },
488
+ {
489
+ "epoch": 2.564102564102564,
490
+ "grad_norm": 12.447614669799805,
491
+ "learning_rate": 8.135391923990499e-06,
492
+ "loss": 0.4888,
493
+ "step": 1600
494
+ },
495
+ {
496
+ "epoch": 2.6041666666666665,
497
+ "grad_norm": 6.291492462158203,
498
+ "learning_rate": 7.393111638954869e-06,
499
+ "loss": 0.4994,
500
+ "step": 1625
501
+ },
502
+ {
503
+ "epoch": 2.644230769230769,
504
+ "grad_norm": 5.843363285064697,
505
+ "learning_rate": 6.6508313539192404e-06,
506
+ "loss": 0.6511,
507
+ "step": 1650
508
+ },
509
+ {
510
+ "epoch": 2.684294871794872,
511
+ "grad_norm": 6.965985298156738,
512
+ "learning_rate": 5.908551068883611e-06,
513
+ "loss": 0.4697,
514
+ "step": 1675
515
+ },
516
+ {
517
+ "epoch": 2.7243589743589745,
518
+ "grad_norm": 7.191352367401123,
519
+ "learning_rate": 5.166270783847981e-06,
520
+ "loss": 0.4224,
521
+ "step": 1700
522
+ },
523
+ {
524
+ "epoch": 2.7644230769230766,
525
+ "grad_norm": 14.876649856567383,
526
+ "learning_rate": 4.4239904988123516e-06,
527
+ "loss": 0.6627,
528
+ "step": 1725
529
+ },
530
+ {
531
+ "epoch": 2.8044871794871797,
532
+ "grad_norm": 6.894955158233643,
533
+ "learning_rate": 3.681710213776722e-06,
534
+ "loss": 0.5652,
535
+ "step": 1750
536
+ },
537
+ {
538
+ "epoch": 2.844551282051282,
539
+ "grad_norm": 5.6652984619140625,
540
+ "learning_rate": 2.9394299287410927e-06,
541
+ "loss": 0.6411,
542
+ "step": 1775
543
+ },
544
+ {
545
+ "epoch": 2.8846153846153846,
546
+ "grad_norm": 8.991388320922852,
547
+ "learning_rate": 2.197149643705463e-06,
548
+ "loss": 0.4483,
549
+ "step": 1800
550
+ },
551
+ {
552
+ "epoch": 2.9246794871794872,
553
+ "grad_norm": 8.376466751098633,
554
+ "learning_rate": 1.4548693586698337e-06,
555
+ "loss": 0.4551,
556
+ "step": 1825
557
+ },
558
+ {
559
+ "epoch": 2.96474358974359,
560
+ "grad_norm": 3.4173240661621094,
561
+ "learning_rate": 7.422802850356294e-07,
562
+ "loss": 0.4227,
563
+ "step": 1850
564
+ },
565
+ {
566
+ "epoch": 3.0,
567
+ "eval_accuracy": 0.8886217948717948,
568
+ "eval_f1_macro": 0.86094260720702,
569
+ "eval_f1_micro": 0.8886217948717948,
570
+ "eval_f1_weighted": 0.883196165156119,
571
+ "eval_loss": 0.37855324149131775,
572
+ "eval_precision_macro": 0.8961444617693151,
573
+ "eval_precision_micro": 0.8886217948717948,
574
+ "eval_precision_weighted": 0.8922651559280282,
575
+ "eval_recall_macro": 0.8524486181675118,
576
+ "eval_recall_micro": 0.8886217948717948,
577
+ "eval_recall_weighted": 0.8886217948717948,
578
+ "eval_runtime": 1323.0713,
579
+ "eval_samples_per_second": 0.943,
580
+ "eval_steps_per_second": 0.059,
581
+ "step": 1872
582
+ }
583
+ ],
584
+ "logging_steps": 25,
585
+ "max_steps": 1872,
586
+ "num_input_tokens_seen": 0,
587
+ "num_train_epochs": 3,
588
+ "save_steps": 500,
589
+ "total_flos": 4.1009193984929587e+18,
590
+ "train_batch_size": 8,
591
+ "trial_name": null,
592
+ "trial_params": null
593
+ }
checkpoint-1872/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db92048ee3b95441a547e425f5207697cec5ea7a83eaed11cf7a1e8f1fd35070
3
+ size 4984
config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "howdyaendra/xblock-large-patch2-224",
3
+ "_num_labels": 11,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 1024,
12
+ "id2label": {
13
+ "0": "altright",
14
+ "1": "bluesky",
15
+ "2": "facebook",
16
+ "3": "fediverse",
17
+ "4": "instagram",
18
+ "5": "irrelevant",
19
+ "6": "ngl",
20
+ "7": "reddit",
21
+ "8": "threads",
22
+ "9": "tumblr",
23
+ "10": "twitter"
24
+ },
25
+ "image_size": 224,
26
+ "initializer_range": 0.02,
27
+ "intermediate_size": 4096,
28
+ "label2id": {
29
+ "altright": 0,
30
+ "bluesky": 1,
31
+ "facebook": 2,
32
+ "fediverse": 3,
33
+ "instagram": 4,
34
+ "irrelevant": 5,
35
+ "ngl": 6,
36
+ "reddit": 7,
37
+ "threads": 8,
38
+ "tumblr": 9,
39
+ "twitter": 10
40
+ },
41
+ "layer_norm_eps": 1e-12,
42
+ "model_type": "vit",
43
+ "num_attention_heads": 16,
44
+ "num_channels": 3,
45
+ "num_hidden_layers": 24,
46
+ "patch_size": 16,
47
+ "problem_type": "single_label_classification",
48
+ "qkv_bias": true,
49
+ "torch_dtype": "float32",
50
+ "transformers_version": "4.40.1"
51
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33a192b221f88e6401d7a41ae47b4bb2306996c5d083d101b46ad946f5a993c5
3
+ size 1213298172
preprocessor_config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_rescale",
8
+ "rescale_factor",
9
+ "do_normalize",
10
+ "image_mean",
11
+ "image_std",
12
+ "return_tensors",
13
+ "data_format",
14
+ "input_data_format"
15
+ ],
16
+ "do_normalize": true,
17
+ "do_rescale": true,
18
+ "do_resize": true,
19
+ "image_mean": [
20
+ 0.5,
21
+ 0.5,
22
+ 0.5
23
+ ],
24
+ "image_processor_type": "ViTImageProcessor",
25
+ "image_std": [
26
+ 0.5,
27
+ 0.5,
28
+ 0.5
29
+ ],
30
+ "resample": 2,
31
+ "rescale_factor": 0.00392156862745098,
32
+ "size": {
33
+ "height": 224,
34
+ "width": 224
35
+ }
36
+ }
runs/May07_01-41-03_d8363a52208a/events.out.tfevents.1715046065.d8363a52208a.122.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3be8f14d8dfa230a87a9477ad8dd1cfa97360918d5bd7e5b6054c792936561d
3
- size 22199
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c7ebf9cd7cf21e11e1c3e7e2759cae8924bbad307734818086f82bfe819f20e
3
+ size 23597
runs/May07_01-41-03_d8363a52208a/events.out.tfevents.1715075176.d8363a52208a.122.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3100f358d35c730f2524b7d059d273b06ec9a65cad5d039ba94fe28e7b730354
3
+ size 921
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db92048ee3b95441a547e425f5207697cec5ea7a83eaed11cf7a1e8f1fd35070
3
+ size 4984
training_params.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "xblock-large-patch3-224/autotrain-data",
3
+ "model": "howdyaendra/xblock-large-patch2-224",
4
+ "username": "howdyaendra",
5
+ "lr": 5e-05,
6
+ "epochs": 3,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "logging_steps": -1,
18
+ "project_name": "xblock-large-patch3-224",
19
+ "auto_find_batch_size": false,
20
+ "mixed_precision": "fp16",
21
+ "save_total_limit": 1,
22
+ "push_to_hub": true,
23
+ "evaluation_strategy": "epoch",
24
+ "image_column": "autotrain_image",
25
+ "target_column": "autotrain_label",
26
+ "log": "tensorboard"
27
+ }