howdyaendra commited on
Commit
35a8681
1 Parent(s): 8ccfaa8

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - image-classification
6
+ widget:
7
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
8
+ example_title: Tiger
9
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
10
+ example_title: Teapot
11
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
12
+ example_title: Palace
13
+ datasets:
14
+ - xblock-social-screenshots-2/autotrain-data
15
+ ---
16
+
17
+ # Model Trained Using AutoTrain
18
+
19
+ - Problem type: Image Classification
20
+
21
+ ## Validation Metrics
22
+ loss: 0.9948348999023438
23
+
24
+ f1_macro: 0.23526955685803672
25
+
26
+ f1_micro: 0.5823986194995686
27
+
28
+ f1_weighted: 0.5233223208450835
29
+
30
+ precision_macro: 0.2973178657776674
31
+
32
+ precision_micro: 0.5823986194995686
33
+
34
+ precision_weighted: 0.5462756602202016
35
+
36
+ recall_macro: 0.24224828041633692
37
+
38
+ recall_micro: 0.5823986194995686
39
+
40
+ recall_weighted: 0.5823986194995686
41
+
42
+ accuracy: 0.5823986194995686
checkpoint-1737/config.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "howdyaendra/autotrain-xblock-twitter-1",
3
+ "_num_labels": 12,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "altright",
14
+ "1": "bluesky",
15
+ "2": "facebook",
16
+ "3": "fediverse",
17
+ "4": "instagram",
18
+ "5": "reddit",
19
+ "6": "social-screenshot",
20
+ "7": "threads",
21
+ "8": "tumblr",
22
+ "9": "twitter",
23
+ "10": "uncategorised",
24
+ "11": "unrecognised-screenshot"
25
+ },
26
+ "image_size": 224,
27
+ "initializer_range": 0.02,
28
+ "intermediate_size": 3072,
29
+ "label2id": {
30
+ "altright": 0,
31
+ "bluesky": 1,
32
+ "facebook": 2,
33
+ "fediverse": 3,
34
+ "instagram": 4,
35
+ "reddit": 5,
36
+ "social-screenshot": 6,
37
+ "threads": 7,
38
+ "tumblr": 8,
39
+ "twitter": 9,
40
+ "uncategorised": 10,
41
+ "unrecognised-screenshot": 11
42
+ },
43
+ "layer_norm_eps": 1e-12,
44
+ "model_type": "vit",
45
+ "num_attention_heads": 12,
46
+ "num_channels": 3,
47
+ "num_hidden_layers": 12,
48
+ "patch_size": 16,
49
+ "problem_type": "single_label_classification",
50
+ "qkv_bias": true,
51
+ "torch_dtype": "float32",
52
+ "transformers_version": "4.39.0"
53
+ }
checkpoint-1737/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2111f184b5b09d458eae2038e8e5e262954fd65eb0fe90c9d8b7c6477652a053
3
+ size 343254736
checkpoint-1737/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9235105f2e1b96a0b71bbbde7cf7bf9813e49f94f88ca816b336dec7a4114fae
3
+ size 686630330
checkpoint-1737/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32498e8d3452a86b1bd0e566d868e6e2e9191c430f6a378c37d3eb335723947d
3
+ size 14244
checkpoint-1737/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d5d60e9b97a61c885fca1690e034c3955261a2c2676f7037ab5f0f2616c7ad2
3
+ size 1064
checkpoint-1737/trainer_state.json ADDED
@@ -0,0 +1,509 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9948348999023438,
3
+ "best_model_checkpoint": "xblock-social-screenshots-2/checkpoint-1737",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1737,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.05,
13
+ "grad_norm": 12.944127082824707,
14
+ "learning_rate": 7.4712643678160925e-06,
15
+ "loss": 2.5595,
16
+ "step": 28
17
+ },
18
+ {
19
+ "epoch": 0.1,
20
+ "grad_norm": 10.593692779541016,
21
+ "learning_rate": 1.5517241379310346e-05,
22
+ "loss": 2.1472,
23
+ "step": 56
24
+ },
25
+ {
26
+ "epoch": 0.15,
27
+ "grad_norm": 9.828369140625,
28
+ "learning_rate": 2.327586206896552e-05,
29
+ "loss": 1.7244,
30
+ "step": 84
31
+ },
32
+ {
33
+ "epoch": 0.19,
34
+ "grad_norm": 8.050122261047363,
35
+ "learning_rate": 3.132183908045977e-05,
36
+ "loss": 1.5718,
37
+ "step": 112
38
+ },
39
+ {
40
+ "epoch": 0.24,
41
+ "grad_norm": 7.566161155700684,
42
+ "learning_rate": 3.936781609195402e-05,
43
+ "loss": 1.4015,
44
+ "step": 140
45
+ },
46
+ {
47
+ "epoch": 0.29,
48
+ "grad_norm": 8.954880714416504,
49
+ "learning_rate": 4.741379310344828e-05,
50
+ "loss": 1.3705,
51
+ "step": 168
52
+ },
53
+ {
54
+ "epoch": 0.34,
55
+ "grad_norm": 8.316436767578125,
56
+ "learning_rate": 4.9392194497760715e-05,
57
+ "loss": 1.3727,
58
+ "step": 196
59
+ },
60
+ {
61
+ "epoch": 0.39,
62
+ "grad_norm": 4.580837726593018,
63
+ "learning_rate": 4.8496481126039675e-05,
64
+ "loss": 1.2824,
65
+ "step": 224
66
+ },
67
+ {
68
+ "epoch": 0.44,
69
+ "grad_norm": 12.312744140625,
70
+ "learning_rate": 4.760076775431862e-05,
71
+ "loss": 1.3119,
72
+ "step": 252
73
+ },
74
+ {
75
+ "epoch": 0.48,
76
+ "grad_norm": 10.076079368591309,
77
+ "learning_rate": 4.670505438259757e-05,
78
+ "loss": 1.252,
79
+ "step": 280
80
+ },
81
+ {
82
+ "epoch": 0.53,
83
+ "grad_norm": 9.073230743408203,
84
+ "learning_rate": 4.580934101087652e-05,
85
+ "loss": 1.3154,
86
+ "step": 308
87
+ },
88
+ {
89
+ "epoch": 0.58,
90
+ "grad_norm": 9.829570770263672,
91
+ "learning_rate": 4.491362763915547e-05,
92
+ "loss": 1.3688,
93
+ "step": 336
94
+ },
95
+ {
96
+ "epoch": 0.63,
97
+ "grad_norm": 7.165198802947998,
98
+ "learning_rate": 4.4017914267434426e-05,
99
+ "loss": 1.4376,
100
+ "step": 364
101
+ },
102
+ {
103
+ "epoch": 0.68,
104
+ "grad_norm": 8.911142349243164,
105
+ "learning_rate": 4.312220089571337e-05,
106
+ "loss": 1.3533,
107
+ "step": 392
108
+ },
109
+ {
110
+ "epoch": 0.73,
111
+ "grad_norm": 10.684996604919434,
112
+ "learning_rate": 4.2226487523992326e-05,
113
+ "loss": 1.2825,
114
+ "step": 420
115
+ },
116
+ {
117
+ "epoch": 0.77,
118
+ "grad_norm": 10.35905647277832,
119
+ "learning_rate": 4.133077415227127e-05,
120
+ "loss": 1.3382,
121
+ "step": 448
122
+ },
123
+ {
124
+ "epoch": 0.82,
125
+ "grad_norm": 9.137449264526367,
126
+ "learning_rate": 4.0435060780550225e-05,
127
+ "loss": 1.3054,
128
+ "step": 476
129
+ },
130
+ {
131
+ "epoch": 0.87,
132
+ "grad_norm": 6.174429416656494,
133
+ "learning_rate": 3.953934740882918e-05,
134
+ "loss": 1.4294,
135
+ "step": 504
136
+ },
137
+ {
138
+ "epoch": 0.92,
139
+ "grad_norm": 9.742380142211914,
140
+ "learning_rate": 3.864363403710813e-05,
141
+ "loss": 1.2502,
142
+ "step": 532
143
+ },
144
+ {
145
+ "epoch": 0.97,
146
+ "grad_norm": 7.022644519805908,
147
+ "learning_rate": 3.774792066538708e-05,
148
+ "loss": 1.2926,
149
+ "step": 560
150
+ },
151
+ {
152
+ "epoch": 1.0,
153
+ "eval_accuracy": 0.5591026747195859,
154
+ "eval_f1_macro": 0.17116779008569874,
155
+ "eval_f1_micro": 0.5591026747195859,
156
+ "eval_f1_weighted": 0.47555675422704724,
157
+ "eval_loss": 1.1608415842056274,
158
+ "eval_precision_macro": 0.17146675898922856,
159
+ "eval_precision_micro": 0.5591026747195859,
160
+ "eval_precision_weighted": 0.4635728502463956,
161
+ "eval_recall_macro": 0.1928392780592314,
162
+ "eval_recall_micro": 0.5591026747195859,
163
+ "eval_recall_weighted": 0.5591026747195859,
164
+ "eval_runtime": 47.359,
165
+ "eval_samples_per_second": 24.473,
166
+ "eval_steps_per_second": 1.541,
167
+ "step": 579
168
+ },
169
+ {
170
+ "epoch": 1.02,
171
+ "grad_norm": 7.127846717834473,
172
+ "learning_rate": 3.685220729366603e-05,
173
+ "loss": 1.161,
174
+ "step": 588
175
+ },
176
+ {
177
+ "epoch": 1.06,
178
+ "grad_norm": 6.150296211242676,
179
+ "learning_rate": 3.5956493921944976e-05,
180
+ "loss": 1.1594,
181
+ "step": 616
182
+ },
183
+ {
184
+ "epoch": 1.11,
185
+ "grad_norm": 6.171022415161133,
186
+ "learning_rate": 3.506078055022393e-05,
187
+ "loss": 1.1253,
188
+ "step": 644
189
+ },
190
+ {
191
+ "epoch": 1.16,
192
+ "grad_norm": 12.212858200073242,
193
+ "learning_rate": 3.416506717850288e-05,
194
+ "loss": 1.2149,
195
+ "step": 672
196
+ },
197
+ {
198
+ "epoch": 1.21,
199
+ "grad_norm": 8.228123664855957,
200
+ "learning_rate": 3.3269353806781835e-05,
201
+ "loss": 1.1905,
202
+ "step": 700
203
+ },
204
+ {
205
+ "epoch": 1.26,
206
+ "grad_norm": 6.095876693725586,
207
+ "learning_rate": 3.237364043506078e-05,
208
+ "loss": 1.1387,
209
+ "step": 728
210
+ },
211
+ {
212
+ "epoch": 1.31,
213
+ "grad_norm": 10.43623161315918,
214
+ "learning_rate": 3.1477927063339734e-05,
215
+ "loss": 1.2243,
216
+ "step": 756
217
+ },
218
+ {
219
+ "epoch": 1.35,
220
+ "grad_norm": 6.94878625869751,
221
+ "learning_rate": 3.058221369161868e-05,
222
+ "loss": 1.0566,
223
+ "step": 784
224
+ },
225
+ {
226
+ "epoch": 1.4,
227
+ "grad_norm": 10.651735305786133,
228
+ "learning_rate": 2.9686500319897637e-05,
229
+ "loss": 1.1425,
230
+ "step": 812
231
+ },
232
+ {
233
+ "epoch": 1.45,
234
+ "grad_norm": 8.426492691040039,
235
+ "learning_rate": 2.8790786948176586e-05,
236
+ "loss": 1.173,
237
+ "step": 840
238
+ },
239
+ {
240
+ "epoch": 1.5,
241
+ "grad_norm": 7.550307273864746,
242
+ "learning_rate": 2.789507357645554e-05,
243
+ "loss": 1.2052,
244
+ "step": 868
245
+ },
246
+ {
247
+ "epoch": 1.55,
248
+ "grad_norm": 9.533821105957031,
249
+ "learning_rate": 2.6999360204734485e-05,
250
+ "loss": 1.2043,
251
+ "step": 896
252
+ },
253
+ {
254
+ "epoch": 1.6,
255
+ "grad_norm": 7.776467323303223,
256
+ "learning_rate": 2.6103646833013435e-05,
257
+ "loss": 1.0199,
258
+ "step": 924
259
+ },
260
+ {
261
+ "epoch": 1.64,
262
+ "grad_norm": 9.101223945617676,
263
+ "learning_rate": 2.5207933461292388e-05,
264
+ "loss": 1.1681,
265
+ "step": 952
266
+ },
267
+ {
268
+ "epoch": 1.69,
269
+ "grad_norm": 7.713497161865234,
270
+ "learning_rate": 2.4312220089571338e-05,
271
+ "loss": 1.1177,
272
+ "step": 980
273
+ },
274
+ {
275
+ "epoch": 1.74,
276
+ "grad_norm": 7.732503414154053,
277
+ "learning_rate": 2.341650671785029e-05,
278
+ "loss": 1.1099,
279
+ "step": 1008
280
+ },
281
+ {
282
+ "epoch": 1.79,
283
+ "grad_norm": 9.252326965332031,
284
+ "learning_rate": 2.252079334612924e-05,
285
+ "loss": 1.2187,
286
+ "step": 1036
287
+ },
288
+ {
289
+ "epoch": 1.84,
290
+ "grad_norm": 7.301546096801758,
291
+ "learning_rate": 2.162507997440819e-05,
292
+ "loss": 1.0414,
293
+ "step": 1064
294
+ },
295
+ {
296
+ "epoch": 1.89,
297
+ "grad_norm": 8.406270027160645,
298
+ "learning_rate": 2.0729366602687143e-05,
299
+ "loss": 1.0479,
300
+ "step": 1092
301
+ },
302
+ {
303
+ "epoch": 1.93,
304
+ "grad_norm": 7.392611980438232,
305
+ "learning_rate": 1.9833653230966092e-05,
306
+ "loss": 1.2113,
307
+ "step": 1120
308
+ },
309
+ {
310
+ "epoch": 1.98,
311
+ "grad_norm": 7.991569995880127,
312
+ "learning_rate": 1.8937939859245045e-05,
313
+ "loss": 1.1279,
314
+ "step": 1148
315
+ },
316
+ {
317
+ "epoch": 2.0,
318
+ "eval_accuracy": 0.548748921484038,
319
+ "eval_f1_macro": 0.23354766712716346,
320
+ "eval_f1_micro": 0.548748921484038,
321
+ "eval_f1_weighted": 0.5206705245021934,
322
+ "eval_loss": 1.076399326324463,
323
+ "eval_precision_macro": 0.2591242803710448,
324
+ "eval_precision_micro": 0.548748921484038,
325
+ "eval_precision_weighted": 0.5086032820438354,
326
+ "eval_recall_macro": 0.2392300811955355,
327
+ "eval_recall_micro": 0.548748921484038,
328
+ "eval_recall_weighted": 0.548748921484038,
329
+ "eval_runtime": 45.7915,
330
+ "eval_samples_per_second": 25.31,
331
+ "eval_steps_per_second": 1.594,
332
+ "step": 1158
333
+ },
334
+ {
335
+ "epoch": 2.03,
336
+ "grad_norm": 8.247747421264648,
337
+ "learning_rate": 1.8042226487523995e-05,
338
+ "loss": 1.0308,
339
+ "step": 1176
340
+ },
341
+ {
342
+ "epoch": 2.08,
343
+ "grad_norm": 5.560178279876709,
344
+ "learning_rate": 1.7146513115802944e-05,
345
+ "loss": 1.0764,
346
+ "step": 1204
347
+ },
348
+ {
349
+ "epoch": 2.13,
350
+ "grad_norm": 6.093762397766113,
351
+ "learning_rate": 1.6250799744081894e-05,
352
+ "loss": 1.1365,
353
+ "step": 1232
354
+ },
355
+ {
356
+ "epoch": 2.18,
357
+ "grad_norm": 4.060959815979004,
358
+ "learning_rate": 1.5355086372360844e-05,
359
+ "loss": 1.0535,
360
+ "step": 1260
361
+ },
362
+ {
363
+ "epoch": 2.22,
364
+ "grad_norm": 7.9812822341918945,
365
+ "learning_rate": 1.4459373000639795e-05,
366
+ "loss": 1.0241,
367
+ "step": 1288
368
+ },
369
+ {
370
+ "epoch": 2.27,
371
+ "grad_norm": 5.996217250823975,
372
+ "learning_rate": 1.3563659628918746e-05,
373
+ "loss": 1.0137,
374
+ "step": 1316
375
+ },
376
+ {
377
+ "epoch": 2.32,
378
+ "grad_norm": 10.663773536682129,
379
+ "learning_rate": 1.2667946257197696e-05,
380
+ "loss": 1.0274,
381
+ "step": 1344
382
+ },
383
+ {
384
+ "epoch": 2.37,
385
+ "grad_norm": 9.829933166503906,
386
+ "learning_rate": 1.1772232885476649e-05,
387
+ "loss": 0.9893,
388
+ "step": 1372
389
+ },
390
+ {
391
+ "epoch": 2.42,
392
+ "grad_norm": 9.35245418548584,
393
+ "learning_rate": 1.0876519513755598e-05,
394
+ "loss": 1.0497,
395
+ "step": 1400
396
+ },
397
+ {
398
+ "epoch": 2.47,
399
+ "grad_norm": 9.116876602172852,
400
+ "learning_rate": 9.98080614203455e-06,
401
+ "loss": 0.972,
402
+ "step": 1428
403
+ },
404
+ {
405
+ "epoch": 2.51,
406
+ "grad_norm": 8.650375366210938,
407
+ "learning_rate": 9.085092770313499e-06,
408
+ "loss": 1.0058,
409
+ "step": 1456
410
+ },
411
+ {
412
+ "epoch": 2.56,
413
+ "grad_norm": 9.76453971862793,
414
+ "learning_rate": 8.18937939859245e-06,
415
+ "loss": 1.0451,
416
+ "step": 1484
417
+ },
418
+ {
419
+ "epoch": 2.61,
420
+ "grad_norm": 5.405215263366699,
421
+ "learning_rate": 7.293666026871402e-06,
422
+ "loss": 0.9628,
423
+ "step": 1512
424
+ },
425
+ {
426
+ "epoch": 2.66,
427
+ "grad_norm": 10.716626167297363,
428
+ "learning_rate": 6.397952655150352e-06,
429
+ "loss": 1.1384,
430
+ "step": 1540
431
+ },
432
+ {
433
+ "epoch": 2.71,
434
+ "grad_norm": 4.47056770324707,
435
+ "learning_rate": 5.502239283429303e-06,
436
+ "loss": 0.9374,
437
+ "step": 1568
438
+ },
439
+ {
440
+ "epoch": 2.76,
441
+ "grad_norm": 7.823177814483643,
442
+ "learning_rate": 4.606525911708254e-06,
443
+ "loss": 1.019,
444
+ "step": 1596
445
+ },
446
+ {
447
+ "epoch": 2.8,
448
+ "grad_norm": 10.679405212402344,
449
+ "learning_rate": 3.7108125399872046e-06,
450
+ "loss": 1.0615,
451
+ "step": 1624
452
+ },
453
+ {
454
+ "epoch": 2.85,
455
+ "grad_norm": 16.2053165435791,
456
+ "learning_rate": 2.815099168266155e-06,
457
+ "loss": 0.959,
458
+ "step": 1652
459
+ },
460
+ {
461
+ "epoch": 2.9,
462
+ "grad_norm": 8.753532409667969,
463
+ "learning_rate": 1.9193857965451054e-06,
464
+ "loss": 1.0041,
465
+ "step": 1680
466
+ },
467
+ {
468
+ "epoch": 2.95,
469
+ "grad_norm": 7.493074417114258,
470
+ "learning_rate": 1.0236724248240563e-06,
471
+ "loss": 0.998,
472
+ "step": 1708
473
+ },
474
+ {
475
+ "epoch": 3.0,
476
+ "grad_norm": 7.273109436035156,
477
+ "learning_rate": 1.2795905310300704e-07,
478
+ "loss": 0.9851,
479
+ "step": 1736
480
+ },
481
+ {
482
+ "epoch": 3.0,
483
+ "eval_accuracy": 0.5823986194995686,
484
+ "eval_f1_macro": 0.23526955685803672,
485
+ "eval_f1_micro": 0.5823986194995686,
486
+ "eval_f1_weighted": 0.5233223208450835,
487
+ "eval_loss": 0.9948348999023438,
488
+ "eval_precision_macro": 0.2973178657776674,
489
+ "eval_precision_micro": 0.5823986194995686,
490
+ "eval_precision_weighted": 0.5462756602202016,
491
+ "eval_recall_macro": 0.24224828041633692,
492
+ "eval_recall_micro": 0.5823986194995686,
493
+ "eval_recall_weighted": 0.5823986194995686,
494
+ "eval_runtime": 46.2253,
495
+ "eval_samples_per_second": 25.073,
496
+ "eval_steps_per_second": 1.579,
497
+ "step": 1737
498
+ }
499
+ ],
500
+ "logging_steps": 28,
501
+ "max_steps": 1737,
502
+ "num_input_tokens_seen": 0,
503
+ "num_train_epochs": 3,
504
+ "save_steps": 500,
505
+ "total_flos": 1.0764602068237517e+18,
506
+ "train_batch_size": 8,
507
+ "trial_name": null,
508
+ "trial_params": null
509
+ }
checkpoint-1737/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c681a71984113cdc446dadfd561c52f82fbe8752214b85f229516f962da6c9d4
3
+ size 4920
config.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "howdyaendra/autotrain-xblock-twitter-1",
3
+ "_num_labels": 12,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "altright",
14
+ "1": "bluesky",
15
+ "2": "facebook",
16
+ "3": "fediverse",
17
+ "4": "instagram",
18
+ "5": "reddit",
19
+ "6": "social-screenshot",
20
+ "7": "threads",
21
+ "8": "tumblr",
22
+ "9": "twitter",
23
+ "10": "uncategorised",
24
+ "11": "unrecognised-screenshot"
25
+ },
26
+ "image_size": 224,
27
+ "initializer_range": 0.02,
28
+ "intermediate_size": 3072,
29
+ "label2id": {
30
+ "altright": 0,
31
+ "bluesky": 1,
32
+ "facebook": 2,
33
+ "fediverse": 3,
34
+ "instagram": 4,
35
+ "reddit": 5,
36
+ "social-screenshot": 6,
37
+ "threads": 7,
38
+ "tumblr": 8,
39
+ "twitter": 9,
40
+ "uncategorised": 10,
41
+ "unrecognised-screenshot": 11
42
+ },
43
+ "layer_norm_eps": 1e-12,
44
+ "model_type": "vit",
45
+ "num_attention_heads": 12,
46
+ "num_channels": 3,
47
+ "num_hidden_layers": 12,
48
+ "patch_size": 16,
49
+ "problem_type": "single_label_classification",
50
+ "qkv_bias": true,
51
+ "torch_dtype": "float32",
52
+ "transformers_version": "4.39.0"
53
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2111f184b5b09d458eae2038e8e5e262954fd65eb0fe90c9d8b7c6477652a053
3
+ size 343254736
preprocessor_config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_rescale",
8
+ "rescale_factor",
9
+ "do_normalize",
10
+ "image_mean",
11
+ "image_std",
12
+ "return_tensors",
13
+ "data_format",
14
+ "input_data_format"
15
+ ],
16
+ "do_normalize": true,
17
+ "do_rescale": true,
18
+ "do_resize": true,
19
+ "image_mean": [
20
+ 0.5,
21
+ 0.5,
22
+ 0.5
23
+ ],
24
+ "image_processor_type": "ViTImageProcessor",
25
+ "image_std": [
26
+ 0.5,
27
+ 0.5,
28
+ 0.5
29
+ ],
30
+ "resample": 2,
31
+ "rescale_factor": 0.00392156862745098,
32
+ "size": {
33
+ "height": 224,
34
+ "width": 224
35
+ }
36
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c681a71984113cdc446dadfd561c52f82fbe8752214b85f229516f962da6c9d4
3
+ size 4920
training_params.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "xblock-social-screenshots-2/autotrain-data",
3
+ "model": "howdyaendra/autotrain-xblock-twitter-1",
4
+ "username": "howdyaendra",
5
+ "lr": 5e-05,
6
+ "epochs": 3,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "logging_steps": -1,
18
+ "project_name": "xblock-social-screenshots-2",
19
+ "auto_find_batch_size": false,
20
+ "mixed_precision": "fp16",
21
+ "save_total_limit": 1,
22
+ "save_strategy": "epoch",
23
+ "push_to_hub": true,
24
+ "repo_id": "howdyaendra/xblock-social-screenshots-2",
25
+ "evaluation_strategy": "epoch",
26
+ "image_column": "autotrain_image",
27
+ "target_column": "autotrain_label",
28
+ "log": "none"
29
+ }