hiendang7613 commited on
Commit
0ac7c42
1 Parent(s): 54e1274

Training in progress, epoch 1

Browse files
Files changed (5) hide show
  1. README.md +48 -0
  2. config.json +428 -0
  3. model.safetensors +3 -0
  4. preprocessor_config.json +22 -0
  5. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: google/vit-large-patch16-224
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: vit-l-tiny-imagenet
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # vit-l-tiny-imagenet
15
+
16
+ This model is a fine-tuned version of [google/vit-large-patch16-224](https://huggingface.co/google/vit-large-patch16-224) on an unknown dataset.
17
+
18
+ ## Model description
19
+
20
+ More information needed
21
+
22
+ ## Intended uses & limitations
23
+
24
+ More information needed
25
+
26
+ ## Training and evaluation data
27
+
28
+ More information needed
29
+
30
+ ## Training procedure
31
+
32
+ ### Training hyperparameters
33
+
34
+ The following hyperparameters were used during training:
35
+ - learning_rate: 5e-05
36
+ - train_batch_size: 32
37
+ - eval_batch_size: 32
38
+ - seed: 42
39
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
+ - lr_scheduler_type: linear
41
+ - num_epochs: 100
42
+
43
+ ### Framework versions
44
+
45
+ - Transformers 4.35.2
46
+ - Pytorch 2.1.0+cu121
47
+ - Datasets 2.17.0
48
+ - Tokenizers 0.15.2
config.json ADDED
@@ -0,0 +1,428 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-large-patch16-224",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 1024,
11
+ "id2label": {
12
+ "0": "n01443537",
13
+ "1": "n01629819",
14
+ "2": "n01641577",
15
+ "3": "n01644900",
16
+ "4": "n01698640",
17
+ "5": "n01742172",
18
+ "6": "n01768244",
19
+ "7": "n01770393",
20
+ "8": "n01774384",
21
+ "9": "n01774750",
22
+ "10": "n01784675",
23
+ "11": "n01882714",
24
+ "12": "n01910747",
25
+ "13": "n01917289",
26
+ "14": "n01944390",
27
+ "15": "n01950731",
28
+ "16": "n01983481",
29
+ "17": "n01984695",
30
+ "18": "n02002724",
31
+ "19": "n02056570",
32
+ "20": "n02058221",
33
+ "21": "n02074367",
34
+ "22": "n02094433",
35
+ "23": "n02099601",
36
+ "24": "n02099712",
37
+ "25": "n02106662",
38
+ "26": "n02113799",
39
+ "27": "n02123045",
40
+ "28": "n02123394",
41
+ "29": "n02124075",
42
+ "30": "n02125311",
43
+ "31": "n02129165",
44
+ "32": "n02132136",
45
+ "33": "n02165456",
46
+ "34": "n02226429",
47
+ "35": "n02231487",
48
+ "36": "n02233338",
49
+ "37": "n02236044",
50
+ "38": "n02268443",
51
+ "39": "n02279972",
52
+ "40": "n02281406",
53
+ "41": "n02321529",
54
+ "42": "n02364673",
55
+ "43": "n02395406",
56
+ "44": "n02403003",
57
+ "45": "n02410509",
58
+ "46": "n02415577",
59
+ "47": "n02423022",
60
+ "48": "n02437312",
61
+ "49": "n02480495",
62
+ "50": "n02481823",
63
+ "51": "n02486410",
64
+ "52": "n02504458",
65
+ "53": "n02509815",
66
+ "54": "n02666347",
67
+ "55": "n02669723",
68
+ "56": "n02699494",
69
+ "57": "n02769748",
70
+ "58": "n02788148",
71
+ "59": "n02791270",
72
+ "60": "n02793495",
73
+ "61": "n02795169",
74
+ "62": "n02802426",
75
+ "63": "n02808440",
76
+ "64": "n02814533",
77
+ "65": "n02814860",
78
+ "66": "n02815834",
79
+ "67": "n02823428",
80
+ "68": "n02837789",
81
+ "69": "n02841315",
82
+ "70": "n02843684",
83
+ "71": "n02883205",
84
+ "72": "n02892201",
85
+ "73": "n02909870",
86
+ "74": "n02917067",
87
+ "75": "n02927161",
88
+ "76": "n02948072",
89
+ "77": "n02950826",
90
+ "78": "n02963159",
91
+ "79": "n02977058",
92
+ "80": "n02988304",
93
+ "81": "n03014705",
94
+ "82": "n03026506",
95
+ "83": "n03042490",
96
+ "84": "n03085013",
97
+ "85": "n03089624",
98
+ "86": "n03100240",
99
+ "87": "n03126707",
100
+ "88": "n03160309",
101
+ "89": "n03179701",
102
+ "90": "n03201208",
103
+ "91": "n03255030",
104
+ "92": "n03355925",
105
+ "93": "n03373237",
106
+ "94": "n03388043",
107
+ "95": "n03393912",
108
+ "96": "n03400231",
109
+ "97": "n03404251",
110
+ "98": "n03424325",
111
+ "99": "n03444034",
112
+ "100": "n03447447",
113
+ "101": "n03544143",
114
+ "102": "n03584254",
115
+ "103": "n03599486",
116
+ "104": "n03617480",
117
+ "105": "n03637318",
118
+ "106": "n03649909",
119
+ "107": "n03662601",
120
+ "108": "n03670208",
121
+ "109": "n03706229",
122
+ "110": "n03733131",
123
+ "111": "n03763968",
124
+ "112": "n03770439",
125
+ "113": "n03796401",
126
+ "114": "n03814639",
127
+ "115": "n03837869",
128
+ "116": "n03838899",
129
+ "117": "n03854065",
130
+ "118": "n03891332",
131
+ "119": "n03902125",
132
+ "120": "n03930313",
133
+ "121": "n03937543",
134
+ "122": "n03970156",
135
+ "123": "n03977966",
136
+ "124": "n03980874",
137
+ "125": "n03983396",
138
+ "126": "n03992509",
139
+ "127": "n04008634",
140
+ "128": "n04023962",
141
+ "129": "n04070727",
142
+ "130": "n04074963",
143
+ "131": "n04099969",
144
+ "132": "n04118538",
145
+ "133": "n04133789",
146
+ "134": "n04146614",
147
+ "135": "n04149813",
148
+ "136": "n04179913",
149
+ "137": "n04251144",
150
+ "138": "n04254777",
151
+ "139": "n04259630",
152
+ "140": "n04265275",
153
+ "141": "n04275548",
154
+ "142": "n04285008",
155
+ "143": "n04311004",
156
+ "144": "n04328186",
157
+ "145": "n04356056",
158
+ "146": "n04366367",
159
+ "147": "n04371430",
160
+ "148": "n04376876",
161
+ "149": "n04398044",
162
+ "150": "n04399382",
163
+ "151": "n04417672",
164
+ "152": "n04456115",
165
+ "153": "n04465666",
166
+ "154": "n04486054",
167
+ "155": "n04487081",
168
+ "156": "n04501370",
169
+ "157": "n04507155",
170
+ "158": "n04532106",
171
+ "159": "n04532670",
172
+ "160": "n04540053",
173
+ "161": "n04560804",
174
+ "162": "n04562935",
175
+ "163": "n04596742",
176
+ "164": "n04598010",
177
+ "165": "n06596364",
178
+ "166": "n07056680",
179
+ "167": "n07583066",
180
+ "168": "n07614500",
181
+ "169": "n07615774",
182
+ "170": "n07646821",
183
+ "171": "n07647870",
184
+ "172": "n07657664",
185
+ "173": "n07695742",
186
+ "174": "n07711569",
187
+ "175": "n07715103",
188
+ "176": "n07720875",
189
+ "177": "n07749582",
190
+ "178": "n07753592",
191
+ "179": "n07768694",
192
+ "180": "n07871810",
193
+ "181": "n07873807",
194
+ "182": "n07875152",
195
+ "183": "n07920052",
196
+ "184": "n07975909",
197
+ "185": "n08496334",
198
+ "186": "n08620881",
199
+ "187": "n08742578",
200
+ "188": "n09193705",
201
+ "189": "n09246464",
202
+ "190": "n09256479",
203
+ "191": "n09332890",
204
+ "192": "n09428293",
205
+ "193": "n12267677",
206
+ "194": "n12520864",
207
+ "195": "n13001041",
208
+ "196": "n13652335",
209
+ "197": "n13652994",
210
+ "198": "n13719102",
211
+ "199": "n14991210"
212
+ },
213
+ "image_size": 224,
214
+ "initializer_range": 0.02,
215
+ "intermediate_size": 4096,
216
+ "label2id": {
217
+ "n01443537": 0,
218
+ "n01629819": 1,
219
+ "n01641577": 2,
220
+ "n01644900": 3,
221
+ "n01698640": 4,
222
+ "n01742172": 5,
223
+ "n01768244": 6,
224
+ "n01770393": 7,
225
+ "n01774384": 8,
226
+ "n01774750": 9,
227
+ "n01784675": 10,
228
+ "n01882714": 11,
229
+ "n01910747": 12,
230
+ "n01917289": 13,
231
+ "n01944390": 14,
232
+ "n01950731": 15,
233
+ "n01983481": 16,
234
+ "n01984695": 17,
235
+ "n02002724": 18,
236
+ "n02056570": 19,
237
+ "n02058221": 20,
238
+ "n02074367": 21,
239
+ "n02094433": 22,
240
+ "n02099601": 23,
241
+ "n02099712": 24,
242
+ "n02106662": 25,
243
+ "n02113799": 26,
244
+ "n02123045": 27,
245
+ "n02123394": 28,
246
+ "n02124075": 29,
247
+ "n02125311": 30,
248
+ "n02129165": 31,
249
+ "n02132136": 32,
250
+ "n02165456": 33,
251
+ "n02226429": 34,
252
+ "n02231487": 35,
253
+ "n02233338": 36,
254
+ "n02236044": 37,
255
+ "n02268443": 38,
256
+ "n02279972": 39,
257
+ "n02281406": 40,
258
+ "n02321529": 41,
259
+ "n02364673": 42,
260
+ "n02395406": 43,
261
+ "n02403003": 44,
262
+ "n02410509": 45,
263
+ "n02415577": 46,
264
+ "n02423022": 47,
265
+ "n02437312": 48,
266
+ "n02480495": 49,
267
+ "n02481823": 50,
268
+ "n02486410": 51,
269
+ "n02504458": 52,
270
+ "n02509815": 53,
271
+ "n02666347": 54,
272
+ "n02669723": 55,
273
+ "n02699494": 56,
274
+ "n02769748": 57,
275
+ "n02788148": 58,
276
+ "n02791270": 59,
277
+ "n02793495": 60,
278
+ "n02795169": 61,
279
+ "n02802426": 62,
280
+ "n02808440": 63,
281
+ "n02814533": 64,
282
+ "n02814860": 65,
283
+ "n02815834": 66,
284
+ "n02823428": 67,
285
+ "n02837789": 68,
286
+ "n02841315": 69,
287
+ "n02843684": 70,
288
+ "n02883205": 71,
289
+ "n02892201": 72,
290
+ "n02909870": 73,
291
+ "n02917067": 74,
292
+ "n02927161": 75,
293
+ "n02948072": 76,
294
+ "n02950826": 77,
295
+ "n02963159": 78,
296
+ "n02977058": 79,
297
+ "n02988304": 80,
298
+ "n03014705": 81,
299
+ "n03026506": 82,
300
+ "n03042490": 83,
301
+ "n03085013": 84,
302
+ "n03089624": 85,
303
+ "n03100240": 86,
304
+ "n03126707": 87,
305
+ "n03160309": 88,
306
+ "n03179701": 89,
307
+ "n03201208": 90,
308
+ "n03255030": 91,
309
+ "n03355925": 92,
310
+ "n03373237": 93,
311
+ "n03388043": 94,
312
+ "n03393912": 95,
313
+ "n03400231": 96,
314
+ "n03404251": 97,
315
+ "n03424325": 98,
316
+ "n03444034": 99,
317
+ "n03447447": 100,
318
+ "n03544143": 101,
319
+ "n03584254": 102,
320
+ "n03599486": 103,
321
+ "n03617480": 104,
322
+ "n03637318": 105,
323
+ "n03649909": 106,
324
+ "n03662601": 107,
325
+ "n03670208": 108,
326
+ "n03706229": 109,
327
+ "n03733131": 110,
328
+ "n03763968": 111,
329
+ "n03770439": 112,
330
+ "n03796401": 113,
331
+ "n03814639": 114,
332
+ "n03837869": 115,
333
+ "n03838899": 116,
334
+ "n03854065": 117,
335
+ "n03891332": 118,
336
+ "n03902125": 119,
337
+ "n03930313": 120,
338
+ "n03937543": 121,
339
+ "n03970156": 122,
340
+ "n03977966": 123,
341
+ "n03980874": 124,
342
+ "n03983396": 125,
343
+ "n03992509": 126,
344
+ "n04008634": 127,
345
+ "n04023962": 128,
346
+ "n04070727": 129,
347
+ "n04074963": 130,
348
+ "n04099969": 131,
349
+ "n04118538": 132,
350
+ "n04133789": 133,
351
+ "n04146614": 134,
352
+ "n04149813": 135,
353
+ "n04179913": 136,
354
+ "n04251144": 137,
355
+ "n04254777": 138,
356
+ "n04259630": 139,
357
+ "n04265275": 140,
358
+ "n04275548": 141,
359
+ "n04285008": 142,
360
+ "n04311004": 143,
361
+ "n04328186": 144,
362
+ "n04356056": 145,
363
+ "n04366367": 146,
364
+ "n04371430": 147,
365
+ "n04376876": 148,
366
+ "n04398044": 149,
367
+ "n04399382": 150,
368
+ "n04417672": 151,
369
+ "n04456115": 152,
370
+ "n04465666": 153,
371
+ "n04486054": 154,
372
+ "n04487081": 155,
373
+ "n04501370": 156,
374
+ "n04507155": 157,
375
+ "n04532106": 158,
376
+ "n04532670": 159,
377
+ "n04540053": 160,
378
+ "n04560804": 161,
379
+ "n04562935": 162,
380
+ "n04596742": 163,
381
+ "n04598010": 164,
382
+ "n06596364": 165,
383
+ "n07056680": 166,
384
+ "n07583066": 167,
385
+ "n07614500": 168,
386
+ "n07615774": 169,
387
+ "n07646821": 170,
388
+ "n07647870": 171,
389
+ "n07657664": 172,
390
+ "n07695742": 173,
391
+ "n07711569": 174,
392
+ "n07715103": 175,
393
+ "n07720875": 176,
394
+ "n07749582": 177,
395
+ "n07753592": 178,
396
+ "n07768694": 179,
397
+ "n07871810": 180,
398
+ "n07873807": 181,
399
+ "n07875152": 182,
400
+ "n07920052": 183,
401
+ "n07975909": 184,
402
+ "n08496334": 185,
403
+ "n08620881": 186,
404
+ "n08742578": 187,
405
+ "n09193705": 188,
406
+ "n09246464": 189,
407
+ "n09256479": 190,
408
+ "n09332890": 191,
409
+ "n09428293": 192,
410
+ "n12267677": 193,
411
+ "n12520864": 194,
412
+ "n13001041": 195,
413
+ "n13652335": 196,
414
+ "n13652994": 197,
415
+ "n13719102": 198,
416
+ "n14991210": 199
417
+ },
418
+ "layer_norm_eps": 1e-12,
419
+ "model_type": "vit",
420
+ "num_attention_heads": 16,
421
+ "num_channels": 3,
422
+ "num_hidden_layers": 24,
423
+ "patch_size": 16,
424
+ "problem_type": "single_label_classification",
425
+ "qkv_bias": true,
426
+ "torch_dtype": "float32",
427
+ "transformers_version": "4.35.2"
428
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6f745f8d01e36996ecd4c43298eca9768a07080d742a9a03533223ff35ace3d
3
+ size 1214073080
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77ceb35871d46766b39f770b5365c238efb375b34d00185089379f315750d245
3
+ size 4536