hemakumari commited on
Commit
9093fbf
1 Parent(s): 7cf21af

Training in progress, epoch 0

Browse files
all_results.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 16.0,
3
+ "eval_accuracy": 0.6666666666666666,
4
+ "eval_loss": 1.2180020809173584,
5
+ "eval_runtime": 1.2421,
6
+ "eval_samples_per_second": 70.045,
7
+ "eval_steps_per_second": 2.415,
8
+ "total_flos": 9.609265079672832e+17,
9
+ "train_loss": 0.6730460596084594,
10
+ "train_runtime": 461.0529,
11
+ "train_samples_per_second": 27.763,
12
+ "train_steps_per_second": 0.217
13
+ }
config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "Alternaria",
13
+ "1": "Anthracnose",
14
+ "2": "Black Mould Rot",
15
+ "3": "Healthy",
16
+ "4": "Stem end Rot"
17
+ },
18
+ "image_size": 224,
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "Alternaria": 0,
23
+ "Anthracnose": 1,
24
+ "Black Mould Rot": 2,
25
+ "Healthy": 3,
26
+ "Stem end Rot": 4
27
+ },
28
+ "layer_norm_eps": 1e-12,
29
+ "model_type": "vit",
30
+ "num_attention_heads": 12,
31
+ "num_channels": 3,
32
+ "num_hidden_layers": 12,
33
+ "patch_size": 16,
34
+ "problem_type": "single_label_classification",
35
+ "qkv_bias": true,
36
+ "torch_dtype": "float32",
37
+ "transformers_version": "4.38.1"
38
+ }
eval_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 16.0,
3
+ "eval_accuracy": 0.6666666666666666,
4
+ "eval_loss": 1.2180020809173584,
5
+ "eval_runtime": 1.2421,
6
+ "eval_samples_per_second": 70.045,
7
+ "eval_steps_per_second": 2.415
8
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79bd521fab41f3a32346779f898fb75c9161e576960ff324f1819d69ab2c747b
3
+ size 343233204
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
runs/Mar13_07-04-33_be0be0d60671/events.out.tfevents.1710313504.be0be0d60671.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f403d39aa9efba36f51da08cf9e65d7c43fc15604b9a1022f8b8b63fc00b45d0
3
+ size 12313
runs/Mar13_07-04-33_be0be0d60671/events.out.tfevents.1710313986.be0be0d60671.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bc876c20e6297473dd243e21c288ae479ded6425957c4f9b96abf468d015f7a
3
+ size 405
runs/Mar13_07-17-00_be0be0d60671/events.out.tfevents.1710314237.be0be0d60671.34.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4df81dad683e6ad2b2bb7da055e58a496b4a728e164727e93af5db2e535f2b3
3
+ size 12407
runs/Mar13_07-17-00_be0be0d60671/events.out.tfevents.1710314711.be0be0d60671.34.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a2330aec3828dbc69fc483ce63d36740898afcc757ac333766650801855dc5e
3
+ size 405
runs/Mar13_07-27-03_be0be0d60671/events.out.tfevents.1710314839.be0be0d60671.34.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fa40ef59184e57c553039f2c504de681d38d21f919225ed571d73bd088ef7a9
3
+ size 5234
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 16.0,
3
+ "total_flos": 9.609265079672832e+17,
4
+ "train_loss": 0.6730460596084594,
5
+ "train_runtime": 461.0529,
6
+ "train_samples_per_second": 27.763,
7
+ "train_steps_per_second": 0.217
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,244 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 1.2180020809173584,
3
+ "best_model_checkpoint": "vit-base-patch16-224-in21k-finetunedmangodisease/checkpoint-87",
4
+ "epoch": 16.0,
5
+ "eval_steps": 500,
6
+ "global_step": 100,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.96,
13
+ "eval_accuracy": 0.6551724137931034,
14
+ "eval_loss": 1.3800822496414185,
15
+ "eval_runtime": 0.8949,
16
+ "eval_samples_per_second": 97.223,
17
+ "eval_steps_per_second": 3.353,
18
+ "step": 6
19
+ },
20
+ {
21
+ "epoch": 1.6,
22
+ "grad_norm": 1.6758902072906494,
23
+ "learning_rate": 5e-05,
24
+ "loss": 0.8237,
25
+ "step": 10
26
+ },
27
+ {
28
+ "epoch": 1.92,
29
+ "eval_accuracy": 0.6781609195402298,
30
+ "eval_loss": 1.349981665611267,
31
+ "eval_runtime": 0.8658,
32
+ "eval_samples_per_second": 100.488,
33
+ "eval_steps_per_second": 3.465,
34
+ "step": 12
35
+ },
36
+ {
37
+ "epoch": 2.88,
38
+ "eval_accuracy": 0.6896551724137931,
39
+ "eval_loss": 1.3429534435272217,
40
+ "eval_runtime": 0.8664,
41
+ "eval_samples_per_second": 100.415,
42
+ "eval_steps_per_second": 3.463,
43
+ "step": 18
44
+ },
45
+ {
46
+ "epoch": 3.2,
47
+ "grad_norm": 1.6055872440338135,
48
+ "learning_rate": 4.4444444444444447e-05,
49
+ "loss": 0.7777,
50
+ "step": 20
51
+ },
52
+ {
53
+ "epoch": 4.0,
54
+ "eval_accuracy": 0.6781609195402298,
55
+ "eval_loss": 1.3140147924423218,
56
+ "eval_runtime": 1.2457,
57
+ "eval_samples_per_second": 69.841,
58
+ "eval_steps_per_second": 2.408,
59
+ "step": 25
60
+ },
61
+ {
62
+ "epoch": 4.8,
63
+ "grad_norm": 1.475545048713684,
64
+ "learning_rate": 3.888888888888889e-05,
65
+ "loss": 0.7066,
66
+ "step": 30
67
+ },
68
+ {
69
+ "epoch": 4.96,
70
+ "eval_accuracy": 0.6896551724137931,
71
+ "eval_loss": 1.2853386402130127,
72
+ "eval_runtime": 0.8536,
73
+ "eval_samples_per_second": 101.919,
74
+ "eval_steps_per_second": 3.514,
75
+ "step": 31
76
+ },
77
+ {
78
+ "epoch": 5.92,
79
+ "eval_accuracy": 0.6781609195402298,
80
+ "eval_loss": 1.2957433462142944,
81
+ "eval_runtime": 0.8698,
82
+ "eval_samples_per_second": 100.02,
83
+ "eval_steps_per_second": 3.449,
84
+ "step": 37
85
+ },
86
+ {
87
+ "epoch": 6.4,
88
+ "grad_norm": 1.3556888103485107,
89
+ "learning_rate": 3.3333333333333335e-05,
90
+ "loss": 0.6696,
91
+ "step": 40
92
+ },
93
+ {
94
+ "epoch": 6.88,
95
+ "eval_accuracy": 0.6551724137931034,
96
+ "eval_loss": 1.2818764448165894,
97
+ "eval_runtime": 0.8627,
98
+ "eval_samples_per_second": 100.841,
99
+ "eval_steps_per_second": 3.477,
100
+ "step": 43
101
+ },
102
+ {
103
+ "epoch": 8.0,
104
+ "grad_norm": 4.025577545166016,
105
+ "learning_rate": 2.777777777777778e-05,
106
+ "loss": 0.673,
107
+ "step": 50
108
+ },
109
+ {
110
+ "epoch": 8.0,
111
+ "eval_accuracy": 0.6781609195402298,
112
+ "eval_loss": 1.285282850265503,
113
+ "eval_runtime": 1.3346,
114
+ "eval_samples_per_second": 65.186,
115
+ "eval_steps_per_second": 2.248,
116
+ "step": 50
117
+ },
118
+ {
119
+ "epoch": 8.96,
120
+ "eval_accuracy": 0.6551724137931034,
121
+ "eval_loss": 1.2612943649291992,
122
+ "eval_runtime": 1.3828,
123
+ "eval_samples_per_second": 62.915,
124
+ "eval_steps_per_second": 2.169,
125
+ "step": 56
126
+ },
127
+ {
128
+ "epoch": 9.6,
129
+ "grad_norm": 1.7103294134140015,
130
+ "learning_rate": 2.2222222222222223e-05,
131
+ "loss": 0.6311,
132
+ "step": 60
133
+ },
134
+ {
135
+ "epoch": 9.92,
136
+ "eval_accuracy": 0.6436781609195402,
137
+ "eval_loss": 1.2490617036819458,
138
+ "eval_runtime": 1.1569,
139
+ "eval_samples_per_second": 75.2,
140
+ "eval_steps_per_second": 2.593,
141
+ "step": 62
142
+ },
143
+ {
144
+ "epoch": 10.88,
145
+ "eval_accuracy": 0.6666666666666666,
146
+ "eval_loss": 1.2408325672149658,
147
+ "eval_runtime": 1.7627,
148
+ "eval_samples_per_second": 49.356,
149
+ "eval_steps_per_second": 1.702,
150
+ "step": 68
151
+ },
152
+ {
153
+ "epoch": 11.2,
154
+ "grad_norm": 1.8741282224655151,
155
+ "learning_rate": 1.6666666666666667e-05,
156
+ "loss": 0.6086,
157
+ "step": 70
158
+ },
159
+ {
160
+ "epoch": 12.0,
161
+ "eval_accuracy": 0.632183908045977,
162
+ "eval_loss": 1.2268433570861816,
163
+ "eval_runtime": 1.4569,
164
+ "eval_samples_per_second": 59.717,
165
+ "eval_steps_per_second": 2.059,
166
+ "step": 75
167
+ },
168
+ {
169
+ "epoch": 12.8,
170
+ "grad_norm": 2.322943925857544,
171
+ "learning_rate": 1.1111111111111112e-05,
172
+ "loss": 0.6374,
173
+ "step": 80
174
+ },
175
+ {
176
+ "epoch": 12.96,
177
+ "eval_accuracy": 0.632183908045977,
178
+ "eval_loss": 1.2218347787857056,
179
+ "eval_runtime": 0.8909,
180
+ "eval_samples_per_second": 97.657,
181
+ "eval_steps_per_second": 3.367,
182
+ "step": 81
183
+ },
184
+ {
185
+ "epoch": 13.92,
186
+ "eval_accuracy": 0.6666666666666666,
187
+ "eval_loss": 1.2180020809173584,
188
+ "eval_runtime": 0.8761,
189
+ "eval_samples_per_second": 99.299,
190
+ "eval_steps_per_second": 3.424,
191
+ "step": 87
192
+ },
193
+ {
194
+ "epoch": 14.4,
195
+ "grad_norm": 1.7907931804656982,
196
+ "learning_rate": 5.555555555555556e-06,
197
+ "loss": 0.599,
198
+ "step": 90
199
+ },
200
+ {
201
+ "epoch": 14.88,
202
+ "eval_accuracy": 0.6666666666666666,
203
+ "eval_loss": 1.2180161476135254,
204
+ "eval_runtime": 1.267,
205
+ "eval_samples_per_second": 68.668,
206
+ "eval_steps_per_second": 2.368,
207
+ "step": 93
208
+ },
209
+ {
210
+ "epoch": 16.0,
211
+ "grad_norm": 3.0324525833129883,
212
+ "learning_rate": 0.0,
213
+ "loss": 0.6037,
214
+ "step": 100
215
+ },
216
+ {
217
+ "epoch": 16.0,
218
+ "eval_accuracy": 0.6666666666666666,
219
+ "eval_loss": 1.219612717628479,
220
+ "eval_runtime": 0.8638,
221
+ "eval_samples_per_second": 100.717,
222
+ "eval_steps_per_second": 3.473,
223
+ "step": 100
224
+ },
225
+ {
226
+ "epoch": 16.0,
227
+ "step": 100,
228
+ "total_flos": 9.609265079672832e+17,
229
+ "train_loss": 0.6730460596084594,
230
+ "train_runtime": 461.0529,
231
+ "train_samples_per_second": 27.763,
232
+ "train_steps_per_second": 0.217
233
+ }
234
+ ],
235
+ "logging_steps": 10,
236
+ "max_steps": 100,
237
+ "num_input_tokens_seen": 0,
238
+ "num_train_epochs": 17,
239
+ "save_steps": 500,
240
+ "total_flos": 9.609265079672832e+17,
241
+ "train_batch_size": 32,
242
+ "trial_name": null,
243
+ "trial_params": null
244
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fcec20e8d1d9a83e6122676b0bb12b093be8b940a355d6ffc12e6ec4ef3fac6
3
+ size 4984