wzk1015 commited on
Commit
3734857
1 Parent(s): 04c1727

cls model,log,config

Browse files
classification/piip_3branch_sbl_320-160-96_cls_token_augreg.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0de2d377bc0bc752958bba4db0fbbea6f999859ec0536bec10e5d37e02a01ed3
3
+ size 1955899841
classification/piip_3branch_sbl_320-160-96_cls_token_augreg.py ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model = dict(
2
+ backbone=dict(
3
+ n_points=4,
4
+ deform_num_heads=16,
5
+ cffn_ratio=0.25,
6
+ deform_ratio=0.5,
7
+ with_cffn=True,
8
+ interact_attn_type='deform',
9
+ interaction_drop_path_rate=0.4,
10
+ separate_head=True,
11
+
12
+ branch1=dict(
13
+ model_type="augreg",
14
+ img_size=96,
15
+ patch_size=16,
16
+ pretrain_img_size=224,
17
+ pretrain_patch_size=16,
18
+ depth=24,
19
+ embed_dim=1024,
20
+ num_heads=16,
21
+ mlp_ratio=4,
22
+ qkv_bias=True,
23
+ drop_path_rate=0.4,
24
+ interaction_indexes=[[0, 1], [2, 3], [4, 5], [6, 7], [8, 9], [10, 11], [12, 13], [14, 15], [16, 17], [18, 19], [20, 21], [22, 23]],
25
+ use_cls_token=True,
26
+ use_flash_attn=True,
27
+ with_cp=True,
28
+ pretrained="pretrained/L_16-i21k-300ep-lr_0.001-aug_medium1-wd_0.1-do_0.1-sd_0.1--imagenet2012-steps_20k-lr_0.01-res_224.pth",
29
+ ),
30
+
31
+ branch2=dict(
32
+ model_type="augreg",
33
+ img_size=160,
34
+ patch_size=16,
35
+ pretrain_img_size=224,
36
+ pretrain_patch_size=16,
37
+ depth=12,
38
+ embed_dim=768,
39
+ num_heads=12,
40
+ mlp_ratio=4,
41
+ qkv_bias=True,
42
+ drop_path_rate=0.2,
43
+ interaction_indexes=[[0, 0], [1, 1], [2, 2], [3, 3], [4, 4], [5, 5], [6, 6], [7, 7], [8, 8], [9, 9], [10, 10], [11, 11]],
44
+ use_cls_token=True,
45
+ use_flash_attn=True,
46
+ with_cp=True,
47
+ pretrained="pretrained/B_16-i21k-300ep-lr_0.001-aug_medium1-wd_0.1-do_0.0-sd_0.0--imagenet2012-steps_20k-lr_0.01-res_224.pth",
48
+ ),
49
+
50
+ branch3=dict(
51
+ model_type="augreg",
52
+ img_size=320,
53
+ patch_size=16,
54
+ pretrain_img_size=224,
55
+ pretrain_patch_size=16,
56
+ depth=12,
57
+ embed_dim=384,
58
+ num_heads=6,
59
+ mlp_ratio=4,
60
+ qkv_bias=True,
61
+ drop_path_rate=0.05,
62
+ interaction_indexes=[[0, 0], [1, 1], [2, 2], [3, 3], [4, 4], [5, 5], [6, 6], [7, 7], [8, 8], [9, 9], [10, 10], [11, 11]],
63
+ use_cls_token=True,
64
+ use_flash_attn=True,
65
+ with_cp=True,
66
+ pretrained="pretrained/S_16-i21k-300ep-lr_0.001-aug_light1-wd_0.03-do_0.0-sd_0.0--imagenet2012-steps_20k-lr_0.03-res_224.pth",
67
+ ),
68
+ ),
69
+ )
classification/piip_3branch_sbl_320-160-96_cls_token_augreg.txt ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"train_lr": 9.999999999999953e-07, "train_loss": 5.362081902085258, "test_loss": 0.8934490169145138, "test_acc1": 78.29200251953125, "test_acc5": 94.83200245117187, "epoch": 0, "n_parameters": "488.8M", "max_acc": 78.29200251953125}
2
+ {"train_lr": 9.999999999999953e-07, "train_loss": 4.663460858124528, "test_loss": 0.7776578705886315, "test_acc1": 79.9260025, "test_acc5": 95.64800221679687, "epoch": 1, "n_parameters": "488.8M", "max_acc": 79.9260025}
3
+ {"train_lr": 6.800000000000116e-06, "train_loss": 3.929186521126307, "test_loss": 0.7632588433579923, "test_acc1": 82.33600251953125, "test_acc5": 96.67600243164063, "epoch": 2, "n_parameters": "488.8M", "max_acc": 82.33600251953125}
4
+ {"train_lr": 1.2599999999999859e-05, "train_loss": 3.505233769139035, "test_loss": 0.7809260524095704, "test_acc1": 83.09000236328124, "test_acc5": 96.94800258789063, "epoch": 3, "n_parameters": "488.8M", "max_acc": 83.09000236328124}
5
+ {"train_lr": 1.8399999999999458e-05, "train_loss": 3.3293923243082206, "test_loss": 0.7268165690232053, "test_acc1": 83.4500025390625, "test_acc5": 97.05400266601562, "epoch": 4, "n_parameters": "488.8M", "max_acc": 83.4500025390625}
6
+ {"train_lr": 2.419999999999938e-05, "train_loss": 3.214345143388787, "test_loss": 0.7509733831288714, "test_acc1": 83.59200241210938, "test_acc5": 97.05800259765626, "epoch": 5, "n_parameters": "488.8M", "max_acc": 83.59200241210938}
7
+ {"train_lr": 2.7071067811864602e-05, "train_loss": 3.1421558767390385, "test_loss": 0.7573428535370078, "test_acc1": 83.67600229492187, "test_acc5": 97.13000262695313, "epoch": 6, "n_parameters": "488.8M", "max_acc": 83.67600229492187}
8
+ {"train_lr": 2.4539904997396135e-05, "train_loss": 3.080557497070848, "test_loss": 0.7054682780499659, "test_acc1": 83.91800288085938, "test_acc5": 97.26600267578125, "epoch": 7, "n_parameters": "488.8M", "max_acc": 83.91800288085938}
9
+ {"train_lr": 2.4539904997396135e-05, "train_loss": 3.0397233152084593, "test_loss": 0.7037731233227755, "test_acc1": 84.10200264648438, "test_acc5": 97.32400249023438, "epoch": 8, "n_parameters": "488.8M", "max_acc": 84.10200264648438}
10
+ {"train_lr": 2.309016994374978e-05, "train_loss": 2.996424310498958, "test_loss": 0.7143044033269773, "test_acc1": 84.24600258789063, "test_acc5": 97.32400254882812, "epoch": 9, "n_parameters": "488.8M", "max_acc": 84.24600258789063}
11
+ {"train_lr": 2.1564344650402722e-05, "train_loss": 2.972179876230985, "test_loss": 0.6699192833398037, "test_acc1": 84.36200245117188, "test_acc5": 97.3720026953125, "epoch": 10, "n_parameters": "488.8M", "max_acc": 84.36200245117188}
12
+ {"train_lr": 1.9999999999999453e-05, "train_loss": 2.9545375070721507, "test_loss": 0.6735954951509205, "test_acc1": 84.46400262695313, "test_acc5": 97.42000234375, "epoch": 11, "n_parameters": "488.8M", "max_acc": 84.46400262695313}
13
+ {"train_lr": 1.843565534959796e-05, "train_loss": 2.9185172934755146, "test_loss": 0.6666356289752142, "test_acc1": 84.5220025, "test_acc5": 97.44400249023437, "epoch": 12, "n_parameters": "488.8M", "max_acc": 84.5220025}
14
+ {"train_lr": 1.6909830056250673e-05, "train_loss": 2.8974646770601553, "test_loss": 0.6738747470223584, "test_acc1": 84.70200258789062, "test_acc5": 97.53600235351563, "epoch": 13, "n_parameters": "488.8M", "max_acc": 84.70200258789062}
15
+ {"train_lr": 1.5460095002604783e-05, "train_loss": 2.8739540015193197, "test_loss": 0.649927597963947, "test_acc1": 84.864002734375, "test_acc5": 97.5240025, "epoch": 14, "n_parameters": "488.8M", "max_acc": 84.864002734375}
16
+ {"train_lr": 1.412214747707529e-05, "train_loss": 2.8637630867181447, "test_loss": 0.6615397283857353, "test_acc1": 84.94600250976562, "test_acc5": 97.52800249023437, "epoch": 15, "n_parameters": "488.8M", "max_acc": 84.94600250976562}
17
+ {"train_lr": 1.2928932188134564e-05, "train_loss": 2.853053854321309, "test_loss": 0.6572925561460955, "test_acc1": 84.94600265625, "test_acc5": 97.50800265625, "epoch": 16, "n_parameters": "488.8M", "max_acc": 84.94600265625}
18
+ {"train_lr": 1.1909830056250373e-05, "train_loss": 2.837055222176724, "test_loss": 0.6545899954563813, "test_acc1": 85.0060025390625, "test_acc5": 97.56800262695313, "epoch": 17, "n_parameters": "488.8M", "max_acc": 85.0060025390625}
19
+ {"train_lr": 1.1089934758116523e-05, "train_loss": 2.8201447446211922, "test_loss": 0.6497874563681212, "test_acc1": 85.01200245117188, "test_acc5": 97.61800240234375, "epoch": 18, "n_parameters": "488.8M", "max_acc": 85.01200245117188}
20
+ {"train_lr": 1.0489434837048593e-05, "train_loss": 2.812104366237311, "test_loss": 0.6435268987184284, "test_acc1": 85.2440026171875, "test_acc5": 97.5940024609375, "epoch": 19, "n_parameters": "488.8M", "max_acc": 85.2440026171875}
classification/piip_3branch_sbl_384-192-128_cls_token_augreg.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:985c232f0aa4dcf4a597541d03c479f5445bd935b89168ac791ed110e5969f68
3
+ size 1956420033
classification/piip_3branch_sbl_384-192-128_cls_token_augreg.py ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model = dict(
2
+ backbone=dict(
3
+ n_points=4,
4
+ deform_num_heads=16,
5
+ cffn_ratio=0.25,
6
+ deform_ratio=0.5,
7
+ with_cffn=True,
8
+ interact_attn_type='deform',
9
+ interaction_drop_path_rate=0.4,
10
+ separate_head=True,
11
+
12
+ branch1=dict(
13
+ model_type="augreg",
14
+ img_size=128,
15
+ patch_size=16,
16
+ pretrain_img_size=224,
17
+ pretrain_patch_size=16,
18
+ depth=24,
19
+ embed_dim=1024,
20
+ num_heads=16,
21
+ mlp_ratio=4,
22
+ qkv_bias=True,
23
+ drop_path_rate=0.4,
24
+ interaction_indexes=[[0, 1], [2, 3], [4, 5], [6, 7], [8, 9], [10, 11], [12, 13], [14, 15], [16, 17], [18, 19], [20, 21], [22, 23]],
25
+ use_cls_token=True,
26
+ use_flash_attn=True,
27
+ with_cp=True,
28
+ pretrained="pretrained/L_16-i21k-300ep-lr_0.001-aug_medium1-wd_0.1-do_0.1-sd_0.1--imagenet2012-steps_20k-lr_0.01-res_224.pth",
29
+ ),
30
+
31
+ branch2=dict(
32
+ model_type="augreg",
33
+ img_size=192,
34
+ patch_size=16,
35
+ pretrain_img_size=224,
36
+ pretrain_patch_size=16,
37
+ depth=12,
38
+ embed_dim=768,
39
+ num_heads=12,
40
+ mlp_ratio=4,
41
+ qkv_bias=True,
42
+ drop_path_rate=0.2,
43
+ interaction_indexes=[[0, 0], [1, 1], [2, 2], [3, 3], [4, 4], [5, 5], [6, 6], [7, 7], [8, 8], [9, 9], [10, 10], [11, 11]],
44
+ use_cls_token=True,
45
+ use_flash_attn=True,
46
+ with_cp=True,
47
+ pretrained="pretrained/B_16-i21k-300ep-lr_0.001-aug_medium1-wd_0.1-do_0.0-sd_0.0--imagenet2012-steps_20k-lr_0.01-res_224.pth",
48
+ ),
49
+
50
+ branch3=dict(
51
+ model_type="augreg",
52
+ img_size=384,
53
+ patch_size=16,
54
+ pretrain_img_size=224,
55
+ pretrain_patch_size=16,
56
+ depth=12,
57
+ embed_dim=384,
58
+ num_heads=6,
59
+ mlp_ratio=4,
60
+ qkv_bias=True,
61
+ drop_path_rate=0.05,
62
+ interaction_indexes=[[0, 0], [1, 1], [2, 2], [3, 3], [4, 4], [5, 5], [6, 6], [7, 7], [8, 8], [9, 9], [10, 10], [11, 11]],
63
+ use_cls_token=True,
64
+ use_flash_attn=True,
65
+ with_cp=True,
66
+ pretrained="pretrained/S_16-i21k-300ep-lr_0.001-aug_light1-wd_0.03-do_0.0-sd_0.0--imagenet2012-steps_20k-lr_0.03-res_224.pth",
67
+ ),
68
+ ),
69
+ )
classification/piip_3branch_sbl_384-192-128_cls_token_augreg.txt ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"train_lr": 9.999999999999953e-07, "train_loss": 5.234040696939214, "test_loss": 0.7643236819354967, "test_acc1": 80.69800254882813, "test_acc5": 95.80800258789063, "epoch": 0, "n_parameters": "488.93M", "max_acc": 80.69800254882813}
2
+ {"train_lr": 9.999999999999953e-07, "train_loss": 4.569301353703491, "test_loss": 0.6868039875423314, "test_acc1": 81.81000241210937, "test_acc5": 96.386002421875, "epoch": 1, "n_parameters": "488.93M", "max_acc": 81.81000241210937}
3
+ {"train_lr": 6.800000000000116e-06, "train_loss": 3.8520777062784664, "test_loss": 0.7450303433041919, "test_acc1": 83.4320025390625, "test_acc5": 97.1280023828125, "epoch": 2, "n_parameters": "488.93M", "max_acc": 83.4320025390625}
4
+ {"train_lr": 1.2599999999999859e-05, "train_loss": 3.4228615775096904, "test_loss": 0.6950710139512102, "test_acc1": 84.14200250976562, "test_acc5": 97.35000255859374, "epoch": 3, "n_parameters": "488.93M", "max_acc": 84.14200250976562}
5
+ {"train_lr": 1.8399999999999458e-05, "train_loss": 3.230861018780324, "test_loss": 0.6867068752475168, "test_acc1": 84.32600256835937, "test_acc5": 97.42400245117187, "epoch": 4, "n_parameters": "488.93M", "max_acc": 84.32600256835937}
6
+ {"train_lr": 2.419999999999938e-05, "train_loss": 3.1228197528708943, "test_loss": 0.6822546062798336, "test_acc1": 84.4180024609375, "test_acc5": 97.47800227539062, "epoch": 5, "n_parameters": "488.93M", "max_acc": 84.4180024609375}
7
+ {"train_lr": 2.7071067811864602e-05, "train_loss": 3.0558222811356437, "test_loss": 0.7145715792060355, "test_acc1": 84.65400235351562, "test_acc5": 97.5200025390625, "epoch": 6, "n_parameters": "488.93M", "max_acc": 84.65400235351562}
8
+ {"train_lr": 2.4539904997396135e-05, "train_loss": 2.994581277648227, "test_loss": 0.662302630957059, "test_acc1": 84.8080023046875, "test_acc5": 97.55600248046875, "epoch": 7, "n_parameters": "488.93M", "max_acc": 84.8080023046875}
9
+ {"train_lr": 2.4539904997396135e-05, "train_loss": 2.9550487662104965, "test_loss": 0.6481138851450777, "test_acc1": 85.00600231445313, "test_acc5": 97.60400245117188, "epoch": 8, "n_parameters": "488.93M", "max_acc": 85.00600231445313}
10
+ {"train_lr": 2.309016994374978e-05, "train_loss": 2.9124116585504334, "test_loss": 0.6609158732881948, "test_acc1": 85.05600260742187, "test_acc5": 97.67200237304688, "epoch": 9, "n_parameters": "488.93M", "max_acc": 85.05600260742187}
11
+ {"train_lr": 2.1564344650402722e-05, "train_loss": 2.8919068921050677, "test_loss": 0.6541004997331977, "test_acc1": 85.2100023046875, "test_acc5": 97.65200245117188, "epoch": 10, "n_parameters": "488.93M", "max_acc": 85.2100023046875}
12
+ {"train_lr": 1.9999999999999453e-05, "train_loss": 2.859698434003728, "test_loss": 0.6424851714422877, "test_acc1": 85.30200254882813, "test_acc5": 97.71600241210938, "epoch": 11, "n_parameters": "488.93M", "max_acc": 85.30200254882813}
13
+ {"train_lr": 1.843565534959796e-05, "train_loss": 2.829253721187155, "test_loss": 0.6302828912077278, "test_acc1": 85.3000025390625, "test_acc5": 97.7280025, "epoch": 12, "n_parameters": "488.93M", "max_acc": 85.30200254882813}
14
+ {"train_lr": 1.6909830056250673e-05, "train_loss": 2.8153070922521093, "test_loss": 0.6341994082790682, "test_acc1": 85.4780025, "test_acc5": 97.75400227539062, "epoch": 13, "n_parameters": "488.93M", "max_acc": 85.4780025}
15
+ {"train_lr": 1.412214747707529e-05, "train_loss": 2.7915727422415593, "test_loss": 0.6172244596755367, "test_acc1": 85.55800256835937, "test_acc5": 97.8040024609375, "epoch": 14, "n_parameters": "488.93M", "max_acc": 85.55800256835937}
16
+ {"train_lr": 1.412214747707529e-05, "train_loss": 2.778482116883893, "test_loss": 0.6217652284094201, "test_acc1": 85.57000250976563, "test_acc5": 97.8180025, "epoch": 15, "n_parameters": "488.93M", "max_acc": 85.57000250976563}
17
+ {"train_lr": 1.2928932188134564e-05, "train_loss": 2.7557977602469452, "test_loss": 0.6224205930114249, "test_acc1": 85.6060025390625, "test_acc5": 97.820002421875, "epoch": 16, "n_parameters": "488.93M", "max_acc": 85.6060025390625}
18
+ {"train_lr": 1.1909830056250373e-05, "train_loss": 2.7540953340742895, "test_loss": 0.6279844817530607, "test_acc1": 85.71400239257812, "test_acc5": 97.86200236328125, "epoch": 17, "n_parameters": "488.93M", "max_acc": 85.71400239257812}
19
+ {"train_lr": 1.1089934758116523e-05, "train_loss": 2.736129855009006, "test_loss": 0.6265762991832133, "test_acc1": 85.73800283203126, "test_acc5": 97.8480023828125, "epoch": 18, "n_parameters": "488.93M", "max_acc": 85.73800283203126}
20
+ {"train_lr": 1.0489434837048593e-05, "train_loss": 2.721998798916285, "test_loss": 0.616532351322101, "test_acc1": 85.85400275390624, "test_acc5": 97.87000234375, "epoch": 19, "n_parameters": "488.93M", "max_acc": 85.85400275390624}
classification/piip_3branch_tsb_368-192-128_cls_token_deit1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93a394ed9cc2414f168d605299aecd4140d710549e79167a2f00b7910aebdf64
3
+ size 576800265
classification/piip_3branch_tsb_368-192-128_cls_token_deit1.py ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model = dict(
2
+ backbone=dict(
3
+ n_points=4,
4
+ deform_num_heads=16,
5
+ cffn_ratio=0.25,
6
+ deform_ratio=0.5,
7
+ with_cffn=True,
8
+ interact_attn_type='deform',
9
+ interaction_drop_path_rate=0.4,
10
+ separate_head=True,
11
+
12
+ branch1=dict(
13
+ img_size=128,
14
+ patch_size=16,
15
+ pretrain_img_size=224,
16
+ pretrain_patch_size=16,
17
+ depth=12,
18
+ embed_dim=768,
19
+ num_heads=12,
20
+ mlp_ratio=4,
21
+ init_scale=1.0,
22
+ qkv_bias=True,
23
+ drop_rate=0.0,
24
+ drop_path_rate=0.2,
25
+ interaction_indexes=[[0, 0], [1, 1], [2, 2], [3, 3], [4, 4], [5, 5], [6, 6], [7, 7], [8, 8], [9, 9], [10, 10], [11, 11]],
26
+ use_cls_token=True,
27
+ use_flash_attn=True,
28
+ with_cp=True,
29
+ pretrained="pretrained/deit_base_patch16_224-b5f2ef4d.pth",
30
+ ),
31
+
32
+ branch2=dict(
33
+ img_size=192,
34
+ patch_size=16,
35
+ pretrain_img_size=224,
36
+ pretrain_patch_size=16,
37
+ depth=12,
38
+ embed_dim=384,
39
+ num_heads=6,
40
+ mlp_ratio=4,
41
+ init_scale=1.0,
42
+ qkv_bias=True,
43
+ drop_rate=0.0,
44
+ drop_path_rate=0.05,
45
+ interaction_indexes=[[0, 0], [1, 1], [2, 2], [3, 3], [4, 4], [5, 5], [6, 6], [7, 7], [8, 8], [9, 9], [10, 10], [11, 11]],
46
+ use_cls_token=True,
47
+ use_flash_attn=True,
48
+ with_cp=True,
49
+ pretrained="pretrained/deit_small_patch16_224-cd65a155.pth",
50
+ ),
51
+
52
+ branch3=dict(
53
+ img_size=368,
54
+ patch_size=16,
55
+ pretrain_img_size=224,
56
+ pretrain_patch_size=16,
57
+ depth=12,
58
+ embed_dim=192,
59
+ num_heads=3,
60
+ mlp_ratio=4,
61
+ init_scale=1.0,
62
+ qkv_bias=True,
63
+ drop_rate=0.0,
64
+ drop_path_rate=0.05,
65
+ interaction_indexes=[[0, 0], [1, 1], [2, 2], [3, 3], [4, 4], [5, 5], [6, 6], [7, 7], [8, 8], [9, 9], [10, 10], [11, 11]],
66
+ use_cls_token=True,
67
+ use_flash_attn=True,
68
+ with_cp=True,
69
+ pretrained="pretrained/deit_tiny_patch16_224-a1311bcf.pth",
70
+ ),
71
+ ),
72
+ )
classification/piip_3branch_tsb_368-192-128_cls_token_deit1.txt ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"train_lr": 9.999999999999953e-07, "train_loss": 4.325746201854244, "test_loss": 0.968449961636715, "test_acc1": 77.6400026171875, "test_acc5": 94.0320025, "epoch": 0, "n_parameters": "144.03M", "max_acc": 77.6400026171875}
2
+ {"train_lr": 9.999999999999953e-07, "train_loss": 3.53042768207576, "test_loss": 0.885815540492763, "test_acc1": 78.67200244140625, "test_acc5": 94.55400229492187, "epoch": 1, "n_parameters": "144.03M", "max_acc": 78.67200244140625}
3
+ {"train_lr": 6.800000000000116e-06, "train_loss": 3.15689085352478, "test_loss": 0.8606244549440698, "test_acc1": 80.66800249023437, "test_acc5": 95.42000250976562, "epoch": 2, "n_parameters": "144.03M", "max_acc": 80.66800249023437}
4
+ {"train_lr": 1.2599999999999859e-05, "train_loss": 3.0174922107173194, "test_loss": 0.8475330369225864, "test_acc1": 80.9340026171875, "test_acc5": 95.67000276367187, "epoch": 3, "n_parameters": "144.03M", "max_acc": 80.9340026171875}
5
+ {"train_lr": 1.8399999999999458e-05, "train_loss": 2.944666322341545, "test_loss": 0.8161598738582655, "test_acc1": 81.30000252929688, "test_acc5": 95.73600258789062, "epoch": 4, "n_parameters": "144.03M", "max_acc": 81.30000252929688}
6
+ {"train_lr": 2.419999999999938e-05, "train_loss": 2.903430822691757, "test_loss": 0.8191158705287509, "test_acc1": 81.28600240234375, "test_acc5": 95.74400268554687, "epoch": 5, "n_parameters": "144.03M", "max_acc": 81.30000252929688}
7
+ {"train_lr": 2.7071067811864602e-05, "train_loss": 2.8867583846469387, "test_loss": 0.8098847847788727, "test_acc1": 81.50000265625, "test_acc5": 95.76600247070313, "epoch": 6, "n_parameters": "144.03M", "max_acc": 81.50000265625}
8
+ {"train_lr": 2.5877852522924498e-05, "train_loss": 2.862130629525577, "test_loss": 0.814881503810371, "test_acc1": 81.5040023828125, "test_acc5": 95.82200268554688, "epoch": 7, "n_parameters": "144.03M", "max_acc": 81.5040023828125}
9
+ {"train_lr": 2.309016994374978e-05, "train_loss": 2.840806765533942, "test_loss": 0.8066507335823615, "test_acc1": 81.69200245117187, "test_acc5": 95.8480028125, "epoch": 8, "n_parameters": "144.03M", "max_acc": 81.69200245117187}
10
+ {"train_lr": 2.309016994374978e-05, "train_loss": 2.824903579329034, "test_loss": 0.7972967231867414, "test_acc1": 81.8100025, "test_acc5": 95.87600274414062, "epoch": 9, "n_parameters": "144.03M", "max_acc": 81.8100025}
11
+ {"train_lr": 2.1564344650402722e-05, "train_loss": 2.813586833999216, "test_loss": 0.8049706921723611, "test_acc1": 81.70400248046874, "test_acc5": 95.83000237304688, "epoch": 10, "n_parameters": "144.03M", "max_acc": 81.8100025}
12
+ {"train_lr": 1.9999999999999453e-05, "train_loss": 2.8057527802140116, "test_loss": 0.7967677732993816, "test_acc1": 81.7060026953125, "test_acc5": 95.89400256835937, "epoch": 11, "n_parameters": "144.03M", "max_acc": 81.8100025}
13
+ {"train_lr": 1.843565534959796e-05, "train_loss": 2.785416632616739, "test_loss": 0.791968504359439, "test_acc1": 81.9300023828125, "test_acc5": 95.94000270507813, "epoch": 12, "n_parameters": "144.03M", "max_acc": 81.9300023828125}
14
+ {"train_lr": 1.6909830056250673e-05, "train_loss": 2.775560283451248, "test_loss": 0.7932183128207123, "test_acc1": 81.86800266601563, "test_acc5": 95.90800266601562, "epoch": 13, "n_parameters": "144.03M", "max_acc": 81.9300023828125}
15
+ {"train_lr": 1.5460095002604783e-05, "train_loss": 2.773722091381498, "test_loss": 0.790430608151973, "test_acc1": 81.95200254882812, "test_acc5": 95.92800255859375, "epoch": 14, "n_parameters": "144.03M", "max_acc": 81.95200254882812}
16
+ {"train_lr": 1.2928932188134564e-05, "train_loss": 2.7609039588654927, "test_loss": 0.7895037974891078, "test_acc1": 82.03600266601562, "test_acc5": 95.98800267578125, "epoch": 15, "n_parameters": "144.03M", "max_acc": 82.03600266601562}
17
+ {"train_lr": 1.2928932188134564e-05, "train_loss": 2.7523779093766576, "test_loss": 0.7834799431292947, "test_acc1": 82.04000245117187, "test_acc5": 95.990002421875, "epoch": 16, "n_parameters": "144.03M", "max_acc": 82.04000245117187}
18
+ {"train_lr": 1.1909830056250373e-05, "train_loss": 2.7483434154094457, "test_loss": 0.7865663038359748, "test_acc1": 82.03600228515624, "test_acc5": 96.034002578125, "epoch": 17, "n_parameters": "144.03M", "max_acc": 82.04000245117187}
19
+ {"train_lr": 1.1089934758116523e-05, "train_loss": 2.748136612830116, "test_loss": 0.7830524395480467, "test_acc1": 82.08200236328125, "test_acc5": 96.03800256835937, "epoch": 18, "n_parameters": "144.03M", "max_acc": 82.08200236328125}
20
+ {"train_lr": 1.0489434837048593e-05, "train_loss": 2.733028691235206, "test_loss": 0.79214631780354, "test_acc1": 82.0480024609375, "test_acc5": 95.98400268554687, "epoch": 19, "n_parameters": "144.03M", "max_acc": 82.08200236328125}