Daniel246 commited on
Commit
80d5c00
1 Parent(s): ef4b518

Upload folder using huggingface_hub

Browse files
args.yaml ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ task: detect
2
+ mode: train
3
+ model: yolov8s.pt
4
+ data: data/drug_class_2023_12_5_s_640/drug_class_2023_12_5_s_640.yaml
5
+ epochs: 1000
6
+ patience: 50
7
+ batch: 64
8
+ imgsz: 640
9
+ save: true
10
+ cache: false
11
+ device: null
12
+ workers: 8
13
+ project: runs/detect
14
+ name: drug_class_2023_12_5_s_640
15
+ exist_ok: false
16
+ pretrained: false
17
+ optimizer: SGD
18
+ verbose: true
19
+ seed: 0
20
+ deterministic: true
21
+ single_cls: false
22
+ image_weights: false
23
+ rect: false
24
+ cos_lr: false
25
+ close_mosaic: 10
26
+ resume: false
27
+ overlap_mask: true
28
+ mask_ratio: 4
29
+ dropout: 0.0
30
+ val: true
31
+ save_json: false
32
+ save_hybrid: false
33
+ conf: null
34
+ iou: 0.7
35
+ max_det: 300
36
+ half: false
37
+ dnn: false
38
+ plots: true
39
+ source: null
40
+ show: false
41
+ save_txt: false
42
+ save_conf: false
43
+ save_crop: false
44
+ hide_labels: false
45
+ hide_conf: false
46
+ vid_stride: 1
47
+ line_thickness: 3
48
+ visualize: false
49
+ augment: false
50
+ agnostic_nms: false
51
+ classes: null
52
+ retina_masks: false
53
+ boxes: true
54
+ format: torchscript
55
+ keras: false
56
+ optimize: false
57
+ int8: false
58
+ dynamic: false
59
+ simplify: false
60
+ opset: null
61
+ workspace: 4
62
+ nms: false
63
+ lr0: 0.01
64
+ lrf: 0.01
65
+ momentum: 0.937
66
+ weight_decay: 0.0005
67
+ warmup_epochs: 3.0
68
+ warmup_momentum: 0.8
69
+ warmup_bias_lr: 0.1
70
+ box: 7.5
71
+ cls: 0.5
72
+ dfl: 1.5
73
+ fl_gamma: 0.0
74
+ label_smoothing: 0.0
75
+ nbs: 64
76
+ hsv_h: 0.015
77
+ hsv_s: 0.7
78
+ hsv_v: 0.4
79
+ degrees: 0.0
80
+ translate: 0.1
81
+ scale: 0.5
82
+ shear: 0.0
83
+ perspective: 0.0
84
+ flipud: 0.0
85
+ fliplr: 0.5
86
+ mosaic: 1.0
87
+ mixup: 0.0
88
+ copy_paste: 0.0
89
+ cfg: null
90
+ v5loader: false
91
+ save_dir: runs/detect/drug_class_2023_12_5_s_640
events.out.tfevents.1701741678.ced569c9f804.1.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e9016ef5284553c6548c6fdeac5d2fde371637c7bb8326a1ecfc432d9897f86
3
+ size 9625534
results.csv ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch, train/box_loss, train/cls_loss, train/dfl_loss, metrics/precision(B), metrics/recall(B), metrics/mAP50(B), metrics/mAP50-95(B), val/box_loss, val/cls_loss, val/dfl_loss, lr/pg0, lr/pg1, lr/pg2
2
+ 0, 1.0648, 0.95599, 1.1481, 0.84447, 0.65332, 0.71167, 0.47169, 1.0361, 0.85667, 1.0629, 0.070022, 0.0033308, 0.0033308
3
+ 1, 1.1018, 0.86778, 1.1585, 0.85809, 0.67569, 0.72911, 0.48727, 1.1097, 0.94623, 1.1066, 0.040016, 0.0066576, 0.0066576
4
+ 2, 1.2159, 1.0068, 1.2307, 0.79804, 0.66629, 0.70557, 0.46379, 1.2251, 1.0844, 1.1897, 0.010003, 0.0099777, 0.0099777
5
+ 3, 1.2384, 1.0304, 1.251, 0.86486, 0.69854, 0.72594, 0.48123, 1.1207, 0.95431, 1.1177, 0.0099703, 0.0099703, 0.0099703
6
+ 4, 1.1852, 0.96311, 1.2181, 0.89192, 0.7057, 0.73796, 0.50947, 1.0681, 0.88188, 1.0974, 0.0099703, 0.0099703, 0.0099703
7
+ 5, 1.1546, 0.92271, 1.1987, 0.905, 0.69582, 0.73812, 0.5128, 1.0366, 0.83952, 1.077, 0.0099604, 0.0099604, 0.0099604
8
+ 6, 1.1374, 0.89981, 1.1863, 0.89674, 0.70683, 0.74243, 0.51734, 1.0237, 0.8102, 1.067, 0.0099505, 0.0099505, 0.0099505
9
+ 7, 1.1212, 0.88004, 1.1778, 0.91234, 0.7089, 0.74775, 0.52638, 1.0052, 0.78836, 1.0592, 0.0099406, 0.0099406, 0.0099406
10
+ 8, 1.1131, 0.87043, 1.1714, 0.91572, 0.71724, 0.75255, 0.52934, 0.99458, 0.77485, 1.055, 0.0099307, 0.0099307, 0.0099307
11
+ 9, 1.1039, 0.86022, 1.1676, 0.92091, 0.71783, 0.75049, 0.52717, 0.98567, 0.76516, 1.0489, 0.0099208, 0.0099208, 0.0099208
12
+ 10, 1.0978, 0.85174, 1.1623, 0.91134, 0.71968, 0.75394, 0.53421, 0.98246, 0.75802, 1.0475, 0.0099109, 0.0099109, 0.0099109
13
+ 11, 1.0921, 0.8481, 1.1582, 0.94051, 0.71896, 0.75489, 0.53865, 0.97662, 0.75198, 1.0451, 0.009901, 0.009901, 0.009901
14
+ 12, 1.0915, 0.84095, 1.156, 0.90711, 0.72445, 0.75626, 0.54249, 0.97299, 0.74766, 1.0432, 0.0098911, 0.0098911, 0.0098911
15
+ 13, 1.0829, 0.83623, 1.1529, 0.90561, 0.71878, 0.75688, 0.54374, 0.96976, 0.74514, 1.0416, 0.0098812, 0.0098812, 0.0098812
16
+ 14, 1.0819, 0.83302, 1.1521, 0.90486, 0.72128, 0.7563, 0.54388, 0.96764, 0.74254, 1.0405, 0.0098713, 0.0098713, 0.0098713
17
+ 15, 1.0781, 0.82918, 1.1497, 0.9295, 0.71771, 0.75691, 0.54315, 0.96667, 0.74057, 1.0398, 0.0098614, 0.0098614, 0.0098614
18
+ 16, 1.0759, 0.82793, 1.1496, 0.92544, 0.72114, 0.75697, 0.54201, 0.96537, 0.73818, 1.0389, 0.0098515, 0.0098515, 0.0098515
19
+ 17, 1.0767, 0.82274, 1.1475, 0.9155, 0.72519, 0.75709, 0.54154, 0.9647, 0.73707, 1.0385, 0.0098416, 0.0098416, 0.0098416
20
+ 18, 1.0742, 0.8232, 1.1474, 0.9171, 0.72574, 0.75717, 0.5408, 0.96388, 0.73565, 1.0384, 0.0098317, 0.0098317, 0.0098317
21
+ 19, 1.071, 0.82129, 1.1465, 0.9012, 0.7261, 0.75735, 0.54106, 0.96281, 0.73424, 1.0381, 0.0098218, 0.0098218, 0.0098218
22
+ 20, 1.07, 0.81875, 1.1462, 0.9027, 0.72612, 0.75799, 0.54112, 0.96196, 0.73294, 1.0377, 0.0098119, 0.0098119, 0.0098119
23
+ 21, 1.0692, 0.81699, 1.1446, 0.90038, 0.7254, 0.75835, 0.54018, 0.96118, 0.73217, 1.0375, 0.009802, 0.009802, 0.009802
24
+ 22, 1.0677, 0.81591, 1.1454, 0.9013, 0.726, 0.75819, 0.54018, 0.9606, 0.73186, 1.0373, 0.0097921, 0.0097921, 0.0097921
25
+ 23, 1.0678, 0.81458, 1.1433, 0.90351, 0.72402, 0.75828, 0.54126, 0.95997, 0.73108, 1.037, 0.0097822, 0.0097822, 0.0097822
26
+ 24, 1.0671, 0.8147, 1.1433, 0.90935, 0.72463, 0.75856, 0.54265, 0.95985, 0.73079, 1.037, 0.0097723, 0.0097723, 0.0097723
27
+ 25, 1.0663, 0.81264, 1.1405, 0.91141, 0.7262, 0.75897, 0.54266, 0.95905, 0.73033, 1.0362, 0.0097624, 0.0097624, 0.0097624
28
+ 26, 1.0651, 0.81122, 1.1409, 0.91052, 0.72685, 0.7598, 0.54305, 0.95842, 0.73004, 1.036, 0.0097525, 0.0097525, 0.0097525
29
+ 27, 1.063, 0.8108, 1.1412, 0.91227, 0.72705, 0.76014, 0.54366, 0.95789, 0.72885, 1.036, 0.0097426, 0.0097426, 0.0097426
30
+ 28, 1.0636, 0.81191, 1.1417, 0.92087, 0.72191, 0.7603, 0.54391, 0.95695, 0.72845, 1.0354, 0.0097327, 0.0097327, 0.0097327
31
+ 29, 1.0629, 0.8115, 1.142, 0.92209, 0.72162, 0.76106, 0.54422, 0.95646, 0.72843, 1.0352, 0.0097228, 0.0097228, 0.0097228
32
+ 30, 1.0623, 0.8109, 1.1408, 0.93031, 0.71713, 0.76105, 0.54522, 0.95558, 0.72827, 1.0346, 0.0097129, 0.0097129, 0.0097129
33
+ 31, 1.0627, 0.8076, 1.1414, 0.93165, 0.71765, 0.7616, 0.54703, 0.95509, 0.72748, 1.0343, 0.009703, 0.009703, 0.009703
34
+ 32, 1.0606, 0.80575, 1.1398, 0.93323, 0.71792, 0.76163, 0.5464, 0.95444, 0.72703, 1.034, 0.0096931, 0.0096931, 0.0096931
35
+ 33, 1.0612, 0.80738, 1.1388, 0.92669, 0.72094, 0.762, 0.54661, 0.9542, 0.72645, 1.0339, 0.0096832, 0.0096832, 0.0096832
36
+ 34, 1.0614, 0.80976, 1.1399, 0.92535, 0.7261, 0.76296, 0.54611, 0.95366, 0.72612, 1.0336, 0.0096733, 0.0096733, 0.0096733
37
+ 35, 1.0604, 0.80585, 1.1386, 0.92623, 0.72619, 0.76335, 0.54783, 0.95317, 0.72618, 1.0333, 0.0096634, 0.0096634, 0.0096634
38
+ 36, 1.06, 0.80504, 1.1371, 0.92379, 0.72615, 0.76363, 0.54722, 0.95246, 0.72625, 1.0327, 0.0096535, 0.0096535, 0.0096535
39
+ 37, 1.0586, 0.80369, 1.1372, 0.92384, 0.72603, 0.76435, 0.54712, 0.95194, 0.7263, 1.0323, 0.0096436, 0.0096436, 0.0096436
40
+ 38, 1.0584, 0.80316, 1.137, 0.92846, 0.72626, 0.76416, 0.54692, 0.95195, 0.72551, 1.0324, 0.0096337, 0.0096337, 0.0096337
41
+ 39, 1.058, 0.80411, 1.1379, 0.92875, 0.72817, 0.76395, 0.54604, 0.95189, 0.72603, 1.0323, 0.0096238, 0.0096238, 0.0096238
42
+ 40, 1.0587, 0.80307, 1.137, 0.9278, 0.72805, 0.76356, 0.54596, 0.9516, 0.72566, 1.0319, 0.0096139, 0.0096139, 0.0096139
43
+ 41, 1.0584, 0.80226, 1.1368, 0.92754, 0.72825, 0.76386, 0.54721, 0.95175, 0.72518, 1.0317, 0.009604, 0.009604, 0.009604
44
+ 42, 1.0597, 0.8032, 1.137, 0.92811, 0.72794, 0.76357, 0.54904, 0.95181, 0.72478, 1.0314, 0.0095941, 0.0095941, 0.0095941
45
+ 43, 1.0559, 0.80144, 1.136, 0.92927, 0.72756, 0.76391, 0.54766, 0.95169, 0.7248, 1.0313, 0.0095842, 0.0095842, 0.0095842
46
+ 44, 1.0566, 0.80112, 1.1351, 0.93607, 0.72721, 0.76381, 0.5479, 0.95159, 0.72483, 1.0314, 0.0095743, 0.0095743, 0.0095743
train_batch0.jpg ADDED
train_batch1.jpg ADDED
train_batch2.jpg ADDED
weights/best.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26b767be5771bf579c3f11793e07b5c56c4b898c8c1ee4ce7d6f1d3b9e5d9a5d
3
+ size 89533858
weights/last.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6606d721cf18579f5b69ffa65f9abbef93468162300748f2e9c28c8a43df8126
3
+ size 89533858