nancyalarabawy commited on
Commit
ce039cb
1 Parent(s): f806bbc

End of training

Browse files
README.md CHANGED
@@ -16,6 +16,19 @@ should probably proofread and complete it, then remove this comment. -->
16
  # segformer-finetuned-segments-riceleafdisease-dec-18
17
 
18
  This model is a fine-tuned version of [nvidia/mit-b0](https://huggingface.co/nvidia/mit-b0) on the nancyalarabawy/RiceLeafDiseases dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  ## Model description
21
 
@@ -34,7 +47,7 @@ More information needed
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
- - learning_rate: 0.0006
38
  - train_batch_size: 2
39
  - eval_batch_size: 2
40
  - seed: 42
@@ -42,6 +55,72 @@ The following hyperparameters were used during training:
42
  - lr_scheduler_type: linear
43
  - num_epochs: 20
44
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  ### Framework versions
46
 
47
  - Transformers 4.36.2
 
16
  # segformer-finetuned-segments-riceleafdisease-dec-18
17
 
18
  This model is a fine-tuned version of [nvidia/mit-b0](https://huggingface.co/nvidia/mit-b0) on the nancyalarabawy/RiceLeafDiseases dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 0.0654
21
+ - Mean Iou: 0.8022
22
+ - Mean Accuracy: 0.8495
23
+ - Overall Accuracy: 0.9793
24
+ - Accuracy Unlabelled: nan
25
+ - Accuracy Healthy: 0.9379
26
+ - Accuracy Disease: 0.6144
27
+ - Accuracy Background: 0.9962
28
+ - Iou Unlabelled: nan
29
+ - Iou Healthy: 0.8897
30
+ - Iou Disease: 0.5318
31
+ - Iou Background: 0.9851
32
 
33
  ## Model description
34
 
 
47
  ### Training hyperparameters
48
 
49
  The following hyperparameters were used during training:
50
+ - learning_rate: 6e-05
51
  - train_batch_size: 2
52
  - eval_batch_size: 2
53
  - seed: 42
 
55
  - lr_scheduler_type: linear
56
  - num_epochs: 20
57
 
58
+ ### Training results
59
+
60
+ | Training Loss | Epoch | Step | Validation Loss | Mean Iou | Mean Accuracy | Overall Accuracy | Accuracy Unlabelled | Accuracy Healthy | Accuracy Disease | Accuracy Background | Iou Unlabelled | Iou Healthy | Iou Disease | Iou Background |
61
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:-------------:|:----------------:|:-------------------:|:----------------:|:----------------:|:-------------------:|:--------------:|:-----------:|:-----------:|:--------------:|
62
+ | 1.0053 | 0.33 | 20 | 1.2067 | 0.3883 | 0.6563 | 0.8803 | nan | 0.9886 | 0.0986 | 0.8816 | 0.0 | 0.6303 | 0.0418 | 0.8810 |
63
+ | 0.8256 | 0.67 | 40 | 0.8465 | 0.4445 | 0.7343 | 0.9192 | nan | 0.9352 | 0.3360 | 0.9317 | 0.0 | 0.7090 | 0.1390 | 0.9300 |
64
+ | 0.4785 | 1.0 | 60 | 0.5391 | 0.4488 | 0.6814 | 0.9466 | nan | 0.9824 | 0.0991 | 0.9626 | 0.0 | 0.7520 | 0.0852 | 0.9579 |
65
+ | 0.5875 | 1.33 | 80 | 0.6045 | 0.6251 | 0.7087 | 0.9510 | nan | 0.9790 | 0.1807 | 0.9664 | nan | 0.7718 | 0.1407 | 0.9627 |
66
+ | 0.5354 | 1.67 | 100 | 0.3814 | 0.6633 | 0.7242 | 0.9596 | nan | 0.9313 | 0.2584 | 0.9831 | nan | 0.7937 | 0.2267 | 0.9694 |
67
+ | 0.5146 | 2.0 | 120 | 0.3455 | 0.6755 | 0.7307 | 0.9650 | nan | 0.9507 | 0.2552 | 0.9862 | nan | 0.8187 | 0.2321 | 0.9756 |
68
+ | 0.5212 | 2.33 | 140 | 0.4321 | 0.6471 | 0.7169 | 0.9598 | nan | 0.9798 | 0.1946 | 0.9764 | nan | 0.7934 | 0.1752 | 0.9727 |
69
+ | 0.4044 | 2.67 | 160 | 0.1609 | 0.6607 | 0.7065 | 0.9607 | nan | 0.9067 | 0.2230 | 0.9898 | nan | 0.7994 | 0.2145 | 0.9681 |
70
+ | 0.4572 | 3.0 | 180 | 0.2172 | 0.6882 | 0.7268 | 0.9657 | nan | 0.9050 | 0.2808 | 0.9946 | nan | 0.8254 | 0.2667 | 0.9724 |
71
+ | 0.2676 | 3.33 | 200 | 0.2950 | 0.6926 | 0.7451 | 0.9690 | nan | 0.9674 | 0.2806 | 0.9874 | nan | 0.8332 | 0.2641 | 0.9804 |
72
+ | 0.2991 | 3.67 | 220 | 0.2523 | 0.6976 | 0.7411 | 0.9695 | nan | 0.9337 | 0.2959 | 0.9937 | nan | 0.8401 | 0.2734 | 0.9792 |
73
+ | 0.5168 | 4.0 | 240 | 0.1013 | 0.6599 | 0.6969 | 0.9651 | nan | 0.9065 | 0.1880 | 0.9961 | nan | 0.8226 | 0.1835 | 0.9735 |
74
+ | 0.3117 | 4.33 | 260 | 0.2323 | 0.6951 | 0.7424 | 0.9696 | nan | 0.9496 | 0.2863 | 0.9912 | nan | 0.8409 | 0.2644 | 0.9800 |
75
+ | 0.2888 | 4.67 | 280 | 0.1264 | 0.7466 | 0.7901 | 0.9736 | nan | 0.9277 | 0.4470 | 0.9957 | nan | 0.8582 | 0.3998 | 0.9817 |
76
+ | 0.1684 | 5.0 | 300 | 0.1291 | 0.7646 | 0.8119 | 0.9758 | nan | 0.9373 | 0.5033 | 0.9951 | nan | 0.8693 | 0.4406 | 0.9837 |
77
+ | 0.2041 | 5.33 | 320 | 0.1804 | 0.7720 | 0.8213 | 0.9773 | nan | 0.9462 | 0.5229 | 0.9948 | nan | 0.8732 | 0.4564 | 0.9863 |
78
+ | 0.179 | 5.67 | 340 | 0.1381 | 0.7937 | 0.8560 | 0.9773 | nan | 0.9280 | 0.6454 | 0.9948 | nan | 0.8727 | 0.5235 | 0.9849 |
79
+ | 0.1444 | 6.0 | 360 | 0.1671 | 0.7393 | 0.7972 | 0.9727 | nan | 0.9750 | 0.4301 | 0.9866 | nan | 0.8485 | 0.3862 | 0.9832 |
80
+ | 0.2365 | 6.33 | 380 | 0.1272 | 0.7813 | 0.8275 | 0.9771 | nan | 0.9354 | 0.5511 | 0.9958 | nan | 0.8722 | 0.4870 | 0.9848 |
81
+ | 0.2216 | 6.67 | 400 | 0.0907 | 0.7923 | 0.8358 | 0.9775 | nan | 0.9224 | 0.5875 | 0.9976 | nan | 0.8761 | 0.5171 | 0.9837 |
82
+ | 0.1437 | 7.0 | 420 | 0.0782 | 0.7715 | 0.8148 | 0.9732 | nan | 0.9167 | 0.5329 | 0.9948 | nan | 0.8591 | 0.4777 | 0.9778 |
83
+ | 0.1065 | 7.33 | 440 | 0.0877 | 0.7537 | 0.7917 | 0.9725 | nan | 0.9187 | 0.4610 | 0.9955 | nan | 0.8549 | 0.4281 | 0.9781 |
84
+ | 0.2535 | 7.67 | 460 | 0.0784 | 0.7457 | 0.7810 | 0.9723 | nan | 0.9078 | 0.4374 | 0.9979 | nan | 0.8593 | 0.4003 | 0.9776 |
85
+ | 0.108 | 8.0 | 480 | 0.1003 | 0.7544 | 0.7984 | 0.9759 | nan | 0.9563 | 0.4455 | 0.9934 | nan | 0.8703 | 0.4084 | 0.9844 |
86
+ | 0.0884 | 8.33 | 500 | 0.0744 | 0.7689 | 0.8108 | 0.9753 | nan | 0.9335 | 0.5039 | 0.9952 | nan | 0.8694 | 0.4557 | 0.9816 |
87
+ | 0.1935 | 8.67 | 520 | 0.1047 | 0.7954 | 0.8373 | 0.9800 | nan | 0.9482 | 0.5672 | 0.9965 | nan | 0.8878 | 0.5111 | 0.9874 |
88
+ | 0.268 | 9.0 | 540 | 0.1281 | 0.8058 | 0.8559 | 0.9816 | nan | 0.9563 | 0.6158 | 0.9957 | nan | 0.8945 | 0.5328 | 0.9900 |
89
+ | 0.0779 | 9.33 | 560 | 0.1393 | 0.7680 | 0.8126 | 0.9774 | nan | 0.9645 | 0.4806 | 0.9928 | nan | 0.8763 | 0.4418 | 0.9859 |
90
+ | 0.1295 | 9.67 | 580 | 0.0749 | 0.7433 | 0.7861 | 0.9741 | nan | 0.9572 | 0.4092 | 0.9920 | nan | 0.8595 | 0.3877 | 0.9826 |
91
+ | 0.6322 | 10.0 | 600 | 0.0792 | 0.7825 | 0.8233 | 0.9776 | nan | 0.9419 | 0.5322 | 0.9957 | nan | 0.8757 | 0.4869 | 0.9849 |
92
+ | 0.1491 | 10.33 | 620 | 0.0685 | 0.7805 | 0.8265 | 0.9760 | nan | 0.9320 | 0.5525 | 0.9950 | nan | 0.8753 | 0.4845 | 0.9816 |
93
+ | 0.0876 | 10.67 | 640 | 0.1347 | 0.7672 | 0.8134 | 0.9783 | nan | 0.9713 | 0.4762 | 0.9928 | nan | 0.8825 | 0.4317 | 0.9873 |
94
+ | 0.3076 | 11.0 | 660 | 0.0989 | 0.7737 | 0.8120 | 0.9745 | nan | 0.9152 | 0.5239 | 0.9970 | nan | 0.8700 | 0.4723 | 0.9788 |
95
+ | 0.08 | 11.33 | 680 | 0.0923 | 0.7914 | 0.8350 | 0.9766 | nan | 0.9212 | 0.5869 | 0.9968 | nan | 0.8780 | 0.5150 | 0.9813 |
96
+ | 0.0816 | 11.67 | 700 | 0.0878 | 0.7948 | 0.8392 | 0.9790 | nan | 0.9446 | 0.5773 | 0.9957 | nan | 0.8891 | 0.5101 | 0.9851 |
97
+ | 0.057 | 12.0 | 720 | 0.0824 | 0.8032 | 0.8537 | 0.9792 | nan | 0.9496 | 0.6175 | 0.9940 | nan | 0.8874 | 0.5368 | 0.9855 |
98
+ | 0.0881 | 12.33 | 740 | 0.0802 | 0.7694 | 0.8051 | 0.9766 | nan | 0.9349 | 0.4833 | 0.9970 | nan | 0.8792 | 0.4466 | 0.9825 |
99
+ | 0.1084 | 12.67 | 760 | 0.1033 | 0.8113 | 0.8662 | 0.9793 | nan | 0.9392 | 0.6646 | 0.9947 | nan | 0.8858 | 0.5626 | 0.9854 |
100
+ | 0.0703 | 13.0 | 780 | 0.0889 | 0.7795 | 0.8234 | 0.9765 | nan | 0.9385 | 0.5367 | 0.9949 | nan | 0.8775 | 0.4785 | 0.9824 |
101
+ | 0.1332 | 13.33 | 800 | 0.0803 | 0.7859 | 0.8332 | 0.9779 | nan | 0.9494 | 0.5562 | 0.9940 | nan | 0.8857 | 0.4881 | 0.9840 |
102
+ | 0.0872 | 13.67 | 820 | 0.1034 | 0.7741 | 0.8173 | 0.9788 | nan | 0.9666 | 0.4913 | 0.9939 | nan | 0.8860 | 0.4492 | 0.9871 |
103
+ | 0.0475 | 14.0 | 840 | 0.0728 | 0.7826 | 0.8295 | 0.9771 | nan | 0.9400 | 0.5536 | 0.9948 | nan | 0.8820 | 0.4829 | 0.9829 |
104
+ | 0.0569 | 14.33 | 860 | 0.0940 | 0.7794 | 0.8236 | 0.9786 | nan | 0.9671 | 0.5108 | 0.9930 | nan | 0.8824 | 0.4690 | 0.9867 |
105
+ | 0.7 | 14.67 | 880 | 0.0753 | 0.8024 | 0.8459 | 0.9797 | nan | 0.9443 | 0.5974 | 0.9961 | nan | 0.8905 | 0.5309 | 0.9858 |
106
+ | 0.0805 | 15.0 | 900 | 0.0738 | 0.8145 | 0.8720 | 0.9810 | nan | 0.9542 | 0.6679 | 0.9940 | nan | 0.8929 | 0.5622 | 0.9884 |
107
+ | 0.601 | 15.33 | 920 | 0.0970 | 0.8117 | 0.8614 | 0.9822 | nan | 0.9685 | 0.6216 | 0.9941 | nan | 0.8990 | 0.5461 | 0.9900 |
108
+ | 0.0844 | 15.67 | 940 | 0.0732 | 0.7691 | 0.8128 | 0.9755 | nan | 0.9479 | 0.4975 | 0.9930 | nan | 0.8731 | 0.4529 | 0.9814 |
109
+ | 0.1097 | 16.0 | 960 | 0.0622 | 0.8170 | 0.8721 | 0.9798 | nan | 0.9321 | 0.6881 | 0.9960 | nan | 0.8893 | 0.5761 | 0.9855 |
110
+ | 0.1446 | 16.33 | 980 | 0.0675 | 0.7983 | 0.8491 | 0.9797 | nan | 0.9621 | 0.5921 | 0.9930 | nan | 0.8900 | 0.5180 | 0.9868 |
111
+ | 0.3657 | 16.67 | 1000 | 0.0696 | 0.7900 | 0.8335 | 0.9777 | nan | 0.9371 | 0.5676 | 0.9957 | nan | 0.8836 | 0.5032 | 0.9831 |
112
+ | 0.2767 | 17.0 | 1020 | 0.0883 | 0.7599 | 0.8009 | 0.9715 | nan | 0.9063 | 0.5008 | 0.9955 | nan | 0.8580 | 0.4468 | 0.9750 |
113
+ | 0.1155 | 17.33 | 1040 | 0.0720 | 0.7929 | 0.8337 | 0.9797 | nan | 0.9527 | 0.5526 | 0.9958 | nan | 0.8882 | 0.5034 | 0.9870 |
114
+ | 0.0765 | 17.67 | 1060 | 0.0733 | 0.7843 | 0.8297 | 0.9767 | nan | 0.9395 | 0.5553 | 0.9944 | nan | 0.8798 | 0.4911 | 0.9821 |
115
+ | 0.1484 | 18.0 | 1080 | 0.0833 | 0.7904 | 0.8384 | 0.9788 | nan | 0.9595 | 0.5626 | 0.9932 | nan | 0.8870 | 0.4985 | 0.9856 |
116
+ | 0.1017 | 18.33 | 1100 | 0.0928 | 0.7866 | 0.8304 | 0.9799 | nan | 0.9707 | 0.5269 | 0.9935 | nan | 0.8888 | 0.4829 | 0.9881 |
117
+ | 0.0745 | 18.67 | 1120 | 0.0662 | 0.8084 | 0.8545 | 0.9806 | nan | 0.9472 | 0.6203 | 0.9961 | nan | 0.8947 | 0.5434 | 0.9870 |
118
+ | 0.093 | 19.0 | 1140 | 0.0673 | 0.8026 | 0.8510 | 0.9799 | nan | 0.9517 | 0.6065 | 0.9947 | nan | 0.8926 | 0.5290 | 0.9863 |
119
+ | 0.0475 | 19.33 | 1160 | 0.0818 | 0.7982 | 0.8447 | 0.9805 | nan | 0.9632 | 0.5768 | 0.9942 | nan | 0.8939 | 0.5130 | 0.9878 |
120
+ | 0.051 | 19.67 | 1180 | 0.0690 | 0.7931 | 0.8368 | 0.9800 | nan | 0.9578 | 0.5573 | 0.9951 | nan | 0.8924 | 0.4996 | 0.9873 |
121
+ | 0.0432 | 20.0 | 1200 | 0.0654 | 0.8022 | 0.8495 | 0.9793 | nan | 0.9379 | 0.6144 | 0.9962 | nan | 0.8897 | 0.5318 | 0.9851 |
122
+
123
+
124
  ### Framework versions
125
 
126
  - Transformers 4.36.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:154584a106a17067c73c36a8dde41cc5846a87c2a60a7fcd2e18b795e669ef12
3
  size 14886832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:342d3472cb353dd3eacb16dcdc2abaa434487502bce0aa9aa9363f6a575e2b30
3
  size 14886832
runs/Dec19_14-24-35_1244ddc99090/events.out.tfevents.1702995880.1244ddc99090.216.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e326e9e3afb3d76dd5c3a75aafc17579036d0e6e6edaf5001db07e720db143a
3
+ size 248037
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5587dff8309babfa030834215fcbd726a7143d50d32d1f2a5a52c0825a39e4c
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68776fbc7c1f4e3500ed832f9b1ed96b590212fba225c4dcd52cc8c22c975e0e
3
  size 4792