--- language: - eng license: wtfpl tags: - multilabel-image-classification - multilabel - generated_from_trainer base_model: facebook/dinov2-large model-index: - name: dinov2-large-2024_05_23-drone_batch-size512_epochs50_freeze results: [] --- DinoVd'eau is a fine-tuned version of [facebook/dinov2-large](https://huggingface.co/facebook/dinov2-large). It achieves the following results on the test set: - Loss: 0.2361 - F1 Micro: 0.7694 - F1 Macro: 0.4048 - Roc Auc: 0.8448 - Accuracy: 0.1449 --- # Model description DinoVd'eau is a model built on top of dinov2 model for underwater multilabel image classification.The classification head is a combination of linear, ReLU, batch normalization, and dropout layers. The source code for training the model can be found in this [Git repository](https://github.com/SeatizenDOI/DinoVdeau). - **Developed by:** [lombardata](https://huggingface.co/lombardata), credits to [César Leblanc](https://huggingface.co/CesarLeblanc) and [Victor Illien](https://huggingface.co/groderg) --- # Intended uses & limitations You can use the raw model for classify diverse marine species, encompassing coral morphotypes classes taken from the Global Coral Reef Monitoring Network (GCRMN), habitats classes and seagrass species. --- # Training and evaluation data Details on the number of images for each class are given in the following table: | Class | train | val | test | Total | |:-------------------------|--------:|------:|-------:|--------:| | Acropore_branched | 1575 | 562 | 565 | 2702 | | Acropore_digitised | 1020 | 356 | 370 | 1746 | | Acropore_sub_massive | 198 | 56 | 60 | 314 | | Acropore_tabular | 659 | 248 | 238 | 1145 | | Algae_assembly | 7175 | 2447 | 2430 | 12052 | | Algae_drawn_up | 439 | 156 | 156 | 751 | | Algae_limestone | 4694 | 1576 | 1523 | 7793 | | Algae_sodding | 7151 | 2460 | 2467 | 12078 | | Bleached_coral | 352 | 162 | 150 | 664 | | Dead_coral | 4615 | 1589 | 1553 | 7757 | | Living_coral | 85 | 37 | 28 | 150 | | Millepore | 860 | 287 | 313 | 1460 | | No_acropore_encrusting | 1978 | 675 | 667 | 3320 | | No_acropore_massive | 4539 | 1613 | 1585 | 7737 | | No_acropore_sub_massive | 3696 | 1245 | 1252 | 6193 | | Rock | 10810 | 3735 | 3718 | 18263 | | Rubble | 9948 | 3429 | 3403 | 16780 | | Sand | 10976 | 3659 | 3659 | 18294 | | Sea_urchins | 400 | 147 | 135 | 682 | | Sponge | 207 | 59 | 56 | 322 | | Thalassodendron_ciliatum | 216 | 74 | 70 | 360 | | Useless | 89 | 30 | 30 | 149 | --- # Training procedure ## Training hyperparameters The following hyperparameters were used during training: - **Number of Epochs**: 50 - **Learning Rate**: 0.001 - **Train Batch Size**: 512 - **Eval Batch Size**: 512 - **Optimizer**: Adam - **LR Scheduler Type**: ReduceLROnPlateau with a patience of 5 epochs and a factor of 0.1 - **Freeze Encoder**: Yes - **Data Augmentation**: Yes ## Data Augmentation Data were augmented using the following transformations : Train Transforms - **PreProcess**: No additional parameters - **Resize**: probability=1.00 - **RandomHorizontalFlip**: probability=0.25 - **RandomVerticalFlip**: probability=0.25 - **ColorJiggle**: probability=0.25 - **RandomPerspective**: probability=0.25 - **Normalize**: probability=1.00 Val Transforms - **PreProcess**: No additional parameters - **Resize**: probability=1.00 - **Normalize**: probability=1.00 ## Training results Epoch | Validation Loss | Accuracy | F1 Macro | F1 Micro | Learning Rate --- | --- | --- | --- | --- | --- 1.0 | 0.5951732397079468 | 0.012405938580435224 | 0.5738973203699311 | 0.40667627783698285 | 0.001 2.0 | 0.4730209410190582 | 0.06975798250966037 | 0.7307120964254151 | 0.4367882492755507 | 0.001 3.0 | 0.3240152895450592 | 0.10738255033557047 | 0.7498981835953409 | 0.37702459211637257 | 0.001 4.0 | 0.2770342230796814 | 0.11795810453528574 | 0.7521195160095482 | 0.3710481900670742 | 0.001 5.0 | 0.25879302620887756 | 0.11958511287370348 | 0.7507292550220328 | 0.3714736793659693 | 0.001 6.0 | 0.25328728556632996 | 0.12182224933902787 | 0.7520252586099456 | 0.36304822534346387 | 0.001 7.0 | 0.25132349133491516 | 0.11531421598535692 | 0.7517183920016662 | 0.3646331511607325 | 0.001 8.0 | 0.2507544159889221 | 0.12283912955053895 | 0.7576399892988702 | 0.38940077262215617 | 0.001 9.0 | 0.24785615503787994 | 0.12751677852348994 | 0.7549859932265752 | 0.38290945223752887 | 0.001 10.0 | 0.2480766475200653 | 0.12649989831197886 | 0.7583163191651716 | 0.37973264961121395 | 0.001 11.0 | 0.24667006731033325 | 0.12426276184665447 | 0.7600958878849345 | 0.3964288145693209 | 0.001 12.0 | 0.2459569126367569 | 0.12507626601586333 | 0.7564640698455339 | 0.3733203958034908 | 0.001 13.0 | 0.245611771941185 | 0.1297539149888143 | 0.7581923944769908 | 0.38618999344086086 | 0.001 14.0 | 0.24649737775325775 | 0.13707545251169412 | 0.7526021832952525 | 0.37084554766098704 | 0.001 15.0 | 0.24523988366127014 | 0.1271100264388855 | 0.7540528606572888 | 0.37953234953900117 | 0.001 16.0 | 0.24370642006397247 | 0.1293471629042099 | 0.7597242635642867 | 0.39042586476441543 | 0.001 17.0 | 0.24466517567634583 | 0.13158429936953428 | 0.7525727259224682 | 0.38542350135117487 | 0.001 17.857142857142858 | N/A | N/A | N/A | N/A | 0.001 18.0 | 0.24544650316238403 | 0.133211307707952 | 0.7534316217590239 | 0.35783734462173733 | 0.001 19.0 | 0.2440878450870514 | 0.13239780353874314 | 0.7568417082268136 | 0.3694145346248099 | 0.001 20.0 | 0.2453632354736328 | 0.13605857230018303 | 0.750895096799091 | 0.3768127127776539 | 0.001 21.0 | 0.24377579987049103 | 0.12487288997356112 | 0.760243826841616 | 0.38961590782494593 | 0.001 22.0 | 0.24192409217357635 | 0.13016066707341875 | 0.7576183975637929 | 0.3715634230883189 | 0.001 23.0 | 0.24348826706409454 | 0.12649989831197886 | 0.7628996647313762 | 0.3880375815747224 | 0.001 24.0 | 0.2413305789232254 | 0.1342281879194631 | 0.7561114991428027 | 0.3896884130115941 | 0.001 25.0 | 0.24189460277557373 | 0.1297539149888143 | 0.7599182173024102 | 0.38267978517684004 | 0.001 26.0 | 0.2437727451324463 | 0.12670327435428105 | 0.7593076827294236 | 0.3971421437602147 | 0.001 27.0 | 0.24182096123695374 | 0.1309741712426276 | 0.761437908496732 | 0.38383597863653807 | 0.001 28.0 | 0.24316559731960297 | 0.13341468375025423 | 0.7498440155769273 | 0.3792682503180625 | 0.001 29.0 | 0.24201267957687378 | 0.1366687004270897 | 0.7621594930458399 | 0.39596794972011545 | 0.001 30.0 | 0.2406790852546692 | 0.14236322961155176 | 0.7619565217391304 | 0.38596411111358153 | 0.001 31.0 | 0.24222084879875183 | 0.13280455562334756 | 0.7611869607298037 | 0.3928781445591724 | 0.001 32.0 | 0.24304261803627014 | 0.13117754728492984 | 0.7516135926480015 | 0.3912203123758987 | 0.001 33.0 | 0.24139608442783356 | 0.13016066707341875 | 0.758885526453094 | 0.38844227395152936 | 0.001 34.0 | 0.24039919674396515 | 0.1354484441732764 | 0.7624706542289075 | 0.4037409737349212 | 0.001 35.0 | 0.24134761095046997 | 0.12995729103111653 | 0.7601615858737297 | 0.3973020120442106 | 0.001 35.714285714285715 | N/A | N/A | N/A | N/A | 0.001 36.0 | 0.24192169308662415 | 0.13565182021557862 | 0.7622066694112803 | 0.38761085480429286 | 0.001 37.0 | 0.2399486005306244 | 0.1342281879194631 | 0.7598352387357096 | 0.3992187594370792 | 0.001 38.0 | 0.24004822969436646 | 0.13300793166564978 | 0.7607364527387098 | 0.3932700433432016 | 0.001 39.0 | 0.24091550707817078 | 0.13890583689241406 | 0.7619087275149901 | 0.4007929579258356 | 0.001 40.0 | 0.23991511762142181 | 0.1354484441732764 | 0.76 | 0.39250375468507387 | 0.001 41.0 | 0.2422637641429901 | 0.12487288997356112 | 0.7639710985018574 | 0.40608061408264917 | 0.001 42.0 | 0.24256455898284912 | 0.1309741712426276 | 0.7568840806286871 | 0.4005098857996497 | 0.001 43.0 | 0.23922023177146912 | 0.13361805979255645 | 0.7594289817122102 | 0.4007981173529554 | 0.001 44.0 | 0.24184103310108185 | 0.13036404311572097 | 0.7576905272903253 | 0.40641694858015515 | 0.001 45.0 | 0.24105145037174225 | 0.13788895668090298 | 0.7591085068536151 | 0.39055068831340933 | 0.001 46.0 | 0.23963303864002228 | 0.13626194834248526 | 0.7653508320819534 | 0.4106196361694743 | 0.001 47.0 | 0.23957742750644684 | 0.13260117958104536 | 0.7575076348829317 | 0.3967990889217657 | 0.001 48.0 | 0.24231907725334167 | 0.12873703477730322 | 0.7563947423325684 | 0.38777990454974365 | 0.001 49.0 | 0.23978127539157867 | 0.13300793166564978 | 0.7608376348147216 | 0.40266317126303974 | 0.001 50.0 | 0.23673731088638306 | 0.14236322961155176 | 0.7652267908369019 | 0.4087415721658059 | 0.0001 --- # CO2 Emissions The estimated CO2 emissions for training this model are documented below: - **Emissions**: 0.02562211166966913 grams of CO2 - **Source**: Code Carbon - **Training Type**: fine-tuning - **Geographical Location**: Brest, France - **Hardware Used**: NVIDIA Tesla V100 PCIe 32 Go --- # Framework Versions - **Transformers**: 4.41.0 - **Pytorch**: 2.3.0+cu118 - **Datasets**: 2.19.1 - **Tokenizers**: 0.19.1