--- license: other base_model: nvidia/mit-b5 tags: - vision - image-segmentation - generated_from_trainer model-index: - name: segformer-b5-finetuned-Hiking results: [] --- # segformer-b5-finetuned-Hiking This model is a fine-tuned version of [nvidia/mit-b5](https://huggingface.co/nvidia/mit-b5) on the twdent/Hiking dataset. It achieves the following results on the evaluation set: - Loss: 0.1401 - Mean Iou: 0.6237 - Mean Accuracy: 0.9673 - Overall Accuracy: 0.9683 - Accuracy Unlabeled: nan - Accuracy Traversable: 0.9641 - Accuracy Non-traversable: 0.9705 - Iou Unlabeled: 0.0 - Iou Traversable: 0.9178 - Iou Non-traversable: 0.9532 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 6e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 50 ### Training results | Training Loss | Epoch | Step | Validation Loss | Mean Iou | Mean Accuracy | Overall Accuracy | Accuracy Unlabeled | Accuracy Traversable | Accuracy Non-traversable | Iou Unlabeled | Iou Traversable | Iou Non-traversable | |:-------------:|:-----:|:----:|:---------------:|:--------:|:-------------:|:----------------:|:------------------:|:--------------------:|:------------------------:|:-------------:|:---------------:|:-------------------:| | 0.2675 | 1.33 | 20 | 0.2742 | 0.6058 | 0.9616 | 0.9550 | nan | 0.9826 | 0.9406 | 0.0 | 0.8853 | 0.9321 | | 0.1418 | 2.67 | 40 | 0.1827 | 0.6073 | 0.9562 | 0.9566 | nan | 0.9548 | 0.9575 | 0.0 | 0.8858 | 0.9360 | | 0.0949 | 4.0 | 60 | 0.1561 | 0.6002 | 0.9382 | 0.9523 | nan | 0.8931 | 0.9832 | 0.0 | 0.8692 | 0.9314 | | 0.0684 | 5.33 | 80 | 0.1364 | 0.6135 | 0.9556 | 0.9614 | nan | 0.9369 | 0.9742 | 0.0 | 0.8967 | 0.9437 | | 0.0627 | 6.67 | 100 | 0.1289 | 0.6122 | 0.9506 | 0.9610 | nan | 0.9177 | 0.9836 | 0.0 | 0.8928 | 0.9438 | | 0.0625 | 8.0 | 120 | 0.1097 | 0.6208 | 0.9610 | 0.9658 | nan | 0.9458 | 0.9762 | 0.0 | 0.9113 | 0.9510 | | 0.0371 | 9.33 | 140 | 0.1361 | 0.6130 | 0.9551 | 0.9610 | nan | 0.9361 | 0.9741 | 0.0 | 0.8959 | 0.9431 | | 0.0409 | 10.67 | 160 | 0.1239 | 0.6194 | 0.9615 | 0.9653 | nan | 0.9494 | 0.9737 | 0.0 | 0.9086 | 0.9494 | | 0.0457 | 12.0 | 180 | 0.0993 | 0.6281 | 0.9715 | 0.9713 | nan | 0.9723 | 0.9707 | 0.0 | 0.9264 | 0.9579 | | 0.0368 | 13.33 | 200 | 0.1354 | 0.6146 | 0.9563 | 0.9617 | nan | 0.9389 | 0.9737 | 0.0 | 0.8993 | 0.9446 | | 0.0667 | 14.67 | 220 | 0.1208 | 0.6171 | 0.9565 | 0.9644 | nan | 0.9316 | 0.9815 | 0.0 | 0.9032 | 0.9482 | | 0.029 | 16.0 | 240 | 0.0946 | 0.6291 | 0.9695 | 0.9724 | nan | 0.9606 | 0.9785 | 0.0 | 0.9276 | 0.9596 | | 0.0467 | 17.33 | 260 | 0.1188 | 0.6224 | 0.9655 | 0.9676 | nan | 0.9589 | 0.9721 | 0.0 | 0.9151 | 0.9522 | | 0.0449 | 18.67 | 280 | 0.1201 | 0.6212 | 0.9638 | 0.9667 | nan | 0.9545 | 0.9731 | 0.0 | 0.9125 | 0.9511 | | 0.0353 | 20.0 | 300 | 0.1285 | 0.6234 | 0.9687 | 0.9681 | nan | 0.9706 | 0.9668 | 0.0 | 0.9174 | 0.9527 | | 0.025 | 21.33 | 320 | 0.1292 | 0.6204 | 0.9641 | 0.9659 | nan | 0.9582 | 0.9699 | 0.0 | 0.9114 | 0.9500 | | 0.0244 | 22.67 | 340 | 0.1352 | 0.6208 | 0.9665 | 0.9664 | nan | 0.9667 | 0.9662 | 0.0 | 0.9124 | 0.9501 | | 0.035 | 24.0 | 360 | 0.1260 | 0.6252 | 0.9699 | 0.9693 | nan | 0.9718 | 0.9681 | 0.0 | 0.9211 | 0.9544 | | 0.0295 | 25.33 | 380 | 0.1190 | 0.6244 | 0.9669 | 0.9688 | nan | 0.9607 | 0.9730 | 0.0 | 0.9190 | 0.9543 | | 0.032 | 26.67 | 400 | 0.1258 | 0.6253 | 0.9694 | 0.9695 | nan | 0.9693 | 0.9695 | 0.0 | 0.9211 | 0.9547 | | 0.0241 | 28.0 | 420 | 0.1255 | 0.6230 | 0.9658 | 0.9678 | nan | 0.9593 | 0.9723 | 0.0 | 0.9164 | 0.9527 | | 0.0246 | 29.33 | 440 | 0.1273 | 0.6238 | 0.9675 | 0.9683 | nan | 0.9651 | 0.9699 | 0.0 | 0.9179 | 0.9534 | | 0.0214 | 30.67 | 460 | 0.1321 | 0.6233 | 0.9670 | 0.9675 | nan | 0.9652 | 0.9687 | 0.0 | 0.9171 | 0.9527 | | 0.0236 | 32.0 | 480 | 0.1289 | 0.6241 | 0.9687 | 0.9685 | nan | 0.9695 | 0.9679 | 0.0 | 0.9189 | 0.9534 | | 0.0238 | 33.33 | 500 | 0.1309 | 0.6234 | 0.9664 | 0.9680 | nan | 0.9612 | 0.9716 | 0.0 | 0.9172 | 0.9529 | | 0.0204 | 34.67 | 520 | 0.1271 | 0.6249 | 0.9681 | 0.9693 | nan | 0.9643 | 0.9719 | 0.0 | 0.9201 | 0.9547 | | 0.0243 | 36.0 | 540 | 0.1264 | 0.6248 | 0.9679 | 0.9693 | nan | 0.9636 | 0.9723 | 0.0 | 0.9196 | 0.9547 | | 0.0259 | 37.33 | 560 | 0.1305 | 0.6226 | 0.9656 | 0.9679 | nan | 0.9582 | 0.9730 | 0.0 | 0.9154 | 0.9525 | | 0.0341 | 38.67 | 580 | 0.1277 | 0.6245 | 0.9674 | 0.9690 | nan | 0.9623 | 0.9725 | 0.0 | 0.9192 | 0.9543 | | 0.0275 | 40.0 | 600 | 0.1369 | 0.6221 | 0.9653 | 0.9672 | nan | 0.9590 | 0.9715 | 0.0 | 0.9147 | 0.9516 | | 0.0303 | 41.33 | 620 | 0.1380 | 0.6235 | 0.9674 | 0.9681 | nan | 0.9650 | 0.9698 | 0.0 | 0.9175 | 0.9530 | | 0.0207 | 42.67 | 640 | 0.1389 | 0.6237 | 0.9677 | 0.9682 | nan | 0.9662 | 0.9692 | 0.0 | 0.9180 | 0.9531 | | 0.0231 | 44.0 | 660 | 0.1369 | 0.6243 | 0.9679 | 0.9688 | nan | 0.9652 | 0.9707 | 0.0 | 0.9190 | 0.9538 | | 0.0249 | 45.33 | 680 | 0.1379 | 0.6237 | 0.9672 | 0.9683 | nan | 0.9640 | 0.9705 | 0.0 | 0.9179 | 0.9532 | | 0.0382 | 46.67 | 700 | 0.1384 | 0.6239 | 0.9677 | 0.9685 | nan | 0.9650 | 0.9704 | 0.0 | 0.9182 | 0.9534 | | 0.0238 | 48.0 | 720 | 0.1420 | 0.6230 | 0.9668 | 0.9677 | nan | 0.9640 | 0.9697 | 0.0 | 0.9166 | 0.9524 | | 0.0212 | 49.33 | 740 | 0.1401 | 0.6237 | 0.9673 | 0.9683 | nan | 0.9641 | 0.9705 | 0.0 | 0.9178 | 0.9532 | ### Framework versions - Transformers 4.35.0.dev0 - Pytorch 2.1.0+cu118 - Datasets 2.14.5 - Tokenizers 0.14.0