lpcv_seg / README.md
gabryland's picture
End of training
57b1aad
---
license: other
base_model: google/deeplabv3_mobilenet_v2_1.0_513
tags:
- generated_from_trainer
model-index:
- name: lpcv_seg
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# lpcv_seg
This model is a fine-tuned version of [google/deeplabv3_mobilenet_v2_1.0_513](https://huggingface.co/google/deeplabv3_mobilenet_v2_1.0_513) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.7747
- Mean Iou: 0.3647
- Mean Accuracy: 0.4742
- Overall Accuracy: 0.7441
- Per Category Iou: [0.7508396000072782, 0.461963906773346, 0.41562431632163865, 0.2643890336606752, 0.20882280410355394, 0.21001420486640948, 0.45776923048905305, 0.6263265430221951, 0.5990132199881534, nan, 0.0, 0.4389678627683, 0.30687238462719907, 0.0]
- Per Category Accuracy: [0.8794950417196551, 0.5122587212045141, 0.47963636761360323, 0.28307093261894156, 0.22726847453969443, 0.817351469679542, 0.5476940642254209, 0.672940072704085, 0.8955934160757268, nan, 0.0, 0.44608425759467085, 0.40330970104326763, 0.0]
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 6e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 30
### Training results
| Training Loss | Epoch | Step | Validation Loss | Mean Iou | Mean Accuracy | Overall Accuracy | Per Category Iou | Per Category Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:-------------:|:----------------:|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|
| 1.0299 | 3.12 | 400 | 1.1445 | 0.1503 | 0.2293 | 0.6846 | [0.7016834423401825, 0.0, 0.28392253175773663, 0.09630649211770352, 0.16489618177561766, 0.05923195228949304, 0.0, 0.47146079872020313, 0.0, nan, 0.0, 0.002537363927640351, 0.17407988046100897, 0.0] | [0.891422563505799, 0.0, 0.7160794493379222, 0.101650104427665, 0.20155241518296468, 0.08566222638071554, 0.0, 0.7801995050268263, 0.0, nan, 0.0, 0.002538359085235112, 0.20133152963007558, 0.0] |
| 1.2946 | 6.25 | 800 | 0.9463 | 0.1514 | 0.2217 | 0.6990 | [0.7200410800543089, 0.025577984693804623, 0.272186783532416, 0.03900872477465889, 0.1420562982408946, 0.13672818472813886, 0.0, 0.3747836709762802, 1.4597782434650593e-05, nan, 0.0, 0.0, 0.2576138826730331, 0.0] | [0.9197800777719178, 0.025581056237151574, 0.3936555737035525, 0.03917595636589459, 0.1525480793198657, 0.16772954472705784, 0.0, 0.8514708072975876, 1.4597782434650593e-05, nan, 0.0, 0.0, 0.33275562889985905, 0.0] |
| 0.8869 | 9.38 | 1200 | 0.9005 | 0.2553 | 0.3463 | 0.7230 | [0.7311911080696011, 0.1716353741804461, 0.32989770226558307, 0.2484733739307282, 0.20678309581325105, 0.16925674004774066, 0.026560396219635213, 0.43309262889720945, 0.4791539954039438, nan, 0.0, 0.1996331281686418, 0.3234492318785704, 0.0] | [0.8817514590505282, 0.1734977169471996, 0.5536498133348888, 0.26666168042981364, 0.2199222972903758, 0.30495278898499545, 0.026735171458080034, 0.8794234634009936, 0.5647817144933271, nan, 0.0, 0.19998287851911137, 0.43016977002783463, 0.0] |
| 0.8056 | 12.5 | 1600 | 0.7305 | 0.3106 | 0.3897 | 0.7595 | [0.7538928589848243, 0.25918801874572633, 0.4203485338620474, 0.2917137053557758, 0.22289503658523782, 0.1500749815626878, 0.08861529960760275, 0.624086456268012, 0.48451939021185103, nan, 0.0, 0.3968018568653773, 0.34541558718052773, 0.0] | [0.9269385933729979, 0.2666068145623601, 0.5690834086799277, 0.323866282726429, 0.23199974868520706, 0.27844356662715314, 0.08945714746397179, 0.7853752936779852, 0.7445501612243971, nan, 0.0, 0.4213536720599331, 0.42861443009823574, 0.0] |
| 0.7534 | 15.62 | 2000 | 0.8091 | 0.2762 | 0.3774 | 0.7408 | [0.7450525081446471, 0.08805507548598117, 0.39296675038083556, 0.17888143033371248, 0.23466648108570068, 0.16079006682931019, 0.03714210117676656, 0.6219865494576018, 0.5133743295405404, nan, 0.0, 0.25318448016278294, 0.3644475822818699, 0.0] | [0.8925789949539187, 0.08813412970942185, 0.6256571559820335, 0.18979883934004887, 0.24639981315291482, 0.4060874245051725, 0.03731593695118816, 0.8003825980603533, 0.8838065177476595, nan, 0.0, 0.25812424222515545, 0.47821461707899227, 0.0] |
| 0.909 | 18.75 | 2400 | 0.7351 | 0.3258 | 0.4109 | 0.7698 | [0.7635658546626002, 0.40487164496967293, 0.3833268397798514, 0.2752089531995694, 0.2867291055853458, 0.019354003421208514, 0.16890635042927862, 0.5748136637050552, 0.6549805047839075, nan, 0.0, 0.41018196642738697, 0.29410209406700305, 0.0] | [0.936836241285118, 0.4435412622122596, 0.6115797993350055, 0.29433416172859234, 0.3084226650431367, 0.01980599690692518, 0.1729223045848531, 0.8637196946532967, 0.9040471540812156, nan, 0.0, 0.44469064868513, 0.3419148529297193, 0.0] |
| 0.6622 | 21.88 | 2800 | 0.6963 | 0.3398 | 0.3986 | 0.7827 | [0.7663761109958838, 0.1332027535954636, 0.4428707076945945, 0.35775923195424675, 0.3733786220955952, 0.24949901778334058, 0.07535404639899677, 0.6174965473057359, 0.5443824880761584, nan, 0.0, 0.4962856002887922, 0.3608184550398603, 0.0] | [0.9493652245312661, 0.13338632341294235, 0.6300075103540804, 0.4173944872618647, 0.41213577281689057, 0.31108969330720987, 0.0759392172499597, 0.6593100168039995, 0.5955570838172228, nan, 0.0, 0.5200351189445205, 0.4777004199984418, 0.0] |
| 1.1202 | 25.0 | 3200 | 1.6352 | 0.1684 | 0.2322 | 0.7026 | [0.6963172586859363, 0.011559652438449069, 0.22669839158942193, 0.05922256175168542, 0.06532726727054458, 0.0023358067869460536, 0.0, 0.509361415684211, 0.38727970470408707, nan, 0.0, 0.09125458640751972, 0.13993367687629527, 0.0] | [0.9487412199772098, 0.011559652438449069, 0.24983867321938985, 0.05932148648786475, 0.06540672362680026, 0.003050864548414594, 0.0, 0.7020197466594538, 0.7380590139684559, nan, 0.0, 0.09130527401337467, 0.14901587211649467, 0.0] |
| 0.6529 | 28.12 | 3600 | 0.7747 | 0.3647 | 0.4742 | 0.7441 | [0.7508396000072782, 0.461963906773346, 0.41562431632163865, 0.2643890336606752, 0.20882280410355394, 0.21001420486640948, 0.45776923048905305, 0.6263265430221951, 0.5990132199881534, nan, 0.0, 0.4389678627683, 0.30687238462719907, 0.0] | [0.8794950417196551, 0.5122587212045141, 0.47963636761360323, 0.28307093261894156, 0.22726847453969443, 0.817351469679542, 0.5476940642254209, 0.672940072704085, 0.8955934160757268, nan, 0.0, 0.44608425759467085, 0.40330970104326763, 0.0] |
### Framework versions
- Transformers 4.32.1
- Pytorch 2.0.1+cu118
- Datasets 2.14.4
- Tokenizers 0.13.3