--- license: mit tags: - generated_from_trainer model-index: - name: predict-perception-xlmr-cause-object results: [] --- # predict-perception-xlmr-cause-object This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.3069 - Rmse: 0.8927 - Rmse Cause::a Causata da un oggetto (es. una pistola): 0.8927 - Mae: 0.5854 - Mae Cause::a Causata da un oggetto (es. una pistola): 0.5854 - R2: 0.5410 - R2 Cause::a Causata da un oggetto (es. una pistola): 0.5410 - Cos: 0.4783 - Pair: 0.0 - Rank: 0.5 - Neighbors: 0.6177 - Rsa: nan ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 20 - eval_batch_size: 8 - seed: 1996 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 30 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rmse | Rmse Cause::a Causata da un oggetto (es. una pistola) | Mae | Mae Cause::a Causata da un oggetto (es. una pistola) | R2 | R2 Cause::a Causata da un oggetto (es. una pistola) | Cos | Pair | Rank | Neighbors | Rsa | |:-------------:|:-----:|:----:|:---------------:|:------:|:-----------------------------------------------------:|:------:|:----------------------------------------------------:|:-------:|:---------------------------------------------------:|:-------:|:----:|:----:|:---------:|:---:| | 1.0329 | 1.0 | 15 | 0.8168 | 1.4564 | 1.4564 | 1.2947 | 1.2947 | -0.2216 | -0.2216 | -0.5652 | 0.0 | 0.5 | 0.5993 | nan | | 1.0096 | 2.0 | 30 | 0.7432 | 1.3893 | 1.3893 | 1.1883 | 1.1883 | -0.1116 | -0.1116 | -0.3913 | 0.0 | 0.5 | 0.6499 | nan | | 0.9323 | 3.0 | 45 | 0.6879 | 1.3366 | 1.3366 | 1.1054 | 1.1054 | -0.0289 | -0.0289 | -0.1304 | 0.0 | 0.5 | 0.5471 | nan | | 0.8636 | 4.0 | 60 | 0.6378 | 1.2870 | 1.2870 | 1.0477 | 1.0477 | 0.0461 | 0.0461 | 0.2174 | 0.0 | 0.5 | 0.3007 | nan | | 0.8041 | 5.0 | 75 | 0.5494 | 1.1945 | 1.1945 | 0.9499 | 0.9499 | 0.1783 | 0.1783 | 0.6522 | 0.0 | 0.5 | 0.6695 | nan | | 0.7413 | 6.0 | 90 | 0.5526 | 1.1980 | 1.1980 | 0.9503 | 0.9503 | 0.1735 | 0.1735 | 0.5652 | 0.0 | 0.5 | 0.3898 | nan | | 0.6397 | 7.0 | 105 | 0.4726 | 1.1078 | 1.1078 | 0.7826 | 0.7826 | 0.2932 | 0.2932 | 0.5652 | 0.0 | 0.5 | 0.3257 | nan | | 0.5556 | 8.0 | 120 | 0.7728 | 1.4167 | 1.4167 | 1.1528 | 1.1528 | -0.1558 | -0.1558 | 0.1304 | 0.0 | 0.5 | 0.4027 | nan | | 0.4972 | 9.0 | 135 | 0.4375 | 1.0659 | 1.0659 | 0.7577 | 0.7577 | 0.3457 | 0.3457 | 0.5652 | 0.0 | 0.5 | 0.5683 | nan | | 0.3691 | 10.0 | 150 | 0.4990 | 1.1383 | 1.1383 | 0.8272 | 0.8272 | 0.2537 | 0.2537 | 0.4783 | 0.0 | 0.5 | 0.4781 | nan | | 0.3381 | 11.0 | 165 | 0.4401 | 1.0690 | 1.0690 | 0.7319 | 0.7319 | 0.3418 | 0.3418 | 0.5652 | 0.0 | 0.5 | 0.5683 | nan | | 0.2966 | 12.0 | 180 | 0.4794 | 1.1158 | 1.1158 | 0.7835 | 0.7835 | 0.2830 | 0.2830 | 0.5652 | 0.0 | 0.5 | 0.5683 | nan | | 0.2324 | 13.0 | 195 | 0.4013 | 1.0208 | 1.0208 | 0.6873 | 0.6873 | 0.3998 | 0.3998 | 0.4783 | 0.0 | 0.5 | 0.5796 | nan | | 0.1848 | 14.0 | 210 | 0.4305 | 1.0574 | 1.0574 | 0.7372 | 0.7372 | 0.3561 | 0.3561 | 0.4783 | 0.0 | 0.5 | 0.5796 | nan | | 0.1621 | 15.0 | 225 | 0.3652 | 0.9738 | 0.9738 | 0.6164 | 0.6164 | 0.4538 | 0.4538 | 0.4783 | 0.0 | 0.5 | 0.6177 | nan | | 0.1762 | 16.0 | 240 | 0.3335 | 0.9307 | 0.9307 | 0.6458 | 0.6458 | 0.5012 | 0.5012 | 0.4783 | 0.0 | 0.5 | 0.5796 | nan | | 0.1404 | 17.0 | 255 | 0.3420 | 0.9424 | 0.9424 | 0.6599 | 0.6599 | 0.4886 | 0.4886 | 0.3913 | 0.0 | 0.5 | 0.5831 | nan | | 0.1379 | 18.0 | 270 | 0.2853 | 0.8608 | 0.8608 | 0.6063 | 0.6063 | 0.5733 | 0.5733 | 0.3913 | 0.0 | 0.5 | 0.5831 | nan | | 0.1322 | 19.0 | 285 | 0.3261 | 0.9203 | 0.9203 | 0.6548 | 0.6548 | 0.5123 | 0.5123 | 0.4783 | 0.0 | 0.5 | 0.5796 | nan | | 0.1067 | 20.0 | 300 | 0.3328 | 0.9296 | 0.9296 | 0.5535 | 0.5535 | 0.5023 | 0.5023 | 0.6522 | 0.0 | 0.5 | 0.6695 | nan | | 0.1038 | 21.0 | 315 | 0.3066 | 0.8924 | 0.8924 | 0.6266 | 0.6266 | 0.5414 | 0.5414 | 0.4783 | 0.0 | 0.5 | 0.5796 | nan | | 0.094 | 22.0 | 330 | 0.2924 | 0.8714 | 0.8714 | 0.5792 | 0.5792 | 0.5626 | 0.5626 | 0.4783 | 0.0 | 0.5 | 0.6177 | nan | | 0.1078 | 23.0 | 345 | 0.3161 | 0.9060 | 0.9060 | 0.6022 | 0.6022 | 0.5272 | 0.5272 | 0.3913 | 0.0 | 0.5 | 0.5831 | nan | | 0.0976 | 24.0 | 360 | 0.3118 | 0.8998 | 0.8998 | 0.6011 | 0.6011 | 0.5337 | 0.5337 | 0.3913 | 0.0 | 0.5 | 0.5831 | nan | | 0.0911 | 25.0 | 375 | 0.3123 | 0.9005 | 0.9005 | 0.5811 | 0.5811 | 0.5330 | 0.5330 | 0.4783 | 0.0 | 0.5 | 0.6177 | nan | | 0.1039 | 26.0 | 390 | 0.3122 | 0.9005 | 0.9005 | 0.5956 | 0.5956 | 0.5330 | 0.5330 | 0.4783 | 0.0 | 0.5 | 0.6177 | nan | | 0.0775 | 27.0 | 405 | 0.3191 | 0.9103 | 0.9103 | 0.6124 | 0.6124 | 0.5228 | 0.5228 | 0.3913 | 0.0 | 0.5 | 0.5831 | nan | | 0.0789 | 28.0 | 420 | 0.3135 | 0.9023 | 0.9023 | 0.5825 | 0.5825 | 0.5311 | 0.5311 | 0.4783 | 0.0 | 0.5 | 0.6177 | nan | | 0.0778 | 29.0 | 435 | 0.3075 | 0.8936 | 0.8936 | 0.5837 | 0.5837 | 0.5401 | 0.5401 | 0.4783 | 0.0 | 0.5 | 0.6177 | nan | | 0.082 | 30.0 | 450 | 0.3069 | 0.8927 | 0.8927 | 0.5854 | 0.5854 | 0.5410 | 0.5410 | 0.4783 | 0.0 | 0.5 | 0.6177 | nan | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.2+cu113 - Datasets 1.18.3 - Tokenizers 0.11.0