--- license: apache-2.0 base_model: ibaucells/RoBERTa-ca-CaWikiTC tags: - generated_from_trainer model-index: - name: test5_balanced_60ep results: [] --- # test5_balanced_60ep This model is a fine-tuned version of [ibaucells/RoBERTa-ca-CaWikiTC](https://huggingface.co/ibaucells/RoBERTa-ca-CaWikiTC) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.8209 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 10 - eval_batch_size: 10 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 60 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.8299 | 1.0 | 63 | 2.8361 | | 2.8512 | 2.0 | 126 | 2.8253 | | 2.8118 | 3.0 | 189 | 2.8376 | | 2.7238 | 4.0 | 252 | 2.7157 | | 2.5914 | 5.0 | 315 | 2.6045 | | 2.5527 | 6.0 | 378 | 2.5580 | | 2.4197 | 7.0 | 441 | 2.6009 | | 2.2561 | 8.0 | 504 | 2.3989 | | 1.9213 | 9.0 | 567 | 2.2129 | | 1.8147 | 10.0 | 630 | 2.1155 | | 1.4336 | 11.0 | 693 | 2.0509 | | 1.2948 | 12.0 | 756 | 1.8276 | | 0.9833 | 13.0 | 819 | 1.8245 | | 0.8577 | 14.0 | 882 | 1.7360 | | 0.665 | 15.0 | 945 | 1.8072 | | 0.5397 | 16.0 | 1008 | 1.8583 | | 0.2861 | 17.0 | 1071 | 1.9073 | | 0.3394 | 18.0 | 1134 | 2.0320 | | 0.2941 | 19.0 | 1197 | 2.1608 | | 0.1868 | 20.0 | 1260 | 2.1440 | | 0.175 | 21.0 | 1323 | 2.2099 | | 0.3357 | 22.0 | 1386 | 2.3903 | | 0.1311 | 23.0 | 1449 | 2.4006 | | 0.103 | 24.0 | 1512 | 2.4874 | | 0.1323 | 25.0 | 1575 | 2.5191 | | 0.0928 | 26.0 | 1638 | 2.3582 | | 0.0872 | 27.0 | 1701 | 2.4494 | | 0.1271 | 28.0 | 1764 | 2.5821 | | 0.1796 | 29.0 | 1827 | 2.6190 | | 0.0479 | 30.0 | 1890 | 2.5549 | | 0.0197 | 31.0 | 1953 | 2.8739 | | 0.0222 | 32.0 | 2016 | 2.7927 | | 0.019 | 33.0 | 2079 | 2.8467 | | 0.0946 | 34.0 | 2142 | 2.9086 | | 0.0084 | 35.0 | 2205 | 2.7976 | | 0.0539 | 36.0 | 2268 | 2.8300 | | 0.0553 | 37.0 | 2331 | 2.9293 | | 0.0124 | 38.0 | 2394 | 3.0047 | | 0.0447 | 39.0 | 2457 | 2.9724 | | 0.0363 | 40.0 | 2520 | 2.9775 | | 0.0056 | 41.0 | 2583 | 2.9972 | | 0.0072 | 42.0 | 2646 | 3.0162 | | 0.0091 | 43.0 | 2709 | 3.0348 | | 0.031 | 44.0 | 2772 | 3.0490 | | 0.0091 | 45.0 | 2835 | 3.0648 | | 0.029 | 46.0 | 2898 | 3.0730 | | 0.0076 | 47.0 | 2961 | 3.0866 | | 0.0115 | 48.0 | 3024 | 3.0979 | | 0.0047 | 49.0 | 3087 | 3.1102 | | 0.0079 | 50.0 | 3150 | 3.1242 | | 0.0039 | 51.0 | 3213 | 3.1299 | | 0.0043 | 52.0 | 3276 | 3.1386 | | 0.0322 | 53.0 | 3339 | 3.1478 | | 0.0041 | 54.0 | 3402 | 3.1478 | | 0.0289 | 55.0 | 3465 | 3.1570 | | 0.0038 | 56.0 | 3528 | 3.1661 | | 0.004 | 57.0 | 3591 | 3.1546 | | 0.0037 | 58.0 | 3654 | 3.1575 | | 0.0069 | 59.0 | 3717 | 3.1587 | | 0.0135 | 60.0 | 3780 | 3.1626 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.1.0+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2