ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k5_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5970
  • Qwk: 0.6711
  • Mse: 0.5970
  • Rmse: 0.7727

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1333 2 4.0121 0.0069 4.0121 2.0030
No log 0.2667 4 2.1405 0.0450 2.1405 1.4630
No log 0.4 6 1.4755 0.0408 1.4755 1.2147
No log 0.5333 8 1.0516 0.2734 1.0516 1.0255
No log 0.6667 10 1.0824 0.1643 1.0824 1.0404
No log 0.8 12 1.2759 -0.1333 1.2759 1.1295
No log 0.9333 14 1.4281 -0.0777 1.4281 1.1951
No log 1.0667 16 1.5161 -0.0709 1.5161 1.2313
No log 1.2 18 1.2867 -0.0394 1.2867 1.1343
No log 1.3333 20 1.1995 0.0967 1.1995 1.0952
No log 1.4667 22 1.2540 -0.0032 1.2540 1.1198
No log 1.6 24 1.1729 0.0820 1.1729 1.0830
No log 1.7333 26 1.0244 0.3326 1.0244 1.0121
No log 1.8667 28 1.0345 0.2697 1.0345 1.0171
No log 2.0 30 1.0688 0.2236 1.0688 1.0338
No log 2.1333 32 1.0751 0.2390 1.0751 1.0369
No log 2.2667 34 1.0323 0.2416 1.0323 1.0160
No log 2.4 36 1.0861 0.1436 1.0861 1.0422
No log 2.5333 38 1.0907 0.1436 1.0907 1.0444
No log 2.6667 40 0.9866 0.2416 0.9866 0.9933
No log 2.8 42 0.9687 0.2265 0.9687 0.9842
No log 2.9333 44 0.9995 0.3372 0.9995 0.9997
No log 3.0667 46 1.0022 0.3687 1.0022 1.0011
No log 3.2 48 1.1401 0.3863 1.1401 1.0678
No log 3.3333 50 1.2943 0.2993 1.2943 1.1377
No log 3.4667 52 1.1018 0.3502 1.1018 1.0497
No log 3.6 54 0.8783 0.4494 0.8783 0.9372
No log 3.7333 56 1.0112 0.3449 1.0112 1.0056
No log 3.8667 58 1.3694 0.0840 1.3694 1.1702
No log 4.0 60 1.4294 0.1379 1.4294 1.1956
No log 4.1333 62 1.1873 0.2857 1.1873 1.0896
No log 4.2667 64 0.8918 0.4265 0.8918 0.9443
No log 4.4 66 0.7505 0.5010 0.7505 0.8663
No log 4.5333 68 0.7444 0.5375 0.7444 0.8628
No log 4.6667 70 0.7857 0.5199 0.7857 0.8864
No log 4.8 72 0.9551 0.3624 0.9551 0.9773
No log 4.9333 74 0.8291 0.5586 0.8291 0.9106
No log 5.0667 76 0.7305 0.4850 0.7305 0.8547
No log 5.2 78 0.8553 0.5230 0.8553 0.9248
No log 5.3333 80 0.7356 0.5163 0.7356 0.8577
No log 5.4667 82 0.6618 0.5235 0.6618 0.8135
No log 5.6 84 0.7038 0.5910 0.7038 0.8389
No log 5.7333 86 0.6963 0.4708 0.6963 0.8344
No log 5.8667 88 0.6581 0.4995 0.6581 0.8112
No log 6.0 90 0.7345 0.5598 0.7345 0.8570
No log 6.1333 92 0.6786 0.5860 0.6786 0.8238
No log 6.2667 94 0.6124 0.6095 0.6124 0.7826
No log 6.4 96 0.5998 0.6306 0.5998 0.7744
No log 6.5333 98 0.6587 0.6573 0.6587 0.8116
No log 6.6667 100 0.6829 0.6573 0.6829 0.8264
No log 6.8 102 0.6049 0.7001 0.6049 0.7778
No log 6.9333 104 0.7245 0.5259 0.7245 0.8512
No log 7.0667 106 1.0516 0.4929 1.0516 1.0255
No log 7.2 108 0.8983 0.5283 0.8983 0.9478
No log 7.3333 110 0.6201 0.5876 0.6201 0.7875
No log 7.4667 112 0.6080 0.7019 0.6080 0.7798
No log 7.6 114 0.5752 0.7074 0.5752 0.7584
No log 7.7333 116 0.6052 0.6234 0.6052 0.7779
No log 7.8667 118 0.7458 0.5664 0.7458 0.8636
No log 8.0 120 0.7472 0.5821 0.7472 0.8644
No log 8.1333 122 0.6138 0.6386 0.6138 0.7834
No log 8.2667 124 0.6058 0.6242 0.6058 0.7783
No log 8.4 126 0.6245 0.6237 0.6245 0.7903
No log 8.5333 128 0.8187 0.5157 0.8187 0.9048
No log 8.6667 130 0.9129 0.4961 0.9129 0.9554
No log 8.8 132 0.7596 0.5012 0.7596 0.8716
No log 8.9333 134 0.6838 0.5438 0.6838 0.8269
No log 9.0667 136 0.6778 0.5327 0.6778 0.8233
No log 9.2 138 0.8229 0.5804 0.8229 0.9072
No log 9.3333 140 0.8636 0.5581 0.8636 0.9293
No log 9.4667 142 0.6708 0.5325 0.6708 0.8190
No log 9.6 144 0.6294 0.6055 0.6294 0.7933
No log 9.7333 146 0.6247 0.6055 0.6247 0.7904
No log 9.8667 148 0.6320 0.6234 0.6320 0.7950
No log 10.0 150 0.7898 0.6057 0.7898 0.8887
No log 10.1333 152 0.7435 0.6074 0.7435 0.8623
No log 10.2667 154 0.5949 0.6555 0.5949 0.7713
No log 10.4 156 0.7074 0.6291 0.7074 0.8410
No log 10.5333 158 0.7434 0.5745 0.7434 0.8622
No log 10.6667 160 0.6047 0.6605 0.6047 0.7776
No log 10.8 162 0.6206 0.5794 0.6206 0.7878
No log 10.9333 164 0.7547 0.5891 0.7547 0.8688
No log 11.0667 166 0.6696 0.6014 0.6696 0.8183
No log 11.2 168 0.5641 0.6667 0.5641 0.7511
No log 11.3333 170 0.5806 0.6623 0.5806 0.7620
No log 11.4667 172 0.5696 0.6335 0.5696 0.7547
No log 11.6 174 0.5697 0.6750 0.5697 0.7548
No log 11.7333 176 0.5665 0.6409 0.5665 0.7527
No log 11.8667 178 0.5597 0.6699 0.5597 0.7481
No log 12.0 180 0.5728 0.6497 0.5728 0.7568
No log 12.1333 182 0.5736 0.6368 0.5736 0.7573
No log 12.2667 184 0.5522 0.6374 0.5522 0.7431
No log 12.4 186 0.5843 0.5865 0.5843 0.7644
No log 12.5333 188 0.6030 0.6050 0.6030 0.7765
No log 12.6667 190 0.5785 0.5926 0.5785 0.7606
No log 12.8 192 0.5797 0.6051 0.5797 0.7614
No log 12.9333 194 0.5528 0.6460 0.5528 0.7435
No log 13.0667 196 0.5152 0.7409 0.5152 0.7178
No log 13.2 198 0.5531 0.6485 0.5531 0.7437
No log 13.3333 200 0.4948 0.7165 0.4948 0.7034
No log 13.4667 202 0.4659 0.7110 0.4659 0.6826
No log 13.6 204 0.4707 0.7158 0.4707 0.6861
No log 13.7333 206 0.4758 0.6748 0.4758 0.6898
No log 13.8667 208 0.5318 0.6572 0.5318 0.7293
No log 14.0 210 0.5964 0.5873 0.5964 0.7722
No log 14.1333 212 0.5735 0.6302 0.5735 0.7573
No log 14.2667 214 0.5857 0.6435 0.5857 0.7653
No log 14.4 216 0.6045 0.6435 0.6045 0.7775
No log 14.5333 218 0.5924 0.5759 0.5924 0.7696
No log 14.6667 220 0.6873 0.5598 0.6873 0.8291
No log 14.8 222 0.7047 0.5590 0.7047 0.8395
No log 14.9333 224 0.6091 0.5876 0.6091 0.7805
No log 15.0667 226 0.5902 0.5817 0.5902 0.7683
No log 15.2 228 0.5756 0.6207 0.5756 0.7587
No log 15.3333 230 0.5807 0.6207 0.5807 0.7620
No log 15.4667 232 0.6104 0.6225 0.6104 0.7813
No log 15.6 234 0.7471 0.5257 0.7471 0.8643
No log 15.7333 236 0.7701 0.5161 0.7701 0.8776
No log 15.8667 238 0.8617 0.4848 0.8617 0.9283
No log 16.0 240 1.0436 0.4228 1.0436 1.0215
No log 16.1333 242 0.9356 0.4583 0.9356 0.9673
No log 16.2667 244 0.6741 0.5138 0.6741 0.8210
No log 16.4 246 0.5975 0.6164 0.5975 0.7730
No log 16.5333 248 0.5897 0.5735 0.5897 0.7680
No log 16.6667 250 0.5949 0.5485 0.5949 0.7713
No log 16.8 252 0.6029 0.5886 0.6029 0.7765
No log 16.9333 254 0.5989 0.5989 0.5989 0.7739
No log 17.0667 256 0.6270 0.6231 0.6270 0.7918
No log 17.2 258 0.6411 0.5970 0.6411 0.8007
No log 17.3333 260 0.6198 0.6301 0.6198 0.7873
No log 17.4667 262 0.6059 0.5797 0.6059 0.7784
No log 17.6 264 0.6213 0.5716 0.6213 0.7882
No log 17.7333 266 0.6171 0.6076 0.6171 0.7855
No log 17.8667 268 0.6073 0.5978 0.6073 0.7793
No log 18.0 270 0.6023 0.6415 0.6023 0.7761
No log 18.1333 272 0.6206 0.6611 0.6206 0.7878
No log 18.2667 274 0.6137 0.6611 0.6137 0.7834
No log 18.4 276 0.5801 0.5978 0.5801 0.7616
No log 18.5333 278 0.5637 0.5921 0.5637 0.7508
No log 18.6667 280 0.5409 0.5879 0.5409 0.7355
No log 18.8 282 0.5098 0.6857 0.5098 0.7140
No log 18.9333 284 0.5099 0.6872 0.5099 0.7141
No log 19.0667 286 0.5038 0.6962 0.5038 0.7098
No log 19.2 288 0.5042 0.7001 0.5042 0.7101
No log 19.3333 290 0.5048 0.7001 0.5048 0.7105
No log 19.4667 292 0.5065 0.6569 0.5065 0.7117
No log 19.6 294 0.5093 0.6886 0.5093 0.7137
No log 19.7333 296 0.5030 0.6970 0.5030 0.7092
No log 19.8667 298 0.5428 0.5876 0.5428 0.7367
No log 20.0 300 0.5609 0.6004 0.5609 0.7489
No log 20.1333 302 0.5215 0.6374 0.5215 0.7221
No log 20.2667 304 0.5244 0.6597 0.5244 0.7241
No log 20.4 306 0.5346 0.6597 0.5346 0.7311
No log 20.5333 308 0.5565 0.5770 0.5565 0.7460
No log 20.6667 310 0.6098 0.5590 0.6098 0.7809
No log 20.8 312 0.5980 0.5781 0.5980 0.7733
No log 20.9333 314 0.5685 0.6072 0.5685 0.7540
No log 21.0667 316 0.5372 0.6488 0.5372 0.7329
No log 21.2 318 0.4858 0.6932 0.4858 0.6970
No log 21.3333 320 0.5223 0.6985 0.5223 0.7227
No log 21.4667 322 0.5254 0.6807 0.5254 0.7248
No log 21.6 324 0.5388 0.6445 0.5388 0.7340
No log 21.7333 326 0.5546 0.6108 0.5546 0.7447
No log 21.8667 328 0.5598 0.5890 0.5598 0.7482
No log 22.0 330 0.5542 0.6337 0.5542 0.7445
No log 22.1333 332 0.5463 0.6296 0.5463 0.7391
No log 22.2667 334 0.5566 0.6771 0.5566 0.7460
No log 22.4 336 0.5623 0.6771 0.5623 0.7499
No log 22.5333 338 0.5598 0.5871 0.5598 0.7482
No log 22.6667 340 0.5994 0.5706 0.5994 0.7742
No log 22.8 342 0.5938 0.5917 0.5938 0.7706
No log 22.9333 344 0.5474 0.6230 0.5474 0.7399
No log 23.0667 346 0.5509 0.6588 0.5509 0.7422
No log 23.2 348 0.6188 0.6295 0.6188 0.7866
No log 23.3333 350 0.6080 0.6101 0.6080 0.7798
No log 23.4667 352 0.5521 0.6597 0.5521 0.7430
No log 23.6 354 0.5831 0.5921 0.5831 0.7636
No log 23.7333 356 0.6460 0.5270 0.6460 0.8038
No log 23.8667 358 0.6107 0.5902 0.6107 0.7815
No log 24.0 360 0.5522 0.6241 0.5522 0.7431
No log 24.1333 362 0.5333 0.6597 0.5333 0.7303
No log 24.2667 364 0.5705 0.5964 0.5705 0.7553
No log 24.4 366 0.5663 0.5763 0.5663 0.7525
No log 24.5333 368 0.5285 0.6942 0.5285 0.7270
No log 24.6667 370 0.5363 0.6087 0.5363 0.7323
No log 24.8 372 0.5395 0.6597 0.5395 0.7345
No log 24.9333 374 0.5726 0.6231 0.5726 0.7567
No log 25.0667 376 0.5697 0.6432 0.5697 0.7548
No log 25.2 378 0.5590 0.6432 0.5590 0.7477
No log 25.3333 380 0.5474 0.6407 0.5474 0.7399
No log 25.4667 382 0.5485 0.6805 0.5485 0.7406
No log 25.6 384 0.5434 0.6579 0.5434 0.7371
No log 25.7333 386 0.5353 0.6518 0.5353 0.7316
No log 25.8667 388 0.5822 0.6723 0.5822 0.7630
No log 26.0 390 0.6098 0.5981 0.6098 0.7809
No log 26.1333 392 0.5785 0.6646 0.5785 0.7606
No log 26.2667 394 0.5308 0.7035 0.5308 0.7285
No log 26.4 396 0.5123 0.6537 0.5123 0.7157
No log 26.5333 398 0.5070 0.6712 0.5070 0.7120
No log 26.6667 400 0.5116 0.6745 0.5116 0.7152
No log 26.8 402 0.5524 0.7035 0.5524 0.7432
No log 26.9333 404 0.5605 0.6916 0.5605 0.7487
No log 27.0667 406 0.5326 0.6697 0.5326 0.7298
No log 27.2 408 0.5157 0.6942 0.5157 0.7181
No log 27.3333 410 0.5035 0.6911 0.5035 0.7096
No log 27.4667 412 0.4885 0.6874 0.4885 0.6989
No log 27.6 414 0.4815 0.6833 0.4815 0.6939
No log 27.7333 416 0.4864 0.6724 0.4864 0.6974
No log 27.8667 418 0.5259 0.6602 0.5259 0.7252
No log 28.0 420 0.5544 0.5917 0.5544 0.7446
No log 28.1333 422 0.5424 0.6588 0.5424 0.7365
No log 28.2667 424 0.5337 0.6724 0.5337 0.7306
No log 28.4 426 0.5421 0.6397 0.5421 0.7363
No log 28.5333 428 0.5635 0.6714 0.5635 0.7507
No log 28.6667 430 0.5495 0.6844 0.5495 0.7413
No log 28.8 432 0.5285 0.6933 0.5285 0.7270
No log 28.9333 434 0.5159 0.6942 0.5159 0.7183
No log 29.0667 436 0.5270 0.6966 0.5270 0.7259
No log 29.2 438 0.5365 0.6581 0.5365 0.7325
No log 29.3333 440 0.5208 0.6919 0.5208 0.7216
No log 29.4667 442 0.5135 0.6736 0.5135 0.7166
No log 29.6 444 0.5396 0.7036 0.5396 0.7346
No log 29.7333 446 0.5573 0.7036 0.5573 0.7465
No log 29.8667 448 0.5681 0.6479 0.5681 0.7537
No log 30.0 450 0.5468 0.7017 0.5468 0.7395
No log 30.1333 452 0.5237 0.6445 0.5237 0.7237
No log 30.2667 454 0.5045 0.6332 0.5045 0.7103
No log 30.4 456 0.4901 0.6874 0.4901 0.7000
No log 30.5333 458 0.4843 0.6959 0.4843 0.6959
No log 30.6667 460 0.4988 0.6894 0.4988 0.7062
No log 30.8 462 0.5173 0.6929 0.5173 0.7193
No log 30.9333 464 0.5202 0.6962 0.5202 0.7212
No log 31.0667 466 0.5327 0.6813 0.5327 0.7298
No log 31.2 468 0.5662 0.6377 0.5662 0.7525
No log 31.3333 470 0.6307 0.5379 0.6307 0.7942
No log 31.4667 472 0.6446 0.5590 0.6446 0.8029
No log 31.6 474 0.6295 0.5793 0.6295 0.7934
No log 31.7333 476 0.5895 0.6507 0.5895 0.7678
No log 31.8667 478 0.5698 0.6390 0.5698 0.7548
No log 32.0 480 0.5623 0.6154 0.5623 0.7499
No log 32.1333 482 0.5583 0.6154 0.5583 0.7472
No log 32.2667 484 0.5606 0.6118 0.5606 0.7488
No log 32.4 486 0.5618 0.6241 0.5618 0.7495
No log 32.5333 488 0.5532 0.5966 0.5532 0.7438
No log 32.6667 490 0.5607 0.6925 0.5607 0.7488
No log 32.8 492 0.6099 0.6833 0.6099 0.7810
No log 32.9333 494 0.6664 0.6117 0.6664 0.8163
No log 33.0667 496 0.6863 0.6004 0.6863 0.8284
No log 33.2 498 0.6442 0.6117 0.6442 0.8026
0.2662 33.3333 500 0.5785 0.6925 0.5785 0.7606
0.2662 33.4667 502 0.5582 0.6028 0.5582 0.7471
0.2662 33.6 504 0.5643 0.6006 0.5643 0.7512
0.2662 33.7333 506 0.5555 0.5831 0.5555 0.7453
0.2662 33.8667 508 0.5668 0.7043 0.5668 0.7529
0.2662 34.0 510 0.6225 0.6637 0.6225 0.7890
0.2662 34.1333 512 0.6818 0.5684 0.6818 0.8257
0.2662 34.2667 514 0.6835 0.5684 0.6835 0.8268
0.2662 34.4 516 0.6469 0.6252 0.6469 0.8043
0.2662 34.5333 518 0.5970 0.6711 0.5970 0.7727

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k5_task5_organization

Finetuned
(4223)
this model