Edit model card

resnet101_rvl-cdip-cnn_rvl_cdip-NK1000_simkd_rand

This model is a fine-tuned version of bdpc/resnet101_rvl-cdip on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4174
  • Accuracy: 0.7665
  • Brier Loss: 0.3263
  • Nll: 2.0962
  • F1 Micro: 0.7665
  • F1 Macro: 0.7661
  • Ece: 0.0504
  • Aurc: 0.0700

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 64
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 50

Training results

Training Loss Epoch Step Validation Loss Accuracy Brier Loss Nll F1 Micro F1 Macro Ece Aurc
No log 1.0 250 0.8645 0.1192 0.9514 3.2233 0.1192 0.0652 0.1115 0.8122
1.2523 2.0 500 0.7139 0.1797 0.8939 3.1527 0.1798 0.1283 0.0795 0.6807
1.2523 3.0 750 0.6662 0.3145 0.8040 6.3258 0.3145 0.2485 0.0647 0.4987
0.6553 4.0 1000 0.6265 0.3738 0.7356 6.0830 0.3738 0.3459 0.0768 0.4070
0.6553 5.0 1250 0.5609 0.531 0.6047 4.7056 0.531 0.5234 0.0639 0.2463
0.5525 6.0 1500 0.5341 0.589 0.5450 3.9772 0.589 0.5948 0.0718 0.1912
0.5525 7.0 1750 0.4938 0.6468 0.4733 3.3676 0.6468 0.6486 0.0670 0.1408
0.4842 8.0 2000 0.4765 0.7 0.4288 2.8692 0.7 0.6960 0.0666 0.1181
0.4842 9.0 2250 0.5359 0.5938 0.5534 3.9887 0.5938 0.6011 0.1211 0.1809
0.4476 10.0 2500 0.4611 0.7037 0.4122 2.7429 0.7037 0.6991 0.0679 0.1097
0.4476 11.0 2750 0.4460 0.7225 0.3913 2.6158 0.7225 0.7240 0.0725 0.0967
0.4219 12.0 3000 0.4387 0.7388 0.3752 2.4639 0.7388 0.7356 0.0696 0.0892
0.4219 13.0 3250 0.4399 0.7378 0.3724 2.4683 0.7378 0.7381 0.0550 0.0898
0.4007 14.0 3500 0.4441 0.737 0.3738 2.4680 0.737 0.7334 0.0581 0.0906
0.4007 15.0 3750 0.4517 0.7248 0.3906 2.5901 0.7248 0.7302 0.0653 0.0961
0.3825 16.0 4000 0.4430 0.737 0.3727 2.5633 0.737 0.7350 0.0595 0.0884
0.3825 17.0 4250 0.4345 0.7482 0.3518 2.3938 0.7482 0.7473 0.0541 0.0784
0.3672 18.0 4500 0.4642 0.7385 0.3690 2.4016 0.7385 0.7367 0.0571 0.0891
0.3672 19.0 4750 0.4309 0.7432 0.3585 2.3331 0.7432 0.7464 0.0558 0.0824
0.3547 20.0 5000 0.4205 0.7602 0.3418 2.2097 0.7602 0.7617 0.0470 0.0744
0.3547 21.0 5250 0.4174 0.7602 0.3387 2.2020 0.7602 0.7594 0.0488 0.0748
0.3442 22.0 5500 0.4207 0.7515 0.3458 2.2370 0.7515 0.7543 0.0540 0.0777
0.3442 23.0 5750 0.4465 0.733 0.3783 2.5113 0.733 0.7295 0.0576 0.0919
0.3355 24.0 6000 0.4391 0.7425 0.3649 2.4598 0.7425 0.7459 0.0534 0.0830
0.3355 25.0 6250 0.4233 0.7598 0.3352 2.2321 0.7598 0.7609 0.0495 0.0729
0.3274 26.0 6500 0.4174 0.7665 0.3305 2.2062 0.7665 0.7673 0.0482 0.0699
0.3274 27.0 6750 0.4153 0.7598 0.3389 2.2158 0.7598 0.7583 0.0549 0.0740
0.3206 28.0 7000 0.4175 0.763 0.3323 2.1843 0.763 0.7610 0.0494 0.0721
0.3206 29.0 7250 0.4201 0.7522 0.3467 2.2627 0.7522 0.7495 0.0576 0.0783
0.3147 30.0 7500 0.4133 0.7625 0.3334 2.1459 0.7625 0.7631 0.0477 0.0733
0.3147 31.0 7750 0.4213 0.7558 0.3421 2.2877 0.7558 0.7535 0.0567 0.0758
0.3092 32.0 8000 0.4136 0.7668 0.3294 2.1791 0.7668 0.7662 0.0465 0.0702
0.3092 33.0 8250 0.4114 0.7638 0.3331 2.1993 0.7638 0.7613 0.0517 0.0722
0.3046 34.0 8500 0.4154 0.764 0.3294 2.1689 0.764 0.7639 0.0489 0.0714
0.3046 35.0 8750 0.4119 0.7638 0.3327 2.1482 0.7638 0.7628 0.0449 0.0725
0.3001 36.0 9000 0.4183 0.759 0.3348 2.1775 0.7590 0.7605 0.0513 0.0731
0.3001 37.0 9250 0.4097 0.7578 0.3344 2.2029 0.7577 0.7571 0.0525 0.0736
0.2964 38.0 9500 0.4126 0.7655 0.3292 2.1374 0.7655 0.7657 0.0481 0.0710
0.2964 39.0 9750 0.4235 0.7642 0.3287 2.1640 0.7642 0.7639 0.0543 0.0707
0.293 40.0 10000 0.4168 0.7678 0.3284 2.1264 0.7678 0.7681 0.0494 0.0702
0.293 41.0 10250 0.4118 0.7682 0.3270 2.1387 0.7682 0.7684 0.0462 0.0702
0.29 42.0 10500 0.4151 0.7618 0.3288 2.1464 0.7618 0.7609 0.0493 0.0718
0.29 43.0 10750 0.4172 0.7608 0.3283 2.1341 0.7608 0.7607 0.0538 0.0708
0.2876 44.0 11000 0.4159 0.7612 0.3278 2.1561 0.7612 0.7601 0.0514 0.0707
0.2876 45.0 11250 0.4173 0.761 0.3291 2.1825 0.761 0.7602 0.0493 0.0711
0.2855 46.0 11500 0.4137 0.761 0.3295 2.1514 0.761 0.7598 0.0507 0.0709
0.2855 47.0 11750 0.4143 0.764 0.3278 2.1414 0.764 0.7630 0.0483 0.0705
0.2841 48.0 12000 0.4162 0.7668 0.3262 2.1191 0.7668 0.7666 0.0451 0.0699
0.2841 49.0 12250 0.4190 0.765 0.3271 2.1267 0.765 0.7647 0.0496 0.0701
0.283 50.0 12500 0.4174 0.7665 0.3263 2.0962 0.7665 0.7661 0.0504 0.0700

Framework versions

  • Transformers 4.26.1
  • Pytorch 1.13.1.post200
  • Datasets 2.9.0
  • Tokenizers 0.13.2
Downloads last month
7