Edit model card

resnet101_rvl-cdip-cnn_rvl_cdip-NK1000_kd_rand

This model is a fine-tuned version of bdpc/resnet101_rvl-cdip on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7141
  • Accuracy: 0.776
  • Brier Loss: 0.3162
  • Nll: 1.9406
  • F1 Micro: 0.776
  • F1 Macro: 0.7755
  • Ece: 0.0376
  • Aurc: 0.0655

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 64
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 50

Training results

Training Loss Epoch Step Validation Loss Accuracy Brier Loss Nll F1 Micro F1 Macro Ece Aurc
No log 1.0 250 3.5746 0.2293 0.8509 3.6359 0.2293 0.1928 0.0536 0.5849
3.6362 2.0 500 3.4699 0.2848 0.8495 4.0612 0.2848 0.2468 0.0791 0.5541
3.6362 3.0 750 2.9284 0.3957 0.7832 3.1766 0.3957 0.3732 0.1412 0.4345
2.5616 4.0 1000 2.4586 0.4497 0.7009 3.0303 0.4497 0.4362 0.0977 0.3388
2.5616 5.0 1250 2.5238 0.4268 0.7479 3.0431 0.4268 0.4163 0.2085 0.3343
1.8525 6.0 1500 1.6707 0.5915 0.5475 2.6137 0.5915 0.5880 0.0929 0.1837
1.8525 7.0 1750 1.3529 0.6385 0.4901 2.3981 0.6385 0.6360 0.0936 0.1504
1.2522 8.0 2000 1.2537 0.6585 0.4640 2.2970 0.6585 0.6591 0.0722 0.1341
1.2522 9.0 2250 1.0462 0.7053 0.4096 2.2117 0.7053 0.7032 0.0668 0.1048
0.8788 10.0 2500 1.2045 0.6743 0.4500 2.2939 0.6743 0.6751 0.0818 0.1245
0.8788 11.0 2750 1.0487 0.7045 0.4107 2.1999 0.7045 0.6995 0.0686 0.1074
0.6325 12.0 3000 1.0625 0.7105 0.4075 2.2241 0.7105 0.7038 0.0670 0.1062
0.6325 13.0 3250 0.9567 0.7262 0.3820 2.1370 0.7262 0.7211 0.0585 0.0950
0.4724 14.0 3500 0.8833 0.744 0.3598 2.0806 0.744 0.7450 0.0477 0.0820
0.4724 15.0 3750 0.9074 0.744 0.3638 2.1065 0.744 0.7447 0.0504 0.0869
0.3827 16.0 4000 0.9355 0.7382 0.3678 2.1484 0.7382 0.7369 0.0490 0.0877
0.3827 17.0 4250 0.8586 0.7502 0.3508 2.0839 0.7502 0.7495 0.0500 0.0803
0.3253 18.0 4500 0.8572 0.7412 0.3579 2.0452 0.7412 0.7419 0.0504 0.0830
0.3253 19.0 4750 0.8663 0.741 0.3581 2.0923 0.7410 0.7405 0.0481 0.0826
0.288 20.0 5000 0.8162 0.7535 0.3431 2.0485 0.7535 0.7547 0.0444 0.0776
0.288 21.0 5250 0.9363 0.735 0.3739 2.1240 0.735 0.7332 0.0482 0.0895
0.2614 22.0 5500 0.8013 0.762 0.3391 2.0235 0.762 0.7598 0.0457 0.0743
0.2614 23.0 5750 0.7936 0.756 0.3391 2.0319 0.756 0.7585 0.0450 0.0748
0.2434 24.0 6000 0.8142 0.7515 0.3476 1.9971 0.7515 0.7483 0.0455 0.0769
0.2434 25.0 6250 0.8032 0.757 0.3441 2.0560 0.757 0.7569 0.0482 0.0766
0.2252 26.0 6500 0.7746 0.7622 0.3305 2.0183 0.7622 0.7648 0.0412 0.0717
0.2252 27.0 6750 0.7663 0.7655 0.3342 1.9947 0.7655 0.7674 0.0371 0.0724
0.213 28.0 7000 0.8283 0.756 0.3460 2.0260 0.756 0.7558 0.0432 0.0770
0.213 29.0 7250 0.7564 0.7638 0.3301 2.0031 0.7638 0.7653 0.0402 0.0716
0.2007 30.0 7500 0.7836 0.761 0.3376 2.0263 0.761 0.7599 0.0380 0.0741
0.2007 31.0 7750 0.7539 0.7668 0.3297 1.9693 0.7668 0.7647 0.0346 0.0699
0.1911 32.0 8000 0.7534 0.7665 0.3294 1.9842 0.7665 0.7645 0.0449 0.0702
0.1911 33.0 8250 0.7608 0.769 0.3276 2.0015 0.769 0.7687 0.0381 0.0701
0.183 34.0 8500 0.7511 0.7675 0.3273 1.9691 0.7675 0.7666 0.0394 0.0687
0.183 35.0 8750 0.7582 0.7615 0.3282 1.9961 0.7615 0.7607 0.0468 0.0717
0.1749 36.0 9000 0.7272 0.7702 0.3228 1.9832 0.7702 0.7706 0.0385 0.0679
0.1749 37.0 9250 0.7446 0.7685 0.3258 2.0006 0.7685 0.7698 0.0391 0.0694
0.1672 38.0 9500 0.7252 0.773 0.3200 1.9722 0.7730 0.7719 0.0423 0.0673
0.1672 39.0 9750 0.7260 0.7738 0.3184 1.9599 0.7738 0.7741 0.0357 0.0660
0.1608 40.0 10000 0.7317 0.7715 0.3191 1.9651 0.7715 0.7717 0.0340 0.0668
0.1608 41.0 10250 0.7268 0.7742 0.3206 1.9666 0.7742 0.7728 0.0429 0.0671
0.1549 42.0 10500 0.7296 0.7745 0.3198 1.9628 0.7745 0.7738 0.0341 0.0666
0.1549 43.0 10750 0.7223 0.773 0.3173 1.9566 0.7730 0.7732 0.0414 0.0661
0.15 44.0 11000 0.7184 0.775 0.3175 1.9566 0.775 0.7746 0.0385 0.0661
0.15 45.0 11250 0.7205 0.7755 0.3181 1.9439 0.7755 0.7749 0.0377 0.0661
0.1458 46.0 11500 0.7174 0.777 0.3173 1.9560 0.777 0.7768 0.0424 0.0660
0.1458 47.0 11750 0.7152 0.7755 0.3164 1.9527 0.7755 0.7750 0.0404 0.0656
0.1429 48.0 12000 0.7173 0.777 0.3161 1.9350 0.777 0.7774 0.0415 0.0656
0.1429 49.0 12250 0.7142 0.7755 0.3159 1.9515 0.7755 0.7748 0.0404 0.0656
0.141 50.0 12500 0.7141 0.776 0.3162 1.9406 0.776 0.7755 0.0376 0.0655

Framework versions

  • Transformers 4.26.1
  • Pytorch 1.13.1.post200
  • Datasets 2.9.0
  • Tokenizers 0.13.2
Downloads last month
6