File size: 7,852 Bytes
eefa2a8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
---
license: apache-2.0
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: resnet101_rvl-cdip-cnn_rvl_cdip-NK1000_hint_rand
  results: []
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# resnet101_rvl-cdip-cnn_rvl_cdip-NK1000_hint_rand

This model is a fine-tuned version of [bdpc/resnet101_rvl-cdip](https://huggingface.co/bdpc/resnet101_rvl-cdip) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 19.9624
- Accuracy: 0.765
- Brier Loss: 0.3910
- Nll: 2.2998
- F1 Micro: 0.765
- F1 Macro: 0.7641
- Ece: 0.1669
- Aurc: 0.0775

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 50

### Training results

| Training Loss | Epoch | Step  | Validation Loss | Accuracy | Brier Loss | Nll    | F1 Micro | F1 Macro | Ece    | Aurc   |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:----------:|:------:|:--------:|:--------:|:------:|:------:|
| No log        | 1.0   | 250   | 26.8546         | 0.2102   | 0.8652     | 3.4225 | 0.2102   | 0.1897   | 0.0659 | 0.6137 |
| 26.8891       | 2.0   | 500   | 25.9803         | 0.2845   | 0.8381     | 3.4360 | 0.2845   | 0.2457   | 0.0762 | 0.5573 |
| 26.8891       | 3.0   | 750   | 25.8252         | 0.337    | 0.8092     | 3.6483 | 0.337    | 0.3168   | 0.1091 | 0.4988 |
| 25.1656       | 4.0   | 1000  | 24.9957         | 0.4088   | 0.7588     | 3.0528 | 0.4088   | 0.3859   | 0.1283 | 0.4163 |
| 25.1656       | 5.0   | 1250  | 24.2209         | 0.5517   | 0.5964     | 2.6661 | 0.5517   | 0.5510   | 0.0726 | 0.2489 |
| 23.8526       | 6.0   | 1500  | 23.4086         | 0.5915   | 0.5431     | 2.5360 | 0.5915   | 0.5840   | 0.0549 | 0.1994 |
| 23.8526       | 7.0   | 1750  | 23.0800         | 0.625    | 0.5049     | 2.4839 | 0.625    | 0.6268   | 0.0608 | 0.1668 |
| 22.7511       | 8.0   | 2000  | 22.9512         | 0.6573   | 0.4660     | 2.4202 | 0.6573   | 0.6564   | 0.0565 | 0.1407 |
| 22.7511       | 9.0   | 2250  | 22.7991         | 0.6783   | 0.4509     | 2.4137 | 0.6783   | 0.6749   | 0.0634 | 0.1315 |
| 21.9881       | 10.0  | 2500  | 22.8533         | 0.6352   | 0.5056     | 2.6549 | 0.6352   | 0.6231   | 0.1085 | 0.1554 |
| 21.9881       | 11.0  | 2750  | 22.7499         | 0.669    | 0.4673     | 2.5291 | 0.669    | 0.6642   | 0.1053 | 0.1347 |
| 21.391        | 12.0  | 3000  | 22.6520         | 0.6757   | 0.4767     | 2.5038 | 0.6757   | 0.6745   | 0.1204 | 0.1355 |
| 21.391        | 13.0  | 3250  | 22.4767         | 0.6737   | 0.4850     | 2.6030 | 0.6737   | 0.6718   | 0.1385 | 0.1380 |
| 20.9347       | 14.0  | 3500  | 22.3023         | 0.6767   | 0.4832     | 2.5438 | 0.6767   | 0.6770   | 0.1594 | 0.1301 |
| 20.9347       | 15.0  | 3750  | 22.1482         | 0.693    | 0.4666     | 2.5622 | 0.693    | 0.6913   | 0.1581 | 0.1209 |
| 20.5776       | 16.0  | 4000  | 22.1655         | 0.6943   | 0.4849     | 2.5685 | 0.6943   | 0.6994   | 0.1766 | 0.1288 |
| 20.5776       | 17.0  | 4250  | 22.0213         | 0.686    | 0.4922     | 2.6576 | 0.686    | 0.6925   | 0.1749 | 0.1250 |
| 20.2836       | 18.0  | 4500  | 21.5434         | 0.7023   | 0.4560     | 2.5508 | 0.7023   | 0.7018   | 0.1720 | 0.1146 |
| 20.2836       | 19.0  | 4750  | 21.7105         | 0.715    | 0.4501     | 2.5953 | 0.715    | 0.7128   | 0.1738 | 0.1083 |
| 20.0339       | 20.0  | 5000  | 21.6301         | 0.7057   | 0.4645     | 2.6131 | 0.7057   | 0.7033   | 0.1701 | 0.1153 |
| 20.0339       | 21.0  | 5250  | 21.9130         | 0.7007   | 0.4825     | 2.7114 | 0.7007   | 0.6989   | 0.1917 | 0.1231 |
| 19.8026       | 22.0  | 5500  | 21.7975         | 0.713    | 0.4702     | 2.7340 | 0.713    | 0.7117   | 0.1879 | 0.1148 |
| 19.8026       | 23.0  | 5750  | 21.5577         | 0.7173   | 0.4621     | 2.7138 | 0.7173   | 0.7100   | 0.1931 | 0.1072 |
| 19.6001       | 24.0  | 6000  | 21.2486         | 0.722    | 0.4491     | 2.5651 | 0.722    | 0.7214   | 0.1853 | 0.1045 |
| 19.6001       | 25.0  | 6250  | 21.0363         | 0.7348   | 0.4344     | 2.4688 | 0.7348   | 0.7364   | 0.1780 | 0.0974 |
| 19.4158       | 26.0  | 6500  | 21.3527         | 0.728    | 0.4495     | 2.7492 | 0.728    | 0.7219   | 0.1864 | 0.1005 |
| 19.4158       | 27.0  | 6750  | 20.8258         | 0.7355   | 0.4339     | 2.4375 | 0.7355   | 0.7352   | 0.1838 | 0.0943 |
| 19.2585       | 28.0  | 7000  | 21.0491         | 0.729    | 0.4465     | 2.6324 | 0.729    | 0.7245   | 0.1953 | 0.1010 |
| 19.2585       | 29.0  | 7250  | 20.7774         | 0.7425   | 0.4283     | 2.4694 | 0.7425   | 0.7410   | 0.1799 | 0.0908 |
| 19.1051       | 30.0  | 7500  | 20.6908         | 0.741    | 0.4311     | 2.4924 | 0.7410   | 0.7405   | 0.1890 | 0.0888 |
| 19.1051       | 31.0  | 7750  | 20.8242         | 0.743    | 0.4264     | 2.5098 | 0.743    | 0.7407   | 0.1826 | 0.0903 |
| 18.9722       | 32.0  | 8000  | 20.6257         | 0.7435   | 0.4288     | 2.4740 | 0.7435   | 0.7432   | 0.1865 | 0.0872 |
| 18.9722       | 33.0  | 8250  | 20.6265         | 0.745    | 0.4289     | 2.4552 | 0.745    | 0.7435   | 0.1862 | 0.0929 |
| 18.854        | 34.0  | 8500  | 20.4251         | 0.7505   | 0.4124     | 2.4631 | 0.7505   | 0.7513   | 0.1790 | 0.0845 |
| 18.854        | 35.0  | 8750  | 20.4164         | 0.741    | 0.4278     | 2.3888 | 0.7410   | 0.7402   | 0.1859 | 0.0889 |
| 18.7477       | 36.0  | 9000  | 20.3432         | 0.751    | 0.4184     | 2.4020 | 0.751    | 0.7485   | 0.1800 | 0.0850 |
| 18.7477       | 37.0  | 9250  | 20.4310         | 0.7555   | 0.4154     | 2.4639 | 0.7555   | 0.7528   | 0.1759 | 0.0842 |
| 18.6548       | 38.0  | 9500  | 20.1987         | 0.7542   | 0.4111     | 2.2921 | 0.7542   | 0.7542   | 0.1792 | 0.0815 |
| 18.6548       | 39.0  | 9750  | 20.2326         | 0.7562   | 0.4017     | 2.3536 | 0.7562   | 0.7537   | 0.1767 | 0.0829 |
| 18.5776       | 40.0  | 10000 | 20.1571         | 0.7575   | 0.3985     | 2.3405 | 0.7575   | 0.7568   | 0.1703 | 0.0811 |
| 18.5776       | 41.0  | 10250 | 20.1580         | 0.7625   | 0.3962     | 2.3855 | 0.7625   | 0.7621   | 0.1713 | 0.0814 |
| 18.5133       | 42.0  | 10500 | 20.0952         | 0.7572   | 0.4038     | 2.3600 | 0.7572   | 0.7563   | 0.1768 | 0.0794 |
| 18.5133       | 43.0  | 10750 | 20.1483         | 0.7575   | 0.4008     | 2.3713 | 0.7575   | 0.7564   | 0.1755 | 0.0820 |
| 18.4613       | 44.0  | 11000 | 20.0749         | 0.762    | 0.3992     | 2.3372 | 0.762    | 0.7618   | 0.1720 | 0.0795 |
| 18.4613       | 45.0  | 11250 | 20.0664         | 0.7578   | 0.4035     | 2.3570 | 0.7577   | 0.7566   | 0.1769 | 0.0795 |
| 18.4218       | 46.0  | 11500 | 19.9611         | 0.7622   | 0.3946     | 2.3399 | 0.7622   | 0.7617   | 0.1674 | 0.0784 |
| 18.4218       | 47.0  | 11750 | 19.9678         | 0.7632   | 0.3907     | 2.3011 | 0.7632   | 0.7635   | 0.1692 | 0.0772 |
| 18.3945       | 48.0  | 12000 | 19.9950         | 0.763    | 0.3910     | 2.2773 | 0.763    | 0.7616   | 0.1695 | 0.0775 |
| 18.3945       | 49.0  | 12250 | 20.0013         | 0.7625   | 0.3911     | 2.2875 | 0.7625   | 0.7618   | 0.1705 | 0.0777 |
| 18.3792       | 50.0  | 12500 | 19.9624         | 0.765    | 0.3910     | 2.2998 | 0.765    | 0.7641   | 0.1669 | 0.0775 |


### Framework versions

- Transformers 4.26.1
- Pytorch 1.13.1.post200
- Datasets 2.9.0
- Tokenizers 0.13.2