File size: 13,256 Bytes
75542dd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e989cce
 
 
 
75542dd
e989cce
 
 
 
 
75542dd
e989cce
 
75542dd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e989cce
75542dd
 
 
9ab14c5
 
e989cce
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
75542dd
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
---
license: apache-2.0
base_model: microsoft/conditional-detr-resnet-50
tags:
- generated_from_trainer
datasets:
- imagefolder
model-index:
- name: ms_detr_finetuned_diana
  results: []
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# ms_detr_finetuned_diana

This model is a fine-tuned version of [microsoft/conditional-detr-resnet-50](https://huggingface.co/microsoft/conditional-detr-resnet-50) on the imagefolder dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3578
- Map: 0.7134
- Map 50: 0.8181
- Map 75: 0.8181
- Map Small: -1.0
- Map Medium: 0.7864
- Map Large: 0.7101
- Mar 1: 0.1236
- Mar 10: 0.7964
- Mar 100: 0.825
- Mar Small: -1.0
- Mar Medium: 0.8
- Mar Large: 0.8302
- Map Per Class: -1.0
- Mar 100 Per Class: -1.0
- Classes: 0

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- num_epochs: 50

### Training results

| Training Loss | Epoch | Step | Validation Loss | Map    | Map 50 | Map 75 | Map Small | Map Medium | Map Large | Mar 1  | Mar 10 | Mar 100 | Mar Small | Mar Medium | Mar Large | Map Per Class | Mar 100 Per Class | Classes |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:----------:|:---------:|:------:|:------:|:-------:|:---------:|:----------:|:---------:|:-------------:|:-----------------:|:-------:|
| 2.1209        | 1.0   | 10   | 2.1209          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 1.4805        | 2.0   | 20   | 1.5062          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 1.205         | 3.0   | 30   | 1.3151          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 1.2767        | 4.0   | 40   | 1.1969          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 1.095         | 5.0   | 50   | 1.0561          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 0.8968        | 6.0   | 60   | 0.9283          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 0.8153        | 7.0   | 70   | 0.8459          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 0.7445        | 8.0   | 80   | 0.7019          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 0.5769        | 9.0   | 90   | 0.6067          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 0.5487        | 10.0  | 100  | 0.5308          | 0.0    | 0.0    | 0.0    | -1.0      | 0.0        | 0.0       | 0.0    | 0.0    | 0.0     | -1.0      | 0.0        | 0.0       | -1.0          | -1.0              | 0       |
| 0.4492        | 11.0  | 110  | 0.5069          | 0.0089 | 0.0099 | 0.0099 | -1.0      | 0.0        | 0.0089    | 0.0064 | 0.0064 | 0.0064  | -1.0      | 0.0        | 0.0078    | -1.0          | -1.0              | 0       |
| 0.3912        | 12.0  | 120  | 0.5000          | 0.0385 | 0.0438 | 0.0438 | -1.0      | 0.0        | 0.0461    | 0.0314 | 0.0386 | 0.0386  | -1.0      | 0.0        | 0.0466    | -1.0          | -1.0              | 0       |
| 0.3875        | 13.0  | 130  | 0.4488          | 0.0901 | 0.1036 | 0.1036 | -1.0      | 0.0        | 0.1072    | 0.0571 | 0.0886 | 0.0886  | -1.0      | 0.0        | 0.1069    | -1.0          | -1.0              | 0       |
| 0.4356        | 14.0  | 140  | 0.4592          | 0.3255 | 0.3789 | 0.3789 | -1.0      | 0.1382     | 0.3656    | 0.1007 | 0.3657 | 0.3657  | -1.0      | 0.1417     | 0.4121    | -1.0          | -1.0              | 0       |
| 0.3536        | 15.0  | 150  | 0.4293          | 0.3127 | 0.3673 | 0.3584 | -1.0      | 0.1914     | 0.3399    | 0.1064 | 0.3629 | 0.3629  | -1.0      | 0.1958     | 0.3974    | -1.0          | -1.0              | 0       |
| 0.3617        | 16.0  | 160  | 0.4128          | 0.4625 | 0.5397 | 0.5288 | -1.0      | 0.3614     | 0.4897    | 0.1164 | 0.5329 | 0.5329  | -1.0      | 0.3625     | 0.5681    | -1.0          | -1.0              | 0       |
| 0.392         | 17.0  | 170  | 0.4258          | 0.4683 | 0.5332 | 0.5332 | -1.0      | 0.3911     | 0.4868    | 0.1179 | 0.5486 | 0.5486  | -1.0      | 0.4        | 0.5793    | -1.0          | -1.0              | 0       |
| 0.3694        | 18.0  | 180  | 0.4563          | 0.4006 | 0.4614 | 0.4614 | -1.0      | 0.3313     | 0.4267    | 0.1157 | 0.4714 | 0.4714  | -1.0      | 0.3333     | 0.5       | -1.0          | -1.0              | 0       |
| 0.3569        | 19.0  | 190  | 0.4160          | 0.4912 | 0.5672 | 0.5672 | -1.0      | 0.386      | 0.5185    | 0.1157 | 0.58   | 0.58    | -1.0      | 0.4        | 0.6172    | -1.0          | -1.0              | 0       |
| 0.3839        | 20.0  | 200  | 0.4665          | 0.5324 | 0.6311 | 0.6212 | -1.0      | 0.4719     | 0.5561    | 0.115  | 0.6114 | 0.6114  | -1.0      | 0.475      | 0.6397    | -1.0          | -1.0              | 0       |
| 0.3123        | 21.0  | 210  | 0.4144          | 0.4808 | 0.5519 | 0.5519 | -1.0      | 0.3279     | 0.5235    | 0.1164 | 0.5643 | 0.5643  | -1.0      | 0.3333     | 0.6121    | -1.0          | -1.0              | 0       |
| 0.2824        | 22.0  | 220  | 0.3918          | 0.5587 | 0.6403 | 0.6403 | -1.0      | 0.468      | 0.5874    | 0.1186 | 0.6557 | 0.6557  | -1.0      | 0.4792     | 0.6922    | -1.0          | -1.0              | 0       |
| 0.2545        | 23.0  | 230  | 0.3530          | 0.5577 | 0.6299 | 0.6299 | -1.0      | 0.448      | 0.5846    | 0.1179 | 0.645  | 0.6514  | -1.0      | 0.4542     | 0.6922    | -1.0          | -1.0              | 0       |
| 0.2716        | 24.0  | 240  | 0.3540          | 0.6501 | 0.7455 | 0.7369 | -1.0      | 0.6292     | 0.6653    | 0.1236 | 0.7486 | 0.77    | -1.0      | 0.6375     | 0.7974    | -1.0          | -1.0              | 0       |
| 0.2631        | 25.0  | 250  | 0.3608          | 0.5918 | 0.6733 | 0.6733 | -1.0      | 0.5879     | 0.6012    | 0.1193 | 0.6936 | 0.6936  | -1.0      | 0.6        | 0.7129    | -1.0          | -1.0              | 0       |
| 0.2628        | 26.0  | 260  | 0.3607          | 0.6089 | 0.6904 | 0.6904 | -1.0      | 0.6516     | 0.608     | 0.1193 | 0.6943 | 0.7114  | -1.0      | 0.6583     | 0.7224    | -1.0          | -1.0              | 0       |
| 0.2653        | 27.0  | 270  | 0.3692          | 0.6648 | 0.7623 | 0.7538 | -1.0      | 0.7795     | 0.6512    | 0.1171 | 0.75   | 0.7771  | -1.0      | 0.8042     | 0.7716    | -1.0          | -1.0              | 0       |
| 0.2272        | 28.0  | 280  | 0.3657          | 0.5998 | 0.6814 | 0.6814 | -1.0      | 0.614      | 0.602     | 0.12   | 0.695  | 0.7007  | -1.0      | 0.6292     | 0.7155    | -1.0          | -1.0              | 0       |
| 0.3795        | 29.0  | 290  | 0.3728          | 0.6409 | 0.7284 | 0.7277 | -1.0      | 0.6901     | 0.6407    | 0.1264 | 0.7364 | 0.7486  | -1.0      | 0.7042     | 0.7578    | -1.0          | -1.0              | 0       |
| 0.2568        | 30.0  | 300  | 0.3724          | 0.6933 | 0.7956 | 0.7854 | -1.0      | 0.7381     | 0.6926    | 0.1236 | 0.7821 | 0.8043  | -1.0      | 0.7542     | 0.8147    | -1.0          | -1.0              | 0       |
| 0.2632        | 31.0  | 310  | 0.3741          | 0.6626 | 0.7614 | 0.7522 | -1.0      | 0.7747     | 0.651     | 0.1243 | 0.7479 | 0.7671  | -1.0      | 0.7958     | 0.7612    | -1.0          | -1.0              | 0       |
| 0.3576        | 32.0  | 320  | 0.3649          | 0.6734 | 0.7746 | 0.7746 | -1.0      | 0.7503     | 0.6686    | 0.1236 | 0.7586 | 0.7757  | -1.0      | 0.7667     | 0.7776    | -1.0          | -1.0              | 0       |
| 0.2254        | 33.0  | 330  | 0.3683          | 0.6991 | 0.8085 | 0.8084 | -1.0      | 0.7949     | 0.691     | 0.1243 | 0.7929 | 0.8121  | -1.0      | 0.8125     | 0.8121    | -1.0          | -1.0              | 0       |
| 0.2495        | 34.0  | 340  | 0.3459          | 0.6975 | 0.811  | 0.8021 | -1.0      | 0.7652     | 0.6919    | 0.1257 | 0.7793 | 0.805   | -1.0      | 0.7833     | 0.8095    | -1.0          | -1.0              | 0       |
| 0.2051        | 35.0  | 350  | 0.3508          | 0.6903 | 0.7939 | 0.7845 | -1.0      | 0.797      | 0.6835    | 0.1243 | 0.7693 | 0.7943  | -1.0      | 0.8167     | 0.7897    | -1.0          | -1.0              | 0       |
| 0.2159        | 36.0  | 360  | 0.3510          | 0.693  | 0.7971 | 0.7971 | -1.0      | 0.7619     | 0.6898    | 0.1214 | 0.7807 | 0.7986  | -1.0      | 0.7792     | 0.8026    | -1.0          | -1.0              | 0       |
| 0.2234        | 37.0  | 370  | 0.3512          | 0.7033 | 0.8062 | 0.8062 | -1.0      | 0.7588     | 0.7014    | 0.1236 | 0.78   | 0.8036  | -1.0      | 0.7708     | 0.8103    | -1.0          | -1.0              | 0       |
| 0.2732        | 38.0  | 380  | 0.3603          | 0.6916 | 0.7917 | 0.7917 | -1.0      | 0.6964     | 0.7019    | 0.1236 | 0.7857 | 0.7993  | -1.0      | 0.7083     | 0.8181    | -1.0          | -1.0              | 0       |
| 0.2397        | 39.0  | 390  | 0.3633          | 0.7141 | 0.8125 | 0.804  | -1.0      | 0.7074     | 0.7255    | 0.1264 | 0.7971 | 0.8186  | -1.0      | 0.7167     | 0.8397    | -1.0          | -1.0              | 0       |
| 0.2534        | 40.0  | 400  | 0.3574          | 0.7115 | 0.8104 | 0.8104 | -1.0      | 0.705      | 0.722     | 0.1236 | 0.7979 | 0.8179  | -1.0      | 0.7125     | 0.8397    | -1.0          | -1.0              | 0       |
| 0.2168        | 41.0  | 410  | 0.3547          | 0.7106 | 0.8087 | 0.8087 | -1.0      | 0.7594     | 0.7141    | 0.1257 | 0.8007 | 0.8229  | -1.0      | 0.7708     | 0.8336    | -1.0          | -1.0              | 0       |
| 0.2237        | 42.0  | 420  | 0.3590          | 0.7055 | 0.8105 | 0.8105 | -1.0      | 0.759      | 0.7089    | 0.1243 | 0.7964 | 0.8186  | -1.0      | 0.7708     | 0.8284    | -1.0          | -1.0              | 0       |
| 0.2152        | 43.0  | 430  | 0.3582          | 0.7132 | 0.82   | 0.82   | -1.0      | 0.7865     | 0.7109    | 0.1243 | 0.7971 | 0.8243  | -1.0      | 0.8        | 0.8293    | -1.0          | -1.0              | 0       |
| 0.1932        | 44.0  | 440  | 0.3612          | 0.7056 | 0.8112 | 0.8112 | -1.0      | 0.7825     | 0.7023    | 0.1229 | 0.7857 | 0.815   | -1.0      | 0.8        | 0.8181    | -1.0          | -1.0              | 0       |
| 0.1897        | 45.0  | 450  | 0.3557          | 0.7077 | 0.8105 | 0.8105 | -1.0      | 0.755      | 0.7111    | 0.1229 | 0.7957 | 0.8186  | -1.0      | 0.7667     | 0.8293    | -1.0          | -1.0              | 0       |
| 0.213         | 46.0  | 460  | 0.3557          | 0.7136 | 0.8193 | 0.8193 | -1.0      | 0.786      | 0.7101    | 0.1236 | 0.7943 | 0.8229  | -1.0      | 0.8        | 0.8276    | -1.0          | -1.0              | 0       |
| 0.2169        | 47.0  | 470  | 0.3567          | 0.7142 | 0.8192 | 0.8192 | -1.0      | 0.7864     | 0.711     | 0.1243 | 0.7957 | 0.8243  | -1.0      | 0.8        | 0.8293    | -1.0          | -1.0              | 0       |
| 0.1971        | 48.0  | 480  | 0.3577          | 0.713  | 0.8181 | 0.8181 | -1.0      | 0.7864     | 0.7097    | 0.1236 | 0.7957 | 0.8243  | -1.0      | 0.8        | 0.8293    | -1.0          | -1.0              | 0       |
| 0.2515        | 49.0  | 490  | 0.3580          | 0.7134 | 0.8181 | 0.8181 | -1.0      | 0.7864     | 0.71      | 0.1236 | 0.7964 | 0.825   | -1.0      | 0.8        | 0.8302    | -1.0          | -1.0              | 0       |
| 0.2874        | 50.0  | 500  | 0.3578          | 0.7134 | 0.8181 | 0.8181 | -1.0      | 0.7864     | 0.7101    | 0.1236 | 0.7964 | 0.825   | -1.0      | 0.8        | 0.8302    | -1.0          | -1.0              | 0       |


### Framework versions

- Transformers 4.41.1
- Pytorch 2.2.1+cu121
- Datasets 2.19.1
- Tokenizers 0.19.1