lombardata commited on
Commit
439b1fe
1 Parent(s): 5481b13

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +165 -120
README.md CHANGED
@@ -1,141 +1,186 @@
 
1
  ---
2
- license: apache-2.0
3
- base_model: facebook/dinov2-large
 
4
  tags:
 
 
5
  - generated_from_trainer
 
6
  model-index:
7
  - name: drone-DinoVdeau-from-probs-large-2024_11_15-batch-size64_freeze_probs
8
  results: []
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
 
14
- # drone-DinoVdeau-from-probs-large-2024_11_15-batch-size64_freeze_probs
15
 
16
- This model is a fine-tuned version of [facebook/dinov2-large](https://huggingface.co/facebook/dinov2-large) on the None dataset.
17
- It achieves the following results on the evaluation set:
18
  - Loss: 0.4672
19
- - Rmse: 0.1553
20
- - Mae: 0.1147
21
- - Kl Divergence: 0.3577
22
- - Explained Variance: 0.4654
23
- - Learning Rate: 0.0000
 
 
 
24
 
25
- ## Model description
26
 
27
- More information needed
28
 
29
- ## Intended uses & limitations
 
 
 
30
 
31
- More information needed
32
 
33
- ## Training and evaluation data
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
- More information needed
36
 
37
- ## Training procedure
38
 
39
- ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
- - learning_rate: 0.001
43
- - train_batch_size: 64
44
- - eval_batch_size: 64
45
- - seed: 42
46
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
- - lr_scheduler_type: linear
48
- - num_epochs: 150
49
- - mixed_precision_training: Native AMP
50
-
51
- ### Training results
52
-
53
- | Training Loss | Epoch | Step | Validation Loss | Rmse | Mae | Kl Divergence | Explained Variance | Rate |
54
- |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:-------------:|:------------------:|:------:|
55
- | No log | 1.0 | 110 | 0.5006 | 0.1904 | 0.1552 | 0.1025 | 0.3284 | 0.001 |
56
- | No log | 2.0 | 220 | 0.4755 | 0.1681 | 0.1245 | 0.5180 | 0.3932 | 0.001 |
57
- | No log | 3.0 | 330 | 0.4745 | 0.1675 | 0.1227 | 0.6862 | 0.3975 | 0.001 |
58
- | No log | 4.0 | 440 | 0.4742 | 0.1672 | 0.1255 | 0.3212 | 0.4024 | 0.001 |
59
- | 0.5081 | 5.0 | 550 | 0.4725 | 0.1653 | 0.1224 | 0.5072 | 0.4118 | 0.001 |
60
- | 0.5081 | 6.0 | 660 | 0.4726 | 0.1657 | 0.1216 | 0.6710 | 0.4101 | 0.001 |
61
- | 0.5081 | 7.0 | 770 | 0.4732 | 0.1655 | 0.1255 | 0.3162 | 0.4183 | 0.001 |
62
- | 0.5081 | 8.0 | 880 | 0.4728 | 0.1651 | 0.1260 | 0.2719 | 0.4234 | 0.001 |
63
- | 0.5081 | 9.0 | 990 | 0.4708 | 0.1639 | 0.1206 | 0.6393 | 0.4237 | 0.001 |
64
- | 0.4668 | 10.0 | 1100 | 0.4733 | 0.1654 | 0.1230 | 0.5359 | 0.4151 | 0.001 |
65
- | 0.4668 | 11.0 | 1210 | 0.4716 | 0.1647 | 0.1253 | 0.2479 | 0.4305 | 0.001 |
66
- | 0.4668 | 12.0 | 1320 | 0.4708 | 0.1631 | 0.1244 | 0.3119 | 0.4358 | 0.001 |
67
- | 0.4668 | 13.0 | 1430 | 0.4715 | 0.1635 | 0.1230 | 0.3694 | 0.4274 | 0.001 |
68
- | 0.4641 | 14.0 | 1540 | 0.4721 | 0.1653 | 0.1216 | 0.5592 | 0.4134 | 0.001 |
69
- | 0.4641 | 15.0 | 1650 | 0.4701 | 0.1628 | 0.1213 | 0.4936 | 0.4314 | 0.001 |
70
- | 0.4641 | 16.0 | 1760 | 0.4719 | 0.1646 | 0.1229 | 0.2820 | 0.4328 | 0.001 |
71
- | 0.4641 | 17.0 | 1870 | 0.4693 | 0.1621 | 0.1200 | 0.5294 | 0.4332 | 0.001 |
72
- | 0.4641 | 18.0 | 1980 | 0.4710 | 0.1635 | 0.1216 | 0.4093 | 0.4294 | 0.001 |
73
- | 0.4618 | 19.0 | 2090 | 0.4698 | 0.1622 | 0.1219 | 0.2918 | 0.4388 | 0.001 |
74
- | 0.4618 | 20.0 | 2200 | 0.4692 | 0.1617 | 0.1190 | 0.4772 | 0.4355 | 0.001 |
75
- | 0.4618 | 21.0 | 2310 | 0.4683 | 0.1606 | 0.1204 | 0.4336 | 0.4424 | 0.001 |
76
- | 0.4618 | 22.0 | 2420 | 0.4724 | 0.1650 | 0.1183 | 0.7962 | 0.4233 | 0.001 |
77
- | 0.4613 | 23.0 | 2530 | 0.4714 | 0.1641 | 0.1223 | 0.2854 | 0.4354 | 0.001 |
78
- | 0.4613 | 24.0 | 2640 | 0.4707 | 0.1633 | 0.1207 | 0.4206 | 0.4280 | 0.001 |
79
- | 0.4613 | 25.0 | 2750 | 0.4679 | 0.1606 | 0.1185 | 0.5436 | 0.4416 | 0.001 |
80
- | 0.4613 | 26.0 | 2860 | 0.4708 | 0.1634 | 0.1192 | 0.4964 | 0.4268 | 0.001 |
81
- | 0.4613 | 27.0 | 2970 | 0.4695 | 0.1625 | 0.1185 | 0.6399 | 0.4301 | 0.001 |
82
- | 0.4607 | 28.0 | 3080 | 0.4701 | 0.1624 | 0.1184 | 0.5737 | 0.4324 | 0.001 |
83
- | 0.4607 | 29.0 | 3190 | 0.4699 | 0.1624 | 0.1200 | 0.4459 | 0.4324 | 0.001 |
84
- | 0.4607 | 30.0 | 3300 | 0.4723 | 0.1643 | 0.1254 | 0.2726 | 0.4308 | 0.001 |
85
- | 0.4607 | 31.0 | 3410 | 0.4696 | 0.1622 | 0.1184 | 0.5308 | 0.4313 | 0.001 |
86
- | 0.4604 | 32.0 | 3520 | 0.4668 | 0.1593 | 0.1175 | 0.4200 | 0.4508 | 0.0001 |
87
- | 0.4604 | 33.0 | 3630 | 0.4663 | 0.1587 | 0.1177 | 0.3529 | 0.4565 | 0.0001 |
88
- | 0.4604 | 34.0 | 3740 | 0.4667 | 0.1592 | 0.1181 | 0.3588 | 0.4542 | 0.0001 |
89
- | 0.4604 | 35.0 | 3850 | 0.4659 | 0.1584 | 0.1160 | 0.4813 | 0.4545 | 0.0001 |
90
- | 0.4604 | 36.0 | 3960 | 0.4658 | 0.1581 | 0.1173 | 0.3504 | 0.4594 | 0.0001 |
91
- | 0.4565 | 37.0 | 4070 | 0.4654 | 0.1578 | 0.1158 | 0.3919 | 0.4608 | 0.0001 |
92
- | 0.4565 | 38.0 | 4180 | 0.4655 | 0.1580 | 0.1166 | 0.4058 | 0.4583 | 0.0001 |
93
- | 0.4565 | 39.0 | 4290 | 0.4658 | 0.1585 | 0.1174 | 0.4118 | 0.4567 | 0.0001 |
94
- | 0.4565 | 40.0 | 4400 | 0.4656 | 0.1579 | 0.1170 | 0.3564 | 0.4607 | 0.0001 |
95
- | 0.4552 | 41.0 | 4510 | 0.4657 | 0.1582 | 0.1171 | 0.3573 | 0.4598 | 0.0001 |
96
- | 0.4552 | 42.0 | 4620 | 0.4652 | 0.1579 | 0.1155 | 0.5042 | 0.4587 | 0.0001 |
97
- | 0.4552 | 43.0 | 4730 | 0.4651 | 0.1575 | 0.1157 | 0.4462 | 0.4613 | 0.0001 |
98
- | 0.4552 | 44.0 | 4840 | 0.4654 | 0.1579 | 0.1166 | 0.4236 | 0.4604 | 0.0001 |
99
- | 0.4552 | 45.0 | 4950 | 0.4649 | 0.1574 | 0.1151 | 0.4510 | 0.4625 | 0.0001 |
100
- | 0.4538 | 46.0 | 5060 | 0.4648 | 0.1575 | 0.1157 | 0.4490 | 0.4619 | 0.0001 |
101
- | 0.4538 | 47.0 | 5170 | 0.4649 | 0.1574 | 0.1152 | 0.4751 | 0.4615 | 0.0001 |
102
- | 0.4538 | 48.0 | 5280 | 0.4648 | 0.1575 | 0.1151 | 0.5305 | 0.4631 | 0.0001 |
103
- | 0.4538 | 49.0 | 5390 | 0.4648 | 0.1574 | 0.1154 | 0.4799 | 0.4630 | 0.0001 |
104
- | 0.4532 | 50.0 | 5500 | 0.4650 | 0.1572 | 0.1172 | 0.2825 | 0.4694 | 0.0001 |
105
- | 0.4532 | 51.0 | 5610 | 0.4656 | 0.1582 | 0.1151 | 0.4879 | 0.4573 | 0.0001 |
106
- | 0.4532 | 52.0 | 5720 | 0.4643 | 0.1566 | 0.1155 | 0.4199 | 0.4674 | 0.0001 |
107
- | 0.4532 | 53.0 | 5830 | 0.4644 | 0.1569 | 0.1156 | 0.3880 | 0.4673 | 0.0001 |
108
- | 0.4532 | 54.0 | 5940 | 0.4646 | 0.1569 | 0.1148 | 0.4229 | 0.4654 | 0.0001 |
109
- | 0.4526 | 55.0 | 6050 | 0.4644 | 0.1569 | 0.1159 | 0.4009 | 0.4659 | 0.0001 |
110
- | 0.4526 | 56.0 | 6160 | 0.4647 | 0.1572 | 0.1164 | 0.3405 | 0.4660 | 0.0001 |
111
- | 0.4526 | 57.0 | 6270 | 0.4645 | 0.1569 | 0.1152 | 0.4188 | 0.4661 | 0.0001 |
112
- | 0.4526 | 58.0 | 6380 | 0.4651 | 0.1576 | 0.1164 | 0.3079 | 0.4659 | 0.0001 |
113
- | 0.4526 | 59.0 | 6490 | 0.4645 | 0.1570 | 0.1150 | 0.4339 | 0.4654 | 1e-05 |
114
- | 0.4514 | 60.0 | 6600 | 0.4642 | 0.1566 | 0.1150 | 0.3894 | 0.4679 | 1e-05 |
115
- | 0.4514 | 61.0 | 6710 | 0.4639 | 0.1563 | 0.1146 | 0.4145 | 0.4693 | 1e-05 |
116
- | 0.4514 | 62.0 | 6820 | 0.4641 | 0.1565 | 0.1148 | 0.4064 | 0.4686 | 1e-05 |
117
- | 0.4514 | 63.0 | 6930 | 0.4643 | 0.1565 | 0.1149 | 0.3542 | 0.4698 | 1e-05 |
118
- | 0.4511 | 64.0 | 7040 | 0.4640 | 0.1564 | 0.1150 | 0.3718 | 0.4702 | 1e-05 |
119
- | 0.4511 | 65.0 | 7150 | 0.4641 | 0.1565 | 0.1152 | 0.4128 | 0.4680 | 1e-05 |
120
- | 0.4511 | 66.0 | 7260 | 0.4644 | 0.1570 | 0.1145 | 0.4988 | 0.4658 | 1e-05 |
121
- | 0.4511 | 67.0 | 7370 | 0.4638 | 0.1562 | 0.1151 | 0.4122 | 0.4697 | 1e-05 |
122
- | 0.4511 | 68.0 | 7480 | 0.4640 | 0.1565 | 0.1144 | 0.4579 | 0.4674 | 1e-05 |
123
- | 0.4508 | 69.0 | 7590 | 0.4638 | 0.1561 | 0.1143 | 0.4197 | 0.4702 | 1e-05 |
124
- | 0.4508 | 70.0 | 7700 | 0.4639 | 0.1563 | 0.1145 | 0.4286 | 0.4695 | 1e-05 |
125
- | 0.4508 | 71.0 | 7810 | 0.4641 | 0.1563 | 0.1153 | 0.3542 | 0.4708 | 1e-05 |
126
- | 0.4508 | 72.0 | 7920 | 0.4642 | 0.1566 | 0.1147 | 0.4250 | 0.4681 | 1e-05 |
127
- | 0.4505 | 73.0 | 8030 | 0.4638 | 0.1561 | 0.1140 | 0.4397 | 0.4700 | 1e-05 |
128
- | 0.4505 | 74.0 | 8140 | 0.4638 | 0.1563 | 0.1145 | 0.4437 | 0.4689 | 1e-05 |
129
- | 0.4505 | 75.0 | 8250 | 0.4638 | 0.1561 | 0.1145 | 0.4049 | 0.4705 | 1e-05 |
130
- | 0.4505 | 76.0 | 8360 | 0.4640 | 0.1565 | 0.1141 | 0.4926 | 0.4675 | 0.0000 |
131
- | 0.4505 | 77.0 | 8470 | 0.4639 | 0.1562 | 0.1142 | 0.4427 | 0.4695 | 0.0000 |
132
- | 0.4505 | 78.0 | 8580 | 0.4639 | 0.1563 | 0.1145 | 0.4293 | 0.4692 | 0.0000 |
133
- | 0.4505 | 79.0 | 8690 | 0.4641 | 0.1564 | 0.1147 | 0.3765 | 0.4700 | 0.0000 |
134
-
135
-
136
- ### Framework versions
137
-
138
- - Transformers 4.41.0
139
- - Pytorch 2.5.0+cu124
140
- - Datasets 3.0.2
141
- - Tokenizers 0.19.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
  ---
3
+ language:
4
+ - eng
5
+ license: cc0-1.0
6
  tags:
7
+ - multilabel-image-classification
8
+ - multilabel
9
  - generated_from_trainer
10
+ base_model: drone-DinoVdeau-from-probs-large-2024_11_15-batch-size64_freeze_probs
11
  model-index:
12
  - name: drone-DinoVdeau-from-probs-large-2024_11_15-batch-size64_freeze_probs
13
  results: []
14
  ---
15
 
16
+ drone-DinoVdeau-from-probs is a fine-tuned version of [drone-DinoVdeau-from-probs-large-2024_11_15-batch-size64_freeze_probs](https://huggingface.co/drone-DinoVdeau-from-probs-large-2024_11_15-batch-size64_freeze_probs). It achieves the following results on the test set:
 
17
 
 
18
 
 
 
19
  - Loss: 0.4672
20
+ - RMSE: 0.1553
21
+ - MAE: 0.1147
22
+ - KL Divergence: 0.3577
23
+
24
+ ---
25
+
26
+ # Model description
27
+ drone-DinoVdeau-from-probs is a model built on top of drone-DinoVdeau-from-probs-large-2024_11_15-batch-size64_freeze_probs model for underwater multilabel image classification.The classification head is a combination of linear, ReLU, batch normalization, and dropout layers.
28
 
29
+ The source code for training the model can be found in this [Git repository](https://github.com/SeatizenDOI/DinoVdeau).
30
 
31
+ - **Developed by:** [lombardata](https://huggingface.co/lombardata), credits to [César Leblanc](https://huggingface.co/CesarLeblanc) and [Victor Illien](https://huggingface.co/groderg)
32
 
33
+ ---
34
+
35
+ # Intended uses & limitations
36
+ You can use the raw model for classify diverse marine species, encompassing coral morphotypes classes taken from the Global Coral Reef Monitoring Network (GCRMN), habitats classes and seagrass species.
37
 
38
+ ---
39
 
40
+ # Training and evaluation data
41
+ Details on the estimated number of images for each class are given in the following table:
42
+ | Class | train | test | val | Total |
43
+ |:------------------------|--------:|-------:|------:|--------:|
44
+ | Acropore_branched | 1220 | 363 | 362 | 1945 |
45
+ | Acropore_digitised | 586 | 195 | 189 | 970 |
46
+ | Acropore_tabular | 308 | 133 | 119 | 560 |
47
+ | Algae | 4777 | 1372 | 1384 | 7533 |
48
+ | Dead_coral | 2513 | 671 | 693 | 3877 |
49
+ | Millepore | 136 | 55 | 59 | 250 |
50
+ | No_acropore_encrusting | 252 | 88 | 93 | 433 |
51
+ | No_acropore_massive | 2158 | 725 | 726 | 3609 |
52
+ | No_acropore_sub_massive | 2036 | 582 | 612 | 3230 |
53
+ | Rock | 5976 | 1941 | 1928 | 9845 |
54
+ | Rubble | 4851 | 1486 | 1474 | 7811 |
55
+ | Sand | 6155 | 2019 | 1990 | 10164 |
56
 
57
+ ---
58
 
59
+ # Training procedure
60
 
61
+ ## Training hyperparameters
62
 
63
  The following hyperparameters were used during training:
64
+
65
+ - **Number of Epochs**: 79.0
66
+ - **Learning Rate**: 0.001
67
+ - **Train Batch Size**: 64
68
+ - **Eval Batch Size**: 64
69
+ - **Optimizer**: Adam
70
+ - **LR Scheduler Type**: ReduceLROnPlateau with a patience of 5 epochs and a factor of 0.1
71
+ - **Freeze Encoder**: Yes
72
+ - **Data Augmentation**: Yes
73
+
74
+
75
+ ## Data Augmentation
76
+ Data were augmented using the following transformations :
77
+
78
+ Train Transforms
79
+ - **PreProcess**: No additional parameters
80
+ - **Resize**: probability=1.00
81
+ - **RandomHorizontalFlip**: probability=0.25
82
+ - **RandomVerticalFlip**: probability=0.25
83
+ - **ColorJiggle**: probability=0.25
84
+ - **RandomPerspective**: probability=0.25
85
+ - **Normalize**: probability=1.00
86
+
87
+ Val Transforms
88
+ - **PreProcess**: No additional parameters
89
+ - **Resize**: probability=1.00
90
+ - **Normalize**: probability=1.00
91
+
92
+
93
+
94
+ ## Training results
95
+ Epoch | Validation Loss | MAE | RMSE | KL div | Learning Rate
96
+ --- | --- | --- | --- | --- | ---
97
+ 1 | 0.5005590319633484 | 0.1552 | 0.1904 | 0.1025 | 0.001
98
+ 2 | 0.47547808289527893 | 0.1245 | 0.1681 | 0.5180 | 0.001
99
+ 3 | 0.47452571988105774 | 0.1227 | 0.1675 | 0.6862 | 0.001
100
+ 4 | 0.47420722246170044 | 0.1255 | 0.1672 | 0.3212 | 0.001
101
+ 5 | 0.47245556116104126 | 0.1224 | 0.1653 | 0.5072 | 0.001
102
+ 6 | 0.4725925624370575 | 0.1216 | 0.1657 | 0.6710 | 0.001
103
+ 7 | 0.4731809198856354 | 0.1255 | 0.1655 | 0.3162 | 0.001
104
+ 8 | 0.47284314036369324 | 0.1260 | 0.1651 | 0.2719 | 0.001
105
+ 9 | 0.4707973003387451 | 0.1206 | 0.1639 | 0.6393 | 0.001
106
+ 10 | 0.4732784628868103 | 0.1230 | 0.1654 | 0.5359 | 0.001
107
+ 11 | 0.47162503004074097 | 0.1253 | 0.1647 | 0.2479 | 0.001
108
+ 12 | 0.47083696722984314 | 0.1244 | 0.1631 | 0.3119 | 0.001
109
+ 13 | 0.47152063250541687 | 0.1230 | 0.1635 | 0.3694 | 0.001
110
+ 14 | 0.47212228178977966 | 0.1216 | 0.1653 | 0.5592 | 0.001
111
+ 15 | 0.47012239694595337 | 0.1213 | 0.1628 | 0.4936 | 0.001
112
+ 16 | 0.4718552827835083 | 0.1229 | 0.1646 | 0.2820 | 0.001
113
+ 17 | 0.46933484077453613 | 0.1200 | 0.1621 | 0.5294 | 0.001
114
+ 18 | 0.4710436165332794 | 0.1216 | 0.1635 | 0.4093 | 0.001
115
+ 19 | 0.4698491394519806 | 0.1219 | 0.1622 | 0.2918 | 0.001
116
+ 20 | 0.4691685736179352 | 0.1190 | 0.1617 | 0.4772 | 0.001
117
+ 21 | 0.46830564737319946 | 0.1204 | 0.1606 | 0.4336 | 0.001
118
+ 22 | 0.47239789366722107 | 0.1183 | 0.1650 | 0.7962 | 0.001
119
+ 23 | 0.47136834263801575 | 0.1223 | 0.1641 | 0.2854 | 0.001
120
+ 24 | 0.4706868529319763 | 0.1207 | 0.1633 | 0.4206 | 0.001
121
+ 25 | 0.46786901354789734 | 0.1185 | 0.1606 | 0.5436 | 0.001
122
+ 26 | 0.47084224224090576 | 0.1192 | 0.1634 | 0.4964 | 0.001
123
+ 27 | 0.4695045053958893 | 0.1185 | 0.1625 | 0.6399 | 0.001
124
+ 28 | 0.4700873792171478 | 0.1184 | 0.1624 | 0.5737 | 0.001
125
+ 29 | 0.4698559045791626 | 0.1200 | 0.1624 | 0.4459 | 0.001
126
+ 30 | 0.4722815454006195 | 0.1254 | 0.1643 | 0.2726 | 0.001
127
+ 31 | 0.46958214044570923 | 0.1184 | 0.1622 | 0.5308 | 0.001
128
+ 32 | 0.46677276492118835 | 0.1175 | 0.1593 | 0.4200 | 0.0001
129
+ 33 | 0.46626824140548706 | 0.1177 | 0.1587 | 0.3529 | 0.0001
130
+ 34 | 0.46665358543395996 | 0.1181 | 0.1592 | 0.3588 | 0.0001
131
+ 35 | 0.46587392687797546 | 0.1160 | 0.1584 | 0.4813 | 0.0001
132
+ 36 | 0.46578526496887207 | 0.1173 | 0.1581 | 0.3504 | 0.0001
133
+ 37 | 0.4654408395290375 | 0.1158 | 0.1578 | 0.3919 | 0.0001
134
+ 38 | 0.46546319127082825 | 0.1166 | 0.1580 | 0.4058 | 0.0001
135
+ 39 | 0.465843141078949 | 0.1174 | 0.1585 | 0.4118 | 0.0001
136
+ 40 | 0.46561121940612793 | 0.1170 | 0.1579 | 0.3564 | 0.0001
137
+ 41 | 0.4657152593135834 | 0.1171 | 0.1582 | 0.3573 | 0.0001
138
+ 42 | 0.4651602804660797 | 0.1155 | 0.1579 | 0.5042 | 0.0001
139
+ 43 | 0.4651065468788147 | 0.1157 | 0.1575 | 0.4462 | 0.0001
140
+ 44 | 0.46537330746650696 | 0.1166 | 0.1579 | 0.4236 | 0.0001
141
+ 45 | 0.46489208936691284 | 0.1151 | 0.1574 | 0.4510 | 0.0001
142
+ 46 | 0.46484702825546265 | 0.1157 | 0.1575 | 0.4490 | 0.0001
143
+ 47 | 0.4648602306842804 | 0.1152 | 0.1574 | 0.4751 | 0.0001
144
+ 48 | 0.4647873342037201 | 0.1151 | 0.1575 | 0.5305 | 0.0001
145
+ 49 | 0.4647849500179291 | 0.1154 | 0.1574 | 0.4799 | 0.0001
146
+ 50 | N/A | 0.0000 | 0.0000 | 0.0000 | 0.0001
147
+ 51 | 0.465638667345047 | 0.1151 | 0.1582 | 0.4879 | 0.0001
148
+ 52 | 0.46429532766342163 | 0.1155 | 0.1566 | 0.4199 | 0.0001
149
+ 53 | 0.46441230177879333 | 0.1156 | 0.1569 | 0.3880 | 0.0001
150
+ 54 | 0.4646008610725403 | 0.1148 | 0.1569 | 0.4229 | 0.0001
151
+ 55 | 0.4644174873828888 | 0.1159 | 0.1569 | 0.4009 | 0.0001
152
+ 56 | 0.464743047952652 | 0.1164 | 0.1572 | 0.3405 | 0.0001
153
+ 57 | 0.4645179808139801 | 0.1152 | 0.1569 | 0.4188 | 0.0001
154
+ 58 | 0.465102881193161 | 0.1164 | 0.1576 | 0.3079 | 0.0001
155
+ 59 | 0.4644688367843628 | 0.1150 | 0.1570 | 0.4339 | 1e-05
156
+ 60 | 0.46417686343193054 | 0.1150 | 0.1566 | 0.3894 | 1e-05
157
+ 61 | 0.4639436900615692 | 0.1146 | 0.1563 | 0.4145 | 1e-05
158
+ 62 | 0.4641311764717102 | 0.1148 | 0.1565 | 0.4064 | 1e-05
159
+ 63 | 0.4643491506576538 | 0.1149 | 0.1565 | 0.3542 | 1e-05
160
+ 64 | 0.46402981877326965 | 0.1150 | 0.1564 | 0.3718 | 1e-05
161
+ 65 | 0.4640822410583496 | 0.1152 | 0.1565 | 0.4128 | 1e-05
162
+ 66 | 0.46441909670829773 | 0.1145 | 0.1570 | 0.4988 | 1e-05
163
+ 67 | 0.46383005380630493 | 0.1151 | 0.1562 | 0.4122 | 1e-05
164
+ 68 | 0.4639807641506195 | 0.1144 | 0.1565 | 0.4579 | 1e-05
165
+ 69 | 0.4637599587440491 | 0.1143 | 0.1561 | 0.4197 | 1e-05
166
+ 70 | 0.46392253041267395 | 0.1145 | 0.1563 | 0.4286 | 1e-05
167
+ 71 | 0.46406444907188416 | 0.1153 | 0.1563 | 0.3542 | 1e-05
168
+ 72 | 0.46417826414108276 | 0.1147 | 0.1566 | 0.4250 | 1e-05
169
+ 73 | 0.4637835919857025 | 0.1140 | 0.1561 | 0.4397 | 1e-05
170
+ 74 | 0.463798850774765 | 0.1145 | 0.1563 | 0.4437 | 1e-05
171
+ 75 | 0.46379053592681885 | 0.1145 | 0.1561 | 0.4049 | 1e-05
172
+ 76 | 0.4639701247215271 | 0.1141 | 0.1565 | 0.4926 | 1.0000000000000002e-06
173
+ 77 | 0.463869571685791 | 0.1142 | 0.1562 | 0.4427 | 1.0000000000000002e-06
174
+ 78 | 0.46388140320777893 | 0.1145 | 0.1563 | 0.4293 | 1.0000000000000002e-06
175
+ 79 | 0.46412238478660583 | 0.1147 | 0.1564 | 0.3765 | 1.0000000000000002e-06
176
+
177
+
178
+ ---
179
+
180
+ # Framework Versions
181
+
182
+ - **Transformers**: 4.41.0
183
+ - **Pytorch**: 2.5.0+cu124
184
+ - **Datasets**: 3.0.2
185
+ - **Tokenizers**: 0.19.1
186
+