Model save
Browse files- README.md +165 -0
- logs/events.out.tfevents.1731741098.AZHAR-PC.1610705.0 +2 -2
- model.safetensors +1 -1
README.md
ADDED
@@ -0,0 +1,165 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
license: apache-2.0
|
4 |
+
base_model: microsoft/swin-large-patch4-window12-384
|
5 |
+
tags:
|
6 |
+
- generated_from_trainer
|
7 |
+
metrics:
|
8 |
+
- accuracy
|
9 |
+
- f1
|
10 |
+
model-index:
|
11 |
+
- name: swin-transformer3
|
12 |
+
results: []
|
13 |
+
---
|
14 |
+
|
15 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
16 |
+
should probably proofread and complete it, then remove this comment. -->
|
17 |
+
|
18 |
+
# swin-transformer3
|
19 |
+
|
20 |
+
This model is a fine-tuned version of [microsoft/swin-large-patch4-window12-384](https://huggingface.co/microsoft/swin-large-patch4-window12-384) on the None dataset.
|
21 |
+
It achieves the following results on the evaluation set:
|
22 |
+
- Loss: 4.1081
|
23 |
+
- Accuracy: 0.5667
|
24 |
+
- F1: 0.5667
|
25 |
+
|
26 |
+
## Model description
|
27 |
+
|
28 |
+
More information needed
|
29 |
+
|
30 |
+
## Intended uses & limitations
|
31 |
+
|
32 |
+
More information needed
|
33 |
+
|
34 |
+
## Training and evaluation data
|
35 |
+
|
36 |
+
More information needed
|
37 |
+
|
38 |
+
## Training procedure
|
39 |
+
|
40 |
+
### Training hyperparameters
|
41 |
+
|
42 |
+
The following hyperparameters were used during training:
|
43 |
+
- learning_rate: 5e-05
|
44 |
+
- train_batch_size: 4
|
45 |
+
- eval_batch_size: 4
|
46 |
+
- seed: 42
|
47 |
+
- gradient_accumulation_steps: 4
|
48 |
+
- total_train_batch_size: 16
|
49 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
50 |
+
- lr_scheduler_type: linear
|
51 |
+
- lr_scheduler_warmup_ratio: 0.1
|
52 |
+
- num_epochs: 100
|
53 |
+
|
54 |
+
### Training results
|
55 |
+
|
56 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|
57 |
+
|:-------------:|:-------:|:----:|:---------------:|:--------:|:------:|
|
58 |
+
| 1.6659 | 0.9925 | 33 | 1.0639 | 0.6333 | 0.6250 |
|
59 |
+
| 0.7561 | 1.9850 | 66 | 0.7258 | 0.5167 | 0.3520 |
|
60 |
+
| 0.7106 | 2.9774 | 99 | 0.7334 | 0.5 | 0.3755 |
|
61 |
+
| 0.6749 | 4.0 | 133 | 0.7088 | 0.4833 | 0.3661 |
|
62 |
+
| 0.751 | 4.9925 | 166 | 0.7356 | 0.4833 | 0.3661 |
|
63 |
+
| 0.7146 | 5.9850 | 199 | 0.7837 | 0.4833 | 0.3150 |
|
64 |
+
| 0.6699 | 6.9774 | 232 | 0.7569 | 0.4833 | 0.3424 |
|
65 |
+
| 0.6521 | 8.0 | 266 | 0.7255 | 0.5333 | 0.4674 |
|
66 |
+
| 0.6885 | 8.9925 | 299 | 0.7253 | 0.5167 | 0.4070 |
|
67 |
+
| 0.6407 | 9.9850 | 332 | 0.6506 | 0.6 | 0.5909 |
|
68 |
+
| 0.6436 | 10.9774 | 365 | 0.6720 | 0.55 | 0.4442 |
|
69 |
+
| 0.7865 | 12.0 | 399 | 0.6606 | 0.55 | 0.4792 |
|
70 |
+
| 0.7191 | 12.9925 | 432 | 0.6407 | 0.65 | 0.6466 |
|
71 |
+
| 0.5889 | 13.9850 | 465 | 0.8008 | 0.4833 | 0.3619 |
|
72 |
+
| 0.5489 | 14.9774 | 498 | 0.7298 | 0.5333 | 0.4674 |
|
73 |
+
| 0.596 | 16.0 | 532 | 0.7465 | 0.6667 | 0.6591 |
|
74 |
+
| 0.6136 | 16.9925 | 565 | 0.9118 | 0.5333 | 0.4692 |
|
75 |
+
| 0.5961 | 17.9850 | 598 | 0.6902 | 0.65 | 0.6298 |
|
76 |
+
| 0.6327 | 18.9774 | 631 | 0.8260 | 0.5667 | 0.5190 |
|
77 |
+
| 0.6518 | 20.0 | 665 | 0.6919 | 0.5833 | 0.5715 |
|
78 |
+
| 0.5551 | 20.9925 | 698 | 1.1780 | 0.55 | 0.516 |
|
79 |
+
| 0.511 | 21.9850 | 731 | 0.7414 | 0.6 | 0.6 |
|
80 |
+
| 0.4749 | 22.9774 | 764 | 0.7978 | 0.6167 | 0.6129 |
|
81 |
+
| 0.4607 | 24.0 | 798 | 0.8087 | 0.55 | 0.5420 |
|
82 |
+
| 0.5837 | 24.9925 | 831 | 0.8271 | 0.5667 | 0.5456 |
|
83 |
+
| 0.4608 | 25.9850 | 864 | 0.8539 | 0.6 | 0.5863 |
|
84 |
+
| 0.536 | 26.9774 | 897 | 0.9802 | 0.5333 | 0.5026 |
|
85 |
+
| 0.4225 | 28.0 | 931 | 0.9275 | 0.6 | 0.5910 |
|
86 |
+
| 0.4325 | 28.9925 | 964 | 0.8834 | 0.6167 | 0.6099 |
|
87 |
+
| 0.4874 | 29.9850 | 997 | 0.8721 | 0.6167 | 0.6168 |
|
88 |
+
| 0.4165 | 30.9774 | 1030 | 1.0360 | 0.6167 | 0.6163 |
|
89 |
+
| 0.4773 | 32.0 | 1064 | 1.2210 | 0.5833 | 0.5759 |
|
90 |
+
| 0.3756 | 32.9925 | 1097 | 1.1291 | 0.5833 | 0.5830 |
|
91 |
+
| 0.636 | 33.9850 | 1130 | 1.0178 | 0.5833 | 0.5830 |
|
92 |
+
| 0.5474 | 34.9774 | 1163 | 0.9479 | 0.5667 | 0.5608 |
|
93 |
+
| 0.3462 | 36.0 | 1197 | 0.9585 | 0.6167 | 0.6163 |
|
94 |
+
| 0.3057 | 36.9925 | 1230 | 1.2014 | 0.6167 | 0.6163 |
|
95 |
+
| 0.2304 | 37.9850 | 1263 | 1.1975 | 0.6333 | 0.6333 |
|
96 |
+
| 0.2628 | 38.9774 | 1296 | 1.5224 | 0.5833 | 0.5793 |
|
97 |
+
| 0.3774 | 40.0 | 1330 | 1.2903 | 0.5667 | 0.5516 |
|
98 |
+
| 0.2604 | 40.9925 | 1363 | 1.4082 | 0.5667 | 0.5608 |
|
99 |
+
| 0.2522 | 41.9850 | 1396 | 1.1783 | 0.6167 | 0.6163 |
|
100 |
+
| 0.1925 | 42.9774 | 1429 | 1.3613 | 0.6167 | 0.6163 |
|
101 |
+
| 0.3436 | 44.0 | 1463 | 1.6383 | 0.5333 | 0.5173 |
|
102 |
+
| 0.1955 | 44.9925 | 1496 | 1.8947 | 0.5 | 0.4829 |
|
103 |
+
| 0.2206 | 45.9850 | 1529 | 1.4390 | 0.6 | 0.6 |
|
104 |
+
| 0.1912 | 46.9774 | 1562 | 1.5288 | 0.65 | 0.6400 |
|
105 |
+
| 0.2794 | 48.0 | 1596 | 1.7393 | 0.55 | 0.5420 |
|
106 |
+
| 0.3166 | 48.9925 | 1629 | 2.0414 | 0.5667 | 0.5608 |
|
107 |
+
| 0.173 | 49.9850 | 1662 | 1.6377 | 0.6 | 0.5991 |
|
108 |
+
| 0.1375 | 50.9774 | 1695 | 1.6228 | 0.6 | 0.6 |
|
109 |
+
| 0.2659 | 52.0 | 1729 | 1.6452 | 0.6333 | 0.6333 |
|
110 |
+
| 0.2045 | 52.9925 | 1762 | 1.9706 | 0.5667 | 0.5608 |
|
111 |
+
| 0.1081 | 53.9850 | 1795 | 1.9546 | 0.6167 | 0.6009 |
|
112 |
+
| 0.1782 | 54.9774 | 1828 | 2.1268 | 0.5667 | 0.5608 |
|
113 |
+
| 0.244 | 56.0 | 1862 | 1.8301 | 0.6167 | 0.6098 |
|
114 |
+
| 0.1783 | 56.9925 | 1895 | 2.5808 | 0.5667 | 0.5071 |
|
115 |
+
| 0.2429 | 57.9850 | 1928 | 2.1214 | 0.6167 | 0.6059 |
|
116 |
+
| 0.2 | 58.9774 | 1961 | 2.2282 | 0.5667 | 0.5657 |
|
117 |
+
| 0.1646 | 60.0 | 1995 | 2.3272 | 0.5833 | 0.5662 |
|
118 |
+
| 0.1663 | 60.9925 | 2028 | 2.4723 | 0.5333 | 0.5323 |
|
119 |
+
| 0.1935 | 61.9850 | 2061 | 2.3384 | 0.6 | 0.5973 |
|
120 |
+
| 0.2079 | 62.9774 | 2094 | 1.9271 | 0.5833 | 0.5830 |
|
121 |
+
| 0.1797 | 64.0 | 2128 | 1.8707 | 0.6167 | 0.6151 |
|
122 |
+
| 0.173 | 64.9925 | 2161 | 2.6292 | 0.5167 | 0.5031 |
|
123 |
+
| 0.1815 | 65.9850 | 2194 | 2.6567 | 0.6 | 0.5973 |
|
124 |
+
| 0.0665 | 66.9774 | 2227 | 3.2104 | 0.5167 | 0.5031 |
|
125 |
+
| 0.1084 | 68.0 | 2261 | 3.6692 | 0.5333 | 0.5228 |
|
126 |
+
| 0.1298 | 68.9925 | 2294 | 3.4104 | 0.55 | 0.5373 |
|
127 |
+
| 0.1338 | 69.9850 | 2327 | 2.8215 | 0.6 | 0.5973 |
|
128 |
+
| 0.0795 | 70.9774 | 2360 | 2.9208 | 0.5833 | 0.5830 |
|
129 |
+
| 0.1138 | 72.0 | 2394 | 3.4277 | 0.5333 | 0.5302 |
|
130 |
+
| 0.1644 | 72.9925 | 2427 | 2.8141 | 0.5833 | 0.5830 |
|
131 |
+
| 0.1659 | 73.9850 | 2460 | 2.8723 | 0.6 | 0.6 |
|
132 |
+
| 0.0453 | 74.9774 | 2493 | 2.8769 | 0.6333 | 0.6309 |
|
133 |
+
| 0.0956 | 76.0 | 2527 | 3.2970 | 0.6167 | 0.6098 |
|
134 |
+
| 0.1581 | 76.9925 | 2560 | 3.6672 | 0.5833 | 0.5816 |
|
135 |
+
| 0.157 | 77.9850 | 2593 | 3.5317 | 0.55 | 0.5501 |
|
136 |
+
| 0.0662 | 78.9774 | 2626 | 3.9003 | 0.55 | 0.5456 |
|
137 |
+
| 0.1954 | 80.0 | 2660 | 3.3000 | 0.5833 | 0.5834 |
|
138 |
+
| 0.0527 | 80.9925 | 2693 | 3.9596 | 0.5667 | 0.5638 |
|
139 |
+
| 0.1578 | 81.9850 | 2726 | 3.6724 | 0.55 | 0.5481 |
|
140 |
+
| 0.0737 | 82.9774 | 2759 | 4.0222 | 0.5167 | 0.5119 |
|
141 |
+
| 0.0617 | 84.0 | 2793 | 3.5510 | 0.5833 | 0.5834 |
|
142 |
+
| 0.0531 | 84.9925 | 2826 | 3.5110 | 0.6 | 0.6 |
|
143 |
+
| 0.0993 | 85.9850 | 2859 | 4.0699 | 0.55 | 0.5481 |
|
144 |
+
| 0.1545 | 86.9774 | 2892 | 3.6860 | 0.5667 | 0.5667 |
|
145 |
+
| 0.0554 | 88.0 | 2926 | 3.4409 | 0.6 | 0.6 |
|
146 |
+
| 0.0641 | 88.9925 | 2959 | 3.8304 | 0.55 | 0.5496 |
|
147 |
+
| 0.0633 | 89.9850 | 2992 | 4.0899 | 0.55 | 0.5456 |
|
148 |
+
| 0.0991 | 90.9774 | 3025 | 3.7344 | 0.6 | 0.6 |
|
149 |
+
| 0.0772 | 92.0 | 3059 | 3.8448 | 0.6 | 0.5991 |
|
150 |
+
| 0.0646 | 92.9925 | 3092 | 3.7794 | 0.6 | 0.5991 |
|
151 |
+
| 0.0562 | 93.9850 | 3125 | 3.9340 | 0.5833 | 0.5830 |
|
152 |
+
| 0.0475 | 94.9774 | 3158 | 4.2388 | 0.55 | 0.5481 |
|
153 |
+
| 0.0715 | 96.0 | 3192 | 4.2732 | 0.5333 | 0.5302 |
|
154 |
+
| 0.0875 | 96.9925 | 3225 | 4.1521 | 0.5667 | 0.5657 |
|
155 |
+
| 0.0253 | 97.9850 | 3258 | 4.0813 | 0.5667 | 0.5667 |
|
156 |
+
| 0.1037 | 98.9774 | 3291 | 4.1074 | 0.5667 | 0.5667 |
|
157 |
+
| 0.1094 | 99.2481 | 3300 | 4.1081 | 0.5667 | 0.5667 |
|
158 |
+
|
159 |
+
|
160 |
+
### Framework versions
|
161 |
+
|
162 |
+
- Transformers 4.45.2
|
163 |
+
- Pytorch 2.4.0+cu121
|
164 |
+
- Datasets 3.1.0
|
165 |
+
- Tokenizers 0.20.1
|
logs/events.out.tfevents.1731741098.AZHAR-PC.1610705.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:545b5e8173acb8985290c9d785367ae9abbf894a75cb387bda919ccb378b84bf
|
3 |
+
size 112496
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 784887780
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ef8523d829f6e0dd71f9e136d03cd2e38a03e71b5a86ecee6de882833d55b693
|
3 |
size 784887780
|