File size: 23,980 Bytes
a017383
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
2023-10-18 23:15:16,265 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,266 Model: "SequenceTagger(
  (embeddings): TransformerWordEmbeddings(
    (model): BertModel(
      (embeddings): BertEmbeddings(
        (word_embeddings): Embedding(32001, 128)
        (position_embeddings): Embedding(512, 128)
        (token_type_embeddings): Embedding(2, 128)
        (LayerNorm): LayerNorm((128,), eps=1e-12, elementwise_affine=True)
        (dropout): Dropout(p=0.1, inplace=False)
      )
      (encoder): BertEncoder(
        (layer): ModuleList(
          (0-1): 2 x BertLayer(
            (attention): BertAttention(
              (self): BertSelfAttention(
                (query): Linear(in_features=128, out_features=128, bias=True)
                (key): Linear(in_features=128, out_features=128, bias=True)
                (value): Linear(in_features=128, out_features=128, bias=True)
                (dropout): Dropout(p=0.1, inplace=False)
              )
              (output): BertSelfOutput(
                (dense): Linear(in_features=128, out_features=128, bias=True)
                (LayerNorm): LayerNorm((128,), eps=1e-12, elementwise_affine=True)
                (dropout): Dropout(p=0.1, inplace=False)
              )
            )
            (intermediate): BertIntermediate(
              (dense): Linear(in_features=128, out_features=512, bias=True)
              (intermediate_act_fn): GELUActivation()
            )
            (output): BertOutput(
              (dense): Linear(in_features=512, out_features=128, bias=True)
              (LayerNorm): LayerNorm((128,), eps=1e-12, elementwise_affine=True)
              (dropout): Dropout(p=0.1, inplace=False)
            )
          )
        )
      )
      (pooler): BertPooler(
        (dense): Linear(in_features=128, out_features=128, bias=True)
        (activation): Tanh()
      )
    )
  )
  (locked_dropout): LockedDropout(p=0.5)
  (linear): Linear(in_features=128, out_features=13, bias=True)
  (loss_function): CrossEntropyLoss()
)"
2023-10-18 23:15:16,266 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,266 MultiCorpus: 5777 train + 722 dev + 723 test sentences
 - NER_ICDAR_EUROPEANA Corpus: 5777 train + 722 dev + 723 test sentences - /root/.flair/datasets/ner_icdar_europeana/nl
2023-10-18 23:15:16,266 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,266 Train:  5777 sentences
2023-10-18 23:15:16,266         (train_with_dev=False, train_with_test=False)
2023-10-18 23:15:16,266 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,266 Training Params:
2023-10-18 23:15:16,266  - learning_rate: "5e-05" 
2023-10-18 23:15:16,266  - mini_batch_size: "8"
2023-10-18 23:15:16,266  - max_epochs: "10"
2023-10-18 23:15:16,266  - shuffle: "True"
2023-10-18 23:15:16,266 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,266 Plugins:
2023-10-18 23:15:16,266  - TensorboardLogger
2023-10-18 23:15:16,266  - LinearScheduler | warmup_fraction: '0.1'
2023-10-18 23:15:16,266 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,266 Final evaluation on model from best epoch (best-model.pt)
2023-10-18 23:15:16,266  - metric: "('micro avg', 'f1-score')"
2023-10-18 23:15:16,267 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,267 Computation:
2023-10-18 23:15:16,267  - compute on device: cuda:0
2023-10-18 23:15:16,267  - embedding storage: none
2023-10-18 23:15:16,267 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,267 Model training base path: "hmbench-icdar/nl-dbmdz/bert-tiny-historic-multilingual-cased-bs8-wsFalse-e10-lr5e-05-poolingfirst-layers-1-crfFalse-5"
2023-10-18 23:15:16,267 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,267 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:16,267 Logging anything other than scalars to TensorBoard is currently not supported.
2023-10-18 23:15:18,125 epoch 1 - iter 72/723 - loss 2.40908332 - time (sec): 1.86 - samples/sec: 9509.35 - lr: 0.000005 - momentum: 0.000000
2023-10-18 23:15:19,840 epoch 1 - iter 144/723 - loss 2.18163332 - time (sec): 3.57 - samples/sec: 9437.85 - lr: 0.000010 - momentum: 0.000000
2023-10-18 23:15:21,665 epoch 1 - iter 216/723 - loss 1.79762825 - time (sec): 5.40 - samples/sec: 9717.50 - lr: 0.000015 - momentum: 0.000000
2023-10-18 23:15:23,570 epoch 1 - iter 288/723 - loss 1.47395731 - time (sec): 7.30 - samples/sec: 9671.73 - lr: 0.000020 - momentum: 0.000000
2023-10-18 23:15:25,411 epoch 1 - iter 360/723 - loss 1.24759954 - time (sec): 9.14 - samples/sec: 9756.94 - lr: 0.000025 - momentum: 0.000000
2023-10-18 23:15:27,208 epoch 1 - iter 432/723 - loss 1.11186962 - time (sec): 10.94 - samples/sec: 9610.13 - lr: 0.000030 - momentum: 0.000000
2023-10-18 23:15:28,981 epoch 1 - iter 504/723 - loss 0.99525058 - time (sec): 12.71 - samples/sec: 9687.61 - lr: 0.000035 - momentum: 0.000000
2023-10-18 23:15:30,762 epoch 1 - iter 576/723 - loss 0.89980120 - time (sec): 14.50 - samples/sec: 9757.34 - lr: 0.000040 - momentum: 0.000000
2023-10-18 23:15:32,561 epoch 1 - iter 648/723 - loss 0.83006529 - time (sec): 16.29 - samples/sec: 9756.67 - lr: 0.000045 - momentum: 0.000000
2023-10-18 23:15:34,301 epoch 1 - iter 720/723 - loss 0.77620201 - time (sec): 18.03 - samples/sec: 9734.29 - lr: 0.000050 - momentum: 0.000000
2023-10-18 23:15:34,369 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:34,369 EPOCH 1 done: loss 0.7741 - lr: 0.000050
2023-10-18 23:15:35,657 DEV : loss 0.2758811414241791 - f1-score (micro avg)  0.0082
2023-10-18 23:15:35,671 saving best model
2023-10-18 23:15:35,701 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:37,598 epoch 2 - iter 72/723 - loss 0.25857317 - time (sec): 1.90 - samples/sec: 9119.75 - lr: 0.000049 - momentum: 0.000000
2023-10-18 23:15:39,477 epoch 2 - iter 144/723 - loss 0.23888287 - time (sec): 3.78 - samples/sec: 9439.22 - lr: 0.000049 - momentum: 0.000000
2023-10-18 23:15:41,200 epoch 2 - iter 216/723 - loss 0.23446606 - time (sec): 5.50 - samples/sec: 9413.80 - lr: 0.000048 - momentum: 0.000000
2023-10-18 23:15:42,964 epoch 2 - iter 288/723 - loss 0.22142250 - time (sec): 7.26 - samples/sec: 9657.27 - lr: 0.000048 - momentum: 0.000000
2023-10-18 23:15:44,709 epoch 2 - iter 360/723 - loss 0.21787192 - time (sec): 9.01 - samples/sec: 9648.55 - lr: 0.000047 - momentum: 0.000000
2023-10-18 23:15:46,447 epoch 2 - iter 432/723 - loss 0.21454913 - time (sec): 10.75 - samples/sec: 9633.59 - lr: 0.000047 - momentum: 0.000000
2023-10-18 23:15:48,218 epoch 2 - iter 504/723 - loss 0.21228027 - time (sec): 12.52 - samples/sec: 9686.73 - lr: 0.000046 - momentum: 0.000000
2023-10-18 23:15:50,039 epoch 2 - iter 576/723 - loss 0.21139498 - time (sec): 14.34 - samples/sec: 9809.15 - lr: 0.000046 - momentum: 0.000000
2023-10-18 23:15:51,821 epoch 2 - iter 648/723 - loss 0.20851881 - time (sec): 16.12 - samples/sec: 9824.09 - lr: 0.000045 - momentum: 0.000000
2023-10-18 23:15:53,645 epoch 2 - iter 720/723 - loss 0.20557912 - time (sec): 17.94 - samples/sec: 9786.99 - lr: 0.000044 - momentum: 0.000000
2023-10-18 23:15:53,709 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:53,710 EPOCH 2 done: loss 0.2053 - lr: 0.000044
2023-10-18 23:15:55,808 DEV : loss 0.2405446618795395 - f1-score (micro avg)  0.1862
2023-10-18 23:15:55,822 saving best model
2023-10-18 23:15:55,858 ----------------------------------------------------------------------------------------------------
2023-10-18 23:15:57,614 epoch 3 - iter 72/723 - loss 0.17701188 - time (sec): 1.76 - samples/sec: 10573.26 - lr: 0.000044 - momentum: 0.000000
2023-10-18 23:15:59,375 epoch 3 - iter 144/723 - loss 0.18046644 - time (sec): 3.52 - samples/sec: 10214.55 - lr: 0.000043 - momentum: 0.000000
2023-10-18 23:16:01,273 epoch 3 - iter 216/723 - loss 0.17924447 - time (sec): 5.41 - samples/sec: 10016.01 - lr: 0.000043 - momentum: 0.000000
2023-10-18 23:16:03,056 epoch 3 - iter 288/723 - loss 0.17534354 - time (sec): 7.20 - samples/sec: 9979.60 - lr: 0.000042 - momentum: 0.000000
2023-10-18 23:16:04,795 epoch 3 - iter 360/723 - loss 0.17549661 - time (sec): 8.94 - samples/sec: 9890.73 - lr: 0.000042 - momentum: 0.000000
2023-10-18 23:16:06,577 epoch 3 - iter 432/723 - loss 0.17806596 - time (sec): 10.72 - samples/sec: 9864.63 - lr: 0.000041 - momentum: 0.000000
2023-10-18 23:16:08,395 epoch 3 - iter 504/723 - loss 0.17649226 - time (sec): 12.54 - samples/sec: 9759.78 - lr: 0.000041 - momentum: 0.000000
2023-10-18 23:16:10,276 epoch 3 - iter 576/723 - loss 0.17718147 - time (sec): 14.42 - samples/sec: 9766.82 - lr: 0.000040 - momentum: 0.000000
2023-10-18 23:16:12,017 epoch 3 - iter 648/723 - loss 0.17435274 - time (sec): 16.16 - samples/sec: 9788.62 - lr: 0.000039 - momentum: 0.000000
2023-10-18 23:16:13,753 epoch 3 - iter 720/723 - loss 0.17453446 - time (sec): 17.89 - samples/sec: 9819.40 - lr: 0.000039 - momentum: 0.000000
2023-10-18 23:16:13,814 ----------------------------------------------------------------------------------------------------
2023-10-18 23:16:13,814 EPOCH 3 done: loss 0.1744 - lr: 0.000039
2023-10-18 23:16:15,580 DEV : loss 0.19726891815662384 - f1-score (micro avg)  0.4418
2023-10-18 23:16:15,595 saving best model
2023-10-18 23:16:15,630 ----------------------------------------------------------------------------------------------------
2023-10-18 23:16:17,465 epoch 4 - iter 72/723 - loss 0.19970285 - time (sec): 1.83 - samples/sec: 9313.18 - lr: 0.000038 - momentum: 0.000000
2023-10-18 23:16:19,232 epoch 4 - iter 144/723 - loss 0.16333378 - time (sec): 3.60 - samples/sec: 9634.99 - lr: 0.000038 - momentum: 0.000000
2023-10-18 23:16:21,027 epoch 4 - iter 216/723 - loss 0.16709855 - time (sec): 5.40 - samples/sec: 9499.46 - lr: 0.000037 - momentum: 0.000000
2023-10-18 23:16:22,919 epoch 4 - iter 288/723 - loss 0.16278167 - time (sec): 7.29 - samples/sec: 9586.69 - lr: 0.000037 - momentum: 0.000000
2023-10-18 23:16:24,700 epoch 4 - iter 360/723 - loss 0.16462834 - time (sec): 9.07 - samples/sec: 9779.30 - lr: 0.000036 - momentum: 0.000000
2023-10-18 23:16:26,498 epoch 4 - iter 432/723 - loss 0.16054808 - time (sec): 10.87 - samples/sec: 9870.03 - lr: 0.000036 - momentum: 0.000000
2023-10-18 23:16:28,289 epoch 4 - iter 504/723 - loss 0.15875214 - time (sec): 12.66 - samples/sec: 9784.37 - lr: 0.000035 - momentum: 0.000000
2023-10-18 23:16:30,002 epoch 4 - iter 576/723 - loss 0.15882572 - time (sec): 14.37 - samples/sec: 9832.90 - lr: 0.000034 - momentum: 0.000000
2023-10-18 23:16:32,220 epoch 4 - iter 648/723 - loss 0.15908778 - time (sec): 16.59 - samples/sec: 9607.35 - lr: 0.000034 - momentum: 0.000000
2023-10-18 23:16:33,961 epoch 4 - iter 720/723 - loss 0.15739102 - time (sec): 18.33 - samples/sec: 9590.24 - lr: 0.000033 - momentum: 0.000000
2023-10-18 23:16:34,026 ----------------------------------------------------------------------------------------------------
2023-10-18 23:16:34,026 EPOCH 4 done: loss 0.1573 - lr: 0.000033
2023-10-18 23:16:35,788 DEV : loss 0.18398496508598328 - f1-score (micro avg)  0.4749
2023-10-18 23:16:35,803 saving best model
2023-10-18 23:16:35,838 ----------------------------------------------------------------------------------------------------
2023-10-18 23:16:37,539 epoch 5 - iter 72/723 - loss 0.16602905 - time (sec): 1.70 - samples/sec: 9888.68 - lr: 0.000033 - momentum: 0.000000
2023-10-18 23:16:39,324 epoch 5 - iter 144/723 - loss 0.15855908 - time (sec): 3.49 - samples/sec: 9666.77 - lr: 0.000032 - momentum: 0.000000
2023-10-18 23:16:41,069 epoch 5 - iter 216/723 - loss 0.15565325 - time (sec): 5.23 - samples/sec: 9589.97 - lr: 0.000032 - momentum: 0.000000
2023-10-18 23:16:42,855 epoch 5 - iter 288/723 - loss 0.15137039 - time (sec): 7.02 - samples/sec: 9638.72 - lr: 0.000031 - momentum: 0.000000
2023-10-18 23:16:44,673 epoch 5 - iter 360/723 - loss 0.15306731 - time (sec): 8.84 - samples/sec: 9834.34 - lr: 0.000031 - momentum: 0.000000
2023-10-18 23:16:46,518 epoch 5 - iter 432/723 - loss 0.14937650 - time (sec): 10.68 - samples/sec: 9838.86 - lr: 0.000030 - momentum: 0.000000
2023-10-18 23:16:48,257 epoch 5 - iter 504/723 - loss 0.14562850 - time (sec): 12.42 - samples/sec: 9845.97 - lr: 0.000029 - momentum: 0.000000
2023-10-18 23:16:49,944 epoch 5 - iter 576/723 - loss 0.14633530 - time (sec): 14.11 - samples/sec: 9882.96 - lr: 0.000029 - momentum: 0.000000
2023-10-18 23:16:51,751 epoch 5 - iter 648/723 - loss 0.14812330 - time (sec): 15.91 - samples/sec: 9847.09 - lr: 0.000028 - momentum: 0.000000
2023-10-18 23:16:53,692 epoch 5 - iter 720/723 - loss 0.14451958 - time (sec): 17.85 - samples/sec: 9841.50 - lr: 0.000028 - momentum: 0.000000
2023-10-18 23:16:53,748 ----------------------------------------------------------------------------------------------------
2023-10-18 23:16:53,748 EPOCH 5 done: loss 0.1448 - lr: 0.000028
2023-10-18 23:16:55,530 DEV : loss 0.1783648133277893 - f1-score (micro avg)  0.4941
2023-10-18 23:16:55,544 saving best model
2023-10-18 23:16:55,580 ----------------------------------------------------------------------------------------------------
2023-10-18 23:16:57,339 epoch 6 - iter 72/723 - loss 0.12412838 - time (sec): 1.76 - samples/sec: 10153.63 - lr: 0.000027 - momentum: 0.000000
2023-10-18 23:16:59,153 epoch 6 - iter 144/723 - loss 0.12845739 - time (sec): 3.57 - samples/sec: 10111.64 - lr: 0.000027 - momentum: 0.000000
2023-10-18 23:17:01,000 epoch 6 - iter 216/723 - loss 0.13059131 - time (sec): 5.42 - samples/sec: 10079.33 - lr: 0.000026 - momentum: 0.000000
2023-10-18 23:17:02,729 epoch 6 - iter 288/723 - loss 0.13011154 - time (sec): 7.15 - samples/sec: 10062.07 - lr: 0.000026 - momentum: 0.000000
2023-10-18 23:17:04,837 epoch 6 - iter 360/723 - loss 0.13108817 - time (sec): 9.26 - samples/sec: 9646.45 - lr: 0.000025 - momentum: 0.000000
2023-10-18 23:17:06,702 epoch 6 - iter 432/723 - loss 0.13199062 - time (sec): 11.12 - samples/sec: 9684.01 - lr: 0.000024 - momentum: 0.000000
2023-10-18 23:17:08,534 epoch 6 - iter 504/723 - loss 0.13635868 - time (sec): 12.95 - samples/sec: 9661.56 - lr: 0.000024 - momentum: 0.000000
2023-10-18 23:17:10,365 epoch 6 - iter 576/723 - loss 0.13682865 - time (sec): 14.78 - samples/sec: 9566.59 - lr: 0.000023 - momentum: 0.000000
2023-10-18 23:17:12,189 epoch 6 - iter 648/723 - loss 0.13499612 - time (sec): 16.61 - samples/sec: 9533.05 - lr: 0.000023 - momentum: 0.000000
2023-10-18 23:17:13,891 epoch 6 - iter 720/723 - loss 0.13337880 - time (sec): 18.31 - samples/sec: 9592.34 - lr: 0.000022 - momentum: 0.000000
2023-10-18 23:17:13,955 ----------------------------------------------------------------------------------------------------
2023-10-18 23:17:13,955 EPOCH 6 done: loss 0.1334 - lr: 0.000022
2023-10-18 23:17:15,731 DEV : loss 0.18174949288368225 - f1-score (micro avg)  0.4952
2023-10-18 23:17:15,745 saving best model
2023-10-18 23:17:15,781 ----------------------------------------------------------------------------------------------------
2023-10-18 23:17:17,561 epoch 7 - iter 72/723 - loss 0.11870149 - time (sec): 1.78 - samples/sec: 9897.22 - lr: 0.000022 - momentum: 0.000000
2023-10-18 23:17:19,339 epoch 7 - iter 144/723 - loss 0.12495624 - time (sec): 3.56 - samples/sec: 9460.86 - lr: 0.000021 - momentum: 0.000000
2023-10-18 23:17:21,174 epoch 7 - iter 216/723 - loss 0.12415400 - time (sec): 5.39 - samples/sec: 9681.31 - lr: 0.000021 - momentum: 0.000000
2023-10-18 23:17:23,015 epoch 7 - iter 288/723 - loss 0.12335361 - time (sec): 7.23 - samples/sec: 9529.01 - lr: 0.000020 - momentum: 0.000000
2023-10-18 23:17:24,834 epoch 7 - iter 360/723 - loss 0.12487733 - time (sec): 9.05 - samples/sec: 9600.12 - lr: 0.000019 - momentum: 0.000000
2023-10-18 23:17:26,642 epoch 7 - iter 432/723 - loss 0.12677615 - time (sec): 10.86 - samples/sec: 9665.79 - lr: 0.000019 - momentum: 0.000000
2023-10-18 23:17:28,481 epoch 7 - iter 504/723 - loss 0.12846292 - time (sec): 12.70 - samples/sec: 9621.48 - lr: 0.000018 - momentum: 0.000000
2023-10-18 23:17:30,414 epoch 7 - iter 576/723 - loss 0.12813188 - time (sec): 14.63 - samples/sec: 9660.86 - lr: 0.000018 - momentum: 0.000000
2023-10-18 23:17:32,221 epoch 7 - iter 648/723 - loss 0.12928188 - time (sec): 16.44 - samples/sec: 9623.10 - lr: 0.000017 - momentum: 0.000000
2023-10-18 23:17:33,963 epoch 7 - iter 720/723 - loss 0.12811456 - time (sec): 18.18 - samples/sec: 9668.44 - lr: 0.000017 - momentum: 0.000000
2023-10-18 23:17:34,024 ----------------------------------------------------------------------------------------------------
2023-10-18 23:17:34,024 EPOCH 7 done: loss 0.1281 - lr: 0.000017
2023-10-18 23:17:35,808 DEV : loss 0.1717141568660736 - f1-score (micro avg)  0.5346
2023-10-18 23:17:35,823 saving best model
2023-10-18 23:17:35,858 ----------------------------------------------------------------------------------------------------
2023-10-18 23:17:38,073 epoch 8 - iter 72/723 - loss 0.12299219 - time (sec): 2.21 - samples/sec: 8057.14 - lr: 0.000016 - momentum: 0.000000
2023-10-18 23:17:39,830 epoch 8 - iter 144/723 - loss 0.11754273 - time (sec): 3.97 - samples/sec: 8740.64 - lr: 0.000016 - momentum: 0.000000
2023-10-18 23:17:41,690 epoch 8 - iter 216/723 - loss 0.11975555 - time (sec): 5.83 - samples/sec: 9103.15 - lr: 0.000015 - momentum: 0.000000
2023-10-18 23:17:43,614 epoch 8 - iter 288/723 - loss 0.11991985 - time (sec): 7.75 - samples/sec: 9207.08 - lr: 0.000014 - momentum: 0.000000
2023-10-18 23:17:45,415 epoch 8 - iter 360/723 - loss 0.12012028 - time (sec): 9.56 - samples/sec: 9293.87 - lr: 0.000014 - momentum: 0.000000
2023-10-18 23:17:47,181 epoch 8 - iter 432/723 - loss 0.12323901 - time (sec): 11.32 - samples/sec: 9411.71 - lr: 0.000013 - momentum: 0.000000
2023-10-18 23:17:48,997 epoch 8 - iter 504/723 - loss 0.12390650 - time (sec): 13.14 - samples/sec: 9515.34 - lr: 0.000013 - momentum: 0.000000
2023-10-18 23:17:50,731 epoch 8 - iter 576/723 - loss 0.12261161 - time (sec): 14.87 - samples/sec: 9531.16 - lr: 0.000012 - momentum: 0.000000
2023-10-18 23:17:52,540 epoch 8 - iter 648/723 - loss 0.12309668 - time (sec): 16.68 - samples/sec: 9526.96 - lr: 0.000012 - momentum: 0.000000
2023-10-18 23:17:54,314 epoch 8 - iter 720/723 - loss 0.12212658 - time (sec): 18.46 - samples/sec: 9518.50 - lr: 0.000011 - momentum: 0.000000
2023-10-18 23:17:54,372 ----------------------------------------------------------------------------------------------------
2023-10-18 23:17:54,372 EPOCH 8 done: loss 0.1221 - lr: 0.000011
2023-10-18 23:17:56,154 DEV : loss 0.17939499020576477 - f1-score (micro avg)  0.5277
2023-10-18 23:17:56,168 ----------------------------------------------------------------------------------------------------
2023-10-18 23:17:57,932 epoch 9 - iter 72/723 - loss 0.10482754 - time (sec): 1.76 - samples/sec: 10587.13 - lr: 0.000011 - momentum: 0.000000
2023-10-18 23:17:59,777 epoch 9 - iter 144/723 - loss 0.11941581 - time (sec): 3.61 - samples/sec: 10252.77 - lr: 0.000010 - momentum: 0.000000
2023-10-18 23:18:01,519 epoch 9 - iter 216/723 - loss 0.11428472 - time (sec): 5.35 - samples/sec: 10270.88 - lr: 0.000009 - momentum: 0.000000
2023-10-18 23:18:03,294 epoch 9 - iter 288/723 - loss 0.11765545 - time (sec): 7.13 - samples/sec: 10286.21 - lr: 0.000009 - momentum: 0.000000
2023-10-18 23:18:05,075 epoch 9 - iter 360/723 - loss 0.12157267 - time (sec): 8.91 - samples/sec: 10049.91 - lr: 0.000008 - momentum: 0.000000
2023-10-18 23:18:06,761 epoch 9 - iter 432/723 - loss 0.11936636 - time (sec): 10.59 - samples/sec: 10058.03 - lr: 0.000008 - momentum: 0.000000
2023-10-18 23:18:08,276 epoch 9 - iter 504/723 - loss 0.11873180 - time (sec): 12.11 - samples/sec: 10211.95 - lr: 0.000007 - momentum: 0.000000
2023-10-18 23:18:09,884 epoch 9 - iter 576/723 - loss 0.11689872 - time (sec): 13.71 - samples/sec: 10369.22 - lr: 0.000007 - momentum: 0.000000
2023-10-18 23:18:11,685 epoch 9 - iter 648/723 - loss 0.11799321 - time (sec): 15.52 - samples/sec: 10271.81 - lr: 0.000006 - momentum: 0.000000
2023-10-18 23:18:13,436 epoch 9 - iter 720/723 - loss 0.11948734 - time (sec): 17.27 - samples/sec: 10181.80 - lr: 0.000006 - momentum: 0.000000
2023-10-18 23:18:13,495 ----------------------------------------------------------------------------------------------------
2023-10-18 23:18:13,495 EPOCH 9 done: loss 0.1197 - lr: 0.000006
2023-10-18 23:18:15,639 DEV : loss 0.17294184863567352 - f1-score (micro avg)  0.5292
2023-10-18 23:18:15,655 ----------------------------------------------------------------------------------------------------
2023-10-18 23:18:17,542 epoch 10 - iter 72/723 - loss 0.12361351 - time (sec): 1.89 - samples/sec: 9438.66 - lr: 0.000005 - momentum: 0.000000
2023-10-18 23:18:19,410 epoch 10 - iter 144/723 - loss 0.11590017 - time (sec): 3.76 - samples/sec: 9446.32 - lr: 0.000004 - momentum: 0.000000
2023-10-18 23:18:21,248 epoch 10 - iter 216/723 - loss 0.11549287 - time (sec): 5.59 - samples/sec: 9439.46 - lr: 0.000004 - momentum: 0.000000
2023-10-18 23:18:23,070 epoch 10 - iter 288/723 - loss 0.11632240 - time (sec): 7.41 - samples/sec: 9349.45 - lr: 0.000003 - momentum: 0.000000
2023-10-18 23:18:24,972 epoch 10 - iter 360/723 - loss 0.12476098 - time (sec): 9.32 - samples/sec: 9447.80 - lr: 0.000003 - momentum: 0.000000
2023-10-18 23:18:26,863 epoch 10 - iter 432/723 - loss 0.12139655 - time (sec): 11.21 - samples/sec: 9515.38 - lr: 0.000002 - momentum: 0.000000
2023-10-18 23:18:28,705 epoch 10 - iter 504/723 - loss 0.12203358 - time (sec): 13.05 - samples/sec: 9588.11 - lr: 0.000002 - momentum: 0.000000
2023-10-18 23:18:30,467 epoch 10 - iter 576/723 - loss 0.12185105 - time (sec): 14.81 - samples/sec: 9508.37 - lr: 0.000001 - momentum: 0.000000
2023-10-18 23:18:32,221 epoch 10 - iter 648/723 - loss 0.12023157 - time (sec): 16.57 - samples/sec: 9540.08 - lr: 0.000001 - momentum: 0.000000
2023-10-18 23:18:34,016 epoch 10 - iter 720/723 - loss 0.11895430 - time (sec): 18.36 - samples/sec: 9573.31 - lr: 0.000000 - momentum: 0.000000
2023-10-18 23:18:34,073 ----------------------------------------------------------------------------------------------------
2023-10-18 23:18:34,073 EPOCH 10 done: loss 0.1190 - lr: 0.000000
2023-10-18 23:18:35,874 DEV : loss 0.1752191036939621 - f1-score (micro avg)  0.5318
2023-10-18 23:18:35,921 ----------------------------------------------------------------------------------------------------
2023-10-18 23:18:35,921 Loading model from best epoch ...
2023-10-18 23:18:36,002 SequenceTagger predicts: Dictionary with 13 tags: O, S-LOC, B-LOC, E-LOC, I-LOC, S-PER, B-PER, E-PER, I-PER, S-ORG, B-ORG, E-ORG, I-ORG
2023-10-18 23:18:37,353 
Results:
- F-score (micro) 0.5439
- F-score (macro) 0.3773
- Accuracy 0.3802

By class:
              precision    recall  f1-score   support

         PER     0.5046    0.4564    0.4793       482
         LOC     0.6705    0.6354    0.6525       458
         ORG     0.0000    0.0000    0.0000        69

   micro avg     0.5874    0.5064    0.5439      1009
   macro avg     0.3917    0.3639    0.3773      1009
weighted avg     0.5454    0.5064    0.5251      1009

2023-10-18 23:18:37,353 ----------------------------------------------------------------------------------------------------