flair-icdar-fr / training.log
stefan-it's picture
Upload folder using huggingface_hub
eb0ba09
2023-10-18 21:25:18,822 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,823 Model: "SequenceTagger(
(embeddings): TransformerWordEmbeddings(
(model): BertModel(
(embeddings): BertEmbeddings(
(word_embeddings): Embedding(32001, 128)
(position_embeddings): Embedding(512, 128)
(token_type_embeddings): Embedding(2, 128)
(LayerNorm): LayerNorm((128,), eps=1e-12, elementwise_affine=True)
(dropout): Dropout(p=0.1, inplace=False)
)
(encoder): BertEncoder(
(layer): ModuleList(
(0-1): 2 x BertLayer(
(attention): BertAttention(
(self): BertSelfAttention(
(query): Linear(in_features=128, out_features=128, bias=True)
(key): Linear(in_features=128, out_features=128, bias=True)
(value): Linear(in_features=128, out_features=128, bias=True)
(dropout): Dropout(p=0.1, inplace=False)
)
(output): BertSelfOutput(
(dense): Linear(in_features=128, out_features=128, bias=True)
(LayerNorm): LayerNorm((128,), eps=1e-12, elementwise_affine=True)
(dropout): Dropout(p=0.1, inplace=False)
)
)
(intermediate): BertIntermediate(
(dense): Linear(in_features=128, out_features=512, bias=True)
(intermediate_act_fn): GELUActivation()
)
(output): BertOutput(
(dense): Linear(in_features=512, out_features=128, bias=True)
(LayerNorm): LayerNorm((128,), eps=1e-12, elementwise_affine=True)
(dropout): Dropout(p=0.1, inplace=False)
)
)
)
)
(pooler): BertPooler(
(dense): Linear(in_features=128, out_features=128, bias=True)
(activation): Tanh()
)
)
)
(locked_dropout): LockedDropout(p=0.5)
(linear): Linear(in_features=128, out_features=13, bias=True)
(loss_function): CrossEntropyLoss()
)"
2023-10-18 21:25:18,823 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,823 MultiCorpus: 7936 train + 992 dev + 992 test sentences
- NER_ICDAR_EUROPEANA Corpus: 7936 train + 992 dev + 992 test sentences - /root/.flair/datasets/ner_icdar_europeana/fr
2023-10-18 21:25:18,823 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,823 Train: 7936 sentences
2023-10-18 21:25:18,823 (train_with_dev=False, train_with_test=False)
2023-10-18 21:25:18,823 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,823 Training Params:
2023-10-18 21:25:18,823 - learning_rate: "5e-05"
2023-10-18 21:25:18,823 - mini_batch_size: "4"
2023-10-18 21:25:18,823 - max_epochs: "10"
2023-10-18 21:25:18,823 - shuffle: "True"
2023-10-18 21:25:18,823 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,823 Plugins:
2023-10-18 21:25:18,823 - TensorboardLogger
2023-10-18 21:25:18,823 - LinearScheduler | warmup_fraction: '0.1'
2023-10-18 21:25:18,823 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,823 Final evaluation on model from best epoch (best-model.pt)
2023-10-18 21:25:18,823 - metric: "('micro avg', 'f1-score')"
2023-10-18 21:25:18,823 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,823 Computation:
2023-10-18 21:25:18,824 - compute on device: cuda:0
2023-10-18 21:25:18,824 - embedding storage: none
2023-10-18 21:25:18,824 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,824 Model training base path: "hmbench-icdar/fr-dbmdz/bert-tiny-historic-multilingual-cased-bs4-wsFalse-e10-lr5e-05-poolingfirst-layers-1-crfFalse-4"
2023-10-18 21:25:18,824 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,824 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:18,824 Logging anything other than scalars to TensorBoard is currently not supported.
2023-10-18 21:25:21,871 epoch 1 - iter 198/1984 - loss 3.05001127 - time (sec): 3.05 - samples/sec: 5386.70 - lr: 0.000005 - momentum: 0.000000
2023-10-18 21:25:24,628 epoch 1 - iter 396/1984 - loss 2.47724953 - time (sec): 5.80 - samples/sec: 5922.14 - lr: 0.000010 - momentum: 0.000000
2023-10-18 21:25:27,667 epoch 1 - iter 594/1984 - loss 1.93619435 - time (sec): 8.84 - samples/sec: 5709.43 - lr: 0.000015 - momentum: 0.000000
2023-10-18 21:25:30,706 epoch 1 - iter 792/1984 - loss 1.58401951 - time (sec): 11.88 - samples/sec: 5618.92 - lr: 0.000020 - momentum: 0.000000
2023-10-18 21:25:33,758 epoch 1 - iter 990/1984 - loss 1.35401175 - time (sec): 14.93 - samples/sec: 5563.79 - lr: 0.000025 - momentum: 0.000000
2023-10-18 21:25:36,794 epoch 1 - iter 1188/1984 - loss 1.20410373 - time (sec): 17.97 - samples/sec: 5524.66 - lr: 0.000030 - momentum: 0.000000
2023-10-18 21:25:39,832 epoch 1 - iter 1386/1984 - loss 1.09303606 - time (sec): 21.01 - samples/sec: 5470.81 - lr: 0.000035 - momentum: 0.000000
2023-10-18 21:25:42,910 epoch 1 - iter 1584/1984 - loss 0.99807468 - time (sec): 24.09 - samples/sec: 5451.75 - lr: 0.000040 - momentum: 0.000000
2023-10-18 21:25:45,923 epoch 1 - iter 1782/1984 - loss 0.92290600 - time (sec): 27.10 - samples/sec: 5442.29 - lr: 0.000045 - momentum: 0.000000
2023-10-18 21:25:49,008 epoch 1 - iter 1980/1984 - loss 0.86043295 - time (sec): 30.18 - samples/sec: 5425.81 - lr: 0.000050 - momentum: 0.000000
2023-10-18 21:25:49,068 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:49,068 EPOCH 1 done: loss 0.8594 - lr: 0.000050
2023-10-18 21:25:50,560 DEV : loss 0.21192067861557007 - f1-score (micro avg) 0.2945
2023-10-18 21:25:50,578 saving best model
2023-10-18 21:25:50,613 ----------------------------------------------------------------------------------------------------
2023-10-18 21:25:53,727 epoch 2 - iter 198/1984 - loss 0.30950366 - time (sec): 3.11 - samples/sec: 4927.98 - lr: 0.000049 - momentum: 0.000000
2023-10-18 21:25:56,796 epoch 2 - iter 396/1984 - loss 0.28151013 - time (sec): 6.18 - samples/sec: 5145.74 - lr: 0.000049 - momentum: 0.000000
2023-10-18 21:25:59,865 epoch 2 - iter 594/1984 - loss 0.28431083 - time (sec): 9.25 - samples/sec: 5342.97 - lr: 0.000048 - momentum: 0.000000
2023-10-18 21:26:02,940 epoch 2 - iter 792/1984 - loss 0.27867074 - time (sec): 12.33 - samples/sec: 5398.12 - lr: 0.000048 - momentum: 0.000000
2023-10-18 21:26:06,023 epoch 2 - iter 990/1984 - loss 0.27356192 - time (sec): 15.41 - samples/sec: 5362.82 - lr: 0.000047 - momentum: 0.000000
2023-10-18 21:26:09,040 epoch 2 - iter 1188/1984 - loss 0.26808423 - time (sec): 18.43 - samples/sec: 5380.47 - lr: 0.000047 - momentum: 0.000000
2023-10-18 21:26:11,822 epoch 2 - iter 1386/1984 - loss 0.26113663 - time (sec): 21.21 - samples/sec: 5451.69 - lr: 0.000046 - momentum: 0.000000
2023-10-18 21:26:14,900 epoch 2 - iter 1584/1984 - loss 0.25788671 - time (sec): 24.29 - samples/sec: 5438.15 - lr: 0.000046 - momentum: 0.000000
2023-10-18 21:26:17,947 epoch 2 - iter 1782/1984 - loss 0.25525919 - time (sec): 27.33 - samples/sec: 5421.86 - lr: 0.000045 - momentum: 0.000000
2023-10-18 21:26:20,762 epoch 2 - iter 1980/1984 - loss 0.25444187 - time (sec): 30.15 - samples/sec: 5429.33 - lr: 0.000044 - momentum: 0.000000
2023-10-18 21:26:20,822 ----------------------------------------------------------------------------------------------------
2023-10-18 21:26:20,822 EPOCH 2 done: loss 0.2546 - lr: 0.000044
2023-10-18 21:26:22,634 DEV : loss 0.16650356352329254 - f1-score (micro avg) 0.4051
2023-10-18 21:26:22,654 saving best model
2023-10-18 21:26:22,690 ----------------------------------------------------------------------------------------------------
2023-10-18 21:26:25,787 epoch 3 - iter 198/1984 - loss 0.21161120 - time (sec): 3.10 - samples/sec: 5483.96 - lr: 0.000044 - momentum: 0.000000
2023-10-18 21:26:28,837 epoch 3 - iter 396/1984 - loss 0.20257772 - time (sec): 6.15 - samples/sec: 5328.22 - lr: 0.000043 - momentum: 0.000000
2023-10-18 21:26:31,767 epoch 3 - iter 594/1984 - loss 0.21363359 - time (sec): 9.08 - samples/sec: 5582.96 - lr: 0.000043 - momentum: 0.000000
2023-10-18 21:26:34,842 epoch 3 - iter 792/1984 - loss 0.20789207 - time (sec): 12.15 - samples/sec: 5470.71 - lr: 0.000042 - momentum: 0.000000
2023-10-18 21:26:37,879 epoch 3 - iter 990/1984 - loss 0.21055604 - time (sec): 15.19 - samples/sec: 5444.07 - lr: 0.000042 - momentum: 0.000000
2023-10-18 21:26:40,954 epoch 3 - iter 1188/1984 - loss 0.20723027 - time (sec): 18.26 - samples/sec: 5444.07 - lr: 0.000041 - momentum: 0.000000
2023-10-18 21:26:44,031 epoch 3 - iter 1386/1984 - loss 0.20657276 - time (sec): 21.34 - samples/sec: 5424.20 - lr: 0.000041 - momentum: 0.000000
2023-10-18 21:26:47,055 epoch 3 - iter 1584/1984 - loss 0.20547288 - time (sec): 24.36 - samples/sec: 5434.88 - lr: 0.000040 - momentum: 0.000000
2023-10-18 21:26:50,070 epoch 3 - iter 1782/1984 - loss 0.20632094 - time (sec): 27.38 - samples/sec: 5413.63 - lr: 0.000039 - momentum: 0.000000
2023-10-18 21:26:53,106 epoch 3 - iter 1980/1984 - loss 0.20495296 - time (sec): 30.41 - samples/sec: 5376.32 - lr: 0.000039 - momentum: 0.000000
2023-10-18 21:26:53,176 ----------------------------------------------------------------------------------------------------
2023-10-18 21:26:53,176 EPOCH 3 done: loss 0.2046 - lr: 0.000039
2023-10-18 21:26:55,394 DEV : loss 0.15022146701812744 - f1-score (micro avg) 0.5321
2023-10-18 21:26:55,413 saving best model
2023-10-18 21:26:55,447 ----------------------------------------------------------------------------------------------------
2023-10-18 21:26:58,566 epoch 4 - iter 198/1984 - loss 0.18075544 - time (sec): 3.12 - samples/sec: 5291.79 - lr: 0.000038 - momentum: 0.000000
2023-10-18 21:27:01,634 epoch 4 - iter 396/1984 - loss 0.18240152 - time (sec): 6.19 - samples/sec: 5143.93 - lr: 0.000038 - momentum: 0.000000
2023-10-18 21:27:04,650 epoch 4 - iter 594/1984 - loss 0.17441944 - time (sec): 9.20 - samples/sec: 5216.19 - lr: 0.000037 - momentum: 0.000000
2023-10-18 21:27:07,674 epoch 4 - iter 792/1984 - loss 0.17622746 - time (sec): 12.23 - samples/sec: 5214.64 - lr: 0.000037 - momentum: 0.000000
2023-10-18 21:27:10,762 epoch 4 - iter 990/1984 - loss 0.17870370 - time (sec): 15.31 - samples/sec: 5266.42 - lr: 0.000036 - momentum: 0.000000
2023-10-18 21:27:13,782 epoch 4 - iter 1188/1984 - loss 0.17730408 - time (sec): 18.33 - samples/sec: 5322.13 - lr: 0.000036 - momentum: 0.000000
2023-10-18 21:27:16,890 epoch 4 - iter 1386/1984 - loss 0.17390393 - time (sec): 21.44 - samples/sec: 5403.97 - lr: 0.000035 - momentum: 0.000000
2023-10-18 21:27:20,088 epoch 4 - iter 1584/1984 - loss 0.17431675 - time (sec): 24.64 - samples/sec: 5366.83 - lr: 0.000034 - momentum: 0.000000
2023-10-18 21:27:23,148 epoch 4 - iter 1782/1984 - loss 0.17801776 - time (sec): 27.70 - samples/sec: 5333.04 - lr: 0.000034 - momentum: 0.000000
2023-10-18 21:27:26,231 epoch 4 - iter 1980/1984 - loss 0.17957717 - time (sec): 30.78 - samples/sec: 5317.52 - lr: 0.000033 - momentum: 0.000000
2023-10-18 21:27:26,291 ----------------------------------------------------------------------------------------------------
2023-10-18 21:27:26,291 EPOCH 4 done: loss 0.1797 - lr: 0.000033
2023-10-18 21:27:28,115 DEV : loss 0.14466138184070587 - f1-score (micro avg) 0.5683
2023-10-18 21:27:28,133 saving best model
2023-10-18 21:27:28,167 ----------------------------------------------------------------------------------------------------
2023-10-18 21:27:31,281 epoch 5 - iter 198/1984 - loss 0.18093443 - time (sec): 3.11 - samples/sec: 5540.53 - lr: 0.000033 - momentum: 0.000000
2023-10-18 21:27:34,301 epoch 5 - iter 396/1984 - loss 0.17976008 - time (sec): 6.13 - samples/sec: 5491.12 - lr: 0.000032 - momentum: 0.000000
2023-10-18 21:27:37,345 epoch 5 - iter 594/1984 - loss 0.17027389 - time (sec): 9.18 - samples/sec: 5444.29 - lr: 0.000032 - momentum: 0.000000
2023-10-18 21:27:40,347 epoch 5 - iter 792/1984 - loss 0.16509799 - time (sec): 12.18 - samples/sec: 5422.24 - lr: 0.000031 - momentum: 0.000000
2023-10-18 21:27:43,351 epoch 5 - iter 990/1984 - loss 0.16344421 - time (sec): 15.18 - samples/sec: 5364.28 - lr: 0.000031 - momentum: 0.000000
2023-10-18 21:27:46,391 epoch 5 - iter 1188/1984 - loss 0.16463321 - time (sec): 18.22 - samples/sec: 5341.67 - lr: 0.000030 - momentum: 0.000000
2023-10-18 21:27:49,456 epoch 5 - iter 1386/1984 - loss 0.16486788 - time (sec): 21.29 - samples/sec: 5405.39 - lr: 0.000029 - momentum: 0.000000
2023-10-18 21:27:52,493 epoch 5 - iter 1584/1984 - loss 0.16396812 - time (sec): 24.33 - samples/sec: 5388.54 - lr: 0.000029 - momentum: 0.000000
2023-10-18 21:27:55,556 epoch 5 - iter 1782/1984 - loss 0.16439615 - time (sec): 27.39 - samples/sec: 5372.41 - lr: 0.000028 - momentum: 0.000000
2023-10-18 21:27:58,492 epoch 5 - iter 1980/1984 - loss 0.16305676 - time (sec): 30.32 - samples/sec: 5399.14 - lr: 0.000028 - momentum: 0.000000
2023-10-18 21:27:58,549 ----------------------------------------------------------------------------------------------------
2023-10-18 21:27:58,549 EPOCH 5 done: loss 0.1629 - lr: 0.000028
2023-10-18 21:28:00,391 DEV : loss 0.1396612972021103 - f1-score (micro avg) 0.6016
2023-10-18 21:28:00,410 saving best model
2023-10-18 21:28:00,443 ----------------------------------------------------------------------------------------------------
2023-10-18 21:28:03,439 epoch 6 - iter 198/1984 - loss 0.16107360 - time (sec): 3.00 - samples/sec: 5139.73 - lr: 0.000027 - momentum: 0.000000
2023-10-18 21:28:06,549 epoch 6 - iter 396/1984 - loss 0.15421664 - time (sec): 6.11 - samples/sec: 5364.46 - lr: 0.000027 - momentum: 0.000000
2023-10-18 21:28:09,674 epoch 6 - iter 594/1984 - loss 0.15915795 - time (sec): 9.23 - samples/sec: 5330.25 - lr: 0.000026 - momentum: 0.000000
2023-10-18 21:28:12,713 epoch 6 - iter 792/1984 - loss 0.15438849 - time (sec): 12.27 - samples/sec: 5364.46 - lr: 0.000026 - momentum: 0.000000
2023-10-18 21:28:15,664 epoch 6 - iter 990/1984 - loss 0.15292486 - time (sec): 15.22 - samples/sec: 5399.57 - lr: 0.000025 - momentum: 0.000000
2023-10-18 21:28:18,518 epoch 6 - iter 1188/1984 - loss 0.15391287 - time (sec): 18.07 - samples/sec: 5466.01 - lr: 0.000024 - momentum: 0.000000
2023-10-18 21:28:21,258 epoch 6 - iter 1386/1984 - loss 0.15391731 - time (sec): 20.81 - samples/sec: 5532.31 - lr: 0.000024 - momentum: 0.000000
2023-10-18 21:28:24,192 epoch 6 - iter 1584/1984 - loss 0.15364826 - time (sec): 23.75 - samples/sec: 5517.99 - lr: 0.000023 - momentum: 0.000000
2023-10-18 21:28:27,215 epoch 6 - iter 1782/1984 - loss 0.15355381 - time (sec): 26.77 - samples/sec: 5506.02 - lr: 0.000023 - momentum: 0.000000
2023-10-18 21:28:30,278 epoch 6 - iter 1980/1984 - loss 0.15357759 - time (sec): 29.84 - samples/sec: 5488.25 - lr: 0.000022 - momentum: 0.000000
2023-10-18 21:28:30,343 ----------------------------------------------------------------------------------------------------
2023-10-18 21:28:30,343 EPOCH 6 done: loss 0.1536 - lr: 0.000022
2023-10-18 21:28:32,160 DEV : loss 0.14352329075336456 - f1-score (micro avg) 0.6138
2023-10-18 21:28:32,179 saving best model
2023-10-18 21:28:32,213 ----------------------------------------------------------------------------------------------------
2023-10-18 21:28:35,313 epoch 7 - iter 198/1984 - loss 0.14010170 - time (sec): 3.10 - samples/sec: 5495.13 - lr: 0.000022 - momentum: 0.000000
2023-10-18 21:28:38,375 epoch 7 - iter 396/1984 - loss 0.14331514 - time (sec): 6.16 - samples/sec: 5523.70 - lr: 0.000021 - momentum: 0.000000
2023-10-18 21:28:41,433 epoch 7 - iter 594/1984 - loss 0.14376651 - time (sec): 9.22 - samples/sec: 5353.76 - lr: 0.000021 - momentum: 0.000000
2023-10-18 21:28:44,525 epoch 7 - iter 792/1984 - loss 0.14327862 - time (sec): 12.31 - samples/sec: 5209.88 - lr: 0.000020 - momentum: 0.000000
2023-10-18 21:28:47,677 epoch 7 - iter 990/1984 - loss 0.14545548 - time (sec): 15.46 - samples/sec: 5195.18 - lr: 0.000019 - momentum: 0.000000
2023-10-18 21:28:50,714 epoch 7 - iter 1188/1984 - loss 0.14341143 - time (sec): 18.50 - samples/sec: 5255.03 - lr: 0.000019 - momentum: 0.000000
2023-10-18 21:28:53,759 epoch 7 - iter 1386/1984 - loss 0.14255578 - time (sec): 21.54 - samples/sec: 5272.56 - lr: 0.000018 - momentum: 0.000000
2023-10-18 21:28:56,904 epoch 7 - iter 1584/1984 - loss 0.14220705 - time (sec): 24.69 - samples/sec: 5262.27 - lr: 0.000018 - momentum: 0.000000
2023-10-18 21:28:59,957 epoch 7 - iter 1782/1984 - loss 0.14227082 - time (sec): 27.74 - samples/sec: 5273.02 - lr: 0.000017 - momentum: 0.000000
2023-10-18 21:29:03,032 epoch 7 - iter 1980/1984 - loss 0.14322288 - time (sec): 30.82 - samples/sec: 5315.07 - lr: 0.000017 - momentum: 0.000000
2023-10-18 21:29:03,091 ----------------------------------------------------------------------------------------------------
2023-10-18 21:29:03,091 EPOCH 7 done: loss 0.1432 - lr: 0.000017
2023-10-18 21:29:04,914 DEV : loss 0.1460934579372406 - f1-score (micro avg) 0.6181
2023-10-18 21:29:04,933 saving best model
2023-10-18 21:29:04,967 ----------------------------------------------------------------------------------------------------
2023-10-18 21:29:07,991 epoch 8 - iter 198/1984 - loss 0.15508407 - time (sec): 3.02 - samples/sec: 5680.49 - lr: 0.000016 - momentum: 0.000000
2023-10-18 21:29:11,034 epoch 8 - iter 396/1984 - loss 0.14857609 - time (sec): 6.07 - samples/sec: 5447.32 - lr: 0.000016 - momentum: 0.000000
2023-10-18 21:29:14,053 epoch 8 - iter 594/1984 - loss 0.14831674 - time (sec): 9.09 - samples/sec: 5539.95 - lr: 0.000015 - momentum: 0.000000
2023-10-18 21:29:16,965 epoch 8 - iter 792/1984 - loss 0.14431698 - time (sec): 12.00 - samples/sec: 5580.93 - lr: 0.000014 - momentum: 0.000000
2023-10-18 21:29:19,967 epoch 8 - iter 990/1984 - loss 0.14072716 - time (sec): 15.00 - samples/sec: 5527.79 - lr: 0.000014 - momentum: 0.000000
2023-10-18 21:29:23,046 epoch 8 - iter 1188/1984 - loss 0.13730687 - time (sec): 18.08 - samples/sec: 5511.35 - lr: 0.000013 - momentum: 0.000000
2023-10-18 21:29:26,169 epoch 8 - iter 1386/1984 - loss 0.13802672 - time (sec): 21.20 - samples/sec: 5447.35 - lr: 0.000013 - momentum: 0.000000
2023-10-18 21:29:29,261 epoch 8 - iter 1584/1984 - loss 0.13646616 - time (sec): 24.29 - samples/sec: 5454.78 - lr: 0.000012 - momentum: 0.000000
2023-10-18 21:29:32,269 epoch 8 - iter 1782/1984 - loss 0.13626809 - time (sec): 27.30 - samples/sec: 5441.22 - lr: 0.000012 - momentum: 0.000000
2023-10-18 21:29:35,338 epoch 8 - iter 1980/1984 - loss 0.13563883 - time (sec): 30.37 - samples/sec: 5385.09 - lr: 0.000011 - momentum: 0.000000
2023-10-18 21:29:35,401 ----------------------------------------------------------------------------------------------------
2023-10-18 21:29:35,401 EPOCH 8 done: loss 0.1354 - lr: 0.000011
2023-10-18 21:29:37,660 DEV : loss 0.14632448554039001 - f1-score (micro avg) 0.6125
2023-10-18 21:29:37,681 ----------------------------------------------------------------------------------------------------
2023-10-18 21:29:40,791 epoch 9 - iter 198/1984 - loss 0.12436606 - time (sec): 3.11 - samples/sec: 4996.85 - lr: 0.000011 - momentum: 0.000000
2023-10-18 21:29:43,807 epoch 9 - iter 396/1984 - loss 0.12130140 - time (sec): 6.12 - samples/sec: 5243.22 - lr: 0.000010 - momentum: 0.000000
2023-10-18 21:29:46,923 epoch 9 - iter 594/1984 - loss 0.12036860 - time (sec): 9.24 - samples/sec: 5222.59 - lr: 0.000009 - momentum: 0.000000
2023-10-18 21:29:49,986 epoch 9 - iter 792/1984 - loss 0.12390287 - time (sec): 12.30 - samples/sec: 5269.18 - lr: 0.000009 - momentum: 0.000000
2023-10-18 21:29:52,890 epoch 9 - iter 990/1984 - loss 0.12172877 - time (sec): 15.21 - samples/sec: 5357.59 - lr: 0.000008 - momentum: 0.000000
2023-10-18 21:29:55,968 epoch 9 - iter 1188/1984 - loss 0.12359398 - time (sec): 18.29 - samples/sec: 5348.49 - lr: 0.000008 - momentum: 0.000000
2023-10-18 21:29:59,087 epoch 9 - iter 1386/1984 - loss 0.12601590 - time (sec): 21.41 - samples/sec: 5335.60 - lr: 0.000007 - momentum: 0.000000
2023-10-18 21:30:02,163 epoch 9 - iter 1584/1984 - loss 0.12731249 - time (sec): 24.48 - samples/sec: 5329.19 - lr: 0.000007 - momentum: 0.000000
2023-10-18 21:30:05,015 epoch 9 - iter 1782/1984 - loss 0.13000702 - time (sec): 27.33 - samples/sec: 5411.53 - lr: 0.000006 - momentum: 0.000000
2023-10-18 21:30:07,908 epoch 9 - iter 1980/1984 - loss 0.13026706 - time (sec): 30.23 - samples/sec: 5415.84 - lr: 0.000006 - momentum: 0.000000
2023-10-18 21:30:07,967 ----------------------------------------------------------------------------------------------------
2023-10-18 21:30:07,968 EPOCH 9 done: loss 0.1302 - lr: 0.000006
2023-10-18 21:30:09,795 DEV : loss 0.14653360843658447 - f1-score (micro avg) 0.6208
2023-10-18 21:30:09,816 saving best model
2023-10-18 21:30:09,851 ----------------------------------------------------------------------------------------------------
2023-10-18 21:30:13,178 epoch 10 - iter 198/1984 - loss 0.10287722 - time (sec): 3.33 - samples/sec: 4987.47 - lr: 0.000005 - momentum: 0.000000
2023-10-18 21:30:16,230 epoch 10 - iter 396/1984 - loss 0.11633178 - time (sec): 6.38 - samples/sec: 5153.20 - lr: 0.000004 - momentum: 0.000000
2023-10-18 21:30:19,312 epoch 10 - iter 594/1984 - loss 0.12476201 - time (sec): 9.46 - samples/sec: 5193.09 - lr: 0.000004 - momentum: 0.000000
2023-10-18 21:30:22,352 epoch 10 - iter 792/1984 - loss 0.12536802 - time (sec): 12.50 - samples/sec: 5209.88 - lr: 0.000003 - momentum: 0.000000
2023-10-18 21:30:25,421 epoch 10 - iter 990/1984 - loss 0.12695762 - time (sec): 15.57 - samples/sec: 5294.90 - lr: 0.000003 - momentum: 0.000000
2023-10-18 21:30:28,513 epoch 10 - iter 1188/1984 - loss 0.12707600 - time (sec): 18.66 - samples/sec: 5283.15 - lr: 0.000002 - momentum: 0.000000
2023-10-18 21:30:31,586 epoch 10 - iter 1386/1984 - loss 0.12684283 - time (sec): 21.73 - samples/sec: 5283.84 - lr: 0.000002 - momentum: 0.000000
2023-10-18 21:30:34,749 epoch 10 - iter 1584/1984 - loss 0.12638192 - time (sec): 24.90 - samples/sec: 5252.68 - lr: 0.000001 - momentum: 0.000000
2023-10-18 21:30:37,835 epoch 10 - iter 1782/1984 - loss 0.12660082 - time (sec): 27.98 - samples/sec: 5271.15 - lr: 0.000001 - momentum: 0.000000
2023-10-18 21:30:40,835 epoch 10 - iter 1980/1984 - loss 0.12678588 - time (sec): 30.98 - samples/sec: 5285.51 - lr: 0.000000 - momentum: 0.000000
2023-10-18 21:30:40,894 ----------------------------------------------------------------------------------------------------
2023-10-18 21:30:40,894 EPOCH 10 done: loss 0.1267 - lr: 0.000000
2023-10-18 21:30:42,746 DEV : loss 0.1484779566526413 - f1-score (micro avg) 0.6175
2023-10-18 21:30:42,793 ----------------------------------------------------------------------------------------------------
2023-10-18 21:30:42,793 Loading model from best epoch ...
2023-10-18 21:30:42,875 SequenceTagger predicts: Dictionary with 13 tags: O, S-PER, B-PER, E-PER, I-PER, S-LOC, B-LOC, E-LOC, I-LOC, S-ORG, B-ORG, E-ORG, I-ORG
2023-10-18 21:30:44,397
Results:
- F-score (micro) 0.6425
- F-score (macro) 0.4918
- Accuracy 0.5084
By class:
precision recall f1-score support
LOC 0.7192 0.7664 0.7421 655
PER 0.4534 0.6547 0.5358 223
ORG 0.3778 0.1339 0.1977 127
micro avg 0.6244 0.6617 0.6425 1005
macro avg 0.5168 0.5183 0.4918 1005
weighted avg 0.6171 0.6617 0.6275 1005
2023-10-18 21:30:44,397 ----------------------------------------------------------------------------------------------------