stefan-it's picture
Upload ./training.log with huggingface_hub
6c3e9d9
2023-10-25 11:54:59,919 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,920 Model: "SequenceTagger(
(embeddings): TransformerWordEmbeddings(
(model): BertModel(
(embeddings): BertEmbeddings(
(word_embeddings): Embedding(64001, 768)
(position_embeddings): Embedding(512, 768)
(token_type_embeddings): Embedding(2, 768)
(LayerNorm): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
(dropout): Dropout(p=0.1, inplace=False)
)
(encoder): BertEncoder(
(layer): ModuleList(
(0-11): 12 x BertLayer(
(attention): BertAttention(
(self): BertSelfAttention(
(query): Linear(in_features=768, out_features=768, bias=True)
(key): Linear(in_features=768, out_features=768, bias=True)
(value): Linear(in_features=768, out_features=768, bias=True)
(dropout): Dropout(p=0.1, inplace=False)
)
(output): BertSelfOutput(
(dense): Linear(in_features=768, out_features=768, bias=True)
(LayerNorm): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
(dropout): Dropout(p=0.1, inplace=False)
)
)
(intermediate): BertIntermediate(
(dense): Linear(in_features=768, out_features=3072, bias=True)
(intermediate_act_fn): GELUActivation()
)
(output): BertOutput(
(dense): Linear(in_features=3072, out_features=768, bias=True)
(LayerNorm): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
(dropout): Dropout(p=0.1, inplace=False)
)
)
)
)
(pooler): BertPooler(
(dense): Linear(in_features=768, out_features=768, bias=True)
(activation): Tanh()
)
)
)
(locked_dropout): LockedDropout(p=0.5)
(linear): Linear(in_features=768, out_features=13, bias=True)
(loss_function): CrossEntropyLoss()
)"
2023-10-25 11:54:59,920 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,920 MultiCorpus: 6183 train + 680 dev + 2113 test sentences
- NER_HIPE_2022 Corpus: 6183 train + 680 dev + 2113 test sentences - /root/.flair/datasets/ner_hipe_2022/v2.1/topres19th/en/with_doc_seperator
2023-10-25 11:54:59,920 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,920 Train: 6183 sentences
2023-10-25 11:54:59,920 (train_with_dev=False, train_with_test=False)
2023-10-25 11:54:59,920 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,920 Training Params:
2023-10-25 11:54:59,920 - learning_rate: "5e-05"
2023-10-25 11:54:59,920 - mini_batch_size: "4"
2023-10-25 11:54:59,920 - max_epochs: "10"
2023-10-25 11:54:59,920 - shuffle: "True"
2023-10-25 11:54:59,920 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,921 Plugins:
2023-10-25 11:54:59,921 - TensorboardLogger
2023-10-25 11:54:59,921 - LinearScheduler | warmup_fraction: '0.1'
2023-10-25 11:54:59,921 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,921 Final evaluation on model from best epoch (best-model.pt)
2023-10-25 11:54:59,921 - metric: "('micro avg', 'f1-score')"
2023-10-25 11:54:59,921 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,921 Computation:
2023-10-25 11:54:59,921 - compute on device: cuda:0
2023-10-25 11:54:59,921 - embedding storage: none
2023-10-25 11:54:59,921 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,921 Model training base path: "hmbench-topres19th/en-dbmdz/bert-base-historic-multilingual-64k-td-cased-bs4-wsFalse-e10-lr5e-05-poolingfirst-layers-1-crfFalse-3"
2023-10-25 11:54:59,921 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,921 ----------------------------------------------------------------------------------------------------
2023-10-25 11:54:59,921 Logging anything other than scalars to TensorBoard is currently not supported.
2023-10-25 11:55:07,914 epoch 1 - iter 154/1546 - loss 1.25390389 - time (sec): 7.99 - samples/sec: 1781.21 - lr: 0.000005 - momentum: 0.000000
2023-10-25 11:55:15,333 epoch 1 - iter 308/1546 - loss 0.76642917 - time (sec): 15.41 - samples/sec: 1690.78 - lr: 0.000010 - momentum: 0.000000
2023-10-25 11:55:23,717 epoch 1 - iter 462/1546 - loss 0.56225091 - time (sec): 23.80 - samples/sec: 1626.05 - lr: 0.000015 - momentum: 0.000000
2023-10-25 11:55:31,627 epoch 1 - iter 616/1546 - loss 0.45626833 - time (sec): 31.70 - samples/sec: 1603.14 - lr: 0.000020 - momentum: 0.000000
2023-10-25 11:55:39,613 epoch 1 - iter 770/1546 - loss 0.39138368 - time (sec): 39.69 - samples/sec: 1593.46 - lr: 0.000025 - momentum: 0.000000
2023-10-25 11:55:47,776 epoch 1 - iter 924/1546 - loss 0.35571490 - time (sec): 47.85 - samples/sec: 1556.21 - lr: 0.000030 - momentum: 0.000000
2023-10-25 11:55:55,520 epoch 1 - iter 1078/1546 - loss 0.32332623 - time (sec): 55.60 - samples/sec: 1557.39 - lr: 0.000035 - momentum: 0.000000
2023-10-25 11:56:03,305 epoch 1 - iter 1232/1546 - loss 0.29722994 - time (sec): 63.38 - samples/sec: 1563.96 - lr: 0.000040 - momentum: 0.000000
2023-10-25 11:56:11,421 epoch 1 - iter 1386/1546 - loss 0.27559241 - time (sec): 71.50 - samples/sec: 1558.87 - lr: 0.000045 - momentum: 0.000000
2023-10-25 11:56:19,164 epoch 1 - iter 1540/1546 - loss 0.25930733 - time (sec): 79.24 - samples/sec: 1561.70 - lr: 0.000050 - momentum: 0.000000
2023-10-25 11:56:19,464 ----------------------------------------------------------------------------------------------------
2023-10-25 11:56:19,464 EPOCH 1 done: loss 0.2589 - lr: 0.000050
2023-10-25 11:56:22,011 DEV : loss 0.0908343568444252 - f1-score (micro avg) 0.7229
2023-10-25 11:56:22,033 saving best model
2023-10-25 11:56:22,639 ----------------------------------------------------------------------------------------------------
2023-10-25 11:56:30,630 epoch 2 - iter 154/1546 - loss 0.09739237 - time (sec): 7.99 - samples/sec: 1474.65 - lr: 0.000049 - momentum: 0.000000
2023-10-25 11:56:38,495 epoch 2 - iter 308/1546 - loss 0.09761310 - time (sec): 15.85 - samples/sec: 1552.78 - lr: 0.000049 - momentum: 0.000000
2023-10-25 11:56:46,873 epoch 2 - iter 462/1546 - loss 0.09917743 - time (sec): 24.23 - samples/sec: 1568.90 - lr: 0.000048 - momentum: 0.000000
2023-10-25 11:56:54,625 epoch 2 - iter 616/1546 - loss 0.09370713 - time (sec): 31.98 - samples/sec: 1598.84 - lr: 0.000048 - momentum: 0.000000
2023-10-25 11:57:02,309 epoch 2 - iter 770/1546 - loss 0.09462855 - time (sec): 39.67 - samples/sec: 1600.11 - lr: 0.000047 - momentum: 0.000000
2023-10-25 11:57:10,081 epoch 2 - iter 924/1546 - loss 0.09486483 - time (sec): 47.44 - samples/sec: 1579.18 - lr: 0.000047 - momentum: 0.000000
2023-10-25 11:57:17,853 epoch 2 - iter 1078/1546 - loss 0.09618995 - time (sec): 55.21 - samples/sec: 1574.82 - lr: 0.000046 - momentum: 0.000000
2023-10-25 11:57:25,684 epoch 2 - iter 1232/1546 - loss 0.09626906 - time (sec): 63.04 - samples/sec: 1578.83 - lr: 0.000046 - momentum: 0.000000
2023-10-25 11:57:33,458 epoch 2 - iter 1386/1546 - loss 0.10061257 - time (sec): 70.82 - samples/sec: 1570.84 - lr: 0.000045 - momentum: 0.000000
2023-10-25 11:57:41,280 epoch 2 - iter 1540/1546 - loss 0.09989085 - time (sec): 78.64 - samples/sec: 1573.54 - lr: 0.000044 - momentum: 0.000000
2023-10-25 11:57:41,600 ----------------------------------------------------------------------------------------------------
2023-10-25 11:57:41,600 EPOCH 2 done: loss 0.0997 - lr: 0.000044
2023-10-25 11:57:44,817 DEV : loss 0.056556034833192825 - f1-score (micro avg) 0.7042
2023-10-25 11:57:44,836 ----------------------------------------------------------------------------------------------------
2023-10-25 11:57:52,516 epoch 3 - iter 154/1546 - loss 0.05181413 - time (sec): 7.68 - samples/sec: 1614.43 - lr: 0.000044 - momentum: 0.000000
2023-10-25 11:58:00,324 epoch 3 - iter 308/1546 - loss 0.05067804 - time (sec): 15.49 - samples/sec: 1602.32 - lr: 0.000043 - momentum: 0.000000
2023-10-25 11:58:08,395 epoch 3 - iter 462/1546 - loss 0.05926773 - time (sec): 23.56 - samples/sec: 1633.22 - lr: 0.000043 - momentum: 0.000000
2023-10-25 11:58:15,574 epoch 3 - iter 616/1546 - loss 0.06021083 - time (sec): 30.74 - samples/sec: 1650.50 - lr: 0.000042 - momentum: 0.000000
2023-10-25 11:58:22,888 epoch 3 - iter 770/1546 - loss 0.06372716 - time (sec): 38.05 - samples/sec: 1638.48 - lr: 0.000042 - momentum: 0.000000
2023-10-25 11:58:30,255 epoch 3 - iter 924/1546 - loss 0.06361460 - time (sec): 45.42 - samples/sec: 1616.96 - lr: 0.000041 - momentum: 0.000000
2023-10-25 11:58:37,724 epoch 3 - iter 1078/1546 - loss 0.06346591 - time (sec): 52.89 - samples/sec: 1633.90 - lr: 0.000041 - momentum: 0.000000
2023-10-25 11:58:45,214 epoch 3 - iter 1232/1546 - loss 0.06338422 - time (sec): 60.38 - samples/sec: 1635.44 - lr: 0.000040 - momentum: 0.000000
2023-10-25 11:58:52,791 epoch 3 - iter 1386/1546 - loss 0.06363303 - time (sec): 67.95 - samples/sec: 1638.00 - lr: 0.000039 - momentum: 0.000000
2023-10-25 11:59:00,043 epoch 3 - iter 1540/1546 - loss 0.06408898 - time (sec): 75.21 - samples/sec: 1648.35 - lr: 0.000039 - momentum: 0.000000
2023-10-25 11:59:00,317 ----------------------------------------------------------------------------------------------------
2023-10-25 11:59:00,318 EPOCH 3 done: loss 0.0642 - lr: 0.000039
2023-10-25 11:59:03,151 DEV : loss 0.10106009989976883 - f1-score (micro avg) 0.7469
2023-10-25 11:59:03,168 saving best model
2023-10-25 11:59:03,853 ----------------------------------------------------------------------------------------------------
2023-10-25 11:59:11,915 epoch 4 - iter 154/1546 - loss 0.04283787 - time (sec): 8.06 - samples/sec: 1594.31 - lr: 0.000038 - momentum: 0.000000
2023-10-25 11:59:19,911 epoch 4 - iter 308/1546 - loss 0.05126655 - time (sec): 16.05 - samples/sec: 1603.31 - lr: 0.000038 - momentum: 0.000000
2023-10-25 11:59:27,978 epoch 4 - iter 462/1546 - loss 0.04992581 - time (sec): 24.12 - samples/sec: 1593.74 - lr: 0.000037 - momentum: 0.000000
2023-10-25 11:59:35,538 epoch 4 - iter 616/1546 - loss 0.05045699 - time (sec): 31.68 - samples/sec: 1602.73 - lr: 0.000037 - momentum: 0.000000
2023-10-25 11:59:43,185 epoch 4 - iter 770/1546 - loss 0.04799193 - time (sec): 39.33 - samples/sec: 1610.64 - lr: 0.000036 - momentum: 0.000000
2023-10-25 11:59:51,099 epoch 4 - iter 924/1546 - loss 0.04985205 - time (sec): 47.24 - samples/sec: 1605.35 - lr: 0.000036 - momentum: 0.000000
2023-10-25 11:59:58,609 epoch 4 - iter 1078/1546 - loss 0.05016705 - time (sec): 54.75 - samples/sec: 1598.42 - lr: 0.000035 - momentum: 0.000000
2023-10-25 12:00:06,623 epoch 4 - iter 1232/1546 - loss 0.05063862 - time (sec): 62.77 - samples/sec: 1579.46 - lr: 0.000034 - momentum: 0.000000
2023-10-25 12:00:14,517 epoch 4 - iter 1386/1546 - loss 0.05249377 - time (sec): 70.66 - samples/sec: 1565.67 - lr: 0.000034 - momentum: 0.000000
2023-10-25 12:00:22,462 epoch 4 - iter 1540/1546 - loss 0.05219490 - time (sec): 78.61 - samples/sec: 1573.98 - lr: 0.000033 - momentum: 0.000000
2023-10-25 12:00:22,782 ----------------------------------------------------------------------------------------------------
2023-10-25 12:00:22,782 EPOCH 4 done: loss 0.0521 - lr: 0.000033
2023-10-25 12:00:25,339 DEV : loss 0.10263549536466599 - f1-score (micro avg) 0.7336
2023-10-25 12:00:25,356 ----------------------------------------------------------------------------------------------------
2023-10-25 12:00:33,332 epoch 5 - iter 154/1546 - loss 0.04033890 - time (sec): 7.97 - samples/sec: 1471.01 - lr: 0.000033 - momentum: 0.000000
2023-10-25 12:00:40,954 epoch 5 - iter 308/1546 - loss 0.03417886 - time (sec): 15.60 - samples/sec: 1537.25 - lr: 0.000032 - momentum: 0.000000
2023-10-25 12:00:48,579 epoch 5 - iter 462/1546 - loss 0.03520716 - time (sec): 23.22 - samples/sec: 1580.31 - lr: 0.000032 - momentum: 0.000000
2023-10-25 12:00:55,977 epoch 5 - iter 616/1546 - loss 0.03469198 - time (sec): 30.62 - samples/sec: 1604.23 - lr: 0.000031 - momentum: 0.000000
2023-10-25 12:01:03,419 epoch 5 - iter 770/1546 - loss 0.03632234 - time (sec): 38.06 - samples/sec: 1592.53 - lr: 0.000031 - momentum: 0.000000
2023-10-25 12:01:10,602 epoch 5 - iter 924/1546 - loss 0.03675886 - time (sec): 45.24 - samples/sec: 1622.62 - lr: 0.000030 - momentum: 0.000000
2023-10-25 12:01:18,263 epoch 5 - iter 1078/1546 - loss 0.03717145 - time (sec): 52.91 - samples/sec: 1620.29 - lr: 0.000029 - momentum: 0.000000
2023-10-25 12:01:25,824 epoch 5 - iter 1232/1546 - loss 0.03733963 - time (sec): 60.47 - samples/sec: 1623.60 - lr: 0.000029 - momentum: 0.000000
2023-10-25 12:01:33,254 epoch 5 - iter 1386/1546 - loss 0.03739261 - time (sec): 67.90 - samples/sec: 1626.40 - lr: 0.000028 - momentum: 0.000000
2023-10-25 12:01:40,791 epoch 5 - iter 1540/1546 - loss 0.03609247 - time (sec): 75.43 - samples/sec: 1643.28 - lr: 0.000028 - momentum: 0.000000
2023-10-25 12:01:41,081 ----------------------------------------------------------------------------------------------------
2023-10-25 12:01:41,081 EPOCH 5 done: loss 0.0361 - lr: 0.000028
2023-10-25 12:01:43,649 DEV : loss 0.10063710063695908 - f1-score (micro avg) 0.7578
2023-10-25 12:01:43,670 saving best model
2023-10-25 12:01:44,401 ----------------------------------------------------------------------------------------------------
2023-10-25 12:01:52,516 epoch 6 - iter 154/1546 - loss 0.02144957 - time (sec): 8.11 - samples/sec: 1503.38 - lr: 0.000027 - momentum: 0.000000
2023-10-25 12:02:00,367 epoch 6 - iter 308/1546 - loss 0.01944377 - time (sec): 15.96 - samples/sec: 1548.06 - lr: 0.000027 - momentum: 0.000000
2023-10-25 12:02:07,813 epoch 6 - iter 462/1546 - loss 0.02131841 - time (sec): 23.41 - samples/sec: 1594.82 - lr: 0.000026 - momentum: 0.000000
2023-10-25 12:02:15,050 epoch 6 - iter 616/1546 - loss 0.02451594 - time (sec): 30.65 - samples/sec: 1608.35 - lr: 0.000026 - momentum: 0.000000
2023-10-25 12:02:22,532 epoch 6 - iter 770/1546 - loss 0.02711661 - time (sec): 38.13 - samples/sec: 1659.82 - lr: 0.000025 - momentum: 0.000000
2023-10-25 12:02:29,884 epoch 6 - iter 924/1546 - loss 0.02668924 - time (sec): 45.48 - samples/sec: 1666.07 - lr: 0.000024 - momentum: 0.000000
2023-10-25 12:02:37,236 epoch 6 - iter 1078/1546 - loss 0.02684771 - time (sec): 52.83 - samples/sec: 1654.02 - lr: 0.000024 - momentum: 0.000000
2023-10-25 12:02:44,824 epoch 6 - iter 1232/1546 - loss 0.02616481 - time (sec): 60.42 - samples/sec: 1648.23 - lr: 0.000023 - momentum: 0.000000
2023-10-25 12:02:52,412 epoch 6 - iter 1386/1546 - loss 0.02831231 - time (sec): 68.01 - samples/sec: 1640.42 - lr: 0.000023 - momentum: 0.000000
2023-10-25 12:02:59,852 epoch 6 - iter 1540/1546 - loss 0.02698839 - time (sec): 75.45 - samples/sec: 1641.36 - lr: 0.000022 - momentum: 0.000000
2023-10-25 12:03:00,129 ----------------------------------------------------------------------------------------------------
2023-10-25 12:03:00,129 EPOCH 6 done: loss 0.0269 - lr: 0.000022
2023-10-25 12:03:02,808 DEV : loss 0.11260586231946945 - f1-score (micro avg) 0.7391
2023-10-25 12:03:02,825 ----------------------------------------------------------------------------------------------------
2023-10-25 12:03:10,873 epoch 7 - iter 154/1546 - loss 0.02745142 - time (sec): 8.05 - samples/sec: 1514.77 - lr: 0.000022 - momentum: 0.000000
2023-10-25 12:03:19,090 epoch 7 - iter 308/1546 - loss 0.02694082 - time (sec): 16.26 - samples/sec: 1527.22 - lr: 0.000021 - momentum: 0.000000
2023-10-25 12:03:27,221 epoch 7 - iter 462/1546 - loss 0.02268879 - time (sec): 24.39 - samples/sec: 1535.08 - lr: 0.000021 - momentum: 0.000000
2023-10-25 12:03:35,458 epoch 7 - iter 616/1546 - loss 0.01949071 - time (sec): 32.63 - samples/sec: 1521.69 - lr: 0.000020 - momentum: 0.000000
2023-10-25 12:03:43,242 epoch 7 - iter 770/1546 - loss 0.02044806 - time (sec): 40.42 - samples/sec: 1518.87 - lr: 0.000019 - momentum: 0.000000
2023-10-25 12:03:50,661 epoch 7 - iter 924/1546 - loss 0.02138991 - time (sec): 47.83 - samples/sec: 1556.48 - lr: 0.000019 - momentum: 0.000000
2023-10-25 12:03:58,015 epoch 7 - iter 1078/1546 - loss 0.02319470 - time (sec): 55.19 - samples/sec: 1566.13 - lr: 0.000018 - momentum: 0.000000
2023-10-25 12:04:06,082 epoch 7 - iter 1232/1546 - loss 0.02195662 - time (sec): 63.25 - samples/sec: 1552.82 - lr: 0.000018 - momentum: 0.000000
2023-10-25 12:04:13,523 epoch 7 - iter 1386/1546 - loss 0.02149172 - time (sec): 70.70 - samples/sec: 1571.28 - lr: 0.000017 - momentum: 0.000000
2023-10-25 12:04:20,843 epoch 7 - iter 1540/1546 - loss 0.02143786 - time (sec): 78.02 - samples/sec: 1587.63 - lr: 0.000017 - momentum: 0.000000
2023-10-25 12:04:21,121 ----------------------------------------------------------------------------------------------------
2023-10-25 12:04:21,121 EPOCH 7 done: loss 0.0214 - lr: 0.000017
2023-10-25 12:04:24,404 DEV : loss 0.12575747072696686 - f1-score (micro avg) 0.7455
2023-10-25 12:04:24,421 ----------------------------------------------------------------------------------------------------
2023-10-25 12:04:32,398 epoch 8 - iter 154/1546 - loss 0.01166650 - time (sec): 7.98 - samples/sec: 1498.85 - lr: 0.000016 - momentum: 0.000000
2023-10-25 12:04:40,449 epoch 8 - iter 308/1546 - loss 0.01080450 - time (sec): 16.03 - samples/sec: 1560.63 - lr: 0.000016 - momentum: 0.000000
2023-10-25 12:04:47,717 epoch 8 - iter 462/1546 - loss 0.00876722 - time (sec): 23.30 - samples/sec: 1617.64 - lr: 0.000015 - momentum: 0.000000
2023-10-25 12:04:54,944 epoch 8 - iter 616/1546 - loss 0.01013665 - time (sec): 30.52 - samples/sec: 1646.31 - lr: 0.000014 - momentum: 0.000000
2023-10-25 12:05:02,132 epoch 8 - iter 770/1546 - loss 0.01169730 - time (sec): 37.71 - samples/sec: 1660.62 - lr: 0.000014 - momentum: 0.000000
2023-10-25 12:05:09,312 epoch 8 - iter 924/1546 - loss 0.01259154 - time (sec): 44.89 - samples/sec: 1651.26 - lr: 0.000013 - momentum: 0.000000
2023-10-25 12:05:17,209 epoch 8 - iter 1078/1546 - loss 0.01208262 - time (sec): 52.79 - samples/sec: 1642.63 - lr: 0.000013 - momentum: 0.000000
2023-10-25 12:05:25,133 epoch 8 - iter 1232/1546 - loss 0.01242729 - time (sec): 60.71 - samples/sec: 1629.92 - lr: 0.000012 - momentum: 0.000000
2023-10-25 12:05:33,093 epoch 8 - iter 1386/1546 - loss 0.01220123 - time (sec): 68.67 - samples/sec: 1618.55 - lr: 0.000012 - momentum: 0.000000
2023-10-25 12:05:40,793 epoch 8 - iter 1540/1546 - loss 0.01251829 - time (sec): 76.37 - samples/sec: 1618.93 - lr: 0.000011 - momentum: 0.000000
2023-10-25 12:05:41,103 ----------------------------------------------------------------------------------------------------
2023-10-25 12:05:41,104 EPOCH 8 done: loss 0.0126 - lr: 0.000011
2023-10-25 12:05:44,000 DEV : loss 0.13706336915493011 - f1-score (micro avg) 0.7418
2023-10-25 12:05:44,018 ----------------------------------------------------------------------------------------------------
2023-10-25 12:05:52,150 epoch 9 - iter 154/1546 - loss 0.00423093 - time (sec): 8.13 - samples/sec: 1508.70 - lr: 0.000011 - momentum: 0.000000
2023-10-25 12:05:59,918 epoch 9 - iter 308/1546 - loss 0.00796578 - time (sec): 15.90 - samples/sec: 1535.43 - lr: 0.000010 - momentum: 0.000000
2023-10-25 12:06:07,701 epoch 9 - iter 462/1546 - loss 0.00742909 - time (sec): 23.68 - samples/sec: 1543.35 - lr: 0.000009 - momentum: 0.000000
2023-10-25 12:06:15,465 epoch 9 - iter 616/1546 - loss 0.00683659 - time (sec): 31.44 - samples/sec: 1554.84 - lr: 0.000009 - momentum: 0.000000
2023-10-25 12:06:23,449 epoch 9 - iter 770/1546 - loss 0.00668404 - time (sec): 39.43 - samples/sec: 1580.63 - lr: 0.000008 - momentum: 0.000000
2023-10-25 12:06:30,775 epoch 9 - iter 924/1546 - loss 0.00669052 - time (sec): 46.75 - samples/sec: 1594.92 - lr: 0.000008 - momentum: 0.000000
2023-10-25 12:06:38,008 epoch 9 - iter 1078/1546 - loss 0.00778673 - time (sec): 53.99 - samples/sec: 1608.11 - lr: 0.000007 - momentum: 0.000000
2023-10-25 12:06:45,365 epoch 9 - iter 1232/1546 - loss 0.00738229 - time (sec): 61.34 - samples/sec: 1626.26 - lr: 0.000007 - momentum: 0.000000
2023-10-25 12:06:52,500 epoch 9 - iter 1386/1546 - loss 0.00725643 - time (sec): 68.48 - samples/sec: 1641.64 - lr: 0.000006 - momentum: 0.000000
2023-10-25 12:06:59,907 epoch 9 - iter 1540/1546 - loss 0.00755278 - time (sec): 75.89 - samples/sec: 1632.10 - lr: 0.000006 - momentum: 0.000000
2023-10-25 12:07:00,187 ----------------------------------------------------------------------------------------------------
2023-10-25 12:07:00,187 EPOCH 9 done: loss 0.0076 - lr: 0.000006
2023-10-25 12:07:02,823 DEV : loss 0.13333649933338165 - f1-score (micro avg) 0.7536
2023-10-25 12:07:02,840 ----------------------------------------------------------------------------------------------------
2023-10-25 12:07:10,967 epoch 10 - iter 154/1546 - loss 0.00165995 - time (sec): 8.13 - samples/sec: 1685.09 - lr: 0.000005 - momentum: 0.000000
2023-10-25 12:07:18,856 epoch 10 - iter 308/1546 - loss 0.00222406 - time (sec): 16.02 - samples/sec: 1592.31 - lr: 0.000004 - momentum: 0.000000
2023-10-25 12:07:26,838 epoch 10 - iter 462/1546 - loss 0.00205192 - time (sec): 24.00 - samples/sec: 1595.86 - lr: 0.000004 - momentum: 0.000000
2023-10-25 12:07:34,915 epoch 10 - iter 616/1546 - loss 0.00309192 - time (sec): 32.07 - samples/sec: 1579.51 - lr: 0.000003 - momentum: 0.000000
2023-10-25 12:07:42,987 epoch 10 - iter 770/1546 - loss 0.00355935 - time (sec): 40.15 - samples/sec: 1584.03 - lr: 0.000003 - momentum: 0.000000
2023-10-25 12:07:51,142 epoch 10 - iter 924/1546 - loss 0.00377660 - time (sec): 48.30 - samples/sec: 1572.64 - lr: 0.000002 - momentum: 0.000000
2023-10-25 12:07:59,104 epoch 10 - iter 1078/1546 - loss 0.00389532 - time (sec): 56.26 - samples/sec: 1560.53 - lr: 0.000002 - momentum: 0.000000
2023-10-25 12:08:07,198 epoch 10 - iter 1232/1546 - loss 0.00374600 - time (sec): 64.36 - samples/sec: 1549.34 - lr: 0.000001 - momentum: 0.000000
2023-10-25 12:08:15,337 epoch 10 - iter 1386/1546 - loss 0.00399484 - time (sec): 72.50 - samples/sec: 1544.18 - lr: 0.000001 - momentum: 0.000000
2023-10-25 12:08:23,099 epoch 10 - iter 1540/1546 - loss 0.00406636 - time (sec): 80.26 - samples/sec: 1543.72 - lr: 0.000000 - momentum: 0.000000
2023-10-25 12:08:23,376 ----------------------------------------------------------------------------------------------------
2023-10-25 12:08:23,377 EPOCH 10 done: loss 0.0042 - lr: 0.000000
2023-10-25 12:08:26,002 DEV : loss 0.14262209832668304 - f1-score (micro avg) 0.7546
2023-10-25 12:08:26,470 ----------------------------------------------------------------------------------------------------
2023-10-25 12:08:26,472 Loading model from best epoch ...
2023-10-25 12:08:28,407 SequenceTagger predicts: Dictionary with 13 tags: O, S-LOC, B-LOC, E-LOC, I-LOC, S-BUILDING, B-BUILDING, E-BUILDING, I-BUILDING, S-STREET, B-STREET, E-STREET, I-STREET
2023-10-25 12:08:37,488
Results:
- F-score (micro) 0.7585
- F-score (macro) 0.5929
- Accuracy 0.6256
By class:
precision recall f1-score support
LOC 0.7988 0.8436 0.8206 946
BUILDING 0.5938 0.3081 0.4057 185
STREET 0.5918 0.5179 0.5524 56
micro avg 0.7727 0.7447 0.7585 1187
macro avg 0.6615 0.5565 0.5929 1187
weighted avg 0.7571 0.7447 0.7433 1187
2023-10-25 12:08:37,488 ----------------------------------------------------------------------------------------------------