File size: 13,571 Bytes
2a843a2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 |
2022-08-25 06:06:10,564 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:10,565 Model: "SequenceTagger(
(embeddings): StackedEmbeddings(
(list_embedding_0): FlairEmbeddings(
(lm): LanguageModel(
(drop): Dropout(p=0.25, inplace=False)
(encoder): Embedding(275, 100)
(rnn): LSTM(100, 2048)
(decoder): Linear(in_features=2048, out_features=275, bias=True)
)
)
(list_embedding_1): FlairEmbeddings(
(lm): LanguageModel(
(drop): Dropout(p=0.25, inplace=False)
(encoder): Embedding(275, 100)
(rnn): LSTM(100, 2048)
(decoder): Linear(in_features=2048, out_features=275, bias=True)
)
)
)
(word_dropout): WordDropout(p=0.05)
(locked_dropout): LockedDropout(p=0.5)
(embedding2nn): Linear(in_features=4096, out_features=4096, bias=True)
(rnn): LSTM(4096, 256, batch_first=True, bidirectional=True)
(linear): Linear(in_features=512, out_features=47, bias=True)
(loss_function): ViterbiLoss()
(crf): CRF()
)"
2022-08-25 06:06:10,567 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:10,567 Corpus: "Corpus: 100 train + 20 dev + 20 test sentences"
2022-08-25 06:06:10,567 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:10,568 Parameters:
2022-08-25 06:06:10,568 - learning_rate: "0.100000"
2022-08-25 06:06:10,568 - mini_batch_size: "32"
2022-08-25 06:06:10,568 - patience: "3"
2022-08-25 06:06:10,569 - anneal_factor: "0.5"
2022-08-25 06:06:10,569 - max_epochs: "10"
2022-08-25 06:06:10,569 - shuffle: "True"
2022-08-25 06:06:10,569 - train_with_dev: "False"
2022-08-25 06:06:10,570 - batch_growth_annealing: "False"
2022-08-25 06:06:10,570 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:10,571 Model training base path: "db/kgc_models/test-kgc-sm-flair/trained_model"
2022-08-25 06:06:10,571 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:10,571 Device: cuda:0
2022-08-25 06:06:10,571 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:10,572 Embeddings storage mode: gpu
2022-08-25 06:06:10,572 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:10,659 epoch 1 - iter 1/4 - loss 4.03916875 - samples/sec: 371.51 - lr: 0.100000
2022-08-25 06:06:10,731 epoch 1 - iter 2/4 - loss 3.89573674 - samples/sec: 443.37 - lr: 0.100000
2022-08-25 06:06:10,805 epoch 1 - iter 3/4 - loss 3.73763981 - samples/sec: 441.41 - lr: 0.100000
2022-08-25 06:06:10,853 epoch 1 - iter 4/4 - loss 3.73055211 - samples/sec: 669.71 - lr: 0.100000
2022-08-25 06:06:10,854 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:10,854 EPOCH 1 done: loss 3.7306 - lr 0.100000
2022-08-25 06:06:10,934 Evaluating as a multi-label problem: False
2022-08-25 06:06:10,941 DEV : loss 3.0921895503997803 - f1-score (micro avg) 0.3087
2022-08-25 06:06:10,942 BAD EPOCHS (no improvement): 0
2022-08-25 06:06:10,944 saving best model
2022-08-25 06:06:12,789 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:12,816 epoch 2 - iter 1/4 - loss 2.91073551 - samples/sec: 1237.62 - lr: 0.100000
2022-08-25 06:06:12,842 epoch 2 - iter 2/4 - loss 2.94079960 - samples/sec: 1251.27 - lr: 0.100000
2022-08-25 06:06:12,874 epoch 2 - iter 3/4 - loss 2.78518824 - samples/sec: 1014.72 - lr: 0.100000
2022-08-25 06:06:12,891 epoch 2 - iter 4/4 - loss 2.80299381 - samples/sec: 1978.97 - lr: 0.100000
2022-08-25 06:06:12,892 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:12,892 EPOCH 2 done: loss 2.8030 - lr 0.100000
2022-08-25 06:06:12,932 Evaluating as a multi-label problem: False
2022-08-25 06:06:12,939 DEV : loss 2.827807903289795 - f1-score (micro avg) 0.2556
2022-08-25 06:06:12,940 BAD EPOCHS (no improvement): 1
2022-08-25 06:06:12,941 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:13,054 epoch 3 - iter 1/4 - loss 2.59037260 - samples/sec: 284.59 - lr: 0.100000
2022-08-25 06:06:13,081 epoch 3 - iter 2/4 - loss 2.60334442 - samples/sec: 1216.48 - lr: 0.100000
2022-08-25 06:06:13,108 epoch 3 - iter 3/4 - loss 2.57638893 - samples/sec: 1264.18 - lr: 0.100000
2022-08-25 06:06:13,126 epoch 3 - iter 4/4 - loss 2.53864971 - samples/sec: 1827.04 - lr: 0.100000
2022-08-25 06:06:13,126 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:13,127 EPOCH 3 done: loss 2.5386 - lr 0.100000
2022-08-25 06:06:13,167 Evaluating as a multi-label problem: False
2022-08-25 06:06:13,173 DEV : loss 2.7169787883758545 - f1-score (micro avg) 0.3168
2022-08-25 06:06:13,174 BAD EPOCHS (no improvement): 0
2022-08-25 06:06:13,175 saving best model
2022-08-25 06:06:14,714 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:14,743 epoch 4 - iter 1/4 - loss 2.71329416 - samples/sec: 1140.13 - lr: 0.100000
2022-08-25 06:06:14,769 epoch 4 - iter 2/4 - loss 2.38254593 - samples/sec: 1254.42 - lr: 0.100000
2022-08-25 06:06:14,803 epoch 4 - iter 3/4 - loss 2.37460295 - samples/sec: 958.15 - lr: 0.100000
2022-08-25 06:06:14,825 epoch 4 - iter 4/4 - loss 2.37499146 - samples/sec: 1527.35 - lr: 0.100000
2022-08-25 06:06:14,826 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:14,826 EPOCH 4 done: loss 2.3750 - lr 0.100000
2022-08-25 06:06:14,864 Evaluating as a multi-label problem: False
2022-08-25 06:06:14,870 DEV : loss 2.749277114868164 - f1-score (micro avg) 0.0
2022-08-25 06:06:14,871 BAD EPOCHS (no improvement): 1
2022-08-25 06:06:14,872 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:14,901 epoch 5 - iter 1/4 - loss 2.32981579 - samples/sec: 1152.14 - lr: 0.100000
2022-08-25 06:06:14,928 epoch 5 - iter 2/4 - loss 2.20069920 - samples/sec: 1219.29 - lr: 0.100000
2022-08-25 06:06:14,965 epoch 5 - iter 3/4 - loss 2.27754946 - samples/sec: 869.53 - lr: 0.100000
2022-08-25 06:06:14,986 epoch 5 - iter 4/4 - loss 2.25410557 - samples/sec: 1547.91 - lr: 0.100000
2022-08-25 06:06:14,987 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:14,987 EPOCH 5 done: loss 2.2541 - lr 0.100000
2022-08-25 06:06:15,029 Evaluating as a multi-label problem: False
2022-08-25 06:06:15,047 DEV : loss 2.7171242237091064 - f1-score (micro avg) 0.2903
2022-08-25 06:06:15,049 BAD EPOCHS (no improvement): 2
2022-08-25 06:06:15,050 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:15,089 epoch 6 - iter 1/4 - loss 2.18811890 - samples/sec: 858.42 - lr: 0.100000
2022-08-25 06:06:15,124 epoch 6 - iter 2/4 - loss 2.14233002 - samples/sec: 929.04 - lr: 0.100000
2022-08-25 06:06:15,156 epoch 6 - iter 3/4 - loss 2.20570734 - samples/sec: 1025.55 - lr: 0.100000
2022-08-25 06:06:15,173 epoch 6 - iter 4/4 - loss 2.18303889 - samples/sec: 1924.88 - lr: 0.100000
2022-08-25 06:06:15,174 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:15,174 EPOCH 6 done: loss 2.1830 - lr 0.100000
2022-08-25 06:06:15,210 Evaluating as a multi-label problem: False
2022-08-25 06:06:15,217 DEV : loss 2.6913464069366455 - f1-score (micro avg) 0.2314
2022-08-25 06:06:15,218 BAD EPOCHS (no improvement): 3
2022-08-25 06:06:15,220 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:15,253 epoch 7 - iter 1/4 - loss 2.14348509 - samples/sec: 988.78 - lr: 0.100000
2022-08-25 06:06:15,281 epoch 7 - iter 2/4 - loss 2.00850553 - samples/sec: 1178.96 - lr: 0.100000
2022-08-25 06:06:15,308 epoch 7 - iter 3/4 - loss 2.06308898 - samples/sec: 1208.59 - lr: 0.100000
2022-08-25 06:06:15,327 epoch 7 - iter 4/4 - loss 2.06392736 - samples/sec: 1723.48 - lr: 0.100000
2022-08-25 06:06:15,327 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:15,328 EPOCH 7 done: loss 2.0639 - lr 0.100000
2022-08-25 06:06:15,368 Evaluating as a multi-label problem: False
2022-08-25 06:06:15,376 DEV : loss 2.2334768772125244 - f1-score (micro avg) 0.5
2022-08-25 06:06:15,377 BAD EPOCHS (no improvement): 0
2022-08-25 06:06:15,381 saving best model
2022-08-25 06:06:16,903 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:16,938 epoch 8 - iter 1/4 - loss 1.97582934 - samples/sec: 924.66 - lr: 0.100000
2022-08-25 06:06:16,971 epoch 8 - iter 2/4 - loss 1.89160690 - samples/sec: 1003.74 - lr: 0.100000
2022-08-25 06:06:17,003 epoch 8 - iter 3/4 - loss 1.92096321 - samples/sec: 1015.15 - lr: 0.100000
2022-08-25 06:06:17,024 epoch 8 - iter 4/4 - loss 1.90466607 - samples/sec: 1613.37 - lr: 0.100000
2022-08-25 06:06:17,024 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:17,025 EPOCH 8 done: loss 1.9047 - lr 0.100000
2022-08-25 06:06:17,076 Evaluating as a multi-label problem: False
2022-08-25 06:06:17,086 DEV : loss 2.3985493183135986 - f1-score (micro avg) 0.3448
2022-08-25 06:06:17,087 BAD EPOCHS (no improvement): 1
2022-08-25 06:06:17,090 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:17,130 epoch 9 - iter 1/4 - loss 1.87079010 - samples/sec: 826.37 - lr: 0.100000
2022-08-25 06:06:17,167 epoch 9 - iter 2/4 - loss 1.85216901 - samples/sec: 864.81 - lr: 0.100000
2022-08-25 06:06:17,203 epoch 9 - iter 3/4 - loss 1.90496054 - samples/sec: 924.19 - lr: 0.100000
2022-08-25 06:06:17,225 epoch 9 - iter 4/4 - loss 1.89038416 - samples/sec: 1452.27 - lr: 0.100000
2022-08-25 06:06:17,226 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:17,226 EPOCH 9 done: loss 1.8904 - lr 0.100000
2022-08-25 06:06:17,276 Evaluating as a multi-label problem: False
2022-08-25 06:06:17,284 DEV : loss 2.121741533279419 - f1-score (micro avg) 0.4627
2022-08-25 06:06:17,285 BAD EPOCHS (no improvement): 2
2022-08-25 06:06:17,288 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:17,323 epoch 10 - iter 1/4 - loss 1.80540347 - samples/sec: 930.59 - lr: 0.100000
2022-08-25 06:06:17,354 epoch 10 - iter 2/4 - loss 1.72573052 - samples/sec: 1054.27 - lr: 0.100000
2022-08-25 06:06:17,386 epoch 10 - iter 3/4 - loss 1.77370743 - samples/sec: 1004.27 - lr: 0.100000
2022-08-25 06:06:17,407 epoch 10 - iter 4/4 - loss 1.75549973 - samples/sec: 1615.25 - lr: 0.100000
2022-08-25 06:06:17,407 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:17,408 EPOCH 10 done: loss 1.7555 - lr 0.100000
2022-08-25 06:06:17,453 Evaluating as a multi-label problem: False
2022-08-25 06:06:17,461 DEV : loss 2.2982170581817627 - f1-score (micro avg) 0.3333
2022-08-25 06:06:17,462 BAD EPOCHS (no improvement): 3
2022-08-25 06:06:19,099 ----------------------------------------------------------------------------------------------------
2022-08-25 06:06:19,100 loading file db/kgc_models/test-kgc-sm-flair/trained_model/best-model.pt
2022-08-25 06:06:27,749 SequenceTagger predicts: Dictionary with 47 tags: O, S-Item, B-Item, E-Item, I-Item, S-Activity, B-Activity, E-Activity, I-Activity, S-Observation, B-Observation, E-Observation, I-Observation, S-Location, B-Location, E-Location, I-Location, S-Consumable, B-Consumable, E-Consumable, I-Consumable, S-Agent, B-Agent, E-Agent, I-Agent, S-Attribute, B-Attribute, E-Attribute, I-Attribute, S-Specifier, B-Specifier, E-Specifier, I-Specifier, S-Cardinality, B-Cardinality, E-Cardinality, I-Cardinality, S-Event, B-Event, E-Event, I-Event, S-Time, B-Time, E-Time, I-Time, <START>, <STOP>
2022-08-25 06:06:28,150 Evaluating as a multi-label problem: False
2022-08-25 06:06:28,159 0.3281 0.3333 0.3307 0.2386
2022-08-25 06:06:28,159
Results:
- F-score (micro) 0.3307
- F-score (macro) 0.125
- Accuracy 0.2386
By class:
precision recall f1-score support
Item 0.2679 0.6250 0.3750 24
Activity 0.8571 0.6667 0.7500 9
Observation 0.0000 0.0000 0.0000 13
Location 0.0000 0.0000 0.0000 5
Consumable 0.0000 0.0000 0.0000 4
Agent 0.0000 0.0000 0.0000 3
Cardinality 0.0000 0.0000 0.0000 2
Time 0.0000 0.0000 0.0000 2
Attribute 0.0000 0.0000 0.0000 1
micro avg 0.3281 0.3333 0.3307 63
macro avg 0.1250 0.1435 0.1250 63
weighted avg 0.2245 0.3333 0.2500 63
2022-08-25 06:06:28,160 ----------------------------------------------------------------------------------------------------
|