File size: 3,002 Bytes
1f0e714 feee2fb afd0df6 cedc360 4076651 198bc82 9be681b 635718c 6f37545 f4868cc af83f0f 01d8801 3821530 c7b074c 918963b b018bd1 63ab209 c5567dd a71b618 a629578 1010e1d c0de51a e561a10 59315ef 6f2a08e f5d7e65 35ad807 f965038 3976f6f 8ef0909 8db24bc 90c7d14 df6f1ec 2878461 e339fae ed57cb7 1f77e8b 753cd9d eccf51c 5b9357d db04a2e 9e3ca02 299f942 bf4c388 cee2e35 3d23b7a 0b6872d 3e0f339 84dc40b 7130898 6dd7293 1f74be9 06651c4 04b59e7 3fae253 6a92ac2 cbf53a8 dfac188 579bd0e 323278b 16e2d4f e840fd5 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 |
Started at: 21:52:14 norbert2, 0.001, 64 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.2395064114320158 - MAE: 0.3744385315534662 Validation loss : 0.20704752658352707 - MAE: 0.34597984502789736 Epoch: 1 Training loss: 0.1767307206085234 - MAE: 0.3192073445602084 Validation loss : 0.17625425801132666 - MAE: 0.3170721942019411 Epoch: 2 Training loss: 0.16618126668412275 - MAE: 0.3092801725863174 Validation loss : 0.17076683383096347 - MAE: 0.3112472933223099 Epoch: 3 Training loss: 0.16126501003299096 - MAE: 0.30584279932575203 Validation loss : 0.16347124337246924 - MAE: 0.3030033733379743 Epoch: 4 Training loss: 0.15625435588034717 - MAE: 0.29963192745208655 Validation loss : 0.15993442521853882 - MAE: 0.3004842166716503 Epoch: 5 Training loss: 0.15695305243887084 - MAE: 0.3010549775416317 Validation loss : 0.16180144634210702 - MAE: 0.3024127057273353 Epoch: 6 Training loss: 0.15556791018355975 - MAE: 0.29878062763401153 Validation loss : 0.15521211989901282 - MAE: 0.2970759091608139 Epoch: 7 Training loss: 0.15237431160428308 - MAE: 0.2953012408610363 Validation loss : 0.1589137056108677 - MAE: 0.29914319058997385 Epoch: 8 Training loss: 0.15153679695695338 - MAE: 0.29506802949496486 Validation loss : 0.15345649150284854 - MAE: 0.29414442352803954 Epoch: 9 Training loss: 0.1525212698663124 - MAE: 0.2963382092044523 Validation loss : 0.1529899320819161 - MAE: 0.29337032979089 Epoch: 10 Training loss: 0.15221869283252293 - MAE: 0.2953835321481793 Validation loss : 0.15726926471247818 - MAE: 0.29654991958475474 Epoch: 11 Training loss: 0.15084269725614124 - MAE: 0.29360040089351036 Validation loss : 0.15456406478628967 - MAE: 0.2942018313412496 Epoch: 12 Training loss: 0.14898980709940496 - MAE: 0.292388587665275 Validation loss : 0.15269274806434457 - MAE: 0.292333423595863 Epoch: 13 Training loss: 0.14591557916366693 - MAE: 0.2891725097986123 Validation loss : 0.1531008021398024 - MAE: 0.29286502452232843 Epoch: 14 Training loss: 0.15047907114329964 - MAE: 0.2928827053558407 Validation loss : 0.15025645133220789 - MAE: 0.2906735936090682 Epoch: 15 Training loss: 0.15035037162026973 - MAE: 0.2942172721654035 Validation loss : 0.1525195986032486 - MAE: 0.2924193384927183 Epoch: 16 Training loss: 0.14917351215174704 - MAE: 0.2934477894984301 Validation loss : 0.15107204539306235 - MAE: 0.2908045947936088 Epoch: 17 Training loss: 0.1479782590059319 - MAE: 0.2918790273580605 Validation loss : 0.15280635677503818 - MAE: 0.29140434010156746 Stopped after 17 epochs. Prediction MAE: 0.2771 Finished at: 21:52:14 Time taken: 5897 s. 0 days 1 hours 38 minutes 17 seconds |