File size: 4,781 Bytes
8ecdf1a 85b6090 e3c6bc8 84b320c f0d1e38 25a1eeb 74bffdd cd3af81 55d7b2d 9c9c762 f656e6b 0a5d8f0 0ebdc8c cb178ff 711c898 f5aa86d 9e46af4 00d2388 d75a248 4cf1834 c4920e5 15a02c7 6461657 73d6e8e 700ac7c 1da0fe1 61b2725 0d6baea cc82b0c 0ff25b2 458124b 2a0d03d 390e53e f1e3eb7 bb1acca f366bc5 381e209 5ac883f 41ffb32 54eb5da 1b31ab2 0c83ab5 60c4a64 ad072a4 aafd56e 6ec2e7d 6d1e465 6918809 dcbb528 b82ce9c c3635dd b75a562 7a4a3d5 5c16c4b 65080ab 34dbfc1 a458191 59915cc 8377a8e 1a87211 866acd0 6788872 7c84aff 6f152f8 557e3f4 c3eb03f 5fea102 fe115eb 2f261be 89b20bc ad41e2d 6944f69 08a7bb0 2970e05 9a70e81 4540483 da65d5a 1084f61 88ac2bd 3d7a869 ab2be0d d824807 210abf0 fd70ab0 9364dc5 ab5c667 4c091a9 da0ae81 ec8d842 09868c5 da37fc8 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 |
Started at: 00:24:23 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.6361208361387253 - MAE: 0.6347873231265873 Validation loss : 0.33865517377853394 - MAE: 0.4740741020180635 Epoch: 1 Training loss: 0.21345499098300935 - MAE: 0.3540912858998663 Validation loss : 0.18473448190424177 - MAE: 0.3282540270842396 Epoch: 2 Training loss: 0.18606460213661194 - MAE: 0.32889828686188877 Validation loss : 0.18045285178555381 - MAE: 0.32713884374037794 Epoch: 3 Training loss: 0.17846946954727172 - MAE: 0.31904566117043975 Validation loss : 0.17773988511827257 - MAE: 0.32482577310153565 Epoch: 4 Training loss: 0.17139309346675874 - MAE: 0.31441827165194275 Validation loss : 0.17418616182274288 - MAE: 0.32107063596093277 Epoch: 5 Training loss: 0.17030848145484925 - MAE: 0.31198532199217943 Validation loss : 0.17218486799134147 - MAE: 0.3190577970415032 Epoch: 6 Training loss: 0.16878579378128053 - MAE: 0.31078672095469284 Validation loss : 0.17055005000697243 - MAE: 0.3174862826881529 Epoch: 7 Training loss: 0.1679440525174141 - MAE: 0.31018192416787727 Validation loss : 0.16925331950187683 - MAE: 0.31617688409581507 Epoch: 8 Training loss: 0.16459781765937806 - MAE: 0.3076978685246546 Validation loss : 0.16737455626328787 - MAE: 0.313686903818875 Epoch: 9 Training loss: 0.1616947764158249 - MAE: 0.3042913169405816 Validation loss : 0.16636485523647732 - MAE: 0.3125500280647287 Started at: 01:47:29 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.6361208361387253 - MAE: 0.6347873231265873 Validation loss : 0.33865517377853394 - MAE: 0.4740741020180635 Epoch: 1 Training loss: 0.21345499098300935 - MAE: 0.3540912858998663 Validation loss : 0.18473448190424177 - MAE: 0.3282540270842396 Epoch: 2 Training loss: 0.18606460213661194 - MAE: 0.32889828686188877 Validation loss : 0.18045285178555381 - MAE: 0.32713884374037794 Epoch: 3 Training loss: 0.17846946954727172 - MAE: 0.31904566117043975 Validation loss : 0.17773988511827257 - MAE: 0.32482577310153565 Epoch: 4 Training loss: 0.17139309346675874 - MAE: 0.31441827165194275 Validation loss : 0.17418616182274288 - MAE: 0.32107063596093277 Epoch: 5 Training loss: 0.17030848145484925 - MAE: 0.31198532199217943 Validation loss : 0.17218486799134147 - MAE: 0.3190577970415032 Epoch: 6 Training loss: 0.16878579378128053 - MAE: 0.31078672095469284 Validation loss : 0.17055005000697243 - MAE: 0.3174862826881529 Epoch: 7 Training loss: 0.1679440525174141 - MAE: 0.31018192416787727 Validation loss : 0.16925331950187683 - MAE: 0.31617688409581507 Epoch: 8 Training loss: 0.16459781765937806 - MAE: 0.3076978685246546 Validation loss : 0.16737455626328787 - MAE: 0.313686903818875 Epoch: 9 Training loss: 0.1616947764158249 - MAE: 0.3042913169405816 Validation loss : 0.16636485523647732 - MAE: 0.3125500280647287 Epoch: 10 Training loss: 0.1614231312274933 - MAE: 0.30281138726351875 Validation loss : 0.16527396109369066 - MAE: 0.310935884275779 Epoch: 11 Training loss: 0.15981832027435303 - MAE: 0.3036158805084463 Validation loss : 0.16393334501319462 - MAE: 0.3090806138434242 Epoch: 12 Training loss: 0.1593213653564453 - MAE: 0.30149370061446074 Validation loss : 0.1633289953072866 - MAE: 0.3085658493030115 Epoch: 13 Training loss: 0.15871796071529387 - MAE: 0.30060354026448305 Validation loss : 0.16265849934683907 - MAE: 0.30747698672849366 Epoch: 14 Training loss: 0.15790118873119355 - MAE: 0.29930967856529156 Validation loss : 0.1618209034204483 - MAE: 0.3064716457801814 Epoch: 15 Training loss: 0.15594214886426927 - MAE: 0.2976610109442401 Validation loss : 0.1610150718026691 - MAE: 0.30535239027482275 Epoch: 16 Training loss: 0.15538664102554323 - MAE: 0.29795636864916086 Validation loss : 0.16068673796123928 - MAE: 0.3052027362150315 Epoch: 17 Training loss: 0.15572097659111023 - MAE: 0.2970189372002837 Validation loss : 0.16038145456049177 - MAE: 0.3048613001870874 Epoch: 18 Training loss: 0.15602077960968017 - MAE: 0.29858003145757656 Validation loss : 0.15998652411831749 - MAE: 0.3046400836871299 |