File size: 2,888 Bytes
7db66ec 1ade577 9457c05 cf6c14b 3c1f5ba b80ec74 45a213a a6c6390 f842217 f935464 0c1ca99 39c51b0 90fc023 3c09443 359c2b1 66f44ee e1e1659 ca7d999 191f729 d0ff4b6 627f2bf 1917344 13010fb 2a6e92b 4c3b633 6f8807e 912542f 863c4c2 2b48916 a64a4c8 d90fe4a 121e0e0 8223f8c d87ef6b 3fc9afa aac7be9 c34ff6e e953db0 db83c3b c99aac5 fb5f173 5c3d37a 7a8dfe2 8cd8763 75df4b2 f8ffcea ceacc6f 0c84e4a ed6891e 667e513 f62d511 5f36335 a2f5139 9100b82 34666f6 9d8cee2 23837db |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 |
Started at: 14:37:27 norbert, 5e-06, 128 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.49241517592560163 - MAE: 0.5587459042828556 Validation loss : 0.35868610159770864 - MAE: 0.4753577830860895 Epoch: 1 Training loss: 0.3175132931633429 - MAE: 0.445445539977737 Validation loss : 0.25557071695456635 - MAE: 0.3910300926165894 Epoch: 2 Training loss: 0.23261894990097393 - MAE: 0.3786371806608561 Validation loss : 0.20166053240363663 - MAE: 0.3493823568582558 Epoch: 3 Training loss: 0.19200442243706095 - MAE: 0.3470611576431532 Validation loss : 0.1713124976770298 - MAE: 0.3275480687779099 Epoch: 4 Training loss: 0.17018077576702292 - MAE: 0.3278047750176803 Validation loss : 0.1574633888698913 - MAE: 0.31174332485895984 Epoch: 5 Training loss: 0.1535857619209723 - MAE: 0.31193839952824526 Validation loss : 0.15219828848903244 - MAE: 0.3075684372060492 Epoch: 6 Training loss: 0.14406093453819102 - MAE: 0.30176564464643024 Validation loss : 0.13898070478761518 - MAE: 0.29244764946070223 Epoch: 7 Training loss: 0.13362374928864565 - MAE: 0.29026767344805254 Validation loss : 0.13048673622511528 - MAE: 0.2813505453732496 Epoch: 8 Training loss: 0.12214575077999722 - MAE: 0.27611727530666424 Validation loss : 0.12706001525795138 - MAE: 0.27665191233280867 Epoch: 9 Training loss: 0.11591105806556615 - MAE: 0.26916014538434074 Validation loss : 0.11628925296905879 - MAE: 0.2643442778374055 Epoch: 10 Training loss: 0.10905246978456323 - MAE: 0.2604420025940941 Validation loss : 0.11431488414873948 - MAE: 0.25982212380330555 Epoch: 11 Training loss: 0.10331852971152826 - MAE: 0.2536274391469398 Validation loss : 0.10815783006113928 - MAE: 0.2503694937076749 Epoch: 12 Training loss: 0.09894988679073073 - MAE: 0.24673591588906577 Validation loss : 0.10309344207918322 - MAE: 0.24489055743019184 Epoch: 13 Training loss: 0.09555300352248279 - MAE: 0.2422491618605809 Validation loss : 0.09517849277000169 - MAE: 0.23459944038365538 Epoch: 14 Training loss: 0.09006806313991547 - MAE: 0.2344158579264854 Validation loss : 0.09714008787193813 - MAE: 0.23673838294794178 Epoch: 15 Training loss: 0.08482879820195112 - MAE: 0.2277674742544599 Validation loss : 0.09199248536213024 - MAE: 0.22825399727560355 Epoch: 16 Training loss: 0.08355843451890078 - MAE: 0.22517612165359246 Validation loss : 0.08988457132835646 - MAE: 0.2259545000346762 Epoch: 17 Training loss: 0.07894870334050873 - MAE: 0.21817967396293864 Validation loss : 0.09049511318271225 - MAE: 0.22485460323935755 |