File size: 3,037 Bytes
7190ad5 4322570 7472bdd b9603da 63d2879 8ef7047 35f59c6 ba49ba0 0d0ed3d 539ad21 4b20ba1 84e7616 f82b0dd 14d8f7a 5f03900 25fc04b 907c700 a9a4db3 2bbc3ac 31669bf 86ec261 166cdd0 ac030e2 3edb54b a4e1a25 6851125 b70cc6b de6eefe 2eacbce ec1a169 10e0fcf ea0dd6b 922e1fe f2b9530 7047d0e b96f08b 13aee63 a71e55f c15727c ffbcad7 9b5c800 938cf69 517369a fd45305 7d397f6 c258b8e 3e96bf6 bb2e1ee 91b7278 3bcc178 10416a6 11d21f7 70a216e 95c2fc1 9bb6d49 74c8cb0 96d57b5 7a4fec5 13d3052 a34a914 8e23044 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 |
Started at: 08:03:19 norbert, 0.001, 320 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.5677837982773781 - MAE: 0.6071012370307216 Validation loss : 0.20541623021875108 - MAE: 0.3553821039391285 Epoch: 1 Training loss: 0.20585036724805833 - MAE: 0.3435565766713197 Validation loss : 0.20508961379528046 - MAE: 0.35595161265108854 Epoch: 2 Training loss: 0.18047694116830826 - MAE: 0.3226156020038149 Validation loss : 0.1755553654261998 - MAE: 0.3231481818337888 Epoch: 3 Training loss: 0.17075181305408477 - MAE: 0.31363742646216203 Validation loss : 0.17132872768810817 - MAE: 0.3185453978044612 Epoch: 4 Training loss: 0.16742072626948357 - MAE: 0.310234020655618 Validation loss : 0.16898912404264724 - MAE: 0.31608206268678873 Epoch: 5 Training loss: 0.16419674679636956 - MAE: 0.3067915436016256 Validation loss : 0.16652749053069524 - MAE: 0.31254405791166223 Epoch: 6 Training loss: 0.161722569167614 - MAE: 0.30472125991941457 Validation loss : 0.16518973452704294 - MAE: 0.31099441861049065 Epoch: 7 Training loss: 0.16122283563017845 - MAE: 0.30372485382572223 Validation loss : 0.16410291833536966 - MAE: 0.30985283497720256 Epoch: 8 Training loss: 0.1606612280011177 - MAE: 0.3034584368058747 Validation loss : 0.16292419178145273 - MAE: 0.3081625656416393 Epoch: 9 Training loss: 0.15927293449640273 - MAE: 0.30205701486209247 Validation loss : 0.16244180713381087 - MAE: 0.30794590258890475 Epoch: 10 Training loss: 0.15609777495265006 - MAE: 0.2989383894238924 Validation loss : 0.16141763968127115 - MAE: 0.3065149170765169 Epoch: 11 Training loss: 0.1561540775001049 - MAE: 0.2985552641022635 Validation loss : 0.16073173497404372 - MAE: 0.30570570883701986 Epoch: 12 Training loss: 0.1551702544093132 - MAE: 0.2968911396481589 Validation loss : 0.16021225707871573 - MAE: 0.30518808422069765 Epoch: 13 Training loss: 0.1527077190577984 - MAE: 0.2943688283146968 Validation loss : 0.15980427605765207 - MAE: 0.3050297453459748 Epoch: 14 Training loss: 0.1537693753838539 - MAE: 0.29611581920427776 Validation loss : 0.1592416518500873 - MAE: 0.3044232209457301 Epoch: 15 Training loss: 0.15109143629670144 - MAE: 0.29332215256492095 Validation loss : 0.15838797922645295 - MAE: 0.3030250668740604 Epoch: 16 Training loss: 0.152120054140687 - MAE: 0.2936483373390902 Validation loss : 0.1587892887847764 - MAE: 0.30440914486462783 Epoch: 17 Training loss: 0.15156445950269698 - MAE: 0.29244372400885443 Validation loss : 0.1583419985004834 - MAE: 0.3039194700887613 Epoch: 18 Training loss: 0.1510380856692791 - MAE: 0.29228889370790057 Validation loss : 0.1578818719301905 - MAE: 0.30315780427375344 Stopped after 18 epochs. |