File size: 3,737 Bytes
71880d9 ca4f91f 96c99f7 417bc43 a4648b4 8dd5c65 f222d7b 8f0231e 04fe6b3 cf72863 948d00e 0772b07 74b4a8e 46ed52d 96c7bc9 538e5f2 10968ce 848d2d2 0247345 08799df 1f168eb 9d83104 8f538ad e5a92fa f14b91f 6bc0ab8 504da7f 45d0acb f25e3d4 5df0eb7 2091070 1aee5cc 56d518d 82f0695 7cb76d3 24cb5cc a560010 691ddb0 0b85b74 e8493c1 4601f19 4d95680 af3093c daee099 36f2c33 f09e09f b180ed5 ea6ffff d54be43 7332226 ba50b37 4798a90 380def6 64d9346 5d98fc4 febac0d 8d058a3 abdaebe 83aaea2 70dc0b9 4b56c01 b74d8dd ad00418 d5560dd f816dbd 1efccdc a6e8158 83077dc f72c6a1 01c1b77 a9104b0 d4f45a1 55fb0d6 ba0cf7a 03d3a01 7aaf651 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 |
Started at: 20:50:40 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.32175946950912476 - MAE: 0.4437903503551826 Validation loss : 0.19621961481041378 - MAE: 0.3388283822410594 Epoch: 1 Training loss: 0.21072698533535003 - MAE: 0.35116530289867337 Validation loss : 0.1876767095592287 - MAE: 0.33628140968919695 Epoch: 2 Training loss: 0.1893942326307297 - MAE: 0.3312522516218666 Validation loss : 0.1748287214173211 - MAE: 0.3213797447248068 Epoch: 3 Training loss: 0.17969159841537474 - MAE: 0.3215382240664255 Validation loss : 0.1697687026527193 - MAE: 0.31446870224950263 Epoch: 4 Training loss: 0.17220519423484804 - MAE: 0.3129526796594085 Validation loss : 0.16811848680178323 - MAE: 0.3117262145282266 Epoch: 5 Training loss: 0.17296144247055054 - MAE: 0.31494873624825814 Validation loss : 0.16849000917540657 - MAE: 0.31104514179599985 Epoch: 6 Training loss: 0.16641035348176955 - MAE: 0.30768475494362546 Validation loss : 0.16421516074074638 - MAE: 0.30715220958263423 Epoch: 7 Training loss: 0.16057054400444032 - MAE: 0.30329494898002457 Validation loss : 0.161027698053254 - MAE: 0.304297376785587 Epoch: 8 Training loss: 0.1566170272231102 - MAE: 0.29957014870206655 Validation loss : 0.15982638630602095 - MAE: 0.30397747682822174 Epoch: 9 Training loss: 0.15673983812332154 - MAE: 0.3000540458404174 Validation loss : 0.159184659520785 - MAE: 0.3035235378542429 Epoch: 10 Training loss: 0.15630604147911073 - MAE: 0.29690365842430627 Validation loss : 0.15921704471111298 - MAE: 0.30240467396157955 Epoch: 11 Training loss: 0.15595020622015 - MAE: 0.29754135005638765 Validation loss : 0.15686986181471083 - MAE: 0.30119996351152656 Epoch: 12 Training loss: 0.15274528950452804 - MAE: 0.2944161972508913 Validation loss : 0.15652166141404045 - MAE: 0.3012713923501961 Epoch: 13 Training loss: 0.15205995708703995 - MAE: 0.29424324063629004 Validation loss : 0.1575678288936615 - MAE: 0.3004794443503115 Epoch: 14 Training loss: 0.15087180227041244 - MAE: 0.292504579327589 Validation loss : 0.15574459234873453 - MAE: 0.29932356111019714 Epoch: 15 Training loss: 0.1518820345401764 - MAE: 0.2940904971897683 Validation loss : 0.15581322544150883 - MAE: 0.30048684662514935 Epoch: 16 Training loss: 0.14916340589523316 - MAE: 0.2900369708605816 Validation loss : 0.15502946575482687 - MAE: 0.2986623058669649 Epoch: 17 Training loss: 0.14969733864068985 - MAE: 0.2916869417468108 Validation loss : 0.15631223718325296 - MAE: 0.2994015598567933 Epoch: 18 Training loss: 0.14621972769498826 - MAE: 0.2875086269286061 Validation loss : 0.1557358337773217 - MAE: 0.29931970436403404 Epoch: 19 Training loss: 0.14798570185899734 - MAE: 0.29024787778757843 Validation loss : 0.15424930387073094 - MAE: 0.2984811820958494 Epoch: 20 Training loss: 0.14769238144159316 - MAE: 0.2896109423923894 Validation loss : 0.1536422868569692 - MAE: 0.2987445844262027 Epoch: 21 Training loss: 0.14361368536949157 - MAE: 0.2842206176667335 Validation loss : 0.1533755792511834 - MAE: 0.29642898867488277 Epoch: 22 Training loss: 0.14456430345773696 - MAE: 0.2860957867005398 Validation loss : 0.1542035871081882 - MAE: 0.2971775973100257 Epoch: 23 Training loss: 0.1426533755660057 - MAE: 0.28512021628758083 Validation loss : 0.15245803362793392 - MAE: 0.2961228783995986 Epoch: 24 Training loss: 0.14542177826166153 - MAE: 0.2855342243309425 |