File size: 3,945 Bytes
b209950 fbbf5ad 45f1b11 8f2a114 5c745af 76f09e8 a43ad13 b6bf34a 4b521a4 c9d0e09 bf1909d 855c551 d6b12f1 39a6869 b135330 07c54a2 68e78a0 1362abc 3170096 cf591ec a8c2817 0882573 702e361 342d348 caddab9 076cb72 15a9813 4fe6668 a3eea35 b32edbf d81668b 9867f74 8ebb769 8f70b79 2e9d3fb f6fe1e4 3367e7e 8816aa7 6edf92f 1383916 dc7fce0 7b508ea 9101793 9a95b4c 523d0cb 1f8f95f f435dbf ece12e4 d23c522 3eed329 53ee702 b14cf47 ab2d562 f4d7a58 6a27fda 7879ee0 c3a2894 ea83bc6 380ea57 568c2af 8c9976d 55b6c55 d953e35 8287bc7 e1e8986 e3d0b6a 34872e8 d3c0392 3f5d36c 61de540 775820c b584db7 22d6325 a7b26f5 1a5e5a1 91c4db0 cafc72a 3353d65 5e291d9 8381f57 91cb371 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 |
Started at: 14:55:50 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.2707270306348801 - MAE: 0.4014548746281838 Validation loss : 0.18827014460283167 - MAE: 0.3350395914193632 Epoch: 1 Training loss: 0.18397963613271714 - MAE: 0.3274942520732981 Validation loss : 0.17896055035731373 - MAE: 0.32625804300542655 Epoch: 2 Training loss: 0.17076618894934653 - MAE: 0.3156013590381467 Validation loss : 0.17423390377970302 - MAE: 0.32323377271249604 Epoch: 3 Training loss: 0.1643083082139492 - MAE: 0.3095641840936019 Validation loss : 0.16728906508754282 - MAE: 0.3149824083534162 Epoch: 4 Training loss: 0.15957146167755126 - MAE: 0.3037565492992936 Validation loss : 0.16289721429347992 - MAE: 0.3089864069106985 Epoch: 5 Training loss: 0.15854436449706555 - MAE: 0.301500893432691 Validation loss : 0.16129758033682318 - MAE: 0.308202317708313 Epoch: 6 Training loss: 0.15570181787014006 - MAE: 0.29918709351469924 Validation loss : 0.16344413292758606 - MAE: 0.30938825983369334 Epoch: 7 Training loss: 0.15292704798281193 - MAE: 0.2961117975066867 Validation loss : 0.16081692103077383 - MAE: 0.3033625257590202 Epoch: 8 Training loss: 0.15216302141547203 - MAE: 0.2959306926556599 Validation loss : 0.15965510378865636 - MAE: 0.30640949969727455 Epoch: 9 Training loss: 0.1456899941712618 - MAE: 0.2888247514317626 Validation loss : 0.15742756864603827 - MAE: 0.3014585494849406 Epoch: 10 Training loss: 0.1467710939049721 - MAE: 0.2907957975079187 Validation loss : 0.15912020206451416 - MAE: 0.30151619716495687 Epoch: 11 Training loss: 0.14659826673567294 - MAE: 0.2893981534314371 Validation loss : 0.1602539621731814 - MAE: 0.3022486893374817 Epoch: 12 Training loss: 0.14781650044023992 - MAE: 0.29275159014985785 Validation loss : 0.15743818177896388 - MAE: 0.3011467784998341 Epoch: 13 Training loss: 0.14709322392940521 - MAE: 0.2905235164202929 Validation loss : 0.15774143706349766 - MAE: 0.30035432342912505 Epoch: 14 Training loss: 0.14297918483614921 - MAE: 0.2870641656789985 Validation loss : 0.16193263846285202 - MAE: 0.3045518551046045 Epoch: 15 Training loss: 0.14410636112093925 - MAE: 0.2889387564652954 Validation loss : 0.15722107492825566 - MAE: 0.299146052737687 Epoch: 16 Training loss: 0.14345928631722926 - MAE: 0.28585450120342 Validation loss : 0.15835655086180744 - MAE: 0.2999799426627903 Epoch: 17 Training loss: 0.14413826659321785 - MAE: 0.2881013153573912 Validation loss : 0.15604389984818065 - MAE: 0.29791508329064814 Epoch: 18 Training loss: 0.14309908427298068 - MAE: 0.2863054251735944 Validation loss : 0.15559008366921367 - MAE: 0.2975231734983981 Epoch: 19 Training loss: 0.14340508081018924 - MAE: 0.2865255988742361 Validation loss : 0.15817041651291006 - MAE: 0.30007397316689965 Epoch: 20 Training loss: 0.14153706684708595 - MAE: 0.28535710740880754 Validation loss : 0.15578039034324534 - MAE: 0.29770187771554596 Epoch: 21 Training loss: 0.1428557775169611 - MAE: 0.285682293954449 Validation loss : 0.16232867451275096 - MAE: 0.3028449341916031 Epoch: 22 Training loss: 0.1438133302330971 - MAE: 0.2875040690451074 Validation loss : 0.15943333638065002 - MAE: 0.30116253984475694 Epoch: 23 Training loss: 0.1429821538925171 - MAE: 0.2868833888762416 Validation loss : 0.16224380889359644 - MAE: 0.3035733609314583 Epoch: 24 Training loss: 0.1443456995487213 - MAE: 0.28863224872068594 Validation loss : 0.16213381728705237 - MAE: 0.30267648828913235 Epoch: 25 Training loss: 0.1441011916846037 - MAE: 0.2873510918998528 Validation loss : 0.1571306345217368 - MAE: 0.29800568097664254 Epoch: 26 |