File size: 3,139 Bytes
797a72b 8b51bb6 882fa36 c15ab4c 1bd332f 742b96a 4465920 6024120 97bf903 8d0db44 6d2e601 1ac98e0 461aac8 e939fb4 8340d3a f0ed3e9 9c88970 2fc70a1 21f0e15 7cd3976 2b34a6b 317869f 1602c7d f8d8c9e 7494fb7 30f96cc 81d92ac 2813fa7 666b9a3 7bbd829 3630a81 e012ec3 a0243b7 4a08212 58e9801 8efa922 5d13d84 f966bf2 d5f9f9d ca60a80 890ba4f ffb81cd bc32ea2 2c85c82 07a7942 78fa631 2906806 63d877a 3e75a18 cbe04e0 ca5aee7 bd2add5 5ece256 0d2e2b8 72c80fd 0682c6d f1463fa 2c3b967 13c59eb db2f4fb b7f62d6 b564e7e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 |
Started at: 10:46:39 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.4224580281972885 - MAE: 0.49800078344951854 Validation loss : 0.18566939760656917 - MAE: 0.32860604518901193 Epoch: 1 Training loss: 0.18062527552247049 - MAE: 0.3249083793427436 Validation loss : 0.17773295325391433 - MAE: 0.32505375557728605 Epoch: 2 Training loss: 0.17239214196801186 - MAE: 0.3161783884766083 Validation loss : 0.1721238993546542 - MAE: 0.31926727109425135 Epoch: 3 Training loss: 0.16808241829276085 - MAE: 0.3116257252131245 Validation loss : 0.16848680245525696 - MAE: 0.31521207737718204 Epoch: 4 Training loss: 0.16556511253118514 - MAE: 0.30872840319635675 Validation loss : 0.16570698601358078 - MAE: 0.31138876910079605 Epoch: 5 Training loss: 0.16236369609832763 - MAE: 0.3063052400904153 Validation loss : 0.16389068610527935 - MAE: 0.30842703017799394 Epoch: 6 Training loss: 0.15821502283215522 - MAE: 0.3002932636657733 Validation loss : 0.16259106204790227 - MAE: 0.3064878174082187 Epoch: 7 Training loss: 0.15610535070300102 - MAE: 0.29970597196416643 Validation loss : 0.16118212920777938 - MAE: 0.30487487035579575 Epoch: 8 Training loss: 0.15775163292884828 - MAE: 0.3012530324315296 Validation loss : 0.16053338874788844 - MAE: 0.30467935002674695 Epoch: 9 Training loss: 0.15300563350319862 - MAE: 0.2966342830465275 Validation loss : 0.1598155660664334 - MAE: 0.3024349783524495 Epoch: 10 Training loss: 0.15308804035186768 - MAE: 0.29689653867839455 Validation loss : 0.15887342218090506 - MAE: 0.30158871157294653 Epoch: 11 Training loss: 0.15205074325203896 - MAE: 0.2968189546038619 Validation loss : 0.15820855384363847 - MAE: 0.3009622242442208 Epoch: 12 Training loss: 0.14986242949962617 - MAE: 0.29354556482882427 Validation loss : 0.15779440937673345 - MAE: 0.3011403656557578 Epoch: 13 Training loss: 0.14915254317224025 - MAE: 0.29414236445220476 Validation loss : 0.1570692671572461 - MAE: 0.29919229552027593 Epoch: 14 Training loss: 0.1507771185040474 - MAE: 0.2942755191435603 Validation loss : 0.15648663569899166 - MAE: 0.2990671428995643 Epoch: 15 Training loss: 0.14960825718939305 - MAE: 0.29282880951675244 Validation loss : 0.15587416028275208 - MAE: 0.29818727256321803 Epoch: 16 Training loss: 0.14761282175779342 - MAE: 0.290322619453174 Validation loss : 0.15559193623416565 - MAE: 0.2986693999701733 Epoch: 17 Training loss: 0.14804843626916409 - MAE: 0.2924950026363286 Validation loss : 0.15508160126559875 - MAE: 0.2970730023460469 Epoch: 18 Training loss: 0.14751567400991916 - MAE: 0.28970414097885494 Validation loss : 0.15513849389903686 - MAE: 0.2973691548251751 Epoch: 19 Training loss: 0.1472933314740658 - MAE: 0.2901643469721373 Validation loss : 0.15472981070770936 - MAE: 0.29793042253066404 |