File size: 2,859 Bytes
a54743f da02fe8 64c7ac4 ef2b635 0bab893 743bf2f 6e0c0fe 4e0ce49 07a14de b58387d cea0a2c 8be760f 6987d12 e76068d f8f2c9f 73ef6e7 db96ccb 51b5543 5f313bc fb33851 c47912b 33fd113 5d2cc42 6a2936c e4284ca b7b2d5b 0128a2d 787d203 25cb336 f03d817 d6c4c1c 2626bc6 500c492 f8d559c 23aa4c6 a45ac6b 70335cf dd9125d c2df8a7 636aeee 8900ea4 dd55b82 aaa6ec3 4a4d811 0edfeff 4d79fd7 1392ae5 18ae2a0 5c3dc1f 726f346 0b4ddac 7586150 8804e7a 6a07917 25f5dfa 646bf85 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 |
Started at: 16:30:49 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.3458641168475151 - MAE: 0.4470721267008891 Validation loss : 0.1788818021030987 - MAE: 0.32197414947064035 Epoch: 1 Training loss: 0.17300812512636185 - MAE: 0.3175820470114712 Validation loss : 0.16896727418198304 - MAE: 0.31332903775222964 Epoch: 2 Training loss: 0.1648126335442066 - MAE: 0.308156299760969 Validation loss : 0.16454754506840424 - MAE: 0.3087028804974932 Epoch: 3 Training loss: 0.1608002531528473 - MAE: 0.3038113721048132 Validation loss : 0.1621274115408168 - MAE: 0.30596198775094147 Epoch: 4 Training loss: 0.15862647637724878 - MAE: 0.3014963360894156 Validation loss : 0.16059190341654947 - MAE: 0.30361069127564794 Epoch: 5 Training loss: 0.15564483717083932 - MAE: 0.29912276695755136 Validation loss : 0.15960544479243896 - MAE: 0.3032567813495726 Epoch: 6 Training loss: 0.15132688611745834 - MAE: 0.29362002153133276 Validation loss : 0.15873736844343297 - MAE: 0.3026340556567635 Epoch: 7 Training loss: 0.15053230985999108 - MAE: 0.2936411939181263 Validation loss : 0.15721423354218989 - MAE: 0.3007017446245457 Epoch: 8 Training loss: 0.15200649231672286 - MAE: 0.29530517620818714 Validation loss : 0.1567032946383252 - MAE: 0.29992379573698985 Epoch: 9 Training loss: 0.14823424234986304 - MAE: 0.29201854833537316 Validation loss : 0.15661324560642242 - MAE: 0.3000581370785886 Epoch: 10 Training loss: 0.1488597995787859 - MAE: 0.29202925516443873 Validation loss : 0.15594959259033203 - MAE: 0.29955980657011466 Epoch: 11 Training loss: 0.1478568846732378 - MAE: 0.2920750352468466 Validation loss : 0.15519277882926605 - MAE: 0.29875912341148037 Epoch: 12 Training loss: 0.14643058851361274 - MAE: 0.2900920148543982 Validation loss : 0.1542835871086401 - MAE: 0.29761316691536144 Epoch: 13 Training loss: 0.145680733025074 - MAE: 0.29051563041966255 Validation loss : 0.1540197721299003 - MAE: 0.2970695665813526 Epoch: 14 Training loss: 0.14762464702129363 - MAE: 0.2914787752307884 Validation loss : 0.15417058765888214 - MAE: 0.2980900478295141 Epoch: 15 Training loss: 0.14693237617611885 - MAE: 0.2902377114756851 Validation loss : 0.15259952054304235 - MAE: 0.2949612107951529 Epoch: 16 Training loss: 0.14457482241094113 - MAE: 0.2876271989725693 Validation loss : 0.15327997593318715 - MAE: 0.2970857583493576 Epoch: 17 Training loss: 0.14548898428678514 - MAE: 0.28977346291513006 Validation loss : 0.15336684838813894 - MAE: 0.29746389708445636 |