File size: 2,794 Bytes
415e1d6 0625bb2 7d08e24 557aa89 450b003 94d6b3c 4f21eea 6a81b5a 5ae1f34 0b261b6 64c82db bf2ade6 5218d77 1f36c26 9a66345 7b89b3f f6ba68d 2a24f47 9d1c1f7 7261512 d01a3d5 d560ed1 b67e2d7 25ca10f ed1681d 1afcba2 88b3981 5433b50 922374e 5a1c0c6 3b92628 ba2995f 3099c28 d2a9fed 91709a5 3c755dc 9f5271e 5e9a2ba 070cc17 a1d6b8d f6d2bf1 acfd692 e0c788c 0729b84 3af0a40 012213c eb44138 8a6167a 1ea0a3b 4ccabd1 a11eff6 9c17bda 1a1685c 76639f7 aa27110 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 |
Started at: 14:22:52 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.6967193094583658 - MAE: 0.6750990625330823 Validation loss : 0.5263436496257782 - MAE: 0.613448569191522 Epoch: 1 Training loss: 0.26826212727106535 - MAE: 0.4087051125792896 Validation loss : 0.22296061515808105 - MAE: 0.3555378972565709 Epoch: 2 Training loss: 0.2124968572304799 - MAE: 0.3476390507866737 Validation loss : 0.20461321175098418 - MAE: 0.3532590132614058 Epoch: 3 Training loss: 0.18429063145930952 - MAE: 0.32677963264803717 Validation loss : 0.17646469473838805 - MAE: 0.3203760276514041 Epoch: 4 Training loss: 0.17768853444319505 - MAE: 0.31684026520272845 Validation loss : 0.17874470055103303 - MAE: 0.3257858562177419 Epoch: 5 Training loss: 0.17304767782871538 - MAE: 0.3139755194179104 Validation loss : 0.1718216747045517 - MAE: 0.31812428100953294 Epoch: 6 Training loss: 0.17041566165593955 - MAE: 0.31176170900212385 Validation loss : 0.17110097408294678 - MAE: 0.3175205683975872 Epoch: 7 Training loss: 0.16822290764405176 - MAE: 0.3093251221653364 Validation loss : 0.17010048031806946 - MAE: 0.3165555384963895 Epoch: 8 Training loss: 0.16749977606993455 - MAE: 0.309235729717861 Validation loss : 0.1681269407272339 - MAE: 0.31405053463780375 Epoch: 9 Training loss: 0.1660219534085347 - MAE: 0.30707150650440435 Validation loss : 0.16768858432769776 - MAE: 0.3134797613244496 Epoch: 10 Training loss: 0.16200380256542793 - MAE: 0.3036754525218348 Validation loss : 0.1665838599205017 - MAE: 0.3121892924671111 Epoch: 11 Training loss: 0.16332647777520692 - MAE: 0.3039458888319623 Validation loss : 0.16546964943408965 - MAE: 0.3107545377167428 Epoch: 12 Training loss: 0.16287456223597893 - MAE: 0.30448042997749564 Validation loss : 0.16524331867694855 - MAE: 0.3106786227351389 Epoch: 13 Training loss: 0.15990304029904878 - MAE: 0.30016709856440854 Validation loss : 0.16489878594875335 - MAE: 0.31031377115036685 Epoch: 14 Training loss: 0.15767766993779403 - MAE: 0.2997244083849813 Validation loss : 0.16339748501777648 - MAE: 0.3083185140265562 Epoch: 15 Training loss: 0.15951268031046942 - MAE: 0.29912324704552734 Validation loss : 0.16375492811203002 - MAE: 0.3090144460164269 Epoch: 16 Training loss: 0.15623677808504838 - MAE: 0.2973088492735953 Validation loss : 0.16299248337745667 - MAE: 0.30820454237494843 Epoch: 17 Training loss: 0.1564337038076841 - MAE: 0.29861382568944933 |