File size: 3,163 Bytes
8ecdf1a 85b6090 e3c6bc8 84b320c f0d1e38 25a1eeb 74bffdd cd3af81 55d7b2d 9c9c762 f656e6b 0a5d8f0 0ebdc8c cb178ff 711c898 f5aa86d 9e46af4 00d2388 d75a248 4cf1834 c4920e5 15a02c7 6461657 73d6e8e 700ac7c 1da0fe1 61b2725 0d6baea cc82b0c 0ff25b2 458124b 2a0d03d 390e53e f1e3eb7 bb1acca f366bc5 381e209 5ac883f 41ffb32 54eb5da 1b31ab2 0c83ab5 60c4a64 ad072a4 aafd56e 6ec2e7d 6d1e465 6918809 dcbb528 b82ce9c c3635dd b75a562 7a4a3d5 5c16c4b 65080ab |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 |
Started at: 00:24:23 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.6361208361387253 - MAE: 0.6347873231265873 Validation loss : 0.33865517377853394 - MAE: 0.4740741020180635 Epoch: 1 Training loss: 0.21345499098300935 - MAE: 0.3540912858998663 Validation loss : 0.18473448190424177 - MAE: 0.3282540270842396 Epoch: 2 Training loss: 0.18606460213661194 - MAE: 0.32889828686188877 Validation loss : 0.18045285178555381 - MAE: 0.32713884374037794 Epoch: 3 Training loss: 0.17846946954727172 - MAE: 0.31904566117043975 Validation loss : 0.17773988511827257 - MAE: 0.32482577310153565 Epoch: 4 Training loss: 0.17139309346675874 - MAE: 0.31441827165194275 Validation loss : 0.17418616182274288 - MAE: 0.32107063596093277 Epoch: 5 Training loss: 0.17030848145484925 - MAE: 0.31198532199217943 Validation loss : 0.17218486799134147 - MAE: 0.3190577970415032 Epoch: 6 Training loss: 0.16878579378128053 - MAE: 0.31078672095469284 Validation loss : 0.17055005000697243 - MAE: 0.3174862826881529 Epoch: 7 Training loss: 0.1679440525174141 - MAE: 0.31018192416787727 Validation loss : 0.16925331950187683 - MAE: 0.31617688409581507 Epoch: 8 Training loss: 0.16459781765937806 - MAE: 0.3076978685246546 Validation loss : 0.16737455626328787 - MAE: 0.313686903818875 Epoch: 9 Training loss: 0.1616947764158249 - MAE: 0.3042913169405816 Validation loss : 0.16636485523647732 - MAE: 0.3125500280647287 Started at: 01:47:29 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.6361208361387253 - MAE: 0.6347873231265873 Validation loss : 0.33865517377853394 - MAE: 0.4740741020180635 Epoch: 1 Training loss: 0.21345499098300935 - MAE: 0.3540912858998663 Validation loss : 0.18473448190424177 - MAE: 0.3282540270842396 Epoch: 2 Training loss: 0.18606460213661194 - MAE: 0.32889828686188877 Validation loss : 0.18045285178555381 - MAE: 0.32713884374037794 Epoch: 3 Training loss: 0.17846946954727172 - MAE: 0.31904566117043975 Validation loss : 0.17773988511827257 - MAE: 0.32482577310153565 Epoch: 4 Training loss: 0.17139309346675874 - MAE: 0.31441827165194275 Validation loss : 0.17418616182274288 - MAE: 0.32107063596093277 Epoch: 5 Training loss: 0.17030848145484925 - MAE: 0.31198532199217943 Validation loss : 0.17218486799134147 - MAE: 0.3190577970415032 Epoch: 6 Training loss: 0.16878579378128053 - MAE: 0.31078672095469284 Validation loss : 0.17055005000697243 - MAE: 0.3174862826881529 |