File size: 3,136 Bytes
fd32488 cbe769a fa18bdc c117686 540d12d 9d7e8c4 8b0eb84 3f32498 5b21d58 f0f6ff0 9aa084e 5eaa3b7 b9da547 8dcf2cc b93ce6f 3074baa cedee85 de1811a 08f3e80 5076d29 59df5ad 8d78666 1d1169d c1ce569 8c4c3e2 da01cef eeaf256 b56f9c5 5099657 9c0ee27 1563e61 7d370b6 ff50cb5 2b54ead 4eff91c b828b77 e984fd8 b6224da 29e7bf2 e44dd3d 7e92b28 ceab61a 36fbf1d 623f746 b58c7ab 4e88a10 55234f3 2465de2 23a5e05 567fb27 c2087b8 70e5cf1 1436710 7f01a79 b9f4314 2d7ed60 05ede4a f2e36e5 8adeed5 c8060bd 1832de7 cfca31b 341bc18 52a6ee1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 |
Started at: 13:10:36 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'afb829e3d0b861bd5f8cda6522b32ca0b097d7eb'}, {}) Epoch: 0 Training loss: 2.7300986201327464 - MSE: 1.5973443396223515 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 1 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 2 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 3 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 4 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 5 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 6 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 7 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 8 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 9 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 10 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 11 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 12 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 13 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 14 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 15 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 16 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 17 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 18 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 19 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 Validation loss : 2.7675703980705957 - MSE: 1.609589820176938 Epoch: 20 Training loss: 2.7547256934461255 - MSE: 1.6089351524297428 |