File size: 2,811 Bytes
521959b c5c12d0 c26b107 f8e2c99 cf0f1d7 7ba0ab5 117d441 87139fd 51f7f0a 404af39 98d2b00 e30b1ec ab10f85 c50c9c1 c596e21 b4bd98a d0fa6c9 586cfd5 0323c9c f0d45c2 19de30a 8ca0cf8 35f0f6c a231784 fce29ad cb9a13b 2713003 fef7c17 c3854d5 8ead488 a35b379 5c54f29 cf9b144 5a70457 8b0c28d f712444 0aa527d f00ed0b 2568793 011b9e5 41e0812 44a25c4 3ec0680 7ad050b e25329d f980ce7 2c0418c ad931c4 e676cec a8396d8 17d00bf 943bcd6 1752b11 d228e66 5131fb0 b02a646 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 |
Started at: 16:58:38 norbert, 1e-06, 128 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.7940866667032241 - MAE: 0.7335820706168809 Validation loss : 0.2009572711061029 - MAE: 0.347151339356108 Epoch: 1 Training loss: 0.19440142974257468 - MAE: 0.33696885945893024 Validation loss : 0.19513289893374725 - MAE: 0.34191775396532464 Epoch: 2 Training loss: 0.1880853456258774 - MAE: 0.32884479290870544 Validation loss : 0.1912411977263058 - MAE: 0.3377114550114297 Epoch: 3 Training loss: 0.18214607641100883 - MAE: 0.32429566931252046 Validation loss : 0.18406631753725164 - MAE: 0.33177290516437835 Epoch: 4 Training loss: 0.17814580470323563 - MAE: 0.3207397197424783 Validation loss : 0.1817777980776394 - MAE: 0.3288081865926719 Epoch: 5 Training loss: 0.17603084936738014 - MAE: 0.3188514731069929 Validation loss : 0.17798463649609508 - MAE: 0.3250048632803593 Epoch: 6 Training loss: 0.17104241520166397 - MAE: 0.3136062115504468 Validation loss : 0.1733261583482518 - MAE: 0.31932644642810926 Epoch: 7 Training loss: 0.16838639885187148 - MAE: 0.3123990239326607 Validation loss : 0.17248562416609595 - MAE: 0.31865962855570096 Epoch: 8 Training loss: 0.1654125490784645 - MAE: 0.3085900949825916 Validation loss : 0.16888807626331553 - MAE: 0.3153836119416533 Epoch: 9 Training loss: 0.1645794503390789 - MAE: 0.30737329828383714 Validation loss : 0.16706455805722406 - MAE: 0.3121367546740557 Epoch: 10 Training loss: 0.1616331295669079 - MAE: 0.30512118468554295 Validation loss : 0.16470523529192982 - MAE: 0.3095945078707051 Epoch: 11 Training loss: 0.16030726864933967 - MAE: 0.3040617280659885 Validation loss : 0.16272581894608104 - MAE: 0.3069817555674387 Epoch: 12 Training loss: 0.1589731414616108 - MAE: 0.3032846674015245 Validation loss : 0.16099156351650462 - MAE: 0.3047476777439796 Epoch: 13 Training loss: 0.15562866285443305 - MAE: 0.29989675715351916 Validation loss : 0.15904783939614014 - MAE: 0.3019063086532548 Epoch: 14 Training loss: 0.15696120083332063 - MAE: 0.30097456977168896 Validation loss : 0.15785902738571167 - MAE: 0.30075530215047686 Epoch: 15 Training loss: 0.151818914860487 - MAE: 0.2954353682482574 Validation loss : 0.1551657340982381 - MAE: 0.29802607762302763 Epoch: 16 Training loss: 0.1526854731142521 - MAE: 0.29552840146002957 Validation loss : 0.1538231990793172 - MAE: 0.29607686650726545 Epoch: 17 Training loss: 0.14886232271790503 - MAE: 0.2912935844145451 |