File size: 3,412 Bytes
fa81dc6 71ad858 99d7cf7 814c1b9 7d58879 0ef1a0b 22d7d47 329febd 33ec83f 053a95c 7ba6fe3 972fbe3 70d24a2 29131a1 b6e2355 9818038 ce807b5 123e0a3 faf2825 5d7c2bd 6b6f711 84b1fda d7da145 a021157 8597192 b7b8840 560f7d4 1cc2296 4219e9e 3f5b7b3 a6606c2 6277510 74592b7 61af910 c80ea84 15e9e58 95fcb20 4954377 d985f9d 73b9f73 5a37191 1681cd1 980daff 875efbe 8c594ad 500e07f 932ce20 d659bbb 46276dc 2d88869 b0cb177 46bc9f5 fa08d34 ad109a5 7ee5def |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 |
Started at: 13:38:19 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Started at: 13:39:29 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Started at: 13:41:56 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.4006912616583017 - MAE: 0.505064727315284 Validation loss : 0.309359085559845 - MAE: 0.44078509929430004 Epoch: 1 Training loss: 0.250718348301374 - MAE: 0.3922167896234202 Validation loss : 0.1936800718307495 - MAE: 0.33703642252306854 Epoch: 2 Training loss: 0.21618989912363198 - MAE: 0.3581347894767805 Validation loss : 0.18340231478214264 - MAE: 0.33186761275995064 Epoch: 3 Training loss: 0.19920441508293152 - MAE: 0.3401495898235937 Validation loss : 0.1788225769996643 - MAE: 0.32578763746292233 Epoch: 4 Training loss: 0.1918170922077619 - MAE: 0.33396066058861956 Validation loss : 0.17603523135185242 - MAE: 0.32383651430230814 Epoch: 5 Training loss: 0.18379446978752428 - MAE: 0.3262024097906551 Validation loss : 0.17055903375148773 - MAE: 0.31736346780813546 Epoch: 6 Training loss: 0.18060297461656424 - MAE: 0.3232976047514557 Validation loss : 0.16787597239017488 - MAE: 0.3139610728375986 Epoch: 7 Training loss: 0.17637471740062421 - MAE: 0.3190542525274511 Validation loss : 0.1664418339729309 - MAE: 0.3116923743658102 Epoch: 8 Training loss: 0.17314187494608071 - MAE: 0.3164703095268888 Validation loss : 0.16483855247497559 - MAE: 0.31011736716454275 Epoch: 9 Training loss: 0.17189548795039838 - MAE: 0.31287147634421525 Validation loss : 0.16325874030590057 - MAE: 0.3085789200886582 Epoch: 10 Training loss: 0.1694010473214663 - MAE: 0.31118425956842755 Validation loss : 0.16111792027950286 - MAE: 0.3070655232845323 Epoch: 11 Training loss: 0.16910729843836564 - MAE: 0.30878990252768745 Validation loss : 0.16129685938358307 - MAE: 0.3079809592227729 Epoch: 12 Training loss: 0.16443614203196305 - MAE: 0.3071422337796407 Validation loss : 0.15956343114376068 - MAE: 0.3059515625935648 Epoch: 13 Training loss: 0.16316631436347961 - MAE: 0.30409203692834197 Validation loss : 0.15849626958370208 - MAE: 0.3045730253138591 Epoch: 14 Training loss: 0.1635306592171009 - MAE: 0.3037401073073271 Validation loss : 0.158303701877594 - MAE: 0.30481573363411335 Epoch: 15 Training loss: 0.15958423912525177 - MAE: 0.30241642588486894 |