File size: 5,952 Bytes
e605cc9 6e5ab3b 1a28e55 f74d1a8 9cc646b 9beb45d 3a45c67 02fba59 3a26325 9028c94 2c04bc4 262ec82 49fb6bc ded09a7 9b47d8b 309ee23 39210e5 9f92c33 35b0752 a00f471 6eabebe b878ae8 b05458b 2875918 0be6538 14e8708 13a9b4e e0536a5 2c8baeb 2c21340 745663d af1790e 78b9ee7 2500b3c d1c9496 deb0663 e38113f c8a10d6 9126c96 9a28174 7089aec 1dcbc02 d79f7f0 f89fbdf ffdccd0 0fa7855 6c3896b 7575574 e529ca1 cf50e90 7a34436 8a78288 737fa55 07d95b2 42b5109 63ecdb9 55747a0 6df23a1 1e82104 ef65d4d dd4b025 82342ad ed06f7d 7e64e24 ccee6b4 db5a4a6 9770972 d958b1a f1ca998 d72fe17 7c741ac 03b51d3 08365b5 bf146c3 c6c7a2d 66a98ea 559b6ee 8f7525d 4f74a77 b5ce660 7e562cf e9be63d beaa65a c79e3c7 e9076e2 c674b53 cec9885 6f78234 4ed14ea 19d6bb9 c869185 e9f9629 18ba125 6278499 ea8cad4 1798613 5f4bf08 c34dc19 af7e869 5d5762c 3ac226e 302bb9e 0ca3f82 04d2e93 e5d6c3c 1cce2ed a8d7acf 495f025 e1f7ce2 25192bc 2f6e557 4ca0d7c e160009 385e574 f010181 03e47a8 e79389b a274968 13918e6 b33b497 5e83a81 34bb8da c0a539e ffe8c7b 07e009f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 |
Started at: 11:39:57 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.2487574193040429 - MSE: 0.36417208558066866 Validation loss : 0.17720953657320052 - MSE: 0.324517725567173 Epoch: 1 Training loss: 0.17969225880306988 - MSE: 0.32316235001284466 Validation loss : 0.18277007442983714 - MSE: 0.32265740191792125 Epoch: 2 Training loss: 0.18366406324702472 - MSE: 0.32790626373068626 Validation loss : 0.176965011904637 - MSE: 0.32382223638952334 Epoch: 3 Training loss: 0.18670494069680948 - MSE: 0.3299349662450387 Validation loss : 0.17702526836232704 - MSE: 0.3240165099565265 Epoch: 4 Training loss: 0.20028187222967897 - MSE: 0.3438661170711591 Validation loss : 0.176956477603226 - MSE: 0.3220238222141861 Epoch: 5 Training loss: 0.18425696651418197 - MSE: 0.3284598663449472 Validation loss : 0.17923579692389024 - MSE: 0.3213620647532961 Epoch: 6 Training loss: 0.18291579756110452 - MSE: 0.32706088243869585 Validation loss : 0.18002789682059578 - MSE: 0.3215154469143273 Epoch: 7 Training loss: 0.18255479792653 - MSE: 0.3266934425025146 Validation loss : 0.18080746585672552 - MSE: 0.3217779787463738 Epoch: 8 Training loss: 0.18238363441476968 - MSE: 0.326510964545854 Validation loss : 0.18103667574398447 - MSE: 0.3218730203910852 Epoch: 9 Training loss: 0.1823012850951725 - MSE: 0.32642306303904944 Validation loss : 0.18114753259402333 - MSE: 0.3219193949429441 Epoch: 10 Training loss: 0.18226306843878654 - MSE: 0.32638221163239944 Validation loss : 0.18120037685289528 - MSE: 0.3219412936734975 Epoch: 11 Training loss: 0.1822457022322011 - MSE: 0.3263632995875595 Validation loss : 0.181226331401955 - MSE: 0.3219524295489856 Epoch: 12 Training loss: 0.18223814496080282 - MSE: 0.326354779336264 Validation loss : 0.18123742317159972 - MSE: 0.3219572314415399 Epoch: 13 Training loss: 0.18223547419240027 - MSE: 0.3263515567617185 Validation loss : 0.18124232102524152 - MSE: 0.3219593503282257 Epoch: 14 Training loss: 0.18223450450122658 - MSE: 0.3263499022376213 Validation loss : 0.1812431158667261 - MSE: 0.3219596943033371 Epoch: 15 Training loss: 0.18223553160347308 - MSE: 0.3263508579780551 Validation loss : 0.18124122296770415 - MSE: 0.3219588756425721 Epoch: 16 Training loss: 0.18223705256197054 - MSE: 0.3263523368442598 Validation loss : 0.18124055501186487 - MSE: 0.32195858670347854 Epoch: 17 Training loss: 0.18223793137118902 - MSE: 0.3263530484795528 Validation loss : 0.18123786794868382 - MSE: 0.32195742406760225 Epoch: 18 Training loss: 0.1822392940143038 - MSE: 0.3263543280507233 Validation loss : 0.18123836092876666 - MSE: 0.3219576373321713 Epoch: 19 Training loss: 0.1822397845234665 - MSE: 0.3263546927109733 Validation loss : 0.181238055229187 - MSE: 0.3219575066216289 Epoch: 20 Training loss: 0.18223992795659805 - MSE: 0.3263547543533027 Validation loss : 0.1812378361143849 - MSE: 0.3219574103085978 Epoch: 21 Training loss: 0.1822402750326292 - MSE: 0.3263550699385248 Validation loss : 0.1812385039573366 - MSE: 0.3219576992476913 Epoch: 22 Training loss: 0.18223916792173675 - MSE: 0.32635392572647925 Validation loss : 0.181226793337952 - MSE: 0.3219526290545502 Epoch: 23 Training loss: 0.18224155640012116 - MSE: 0.32635636567848586 Validation loss : 0.18123600643241045 - MSE: 0.3219566191658417 Epoch: 24 Training loss: 0.18224167383186102 - MSE: 0.32635633417493376 Validation loss : 0.18123632443673682 - MSE: 0.32195675675588625 Epoch: 25 Training loss: 0.18224162090248264 - MSE: 0.3263562383563682 Validation loss : 0.181236690079624 - MSE: 0.32195691498443746 Epoch: 26 Training loss: 0.18224160620933258 - MSE: 0.3263561929586502 Validation loss : 0.18123645154815732 - MSE: 0.32195681179190405 Epoch: 27 Training loss: 0.1822416710520759 - MSE: 0.3263562471136733 Validation loss : 0.18123681831992033 - MSE: 0.32195697002045526 Epoch: 28 Training loss: 0.18224165091281616 - MSE: 0.32635619502476765 Validation loss : 0.18123680296720882 - MSE: 0.321956963140953 Epoch: 29 Training loss: 0.18224163323186982 - MSE: 0.3263561882008935 Validation loss : 0.18123737508148857 - MSE: 0.3219572108030332 Epoch: 30 Training loss: 0.1822399182746253 - MSE: 0.3263545064950311 Validation loss : 0.18123632443673682 - MSE: 0.32195675675588625 Epoch: 31 Training loss: 0.18224117341380433 - MSE: 0.32635554491807284 Validation loss : 0.1812329568871946 - MSE: 0.3219552983014142 Epoch: 32 Training loss: 0.1822413879300132 - MSE: 0.3263558743974608 Validation loss : 0.1812364367598837 - MSE: 0.3219568049124018 Epoch: 33 Training loss: 0.18224198902545846 - MSE: 0.32635650742551514 Validation loss : 0.1812365473897168 - MSE: 0.3219568530689174 Epoch: 34 Training loss: 0.18224186066367903 - MSE: 0.3263563794399653 Validation loss : 0.18123678772738486 - MSE: 0.3219569562614508 Epoch: 35 Training loss: 0.18224179525363263 - MSE: 0.3263562783328973 Validation loss : 0.18123700921282623 - MSE: 0.321957052574482 Epoch: 36 Training loss: 0.18224171250302174 - MSE: 0.32635620431281853 Validation loss : 0.18123708778258527 - MSE: 0.3219570869719931 Epoch: 37 Training loss: 0.18223311156412672 - MSE: 0.3263353195428408 Validation loss : 0.18125040219588714 - MSE: 0.32196284511535694 Epoch: 38 Training loss: 0.1822478898285609 - MSE: 0.32636291019076585 Validation loss : 0.18123621369401613 - MSE: 0.32195670859937064 Epoch: 39 Training loss: 0.18224211545840738 - MSE: 0.32635662244570424 Validation loss : 0.18123670701276173 - MSE: 0.32195692186393965 Epoch: 40 Training loss: 0.18224123697066066 - MSE: 0.32635547624336203 Validation loss : 0.181234287944707 - MSE: 0.32195587617960125 |