File size: 6,863 Bytes
4001542 f70081d bfe4889 57c6f06 420046c 5e0a3da e8ebc03 c4038c0 7e2b43b 6bc8b46 7184377 c7792bd e70cc41 7d1a920 a947808 e8e06a9 18308cf d083566 4b2416e e7ba8a8 2a1d7ec a4b8a87 08546e8 c040908 ccb9371 dd2283a 2a94064 8d19120 d0fd6fa d971075 2f11030 00ffef7 7589303 9f7427d 215ca20 472f0c7 8e325e8 3154351 86b7e6d b7e6a40 15b9ad9 ac7d759 8d91f1c e249531 6875df7 34108ab 0325135 2587462 08b7760 e091124 e09f6a5 f0958b0 df09ca9 aad6266 404b115 0b31c62 907ef70 f1817ea dbcf0fd 19b5dda 8671bd3 c85fb05 4489afc 2ad0729 874f548 452dfb0 337893d 2476c4c e9aa6f8 27a195a 41958c8 862d359 5aef33c 0d10d1f a96eacb f9a157f 97d8f0d e522da0 20c32b5 ac6fb45 c566ccf e844f74 f61e0a0 1657b87 a5fbc8d ef27d64 72db1b8 87233f9 dd350e2 2ef68ad ee18590 18e3705 b124356 b24e3bf 3af3e90 da19f8b 1506dfb 678cf05 55327a6 ef20459 9a091bb 2e13c85 054204f 2fe798a a06a528 3850eb3 3fa42ab ff27786 1f2e6f6 77e43c9 bbd8ca8 416fd8c 6c0d046 9aac2df dab6b41 313f18f f1124ca 680d38d f3ac684 b3ffc7f 828762d 0485ce5 0537688 0137ad4 884f1c6 94c9f76 18b2dee 4516998 70eb2e0 67f31f8 4e82dde 5595ae3 81b0a83 152a3e7 21fb2a3 1ca7565 213c584 83fda25 25a1929 7c75cc0 b10d97b 6999f11 829f7f7 6efe931 99b128e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 |
Started at: 14:58:05 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '44815f7e109b53547cccdf3c6847f4c28b989816'}, {}) Epoch: 0 Training loss: 0.25046255227726727 - MSE: 0.36727700175627953 Validation loss : 0.17648151460470576 - MSE: 0.32341695900998885 Epoch: 1 Training loss: 0.17091115193469875 - MSE: 0.31430697735027086 Validation loss : 0.16562944582917474 - MSE: 0.31155962633828027 Epoch: 2 Training loss: 0.16704481191517132 - MSE: 0.3114155539092336 Validation loss : 0.1604571777085463 - MSE: 0.30340937581848676 Epoch: 3 Training loss: 0.16466754909396777 - MSE: 0.30768393210761735 Validation loss : 0.1632813240091006 - MSE: 0.3109833563942495 Epoch: 4 Training loss: 0.16455070379421796 - MSE: 0.3093531171941453 Validation loss : 0.16564318244204376 - MSE: 0.30289557314252813 Epoch: 5 Training loss: 0.16568223224376058 - MSE: 0.30846559528561224 Validation loss : 0.1605968801022479 - MSE: 0.2988926674776124 Epoch: 6 Training loss: 0.15717696056223762 - MSE: 0.3022554811747332 Validation loss : 0.16166257779255058 - MSE: 0.29849396869512723 Epoch: 7 Training loss: 0.15822318865714338 - MSE: 0.30195677281494104 Validation loss : 0.1757645166733048 - MSE: 0.3080704490543514 Epoch: 8 Training loss: 0.15552518461939646 - MSE: 0.2998732285398632 Validation loss : 0.1911594748045459 - MSE: 0.32289788546992626 Epoch: 9 Training loss: 0.15594329739812063 - MSE: 0.3009645091966726 Validation loss : 0.19870018710692725 - MSE: 0.33138930551919304 Epoch: 10 Training loss: 0.1547534751521452 - MSE: 0.2992649989557918 Validation loss : 0.19730406333551262 - MSE: 0.3297122567265123 Epoch: 11 Training loss: 0.15529195780908395 - MSE: 0.29975488127620853 Validation loss : 0.18655535093311107 - MSE: 0.3217206286243775 Epoch: 12 Training loss: 0.15309068918001228 - MSE: 0.2969653755509521 Validation loss : 0.17314820917266788 - MSE: 0.30849219803901623 Epoch: 13 Training loss: 0.15322715030784534 - MSE: 0.29618723127423563 Validation loss : 0.1593460923794544 - MSE: 0.29643875794480956 Epoch: 14 Training loss: 0.1519091561209732 - MSE: 0.2967038231193114 Validation loss : 0.15947996328274408 - MSE: 0.2962177626613215 Epoch: 15 Training loss: 0.1513304633234963 - MSE: 0.2948141613710868 Validation loss : 0.17381140318783847 - MSE: 0.30743163511355315 Epoch: 16 Training loss: 0.15304585124559814 - MSE: 0.2953887584017137 Validation loss : 0.15291882255537945 - MSE: 0.2937530252336034 Epoch: 17 Training loss: 0.15207511008844762 - MSE: 0.29601101879974745 Validation loss : 0.15711521870936407 - MSE: 0.2956258048557993 Epoch: 18 Training loss: 0.1522078831032448 - MSE: 0.2947807228620845 Validation loss : 0.1593479738768303 - MSE: 0.2975361759408073 Epoch: 19 Training loss: 0.15030200151622597 - MSE: 0.29364889133341376 Validation loss : 0.15571271227390476 - MSE: 0.2947020680712253 Epoch: 20 Training loss: 0.14997340040990545 - MSE: 0.29321906843107093 Validation loss : 0.16053046500592522 - MSE: 0.2972562162079979 Epoch: 21 Training loss: 0.1503814266153096 - MSE: 0.29498264967501453 Validation loss : 0.18625259749365575 - MSE: 0.3214374853139268 Epoch: 22 Training loss: 0.15071523613130985 - MSE: 0.2931879129534037 Validation loss : 0.15868466793361938 - MSE: 0.29779115185817956 Epoch: 23 Training loss: 0.14925291788230088 - MSE: 0.29253552849679587 Validation loss : 0.17538098990917206 - MSE: 0.30947426368924724 Epoch: 24 Training loss: 0.15000015682526652 - MSE: 0.2927369972856095 Validation loss : 0.1700212608909968 - MSE: 0.3071491417052287 Epoch: 25 Training loss: 0.15024784067409291 - MSE: 0.29311721077485986 Validation loss : 0.19164935623606047 - MSE: 0.32203448128239803 Epoch: 26 Training loss: 0.15137434639225755 - MSE: 0.29379136046127924 Validation loss : 0.1671447205272588 - MSE: 0.30391976326317943 Epoch: 27 Training loss: 0.15017312853998943 - MSE: 0.29455603880812276 Validation loss : 0.18036871166391807 - MSE: 0.3125465653289461 Epoch: 28 Training loss: 0.14778636641762583 - MSE: 0.2905298917289279 Validation loss : 0.16160754949757547 - MSE: 0.2979193112805207 Epoch: 29 Training loss: 0.1477772275241196 - MSE: 0.2912173747535554 Validation loss : 0.17026665750326533 - MSE: 0.30567947675889745 Epoch: 30 Training loss: 0.14744956316757324 - MSE: 0.291312328070814 Validation loss : 0.16202468379880441 - MSE: 0.29700271247754034 Epoch: 31 Training loss: 0.1484179409778663 - MSE: 0.2921315750433443 Validation loss : 0.16338278116150337 - MSE: 0.2987860072168981 Epoch: 32 Training loss: 0.14738556740777142 - MSE: 0.29104356719237584 Validation loss : 0.15635579254365328 - MSE: 0.29443009978003276 Epoch: 33 Training loss: 0.148553161413839 - MSE: 0.29089201478673565 Validation loss : 0.15951512471744508 - MSE: 0.2966586790896474 Epoch: 34 Training loss: 0.15104458401650947 - MSE: 0.2934267254216827 Validation loss : 0.17196247625080022 - MSE: 0.3055353329027785 Epoch: 35 Training loss: 0.1503695050623211 - MSE: 0.29389694550707784 Validation loss : 0.16170481603705522 - MSE: 0.29835679015904104 Epoch: 36 Training loss: 0.14925730479928442 - MSE: 0.2924833646517312 Validation loss : 0.17856141236243825 - MSE: 0.31225271412806316 Epoch: 37 Training loss: 0.15118206122199895 - MSE: 0.29395904130187084 Validation loss : 0.17895790042750764 - MSE: 0.3121755853923419 Epoch: 38 Training loss: 0.1497652710505246 - MSE: 0.2912881449244226 Validation loss : 0.15206407851567774 - MSE: 0.2909835362196566 Epoch: 39 Training loss: 0.14738748415472544 - MSE: 0.29036707520169275 Validation loss : 0.16739020203099106 - MSE: 0.30124313077212783 Epoch: 40 Training loss: 0.14675953452886664 - MSE: 0.2900500768121404 Validation loss : 0.17623373211333246 - MSE: 0.30992423744210557 Epoch: 41 Training loss: 0.14824477195512825 - MSE: 0.29090776402279417 Validation loss : 0.16109058317361455 - MSE: 0.2974189646320188 Epoch: 42 Training loss: 0.14692961365967838 - MSE: 0.2904768427610488 Validation loss : 0.1707474204401175 - MSE: 0.30493211675533977 Epoch: 43 Training loss: 0.14761571415941122 - MSE: 0.2907837329648465 Validation loss : 0.1536927172970591 - MSE: 0.2910231603894442 Epoch: 44 Training loss: 0.15164254637917285 - MSE: 0.2942169462116838 Validation loss : 0.15892887950846643 - MSE: 0.2960443273887821 Epoch: 45 Training loss: 0.14786733518624065 - MSE: 0.2917048977763665 Validation loss : 0.1669148841139042 - MSE: 0.30125252101874894 Epoch: 46 Training loss: 0.14599600882472727 - MSE: 0.286971878648382 Validation loss : 0.15723810294135052 - MSE: 0.29333694749044353 Epoch: 47 Training loss: 0.14816145228250377 - MSE: 0.28941392496829527 |