File size: 11,849 Bytes
ee53a4a 50f1745 8a33155 79fb205 22be4a5 3f1ff0e bbed904 14c983c bb23377 0ea8477 8597b5f a529489 8f5752b d444cf9 2dfa794 782f4ba 2665515 8c5051b 84a2460 36464be b4eab8a b4c7480 bd5a96e c248496 a3f3665 3572268 8f95e8a 0e3f82d bd48add 1da9fd2 3759f70 e09bf1b ec4c3b8 b891788 9c846de c36a8be 3f890ba aaf9dcb 9442814 64e656f 47314ea b48fff4 c931c86 ec8e58d 28782de d61067e 2cde297 dce64ed 1a44567 a894d88 f8b0b44 63524f2 3dc095f 915dd82 a5a47b9 15b72bc 5ba086a 6a406c0 fd9bd7c f2205a0 712111f 21fd887 992268b 6fd095f 71d7764 c0f3990 73015a7 ad81682 5a521e7 daf7029 3060a2d 4be9431 8aa7919 c895440 9a9cef0 f236bc8 66cca27 7e49157 a294d56 febe9a0 83c1bd3 9469a62 90a233b 539cc86 777a580 52b7ba5 f7119ee 9186c3b e2d4e0d c7d4106 edc5e85 7e70a20 720f4a0 61139b9 86ab2b7 9401e87 fd467de ee43001 1078eb3 6a7636d 5bac8c4 5d0721b a1c0044 35c229f 4222c2e 7d91c74 01aac9c 9c3e95d 050c19d 3d484a1 229d040 53d9b60 fde77a5 af0d9dd 999f338 9581137 1591024 f2c7f34 03b454e cefb7da 3b85915 14be623 c5851b0 65d6066 978a32c 9f52222 f5a0354 05e9788 4f6505f e305e3a a4f6d7a bdd2cfd bbef9aa 1054174 63b42b9 6f00408 055694c 90842d0 5c96627 c606fb2 e2dfc21 db90dbb 77c5e49 5092051 973f642 c1bd784 dc3afdd 4ca984a 47b850e b7433ab 2d21015 47f85d1 cb32d0b 1c54cf7 da0b33f b0afa4f 9d74bc6 7ee2037 3998ccb 70afe20 f4caccd 25de5bf 868fb88 18839b0 1f375f0 de69b21 4fc87fa d76951b 73567db 6b22e36 6563824 2821164 47066b2 9e6661c 6969ff2 bb021a5 8a82f84 c29e1ae 5b4d801 e77f0b0 3308056 167c9b8 d4d0dca af54614 bf3bfdc 0c542e9 45ea433 fec4d59 812e1ce ea5c962 4650730 584fed8 f959966 20b3d42 a522578 98c26bc 4eebd73 9df3bbb dcf18b1 e45d4b7 adc4708 3d49f77 afc4293 44b4642 4468692 b117e81 c164e20 c6e9dfa 3c962b0 7e2025c 273aa2c cba3233 7413f0f e18429c f770113 e3463ad 5bec24e 035092f 62d72e7 eb1e452 d318627 0b780fe a68276d 78238ba 08cee8f 6c9e183 ed5dc41 5b21de6 672c326 f5fda09 081e18e b2cf9c1 fecae89 0cf1633 4ccd826 2f62a9d 2c58b7f 932c441 9a4ecec 68e16b9 803efb0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 |
Started at: 17:01:49 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Started at: 17:03:05 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Started at: 17:03:18 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 1.5561567282676696 - MAE: 1.1770511100698875 Validation loss : 0.939958651860555 - MAE: 0.8876047142856601 Epoch: 1 Training loss: 0.6363586246967315 - MAE: 0.6928943697649724 Validation loss : 0.33423560195498997 - MAE: 0.45375851608477785 Epoch: 2 Training loss: 0.2759279823303223 - MAE: 0.4052092643476477 Validation loss : 0.20369169281588662 - MAE: 0.3433357920074413 Epoch: 3 Training loss: 0.20272068321704864 - MAE: 0.3399617345503301 Validation loss : 0.19545602964030373 - MAE: 0.3416856898288076 Epoch: 4 Training loss: 0.1912100625038147 - MAE: 0.3326939730215834 Validation loss : 0.19473741783036125 - MAE: 0.3417824430892937 Epoch: 5 Training loss: 0.19083771824836732 - MAE: 0.33137282498844967 Validation loss : 0.19215407967567444 - MAE: 0.33912431491105655 Epoch: 6 Training loss: 0.18923945665359498 - MAE: 0.33146465315973606 Validation loss : 0.18989400565624237 - MAE: 0.33674625605103725 Epoch: 7 Training loss: 0.18950583219528197 - MAE: 0.33087934479225867 Validation loss : 0.18789740900198618 - MAE: 0.33464578460452915 Epoch: 8 Training loss: 0.1856936401128769 - MAE: 0.3275132613322806 Validation loss : 0.1864797506067488 - MAE: 0.33324104491047596 Epoch: 9 Training loss: 0.1836453753709793 - MAE: 0.3255223596708951 Validation loss : 0.18501530918810102 - MAE: 0.3316937400385192 Epoch: 10 Training loss: 0.18246697008609772 - MAE: 0.32322550938186817 Validation loss : 0.18376492957274118 - MAE: 0.33042730021623856 Epoch: 11 Training loss: 0.1811656492948532 - MAE: 0.32419844121595726 Validation loss : 0.18262643449836308 - MAE: 0.32929917485750665 Epoch: 12 Training loss: 0.18094802320003509 - MAE: 0.32277421430430436 Validation loss : 0.18164582550525665 - MAE: 0.3283432280968596 Epoch: 13 Training loss: 0.1792392772436142 - MAE: 0.3209453182214294 Validation loss : 0.18065459695127276 - MAE: 0.3273255372779676 Epoch: 14 Training loss: 0.17940414905548097 - MAE: 0.3216579290630419 Validation loss : 0.17975443104902902 - MAE: 0.3264252069830345 Epoch: 15 Training loss: 0.17663683056831359 - MAE: 0.31768596289501483 Validation loss : 0.17874346673488617 - MAE: 0.3253559679886541 Epoch: 16 Training loss: 0.17649461567401886 - MAE: 0.3184662269845276 Validation loss : 0.17812747756640115 - MAE: 0.32483073473627894 Epoch: 17 Training loss: 0.17691377580165862 - MAE: 0.31822431281431196 Validation loss : 0.17722326185968187 - MAE: 0.3238687376329823 Epoch: 18 Training loss: 0.1768798279762268 - MAE: 0.3188422584293619 Validation loss : 0.17655367155869803 - MAE: 0.32321426470152304 Epoch: 19 Training loss: 0.17445733308792113 - MAE: 0.31640191359506836 Validation loss : 0.17601815693908268 - MAE: 0.3227230412018289 Epoch: 20 Training loss: 0.17480538487434388 - MAE: 0.3161891511192775 Validation loss : 0.1753774748908149 - MAE: 0.32208324574836594 Epoch: 21 Training loss: 0.1732891607284546 - MAE: 0.31558897873451675 Validation loss : 0.1746254563331604 - MAE: 0.3213051447966539 Epoch: 22 Training loss: 0.17241579234600068 - MAE: 0.3140265985892311 Validation loss : 0.17429026464621225 - MAE: 0.3210719014371899 Epoch: 23 Training loss: 0.17162418186664583 - MAE: 0.31408943363705194 Validation loss : 0.17367220256063673 - MAE: 0.32039953149205447 Epoch: 24 Training loss: 0.17133045554161072 - MAE: 0.3127541530123475 Validation loss : 0.17310193015469444 - MAE: 0.31980652640276674 Epoch: 25 Training loss: 0.16942350268363954 - MAE: 0.3121290878409379 Validation loss : 0.17268804874685076 - MAE: 0.31942688696053373 Epoch: 26 Training loss: 0.17049773633480073 - MAE: 0.31363952840493803 Validation loss : 0.172120107544793 - MAE: 0.31881123537483563 Epoch: 27 Training loss: 0.16871756613254546 - MAE: 0.3115901409937759 Validation loss : 0.17177432609928978 - MAE: 0.31851900529837285 Epoch: 28 Training loss: 0.16907432436943054 - MAE: 0.31170754064878 Validation loss : 0.17153260277377236 - MAE: 0.31832788132070616 Epoch: 29 Training loss: 0.16863942086696626 - MAE: 0.30940055904153907 Validation loss : 0.17075370252132416 - MAE: 0.31735976568837887 Epoch: 30 Training loss: 0.16562108039855958 - MAE: 0.3078162592146858 Validation loss : 0.17026153869099087 - MAE: 0.3168285940423929 Epoch: 31 Training loss: 0.16831409573554992 - MAE: 0.3108275862555087 Validation loss : 0.17004851169056362 - MAE: 0.3166782612622688 Epoch: 32 Training loss: 0.16607610881328583 - MAE: 0.3073008856775552 Validation loss : 0.169812281926473 - MAE: 0.3164999426442513 Epoch: 33 Training loss: 0.16536492824554444 - MAE: 0.30730756893246436 Validation loss : 0.16954073972172207 - MAE: 0.3162310921518069 Epoch: 34 Training loss: 0.16565500557422638 - MAE: 0.30880801031262106 Validation loss : 0.16896353165308634 - MAE: 0.3154897166767324 Epoch: 35 Training loss: 0.16480786919593812 - MAE: 0.3066865369170677 Validation loss : 0.16894808411598206 - MAE: 0.315587755169499 Epoch: 36 Training loss: 0.16488760769367217 - MAE: 0.3062544913178101 Validation loss : 0.1683160811662674 - MAE: 0.3147328218692977 Epoch: 37 Training loss: 0.16296948552131651 - MAE: 0.30529006908692957 Validation loss : 0.16822955012321472 - MAE: 0.31474619083221317 Epoch: 38 Training loss: 0.16405799746513366 - MAE: 0.3065666077277542 Validation loss : 0.16793323556582132 - MAE: 0.3143827065817725 Epoch: 39 Training loss: 0.16395046710968017 - MAE: 0.3061178662220702 Validation loss : 0.16756848825348747 - MAE: 0.31392760086070465 Epoch: 40 Training loss: 0.16225593984127046 - MAE: 0.30391575924794567 Validation loss : 0.16728156473901537 - MAE: 0.31359587016220536 Epoch: 41 Training loss: 0.162018803358078 - MAE: 0.3050126817417 Validation loss : 0.16726960241794586 - MAE: 0.3136659181800001 Epoch: 42 Training loss: 0.1616269725561142 - MAE: 0.3047072823079763 Validation loss : 0.1666801455948088 - MAE: 0.31286053951971926 Epoch: 43 Training loss: 0.16371312737464905 - MAE: 0.3072663025489704 Validation loss : 0.16660702228546143 - MAE: 0.3128544246563194 Epoch: 44 Training loss: 0.16274581968784332 - MAE: 0.3055276315731144 Validation loss : 0.16622301439444223 - MAE: 0.3123737118409309 Epoch: 45 Training loss: 0.16198877573013307 - MAE: 0.30387318215935566 Validation loss : 0.16592974132961696 - MAE: 0.3120266870766228 Epoch: 46 Training loss: 0.1613679423928261 - MAE: 0.30285112156406757 Validation loss : 0.16592562860912746 - MAE: 0.3120895779099974 Epoch: 47 Training loss: 0.1596740049123764 - MAE: 0.30196700285423644 Validation loss : 0.1656825906700558 - MAE: 0.3117942734701369 Epoch: 48 Training loss: 0.16010645925998687 - MAE: 0.3010781145976006 Validation loss : 0.1654240306880739 - MAE: 0.3114403430988788 Epoch: 49 Training loss: 0.16135582566261292 - MAE: 0.3033602255531342 Validation loss : 0.16523137191931406 - MAE: 0.3111766545135443 Epoch: 50 Training loss: 0.15887635439634323 - MAE: 0.30139360974455354 Validation loss : 0.16520478659205967 - MAE: 0.31118266518707904 Epoch: 51 Training loss: 0.1623841804265976 - MAE: 0.3041726753025802 Validation loss : 0.16488962206575605 - MAE: 0.3107705042403597 Epoch: 52 Training loss: 0.15951467990875245 - MAE: 0.3024485706726396 Validation loss : 0.16450322336620754 - MAE: 0.3101646703863994 Epoch: 53 Training loss: 0.1593264466524124 - MAE: 0.3019928035241633 Validation loss : 0.16450072824954987 - MAE: 0.3102381133991464 Epoch: 54 Training loss: 0.15892296969890596 - MAE: 0.3010730142919314 Validation loss : 0.1644514716333813 - MAE: 0.31025052996699415 Epoch: 55 Training loss: 0.15801858335733412 - MAE: 0.300129276806219 Validation loss : 0.1639546040031645 - MAE: 0.30954432283469774 Epoch: 56 Training loss: 0.15852347761392593 - MAE: 0.30000137034695895 Validation loss : 0.16394910713036856 - MAE: 0.3096252551544829 Epoch: 57 Training loss: 0.15989623069763184 - MAE: 0.3008597511380722 Validation loss : 0.1637496915128496 - MAE: 0.30938490975556004 Epoch: 58 Training loss: 0.15900149762630464 - MAE: 0.30091532233302887 Validation loss : 0.16357198854287466 - MAE: 0.3091647551750498 Epoch: 59 Training loss: 0.15815627008676528 - MAE: 0.2998753613511919 Validation loss : 0.16364828579955631 - MAE: 0.30934114039380045 Epoch: 60 Training loss: 0.15863654345273973 - MAE: 0.2997401767468425 Validation loss : 0.16360284553633797 - MAE: 0.3093700140804052 Epoch: 61 Training loss: 0.1569861376285553 - MAE: 0.3001724729244438 Validation loss : 0.1631663309203254 - MAE: 0.3087389482069802 Epoch: 62 Training loss: 0.15764495313167573 - MAE: 0.29931992861180673 Validation loss : 0.1628624697526296 - MAE: 0.3082940994455321 Epoch: 63 Training loss: 0.15620498329401017 - MAE: 0.2983585146384368 Validation loss : 0.16287368204858568 - MAE: 0.30839281732122203 Epoch: 64 Training loss: 0.15624627590179443 - MAE: 0.29862051694345537 Validation loss : 0.16266994840568966 - MAE: 0.308149927566448 Epoch: 65 Training loss: 0.15566111862659454 - MAE: 0.29753564913606617 Validation loss : 0.1626807086997562 - MAE: 0.30822756782090394 Epoch: 66 Training loss: 0.15647490918636323 - MAE: 0.29872063729818527 Validation loss : 0.16234126521481407 - MAE: 0.3077123779358625 Epoch: 67 Training loss: 0.15619896858930588 - MAE: 0.297556687519032 Validation loss : 0.16227295829190147 - MAE: 0.30769900980501297 Epoch: 68 Training loss: 0.15651787519454957 - MAE: 0.2983152167118014 Validation loss : 0.16212009721332127 - MAE: 0.30753654402945435 Epoch: 69 Training loss: 0.15651776432991027 - MAE: 0.299202365307033 Validation loss : 0.16212468677096897 - MAE: 0.3076282798695557 Epoch: 70 Training loss: 0.15673158675432206 - MAE: 0.29856837209150694 Validation loss : 0.16198681791623434 - MAE: 0.30741784086039886 Epoch: 71 Training loss: 0.15566783308982848 - MAE: 0.2983743266636147 Validation loss : 0.1617990599738227 - MAE: 0.3071957299555147 Epoch: 72 Training loss: 0.1555081295967102 - MAE: 0.296568577820393 Validation loss : 0.16165704197353786 - MAE: 0.30704654154909133 Epoch: 73 Training loss: 0.1559882402420044 - MAE: 0.2972687981178256 Validation loss : 0.16142904261747995 - MAE: 0.3066734296802609 Epoch: 74 Training loss: 0.1546381199359894 - MAE: 0.29657499645115093 Validation loss : 0.1616316239039103 - MAE: 0.30704220168092655 Epoch: 75 Training loss: 0.15564891695976257 - MAE: 0.29796052691901004 Validation loss : 0.16111160814762115 - MAE: 0.30627361301565775 Epoch: 76 Training loss: 0.15516877621412278 - MAE: 0.29683828489531877 Validation loss : 0.16111387809117636 - MAE: 0.3063159431672632 Epoch: 77 Training loss: 0.15396408140659332 - MAE: 0.2969101587330462 Validation loss : 0.1609818802939521 - MAE: 0.30617236589118246 |