File size: 8,201 Bytes
ee53a4a 50f1745 8a33155 79fb205 22be4a5 3f1ff0e bbed904 14c983c bb23377 0ea8477 8597b5f a529489 8f5752b d444cf9 2dfa794 782f4ba 2665515 8c5051b 84a2460 36464be b4eab8a b4c7480 bd5a96e c248496 a3f3665 3572268 8f95e8a 0e3f82d bd48add 1da9fd2 3759f70 e09bf1b ec4c3b8 b891788 9c846de c36a8be 3f890ba aaf9dcb 9442814 64e656f 47314ea b48fff4 c931c86 ec8e58d 28782de d61067e 2cde297 dce64ed 1a44567 a894d88 f8b0b44 63524f2 3dc095f 915dd82 a5a47b9 15b72bc 5ba086a 6a406c0 fd9bd7c f2205a0 712111f 21fd887 992268b 6fd095f 71d7764 c0f3990 73015a7 ad81682 5a521e7 daf7029 3060a2d 4be9431 8aa7919 c895440 9a9cef0 f236bc8 66cca27 7e49157 a294d56 febe9a0 83c1bd3 9469a62 90a233b 539cc86 777a580 52b7ba5 f7119ee 9186c3b e2d4e0d c7d4106 edc5e85 7e70a20 720f4a0 61139b9 86ab2b7 9401e87 fd467de ee43001 1078eb3 6a7636d 5bac8c4 5d0721b a1c0044 35c229f 4222c2e 7d91c74 01aac9c 9c3e95d 050c19d 3d484a1 229d040 53d9b60 fde77a5 af0d9dd 999f338 9581137 1591024 f2c7f34 03b454e cefb7da 3b85915 14be623 c5851b0 65d6066 978a32c 9f52222 f5a0354 05e9788 4f6505f e305e3a a4f6d7a bdd2cfd bbef9aa 1054174 63b42b9 6f00408 055694c 90842d0 5c96627 c606fb2 e2dfc21 db90dbb 77c5e49 5092051 973f642 c1bd784 dc3afdd 4ca984a 47b850e b7433ab 2d21015 47f85d1 cb32d0b 1c54cf7 da0b33f b0afa4f 9d74bc6 7ee2037 3998ccb 70afe20 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 |
Started at: 17:01:49 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Started at: 17:03:05 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Started at: 17:03:18 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 1.5561567282676696 - MAE: 1.1770511100698875 Validation loss : 0.939958651860555 - MAE: 0.8876047142856601 Epoch: 1 Training loss: 0.6363586246967315 - MAE: 0.6928943697649724 Validation loss : 0.33423560195498997 - MAE: 0.45375851608477785 Epoch: 2 Training loss: 0.2759279823303223 - MAE: 0.4052092643476477 Validation loss : 0.20369169281588662 - MAE: 0.3433357920074413 Epoch: 3 Training loss: 0.20272068321704864 - MAE: 0.3399617345503301 Validation loss : 0.19545602964030373 - MAE: 0.3416856898288076 Epoch: 4 Training loss: 0.1912100625038147 - MAE: 0.3326939730215834 Validation loss : 0.19473741783036125 - MAE: 0.3417824430892937 Epoch: 5 Training loss: 0.19083771824836732 - MAE: 0.33137282498844967 Validation loss : 0.19215407967567444 - MAE: 0.33912431491105655 Epoch: 6 Training loss: 0.18923945665359498 - MAE: 0.33146465315973606 Validation loss : 0.18989400565624237 - MAE: 0.33674625605103725 Epoch: 7 Training loss: 0.18950583219528197 - MAE: 0.33087934479225867 Validation loss : 0.18789740900198618 - MAE: 0.33464578460452915 Epoch: 8 Training loss: 0.1856936401128769 - MAE: 0.3275132613322806 Validation loss : 0.1864797506067488 - MAE: 0.33324104491047596 Epoch: 9 Training loss: 0.1836453753709793 - MAE: 0.3255223596708951 Validation loss : 0.18501530918810102 - MAE: 0.3316937400385192 Epoch: 10 Training loss: 0.18246697008609772 - MAE: 0.32322550938186817 Validation loss : 0.18376492957274118 - MAE: 0.33042730021623856 Epoch: 11 Training loss: 0.1811656492948532 - MAE: 0.32419844121595726 Validation loss : 0.18262643449836308 - MAE: 0.32929917485750665 Epoch: 12 Training loss: 0.18094802320003509 - MAE: 0.32277421430430436 Validation loss : 0.18164582550525665 - MAE: 0.3283432280968596 Epoch: 13 Training loss: 0.1792392772436142 - MAE: 0.3209453182214294 Validation loss : 0.18065459695127276 - MAE: 0.3273255372779676 Epoch: 14 Training loss: 0.17940414905548097 - MAE: 0.3216579290630419 Validation loss : 0.17975443104902902 - MAE: 0.3264252069830345 Epoch: 15 Training loss: 0.17663683056831359 - MAE: 0.31768596289501483 Validation loss : 0.17874346673488617 - MAE: 0.3253559679886541 Epoch: 16 Training loss: 0.17649461567401886 - MAE: 0.3184662269845276 Validation loss : 0.17812747756640115 - MAE: 0.32483073473627894 Epoch: 17 Training loss: 0.17691377580165862 - MAE: 0.31822431281431196 Validation loss : 0.17722326185968187 - MAE: 0.3238687376329823 Epoch: 18 Training loss: 0.1768798279762268 - MAE: 0.3188422584293619 Validation loss : 0.17655367155869803 - MAE: 0.32321426470152304 Epoch: 19 Training loss: 0.17445733308792113 - MAE: 0.31640191359506836 Validation loss : 0.17601815693908268 - MAE: 0.3227230412018289 Epoch: 20 Training loss: 0.17480538487434388 - MAE: 0.3161891511192775 Validation loss : 0.1753774748908149 - MAE: 0.32208324574836594 Epoch: 21 Training loss: 0.1732891607284546 - MAE: 0.31558897873451675 Validation loss : 0.1746254563331604 - MAE: 0.3213051447966539 Epoch: 22 Training loss: 0.17241579234600068 - MAE: 0.3140265985892311 Validation loss : 0.17429026464621225 - MAE: 0.3210719014371899 Epoch: 23 Training loss: 0.17162418186664583 - MAE: 0.31408943363705194 Validation loss : 0.17367220256063673 - MAE: 0.32039953149205447 Epoch: 24 Training loss: 0.17133045554161072 - MAE: 0.3127541530123475 Validation loss : 0.17310193015469444 - MAE: 0.31980652640276674 Epoch: 25 Training loss: 0.16942350268363954 - MAE: 0.3121290878409379 Validation loss : 0.17268804874685076 - MAE: 0.31942688696053373 Epoch: 26 Training loss: 0.17049773633480073 - MAE: 0.31363952840493803 Validation loss : 0.172120107544793 - MAE: 0.31881123537483563 Epoch: 27 Training loss: 0.16871756613254546 - MAE: 0.3115901409937759 Validation loss : 0.17177432609928978 - MAE: 0.31851900529837285 Epoch: 28 Training loss: 0.16907432436943054 - MAE: 0.31170754064878 Validation loss : 0.17153260277377236 - MAE: 0.31832788132070616 Epoch: 29 Training loss: 0.16863942086696626 - MAE: 0.30940055904153907 Validation loss : 0.17075370252132416 - MAE: 0.31735976568837887 Epoch: 30 Training loss: 0.16562108039855958 - MAE: 0.3078162592146858 Validation loss : 0.17026153869099087 - MAE: 0.3168285940423929 Epoch: 31 Training loss: 0.16831409573554992 - MAE: 0.3108275862555087 Validation loss : 0.17004851169056362 - MAE: 0.3166782612622688 Epoch: 32 Training loss: 0.16607610881328583 - MAE: 0.3073008856775552 Validation loss : 0.169812281926473 - MAE: 0.3164999426442513 Epoch: 33 Training loss: 0.16536492824554444 - MAE: 0.30730756893246436 Validation loss : 0.16954073972172207 - MAE: 0.3162310921518069 Epoch: 34 Training loss: 0.16565500557422638 - MAE: 0.30880801031262106 Validation loss : 0.16896353165308634 - MAE: 0.3154897166767324 Epoch: 35 Training loss: 0.16480786919593812 - MAE: 0.3066865369170677 Validation loss : 0.16894808411598206 - MAE: 0.315587755169499 Epoch: 36 Training loss: 0.16488760769367217 - MAE: 0.3062544913178101 Validation loss : 0.1683160811662674 - MAE: 0.3147328218692977 Epoch: 37 Training loss: 0.16296948552131651 - MAE: 0.30529006908692957 Validation loss : 0.16822955012321472 - MAE: 0.31474619083221317 Epoch: 38 Training loss: 0.16405799746513366 - MAE: 0.3065666077277542 Validation loss : 0.16793323556582132 - MAE: 0.3143827065817725 Epoch: 39 Training loss: 0.16395046710968017 - MAE: 0.3061178662220702 Validation loss : 0.16756848825348747 - MAE: 0.31392760086070465 Epoch: 40 Training loss: 0.16225593984127046 - MAE: 0.30391575924794567 Validation loss : 0.16728156473901537 - MAE: 0.31359587016220536 Epoch: 41 Training loss: 0.162018803358078 - MAE: 0.3050126817417 Validation loss : 0.16726960241794586 - MAE: 0.3136659181800001 Epoch: 42 Training loss: 0.1616269725561142 - MAE: 0.3047072823079763 Validation loss : 0.1666801455948088 - MAE: 0.31286053951971926 Epoch: 43 Training loss: 0.16371312737464905 - MAE: 0.3072663025489704 Validation loss : 0.16660702228546143 - MAE: 0.3128544246563194 Epoch: 44 Training loss: 0.16274581968784332 - MAE: 0.3055276315731144 Validation loss : 0.16622301439444223 - MAE: 0.3123737118409309 Epoch: 45 Training loss: 0.16198877573013307 - MAE: 0.30387318215935566 Validation loss : 0.16592974132961696 - MAE: 0.3120266870766228 Epoch: 46 Training loss: 0.1613679423928261 - MAE: 0.30285112156406757 Validation loss : 0.16592562860912746 - MAE: 0.3120895779099974 Epoch: 47 Training loss: 0.1596740049123764 - MAE: 0.30196700285423644 Validation loss : 0.1656825906700558 - MAE: 0.3117942734701369 Epoch: 48 Training loss: 0.16010645925998687 - MAE: 0.3010781145976006 Validation loss : 0.1654240306880739 - MAE: 0.3114403430988788 Epoch: 49 Training loss: 0.16135582566261292 - MAE: 0.3033602255531342 Validation loss : 0.16523137191931406 - MAE: 0.3111766545135443 Epoch: 50 Training loss: 0.15887635439634323 - MAE: 0.30139360974455354 Validation loss : 0.16520478659205967 - MAE: 0.31118266518707904 |