Started at: 17:01:49 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Started at: 17:03:05 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Started at: 17:03:18 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 1.5561567282676696 - MAE: 1.1770511100698875 Validation loss : 0.939958651860555 - MAE: 0.8876047142856601 Epoch: 1 Training loss: 0.6363586246967315 - MAE: 0.6928943697649724 Validation loss : 0.33423560195498997 - MAE: 0.45375851608477785 Epoch: 2 Training loss: 0.2759279823303223 - MAE: 0.4052092643476477 Validation loss : 0.20369169281588662 - MAE: 0.3433357920074413 Epoch: 3 Training loss: 0.20272068321704864 - MAE: 0.3399617345503301 Validation loss : 0.19545602964030373 - MAE: 0.3416856898288076 Epoch: 4 Training loss: 0.1912100625038147 - MAE: 0.3326939730215834 Validation loss : 0.19473741783036125 - MAE: 0.3417824430892937 Epoch: 5 Training loss: 0.19083771824836732 - MAE: 0.33137282498844967 Validation loss : 0.19215407967567444 - MAE: 0.33912431491105655 Epoch: 6 Training loss: 0.18923945665359498 - MAE: 0.33146465315973606 Validation loss : 0.18989400565624237 - MAE: 0.33674625605103725 Epoch: 7 Training loss: 0.18950583219528197 - MAE: 0.33087934479225867 Validation loss : 0.18789740900198618 - MAE: 0.33464578460452915 Epoch: 8 Training loss: 0.1856936401128769 - MAE: 0.3275132613322806 Validation loss : 0.1864797506067488 - MAE: 0.33324104491047596 Epoch: 9 Training loss: 0.1836453753709793 - MAE: 0.3255223596708951 Validation loss : 0.18501530918810102 - MAE: 0.3316937400385192 Epoch: 10 Training loss: 0.18246697008609772 - MAE: 0.32322550938186817 Validation loss : 0.18376492957274118 - MAE: 0.33042730021623856 Epoch: 11 Training loss: 0.1811656492948532 - MAE: 0.32419844121595726 Validation loss : 0.18262643449836308 - MAE: 0.32929917485750665 Epoch: 12 Training loss: 0.18094802320003509 - MAE: 0.32277421430430436 Validation loss : 0.18164582550525665 - MAE: 0.3283432280968596 Epoch: 13 Training loss: 0.1792392772436142 - MAE: 0.3209453182214294 Validation loss : 0.18065459695127276 - MAE: 0.3273255372779676 Epoch: 14 Training loss: 0.17940414905548097 - MAE: 0.3216579290630419 Validation loss : 0.17975443104902902 - MAE: 0.3264252069830345 Epoch: 15 Training loss: 0.17663683056831359 - MAE: 0.31768596289501483 Validation loss : 0.17874346673488617 - MAE: 0.3253559679886541 Epoch: 16 Training loss: 0.17649461567401886 - MAE: 0.3184662269845276 Validation loss : 0.17812747756640115 - MAE: 0.32483073473627894 Epoch: 17 Training loss: 0.17691377580165862 - MAE: 0.31822431281431196 Validation loss : 0.17722326185968187 - MAE: 0.3238687376329823 Epoch: 18 Training loss: 0.1768798279762268 - MAE: 0.3188422584293619 Validation loss : 0.17655367155869803 - MAE: 0.32321426470152304 Epoch: 19 Training loss: 0.17445733308792113 - MAE: 0.31640191359506836 Validation loss : 0.17601815693908268 - MAE: 0.3227230412018289 Epoch: 20 Training loss: 0.17480538487434388 - MAE: 0.3161891511192775 Validation loss : 0.1753774748908149 - MAE: 0.32208324574836594 Epoch: 21 Training loss: 0.1732891607284546 - MAE: 0.31558897873451675 Validation loss : 0.1746254563331604 - MAE: 0.3213051447966539 Epoch: 22 Training loss: 0.17241579234600068 - MAE: 0.3140265985892311 Validation loss : 0.17429026464621225 - MAE: 0.3210719014371899 Epoch: 23 Training loss: 0.17162418186664583 - MAE: 0.31408943363705194 Validation loss : 0.17367220256063673 - MAE: 0.32039953149205447 Epoch: 24 Training loss: 0.17133045554161072 - MAE: 0.3127541530123475 Validation loss : 0.17310193015469444 - MAE: 0.31980652640276674 Epoch: 25 Training loss: 0.16942350268363954 - MAE: 0.3121290878409379 Validation loss : 0.17268804874685076 - MAE: 0.31942688696053373 Epoch: 26 Training loss: 0.17049773633480073 - MAE: 0.31363952840493803 Validation loss : 0.172120107544793 - MAE: 0.31881123537483563 Epoch: 27 Training loss: 0.16871756613254546 - MAE: 0.3115901409937759 Validation loss : 0.17177432609928978 - MAE: 0.31851900529837285 Epoch: 28 Training loss: 0.16907432436943054 - MAE: 0.31170754064878 Validation loss : 0.17153260277377236 - MAE: 0.31832788132070616 Epoch: 29 Training loss: 0.16863942086696626 - MAE: 0.30940055904153907 Validation loss : 0.17075370252132416 - MAE: 0.31735976568837887 Epoch: 30 Training loss: 0.16562108039855958 - MAE: 0.3078162592146858 Validation loss : 0.17026153869099087 - MAE: 0.3168285940423929 Epoch: 31 Training loss: 0.16831409573554992 - MAE: 0.3108275862555087 Validation loss : 0.17004851169056362 - MAE: 0.3166782612622688 Epoch: 32 Training loss: 0.16607610881328583 - MAE: 0.3073008856775552 Validation loss : 0.169812281926473 - MAE: 0.3164999426442513 Epoch: 33 Training loss: 0.16536492824554444 - MAE: 0.30730756893246436 Validation loss : 0.16954073972172207 - MAE: 0.3162310921518069 Epoch: 34 Training loss: 0.16565500557422638 - MAE: 0.30880801031262106 Validation loss : 0.16896353165308634 - MAE: 0.3154897166767324 Epoch: 35 Training loss: 0.16480786919593812 - MAE: 0.3066865369170677 Validation loss : 0.16894808411598206 - MAE: 0.315587755169499 Epoch: 36 Training loss: 0.16488760769367217 - MAE: 0.3062544913178101 Validation loss : 0.1683160811662674 - MAE: 0.3147328218692977 Epoch: 37 Training loss: 0.16296948552131651 - MAE: 0.30529006908692957 Validation loss : 0.16822955012321472 - MAE: 0.31474619083221317 Epoch: 38 Training loss: 0.16405799746513366 - MAE: 0.3065666077277542 Validation loss : 0.16793323556582132 - MAE: 0.3143827065817725 Epoch: 39 Training loss: 0.16395046710968017 - MAE: 0.3061178662220702 Validation loss : 0.16756848825348747 - MAE: 0.31392760086070465 Epoch: 40 Training loss: 0.16225593984127046 - MAE: 0.30391575924794567 Validation loss : 0.16728156473901537 - MAE: 0.31359587016220536 Epoch: 41 Training loss: 0.162018803358078 - MAE: 0.3050126817417 Validation loss : 0.16726960241794586 - MAE: 0.3136659181800001 Epoch: 42 Training loss: 0.1616269725561142 - MAE: 0.3047072823079763 Validation loss : 0.1666801455948088 - MAE: 0.31286053951971926 Epoch: 43 Training loss: 0.16371312737464905 - MAE: 0.3072663025489704 Validation loss : 0.16660702228546143 - MAE: 0.3128544246563194 Epoch: 44 Training loss: 0.16274581968784332 - MAE: 0.3055276315731144 Validation loss : 0.16622301439444223 - MAE: 0.3123737118409309 Epoch: 45 Training loss: 0.16198877573013307 - MAE: 0.30387318215935566 Validation loss : 0.16592974132961696 - MAE: 0.3120266870766228 Epoch: 46 Training loss: 0.1613679423928261 - MAE: 0.30285112156406757 Validation loss : 0.16592562860912746 - MAE: 0.3120895779099974 Epoch: 47 Training loss: 0.1596740049123764 - MAE: 0.30196700285423644 Validation loss : 0.1656825906700558 - MAE: 0.3117942734701369