home-standard-results / 06-05-2023_norbert_batchsize_256_all_units_0.0001_1_200_freeze_True_earlystop_3.txt
ececet's picture
commit files to HF hub
c606fb2
raw
history blame
7.27 kB
Started at: 17:01:49
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Started at: 17:03:05
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Started at: 17:03:18
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 1.5561567282676696 - MAE: 1.1770511100698875
Validation loss : 0.939958651860555 - MAE: 0.8876047142856601
Epoch: 1
Training loss: 0.6363586246967315 - MAE: 0.6928943697649724
Validation loss : 0.33423560195498997 - MAE: 0.45375851608477785
Epoch: 2
Training loss: 0.2759279823303223 - MAE: 0.4052092643476477
Validation loss : 0.20369169281588662 - MAE: 0.3433357920074413
Epoch: 3
Training loss: 0.20272068321704864 - MAE: 0.3399617345503301
Validation loss : 0.19545602964030373 - MAE: 0.3416856898288076
Epoch: 4
Training loss: 0.1912100625038147 - MAE: 0.3326939730215834
Validation loss : 0.19473741783036125 - MAE: 0.3417824430892937
Epoch: 5
Training loss: 0.19083771824836732 - MAE: 0.33137282498844967
Validation loss : 0.19215407967567444 - MAE: 0.33912431491105655
Epoch: 6
Training loss: 0.18923945665359498 - MAE: 0.33146465315973606
Validation loss : 0.18989400565624237 - MAE: 0.33674625605103725
Epoch: 7
Training loss: 0.18950583219528197 - MAE: 0.33087934479225867
Validation loss : 0.18789740900198618 - MAE: 0.33464578460452915
Epoch: 8
Training loss: 0.1856936401128769 - MAE: 0.3275132613322806
Validation loss : 0.1864797506067488 - MAE: 0.33324104491047596
Epoch: 9
Training loss: 0.1836453753709793 - MAE: 0.3255223596708951
Validation loss : 0.18501530918810102 - MAE: 0.3316937400385192
Epoch: 10
Training loss: 0.18246697008609772 - MAE: 0.32322550938186817
Validation loss : 0.18376492957274118 - MAE: 0.33042730021623856
Epoch: 11
Training loss: 0.1811656492948532 - MAE: 0.32419844121595726
Validation loss : 0.18262643449836308 - MAE: 0.32929917485750665
Epoch: 12
Training loss: 0.18094802320003509 - MAE: 0.32277421430430436
Validation loss : 0.18164582550525665 - MAE: 0.3283432280968596
Epoch: 13
Training loss: 0.1792392772436142 - MAE: 0.3209453182214294
Validation loss : 0.18065459695127276 - MAE: 0.3273255372779676
Epoch: 14
Training loss: 0.17940414905548097 - MAE: 0.3216579290630419
Validation loss : 0.17975443104902902 - MAE: 0.3264252069830345
Epoch: 15
Training loss: 0.17663683056831359 - MAE: 0.31768596289501483
Validation loss : 0.17874346673488617 - MAE: 0.3253559679886541
Epoch: 16
Training loss: 0.17649461567401886 - MAE: 0.3184662269845276
Validation loss : 0.17812747756640115 - MAE: 0.32483073473627894
Epoch: 17
Training loss: 0.17691377580165862 - MAE: 0.31822431281431196
Validation loss : 0.17722326185968187 - MAE: 0.3238687376329823
Epoch: 18
Training loss: 0.1768798279762268 - MAE: 0.3188422584293619
Validation loss : 0.17655367155869803 - MAE: 0.32321426470152304
Epoch: 19
Training loss: 0.17445733308792113 - MAE: 0.31640191359506836
Validation loss : 0.17601815693908268 - MAE: 0.3227230412018289
Epoch: 20
Training loss: 0.17480538487434388 - MAE: 0.3161891511192775
Validation loss : 0.1753774748908149 - MAE: 0.32208324574836594
Epoch: 21
Training loss: 0.1732891607284546 - MAE: 0.31558897873451675
Validation loss : 0.1746254563331604 - MAE: 0.3213051447966539
Epoch: 22
Training loss: 0.17241579234600068 - MAE: 0.3140265985892311
Validation loss : 0.17429026464621225 - MAE: 0.3210719014371899
Epoch: 23
Training loss: 0.17162418186664583 - MAE: 0.31408943363705194
Validation loss : 0.17367220256063673 - MAE: 0.32039953149205447
Epoch: 24
Training loss: 0.17133045554161072 - MAE: 0.3127541530123475
Validation loss : 0.17310193015469444 - MAE: 0.31980652640276674
Epoch: 25
Training loss: 0.16942350268363954 - MAE: 0.3121290878409379
Validation loss : 0.17268804874685076 - MAE: 0.31942688696053373
Epoch: 26
Training loss: 0.17049773633480073 - MAE: 0.31363952840493803
Validation loss : 0.172120107544793 - MAE: 0.31881123537483563
Epoch: 27
Training loss: 0.16871756613254546 - MAE: 0.3115901409937759
Validation loss : 0.17177432609928978 - MAE: 0.31851900529837285
Epoch: 28
Training loss: 0.16907432436943054 - MAE: 0.31170754064878
Validation loss : 0.17153260277377236 - MAE: 0.31832788132070616
Epoch: 29
Training loss: 0.16863942086696626 - MAE: 0.30940055904153907
Validation loss : 0.17075370252132416 - MAE: 0.31735976568837887
Epoch: 30
Training loss: 0.16562108039855958 - MAE: 0.3078162592146858
Validation loss : 0.17026153869099087 - MAE: 0.3168285940423929
Epoch: 31
Training loss: 0.16831409573554992 - MAE: 0.3108275862555087
Validation loss : 0.17004851169056362 - MAE: 0.3166782612622688
Epoch: 32
Training loss: 0.16607610881328583 - MAE: 0.3073008856775552
Validation loss : 0.169812281926473 - MAE: 0.3164999426442513
Epoch: 33
Training loss: 0.16536492824554444 - MAE: 0.30730756893246436
Validation loss : 0.16954073972172207 - MAE: 0.3162310921518069
Epoch: 34
Training loss: 0.16565500557422638 - MAE: 0.30880801031262106
Validation loss : 0.16896353165308634 - MAE: 0.3154897166767324
Epoch: 35
Training loss: 0.16480786919593812 - MAE: 0.3066865369170677
Validation loss : 0.16894808411598206 - MAE: 0.315587755169499
Epoch: 36
Training loss: 0.16488760769367217 - MAE: 0.3062544913178101
Validation loss : 0.1683160811662674 - MAE: 0.3147328218692977
Epoch: 37
Training loss: 0.16296948552131651 - MAE: 0.30529006908692957
Validation loss : 0.16822955012321472 - MAE: 0.31474619083221317
Epoch: 38
Training loss: 0.16405799746513366 - MAE: 0.3065666077277542
Validation loss : 0.16793323556582132 - MAE: 0.3143827065817725
Epoch: 39
Training loss: 0.16395046710968017 - MAE: 0.3061178662220702
Validation loss : 0.16756848825348747 - MAE: 0.31392760086070465
Epoch: 40
Training loss: 0.16225593984127046 - MAE: 0.30391575924794567
Validation loss : 0.16728156473901537 - MAE: 0.31359587016220536
Epoch: 41
Training loss: 0.162018803358078 - MAE: 0.3050126817417
Validation loss : 0.16726960241794586 - MAE: 0.3136659181800001
Epoch: 42
Training loss: 0.1616269725561142 - MAE: 0.3047072823079763
Validation loss : 0.1666801455948088 - MAE: 0.31286053951971926
Epoch: 43
Training loss: 0.16371312737464905 - MAE: 0.3072663025489704
Validation loss : 0.16660702228546143 - MAE: 0.3128544246563194
Epoch: 44