home-standard-results / 06-05-2023_norbert_batchsize_256_all_units_0.0001_1_200_freeze_True_earlystop_3.txt
ececet's picture
commit files to HF hub
2c58b7f
raw
history blame
11.7 kB
Started at: 17:01:49
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Started at: 17:03:05
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Started at: 17:03:18
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 1.5561567282676696 - MAE: 1.1770511100698875
Validation loss : 0.939958651860555 - MAE: 0.8876047142856601
Epoch: 1
Training loss: 0.6363586246967315 - MAE: 0.6928943697649724
Validation loss : 0.33423560195498997 - MAE: 0.45375851608477785
Epoch: 2
Training loss: 0.2759279823303223 - MAE: 0.4052092643476477
Validation loss : 0.20369169281588662 - MAE: 0.3433357920074413
Epoch: 3
Training loss: 0.20272068321704864 - MAE: 0.3399617345503301
Validation loss : 0.19545602964030373 - MAE: 0.3416856898288076
Epoch: 4
Training loss: 0.1912100625038147 - MAE: 0.3326939730215834
Validation loss : 0.19473741783036125 - MAE: 0.3417824430892937
Epoch: 5
Training loss: 0.19083771824836732 - MAE: 0.33137282498844967
Validation loss : 0.19215407967567444 - MAE: 0.33912431491105655
Epoch: 6
Training loss: 0.18923945665359498 - MAE: 0.33146465315973606
Validation loss : 0.18989400565624237 - MAE: 0.33674625605103725
Epoch: 7
Training loss: 0.18950583219528197 - MAE: 0.33087934479225867
Validation loss : 0.18789740900198618 - MAE: 0.33464578460452915
Epoch: 8
Training loss: 0.1856936401128769 - MAE: 0.3275132613322806
Validation loss : 0.1864797506067488 - MAE: 0.33324104491047596
Epoch: 9
Training loss: 0.1836453753709793 - MAE: 0.3255223596708951
Validation loss : 0.18501530918810102 - MAE: 0.3316937400385192
Epoch: 10
Training loss: 0.18246697008609772 - MAE: 0.32322550938186817
Validation loss : 0.18376492957274118 - MAE: 0.33042730021623856
Epoch: 11
Training loss: 0.1811656492948532 - MAE: 0.32419844121595726
Validation loss : 0.18262643449836308 - MAE: 0.32929917485750665
Epoch: 12
Training loss: 0.18094802320003509 - MAE: 0.32277421430430436
Validation loss : 0.18164582550525665 - MAE: 0.3283432280968596
Epoch: 13
Training loss: 0.1792392772436142 - MAE: 0.3209453182214294
Validation loss : 0.18065459695127276 - MAE: 0.3273255372779676
Epoch: 14
Training loss: 0.17940414905548097 - MAE: 0.3216579290630419
Validation loss : 0.17975443104902902 - MAE: 0.3264252069830345
Epoch: 15
Training loss: 0.17663683056831359 - MAE: 0.31768596289501483
Validation loss : 0.17874346673488617 - MAE: 0.3253559679886541
Epoch: 16
Training loss: 0.17649461567401886 - MAE: 0.3184662269845276
Validation loss : 0.17812747756640115 - MAE: 0.32483073473627894
Epoch: 17
Training loss: 0.17691377580165862 - MAE: 0.31822431281431196
Validation loss : 0.17722326185968187 - MAE: 0.3238687376329823
Epoch: 18
Training loss: 0.1768798279762268 - MAE: 0.3188422584293619
Validation loss : 0.17655367155869803 - MAE: 0.32321426470152304
Epoch: 19
Training loss: 0.17445733308792113 - MAE: 0.31640191359506836
Validation loss : 0.17601815693908268 - MAE: 0.3227230412018289
Epoch: 20
Training loss: 0.17480538487434388 - MAE: 0.3161891511192775
Validation loss : 0.1753774748908149 - MAE: 0.32208324574836594
Epoch: 21
Training loss: 0.1732891607284546 - MAE: 0.31558897873451675
Validation loss : 0.1746254563331604 - MAE: 0.3213051447966539
Epoch: 22
Training loss: 0.17241579234600068 - MAE: 0.3140265985892311
Validation loss : 0.17429026464621225 - MAE: 0.3210719014371899
Epoch: 23
Training loss: 0.17162418186664583 - MAE: 0.31408943363705194
Validation loss : 0.17367220256063673 - MAE: 0.32039953149205447
Epoch: 24
Training loss: 0.17133045554161072 - MAE: 0.3127541530123475
Validation loss : 0.17310193015469444 - MAE: 0.31980652640276674
Epoch: 25
Training loss: 0.16942350268363954 - MAE: 0.3121290878409379
Validation loss : 0.17268804874685076 - MAE: 0.31942688696053373
Epoch: 26
Training loss: 0.17049773633480073 - MAE: 0.31363952840493803
Validation loss : 0.172120107544793 - MAE: 0.31881123537483563
Epoch: 27
Training loss: 0.16871756613254546 - MAE: 0.3115901409937759
Validation loss : 0.17177432609928978 - MAE: 0.31851900529837285
Epoch: 28
Training loss: 0.16907432436943054 - MAE: 0.31170754064878
Validation loss : 0.17153260277377236 - MAE: 0.31832788132070616
Epoch: 29
Training loss: 0.16863942086696626 - MAE: 0.30940055904153907
Validation loss : 0.17075370252132416 - MAE: 0.31735976568837887
Epoch: 30
Training loss: 0.16562108039855958 - MAE: 0.3078162592146858
Validation loss : 0.17026153869099087 - MAE: 0.3168285940423929
Epoch: 31
Training loss: 0.16831409573554992 - MAE: 0.3108275862555087
Validation loss : 0.17004851169056362 - MAE: 0.3166782612622688
Epoch: 32
Training loss: 0.16607610881328583 - MAE: 0.3073008856775552
Validation loss : 0.169812281926473 - MAE: 0.3164999426442513
Epoch: 33
Training loss: 0.16536492824554444 - MAE: 0.30730756893246436
Validation loss : 0.16954073972172207 - MAE: 0.3162310921518069
Epoch: 34
Training loss: 0.16565500557422638 - MAE: 0.30880801031262106
Validation loss : 0.16896353165308634 - MAE: 0.3154897166767324
Epoch: 35
Training loss: 0.16480786919593812 - MAE: 0.3066865369170677
Validation loss : 0.16894808411598206 - MAE: 0.315587755169499
Epoch: 36
Training loss: 0.16488760769367217 - MAE: 0.3062544913178101
Validation loss : 0.1683160811662674 - MAE: 0.3147328218692977
Epoch: 37
Training loss: 0.16296948552131651 - MAE: 0.30529006908692957
Validation loss : 0.16822955012321472 - MAE: 0.31474619083221317
Epoch: 38
Training loss: 0.16405799746513366 - MAE: 0.3065666077277542
Validation loss : 0.16793323556582132 - MAE: 0.3143827065817725
Epoch: 39
Training loss: 0.16395046710968017 - MAE: 0.3061178662220702
Validation loss : 0.16756848825348747 - MAE: 0.31392760086070465
Epoch: 40
Training loss: 0.16225593984127046 - MAE: 0.30391575924794567
Validation loss : 0.16728156473901537 - MAE: 0.31359587016220536
Epoch: 41
Training loss: 0.162018803358078 - MAE: 0.3050126817417
Validation loss : 0.16726960241794586 - MAE: 0.3136659181800001
Epoch: 42
Training loss: 0.1616269725561142 - MAE: 0.3047072823079763
Validation loss : 0.1666801455948088 - MAE: 0.31286053951971926
Epoch: 43
Training loss: 0.16371312737464905 - MAE: 0.3072663025489704
Validation loss : 0.16660702228546143 - MAE: 0.3128544246563194
Epoch: 44
Training loss: 0.16274581968784332 - MAE: 0.3055276315731144
Validation loss : 0.16622301439444223 - MAE: 0.3123737118409309
Epoch: 45
Training loss: 0.16198877573013307 - MAE: 0.30387318215935566
Validation loss : 0.16592974132961696 - MAE: 0.3120266870766228
Epoch: 46
Training loss: 0.1613679423928261 - MAE: 0.30285112156406757
Validation loss : 0.16592562860912746 - MAE: 0.3120895779099974
Epoch: 47
Training loss: 0.1596740049123764 - MAE: 0.30196700285423644
Validation loss : 0.1656825906700558 - MAE: 0.3117942734701369
Epoch: 48
Training loss: 0.16010645925998687 - MAE: 0.3010781145976006
Validation loss : 0.1654240306880739 - MAE: 0.3114403430988788
Epoch: 49
Training loss: 0.16135582566261292 - MAE: 0.3033602255531342
Validation loss : 0.16523137191931406 - MAE: 0.3111766545135443
Epoch: 50
Training loss: 0.15887635439634323 - MAE: 0.30139360974455354
Validation loss : 0.16520478659205967 - MAE: 0.31118266518707904
Epoch: 51
Training loss: 0.1623841804265976 - MAE: 0.3041726753025802
Validation loss : 0.16488962206575605 - MAE: 0.3107705042403597
Epoch: 52
Training loss: 0.15951467990875245 - MAE: 0.3024485706726396
Validation loss : 0.16450322336620754 - MAE: 0.3101646703863994
Epoch: 53
Training loss: 0.1593264466524124 - MAE: 0.3019928035241633
Validation loss : 0.16450072824954987 - MAE: 0.3102381133991464
Epoch: 54
Training loss: 0.15892296969890596 - MAE: 0.3010730142919314
Validation loss : 0.1644514716333813 - MAE: 0.31025052996699415
Epoch: 55
Training loss: 0.15801858335733412 - MAE: 0.300129276806219
Validation loss : 0.1639546040031645 - MAE: 0.30954432283469774
Epoch: 56
Training loss: 0.15852347761392593 - MAE: 0.30000137034695895
Validation loss : 0.16394910713036856 - MAE: 0.3096252551544829
Epoch: 57
Training loss: 0.15989623069763184 - MAE: 0.3008597511380722
Validation loss : 0.1637496915128496 - MAE: 0.30938490975556004
Epoch: 58
Training loss: 0.15900149762630464 - MAE: 0.30091532233302887
Validation loss : 0.16357198854287466 - MAE: 0.3091647551750498
Epoch: 59
Training loss: 0.15815627008676528 - MAE: 0.2998753613511919
Validation loss : 0.16364828579955631 - MAE: 0.30934114039380045
Epoch: 60
Training loss: 0.15863654345273973 - MAE: 0.2997401767468425
Validation loss : 0.16360284553633797 - MAE: 0.3093700140804052
Epoch: 61
Training loss: 0.1569861376285553 - MAE: 0.3001724729244438
Validation loss : 0.1631663309203254 - MAE: 0.3087389482069802
Epoch: 62
Training loss: 0.15764495313167573 - MAE: 0.29931992861180673
Validation loss : 0.1628624697526296 - MAE: 0.3082940994455321
Epoch: 63
Training loss: 0.15620498329401017 - MAE: 0.2983585146384368
Validation loss : 0.16287368204858568 - MAE: 0.30839281732122203
Epoch: 64
Training loss: 0.15624627590179443 - MAE: 0.29862051694345537
Validation loss : 0.16266994840568966 - MAE: 0.308149927566448
Epoch: 65
Training loss: 0.15566111862659454 - MAE: 0.29753564913606617
Validation loss : 0.1626807086997562 - MAE: 0.30822756782090394
Epoch: 66
Training loss: 0.15647490918636323 - MAE: 0.29872063729818527
Validation loss : 0.16234126521481407 - MAE: 0.3077123779358625
Epoch: 67
Training loss: 0.15619896858930588 - MAE: 0.297556687519032
Validation loss : 0.16227295829190147 - MAE: 0.30769900980501297
Epoch: 68
Training loss: 0.15651787519454957 - MAE: 0.2983152167118014
Validation loss : 0.16212009721332127 - MAE: 0.30753654402945435
Epoch: 69
Training loss: 0.15651776432991027 - MAE: 0.299202365307033
Validation loss : 0.16212468677096897 - MAE: 0.3076282798695557
Epoch: 70
Training loss: 0.15673158675432206 - MAE: 0.29856837209150694
Validation loss : 0.16198681791623434 - MAE: 0.30741784086039886
Epoch: 71
Training loss: 0.15566783308982848 - MAE: 0.2983743266636147
Validation loss : 0.1617990599738227 - MAE: 0.3071957299555147
Epoch: 72
Training loss: 0.1555081295967102 - MAE: 0.296568577820393
Validation loss : 0.16165704197353786 - MAE: 0.30704654154909133
Epoch: 73
Training loss: 0.1559882402420044 - MAE: 0.2972687981178256
Validation loss : 0.16142904261747995 - MAE: 0.3066734296802609
Epoch: 74
Training loss: 0.1546381199359894 - MAE: 0.29657499645115093
Validation loss : 0.1616316239039103 - MAE: 0.30704220168092655
Epoch: 75
Training loss: 0.15564891695976257 - MAE: 0.29796052691901004
Validation loss : 0.16111160814762115 - MAE: 0.30627361301565775
Epoch: 76
Training loss: 0.15516877621412278 - MAE: 0.29683828489531877