home-standard-results / 07-05-2023_nb-bert-base_batchsize_256_all_units_0.0001_1_200_freeze_True_earlystop_3.txt
ececet's picture
commit files to HF hub
3a2942b
Started at: 15:31:13
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 1.2954848909378052 - MAE: 1.050244734015666
Validation loss : 0.8470989664395651 - MAE: 0.8281701794455048
Epoch: 1
Training loss: 0.4445240414142609 - MAE: 0.5411769913687082
Validation loss : 0.324721054898368 - MAE: 0.43869614013342645
Epoch: 2
Training loss: 0.25153143346309664 - MAE: 0.3816801249414825
Validation loss : 0.23877331614494324 - MAE: 0.3700370086786411
Epoch: 3
Training loss: 0.23011713743209838 - MAE: 0.36750653866750815
Validation loss : 0.22173692286014557 - MAE: 0.3575197016352192
Epoch: 4
Training loss: 0.21643910229206084 - MAE: 0.35473120380428824
Validation loss : 0.2111869735850228 - MAE: 0.3495476388776553
Epoch: 5
Training loss: 0.20841679513454436 - MAE: 0.3475277955848527
Validation loss : 0.2027071151468489 - MAE: 0.3434146884806285
Epoch: 6
Training loss: 0.20075951278209686 - MAE: 0.3414739466792711
Validation loss : 0.19626588291592068 - MAE: 0.3390224698180928
Epoch: 7
Training loss: 0.1959322625398636 - MAE: 0.33718433600506764
Validation loss : 0.19155074821578133 - MAE: 0.33601220203771287
Epoch: 8
Training loss: 0.19064979493618012 - MAE: 0.33252034985072276
Validation loss : 0.18807757563061184 - MAE: 0.33369018106396875
Epoch: 9
Training loss: 0.18758945107460023 - MAE: 0.3310842562717609
Validation loss : 0.18544620275497437 - MAE: 0.33179714328905824
Epoch: 10
Training loss: 0.18699397385120392 - MAE: 0.3293496073915552
Validation loss : 0.1833840227789349 - MAE: 0.3305601596359932
Epoch: 11
Training loss: 0.1838909602165222 - MAE: 0.326863464187819
Validation loss : 0.18136033746931288 - MAE: 0.32891395485147645
Epoch: 12
Training loss: 0.18286386430263518 - MAE: 0.325863773872513
Validation loss : 0.17971531218952602 - MAE: 0.32756646233438397
Epoch: 13
Training loss: 0.1797239762544632 - MAE: 0.32309299652089063
Validation loss : 0.17807276712523568 - MAE: 0.32608809156272295
Epoch: 14
Training loss: 0.17857051372528077 - MAE: 0.3213261058182539
Validation loss : 0.1767602562904358 - MAE: 0.3250377233246472
Epoch: 15
Training loss: 0.17911633133888244 - MAE: 0.3217271718139446
Validation loss : 0.1751896556880739 - MAE: 0.3234474552955444
Epoch: 16
Training loss: 0.17600405514240264 - MAE: 0.3189804320611656
Validation loss : 0.17363134854369694 - MAE: 0.3217130704298945
Epoch: 17
Training loss: 0.17391868591308593 - MAE: 0.3165747054747144
Validation loss : 0.17256545523802438 - MAE: 0.3208692052839463
Epoch: 18
Training loss: 0.17313318729400634 - MAE: 0.3169905761419007
Validation loss : 0.17127088208993277 - MAE: 0.3194913697705306
Epoch: 19
Training loss: 0.17114326536655425 - MAE: 0.31336970082229104
Validation loss : 0.17016845610406664 - MAE: 0.31830314372899243
Epoch: 20
Training loss: 0.1727628666162491 - MAE: 0.31565258354376935
Validation loss : 0.16911302506923676 - MAE: 0.31717762732632004
Epoch: 21
Training loss: 0.17012356221675873 - MAE: 0.31316639736858143
Validation loss : 0.1677489462825987 - MAE: 0.31546195651667186
Epoch: 22
Training loss: 0.16859103083610535 - MAE: 0.3127057693660738
Validation loss : 0.16697113381491768 - MAE: 0.31471562544725196
Epoch: 23
Training loss: 0.1686873698234558 - MAE: 0.311427263715106
Validation loss : 0.16612334880563948 - MAE: 0.3138178950174241
Epoch: 24
Training loss: 0.16640773981809617 - MAE: 0.3095450918700528
Validation loss : 0.16508907741970485 - MAE: 0.3125503103864031
Epoch: 25
Training loss: 0.1658923715353012 - MAE: 0.3082243922152051
Validation loss : 0.16451590259869894 - MAE: 0.3120238609455365
Epoch: 26
Training loss: 0.16436728775501253 - MAE: 0.3071946399101465
Validation loss : 0.16374372442563376 - MAE: 0.3111470200747217
Epoch: 27
Training loss: 0.16651157915592193 - MAE: 0.30963943402408695
Validation loss : 0.16270800100432503 - MAE: 0.3097923430317113
Epoch: 28
Training loss: 0.16491072356700898 - MAE: 0.3070697425803366
Validation loss : 0.1623369190427992 - MAE: 0.3095072696428577
Epoch: 29
Training loss: 0.16389427542686463 - MAE: 0.30555235896137595
Validation loss : 0.16124575005637276 - MAE: 0.308045409739515
Epoch: 30
Training loss: 0.16372050642967223 - MAE: 0.3068781009000078
Validation loss : 0.1606789148516125 - MAE: 0.3074341210225871
Epoch: 31
Training loss: 0.16316969752311705 - MAE: 0.3060961831725536
Validation loss : 0.16020016206635368 - MAE: 0.3069771875888861
Epoch: 32
Training loss: 0.1627587217092514 - MAE: 0.30426592400661
Validation loss : 0.15954788691467708 - MAE: 0.3061654847578481
Epoch: 33
Training loss: 0.16183113038539887 - MAE: 0.30331386520448744
Validation loss : 0.15876509911484188 - MAE: 0.30512034015868056