File size: 2,910 Bytes
ffdff51 ca93d39 8a9dc10 54e7f6f 8344db1 38f51a1 480fedd f82f0a4 c45369e 5bcb629 96d8a52 f9c61ff 05f907e a04b204 d682b41 656c0a1 30b5648 242651e e148f83 7233b82 67b6d25 c14add0 2ee0943 ab40d12 df9b96f 4bc1456 ed075b0 e5cdd3f 4b32231 f3329cf f98b900 a6865f2 1ad68f3 a3244ec 41f96c1 ad2af4f 67df123 00e4480 87e8292 1ab51e5 c65a2e6 a64d778 9abe17c e6cbaa5 850f5b8 97b8fdb e8d10b7 56253a9 4e2e70f 36f6375 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 |
Started at: 14:47:29 nb-bert-base, 5e-06, 128 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 0.46638986387036063 - MAE: 0.5398754368007034 Validation loss : 0.2809767316322069 - MAE: 0.4124626148131228 Epoch: 1 Training loss: 0.21324598220261662 - MAE: 0.36342001646930033 Validation loss : 0.17091604342331757 - MAE: 0.32559828894035 Epoch: 2 Training loss: 0.15148694582960823 - MAE: 0.3101711238743909 Validation loss : 0.14158385267128815 - MAE: 0.29382737678226023 Epoch: 3 Training loss: 0.124982850253582 - MAE: 0.2813409221247067 Validation loss : 0.11344313339607136 - MAE: 0.25976834796862297 Epoch: 4 Training loss: 0.10359308848326856 - MAE: 0.25381723494987607 Validation loss : 0.10265538237384848 - MAE: 0.24230259616167318 Epoch: 5 Training loss: 0.08948654772883112 - MAE: 0.23498061801274692 Validation loss : 0.09245944788327089 - MAE: 0.2265152617992811 Epoch: 6 Training loss: 0.07607155100188472 - MAE: 0.2157861911084654 Validation loss : 0.08652295635358707 - MAE: 0.2184979559256568 Epoch: 7 Training loss: 0.06690294654531913 - MAE: 0.2011048192003143 Validation loss : 0.08006738347781671 - MAE: 0.20853752498405181 Epoch: 8 Training loss: 0.05949505604803562 - MAE: 0.1890938053594342 Validation loss : 0.06948294661737778 - MAE: 0.1920086602951222 Epoch: 9 Training loss: 0.05425354472615502 - MAE: 0.17931217790377968 Validation loss : 0.06705979368573911 - MAE: 0.1872916947876361 Epoch: 10 Training loss: 0.051136760142716495 - MAE: 0.17425557358320554 Validation loss : 0.07039306645055075 - MAE: 0.1957787938835992 Epoch: 11 Training loss: 0.049823644300076096 - MAE: 0.17265991846906573 Validation loss : 0.059706183905537065 - MAE: 0.17449212548911355 Epoch: 12 Training loss: 0.04213783075525002 - MAE: 0.15748476174497095 Validation loss : 0.06050218903535121 - MAE: 0.17571373887778519 Epoch: 13 Training loss: 0.03851394113153219 - MAE: 0.1512245667258023 Validation loss : 0.05904384138616356 - MAE: 0.1736026045715516 Epoch: 14 Training loss: 0.03739640076052059 - MAE: 0.14833851531891173 Validation loss : 0.05757016124757561 - MAE: 0.17082927729316882 Epoch: 15 Training loss: 0.03432362365790389 - MAE: 0.14270446314923024 |