File size: 2,027 Bytes
ffdff51 ca93d39 8a9dc10 54e7f6f 8344db1 38f51a1 480fedd f82f0a4 c45369e 5bcb629 96d8a52 f9c61ff 05f907e a04b204 d682b41 656c0a1 30b5648 242651e e148f83 7233b82 67b6d25 c14add0 2ee0943 ab40d12 df9b96f 4bc1456 ed075b0 e5cdd3f 4b32231 f3329cf f98b900 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 |
Started at: 14:47:29 nb-bert-base, 5e-06, 128 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 0.46638986387036063 - MAE: 0.5398754368007034 Validation loss : 0.2809767316322069 - MAE: 0.4124626148131228 Epoch: 1 Training loss: 0.21324598220261662 - MAE: 0.36342001646930033 Validation loss : 0.17091604342331757 - MAE: 0.32559828894035 Epoch: 2 Training loss: 0.15148694582960823 - MAE: 0.3101711238743909 Validation loss : 0.14158385267128815 - MAE: 0.29382737678226023 Epoch: 3 Training loss: 0.124982850253582 - MAE: 0.2813409221247067 Validation loss : 0.11344313339607136 - MAE: 0.25976834796862297 Epoch: 4 Training loss: 0.10359308848326856 - MAE: 0.25381723494987607 Validation loss : 0.10265538237384848 - MAE: 0.24230259616167318 Epoch: 5 Training loss: 0.08948654772883112 - MAE: 0.23498061801274692 Validation loss : 0.09245944788327089 - MAE: 0.2265152617992811 Epoch: 6 Training loss: 0.07607155100188472 - MAE: 0.2157861911084654 Validation loss : 0.08652295635358707 - MAE: 0.2184979559256568 Epoch: 7 Training loss: 0.06690294654531913 - MAE: 0.2011048192003143 Validation loss : 0.08006738347781671 - MAE: 0.20853752498405181 Epoch: 8 Training loss: 0.05949505604803562 - MAE: 0.1890938053594342 Validation loss : 0.06948294661737778 - MAE: 0.1920086602951222 Epoch: 9 |