File size: 1,327 Bytes
30d11eb
b4cd124
8923fc8
0696b08
9af260c
4cccc11
a0db37a
26964eb
1e60df2
7952152
efbc86a
d212fdc
c52a128
61608eb
d8d9a8c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
Started at: 10:14:01
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.2541003803713153 - MAE: 0.3777772562039763
Validation loss : 0.1611750049121452 - MAE: 0.3084891751906978
Epoch: 1
Training loss: 0.16133709477655817 - MAE: 0.3044365154854021
Validation loss : 0.1503054533492435 - MAE: 0.2940590320068253
Epoch: 2
Training loss: 0.15391364648486627 - MAE: 0.2958381596812475
Validation loss : 0.1500918763605031 - MAE: 0.2909181862900977
Epoch: 3
Training loss: 0.14992259344970338 - MAE: 0.29211136159093676
Validation loss : 0.15011596024939508 - MAE: 0.2891221351645873
Epoch: 4