Started at: 11:52:44 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'afb829e3d0b861bd5f8cda6522b32ca0b097d7eb'}, {}) Epoch: 0 Training loss: 0.19169998451283105 - MSE: 0.3224822171590252 Validation loss : 0.1576854281593114 - MSE: 0.3052585763866773 Epoch: 1 Training loss: 0.170721908933238 - MSE: 0.3106345883186025 Validation loss : 0.14974205638282 - MSE: 0.30164828824445067 Epoch: 2 Training loss: 0.16939535588026047 - MSE: 0.31008999561987205 Validation loss : 0.15034097363241017 - MSE: 0.30232355450380055 Epoch: 3 Training loss: 0.1694463299293267 - MSE: 0.3101087405316537 Validation loss : 0.15061823884025216 - MSE: 0.3026353720119914 Epoch: 4 Training loss: 0.16941545233130456 - MSE: 0.310085018434928 Validation loss : 0.15079936920665205 - MSE: 0.30283750300395695 Epoch: 5 Training loss: 0.16935620966710543 - MSE: 0.31004690146169944 Validation loss : 0.15084448363631964 - MSE: 0.3028875614759272 Epoch: 6 Training loss: 0.16949052591072886 - MSE: 0.3101781524663373 Validation loss : 0.15096625522710383 - MSE: 0.30302316902771054 Epoch: 7 Training loss: 0.16949893361643742 - MSE: 0.310188779241269 Validation loss : 0.15100383502431214 - MSE: 0.30306482790138034 Epoch: 8 Training loss: 0.1694979808440334 - MSE: 0.3101921144601653 Validation loss : 0.15103243081830442 - MSE: 0.30309663850448487 Epoch: 9 Training loss: 0.1694973421724219 - MSE: 0.3101948706230913 Validation loss : 0.1510547660291195 - MSE: 0.303121549986372 Epoch: 10 Training loss: 0.16949688792228698 - MSE: 0.31019701891325113 Validation loss : 0.15107234381139278 - MSE: 0.3031413168423569 Epoch: 11 Training loss: 0.16949651574617938 - MSE: 0.3101986945843064 Validation loss : 0.15108662215061486 - MSE: 0.3031575035779497 Epoch: 12 Training loss: 0.16949623556513535 - MSE: 0.3102000477910796 Validation loss : 0.1510984308551997 - MSE: 0.30317088319088725 Epoch: 13 Training loss: 0.16949602914483924 - MSE: 0.3102011731046972 Validation loss : 0.15110818133689463 - MSE: 0.30318192262302546 Epoch: 14 Training loss: 0.16949585684035953 - MSE: 0.31020210549164046 Validation loss : 0.15111643797717988 - MSE: 0.30319126495260207 Epoch: 15 Training loss: 0.1694957421798455 - MSE: 0.3102031148284955 Validation loss : 0.15112340752966702 - MSE: 0.30319914952951876 Epoch: 16 Training loss: 0.1694956130103061 - MSE: 0.3102041142706642 Validation loss : 0.15112938289530575 - MSE: 0.3032059101165032 Epoch: 17 Training loss: 0.16949550995701237 - MSE: 0.31020503356981505 Validation loss : 0.1511345561593771 - MSE: 0.30321175591188876 Epoch: 18 Training loss: 0.16949546227329657 - MSE: 0.31020586811368406 Validation loss : 0.15113910171203315 - MSE: 0.3032168940185329 Epoch: 19 Training loss: 0.1694954017275258 - MSE: 0.3102065734353139 Validation loss : 0.15114298881962895 - MSE: 0.30322128473881094 Epoch: 20 Training loss: 0.16949532439834195 - MSE: 0.3102071649483396 Validation loss : 0.15114647196605802 - MSE: 0.30322521829612015 Epoch: 21 Training loss: 0.16949528412599313 - MSE: 0.3102077097474076 Validation loss : 0.15114951902069151 - MSE: 0.3032286566226503 Epoch: 22