File size: 5,296 Bytes
841080a 53ac994 d876328 a667809 b252260 36060c1 b970158 dfb5123 9ab82d4 e544147 af506e4 cad50a0 990e359 0965f2d 5232ecc 162f5ae 9bd8049 1245df1 40db2a9 f1861e8 ca66ff9 90a84d3 a0e5fd9 7753880 76cd010 6988c4e da01ac6 d5d07e2 17fa806 bd661cc a75bb70 1a76d41 b130ddf 12bb8d9 645b211 2715ef8 4418613 3c51552 acbbf33 fdc846d c5d66e9 79fd51d cd09b45 7cb372e 93d3961 7921e72 9bd6a81 e18d985 afb3eed 012a19d 0a3911e 80f044a 20a81b2 cd6e805 10888ce 0caf18f 46a2c3e aa8b0de 1e5ea82 bc05418 18b0294 eef8046 a0547fe a22e795 1476aa3 efae211 b6fe52b fdd38aa 62e595e d129992 05e35d6 7135bf6 c87ee6b 2c86dfc bcadaae 06c7ca0 a63fe2f cb2a9fa f8a6bff fd2e4f1 dd00538 9de94e7 72a0310 897bef8 dc3df3b 757aabf 8811f92 555e0fd fd5c3cb c395f7c a2d5860 ca65a32 44f4c55 2ef0bfb 19eac4f 0cd6ec1 7f40813 ca7bb25 ea17e7c 44ebf93 b5bb883 8238019 32c7b3a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 |
Started at: 15:31:13 ({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {}) Epoch: 0 Training loss: 1.2954848909378052 - MAE: 1.050244734015666 Validation loss : 0.8470989664395651 - MAE: 0.8281701794455048 Epoch: 1 Training loss: 0.4445240414142609 - MAE: 0.5411769913687082 Validation loss : 0.324721054898368 - MAE: 0.43869614013342645 Epoch: 2 Training loss: 0.25153143346309664 - MAE: 0.3816801249414825 Validation loss : 0.23877331614494324 - MAE: 0.3700370086786411 Epoch: 3 Training loss: 0.23011713743209838 - MAE: 0.36750653866750815 Validation loss : 0.22173692286014557 - MAE: 0.3575197016352192 Epoch: 4 Training loss: 0.21643910229206084 - MAE: 0.35473120380428824 Validation loss : 0.2111869735850228 - MAE: 0.3495476388776553 Epoch: 5 Training loss: 0.20841679513454436 - MAE: 0.3475277955848527 Validation loss : 0.2027071151468489 - MAE: 0.3434146884806285 Epoch: 6 Training loss: 0.20075951278209686 - MAE: 0.3414739466792711 Validation loss : 0.19626588291592068 - MAE: 0.3390224698180928 Epoch: 7 Training loss: 0.1959322625398636 - MAE: 0.33718433600506764 Validation loss : 0.19155074821578133 - MAE: 0.33601220203771287 Epoch: 8 Training loss: 0.19064979493618012 - MAE: 0.33252034985072276 Validation loss : 0.18807757563061184 - MAE: 0.33369018106396875 Epoch: 9 Training loss: 0.18758945107460023 - MAE: 0.3310842562717609 Validation loss : 0.18544620275497437 - MAE: 0.33179714328905824 Epoch: 10 Training loss: 0.18699397385120392 - MAE: 0.3293496073915552 Validation loss : 0.1833840227789349 - MAE: 0.3305601596359932 Epoch: 11 Training loss: 0.1838909602165222 - MAE: 0.326863464187819 Validation loss : 0.18136033746931288 - MAE: 0.32891395485147645 Epoch: 12 Training loss: 0.18286386430263518 - MAE: 0.325863773872513 Validation loss : 0.17971531218952602 - MAE: 0.32756646233438397 Epoch: 13 Training loss: 0.1797239762544632 - MAE: 0.32309299652089063 Validation loss : 0.17807276712523568 - MAE: 0.32608809156272295 Epoch: 14 Training loss: 0.17857051372528077 - MAE: 0.3213261058182539 Validation loss : 0.1767602562904358 - MAE: 0.3250377233246472 Epoch: 15 Training loss: 0.17911633133888244 - MAE: 0.3217271718139446 Validation loss : 0.1751896556880739 - MAE: 0.3234474552955444 Epoch: 16 Training loss: 0.17600405514240264 - MAE: 0.3189804320611656 Validation loss : 0.17363134854369694 - MAE: 0.3217130704298945 Epoch: 17 Training loss: 0.17391868591308593 - MAE: 0.3165747054747144 Validation loss : 0.17256545523802438 - MAE: 0.3208692052839463 Epoch: 18 Training loss: 0.17313318729400634 - MAE: 0.3169905761419007 Validation loss : 0.17127088208993277 - MAE: 0.3194913697705306 Epoch: 19 Training loss: 0.17114326536655425 - MAE: 0.31336970082229104 Validation loss : 0.17016845610406664 - MAE: 0.31830314372899243 Epoch: 20 Training loss: 0.1727628666162491 - MAE: 0.31565258354376935 Validation loss : 0.16911302506923676 - MAE: 0.31717762732632004 Epoch: 21 Training loss: 0.17012356221675873 - MAE: 0.31316639736858143 Validation loss : 0.1677489462825987 - MAE: 0.31546195651667186 Epoch: 22 Training loss: 0.16859103083610535 - MAE: 0.3127057693660738 Validation loss : 0.16697113381491768 - MAE: 0.31471562544725196 Epoch: 23 Training loss: 0.1686873698234558 - MAE: 0.311427263715106 Validation loss : 0.16612334880563948 - MAE: 0.3138178950174241 Epoch: 24 Training loss: 0.16640773981809617 - MAE: 0.3095450918700528 Validation loss : 0.16508907741970485 - MAE: 0.3125503103864031 Epoch: 25 Training loss: 0.1658923715353012 - MAE: 0.3082243922152051 Validation loss : 0.16451590259869894 - MAE: 0.3120238609455365 Epoch: 26 Training loss: 0.16436728775501253 - MAE: 0.3071946399101465 Validation loss : 0.16374372442563376 - MAE: 0.3111470200747217 Epoch: 27 Training loss: 0.16651157915592193 - MAE: 0.30963943402408695 Validation loss : 0.16270800100432503 - MAE: 0.3097923430317113 Epoch: 28 Training loss: 0.16491072356700898 - MAE: 0.3070697425803366 Validation loss : 0.1623369190427992 - MAE: 0.3095072696428577 Epoch: 29 Training loss: 0.16389427542686463 - MAE: 0.30555235896137595 Validation loss : 0.16124575005637276 - MAE: 0.308045409739515 Epoch: 30 Training loss: 0.16372050642967223 - MAE: 0.3068781009000078 Validation loss : 0.1606789148516125 - MAE: 0.3074341210225871 Epoch: 31 Training loss: 0.16316969752311705 - MAE: 0.3060961831725536 Validation loss : 0.16020016206635368 - MAE: 0.3069771875888861 Epoch: 32 Training loss: 0.1627587217092514 - MAE: 0.30426592400661 Validation loss : 0.15954788691467708 - MAE: 0.3061654847578481 Epoch: 33 Training loss: 0.16183113038539887 - MAE: 0.30331386520448744 |