--- tags: - generated_from_trainer model-index: - name: dna_bert_3_1000seq-finetuned results: [] --- # dna_bert_3_1000seq-finetuned This model is a fine-tuned version of [armheb/DNA_bert_3](https://huggingface.co/armheb/DNA_bert_3) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.4684 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 100 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.8607 | 1.0 | 62 | 0.6257 | | 0.6177 | 2.0 | 124 | 0.6120 | | 0.6098 | 3.0 | 186 | 0.6062 | | 0.604 | 4.0 | 248 | 0.6052 | | 0.5999 | 5.0 | 310 | 0.6040 | | 0.5982 | 6.0 | 372 | 0.5996 | | 0.5985 | 7.0 | 434 | 0.5985 | | 0.5956 | 8.0 | 496 | 0.5968 | | 0.5936 | 9.0 | 558 | 0.5950 | | 0.5908 | 10.0 | 620 | 0.5941 | | 0.5904 | 11.0 | 682 | 0.5932 | | 0.59 | 12.0 | 744 | 0.5917 | | 0.5877 | 13.0 | 806 | 0.5921 | | 0.5847 | 14.0 | 868 | 0.5903 | | 0.5831 | 15.0 | 930 | 0.5887 | | 0.5852 | 16.0 | 992 | 0.5878 | | 0.5805 | 17.0 | 1054 | 0.5872 | | 0.5795 | 18.0 | 1116 | 0.5853 | | 0.5754 | 19.0 | 1178 | 0.5869 | | 0.5757 | 20.0 | 1240 | 0.5839 | | 0.5722 | 21.0 | 1302 | 0.5831 | | 0.5693 | 22.0 | 1364 | 0.5811 | | 0.5667 | 23.0 | 1426 | 0.5802 | | 0.5652 | 24.0 | 1488 | 0.5775 | | 0.5608 | 25.0 | 1550 | 0.5788 | | 0.5591 | 26.0 | 1612 | 0.5724 | | 0.5538 | 27.0 | 1674 | 0.5736 | | 0.552 | 28.0 | 1736 | 0.5689 | | 0.5483 | 29.0 | 1798 | 0.5689 | | 0.5442 | 30.0 | 1860 | 0.5671 | | 0.5405 | 31.0 | 1922 | 0.5658 | | 0.537 | 32.0 | 1984 | 0.5605 | | 0.5349 | 33.0 | 2046 | 0.5575 | | 0.5275 | 34.0 | 2108 | 0.5569 | | 0.5227 | 35.0 | 2170 | 0.5537 | | 0.52 | 36.0 | 2232 | 0.5509 | | 0.5173 | 37.0 | 2294 | 0.5504 | | 0.5123 | 38.0 | 2356 | 0.5435 | | 0.5088 | 39.0 | 2418 | 0.5472 | | 0.5037 | 40.0 | 2480 | 0.5383 | | 0.501 | 41.0 | 2542 | 0.5379 | | 0.4931 | 42.0 | 2604 | 0.5365 | | 0.4923 | 43.0 | 2666 | 0.5328 | | 0.4879 | 44.0 | 2728 | 0.5301 | | 0.482 | 45.0 | 2790 | 0.5295 | | 0.4805 | 46.0 | 2852 | 0.5261 | | 0.4772 | 47.0 | 2914 | 0.5221 | | 0.4738 | 48.0 | 2976 | 0.5234 | | 0.4674 | 49.0 | 3038 | 0.5210 | | 0.4646 | 50.0 | 3100 | 0.5169 | | 0.4621 | 51.0 | 3162 | 0.5142 | | 0.4574 | 52.0 | 3224 | 0.5129 | | 0.4552 | 53.0 | 3286 | 0.5127 | | 0.4539 | 54.0 | 3348 | 0.5124 | | 0.4506 | 55.0 | 3410 | 0.5076 | | 0.4457 | 56.0 | 3472 | 0.5082 | | 0.4454 | 57.0 | 3534 | 0.5027 | | 0.4398 | 58.0 | 3596 | 0.5019 | | 0.4386 | 59.0 | 3658 | 0.4998 | | 0.4332 | 60.0 | 3720 | 0.4970 | | 0.4277 | 61.0 | 3782 | 0.4995 | | 0.4273 | 62.0 | 3844 | 0.4962 | | 0.4235 | 63.0 | 3906 | 0.4909 | | 0.4201 | 64.0 | 3968 | 0.4913 | | 0.4198 | 65.0 | 4030 | 0.4899 | | 0.4182 | 66.0 | 4092 | 0.4919 | | 0.4157 | 67.0 | 4154 | 0.4902 | | 0.4104 | 68.0 | 4216 | 0.4881 | | 0.4095 | 69.0 | 4278 | 0.4881 | | 0.4077 | 70.0 | 4340 | 0.4861 | | 0.4064 | 71.0 | 4402 | 0.4868 | | 0.4041 | 72.0 | 4464 | 0.4826 | | 0.4029 | 73.0 | 4526 | 0.4833 | | 0.3976 | 74.0 | 4588 | 0.4819 | | 0.3997 | 75.0 | 4650 | 0.4809 | | 0.3974 | 76.0 | 4712 | 0.4801 | | 0.3953 | 77.0 | 4774 | 0.4783 | | 0.3938 | 78.0 | 4836 | 0.4775 | | 0.3934 | 79.0 | 4898 | 0.4762 | | 0.3923 | 80.0 | 4960 | 0.4742 | | 0.3893 | 81.0 | 5022 | 0.4742 | | 0.3909 | 82.0 | 5084 | 0.4740 | | 0.3856 | 83.0 | 5146 | 0.4739 | | 0.3904 | 84.0 | 5208 | 0.4740 | | 0.3883 | 85.0 | 5270 | 0.4701 | | 0.3865 | 86.0 | 5332 | 0.4727 | | 0.3809 | 87.0 | 5394 | 0.4736 | | 0.3853 | 88.0 | 5456 | 0.4704 | | 0.3821 | 89.0 | 5518 | 0.4704 | | 0.3809 | 90.0 | 5580 | 0.4701 | | 0.3814 | 91.0 | 5642 | 0.4698 | | 0.3795 | 92.0 | 5704 | 0.4702 | | 0.3804 | 93.0 | 5766 | 0.4692 | | 0.377 | 94.0 | 5828 | 0.4683 | | 0.3812 | 95.0 | 5890 | 0.4692 | | 0.3806 | 96.0 | 5952 | 0.4683 | | 0.3745 | 97.0 | 6014 | 0.4690 | | 0.3825 | 98.0 | 6076 | 0.4684 | | 0.374 | 99.0 | 6138 | 0.4687 | | 0.3795 | 100.0 | 6200 | 0.4684 | ### Framework versions - Transformers 4.21.1 - Pytorch 1.12.0+cu113 - Datasets 2.4.0 - Tokenizers 0.12.1