Edit model card

GUE_splice_reconstructed-seqsight_8192_512_30M-L32_all

This model is a fine-tuned version of mahdibaghbanzadeh/seqsight_8192_512_30M on the mahdibaghbanzadeh/GUE_splice_reconstructed dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9664
  • F1 Score: 0.7014
  • Accuracy: 0.7030

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0005
  • train_batch_size: 2048
  • eval_batch_size: 2048
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • training_steps: 10000

Training results

Training Loss Epoch Step Validation Loss F1 Score Accuracy
0.9343 11.11 200 0.8346 0.5809 0.6241
0.7818 22.22 400 0.7676 0.6464 0.6611
0.7135 33.33 600 0.7563 0.6612 0.6620
0.6651 44.44 800 0.7568 0.6691 0.6710
0.627 55.56 1000 0.7487 0.6738 0.6767
0.5948 66.67 1200 0.7433 0.6800 0.6850
0.5688 77.78 1400 0.7494 0.6815 0.6870
0.5437 88.89 1600 0.7587 0.6811 0.6815
0.5212 100.0 1800 0.7816 0.6787 0.6789
0.5014 111.11 2000 0.7846 0.6859 0.6857
0.4824 122.22 2200 0.7872 0.6852 0.6859
0.4664 133.33 2400 0.7993 0.6910 0.6951
0.4508 144.44 2600 0.8120 0.6901 0.6949
0.4369 155.56 2800 0.8342 0.6886 0.6898
0.4241 166.67 3000 0.8359 0.6855 0.6872
0.41 177.78 3200 0.8491 0.6895 0.6903
0.3981 188.89 3400 0.8528 0.6881 0.6874
0.3863 200.0 3600 0.8736 0.6874 0.6914
0.3748 211.11 3800 0.8668 0.6905 0.6940
0.3662 222.22 4000 0.8681 0.6891 0.6896
0.355 233.33 4200 0.8869 0.6923 0.6942
0.3465 244.44 4400 0.8918 0.6886 0.6911
0.3395 255.56 4600 0.9159 0.6880 0.6887
0.3315 266.67 4800 0.9279 0.6934 0.6953
0.3231 277.78 5000 0.9232 0.6917 0.6927
0.3162 288.89 5200 0.9350 0.6913 0.6925
0.3108 300.0 5400 0.9520 0.6959 0.6979
0.3042 311.11 5600 0.9396 0.6899 0.6918
0.3 322.22 5800 0.9521 0.6927 0.6982
0.2898 333.33 6000 0.9616 0.6928 0.6951
0.2887 344.44 6200 0.9716 0.6937 0.6960
0.2819 355.56 6400 0.9720 0.6886 0.6876
0.2774 366.67 6600 0.9822 0.6907 0.6918
0.2727 377.78 6800 0.9960 0.6888 0.6907
0.2694 388.89 7000 0.9855 0.6936 0.6957
0.2635 400.0 7200 0.9976 0.6907 0.6918
0.261 411.11 7400 1.0112 0.6915 0.6938
0.2582 422.22 7600 1.0083 0.6880 0.6885
0.2557 433.33 7800 1.0171 0.6923 0.6936
0.2534 444.44 8000 1.0165 0.6962 0.6977
0.2498 455.56 8200 1.0167 0.6922 0.6929
0.2468 466.67 8400 1.0320 0.6914 0.6927
0.2447 477.78 8600 1.0256 0.6906 0.6916
0.2424 488.89 8800 1.0227 0.6895 0.6903
0.2401 500.0 9000 1.0330 0.6923 0.6938
0.2397 511.11 9200 1.0332 0.6920 0.6933
0.2387 522.22 9400 1.0357 0.6937 0.6951
0.2369 533.33 9600 1.0390 0.6926 0.6940
0.236 544.44 9800 1.0405 0.6924 0.6933
0.2364 555.56 10000 1.0352 0.6910 0.6920

Framework versions

  • PEFT 0.9.0
  • Transformers 4.38.2
  • Pytorch 2.2.0+cu121
  • Datasets 2.17.1
  • Tokenizers 0.15.2
Downloads last month
0
Unable to determine this model’s pipeline type. Check the docs .
Invalid base_model specified in model card metadata. Needs to be a model id from hf.co/models.