Edit model card

GUE_EMP_H3-seqsight_16384_512_56M-L32_f

This model is a fine-tuned version of mahdibaghbanzadeh/seqsight_16384_512_56M on the mahdibaghbanzadeh/GUE_EMP_H3 dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2646
  • F1 Score: 0.8951
  • Accuracy: 0.8951

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0005
  • train_batch_size: 128
  • eval_batch_size: 128
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • training_steps: 10000

Training results

Training Loss Epoch Step Validation Loss F1 Score Accuracy
0.3885 2.13 200 0.3562 0.8477 0.8477
0.2813 4.26 400 0.3305 0.8675 0.8677
0.2493 6.38 600 0.3649 0.8522 0.8524
0.2349 8.51 800 0.3031 0.8838 0.8838
0.2193 10.64 1000 0.3812 0.8577 0.8584
0.2032 12.77 1200 0.3416 0.8764 0.8764
0.1925 14.89 1400 0.3750 0.8708 0.8711
0.1779 17.02 1600 0.3903 0.8597 0.8597
0.1674 19.15 1800 0.3564 0.8724 0.8724
0.1489 21.28 2000 0.4619 0.8612 0.8617
0.1423 23.4 2200 0.4485 0.8735 0.8737
0.1215 25.53 2400 0.4759 0.8784 0.8784
0.1185 27.66 2600 0.5499 0.8436 0.8444
0.0993 29.79 2800 0.5338 0.8520 0.8524
0.0962 31.91 3000 0.5457 0.8514 0.8517
0.0823 34.04 3200 0.5406 0.8577 0.8577
0.0787 36.17 3400 0.6370 0.8559 0.8564
0.0708 38.3 3600 0.6247 0.8574 0.8577
0.0674 40.43 3800 0.6834 0.8478 0.8484
0.057 42.55 4000 0.8145 0.8462 0.8470
0.0536 44.68 4200 0.7901 0.8400 0.8410
0.0505 46.81 4400 0.7505 0.8659 0.8664
0.0463 48.94 4600 0.7752 0.8490 0.8497
0.0449 51.06 4800 0.7215 0.8601 0.8604
0.0384 53.19 5000 0.8821 0.8376 0.8383
0.0351 55.32 5200 0.9139 0.8465 0.8470
0.0349 57.45 5400 0.9360 0.8387 0.8397
0.0361 59.57 5600 0.8710 0.8575 0.8577
0.0308 61.7 5800 0.8229 0.8597 0.8597
0.0294 63.83 6000 0.9199 0.8517 0.8524
0.0293 65.96 6200 0.8718 0.8588 0.8591
0.0271 68.09 6400 0.8787 0.8617 0.8617
0.0238 70.21 6600 0.9513 0.8581 0.8584
0.0241 72.34 6800 0.9352 0.8629 0.8631
0.0225 74.47 7000 0.9943 0.8548 0.8550
0.0231 76.6 7200 0.9241 0.8602 0.8604
0.0204 78.72 7400 1.0017 0.8622 0.8624
0.0206 80.85 7600 1.0763 0.8498 0.8504
0.0182 82.98 7800 1.0418 0.8575 0.8577
0.0166 85.11 8000 1.0393 0.8567 0.8570
0.0172 87.23 8200 1.0861 0.8492 0.8497
0.0167 89.36 8400 1.1617 0.8470 0.8477
0.015 91.49 8600 1.0801 0.8621 0.8624
0.0151 93.62 8800 1.1022 0.8541 0.8544
0.014 95.74 9000 1.1847 0.8438 0.8444
0.0125 97.87 9200 1.1438 0.8534 0.8537
0.0131 100.0 9400 1.1487 0.8554 0.8557
0.0121 102.13 9600 1.1538 0.8533 0.8537
0.0124 104.26 9800 1.1753 0.8513 0.8517
0.0121 106.38 10000 1.1525 0.8501 0.8504

Framework versions

  • PEFT 0.9.0
  • Transformers 4.38.2
  • Pytorch 2.2.0+cu121
  • Datasets 2.17.1
  • Tokenizers 0.15.2
Downloads last month
0
Unable to determine this model’s pipeline type. Check the docs .
Invalid base_model specified in model card metadata. Needs to be a model id from hf.co/models.