Edit model card

GUE_prom_prom_300_all-seqsight_16384_512_22M-L32_all

This model is a fine-tuned version of mahdibaghbanzadeh/seqsight_16384_512_22M on the mahdibaghbanzadeh/GUE_prom_prom_300_all dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4221
  • F1 Score: 0.8280
  • Accuracy: 0.8280

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0005
  • train_batch_size: 1536
  • eval_batch_size: 1536
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • training_steps: 10000

Training results

Training Loss Epoch Step Validation Loss F1 Score Accuracy
0.5993 6.45 200 0.5279 0.7420 0.7424
0.511 12.9 400 0.5012 0.7558 0.7568
0.4874 19.35 600 0.4878 0.7699 0.7703
0.4675 25.81 800 0.4767 0.7743 0.7752
0.4472 32.26 1000 0.4560 0.7845 0.7850
0.4209 38.71 1200 0.4419 0.7936 0.7943
0.4041 45.16 1400 0.4286 0.8064 0.8064
0.391 51.61 1600 0.4183 0.8108 0.8108
0.3826 58.06 1800 0.4144 0.8116 0.8117
0.3731 64.52 2000 0.4179 0.8139 0.8140
0.3664 70.97 2200 0.4126 0.8133 0.8135
0.36 77.42 2400 0.4184 0.8099 0.8103
0.3538 83.87 2600 0.4093 0.8168 0.8169
0.3482 90.32 2800 0.4159 0.8165 0.8166
0.3418 96.77 3000 0.4082 0.8214 0.8215
0.3369 103.23 3200 0.4192 0.8204 0.8206
0.3321 109.68 3400 0.4123 0.8200 0.8203
0.3266 116.13 3600 0.4095 0.8210 0.8211
0.3241 122.58 3800 0.4094 0.8224 0.8225
0.3213 129.03 4000 0.4024 0.8233 0.8235
0.3168 135.48 4200 0.4072 0.8249 0.825
0.3121 141.94 4400 0.4084 0.8259 0.8260
0.3107 148.39 4600 0.4125 0.8266 0.8267
0.3074 154.84 4800 0.4168 0.8231 0.8233
0.3051 161.29 5000 0.4144 0.8260 0.8262
0.3034 167.74 5200 0.4244 0.8241 0.8243
0.2992 174.19 5400 0.4163 0.8295 0.8296
0.2985 180.65 5600 0.4101 0.8268 0.8269
0.2959 187.1 5800 0.4233 0.8252 0.8253
0.2944 193.55 6000 0.4147 0.8268 0.8269
0.2926 200.0 6200 0.4145 0.8309 0.8309
0.2907 206.45 6400 0.4186 0.8252 0.8253
0.2891 212.9 6600 0.4275 0.8265 0.8267
0.288 219.35 6800 0.4174 0.8264 0.8265
0.2861 225.81 7000 0.4149 0.8270 0.8270
0.2833 232.26 7200 0.4089 0.8287 0.8287
0.2842 238.71 7400 0.4158 0.8267 0.8267
0.2828 245.16 7600 0.4135 0.8286 0.8287
0.2819 251.61 7800 0.4157 0.8272 0.8272
0.2797 258.06 8000 0.4160 0.8296 0.8296
0.2785 264.52 8200 0.4180 0.8249 0.825
0.2785 270.97 8400 0.4247 0.8269 0.8270
0.278 277.42 8600 0.4147 0.8271 0.8272
0.2767 283.87 8800 0.4157 0.8261 0.8262
0.2769 290.32 9000 0.4172 0.8249 0.825
0.2757 296.77 9200 0.4173 0.8258 0.8258
0.2763 303.23 9400 0.4180 0.8259 0.8260
0.2755 309.68 9600 0.4202 0.8269 0.8270
0.2764 316.13 9800 0.4165 0.8258 0.8258
0.2741 322.58 10000 0.4192 0.8258 0.8258

Framework versions

  • PEFT 0.9.0
  • Transformers 4.38.2
  • Pytorch 2.2.0+cu121
  • Datasets 2.17.1
  • Tokenizers 0.15.2
Downloads last month
0
Unable to determine this model’s pipeline type. Check the docs .
Invalid base_model specified in model card metadata. Needs to be a model id from hf.co/models.