Edit model card

DarkBERT-finetuned-ner

This model is a fine-tuned version of s2w-ai/DarkBERT on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.3864
  • Precision: 0.2772
  • Recall: 0.3552
  • F1: 0.3114
  • Accuracy: 0.8673

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-06
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 16
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 40

Training results

Training Loss Epoch Step Validation Loss Precision Recall F1 Accuracy
No log 0.9950 100 1.2142 0.1429 0.0011 0.0022 0.8479
No log 2.0 201 0.6880 0.0 0.0 0.0 0.8482
No log 2.9950 301 0.6185 0.0483 0.0076 0.0131 0.8522
No log 4.0 402 0.5666 0.1453 0.0542 0.0790 0.8624
0.903 4.9950 502 0.5301 0.1633 0.0792 0.1066 0.8650
0.903 6.0 603 0.5085 0.1892 0.1258 0.1511 0.8674
0.903 6.9950 703 0.4901 0.2038 0.1518 0.1740 0.8684
0.903 8.0 804 0.4668 0.2222 0.1670 0.1907 0.8721
0.903 8.9950 904 0.4513 0.2301 0.1822 0.2034 0.8744
0.5374 10.0 1005 0.4432 0.2271 0.2039 0.2149 0.8748
0.5374 10.9950 1105 0.4300 0.2457 0.2321 0.2387 0.8774
0.5374 12.0 1206 0.4181 0.2673 0.2592 0.2632 0.8807
0.5374 12.9950 1306 0.4090 0.2888 0.2928 0.2908 0.8834
0.5374 14.0 1407 0.4015 0.3086 0.3286 0.3183 0.8861
0.4589 14.9950 1507 0.3967 0.3127 0.3460 0.3285 0.8874
0.4589 16.0 1608 0.3938 0.3149 0.3709 0.3406 0.8875
0.4589 16.9950 1708 0.3837 0.3234 0.3655 0.3432 0.8910
0.4589 18.0 1809 0.3827 0.3177 0.3753 0.3441 0.8895
0.4589 18.9950 1909 0.3785 0.3181 0.3785 0.3457 0.8906
0.4201 20.0 2010 0.3785 0.3096 0.3970 0.3479 0.8886
0.4201 20.9950 2110 0.3684 0.3361 0.3926 0.3622 0.8938
0.4201 22.0 2211 0.3689 0.3218 0.3948 0.3546 0.8911
0.4201 22.9950 2311 0.3671 0.3248 0.4013 0.3590 0.8910
0.4201 24.0 2412 0.3649 0.3278 0.4046 0.3621 0.8916
0.3927 24.9950 2512 0.3609 0.3378 0.4067 0.3691 0.8942
0.3927 26.0 2613 0.3606 0.3301 0.4121 0.3666 0.8923
0.3927 26.9950 2713 0.3590 0.3307 0.4132 0.3674 0.8933
0.3927 28.0 2814 0.3607 0.3325 0.4187 0.3706 0.8912
0.3927 28.9950 2914 0.3582 0.3353 0.4230 0.3741 0.8925
0.3796 30.0 3015 0.3576 0.3362 0.4230 0.3746 0.8924
0.3796 30.9950 3115 0.3556 0.3359 0.4208 0.3736 0.8933
0.3796 32.0 3216 0.3544 0.3401 0.4208 0.3762 0.8935
0.3796 32.9950 3316 0.3548 0.3377 0.4241 0.3760 0.8931
0.3796 34.0 3417 0.3545 0.3408 0.4273 0.3792 0.8930
0.3685 34.9950 3517 0.3536 0.3447 0.4284 0.3820 0.8935
0.3685 36.0 3618 0.3531 0.3444 0.4284 0.3818 0.8935
0.3685 36.9950 3718 0.3534 0.3432 0.4295 0.3815 0.8934
0.3685 38.0 3819 0.3535 0.3458 0.4328 0.3844 0.8934
0.3685 38.9950 3919 0.3531 0.3458 0.4317 0.3840 0.8935
0.3666 39.8010 4000 0.3528 0.3467 0.4317 0.3845 0.8937

Framework versions

  • Transformers 4.41.0
  • Pytorch 2.3.0+cu121
  • Datasets 2.19.1
  • Tokenizers 0.19.1
Downloads last month
15
Safetensors
Model size
124M params
Tensor type
F32
·

Finetuned from

Space using guidobenb/DarkBERT-finetuned-ner 1