Edit model card

robbert-2023-dutch-large-ft-nlp-xxl

This model is a fine-tuned version of Tweeties/tweety-7b-dutch-v24a on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 3.6377

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • training_steps: 2000

Training results

Training Loss Epoch Step Validation Loss
4.5391 0.0424 50 4.0393
4.0619 0.0849 100 3.9164
3.9954 0.1273 150 3.8674
3.9471 0.1698 200 3.8336
3.9038 0.2122 250 3.8172
3.9226 0.2547 300 3.7907
3.8362 0.2971 350 3.7700
3.8002 0.3396 400 3.7757
3.8245 0.3820 450 3.7469
3.7602 0.4244 500 3.7271
3.7623 0.4669 550 3.7159
3.7735 0.5093 600 3.7080
3.7444 0.5518 650 3.7013
3.7844 0.5942 700 3.6899
3.7439 0.6367 750 3.6784
3.6924 0.6791 800 3.6739
3.7393 0.7216 850 3.6651
3.7169 0.7640 900 3.6671
3.718 0.8065 950 3.6556
3.7342 0.8489 1000 3.6514
3.7327 0.8913 1050 3.6444
3.7048 0.9338 1100 3.6417
3.7166 0.9762 1150 3.6396
3.5797 1.0187 1200 3.6392
3.4572 1.0611 1250 3.6405
3.4068 1.1036 1300 3.6497
3.4344 1.1460 1350 3.6510
3.4347 1.1885 1400 3.6452
3.3956 1.2309 1450 3.6482
3.4001 1.2733 1500 3.6559
3.394 1.3158 1550 3.6457
3.421 1.3582 1600 3.6468
3.4062 1.4007 1650 3.6457
3.3814 1.4431 1700 3.6435
3.3415 1.4856 1750 3.6458
3.3699 1.5280 1800 3.6402
3.3902 1.5705 1850 3.6390
3.3718 1.6129 1900 3.6382
3.4026 1.6553 1950 3.6380
3.3431 1.6978 2000 3.6377

Framework versions

  • PEFT 0.11.1
  • Transformers 4.41.2
  • Pytorch 2.1.1+cu121
  • Datasets 2.19.2
  • Tokenizers 0.19.1
Downloads last month
1
Unable to determine this model’s pipeline type. Check the docs .

Adapter for