Kielak2's picture
End of training
2a1156f verified
|
raw
history blame
3.24 kB
metadata
tags:
  - generated_from_trainer
model-index:
  - name: calculator_model_test
    results: []

calculator_model_test

This model is a fine-tuned version of on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7602

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.001
  • train_batch_size: 64
  • eval_batch_size: 64
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 40
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss
1.6045 1.0 3 1.5511
1.3068 2.0 6 1.6045
1.4803 3.0 9 1.3825
1.3351 4.0 12 1.3222
1.259 5.0 15 1.4317
1.237 6.0 18 1.3714
1.2416 7.0 21 1.3284
1.1719 8.0 24 1.2900
1.1413 9.0 27 1.2300
1.0976 10.0 30 1.2427
1.1137 11.0 33 1.1367
1.0512 12.0 36 1.0890
1.0176 13.0 39 1.1537
1.0737 14.0 42 1.1180
0.9744 15.0 45 1.0935
0.9448 16.0 48 0.9942
0.9564 17.0 51 1.0161
0.9895 18.0 54 0.9486
0.9088 19.0 57 0.9694
0.9169 20.0 60 0.9542
0.8814 21.0 63 0.9105
0.8952 22.0 66 0.9043
0.8635 23.0 69 0.8930
0.8357 24.0 72 0.9285
0.8266 25.0 75 0.8613
0.8271 26.0 78 0.9072
0.7921 27.0 81 0.8674
0.775 28.0 84 0.8476
0.7766 29.0 87 0.8600
0.7682 30.0 90 0.8438
0.7844 31.0 93 0.8022
0.7114 32.0 96 0.8131
0.7391 33.0 99 0.7957
0.7347 34.0 102 0.7810
0.7104 35.0 105 0.7740
0.7248 36.0 108 0.7665
0.7359 37.0 111 0.7819
0.7358 38.0 114 0.7668
0.7235 39.0 117 0.7601
0.7203 40.0 120 0.7602

Framework versions

  • Transformers 4.38.1
  • Pytorch 2.1.0+cu121
  • Datasets 2.18.0
  • Tokenizers 0.15.2