pszmk's picture
End of training
b2eea4d verified
metadata
tags:
  - generated_from_trainer
model-index:
  - name: calculator_model_test
    results: []

calculator_model_test

This model is a fine-tuned version of on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7556

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.001
  • train_batch_size: 512
  • eval_batch_size: 512
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 40

Training results

Training Loss Epoch Step Validation Loss
3.1456 1.0 6 2.5517
2.1398 2.0 12 1.8461
1.7232 3.0 18 1.6222
1.6435 4.0 24 1.6869
1.6417 5.0 30 1.5262
1.4634 6.0 36 1.4824
1.4507 7.0 42 1.8991
1.54 8.0 48 1.4370
1.483 9.0 54 1.4203
1.4086 10.0 60 1.4349
1.4063 11.0 66 1.5677
1.4235 12.0 72 1.3646
1.2839 13.0 78 1.2074
1.2127 14.0 84 1.2050
1.1447 15.0 90 1.1797
1.1515 16.0 96 1.1440
1.1138 17.0 102 1.1293
1.07 18.0 108 1.1061
1.0295 19.0 114 1.1572
1.1086 20.0 120 1.3547
1.1835 21.0 126 1.1306
1.0968 22.0 132 1.0152
1.0012 23.0 138 1.0467
1.0088 24.0 144 0.9831
0.9616 25.0 150 0.9141
0.9385 26.0 156 0.9093
0.8887 27.0 162 0.9188
0.9549 28.0 168 0.9369
0.8972 29.0 174 0.8518
0.8973 30.0 180 0.8424
0.8419 31.0 186 0.8858
0.8887 32.0 192 0.8781
0.8509 33.0 198 0.8372
0.8473 34.0 204 0.8021
0.8361 35.0 210 0.7912
0.8037 36.0 216 0.7771
0.7776 37.0 222 0.7742
0.7767 38.0 228 0.7602
0.795 39.0 234 0.7568
0.8284 40.0 240 0.7556

Framework versions

  • Transformers 4.38.1
  • Pytorch 2.1.0+cu121
  • Datasets 2.18.0
  • Tokenizers 0.15.2