mlm
This model is a fine-tuned version of on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 6.2470
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 64
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 200
Training results
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
6.0795 | 10.87 | 500 | 6.3069 |
6.0066 | 21.74 | 1000 | 6.2240 |
5.9834 | 32.61 | 1500 | 6.3617 |
5.9497 | 43.48 | 2000 | 6.2711 |
5.9994 | 54.35 | 2500 | 6.2702 |
6.0073 | 65.22 | 3000 | 6.2237 |
5.9752 | 76.09 | 3500 | 6.1862 |
5.9932 | 86.96 | 4000 | 6.3695 |
5.9819 | 97.83 | 4500 | 6.3884 |
5.9803 | 108.7 | 5000 | 6.3632 |
5.9649 | 119.57 | 5500 | 6.2496 |
5.9697 | 130.43 | 6000 | 6.2470 |
Framework versions
- Transformers 4.35.2
- Pytorch 2.1.0+cu121
- Datasets 2.16.0
- Tokenizers 0.15.0
- Downloads last month
- 2