PuxAI commited on
Commit
7016efd
·
verified ·
1 Parent(s): db8cde9

End of training

Browse files
Files changed (2) hide show
  1. README.md +15 -13
  2. model.safetensors +1 -1
README.md CHANGED
@@ -21,11 +21,11 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  This model is a fine-tuned version of [distilbert-base-multilingual-cased](https://huggingface.co/distilbert-base-multilingual-cased) on an unknown dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.2739
25
- - Accuracy: 0.8794
26
- - F1: 0.8786
27
- - Precision: 0.9279
28
- - Recall: 0.8342
29
 
30
  ## Model description
31
 
@@ -44,26 +44,28 @@ More information needed
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
- - learning_rate: 2e-05
48
  - train_batch_size: 32
49
  - eval_batch_size: 64
50
  - seed: 42
51
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
- - num_epochs: 3
54
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
58
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
59
- | 0.4024 | 1.0 | 506 | 0.2979 | 0.8517 | 0.8521 | 0.8900 | 0.8172 |
60
- | 0.2601 | 2.0 | 1012 | 0.2740 | 0.8783 | 0.8774 | 0.9267 | 0.8332 |
61
- | 0.2324 | 3.0 | 1518 | 0.2722 | 0.8761 | 0.8756 | 0.9214 | 0.8342 |
 
 
62
 
63
 
64
  ### Framework versions
65
 
66
- - Transformers 5.3.0
67
- - Pytorch 2.10.0+cu128
68
  - Datasets 4.8.2
69
- - Tokenizers 0.22.2
 
21
 
22
  This model is a fine-tuned version of [distilbert-base-multilingual-cased](https://huggingface.co/distilbert-base-multilingual-cased) on an unknown dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.2337
25
+ - Accuracy: 0.8972
26
+ - F1: 0.8892
27
+ - Precision: 0.9345
28
+ - Recall: 0.848
29
 
30
  ## Model description
31
 
 
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
+ - learning_rate: 3e-05
48
  - train_batch_size: 32
49
  - eval_batch_size: 64
50
  - seed: 42
51
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
+ - num_epochs: 5
54
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
58
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
59
+ | 0.4157 | 1.0 | 506 | 0.2809 | 0.8789 | 0.8728 | 0.8915 | 0.8549 |
60
+ | 0.2653 | 2.0 | 1012 | 0.2862 | 0.87 | 0.8675 | 0.8597 | 0.8754 |
61
+ | 0.2374 | 3.0 | 1518 | 0.2337 | 0.8972 | 0.8892 | 0.9345 | 0.848 |
62
+ | 0.2209 | 4.0 | 2024 | 0.2392 | 0.8933 | 0.8857 | 0.9242 | 0.8503 |
63
+ | 0.2086 | 5.0 | 2530 | 0.2499 | 0.8906 | 0.8835 | 0.9154 | 0.8537 |
64
 
65
 
66
  ### Framework versions
67
 
68
+ - Transformers 4.56.0
69
+ - Pytorch 2.8.0+cu129
70
  - Datasets 4.8.2
71
+ - Tokenizers 0.22.0
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:50f9350531790b36c0ac76cd2e884e410a8749a4f5d36e3c37af6b7ac6b28f87
3
  size 541323512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b152c87ad1404b69b8c6fc04c003983ec039f8528aea2449fa51daa4334b54e
3
  size 541323512