lBober commited on
Commit
8d9ea3a
1 Parent(s): 1dcaa76

lBober/my-model-MiniLM-Area

Browse files
README.md CHANGED
@@ -20,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [Microsoft/Multilingual-MiniLM-L12-H384](https://huggingface.co/Microsoft/Multilingual-MiniLM-L12-H384) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.7030
24
- - Accuracy: 0.4604
25
- - F1: 0.4211
26
- - Precision: 0.4635
27
- - Recall: 0.4604
28
 
29
  ## Model description
30
 
@@ -43,38 +43,23 @@ More information needed
43
  ### Training hyperparameters
44
 
45
  The following hyperparameters were used during training:
46
- - learning_rate: 3e-05
47
- - train_batch_size: 8
48
- - eval_batch_size: 8
49
  - seed: 42
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
- - num_epochs: 20
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
57
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
58
- | 1.8269 | 1.0 | 81 | 1.7906 | 0.3022 | 0.1832 | 0.1444 | 0.3022 |
59
- | 1.6732 | 2.0 | 162 | 1.7044 | 0.3381 | 0.2021 | 0.1453 | 0.3381 |
60
- | 1.5682 | 3.0 | 243 | 1.6712 | 0.3381 | 0.2041 | 0.1462 | 0.3381 |
61
- | 1.474 | 4.0 | 324 | 1.6753 | 0.3165 | 0.1918 | 0.1435 | 0.3165 |
62
- | 1.3727 | 5.0 | 405 | 1.6871 | 0.3165 | 0.1913 | 0.1424 | 0.3165 |
63
- | 1.2838 | 6.0 | 486 | 1.7624 | 0.3094 | 0.1875 | 0.1450 | 0.3094 |
64
- | 1.2015 | 7.0 | 567 | 1.7454 | 0.3381 | 0.2555 | 0.2103 | 0.3381 |
65
- | 1.1481 | 8.0 | 648 | 1.6219 | 0.4173 | 0.3319 | 0.2848 | 0.4173 |
66
- | 1.0393 | 9.0 | 729 | 1.6179 | 0.4101 | 0.3515 | 0.4186 | 0.4101 |
67
- | 0.9609 | 10.0 | 810 | 1.5942 | 0.4460 | 0.4154 | 0.4790 | 0.4460 |
68
- | 0.8892 | 11.0 | 891 | 1.5765 | 0.4388 | 0.3927 | 0.3892 | 0.4388 |
69
- | 0.8146 | 12.0 | 972 | 1.5880 | 0.4532 | 0.4104 | 0.4634 | 0.4532 |
70
- | 0.7427 | 13.0 | 1053 | 1.6326 | 0.4604 | 0.4153 | 0.4649 | 0.4604 |
71
- | 0.6991 | 14.0 | 1134 | 1.8251 | 0.4317 | 0.3888 | 0.4749 | 0.4317 |
72
- | 0.641 | 15.0 | 1215 | 1.6103 | 0.4604 | 0.4210 | 0.4605 | 0.4604 |
73
- | 0.5892 | 16.0 | 1296 | 1.6240 | 0.4892 | 0.4462 | 0.4940 | 0.4892 |
74
- | 0.5565 | 17.0 | 1377 | 1.6828 | 0.4460 | 0.4002 | 0.4503 | 0.4460 |
75
- | 0.5202 | 18.0 | 1458 | 1.6728 | 0.4604 | 0.4204 | 0.4452 | 0.4604 |
76
- | 0.498 | 19.0 | 1539 | 1.6580 | 0.4964 | 0.4641 | 0.5068 | 0.4964 |
77
- | 0.48 | 20.0 | 1620 | 1.7030 | 0.4604 | 0.4211 | 0.4635 | 0.4604 |
78
 
79
 
80
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [Microsoft/Multilingual-MiniLM-L12-H384](https://huggingface.co/Microsoft/Multilingual-MiniLM-L12-H384) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 1.7970
24
+ - Accuracy: 0.2662
25
+ - F1: 0.1237
26
+ - Precision: 0.1887
27
+ - Recall: 0.2662
28
 
29
  ## Model description
30
 
 
43
  ### Training hyperparameters
44
 
45
  The following hyperparameters were used during training:
46
+ - learning_rate: 5e-05
47
+ - train_batch_size: 100
48
+ - eval_batch_size: 20
49
  - seed: 42
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
+ - num_epochs: 5
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall |
57
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:|
58
+ | 1.936 | 1.0 | 7 | 1.9121 | 0.2374 | 0.0911 | 0.0564 | 0.2374 |
59
+ | 1.8873 | 2.0 | 14 | 1.8516 | 0.2374 | 0.0911 | 0.0564 | 0.2374 |
60
+ | 1.843 | 3.0 | 21 | 1.8171 | 0.2662 | 0.1340 | 0.1492 | 0.2662 |
61
+ | 1.8229 | 4.0 | 28 | 1.8018 | 0.2662 | 0.1237 | 0.1887 | 0.2662 |
62
+ | 1.8163 | 5.0 | 35 | 1.7970 | 0.2662 | 0.1237 | 0.1887 | 0.2662 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
63
 
64
 
65
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1269508880b89e91f8832d3604d27a959c647b2f3027dedf421fdc5c258bafe
3
  size 470649364
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb18b6e642bf767cffe069d0342f9a381ae071f7f491ac7cc81cc680dc0b5e02
3
  size 470649364
runs/May25_20-38-58_54438e66731d/events.out.tfevents.1716669539.54438e66731d.25372.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5dd1f6e10bf26ba1e5cad776eb9cd8f02958ec47bb12c69c124cbd73b82d9b0
3
+ size 14654
runs/May25_20-50-17_54438e66731d/events.out.tfevents.1716670217.54438e66731d.28640.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab26f6eb930310a1c78011ea30528ff0439ac74106a7e0bea272f43c960db162
3
+ size 8804
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9d4d2f92da563386822c07914e44422e17497eb2e36a947a3ba0b9f8a368a94
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce61c0a500a8d338942e70b42929fb9edf274d3840f2d1af27c4e18ce687feb3
3
  size 5112