jaggernaut007 commited on
Commit
be193b9
1 Parent(s): d166a4e

End of training

Browse files
Files changed (3) hide show
  1. README.md +18 -21
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
README.md CHANGED
@@ -20,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [surrey-nlp/albert-large-v2-finetuned-abbDet](https://huggingface.co/surrey-nlp/albert-large-v2-finetuned-abbDet) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.1013
24
- - Precision: 0.9788
25
- - Recall: 0.9756
26
- - F1: 0.9772
27
- - Accuracy: 0.9754
28
 
29
  ## Model description
30
 
@@ -44,32 +44,29 @@ More information needed
44
 
45
  The following hyperparameters were used during training:
46
  - learning_rate: 2e-06
47
- - train_batch_size: 2
48
  - eval_batch_size: 4
49
  - seed: 42
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
  - num_epochs: 10
 
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
- | No log | 0.19 | 100 | 0.1850 | 0.9561 | 0.9550 | 0.9555 | 0.9554 |
59
- | No log | 0.37 | 200 | 0.1138 | 0.9716 | 0.9673 | 0.9695 | 0.9675 |
60
- | No log | 0.56 | 300 | 0.1021 | 0.9753 | 0.9720 | 0.9737 | 0.9714 |
61
- | No log | 0.75 | 400 | 0.1034 | 0.9758 | 0.9719 | 0.9739 | 0.9714 |
62
- | 0.2096 | 0.93 | 500 | 0.0964 | 0.9775 | 0.9740 | 0.9757 | 0.9737 |
63
- | 0.2096 | 1.12 | 600 | 0.0984 | 0.9781 | 0.9755 | 0.9768 | 0.9748 |
64
- | 0.2096 | 1.31 | 700 | 0.1016 | 0.9786 | 0.9758 | 0.9772 | 0.9753 |
65
- | 0.2096 | 1.49 | 800 | 0.1034 | 0.9781 | 0.9758 | 0.9769 | 0.9750 |
66
- | 0.2096 | 1.68 | 900 | 0.1020 | 0.9775 | 0.9755 | 0.9765 | 0.9746 |
67
- | 0.0669 | 1.87 | 1000 | 0.1006 | 0.9778 | 0.9749 | 0.9764 | 0.9744 |
68
- | 0.0669 | 2.05 | 1100 | 0.0973 | 0.9788 | 0.9760 | 0.9774 | 0.9755 |
69
- | 0.0669 | 2.24 | 1200 | 0.1013 | 0.9788 | 0.9756 | 0.9772 | 0.9754 |
70
- | 0.0669 | 2.43 | 1300 | 0.0995 | 0.9782 | 0.9758 | 0.9770 | 0.9750 |
71
- | 0.0669 | 2.61 | 1400 | 0.1079 | 0.9782 | 0.9759 | 0.9771 | 0.9752 |
72
- | 0.051 | 2.8 | 1500 | 0.1092 | 0.9781 | 0.9760 | 0.9771 | 0.9753 |
73
 
74
 
75
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [surrey-nlp/albert-large-v2-finetuned-abbDet](https://huggingface.co/surrey-nlp/albert-large-v2-finetuned-abbDet) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.0950
24
+ - Precision: 0.9784
25
+ - Recall: 0.9763
26
+ - F1: 0.9773
27
+ - Accuracy: 0.9757
28
 
29
  ## Model description
30
 
 
44
 
45
  The following hyperparameters were used during training:
46
  - learning_rate: 2e-06
47
+ - train_batch_size: 4
48
  - eval_batch_size: 4
49
  - seed: 42
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
  - num_epochs: 10
53
+ - mixed_precision_training: Native AMP
54
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
58
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
59
+ | No log | 0.37 | 100 | 0.1655 | 0.9638 | 0.9621 | 0.9629 | 0.9622 |
60
+ | No log | 0.75 | 200 | 0.1073 | 0.9752 | 0.9705 | 0.9729 | 0.9709 |
61
+ | No log | 1.12 | 300 | 0.0951 | 0.9776 | 0.9742 | 0.9759 | 0.9740 |
62
+ | No log | 1.49 | 400 | 0.0952 | 0.9778 | 0.9752 | 0.9765 | 0.9748 |
63
+ | 0.1901 | 1.87 | 500 | 0.0948 | 0.9780 | 0.9745 | 0.9763 | 0.9746 |
64
+ | 0.1901 | 2.24 | 600 | 0.0947 | 0.9788 | 0.9758 | 0.9773 | 0.9755 |
65
+ | 0.1901 | 2.61 | 700 | 0.0962 | 0.9789 | 0.9766 | 0.9778 | 0.9758 |
66
+ | 0.1901 | 2.99 | 800 | 0.0950 | 0.9784 | 0.9763 | 0.9773 | 0.9757 |
67
+ | 0.1901 | 3.36 | 900 | 0.0984 | 0.9784 | 0.9763 | 0.9773 | 0.9755 |
68
+ | 0.0493 | 3.73 | 1000 | 0.1012 | 0.9781 | 0.9759 | 0.9770 | 0.9752 |
69
+ | 0.0493 | 4.1 | 1100 | 0.1029 | 0.9781 | 0.9763 | 0.9772 | 0.9754 |
 
 
 
 
70
 
71
 
72
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae8c7992f42141781a6215d0eeddc78a897bacba8971a797713b9fa858a3651e
3
  size 66557136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a95b9de274ef18ff926af023c0c903c0db563b344555802f2f32c5d2f21cee7e
3
  size 66557136
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f625f4bf981bf830bce52fd6fec7e0585d7f9ac35b168e5f3651348b053067f
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d8760de0c03cf801906c27cf2357b08b72f43b3a2df29ed6230ac59f49992ba
3
  size 4984