moska commited on
Commit
c18a5f5
1 Parent(s): 47048c1

Model save

Browse files
Files changed (1) hide show
  1. README.md +95 -0
README.md ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ base_model: allegro/plt5-small
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ - recall
9
+ - f1
10
+ - precision
11
+ model-index:
12
+ - name: plt5-seq-clf-with-entities-updated-finetuned
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # plt5-seq-clf-with-entities-updated-finetuned
20
+
21
+ This model is a fine-tuned version of [allegro/plt5-small](https://huggingface.co/allegro/plt5-small) on an unknown dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 0.9863
24
+ - Accuracy: {'accuracy': 0.6016129032258064}
25
+ - Recall: {'recall': 0.6016129032258064}
26
+ - F1: {'f1': 0.6090459454706235}
27
+ - Precision: {'precision': 0.6487538544674235}
28
+
29
+ ## Model description
30
+
31
+ More information needed
32
+
33
+ ## Intended uses & limitations
34
+
35
+ More information needed
36
+
37
+ ## Training and evaluation data
38
+
39
+ More information needed
40
+
41
+ ## Training procedure
42
+
43
+ ### Training hyperparameters
44
+
45
+ The following hyperparameters were used during training:
46
+ - learning_rate: 5.6e-05
47
+ - train_batch_size: 8
48
+ - eval_batch_size: 8
49
+ - seed: 42
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: linear
52
+ - num_epochs: 30
53
+
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Recall | F1 | Precision |
57
+ |:-------------:|:-----:|:-----:|:---------------:|:---------------------------------:|:-------------------------------:|:---------------------------:|:----------------------------------:|
58
+ | 1.7248 | 1.0 | 718 | 1.6722 | {'accuracy': 0.47258064516129034} | {'recall': 0.47258064516129034} | {'f1': 0.30332120269936047} | {'precision': 0.2233324661810614} |
59
+ | 1.6984 | 2.0 | 1436 | 1.6216 | {'accuracy': 0.47258064516129034} | {'recall': 0.47258064516129034} | {'f1': 0.30332120269936047} | {'precision': 0.2233324661810614} |
60
+ | 1.6839 | 3.0 | 2154 | 1.6344 | {'accuracy': 0.47258064516129034} | {'recall': 0.47258064516129034} | {'f1': 0.30332120269936047} | {'precision': 0.2233324661810614} |
61
+ | 1.6882 | 4.0 | 2872 | 1.6261 | {'accuracy': 0.47258064516129034} | {'recall': 0.47258064516129034} | {'f1': 0.30332120269936047} | {'precision': 0.2233324661810614} |
62
+ | 1.68 | 5.0 | 3590 | 1.6223 | {'accuracy': 0.47258064516129034} | {'recall': 0.47258064516129034} | {'f1': 0.30332120269936047} | {'precision': 0.2233324661810614} |
63
+ | 1.6777 | 6.0 | 4308 | 1.6521 | {'accuracy': 0.4854838709677419} | {'recall': 0.4854838709677419} | {'f1': 0.36285141031634727} | {'precision': 0.30095036265938396} |
64
+ | 1.6681 | 7.0 | 5026 | 1.6165 | {'accuracy': 0.47096774193548385} | {'recall': 0.47096774193548385} | {'f1': 0.3142909197822259} | {'precision': 0.27758817356390014} |
65
+ | 1.6585 | 8.0 | 5744 | 1.5583 | {'accuracy': 0.47580645161290325} | {'recall': 0.47580645161290325} | {'f1': 0.3179514906044033} | {'precision': 0.274054689168981} |
66
+ | 1.6399 | 9.0 | 6462 | 1.6084 | {'accuracy': 0.3564516129032258} | {'recall': 0.3564516129032258} | {'f1': 0.30977675417942074} | {'precision': 0.3086887092441926} |
67
+ | 1.6158 | 10.0 | 7180 | 1.6613 | {'accuracy': 0.3225806451612903} | {'recall': 0.3225806451612903} | {'f1': 0.2777093706693196} | {'precision': 0.5070305497722745} |
68
+ | 1.5835 | 11.0 | 7898 | 1.6525 | {'accuracy': 0.3370967741935484} | {'recall': 0.3370967741935484} | {'f1': 0.2946753320835634} | {'precision': 0.4987499213117131} |
69
+ | 1.5443 | 12.0 | 8616 | 1.5433 | {'accuracy': 0.39838709677419354} | {'recall': 0.39838709677419354} | {'f1': 0.37257538542456536} | {'precision': 0.5472359482869795} |
70
+ | 1.4792 | 13.0 | 9334 | 1.4685 | {'accuracy': 0.4290322580645161} | {'recall': 0.4290322580645161} | {'f1': 0.3843028777529311} | {'precision': 0.5497170294652844} |
71
+ | 1.419 | 14.0 | 10052 | 1.5534 | {'accuracy': 0.4032258064516129} | {'recall': 0.4032258064516129} | {'f1': 0.35189485350144095} | {'precision': 0.5701307405449848} |
72
+ | 1.3881 | 15.0 | 10770 | 1.3641 | {'accuracy': 0.4790322580645161} | {'recall': 0.4790322580645161} | {'f1': 0.4461803399889066} | {'precision': 0.5258731490942117} |
73
+ | 1.3582 | 16.0 | 11488 | 1.3837 | {'accuracy': 0.43870967741935485} | {'recall': 0.43870967741935485} | {'f1': 0.3975785817347331} | {'precision': 0.5481481481481482} |
74
+ | 1.3074 | 17.0 | 12206 | 1.2409 | {'accuracy': 0.5177419354838709} | {'recall': 0.5177419354838709} | {'f1': 0.49737440159156987} | {'precision': 0.5439755251062998} |
75
+ | 1.2529 | 18.0 | 12924 | 1.2490 | {'accuracy': 0.5241935483870968} | {'recall': 0.5241935483870968} | {'f1': 0.5075488601971412} | {'precision': 0.5801964826379877} |
76
+ | 1.2223 | 19.0 | 13642 | 1.1680 | {'accuracy': 0.5435483870967742} | {'recall': 0.5435483870967742} | {'f1': 0.5172098120467532} | {'precision': 0.5483692723442298} |
77
+ | 1.1881 | 20.0 | 14360 | 1.1325 | {'accuracy': 0.5467741935483871} | {'recall': 0.5467741935483871} | {'f1': 0.528976565119481} | {'precision': 0.5918362760770626} |
78
+ | 1.1524 | 21.0 | 15078 | 1.1075 | {'accuracy': 0.5338709677419354} | {'recall': 0.5338709677419354} | {'f1': 0.5363641334830415} | {'precision': 0.6113524377471905} |
79
+ | 1.1307 | 22.0 | 15796 | 1.0685 | {'accuracy': 0.5612903225806452} | {'recall': 0.5612903225806452} | {'f1': 0.567131293394492} | {'precision': 0.6230821316117012} |
80
+ | 1.1198 | 23.0 | 16514 | 1.0978 | {'accuracy': 0.5564516129032258} | {'recall': 0.5564516129032258} | {'f1': 0.5596055517552543} | {'precision': 0.6285694241881432} |
81
+ | 1.0856 | 24.0 | 17232 | 1.0779 | {'accuracy': 0.5532258064516129} | {'recall': 0.5532258064516129} | {'f1': 0.5591833153283243} | {'precision': 0.6338935526492327} |
82
+ | 1.0829 | 25.0 | 17950 | 1.0175 | {'accuracy': 0.5903225806451613} | {'recall': 0.5903225806451613} | {'f1': 0.5964860501094582} | {'precision': 0.6422535611112073} |
83
+ | 1.0613 | 26.0 | 18668 | 1.0426 | {'accuracy': 0.567741935483871} | {'recall': 0.567741935483871} | {'f1': 0.5748961882147833} | {'precision': 0.6378855920377489} |
84
+ | 1.0363 | 27.0 | 19386 | 0.9920 | {'accuracy': 0.5935483870967742} | {'recall': 0.5935483870967742} | {'f1': 0.6001368374403852} | {'precision': 0.6385480642288512} |
85
+ | 1.0412 | 28.0 | 20104 | 1.0210 | {'accuracy': 0.5758064516129032} | {'recall': 0.5758064516129032} | {'f1': 0.5836230006413563} | {'precision': 0.6487093843541626} |
86
+ | 1.0256 | 29.0 | 20822 | 0.9992 | {'accuracy': 0.5870967741935483} | {'recall': 0.5870967741935483} | {'f1': 0.5944960724933464} | {'precision': 0.6439234847872369} |
87
+ | 1.0354 | 30.0 | 21540 | 0.9863 | {'accuracy': 0.6016129032258064} | {'recall': 0.6016129032258064} | {'f1': 0.6090459454706235} | {'precision': 0.6487538544674235} |
88
+
89
+
90
+ ### Framework versions
91
+
92
+ - Transformers 4.32.0
93
+ - Pytorch 2.0.1+cu117
94
+ - Datasets 2.14.4
95
+ - Tokenizers 0.13.3