María Navas Loro commited on
Commit
b314ae4
1 Parent(s): 9a6c7bb

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -18
README.md CHANGED
@@ -15,19 +15,26 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # roberta-finetuned-CPV_Spanish
17
 
18
- This model is a fine-tuned version of [PlanTL-GOB-ES/roberta-base-bne](https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne) on a dataset derived from Spanish Public Procurement documents from 2019. The whole fine-tuning process is available in the following [Kaggle notebook](https://www.kaggle.com/code/marianavasloro/fine-tuned-roberta-for-spanish-cpv-codes).
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.0460
21
- - F1: 0.7937
22
- - Roc Auc: 0.8857
23
- - Accuracy: 0.7398
24
- - Coverage Error: 10.3171
25
- - Label Ranking Average Precision Score: 0.7977
 
 
 
 
26
 
27
  ## Intended uses & limitations
28
 
29
- This model only predicts the first two digits of the CPV codes.
 
 
30
 
 
31
 
32
  ## Training procedure
33
 
@@ -46,16 +53,16 @@ The following hyperparameters were used during training:
46
 
47
  | Training Loss | Epoch | Step | Validation Loss | F1 | Roc Auc | Accuracy | Coverage Error | Label Ranking Average Precision Score |
48
  |:-------------:|:-----:|:-----:|:---------------:|:------:|:-------:|:--------:|:--------------:|:-------------------------------------:|
49
- | 0.0359 | 1.0 | 9054 | 0.0368 | 0.7527 | 0.8361 | 0.6920 | 14.2585 | 0.7318 |
50
- | 0.0314 | 2.0 | 18108 | 0.0332 | 0.7753 | 0.8518 | 0.7198 | 12.9053 | 0.7612 |
51
- | 0.0235 | 3.0 | 27162 | 0.0332 | 0.7824 | 0.8656 | 0.7284 | 11.8961 | 0.7767 |
52
- | 0.0166 | 4.0 | 36216 | 0.0348 | 0.7824 | 0.8725 | 0.7289 | 11.3928 | 0.7821 |
53
- | 0.0114 | 5.0 | 45270 | 0.0371 | 0.7825 | 0.8799 | 0.7271 | 10.8051 | 0.7871 |
54
- | 0.0079 | 6.0 | 54324 | 0.0398 | 0.7829 | 0.8765 | 0.7260 | 11.0922 | 0.7831 |
55
- | 0.0042 | 7.0 | 63378 | 0.0414 | 0.7889 | 0.8798 | 0.7317 | 10.7793 | 0.7891 |
56
- | 0.0025 | 8.0 | 72432 | 0.0434 | 0.7895 | 0.8847 | 0.7317 | 10.3856 | 0.7924 |
57
- | 0.0014 | 9.0 | 81486 | 0.0451 | 0.7928 | 0.8860 | 0.7356 | 10.3086 | 0.7960 |
58
- | 0.001 | 10.0 | 90540 | 0.0460 | 0.7937 | 0.8857 | 0.7398 | 10.3171 | 0.7977 |
59
 
60
 
61
  ### Framework versions
 
15
 
16
  # roberta-finetuned-CPV_Spanish
17
 
18
+ This model is a fine-tuned version of [PlanTL-GOB-ES/roberta-base-bne](https://huggingface.co/PlanTL-GOB-ES/roberta-base-bne) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.0465
21
+ - F1: 0.7918
22
+ - Roc Auc: 0.8860
23
+ - Accuracy: 0.7376
24
+ - Coverage Error: 10.2744
25
+ - Label Ranking Average Precision Score: 0.7973
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
 
31
  ## Intended uses & limitations
32
 
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
 
37
+ More information needed
38
 
39
  ## Training procedure
40
 
 
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | F1 | Roc Auc | Accuracy | Coverage Error | Label Ranking Average Precision Score |
55
  |:-------------:|:-----:|:-----:|:---------------:|:------:|:-------:|:--------:|:--------------:|:-------------------------------------:|
56
+ | 0.0354 | 1.0 | 9054 | 0.0362 | 0.7560 | 0.8375 | 0.6963 | 14.0835 | 0.7357 |
57
+ | 0.0311 | 2.0 | 18108 | 0.0331 | 0.7756 | 0.8535 | 0.7207 | 12.7880 | 0.7633 |
58
+ | 0.0235 | 3.0 | 27162 | 0.0333 | 0.7823 | 0.8705 | 0.7283 | 11.5179 | 0.7811 |
59
+ | 0.0157 | 4.0 | 36216 | 0.0348 | 0.7821 | 0.8699 | 0.7274 | 11.5836 | 0.7798 |
60
+ | 0.011 | 5.0 | 45270 | 0.0377 | 0.7799 | 0.8787 | 0.7239 | 10.9173 | 0.7841 |
61
+ | 0.008 | 6.0 | 54324 | 0.0395 | 0.7854 | 0.8787 | 0.7309 | 10.9042 | 0.7879 |
62
+ | 0.0042 | 7.0 | 63378 | 0.0421 | 0.7872 | 0.8823 | 0.7300 | 10.5687 | 0.7903 |
63
+ | 0.0025 | 8.0 | 72432 | 0.0439 | 0.7884 | 0.8867 | 0.7305 | 10.2220 | 0.7934 |
64
+ | 0.0015 | 9.0 | 81486 | 0.0456 | 0.7889 | 0.8872 | 0.7316 | 10.1781 | 0.7945 |
65
+ | 0.001 | 10.0 | 90540 | 0.0465 | 0.7918 | 0.8860 | 0.7376 | 10.2744 | 0.7973 |
66
 
67
 
68
  ### Framework versions