pritmanvar
commited on
End of training
Browse files
README.md
CHANGED
@@ -5,9 +5,36 @@ tags:
|
|
5 |
- generated_from_trainer
|
6 |
datasets:
|
7 |
- shipping_label_ner
|
|
|
|
|
|
|
|
|
|
|
8 |
model-index:
|
9 |
- name: ner_roberta_model
|
10 |
-
results:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -17,16 +44,11 @@ should probably proofread and complete it, then remove this comment. -->
|
|
17 |
|
18 |
This model is a fine-tuned version of [xlm-roberta-large](https://huggingface.co/xlm-roberta-large) on the shipping_label_ner dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
-
-
|
21 |
-
-
|
22 |
-
-
|
23 |
-
-
|
24 |
-
-
|
25 |
-
- eval_runtime: 0.3242
|
26 |
-
- eval_samples_per_second: 15.424
|
27 |
-
- eval_steps_per_second: 9.255
|
28 |
-
- epoch: 23.0
|
29 |
-
- step: 230
|
30 |
|
31 |
## Model description
|
32 |
|
@@ -51,7 +73,63 @@ The following hyperparameters were used during training:
|
|
51 |
- seed: 42
|
52 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
53 |
- lr_scheduler_type: linear
|
54 |
-
- num_epochs:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
55 |
|
56 |
### Framework versions
|
57 |
|
|
|
5 |
- generated_from_trainer
|
6 |
datasets:
|
7 |
- shipping_label_ner
|
8 |
+
metrics:
|
9 |
+
- precision
|
10 |
+
- recall
|
11 |
+
- f1
|
12 |
+
- accuracy
|
13 |
model-index:
|
14 |
- name: ner_roberta_model
|
15 |
+
results:
|
16 |
+
- task:
|
17 |
+
name: Token Classification
|
18 |
+
type: token-classification
|
19 |
+
dataset:
|
20 |
+
name: shipping_label_ner
|
21 |
+
type: shipping_label_ner
|
22 |
+
config: shipping_label_ner
|
23 |
+
split: validation
|
24 |
+
args: shipping_label_ner
|
25 |
+
metrics:
|
26 |
+
- name: Precision
|
27 |
+
type: precision
|
28 |
+
value: 0.5272727272727272
|
29 |
+
- name: Recall
|
30 |
+
type: recall
|
31 |
+
value: 0.7837837837837838
|
32 |
+
- name: F1
|
33 |
+
type: f1
|
34 |
+
value: 0.6304347826086956
|
35 |
+
- name: Accuracy
|
36 |
+
type: accuracy
|
37 |
+
value: 0.7796610169491526
|
38 |
---
|
39 |
|
40 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
44 |
|
45 |
This model is a fine-tuned version of [xlm-roberta-large](https://huggingface.co/xlm-roberta-large) on the shipping_label_ner dataset.
|
46 |
It achieves the following results on the evaluation set:
|
47 |
+
- Loss: 2.0623
|
48 |
+
- Precision: 0.5273
|
49 |
+
- Recall: 0.7838
|
50 |
+
- F1: 0.6304
|
51 |
+
- Accuracy: 0.7797
|
|
|
|
|
|
|
|
|
|
|
52 |
|
53 |
## Model description
|
54 |
|
|
|
73 |
- seed: 42
|
74 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
75 |
- lr_scheduler_type: linear
|
76 |
+
- num_epochs: 50
|
77 |
+
|
78 |
+
### Training results
|
79 |
+
|
80 |
+
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
81 |
+
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
82 |
+
| No log | 1.0 | 14 | 1.1206 | 0.3125 | 0.4054 | 0.3529 | 0.6610 |
|
83 |
+
| No log | 2.0 | 28 | 0.7363 | 0.5128 | 0.5405 | 0.5263 | 0.7119 |
|
84 |
+
| No log | 3.0 | 42 | 0.6219 | 0.5333 | 0.6486 | 0.5854 | 0.7542 |
|
85 |
+
| No log | 4.0 | 56 | 0.7328 | 0.4727 | 0.7027 | 0.5652 | 0.7627 |
|
86 |
+
| No log | 5.0 | 70 | 0.8181 | 0.5 | 0.7297 | 0.5934 | 0.7542 |
|
87 |
+
| No log | 6.0 | 84 | 0.8485 | 0.5185 | 0.7568 | 0.6154 | 0.7627 |
|
88 |
+
| No log | 7.0 | 98 | 0.9692 | 0.5 | 0.7027 | 0.5843 | 0.7542 |
|
89 |
+
| No log | 8.0 | 112 | 0.9842 | 0.4915 | 0.7838 | 0.6042 | 0.7458 |
|
90 |
+
| No log | 9.0 | 126 | 1.1196 | 0.5 | 0.7838 | 0.6105 | 0.7542 |
|
91 |
+
| No log | 10.0 | 140 | 1.2147 | 0.5 | 0.7838 | 0.6105 | 0.7542 |
|
92 |
+
| No log | 11.0 | 154 | 1.4110 | 0.5 | 0.7568 | 0.6022 | 0.7712 |
|
93 |
+
| No log | 12.0 | 168 | 1.2104 | 0.5370 | 0.7838 | 0.6374 | 0.7881 |
|
94 |
+
| No log | 13.0 | 182 | 1.4145 | 0.5283 | 0.7568 | 0.6222 | 0.7797 |
|
95 |
+
| No log | 14.0 | 196 | 1.4939 | 0.5179 | 0.7838 | 0.6237 | 0.7712 |
|
96 |
+
| No log | 15.0 | 210 | 1.5558 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
|
97 |
+
| No log | 16.0 | 224 | 1.5639 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
|
98 |
+
| No log | 17.0 | 238 | 1.5208 | 0.5179 | 0.7838 | 0.6237 | 0.7712 |
|
99 |
+
| No log | 18.0 | 252 | 1.4787 | 0.5918 | 0.7838 | 0.6744 | 0.7966 |
|
100 |
+
| No log | 19.0 | 266 | 1.3946 | 0.5283 | 0.7568 | 0.6222 | 0.7797 |
|
101 |
+
| No log | 20.0 | 280 | 1.6672 | 0.5370 | 0.7838 | 0.6374 | 0.7881 |
|
102 |
+
| No log | 21.0 | 294 | 1.5746 | 0.5185 | 0.7568 | 0.6154 | 0.7712 |
|
103 |
+
| No log | 22.0 | 308 | 1.8881 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
104 |
+
| No log | 23.0 | 322 | 1.5084 | 0.5370 | 0.7838 | 0.6374 | 0.7881 |
|
105 |
+
| No log | 24.0 | 336 | 1.7922 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
106 |
+
| No log | 25.0 | 350 | 1.7265 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
|
107 |
+
| No log | 26.0 | 364 | 1.7467 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
|
108 |
+
| No log | 27.0 | 378 | 2.0162 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
|
109 |
+
| No log | 28.0 | 392 | 1.9460 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
|
110 |
+
| No log | 29.0 | 406 | 1.8957 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
111 |
+
| No log | 30.0 | 420 | 1.9941 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
|
112 |
+
| No log | 31.0 | 434 | 1.9095 | 0.5 | 0.7568 | 0.6022 | 0.7712 |
|
113 |
+
| No log | 32.0 | 448 | 1.8920 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
|
114 |
+
| No log | 33.0 | 462 | 1.9310 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
115 |
+
| No log | 34.0 | 476 | 1.9830 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
116 |
+
| No log | 35.0 | 490 | 2.0445 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
117 |
+
| 0.2599 | 36.0 | 504 | 2.1138 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
118 |
+
| 0.2599 | 37.0 | 518 | 2.0024 | 0.5091 | 0.7568 | 0.6087 | 0.7797 |
|
119 |
+
| 0.2599 | 38.0 | 532 | 2.0004 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
120 |
+
| 0.2599 | 39.0 | 546 | 2.0725 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
121 |
+
| 0.2599 | 40.0 | 560 | 2.0507 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
122 |
+
| 0.2599 | 41.0 | 574 | 2.0548 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
123 |
+
| 0.2599 | 42.0 | 588 | 2.1176 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
124 |
+
| 0.2599 | 43.0 | 602 | 2.0946 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
125 |
+
| 0.2599 | 44.0 | 616 | 2.1211 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
|
126 |
+
| 0.2599 | 45.0 | 630 | 2.1103 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
|
127 |
+
| 0.2599 | 46.0 | 644 | 2.0876 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
|
128 |
+
| 0.2599 | 47.0 | 658 | 2.0910 | 0.5179 | 0.7838 | 0.6237 | 0.7712 |
|
129 |
+
| 0.2599 | 48.0 | 672 | 2.0800 | 0.5179 | 0.7838 | 0.6237 | 0.7712 |
|
130 |
+
| 0.2599 | 49.0 | 686 | 2.0584 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
|
131 |
+
| 0.2599 | 50.0 | 700 | 2.0623 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
|
132 |
+
|
133 |
|
134 |
### Framework versions
|
135 |
|
config.json
CHANGED
@@ -12,38 +12,34 @@
|
|
12 |
"hidden_size": 1024,
|
13 |
"id2label": {
|
14 |
"0": "O",
|
15 |
-
"1": "B-
|
16 |
-
"2": "I-
|
17 |
-
"3": "B-
|
18 |
-
"4": "I-
|
19 |
-
"5": "B-
|
20 |
-
"6": "I-
|
21 |
-
"7": "B-
|
22 |
-
"8": "I-
|
23 |
"9": "B-LOCATION",
|
24 |
"10": "I-LOCATION",
|
25 |
-
"11": "B-
|
26 |
-
"12": "I-
|
27 |
-
"13": "B-BARCODE",
|
28 |
-
"14": "I-BARCODE"
|
29 |
},
|
30 |
"initializer_range": 0.02,
|
31 |
"intermediate_size": 4096,
|
32 |
"label2id": {
|
33 |
-
"B-
|
34 |
-
"B-
|
35 |
-
"B-
|
36 |
-
"B-GCNUM": 1,
|
37 |
"B-LOCATION": 9,
|
38 |
-
"B-NAME":
|
39 |
-
"B-ORG":
|
40 |
-
"I-
|
41 |
-
"I-
|
42 |
-
"I-
|
43 |
-
"I-GCNUM": 2,
|
44 |
"I-LOCATION": 10,
|
45 |
-
"I-NAME":
|
46 |
-
"I-ORG":
|
47 |
"O": 0
|
48 |
},
|
49 |
"layer_norm_eps": 1e-05,
|
|
|
12 |
"hidden_size": 1024,
|
13 |
"id2label": {
|
14 |
"0": "O",
|
15 |
+
"1": "B-ORG",
|
16 |
+
"2": "I-ORG",
|
17 |
+
"3": "B-GCNUMBER",
|
18 |
+
"4": "I-GCNUMBER",
|
19 |
+
"5": "B-BGNUMBER",
|
20 |
+
"6": "I-BGNUMBER",
|
21 |
+
"7": "B-NAME",
|
22 |
+
"8": "I-NAME",
|
23 |
"9": "B-LOCATION",
|
24 |
"10": "I-LOCATION",
|
25 |
+
"11": "B-COUNTRY",
|
26 |
+
"12": "I-COUNTRY"
|
|
|
|
|
27 |
},
|
28 |
"initializer_range": 0.02,
|
29 |
"intermediate_size": 4096,
|
30 |
"label2id": {
|
31 |
+
"B-BGNUMBER": 5,
|
32 |
+
"B-COUNTRY": 11,
|
33 |
+
"B-GCNUMBER": 3,
|
|
|
34 |
"B-LOCATION": 9,
|
35 |
+
"B-NAME": 7,
|
36 |
+
"B-ORG": 1,
|
37 |
+
"I-BGNUMBER": 6,
|
38 |
+
"I-COUNTRY": 12,
|
39 |
+
"I-GCNUMBER": 4,
|
|
|
40 |
"I-LOCATION": 10,
|
41 |
+
"I-NAME": 8,
|
42 |
+
"I-ORG": 2,
|
43 |
"O": 0
|
44 |
},
|
45 |
"layer_norm_eps": 1e-05,
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fd5d056b81bfca346f30cb0788e39cabe2dbbf25f4f935ee0c5dbf23e1291389
|
3 |
+
size 2235465156
|
runs/Mar27_05-21-34_757ff407108b/events.out.tfevents.1711516907.757ff407108b.1998.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d96d9a51251d29b130228aff1df69407699a2613fd54a77ed55bc25a0e1bcec
|
3 |
+
size 5656
|
runs/Mar27_05-22-19_757ff407108b/events.out.tfevents.1711516944.757ff407108b.1998.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3742923b4c2086dd911be2c35eb0b705c76e20eec9153d4dda99130966e1f228
|
3 |
+
size 34460
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4920
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:221ba8edfb13f553d7351e61bf3005f4f969c530104873c32c4775901cc10b53
|
3 |
size 4920
|