pritmanvar commited on
Commit
4c71718
1 Parent(s): cde809d

End of training

Browse files
README.md CHANGED
@@ -5,9 +5,36 @@ tags:
5
  - generated_from_trainer
6
  datasets:
7
  - shipping_label_ner
 
 
 
 
 
8
  model-index:
9
  - name: ner_roberta_model
10
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -17,16 +44,11 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [xlm-roberta-large](https://huggingface.co/xlm-roberta-large) on the shipping_label_ner dataset.
19
  It achieves the following results on the evaluation set:
20
- - eval_loss: 0.0031
21
- - eval_precision: 1.0
22
- - eval_recall: 1.0
23
- - eval_f1: 1.0
24
- - eval_accuracy: 1.0
25
- - eval_runtime: 0.3242
26
- - eval_samples_per_second: 15.424
27
- - eval_steps_per_second: 9.255
28
- - epoch: 23.0
29
- - step: 230
30
 
31
  ## Model description
32
 
@@ -51,7 +73,63 @@ The following hyperparameters were used during training:
51
  - seed: 42
52
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
  - lr_scheduler_type: linear
54
- - num_epochs: 100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
55
 
56
  ### Framework versions
57
 
 
5
  - generated_from_trainer
6
  datasets:
7
  - shipping_label_ner
8
+ metrics:
9
+ - precision
10
+ - recall
11
+ - f1
12
+ - accuracy
13
  model-index:
14
  - name: ner_roberta_model
15
+ results:
16
+ - task:
17
+ name: Token Classification
18
+ type: token-classification
19
+ dataset:
20
+ name: shipping_label_ner
21
+ type: shipping_label_ner
22
+ config: shipping_label_ner
23
+ split: validation
24
+ args: shipping_label_ner
25
+ metrics:
26
+ - name: Precision
27
+ type: precision
28
+ value: 0.5272727272727272
29
+ - name: Recall
30
+ type: recall
31
+ value: 0.7837837837837838
32
+ - name: F1
33
+ type: f1
34
+ value: 0.6304347826086956
35
+ - name: Accuracy
36
+ type: accuracy
37
+ value: 0.7796610169491526
38
  ---
39
 
40
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
44
 
45
  This model is a fine-tuned version of [xlm-roberta-large](https://huggingface.co/xlm-roberta-large) on the shipping_label_ner dataset.
46
  It achieves the following results on the evaluation set:
47
+ - Loss: 2.0623
48
+ - Precision: 0.5273
49
+ - Recall: 0.7838
50
+ - F1: 0.6304
51
+ - Accuracy: 0.7797
 
 
 
 
 
52
 
53
  ## Model description
54
 
 
73
  - seed: 42
74
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
75
  - lr_scheduler_type: linear
76
+ - num_epochs: 50
77
+
78
+ ### Training results
79
+
80
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
81
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
82
+ | No log | 1.0 | 14 | 1.1206 | 0.3125 | 0.4054 | 0.3529 | 0.6610 |
83
+ | No log | 2.0 | 28 | 0.7363 | 0.5128 | 0.5405 | 0.5263 | 0.7119 |
84
+ | No log | 3.0 | 42 | 0.6219 | 0.5333 | 0.6486 | 0.5854 | 0.7542 |
85
+ | No log | 4.0 | 56 | 0.7328 | 0.4727 | 0.7027 | 0.5652 | 0.7627 |
86
+ | No log | 5.0 | 70 | 0.8181 | 0.5 | 0.7297 | 0.5934 | 0.7542 |
87
+ | No log | 6.0 | 84 | 0.8485 | 0.5185 | 0.7568 | 0.6154 | 0.7627 |
88
+ | No log | 7.0 | 98 | 0.9692 | 0.5 | 0.7027 | 0.5843 | 0.7542 |
89
+ | No log | 8.0 | 112 | 0.9842 | 0.4915 | 0.7838 | 0.6042 | 0.7458 |
90
+ | No log | 9.0 | 126 | 1.1196 | 0.5 | 0.7838 | 0.6105 | 0.7542 |
91
+ | No log | 10.0 | 140 | 1.2147 | 0.5 | 0.7838 | 0.6105 | 0.7542 |
92
+ | No log | 11.0 | 154 | 1.4110 | 0.5 | 0.7568 | 0.6022 | 0.7712 |
93
+ | No log | 12.0 | 168 | 1.2104 | 0.5370 | 0.7838 | 0.6374 | 0.7881 |
94
+ | No log | 13.0 | 182 | 1.4145 | 0.5283 | 0.7568 | 0.6222 | 0.7797 |
95
+ | No log | 14.0 | 196 | 1.4939 | 0.5179 | 0.7838 | 0.6237 | 0.7712 |
96
+ | No log | 15.0 | 210 | 1.5558 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
97
+ | No log | 16.0 | 224 | 1.5639 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
98
+ | No log | 17.0 | 238 | 1.5208 | 0.5179 | 0.7838 | 0.6237 | 0.7712 |
99
+ | No log | 18.0 | 252 | 1.4787 | 0.5918 | 0.7838 | 0.6744 | 0.7966 |
100
+ | No log | 19.0 | 266 | 1.3946 | 0.5283 | 0.7568 | 0.6222 | 0.7797 |
101
+ | No log | 20.0 | 280 | 1.6672 | 0.5370 | 0.7838 | 0.6374 | 0.7881 |
102
+ | No log | 21.0 | 294 | 1.5746 | 0.5185 | 0.7568 | 0.6154 | 0.7712 |
103
+ | No log | 22.0 | 308 | 1.8881 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
104
+ | No log | 23.0 | 322 | 1.5084 | 0.5370 | 0.7838 | 0.6374 | 0.7881 |
105
+ | No log | 24.0 | 336 | 1.7922 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
106
+ | No log | 25.0 | 350 | 1.7265 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
107
+ | No log | 26.0 | 364 | 1.7467 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
108
+ | No log | 27.0 | 378 | 2.0162 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
109
+ | No log | 28.0 | 392 | 1.9460 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
110
+ | No log | 29.0 | 406 | 1.8957 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
111
+ | No log | 30.0 | 420 | 1.9941 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
112
+ | No log | 31.0 | 434 | 1.9095 | 0.5 | 0.7568 | 0.6022 | 0.7712 |
113
+ | No log | 32.0 | 448 | 1.8920 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
114
+ | No log | 33.0 | 462 | 1.9310 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
115
+ | No log | 34.0 | 476 | 1.9830 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
116
+ | No log | 35.0 | 490 | 2.0445 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
117
+ | 0.2599 | 36.0 | 504 | 2.1138 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
118
+ | 0.2599 | 37.0 | 518 | 2.0024 | 0.5091 | 0.7568 | 0.6087 | 0.7797 |
119
+ | 0.2599 | 38.0 | 532 | 2.0004 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
120
+ | 0.2599 | 39.0 | 546 | 2.0725 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
121
+ | 0.2599 | 40.0 | 560 | 2.0507 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
122
+ | 0.2599 | 41.0 | 574 | 2.0548 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
123
+ | 0.2599 | 42.0 | 588 | 2.1176 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
124
+ | 0.2599 | 43.0 | 602 | 2.0946 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
125
+ | 0.2599 | 44.0 | 616 | 2.1211 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
126
+ | 0.2599 | 45.0 | 630 | 2.1103 | 0.5091 | 0.7568 | 0.6087 | 0.7712 |
127
+ | 0.2599 | 46.0 | 644 | 2.0876 | 0.5 | 0.7568 | 0.6022 | 0.7627 |
128
+ | 0.2599 | 47.0 | 658 | 2.0910 | 0.5179 | 0.7838 | 0.6237 | 0.7712 |
129
+ | 0.2599 | 48.0 | 672 | 2.0800 | 0.5179 | 0.7838 | 0.6237 | 0.7712 |
130
+ | 0.2599 | 49.0 | 686 | 2.0584 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
131
+ | 0.2599 | 50.0 | 700 | 2.0623 | 0.5273 | 0.7838 | 0.6304 | 0.7797 |
132
+
133
 
134
  ### Framework versions
135
 
config.json CHANGED
@@ -12,38 +12,34 @@
12
  "hidden_size": 1024,
13
  "id2label": {
14
  "0": "O",
15
- "1": "B-GCNUM",
16
- "2": "I-GCNUM",
17
- "3": "B-BGNUM",
18
- "4": "I-BGNUM",
19
- "5": "B-DATE",
20
- "6": "I-DATE",
21
- "7": "B-ORG",
22
- "8": "I-ORG",
23
  "9": "B-LOCATION",
24
  "10": "I-LOCATION",
25
- "11": "B-NAME",
26
- "12": "I-NAME",
27
- "13": "B-BARCODE",
28
- "14": "I-BARCODE"
29
  },
30
  "initializer_range": 0.02,
31
  "intermediate_size": 4096,
32
  "label2id": {
33
- "B-BARCODE": 13,
34
- "B-BGNUM": 3,
35
- "B-DATE": 5,
36
- "B-GCNUM": 1,
37
  "B-LOCATION": 9,
38
- "B-NAME": 11,
39
- "B-ORG": 7,
40
- "I-BARCODE": 14,
41
- "I-BGNUM": 4,
42
- "I-DATE": 6,
43
- "I-GCNUM": 2,
44
  "I-LOCATION": 10,
45
- "I-NAME": 12,
46
- "I-ORG": 8,
47
  "O": 0
48
  },
49
  "layer_norm_eps": 1e-05,
 
12
  "hidden_size": 1024,
13
  "id2label": {
14
  "0": "O",
15
+ "1": "B-ORG",
16
+ "2": "I-ORG",
17
+ "3": "B-GCNUMBER",
18
+ "4": "I-GCNUMBER",
19
+ "5": "B-BGNUMBER",
20
+ "6": "I-BGNUMBER",
21
+ "7": "B-NAME",
22
+ "8": "I-NAME",
23
  "9": "B-LOCATION",
24
  "10": "I-LOCATION",
25
+ "11": "B-COUNTRY",
26
+ "12": "I-COUNTRY"
 
 
27
  },
28
  "initializer_range": 0.02,
29
  "intermediate_size": 4096,
30
  "label2id": {
31
+ "B-BGNUMBER": 5,
32
+ "B-COUNTRY": 11,
33
+ "B-GCNUMBER": 3,
 
34
  "B-LOCATION": 9,
35
+ "B-NAME": 7,
36
+ "B-ORG": 1,
37
+ "I-BGNUMBER": 6,
38
+ "I-COUNTRY": 12,
39
+ "I-GCNUMBER": 4,
 
40
  "I-LOCATION": 10,
41
+ "I-NAME": 8,
42
+ "I-ORG": 2,
43
  "O": 0
44
  },
45
  "layer_norm_eps": 1e-05,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbdde8eedbd88fe9d75ae1062180b98de398503ddda6a844c3dba50dc877c1b9
3
- size 2235473356
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd5d056b81bfca346f30cb0788e39cabe2dbbf25f4f935ee0c5dbf23e1291389
3
+ size 2235465156
runs/Mar27_05-21-34_757ff407108b/events.out.tfevents.1711516907.757ff407108b.1998.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d96d9a51251d29b130228aff1df69407699a2613fd54a77ed55bc25a0e1bcec
3
+ size 5656
runs/Mar27_05-22-19_757ff407108b/events.out.tfevents.1711516944.757ff407108b.1998.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3742923b4c2086dd911be2c35eb0b705c76e20eec9153d4dda99130966e1f228
3
+ size 34460
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c88ed8958463d7430fbe65967a6974a47a85bdee475e4eb02eb56fe84379197
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:221ba8edfb13f553d7351e61bf3005f4f969c530104873c32c4775901cc10b53
3
  size 4920