End of training
Browse files
README.md
CHANGED
@@ -4,7 +4,7 @@ base_model: bert-base-cased
|
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
datasets:
|
7 |
-
-
|
8 |
metrics:
|
9 |
- precision
|
10 |
- recall
|
@@ -17,24 +17,24 @@ model-index:
|
|
17 |
name: Token Classification
|
18 |
type: token-classification
|
19 |
dataset:
|
20 |
-
name:
|
21 |
-
type:
|
22 |
-
config:
|
23 |
split: validation
|
24 |
-
args:
|
25 |
metrics:
|
26 |
- name: Precision
|
27 |
type: precision
|
28 |
-
value: 0.
|
29 |
- name: Recall
|
30 |
type: recall
|
31 |
-
value: 0.
|
32 |
- name: F1
|
33 |
type: f1
|
34 |
-
value: 0.
|
35 |
- name: Accuracy
|
36 |
type: accuracy
|
37 |
-
value: 0.
|
38 |
---
|
39 |
|
40 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -42,13 +42,13 @@ should probably proofread and complete it, then remove this comment. -->
|
|
42 |
|
43 |
# bert-base-cased-finetuned-ner
|
44 |
|
45 |
-
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the
|
46 |
It achieves the following results on the evaluation set:
|
47 |
-
- Loss:
|
48 |
-
- Precision: 0.
|
49 |
-
- Recall: 0.
|
50 |
-
- F1: 0.
|
51 |
-
- Accuracy: 0.
|
52 |
|
53 |
## Model description
|
54 |
|
@@ -68,62 +68,32 @@ More information needed
|
|
68 |
|
69 |
The following hyperparameters were used during training:
|
70 |
- learning_rate: 2e-05
|
71 |
-
- train_batch_size:
|
72 |
-
- eval_batch_size:
|
73 |
- seed: 42
|
74 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
75 |
- lr_scheduler_type: linear
|
76 |
-
- num_epochs:
|
77 |
|
78 |
### Training results
|
79 |
|
80 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
81 |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
82 |
-
| No log | 1.0 |
|
83 |
-
|
|
84 |
-
|
|
85 |
-
|
|
86 |
-
|
|
87 |
-
|
|
88 |
-
|
|
89 |
-
|
|
90 |
-
|
|
91 |
-
|
|
92 |
-
| No log | 11.0 | 44 | 0.5919 | 0.1680 | 0.1232 | 0.1421 | 0.8350 |
|
93 |
-
| No log | 12.0 | 48 | 0.5789 | 0.2051 | 0.1375 | 0.1647 | 0.8384 |
|
94 |
-
| No log | 13.0 | 52 | 0.5827 | 0.1611 | 0.1375 | 0.1484 | 0.8355 |
|
95 |
-
| No log | 14.0 | 56 | 0.5638 | 0.2281 | 0.1862 | 0.2050 | 0.8433 |
|
96 |
-
| No log | 15.0 | 60 | 0.5576 | 0.1879 | 0.1691 | 0.1780 | 0.8420 |
|
97 |
-
| No log | 16.0 | 64 | 0.5485 | 0.2110 | 0.1862 | 0.1979 | 0.8456 |
|
98 |
-
| No log | 17.0 | 68 | 0.5479 | 0.2401 | 0.2264 | 0.2330 | 0.8500 |
|
99 |
-
| No log | 18.0 | 72 | 0.5460 | 0.2406 | 0.2378 | 0.2392 | 0.8503 |
|
100 |
-
| No log | 19.0 | 76 | 0.5374 | 0.2531 | 0.2350 | 0.2437 | 0.8542 |
|
101 |
-
| No log | 20.0 | 80 | 0.5365 | 0.2364 | 0.2493 | 0.2427 | 0.8539 |
|
102 |
-
| No log | 21.0 | 84 | 0.5284 | 0.2462 | 0.2350 | 0.2405 | 0.8552 |
|
103 |
-
| No log | 22.0 | 88 | 0.5306 | 0.2812 | 0.2837 | 0.2825 | 0.8601 |
|
104 |
-
| No log | 23.0 | 92 | 0.5262 | 0.2722 | 0.2722 | 0.2722 | 0.8573 |
|
105 |
-
| No log | 24.0 | 96 | 0.5306 | 0.2447 | 0.2665 | 0.2551 | 0.8555 |
|
106 |
-
| No log | 25.0 | 100 | 0.5249 | 0.2785 | 0.3009 | 0.2893 | 0.8594 |
|
107 |
-
| No log | 26.0 | 104 | 0.5201 | 0.2801 | 0.2865 | 0.2833 | 0.8586 |
|
108 |
-
| No log | 27.0 | 108 | 0.5213 | 0.2806 | 0.2894 | 0.2849 | 0.8604 |
|
109 |
-
| No log | 28.0 | 112 | 0.5207 | 0.2732 | 0.2951 | 0.2837 | 0.8612 |
|
110 |
-
| No log | 29.0 | 116 | 0.5144 | 0.3027 | 0.3209 | 0.3115 | 0.8630 |
|
111 |
-
| No log | 30.0 | 120 | 0.5135 | 0.3073 | 0.3381 | 0.3220 | 0.8648 |
|
112 |
-
| No log | 31.0 | 124 | 0.5147 | 0.2953 | 0.3266 | 0.3102 | 0.8651 |
|
113 |
-
| No log | 32.0 | 128 | 0.5121 | 0.2937 | 0.3181 | 0.3054 | 0.8645 |
|
114 |
-
| No log | 33.0 | 132 | 0.5092 | 0.3061 | 0.3324 | 0.3187 | 0.8645 |
|
115 |
-
| No log | 34.0 | 136 | 0.5064 | 0.3342 | 0.3696 | 0.3510 | 0.8677 |
|
116 |
-
| No log | 35.0 | 140 | 0.5056 | 0.3191 | 0.3438 | 0.3310 | 0.8674 |
|
117 |
-
| No log | 36.0 | 144 | 0.5091 | 0.3023 | 0.3352 | 0.3179 | 0.8661 |
|
118 |
-
| No log | 37.0 | 148 | 0.5104 | 0.3061 | 0.3324 | 0.3187 | 0.8658 |
|
119 |
-
| No log | 38.0 | 152 | 0.5100 | 0.3152 | 0.3324 | 0.3236 | 0.8677 |
|
120 |
-
| No log | 39.0 | 156 | 0.5102 | 0.3243 | 0.3410 | 0.3324 | 0.8684 |
|
121 |
-
| No log | 40.0 | 160 | 0.5103 | 0.3251 | 0.3410 | 0.3329 | 0.8684 |
|
122 |
|
123 |
|
124 |
### Framework versions
|
125 |
|
126 |
-
- Transformers 4.
|
127 |
-
- Pytorch 2.
|
128 |
-
- Datasets 2.
|
129 |
-
- Tokenizers 0.
|
|
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
datasets:
|
7 |
+
- lener_br
|
8 |
metrics:
|
9 |
- precision
|
10 |
- recall
|
|
|
17 |
name: Token Classification
|
18 |
type: token-classification
|
19 |
dataset:
|
20 |
+
name: lener_br
|
21 |
+
type: lener_br
|
22 |
+
config: lener_br
|
23 |
split: validation
|
24 |
+
args: lener_br
|
25 |
metrics:
|
26 |
- name: Precision
|
27 |
type: precision
|
28 |
+
value: 0.7640519805855644
|
29 |
- name: Recall
|
30 |
type: recall
|
31 |
+
value: 0.818242790073776
|
32 |
- name: F1
|
33 |
type: f1
|
34 |
+
value: 0.7902194154319487
|
35 |
- name: Accuracy
|
36 |
type: accuracy
|
37 |
+
value: 0.9615441099339138
|
38 |
---
|
39 |
|
40 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
42 |
|
43 |
# bert-base-cased-finetuned-ner
|
44 |
|
45 |
+
This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the lener_br dataset.
|
46 |
It achieves the following results on the evaluation set:
|
47 |
+
- Loss: nan
|
48 |
+
- Precision: 0.7641
|
49 |
+
- Recall: 0.8182
|
50 |
+
- F1: 0.7902
|
51 |
+
- Accuracy: 0.9615
|
52 |
|
53 |
## Model description
|
54 |
|
|
|
68 |
|
69 |
The following hyperparameters were used during training:
|
70 |
- learning_rate: 2e-05
|
71 |
+
- train_batch_size: 8
|
72 |
+
- eval_batch_size: 8
|
73 |
- seed: 42
|
74 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
75 |
- lr_scheduler_type: linear
|
76 |
+
- num_epochs: 10
|
77 |
|
78 |
### Training results
|
79 |
|
80 |
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|
81 |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
|
82 |
+
| No log | 1.0 | 432 | nan | 0.6807 | 0.7773 | 0.7258 | 0.9450 |
|
83 |
+
| 0.3019 | 2.0 | 864 | nan | 0.7244 | 0.7725 | 0.7476 | 0.9531 |
|
84 |
+
| 0.0871 | 3.0 | 1296 | nan | 0.7352 | 0.8192 | 0.7749 | 0.9571 |
|
85 |
+
| 0.0527 | 4.0 | 1728 | nan | 0.7455 | 0.7864 | 0.7654 | 0.9557 |
|
86 |
+
| 0.031 | 5.0 | 2160 | nan | 0.7334 | 0.7976 | 0.7642 | 0.9544 |
|
87 |
+
| 0.0223 | 6.0 | 2592 | nan | 0.7703 | 0.8343 | 0.8010 | 0.9624 |
|
88 |
+
| 0.0171 | 7.0 | 3024 | nan | 0.7279 | 0.8119 | 0.7676 | 0.9569 |
|
89 |
+
| 0.0171 | 8.0 | 3456 | nan | 0.7609 | 0.8067 | 0.7831 | 0.9613 |
|
90 |
+
| 0.012 | 9.0 | 3888 | nan | 0.7585 | 0.8152 | 0.7858 | 0.9608 |
|
91 |
+
| 0.0097 | 10.0 | 4320 | nan | 0.7641 | 0.8182 | 0.7902 | 0.9615 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
92 |
|
93 |
|
94 |
### Framework versions
|
95 |
|
96 |
+
- Transformers 4.35.2
|
97 |
+
- Pytorch 2.1.0+cu118
|
98 |
+
- Datasets 2.15.0
|
99 |
+
- Tokenizers 0.15.0
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 430942044
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9930c00dc43986c7cb69c7cb16dee49ad09283bca51263c251ddcf3a09b7f074
|
3 |
size 430942044
|
runs/Dec02_03-31-50_397da82fe218/events.out.tfevents.1701487923.397da82fe218.488.2
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1c1b6f48309d1ae41f79dbe45ea1578163be937793fb65dc0c83fd6cc533d74f
|
3 |
+
size 11115
|
runs/Dec02_03-31-50_397da82fe218/events.out.tfevents.1701489755.397da82fe218.488.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8174818a6fb5f3cea0263e28b8df8feef8dde4f5518eb55be758e8e75b6731d2
|
3 |
+
size 560
|