Spaces:
Runtime error
Runtime error
Commit
•
fb31a67
1
Parent(s):
71f03d0
Upload 17 files
Browse files- training_trf/model-best/config.cfg +5 -4
- training_trf/model-best/meta.json +24 -24
- training_trf/model-best/tokenizer +0 -0
- training_trf/model-best/torch_ner/model +0 -0
- training_trf/model-best/transformer/model/config.json +1 -1
- training_trf/model-best/transformer/model/pytorch_model.bin +2 -2
- training_trf/model-best/transformer/model/tokenizer_config.json +1 -1
- training_trf/model-best/vocab/strings.json +0 -0
training_trf/model-best/config.cfg
CHANGED
@@ -28,8 +28,8 @@ factory = "torch_ner"
|
|
28 |
@architectures = "TorchEntityRecognizer.v1"
|
29 |
hidden_width = 48
|
30 |
nO = 11
|
31 |
-
dropout = 0.1
|
32 |
width = 768
|
|
|
33 |
|
34 |
[components.torch_ner.model.tok2vec]
|
35 |
@architectures = "spacy-transformers.TransformerListener.v1"
|
@@ -79,12 +79,13 @@ dropout = 0.1
|
|
79 |
accumulate_gradient = 1
|
80 |
patience = 1600000
|
81 |
max_epochs = 0
|
82 |
-
max_steps =
|
83 |
eval_frequency = 100
|
84 |
frozen_components = []
|
85 |
dev_corpus = "corpora.dev"
|
86 |
train_corpus = "corpora.train"
|
87 |
before_to_disk = null
|
|
|
88 |
|
89 |
[training.batcher]
|
90 |
@batchers = "spacy.batch_by_padded.v1"
|
@@ -110,14 +111,14 @@ eps = 0.00000001
|
|
110 |
[training.optimizer.learn_rate]
|
111 |
@schedules = "warmup_linear.v1"
|
112 |
warmup_steps = 250
|
113 |
-
total_steps =
|
114 |
initial_rate = 0.00005
|
115 |
|
116 |
[training.score_weights]
|
117 |
-
ents_per_type = null
|
118 |
ents_f = 1.0
|
119 |
ents_p = 0.0
|
120 |
ents_r = 0.0
|
|
|
121 |
|
122 |
[pretraining]
|
123 |
|
|
|
28 |
@architectures = "TorchEntityRecognizer.v1"
|
29 |
hidden_width = 48
|
30 |
nO = 11
|
|
|
31 |
width = 768
|
32 |
+
dropout = 0.1
|
33 |
|
34 |
[components.torch_ner.model.tok2vec]
|
35 |
@architectures = "spacy-transformers.TransformerListener.v1"
|
|
|
79 |
accumulate_gradient = 1
|
80 |
patience = 1600000
|
81 |
max_epochs = 0
|
82 |
+
max_steps = 20000
|
83 |
eval_frequency = 100
|
84 |
frozen_components = []
|
85 |
dev_corpus = "corpora.dev"
|
86 |
train_corpus = "corpora.train"
|
87 |
before_to_disk = null
|
88 |
+
annotating_components = []
|
89 |
|
90 |
[training.batcher]
|
91 |
@batchers = "spacy.batch_by_padded.v1"
|
|
|
111 |
[training.optimizer.learn_rate]
|
112 |
@schedules = "warmup_linear.v1"
|
113 |
warmup_steps = 250
|
114 |
+
total_steps = 20000
|
115 |
initial_rate = 0.00005
|
116 |
|
117 |
[training.score_weights]
|
|
|
118 |
ents_f = 1.0
|
119 |
ents_p = 0.0
|
120 |
ents_r = 0.0
|
121 |
+
ents_per_type = null
|
122 |
|
123 |
[pretraining]
|
124 |
|
training_trf/model-best/meta.json
CHANGED
@@ -2,13 +2,13 @@
|
|
2 |
"lang":"en",
|
3 |
"name":"pipeline",
|
4 |
"version":"0.0.0",
|
5 |
-
"spacy_version":">=3.
|
6 |
"description":"",
|
7 |
"author":"",
|
8 |
"email":"",
|
9 |
"url":"",
|
10 |
"license":"",
|
11 |
-
"spacy_git_version":"
|
12 |
"vectors":{
|
13 |
"width":0,
|
14 |
"vectors":0,
|
@@ -45,37 +45,37 @@
|
|
45 |
|
46 |
],
|
47 |
"performance":{
|
|
|
|
|
|
|
48 |
"ents_per_type":{
|
49 |
"test":{
|
50 |
-
"p":0.
|
51 |
-
"r":0.
|
52 |
-
"f":0.
|
53 |
-
},
|
54 |
-
"treatment":{
|
55 |
-
"p":0.6986449864,
|
56 |
-
"r":0.8397394137,
|
57 |
-
"f":0.7627218935
|
58 |
},
|
59 |
"pronoun":{
|
60 |
-
"p":0.
|
61 |
-
"r":0.
|
62 |
-
"f":0.
|
63 |
},
|
64 |
"person":{
|
65 |
-
"p":0.
|
66 |
-
"r":0.
|
67 |
-
"f":0.
|
68 |
},
|
69 |
"problem":{
|
70 |
-
"p":0.
|
71 |
-
"r":0.
|
72 |
-
"f":0.
|
|
|
|
|
|
|
|
|
|
|
73 |
}
|
74 |
},
|
75 |
-
"
|
76 |
-
"
|
77 |
-
"ents_r":0.8559070368,
|
78 |
-
"transformer_loss":0.0232139751,
|
79 |
-
"torch_ner_loss":2.7205681801
|
80 |
}
|
81 |
}
|
|
|
2 |
"lang":"en",
|
3 |
"name":"pipeline",
|
4 |
"version":"0.0.0",
|
5 |
+
"spacy_version":">=3.1.1,<3.2.0",
|
6 |
"description":"",
|
7 |
"author":"",
|
8 |
"email":"",
|
9 |
"url":"",
|
10 |
"license":"",
|
11 |
+
"spacy_git_version":"ffaead8fe",
|
12 |
"vectors":{
|
13 |
"width":0,
|
14 |
"vectors":0,
|
|
|
45 |
|
46 |
],
|
47 |
"performance":{
|
48 |
+
"ents_f":0.8339137593,
|
49 |
+
"ents_p":0.8182157058,
|
50 |
+
"ents_r":0.8502259522,
|
51 |
"ents_per_type":{
|
52 |
"test":{
|
53 |
+
"p":0.8082727781,
|
54 |
+
"r":0.8363215732,
|
55 |
+
"f":0.8220579875
|
|
|
|
|
|
|
|
|
|
|
56 |
},
|
57 |
"pronoun":{
|
58 |
+
"p":0.9892473118,
|
59 |
+
"r":0.9787234043,
|
60 |
+
"f":0.9839572193
|
61 |
},
|
62 |
"person":{
|
63 |
+
"p":0.9094579008,
|
64 |
+
"r":0.9047619048,
|
65 |
+
"f":0.9071038251
|
66 |
},
|
67 |
"problem":{
|
68 |
+
"p":0.744809362,
|
69 |
+
"r":0.8353090601,
|
70 |
+
"f":0.7874675713
|
71 |
+
},
|
72 |
+
"treatment":{
|
73 |
+
"p":0.8118351064,
|
74 |
+
"r":0.7954397394,
|
75 |
+
"f":0.8035538006
|
76 |
}
|
77 |
},
|
78 |
+
"transformer_loss":0.3640895431,
|
79 |
+
"torch_ner_loss":404.4625749518
|
|
|
|
|
|
|
80 |
}
|
81 |
}
|
training_trf/model-best/tokenizer
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
training_trf/model-best/torch_ner/model
CHANGED
Binary files a/training_trf/model-best/torch_ner/model and b/training_trf/model-best/torch_ner/model differ
|
|
training_trf/model-best/transformer/model/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"RobertaForMaskedLM"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "roberta-base",
|
3 |
"architectures": [
|
4 |
"RobertaForMaskedLM"
|
5 |
],
|
training_trf/model-best/transformer/model/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96ad2bf189ddf1ab1345f9cc17577a02acb5e225d79951b1d777910a13002107
|
3 |
+
size 498667479
|
training_trf/model-best/transformer/model/tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "
|
|
|
1 |
+
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "roberta-base", "tokenizer_class": "RobertaTokenizer"}
|
training_trf/model-best/vocab/strings.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|