iproskurina commited on
Commit
513a87e
1 Parent(s): 3f63f0a

Update spaCy pipeline

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +6 -0
  2. README.md +18 -0
  3. config.cfg +116 -0
  4. en_grammar_checker-any-py3-none-any.whl +3 -0
  5. meta.json +153 -0
  6. tokenizer +3 -0
  7. training/articles/config.cfg +157 -0
  8. training/articles/meta.json +46 -0
  9. training/articles/spancat/cfg +8 -0
  10. training/articles/spancat/model +3 -0
  11. training/articles/tokenizer +3 -0
  12. training/articles/transformer/cfg +3 -0
  13. training/articles/transformer/model +3 -0
  14. training/articles/vocab/key2row +1 -0
  15. training/articles/vocab/lookups.bin +3 -0
  16. training/articles/vocab/strings.json +3 -0
  17. training/articles/vocab/vectors +3 -0
  18. training/articles/vocab/vectors.cfg +3 -0
  19. training/grammar_major/config.cfg +160 -0
  20. training/grammar_major/meta.json +52 -0
  21. training/grammar_major/spancat/cfg +11 -0
  22. training/grammar_major/spancat/model +3 -0
  23. training/grammar_major/tokenizer +3 -0
  24. training/grammar_major/transformer/cfg +3 -0
  25. training/grammar_major/transformer/model +3 -0
  26. training/grammar_major/vocab/key2row +1 -0
  27. training/grammar_major/vocab/lookups.bin +3 -0
  28. training/grammar_major/vocab/strings.json +3 -0
  29. training/grammar_major/vocab/vectors +3 -0
  30. training/grammar_major/vocab/vectors.cfg +3 -0
  31. training/grammar_minor/config.cfg +161 -0
  32. training/grammar_minor/meta.json +54 -0
  33. training/grammar_minor/spancat/cfg +12 -0
  34. training/grammar_minor/spancat/model +3 -0
  35. training/grammar_minor/tokenizer +3 -0
  36. training/grammar_minor/transformer/cfg +3 -0
  37. training/grammar_minor/transformer/model +3 -0
  38. training/grammar_minor/vocab/key2row +1 -0
  39. training/grammar_minor/vocab/lookups.bin +3 -0
  40. training/grammar_minor/vocab/strings.json +3 -0
  41. training/grammar_minor/vocab/vectors +3 -0
  42. training/grammar_minor/vocab/vectors.cfg +3 -0
  43. training/punctuation/config.cfg +157 -0
  44. training/punctuation/meta.json +46 -0
  45. training/punctuation/spancat/cfg +8 -0
  46. training/punctuation/spancat/model +3 -0
  47. training/punctuation/tokenizer +3 -0
  48. training/punctuation/transformer/cfg +3 -0
  49. training/punctuation/transformer/model +3 -0
  50. training/punctuation/vocab/key2row +1 -0
.gitattributes CHANGED
@@ -25,3 +25,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
  *.zstandard filter=lfs diff=lfs merge=lfs -text
27
  *tfevents* filter=lfs diff=lfs merge=lfs -text
28
+ *.whl filter=lfs diff=lfs merge=lfs -text
29
+ *.npz filter=lfs diff=lfs merge=lfs -text
30
+ *strings.json filter=lfs diff=lfs merge=lfs -text
31
+ vectors filter=lfs diff=lfs merge=lfs -text
32
+ model filter=lfs diff=lfs merge=lfs -text
33
+ tokenizer filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - spacy
4
+ language:
5
+ - en
6
+ license: cc-by-sa-3.0
7
+ ---
8
+ | Feature | Description |
9
+ | --- | --- |
10
+ | **Name** | `en_grammar_checker` |
11
+ | **Version** | `1.0.0` |
12
+ | **spaCy** | `>=3.3.0,<3.4.0` |
13
+ | **Default Pipeline** | `punctuation`, `spelling`, `articles`, `grammar_major`, `grammar_minor`, `vocabulary` |
14
+ | **Components** | `punctuation`, `spelling`, `articles`, `grammar_major`, `grammar_minor`, `vocabulary` |
15
+ | **Vectors** | 0 keys, 0 unique vectors (0 dimensions) |
16
+ | **Sources** | n/a |
17
+ | **License** | n/a |
18
+ | **Author** | [n/a]() |
config.cfg ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = null
3
+ dev = null
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ seed = 0
9
+ gpu_allocator = null
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["punctuation","spelling","articles","grammar_major","grammar_minor","vocabulary"]
14
+ disabled = []
15
+ before_creation = null
16
+ after_creation = null
17
+ after_pipeline_creation = null
18
+ batch_size = 1000
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.articles]
24
+ factory = "articles"
25
+
26
+ [components.grammar_major]
27
+ factory = "grammar_major"
28
+
29
+ [components.grammar_minor]
30
+ factory = "grammar_minor"
31
+
32
+ [components.punctuation]
33
+ factory = "punctuation"
34
+
35
+ [components.spelling]
36
+ factory = "spelling"
37
+
38
+ [components.vocabulary]
39
+ factory = "vocabulary"
40
+
41
+ [corpora]
42
+
43
+ [corpora.dev]
44
+ @readers = "spacy.Corpus.v1"
45
+ path = ${paths.dev}
46
+ gold_preproc = false
47
+ max_length = 0
48
+ limit = 0
49
+ augmenter = null
50
+
51
+ [corpora.train]
52
+ @readers = "spacy.Corpus.v1"
53
+ path = ${paths.train}
54
+ gold_preproc = false
55
+ max_length = 0
56
+ limit = 0
57
+ augmenter = null
58
+
59
+ [training]
60
+ seed = ${system.seed}
61
+ gpu_allocator = ${system.gpu_allocator}
62
+ dropout = 0.1
63
+ accumulate_gradient = 1
64
+ patience = 1600
65
+ max_epochs = 0
66
+ max_steps = 20000
67
+ eval_frequency = 200
68
+ frozen_components = []
69
+ annotating_components = []
70
+ dev_corpus = "corpora.dev"
71
+ train_corpus = "corpora.train"
72
+ before_to_disk = null
73
+
74
+ [training.batcher]
75
+ @batchers = "spacy.batch_by_words.v1"
76
+ discard_oversize = false
77
+ tolerance = 0.2
78
+ get_length = null
79
+
80
+ [training.batcher.size]
81
+ @schedules = "compounding.v1"
82
+ start = 100
83
+ stop = 1000
84
+ compound = 1.001
85
+ t = 0.0
86
+
87
+ [training.logger]
88
+ @loggers = "spacy.ConsoleLogger.v1"
89
+ progress_bar = false
90
+
91
+ [training.optimizer]
92
+ @optimizers = "Adam.v1"
93
+ beta1 = 0.9
94
+ beta2 = 0.999
95
+ L2_is_weight_decay = true
96
+ L2 = 0.01
97
+ grad_clip = 1.0
98
+ use_averages = false
99
+ eps = 0.00000001
100
+ learn_rate = 0.001
101
+
102
+ [training.score_weights]
103
+
104
+ [pretraining]
105
+
106
+ [initialize]
107
+ vectors = ${paths.vectors}
108
+ init_tok2vec = ${paths.init_tok2vec}
109
+ vocab_data = null
110
+ lookups = null
111
+ before_init = null
112
+ after_init = null
113
+
114
+ [initialize.components]
115
+
116
+ [initialize.tokenizer]
en_grammar_checker-any-py3-none-any.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0a808972cae10f9b445a2435a70ee3664f8433086d3b3862febe7679cfb321e
3
+ size 2441454414
meta.json ADDED
@@ -0,0 +1,153 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lang":"en",
3
+ "name":"grammar_checker",
4
+ "version":"1.0.0",
5
+ "description":"Essay Grammar Checker",
6
+ "author":"Irina Proskurina",
7
+ "email":"",
8
+ "url":"",
9
+ "license":"CC BY-SA 3.0",
10
+ "spacy_version":">=3.3.0,<3.4.0",
11
+ "parent_package":"spacy",
12
+ "requirements":[
13
+ "spacy-transformers>=1.0.0,<1.1.0"
14
+ ],
15
+ "sources":[
16
+ {
17
+ "license":"MIT"
18
+ }
19
+ ],
20
+ "vectors":{
21
+ "width":0,
22
+ "vectors":0,
23
+ "keys":0,
24
+ "name":null
25
+ },
26
+ "pipeline":[
27
+ "punctuation",
28
+ "spelling",
29
+ "articles",
30
+ "grammar_major",
31
+ "grammar_minor",
32
+ "vocabulary"
33
+ ],
34
+ "components":[
35
+ "punctuation",
36
+ "spelling",
37
+ "articles",
38
+ "grammar_major",
39
+ "grammar_minor",
40
+ "vocabulary"
41
+ ],
42
+ "labels":{
43
+
44
+ },
45
+ "performance":{
46
+ "spans_articles_p":0.8274481176,
47
+ "spans_articles_r":0.8197202306,
48
+ "spans_articles_f":0.823566046,
49
+ "spans_punctuation_p":0.8723066455,
50
+ "spans_punctuation_r":0.6734372264,
51
+ "spans_punctuation_f":0.7600790514,
52
+ "spans_spelling_p":0.9037660231,
53
+ "spans_spelling_r":0.9159588014,
54
+ "spans_spelling_f":0.9098215644,
55
+ "spans_spelling_per_type":{
56
+ "Spelling":{
57
+ "p":0.9044968253,
58
+ "r":0.9227970539,
59
+ "f":0.9135553018
60
+ },
61
+ "Capitalisation":{
62
+ "p":0.8959762655,
63
+ "r":0.8483146067,
64
+ "f":0.8714942736
65
+ }
66
+ },
67
+ "spans_vocabulary_p":0.8326700274,
68
+ "spans_vocabulary_r":0.4697166234,
69
+ "spans_vocabulary_f":0.6006188,
70
+ "spans_vocabulary_per_type":{
71
+ "lex_part_choice":{
72
+ "p":0.8783783784,
73
+ "r":0.0741444867,
74
+ "f":0.1367461431
75
+ },
76
+ "lex_item_choice":{
77
+ "p":0.8219627873,
78
+ "r":0.5290802348,
79
+ "f":0.6437755977
80
+ },
81
+ "Category_confusion":{
82
+ "p":0.84472708,
83
+ "r":0.5153572583,
84
+ "f":0.6401606426
85
+ },
86
+ "Formational_affixes":{
87
+ "p":0.8892307692,
88
+ "r":0.5608020699,
89
+ "f":0.6878222927
90
+ }
91
+ },
92
+ "spans_grammar_minor_p":0.8690835299,
93
+ "spans_grammar_minor_r":0.7119557892,
94
+ "spans_grammar_minor_f":0.7827117807,
95
+ "spans_grammar_minor_per_type":{
96
+ "Verb_pattern":{
97
+ "p":0.7399617591,
98
+ "r":0.2491951062,
99
+ "f":0.3728323699
100
+ },
101
+ "Noun_number":{
102
+ "p":0.8889171489,
103
+ "r":0.9032758286,
104
+ "f":0.8960389694
105
+ },
106
+ "Word_order":{
107
+ "p":0.8179384203,
108
+ "r":0.2189179506,
109
+ "f":0.3453928773
110
+ },
111
+ "Numerals":{
112
+ "p":0.7062761506,
113
+ "r":0.602426838,
114
+ "f":0.6502311248
115
+ },
116
+ "Determiners":{
117
+ "p":0.5853658537,
118
+ "r":0.0231884058,
119
+ "f":0.0446096654
120
+ }
121
+ },
122
+ "spans_grammar_major_p":0.8694734373,
123
+ "spans_grammar_major_r":0.6123730074,
124
+ "spans_grammar_major_f":0.7186197538,
125
+ "spans_grammar_major_per_type":{
126
+ "Agreement_errors":{
127
+ "p":0.8953090397,
128
+ "r":0.7739296902,
129
+ "f":0.8302062914
130
+ },
131
+ "Prepositions":{
132
+ "p":0.881741712,
133
+ "r":0.5715202053,
134
+ "f":0.6935201401
135
+ },
136
+ "Redundant_comp":{
137
+ "p":0.8028656126,
138
+ "r":0.314191802,
139
+ "f":0.4516397999
140
+ },
141
+ "Tense_choice":{
142
+ "p":0.86231523,
143
+ "r":0.7354107649,
144
+ "f":0.7938231022
145
+ }
146
+ }
147
+ },
148
+ "speed":5613.7287946692,
149
+ "spacy_git_version":"61dfdd9fb",
150
+ "disabled":[
151
+
152
+ ]
153
+ }
tokenizer ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/articles/config.cfg ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "./realec/articles_train.spacy"
3
+ dev = "./realec/articles_dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ gpu_allocator = "pytorch"
9
+ seed = 0
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["transformer","spancat"]
14
+ batch_size = 128
15
+ disabled = []
16
+ before_creation = null
17
+ after_creation = null
18
+ after_pipeline_creation = null
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.spancat]
24
+ factory = "spancat"
25
+ max_positive = null
26
+ scorer = {"@scorers":"spacy.spancat_scorer.v1"}
27
+ spans_key = "articles"
28
+ threshold = 0.5
29
+
30
+ [components.spancat.model]
31
+ @architectures = "spacy.SpanCategorizer.v1"
32
+
33
+ [components.spancat.model.reducer]
34
+ @layers = "spacy.mean_max_reducer.v1"
35
+ hidden_size = 128
36
+
37
+ [components.spancat.model.scorer]
38
+ @layers = "spacy.LinearLogistic.v1"
39
+ nO = null
40
+ nI = null
41
+
42
+ [components.spancat.model.tok2vec]
43
+ @architectures = "spacy-transformers.TransformerListener.v1"
44
+ grad_factor = 1.0
45
+ pooling = {"@layers":"reduce_mean.v1"}
46
+ upstream = "*"
47
+
48
+ [components.spancat.suggester]
49
+ @misc = "spacy.ngram_suggester.v1"
50
+ sizes = [1,2,3]
51
+
52
+ [components.transformer]
53
+ factory = "transformer"
54
+ max_batch_items = 4096
55
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
56
+
57
+ [components.transformer.model]
58
+ @architectures = "spacy-transformers.TransformerModel.v3"
59
+ name = "bert-base-cased"
60
+ mixed_precision = false
61
+
62
+ [components.transformer.model.get_spans]
63
+ @span_getters = "spacy-transformers.strided_spans.v1"
64
+ window = 128
65
+ stride = 96
66
+
67
+ [components.transformer.model.grad_scaler_config]
68
+
69
+ [components.transformer.model.tokenizer_config]
70
+ use_fast = true
71
+
72
+ [components.transformer.model.transformer_config]
73
+
74
+ [corpora]
75
+
76
+ [corpora.dev]
77
+ @readers = "spacy.Corpus.v1"
78
+ path = "./realec/articles_dev.spacy"
79
+ max_length = 0
80
+ gold_preproc = false
81
+ limit = 0
82
+ augmenter = null
83
+
84
+ [corpora.train]
85
+ @readers = "spacy.Corpus.v1"
86
+ path = "./realec/articles_train.spacy"
87
+ max_length = 0
88
+ gold_preproc = false
89
+ limit = 0
90
+ augmenter = null
91
+
92
+ [training]
93
+ accumulate_gradient = 3
94
+ dev_corpus = "corpora.dev"
95
+ train_corpus = "corpora.train"
96
+ frozen_components = []
97
+ seed = 0
98
+ gpu_allocator = "pytorch"
99
+ dropout = 0.1
100
+ patience = 1600
101
+ max_epochs = 0
102
+ max_steps = 20000
103
+ eval_frequency = 200
104
+ annotating_components = []
105
+ before_to_disk = null
106
+
107
+ [training.batcher]
108
+ @batchers = "spacy.batch_by_padded.v1"
109
+ discard_oversize = true
110
+ size = 2000
111
+ buffer = 256
112
+ get_length = null
113
+
114
+ [training.logger]
115
+ @loggers = "spacy.WandbLogger.v3"
116
+ project_name = "adwiser"
117
+ remove_config_values = ["paths.train","paths.dev","corpora.train.path","corpora.dev.path"]
118
+ model_log_interval = 1000
119
+ log_dataset_dir = null
120
+ entity = null
121
+ run_name = "articles"
122
+
123
+ [training.optimizer]
124
+ @optimizers = "Adam.v1"
125
+ beta1 = 0.9
126
+ beta2 = 0.999
127
+ L2_is_weight_decay = true
128
+ L2 = 0.01
129
+ grad_clip = 1.0
130
+ use_averages = false
131
+ eps = 0.00000001
132
+
133
+ [training.optimizer.learn_rate]
134
+ @schedules = "warmup_linear.v1"
135
+ warmup_steps = 250
136
+ total_steps = 20000
137
+ initial_rate = 0.00005
138
+
139
+ [training.score_weights]
140
+ spans_sc_f = 0.5
141
+ spans_sc_p = 0.0
142
+ spans_sc_r = 0.0
143
+ spans_Articles_f = 0.5
144
+
145
+ [pretraining]
146
+
147
+ [initialize]
148
+ vectors = null
149
+ init_tok2vec = null
150
+ vocab_data = null
151
+ lookups = null
152
+ before_init = null
153
+ after_init = null
154
+
155
+ [initialize.components]
156
+
157
+ [initialize.tokenizer]
training/articles/meta.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lang":"en",
3
+ "name":"pipeline",
4
+ "version":"0.0.0",
5
+ "spacy_version":">=3.3.0,<3.4.0",
6
+ "description":"",
7
+ "author":"",
8
+ "email":"",
9
+ "url":"",
10
+ "license":"",
11
+ "spacy_git_version":"Unknown",
12
+ "vectors":{
13
+ "width":0,
14
+ "vectors":0,
15
+ "keys":0,
16
+ "name":null,
17
+ "mode":"default"
18
+ },
19
+ "labels":{
20
+ "transformer":[
21
+
22
+ ],
23
+ "spancat":[
24
+ "Articles"
25
+ ]
26
+ },
27
+ "pipeline":[
28
+ "transformer",
29
+ "spancat"
30
+ ],
31
+ "components":[
32
+ "transformer",
33
+ "spancat"
34
+ ],
35
+ "disabled":[
36
+
37
+ ],
38
+ "performance":{
39
+ "spans_sc_f":0.0,
40
+ "spans_sc_p":0.0,
41
+ "spans_sc_r":0.0,
42
+ "spans_Articles_f":0.0,
43
+ "transformer_loss":100.8329219741,
44
+ "spancat_loss":2964.9278720927
45
+ }
46
+ }
training/articles/spancat/cfg ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels":[
3
+ "Articles"
4
+ ],
5
+ "spans_key":"articles",
6
+ "threshold":0.5,
7
+ "max_positive":null
8
+ }
training/articles/spancat/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bcd34ab7d5ff9adb0214059514deb98fb9c131f29d0a0a7a4c520fcafd3e875
3
+ size 4723491
training/articles/tokenizer ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/articles/transformer/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "max_batch_items":4096
3
+ }
training/articles/transformer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2036ce8a5df6924d046d42dd67dc03f670b4d1d25c2d8073de95822165e9a5ed
3
+ size 434208023
training/articles/vocab/key2row ADDED
@@ -0,0 +1 @@
 
1
+
training/articles/vocab/lookups.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76be8b528d0075f7aae98d6fa57a6d3c83ae480a8469e668d7b0af968995ac71
3
+ size 1
training/articles/vocab/strings.json ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6c403d7808ec8eeeafe465538d7df7f45c1202362b29c9c627a7ed8302dc210
3
+ size 287029
training/articles/vocab/vectors ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14772b683e726436d5948ad3fff2b43d036ef2ebbe3458aafed6004e05a40706
3
+ size 128
training/articles/vocab/vectors.cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "mode":"default"
3
+ }
training/grammar_major/config.cfg ADDED
@@ -0,0 +1,160 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "./realec/grammar_major_train.spacy"
3
+ dev = "./realec/grammar_major_dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ gpu_allocator = "pytorch"
9
+ seed = 0
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["transformer","spancat"]
14
+ batch_size = 128
15
+ disabled = []
16
+ before_creation = null
17
+ after_creation = null
18
+ after_pipeline_creation = null
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.spancat]
24
+ factory = "spancat"
25
+ max_positive = null
26
+ scorer = {"@scorers":"spacy.spancat_scorer.v1"}
27
+ spans_key = "grammar_major"
28
+ threshold = 0.5
29
+
30
+ [components.spancat.model]
31
+ @architectures = "spacy.SpanCategorizer.v1"
32
+
33
+ [components.spancat.model.reducer]
34
+ @layers = "spacy.mean_max_reducer.v1"
35
+ hidden_size = 128
36
+
37
+ [components.spancat.model.scorer]
38
+ @layers = "spacy.LinearLogistic.v1"
39
+ nO = null
40
+ nI = null
41
+
42
+ [components.spancat.model.tok2vec]
43
+ @architectures = "spacy-transformers.TransformerListener.v1"
44
+ grad_factor = 1.0
45
+ pooling = {"@layers":"reduce_mean.v1"}
46
+ upstream = "*"
47
+
48
+ [components.spancat.suggester]
49
+ @misc = "spacy.ngram_suggester.v1"
50
+ sizes = [1,2,3]
51
+
52
+ [components.transformer]
53
+ factory = "transformer"
54
+ max_batch_items = 4096
55
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
56
+
57
+ [components.transformer.model]
58
+ @architectures = "spacy-transformers.TransformerModel.v3"
59
+ name = "bert-base-cased"
60
+ mixed_precision = false
61
+
62
+ [components.transformer.model.get_spans]
63
+ @span_getters = "spacy-transformers.strided_spans.v1"
64
+ window = 128
65
+ stride = 96
66
+
67
+ [components.transformer.model.grad_scaler_config]
68
+
69
+ [components.transformer.model.tokenizer_config]
70
+ use_fast = true
71
+
72
+ [components.transformer.model.transformer_config]
73
+
74
+ [corpora]
75
+
76
+ [corpora.dev]
77
+ @readers = "spacy.Corpus.v1"
78
+ path = "./realec/grammar_major_dev.spacy"
79
+ max_length = 0
80
+ gold_preproc = false
81
+ limit = 0
82
+ augmenter = null
83
+
84
+ [corpora.train]
85
+ @readers = "spacy.Corpus.v1"
86
+ path = "./realec/grammar_major_train.spacy"
87
+ max_length = 0
88
+ gold_preproc = false
89
+ limit = 0
90
+ augmenter = null
91
+
92
+ [training]
93
+ accumulate_gradient = 3
94
+ dev_corpus = "corpora.dev"
95
+ train_corpus = "corpora.train"
96
+ frozen_components = []
97
+ seed = 0
98
+ gpu_allocator = "pytorch"
99
+ dropout = 0.1
100
+ patience = 1600
101
+ max_epochs = 0
102
+ max_steps = 20000
103
+ eval_frequency = 200
104
+ annotating_components = []
105
+ before_to_disk = null
106
+
107
+ [training.batcher]
108
+ @batchers = "spacy.batch_by_padded.v1"
109
+ discard_oversize = true
110
+ size = 2000
111
+ buffer = 256
112
+ get_length = null
113
+
114
+ [training.logger]
115
+ @loggers = "spacy.WandbLogger.v3"
116
+ project_name = "adwiser"
117
+ remove_config_values = ["paths.train","paths.dev","corpora.train.path","corpora.dev.path"]
118
+ model_log_interval = 1000
119
+ log_dataset_dir = null
120
+ entity = null
121
+ run_name = "grammar_major"
122
+
123
+ [training.optimizer]
124
+ @optimizers = "Adam.v1"
125
+ beta1 = 0.9
126
+ beta2 = 0.999
127
+ L2_is_weight_decay = true
128
+ L2 = 0.01
129
+ grad_clip = 1.0
130
+ use_averages = false
131
+ eps = 0.00000001
132
+
133
+ [training.optimizer.learn_rate]
134
+ @schedules = "warmup_linear.v1"
135
+ warmup_steps = 250
136
+ total_steps = 20000
137
+ initial_rate = 0.00005
138
+
139
+ [training.score_weights]
140
+ spans_sc_f = 0.5
141
+ spans_sc_p = 0.0
142
+ spans_sc_r = 0.0
143
+ spans_Agreement_errors_f = 0.13
144
+ spans_Prepositions_f = 0.12
145
+ spans_Redundant_comp_f = 0.14
146
+ spans_Tense_choice_f = 0.11
147
+
148
+ [pretraining]
149
+
150
+ [initialize]
151
+ vectors = null
152
+ init_tok2vec = null
153
+ vocab_data = null
154
+ lookups = null
155
+ before_init = null
156
+ after_init = null
157
+
158
+ [initialize.components]
159
+
160
+ [initialize.tokenizer]
training/grammar_major/meta.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lang":"en",
3
+ "name":"pipeline",
4
+ "version":"0.0.0",
5
+ "spacy_version":">=3.3.0,<3.4.0",
6
+ "description":"",
7
+ "author":"",
8
+ "email":"",
9
+ "url":"",
10
+ "license":"",
11
+ "spacy_git_version":"Unknown",
12
+ "vectors":{
13
+ "width":0,
14
+ "vectors":0,
15
+ "keys":0,
16
+ "name":null,
17
+ "mode":"default"
18
+ },
19
+ "labels":{
20
+ "transformer":[
21
+
22
+ ],
23
+ "spancat":[
24
+ "Agreement_errors",
25
+ "Prepositions",
26
+ "Tense_choice",
27
+ "Redundant_comp"
28
+ ]
29
+ },
30
+ "pipeline":[
31
+ "transformer",
32
+ "spancat"
33
+ ],
34
+ "components":[
35
+ "transformer",
36
+ "spancat"
37
+ ],
38
+ "disabled":[
39
+
40
+ ],
41
+ "performance":{
42
+ "spans_sc_f":0.0,
43
+ "spans_sc_p":0.0,
44
+ "spans_sc_r":0.0,
45
+ "spans_Agreement_errors_f":0.0,
46
+ "spans_Prepositions_f":0.0,
47
+ "spans_Redundant_comp_f":0.0,
48
+ "spans_Tense_choice_f":0.0,
49
+ "transformer_loss":155.2905472092,
50
+ "spancat_loss":4539.9674681051
51
+ }
52
+ }
training/grammar_major/spancat/cfg ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels":[
3
+ "Agreement_errors",
4
+ "Prepositions",
5
+ "Tense_choice",
6
+ "Redundant_comp"
7
+ ],
8
+ "spans_key":"grammar_major",
9
+ "threshold":0.5,
10
+ "max_positive":null
11
+ }
training/grammar_major/spancat/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e9fbf9b48ca97b5fe06e25c549763cea2b182321819a81a2ea9b61eede20cfa
3
+ size 4725039
training/grammar_major/tokenizer ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/grammar_major/transformer/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "max_batch_items":4096
3
+ }
training/grammar_major/transformer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:745dee6d568bdaf5acde6d3cb6fd6cccfed39741771b1f36ea1d94968d43b2ff
3
+ size 434208023
training/grammar_major/vocab/key2row ADDED
@@ -0,0 +1 @@
 
1
+
training/grammar_major/vocab/lookups.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76be8b528d0075f7aae98d6fa57a6d3c83ae480a8469e668d7b0af968995ac71
3
+ size 1
training/grammar_major/vocab/strings.json ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef59bf11f594ff3dbb5802fd852d888513130a338f18d441d0738d16cfd1d9f
3
+ size 173293
training/grammar_major/vocab/vectors ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14772b683e726436d5948ad3fff2b43d036ef2ebbe3458aafed6004e05a40706
3
+ size 128
training/grammar_major/vocab/vectors.cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "mode":"default"
3
+ }
training/grammar_minor/config.cfg ADDED
@@ -0,0 +1,161 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "./realec/grammar_minor_train.spacy"
3
+ dev = "./realec/grammar_minor_dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ gpu_allocator = "pytorch"
9
+ seed = 0
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["transformer","spancat"]
14
+ batch_size = 128
15
+ disabled = []
16
+ before_creation = null
17
+ after_creation = null
18
+ after_pipeline_creation = null
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.spancat]
24
+ factory = "spancat"
25
+ max_positive = null
26
+ scorer = {"@scorers":"spacy.spancat_scorer.v1"}
27
+ spans_key = "grammar_minor"
28
+ threshold = 0.5
29
+
30
+ [components.spancat.model]
31
+ @architectures = "spacy.SpanCategorizer.v1"
32
+
33
+ [components.spancat.model.reducer]
34
+ @layers = "spacy.mean_max_reducer.v1"
35
+ hidden_size = 128
36
+
37
+ [components.spancat.model.scorer]
38
+ @layers = "spacy.LinearLogistic.v1"
39
+ nO = null
40
+ nI = null
41
+
42
+ [components.spancat.model.tok2vec]
43
+ @architectures = "spacy-transformers.TransformerListener.v1"
44
+ grad_factor = 1.0
45
+ pooling = {"@layers":"reduce_mean.v1"}
46
+ upstream = "*"
47
+
48
+ [components.spancat.suggester]
49
+ @misc = "spacy.ngram_suggester.v1"
50
+ sizes = [1,2,3]
51
+
52
+ [components.transformer]
53
+ factory = "transformer"
54
+ max_batch_items = 4096
55
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
56
+
57
+ [components.transformer.model]
58
+ @architectures = "spacy-transformers.TransformerModel.v3"
59
+ name = "bert-base-cased"
60
+ mixed_precision = false
61
+
62
+ [components.transformer.model.get_spans]
63
+ @span_getters = "spacy-transformers.strided_spans.v1"
64
+ window = 128
65
+ stride = 96
66
+
67
+ [components.transformer.model.grad_scaler_config]
68
+
69
+ [components.transformer.model.tokenizer_config]
70
+ use_fast = true
71
+
72
+ [components.transformer.model.transformer_config]
73
+
74
+ [corpora]
75
+
76
+ [corpora.dev]
77
+ @readers = "spacy.Corpus.v1"
78
+ path = "./realec/grammar_minor_dev.spacy"
79
+ max_length = 0
80
+ gold_preproc = false
81
+ limit = 0
82
+ augmenter = null
83
+
84
+ [corpora.train]
85
+ @readers = "spacy.Corpus.v1"
86
+ path = "./realec/grammar_minor_train.spacy"
87
+ max_length = 0
88
+ gold_preproc = false
89
+ limit = 0
90
+ augmenter = null
91
+
92
+ [training]
93
+ accumulate_gradient = 3
94
+ dev_corpus = "corpora.dev"
95
+ train_corpus = "corpora.train"
96
+ frozen_components = []
97
+ seed = 0
98
+ gpu_allocator = "pytorch"
99
+ dropout = 0.1
100
+ patience = 1600
101
+ max_epochs = 0
102
+ max_steps = 20000
103
+ eval_frequency = 200
104
+ annotating_components = []
105
+ before_to_disk = null
106
+
107
+ [training.batcher]
108
+ @batchers = "spacy.batch_by_padded.v1"
109
+ discard_oversize = true
110
+ size = 2000
111
+ buffer = 256
112
+ get_length = null
113
+
114
+ [training.logger]
115
+ @loggers = "spacy.WandbLogger.v3"
116
+ project_name = "adwiser"
117
+ remove_config_values = ["paths.train","paths.dev","corpora.train.path","corpora.dev.path"]
118
+ model_log_interval = 1000
119
+ log_dataset_dir = null
120
+ entity = null
121
+ run_name = "grammar_minor"
122
+
123
+ [training.optimizer]
124
+ @optimizers = "Adam.v1"
125
+ beta1 = 0.9
126
+ beta2 = 0.999
127
+ L2_is_weight_decay = true
128
+ L2 = 0.01
129
+ grad_clip = 1.0
130
+ use_averages = false
131
+ eps = 0.00000001
132
+
133
+ [training.optimizer.learn_rate]
134
+ @schedules = "warmup_linear.v1"
135
+ warmup_steps = 250
136
+ total_steps = 20000
137
+ initial_rate = 0.00005
138
+
139
+ [training.score_weights]
140
+ spans_sc_f = 0.5
141
+ spans_sc_p = 0.0
142
+ spans_sc_r = 0.0
143
+ spans_Determiners_f = 0.18
144
+ spans_Noun_number_f = 0.01
145
+ spans_Numerals_f = 0.13
146
+ spans_Verb_pattern_f = 0.12
147
+ spans_Word_order_f = 0.07
148
+
149
+ [pretraining]
150
+
151
+ [initialize]
152
+ vectors = null
153
+ init_tok2vec = null
154
+ vocab_data = null
155
+ lookups = null
156
+ before_init = null
157
+ after_init = null
158
+
159
+ [initialize.components]
160
+
161
+ [initialize.tokenizer]
training/grammar_minor/meta.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lang":"en",
3
+ "name":"pipeline",
4
+ "version":"0.0.0",
5
+ "spacy_version":">=3.3.0,<3.4.0",
6
+ "description":"",
7
+ "author":"",
8
+ "email":"",
9
+ "url":"",
10
+ "license":"",
11
+ "spacy_git_version":"Unknown",
12
+ "vectors":{
13
+ "width":0,
14
+ "vectors":0,
15
+ "keys":0,
16
+ "name":null,
17
+ "mode":"default"
18
+ },
19
+ "labels":{
20
+ "transformer":[
21
+
22
+ ],
23
+ "spancat":[
24
+ "Verb_pattern",
25
+ "Determiners",
26
+ "Numerals",
27
+ "Word_order",
28
+ "Noun_number"
29
+ ]
30
+ },
31
+ "pipeline":[
32
+ "transformer",
33
+ "spancat"
34
+ ],
35
+ "components":[
36
+ "transformer",
37
+ "spancat"
38
+ ],
39
+ "disabled":[
40
+
41
+ ],
42
+ "performance":{
43
+ "spans_sc_f":0.0,
44
+ "spans_sc_p":0.0,
45
+ "spans_sc_r":0.0,
46
+ "spans_Determiners_f":0.0,
47
+ "spans_Noun_number_f":0.0,
48
+ "spans_Numerals_f":0.0,
49
+ "spans_Verb_pattern_f":0.0,
50
+ "spans_Word_order_f":0.0,
51
+ "transformer_loss":74.2849839284,
52
+ "spancat_loss":3224.212622695
53
+ }
54
+ }
training/grammar_minor/spancat/cfg ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels":[
3
+ "Verb_pattern",
4
+ "Determiners",
5
+ "Numerals",
6
+ "Word_order",
7
+ "Noun_number"
8
+ ],
9
+ "spans_key":"grammar_minor",
10
+ "threshold":0.5,
11
+ "max_positive":null
12
+ }
training/grammar_minor/spancat/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e0283722a11d5a844cd733d76a12574c63e598f946c2d5843a6f81d80ec9dbe
3
+ size 4725555
training/grammar_minor/tokenizer ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/grammar_minor/transformer/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "max_batch_items":4096
3
+ }
training/grammar_minor/transformer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a214501cb8a98fc0d7934072baa83b46d144b0f255f60afd65c190c2ed876d61
3
+ size 434208023
training/grammar_minor/vocab/key2row ADDED
@@ -0,0 +1 @@
 
1
+
training/grammar_minor/vocab/lookups.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76be8b528d0075f7aae98d6fa57a6d3c83ae480a8469e668d7b0af968995ac71
3
+ size 1
training/grammar_minor/vocab/strings.json ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:169ca15549fd804567ae7d8222e75f320265558fb39b5b90add91046eb7a3723
3
+ size 201573
training/grammar_minor/vocab/vectors ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14772b683e726436d5948ad3fff2b43d036ef2ebbe3458aafed6004e05a40706
3
+ size 128
training/grammar_minor/vocab/vectors.cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "mode":"default"
3
+ }
training/punctuation/config.cfg ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "./realec/punctuation_train.spacy"
3
+ dev = "./realec/punctuation_dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ gpu_allocator = "pytorch"
9
+ seed = 0
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["transformer","spancat"]
14
+ batch_size = 128
15
+ disabled = []
16
+ before_creation = null
17
+ after_creation = null
18
+ after_pipeline_creation = null
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.spancat]
24
+ factory = "spancat"
25
+ max_positive = null
26
+ scorer = {"@scorers":"spacy.spancat_scorer.v1"}
27
+ spans_key = "punctuation"
28
+ threshold = 0.5
29
+
30
+ [components.spancat.model]
31
+ @architectures = "spacy.SpanCategorizer.v1"
32
+
33
+ [components.spancat.model.reducer]
34
+ @layers = "spacy.mean_max_reducer.v1"
35
+ hidden_size = 128
36
+
37
+ [components.spancat.model.scorer]
38
+ @layers = "spacy.LinearLogistic.v1"
39
+ nO = null
40
+ nI = null
41
+
42
+ [components.spancat.model.tok2vec]
43
+ @architectures = "spacy-transformers.TransformerListener.v1"
44
+ grad_factor = 1.0
45
+ pooling = {"@layers":"reduce_mean.v1"}
46
+ upstream = "*"
47
+
48
+ [components.spancat.suggester]
49
+ @misc = "spacy.ngram_suggester.v1"
50
+ sizes = [1,2,3]
51
+
52
+ [components.transformer]
53
+ factory = "transformer"
54
+ max_batch_items = 4096
55
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
56
+
57
+ [components.transformer.model]
58
+ @architectures = "spacy-transformers.TransformerModel.v3"
59
+ name = "bert-base-cased"
60
+ mixed_precision = false
61
+
62
+ [components.transformer.model.get_spans]
63
+ @span_getters = "spacy-transformers.strided_spans.v1"
64
+ window = 128
65
+ stride = 96
66
+
67
+ [components.transformer.model.grad_scaler_config]
68
+
69
+ [components.transformer.model.tokenizer_config]
70
+ use_fast = true
71
+
72
+ [components.transformer.model.transformer_config]
73
+
74
+ [corpora]
75
+
76
+ [corpora.dev]
77
+ @readers = "spacy.Corpus.v1"
78
+ path = "./realec/punctuation_dev.spacy"
79
+ max_length = 0
80
+ gold_preproc = false
81
+ limit = 0
82
+ augmenter = null
83
+
84
+ [corpora.train]
85
+ @readers = "spacy.Corpus.v1"
86
+ path = "./realec/punctuation_train.spacy"
87
+ max_length = 0
88
+ gold_preproc = false
89
+ limit = 0
90
+ augmenter = null
91
+
92
+ [training]
93
+ accumulate_gradient = 3
94
+ dev_corpus = "corpora.dev"
95
+ train_corpus = "corpora.train"
96
+ frozen_components = []
97
+ seed = 0
98
+ gpu_allocator = "pytorch"
99
+ dropout = 0.1
100
+ patience = 1600
101
+ max_epochs = 0
102
+ max_steps = 20000
103
+ eval_frequency = 200
104
+ annotating_components = []
105
+ before_to_disk = null
106
+
107
+ [training.batcher]
108
+ @batchers = "spacy.batch_by_padded.v1"
109
+ discard_oversize = true
110
+ size = 2000
111
+ buffer = 256
112
+ get_length = null
113
+
114
+ [training.logger]
115
+ @loggers = "spacy.WandbLogger.v3"
116
+ project_name = "adwiser"
117
+ remove_config_values = ["paths.train","paths.dev","corpora.train.path","corpora.dev.path"]
118
+ model_log_interval = 1000
119
+ log_dataset_dir = null
120
+ entity = null
121
+ run_name = "punctuation"
122
+
123
+ [training.optimizer]
124
+ @optimizers = "Adam.v1"
125
+ beta1 = 0.9
126
+ beta2 = 0.999
127
+ L2_is_weight_decay = true
128
+ L2 = 0.01
129
+ grad_clip = 1.0
130
+ use_averages = false
131
+ eps = 0.00000001
132
+
133
+ [training.optimizer.learn_rate]
134
+ @schedules = "warmup_linear.v1"
135
+ warmup_steps = 250
136
+ total_steps = 20000
137
+ initial_rate = 0.00005
138
+
139
+ [training.score_weights]
140
+ spans_sc_f = 0.5
141
+ spans_sc_p = 0.0
142
+ spans_sc_r = 0.0
143
+ spans_Punctuation_f = 0.5
144
+
145
+ [pretraining]
146
+
147
+ [initialize]
148
+ vectors = null
149
+ init_tok2vec = null
150
+ vocab_data = null
151
+ lookups = null
152
+ before_init = null
153
+ after_init = null
154
+
155
+ [initialize.components]
156
+
157
+ [initialize.tokenizer]
training/punctuation/meta.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lang":"en",
3
+ "name":"pipeline",
4
+ "version":"0.0.0",
5
+ "spacy_version":">=3.3.0,<3.4.0",
6
+ "description":"",
7
+ "author":"",
8
+ "email":"",
9
+ "url":"",
10
+ "license":"",
11
+ "spacy_git_version":"Unknown",
12
+ "vectors":{
13
+ "width":0,
14
+ "vectors":0,
15
+ "keys":0,
16
+ "name":null,
17
+ "mode":"default"
18
+ },
19
+ "labels":{
20
+ "transformer":[
21
+
22
+ ],
23
+ "spancat":[
24
+ "Punctuation"
25
+ ]
26
+ },
27
+ "pipeline":[
28
+ "transformer",
29
+ "spancat"
30
+ ],
31
+ "components":[
32
+ "transformer",
33
+ "spancat"
34
+ ],
35
+ "disabled":[
36
+
37
+ ],
38
+ "performance":{
39
+ "spans_sc_f":0.0,
40
+ "spans_sc_p":0.0,
41
+ "spans_sc_r":0.0,
42
+ "spans_Punctuation_f":0.0,
43
+ "transformer_loss":155.9655457946,
44
+ "spancat_loss":2200.9864988542
45
+ }
46
+ }
training/punctuation/spancat/cfg ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels":[
3
+ "Punctuation"
4
+ ],
5
+ "spans_key":"punctuation",
6
+ "threshold":0.5,
7
+ "max_positive":null
8
+ }
training/punctuation/spancat/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80d8d230b00ce3dfdf04de2e52fc4d14b2ac8e98e0c785e889c5e795940c06b2
3
+ size 4723491
training/punctuation/tokenizer ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/punctuation/transformer/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "max_batch_items":4096
3
+ }
training/punctuation/transformer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a3cab4830ec3d78cf951043924a896eb0146f4647f4adf3160ac035127f6fd6
3
+ size 434208023
training/punctuation/vocab/key2row ADDED
@@ -0,0 +1 @@
 
1
+