iproskurina commited on
Commit
c417fef
1 Parent(s): ce365be

Update spaCy pipeline

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +6 -16
  2. config.cfg +41 -92
  3. custom_factories.py +30 -5
  4. en_grammar_checker-any-py3-none-any.whl +2 -2
  5. meta.json +102 -51
  6. tokenizer +2 -2
  7. training/articles/config.cfg +157 -0
  8. training/articles/meta.json +46 -0
  9. training/articles/spancat/cfg +8 -0
  10. training/articles/spancat/model +3 -0
  11. training/articles/tokenizer +3 -0
  12. training/articles/transformer/cfg +3 -0
  13. training/articles/transformer/model +3 -0
  14. training/articles/vocab/key2row +1 -0
  15. training/articles/vocab/lookups.bin +3 -0
  16. training/articles/vocab/strings.json +3 -0
  17. training/articles/vocab/vectors +3 -0
  18. training/articles/vocab/vectors.cfg +3 -0
  19. training/grammar_major/config.cfg +160 -0
  20. training/grammar_major/meta.json +52 -0
  21. training/grammar_major/spancat/cfg +11 -0
  22. training/grammar_major/spancat/model +3 -0
  23. training/grammar_major/tokenizer +3 -0
  24. training/grammar_major/transformer/cfg +3 -0
  25. training/grammar_major/transformer/model +3 -0
  26. training/grammar_major/vocab/key2row +1 -0
  27. training/grammar_major/vocab/lookups.bin +3 -0
  28. training/grammar_major/vocab/strings.json +3 -0
  29. training/grammar_major/vocab/vectors +3 -0
  30. training/grammar_major/vocab/vectors.cfg +3 -0
  31. training/grammar_minor/config.cfg +161 -0
  32. training/grammar_minor/meta.json +54 -0
  33. training/grammar_minor/spancat/cfg +12 -0
  34. training/grammar_minor/spancat/model +3 -0
  35. training/grammar_minor/tokenizer +3 -0
  36. training/grammar_minor/transformer/cfg +3 -0
  37. training/grammar_minor/transformer/model +3 -0
  38. training/grammar_minor/vocab/key2row +1 -0
  39. training/grammar_minor/vocab/lookups.bin +3 -0
  40. training/grammar_minor/vocab/strings.json +3 -0
  41. training/grammar_minor/vocab/vectors +3 -0
  42. training/grammar_minor/vocab/vectors.cfg +3 -0
  43. training/punctuation/config.cfg +157 -0
  44. training/punctuation/meta.json +46 -0
  45. training/punctuation/spancat/cfg +8 -0
  46. training/punctuation/spancat/model +3 -0
  47. training/punctuation/tokenizer +3 -0
  48. training/punctuation/transformer/cfg +3 -0
  49. training/punctuation/transformer/model +3 -0
  50. training/punctuation/vocab/key2row +1 -0
README.md CHANGED
@@ -8,24 +8,14 @@ model-index:
8
  - name: en_grammar_checker
9
  results: []
10
  ---
11
- Essay Grammar Checker
12
-
13
  | Feature | Description |
14
  | --- | --- |
15
  | **Name** | `en_grammar_checker` |
16
- | **Version** | `1.0.1` |
17
- | **spaCy** | `>=3.4.4,<3.5.0` |
18
- | **Default Pipeline** | `errors` |
19
- | **Components** | `errors` |
20
  | **Vectors** | 0 keys, 0 unique vectors (0 dimensions) |
21
  | **Sources** | n/a |
22
- | **License** | `CC BY-SA 3.0` |
23
- | **Author** | [Irina Proskurina]() |
24
-
25
- ### Accuracy
26
-
27
- | Type | Score |
28
- | --- | --- |
29
- | `SPANS_ERRORS_P` | 79.38 |
30
- | `SPANS_ERRORS_R` | 44.77 |
31
- | `SPANS_ERRORS_F` | 57.25 |
8
  - name: en_grammar_checker
9
  results: []
10
  ---
 
 
11
  | Feature | Description |
12
  | --- | --- |
13
  | **Name** | `en_grammar_checker` |
14
+ | **Version** | `1.0.0` |
15
+ | **spaCy** | `>=3.3.0,<3.4.0` |
16
+ | **Default Pipeline** | `punctuation`, `spelling`, `articles`, `grammar_major`, `grammar_minor`, `vocabulary` |
17
+ | **Components** | `punctuation`, `spelling`, `articles`, `grammar_major`, `grammar_minor`, `vocabulary` |
18
  | **Vectors** | 0 keys, 0 unique vectors (0 dimensions) |
19
  | **Sources** | n/a |
20
+ | **License** | n/a |
21
+ | **Author** | [n/a]() |
 
 
 
 
 
 
 
 
config.cfg CHANGED
@@ -1,124 +1,92 @@
1
  [paths]
2
- train = "./realec/train.spacy"
3
- dev = "./realec/dev.spacy"
4
  vectors = null
5
  init_tok2vec = null
6
 
7
  [system]
8
- gpu_allocator = "pytorch"
9
  seed = 0
 
10
 
11
  [nlp]
12
  lang = "en"
13
- pipeline = ["transformer","spancat"]
14
- batch_size = 16
15
  disabled = []
16
  before_creation = null
17
  after_creation = null
18
  after_pipeline_creation = null
 
19
  tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
 
21
  [components]
22
 
23
- [components.spancat]
24
- factory = "spancat"
25
- max_positive = null
26
- scorer = {"@scorers":"spacy.spancat_scorer.v1"}
27
- spans_key = "errors"
28
- threshold = 0.5
29
 
30
- [components.spancat.model]
31
- @architectures = "spacy.SpanCategorizer.v1"
32
 
33
- [components.spancat.model.reducer]
34
- @layers = "spacy.mean_max_reducer.v1"
35
- hidden_size = 128
36
 
37
- [components.spancat.model.scorer]
38
- @layers = "spacy.LinearLogistic.v1"
39
- nO = null
40
- nI = null
41
 
42
- [components.spancat.model.tok2vec]
43
- @architectures = "spacy-transformers.TransformerListener.v1"
44
- grad_factor = 1.0
45
- pooling = {"@layers":"reduce_mean.v1"}
46
- upstream = "*"
47
 
48
- [components.spancat.suggester]
49
- @misc = "spacy.ngram_suggester.v1"
50
- sizes = [1,2,3]
51
-
52
- [components.transformer]
53
- factory = "transformer"
54
- max_batch_items = 4096
55
- set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
56
-
57
- [components.transformer.model]
58
- @architectures = "spacy-transformers.TransformerModel.v3"
59
- name = "bert-base-cased"
60
- mixed_precision = false
61
-
62
- [components.transformer.model.get_spans]
63
- @span_getters = "spacy-transformers.strided_spans.v1"
64
- window = 128
65
- stride = 96
66
-
67
- [components.transformer.model.grad_scaler_config]
68
-
69
- [components.transformer.model.tokenizer_config]
70
- use_fast = true
71
-
72
- [components.transformer.model.transformer_config]
73
 
74
  [corpora]
75
 
76
  [corpora.dev]
77
  @readers = "spacy.Corpus.v1"
78
- path = "./realec/dev.spacy"
79
- max_length = 0
80
  gold_preproc = false
 
81
  limit = 0
82
  augmenter = null
83
 
84
  [corpora.train]
85
  @readers = "spacy.Corpus.v1"
86
- path = "./realec/train.spacy"
87
- max_length = 0
88
  gold_preproc = false
 
89
  limit = 0
90
  augmenter = null
91
 
92
  [training]
93
- accumulate_gradient = 3
94
- dev_corpus = "corpora.dev"
95
- train_corpus = "corpora.train"
96
- seed = 0
97
- gpu_allocator = "pytorch"
98
  dropout = 0.1
 
99
  patience = 1600
100
  max_epochs = 0
101
  max_steps = 20000
102
  eval_frequency = 200
103
  frozen_components = []
104
  annotating_components = []
 
 
105
  before_to_disk = null
106
 
107
  [training.batcher]
108
- @batchers = "spacy.batch_by_padded.v1"
109
- discard_oversize = true
110
- size = 2000
111
- buffer = 256
112
  get_length = null
113
 
 
 
 
 
 
 
 
114
  [training.logger]
115
- @loggers = "spacy.WandbLogger.v3"
116
- project_name = "my-awesome-project"
117
- remove_config_values = ["paths.train","paths.dev","corpora.train.path","corpora.dev.path"]
118
- log_dataset_dir = null
119
- entity = null
120
- run_name = "grammar-checker"
121
- model_log_interval = null
122
 
123
  [training.optimizer]
124
  @optimizers = "Adam.v1"
@@ -129,34 +97,15 @@ L2 = 0.01
129
  grad_clip = 1.0
130
  use_averages = false
131
  eps = 0.00000001
132
-
133
- [training.optimizer.learn_rate]
134
- @schedules = "warmup_linear.v1"
135
- warmup_steps = 250
136
- total_steps = 20000
137
- initial_rate = 0.00005
138
 
139
  [training.score_weights]
140
- spans_sc_f = 0.5
141
- spans_sc_p = 0.0
142
- spans_sc_r = 0.0
143
- spans_Agreement_errors_f = 0.06
144
- spans_Articles_f = 0.03
145
- spans_Capitalisation_f = 0.05
146
- spans_Formational_affixes_f = 0.1
147
- spans_Noun_number_f = 0.04
148
- spans_Numerals_f = 0.06
149
- spans_Prepositions_f = 0.05
150
- spans_Punctuation_f = 0.03
151
- spans_Spelling_f = 0.02
152
- spans_Tense_choice_f = 0.03
153
- spans_lex_item_choice_f = 0.03
154
 
155
  [pretraining]
156
 
157
  [initialize]
158
- vectors = null
159
- init_tok2vec = null
160
  vocab_data = null
161
  lookups = null
162
  before_init = null
1
  [paths]
2
+ train = null
3
+ dev = null
4
  vectors = null
5
  init_tok2vec = null
6
 
7
  [system]
 
8
  seed = 0
9
+ gpu_allocator = null
10
 
11
  [nlp]
12
  lang = "en"
13
+ pipeline = ["punctuation","spelling","articles","grammar_major","grammar_minor","vocabulary"]
 
14
  disabled = []
15
  before_creation = null
16
  after_creation = null
17
  after_pipeline_creation = null
18
+ batch_size = 1000
19
  tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
 
21
  [components]
22
 
23
+ [components.articles]
24
+ factory = "articles"
 
 
 
 
25
 
26
+ [components.grammar_major]
27
+ factory = "grammar_major"
28
 
29
+ [components.grammar_minor]
30
+ factory = "grammar_minor"
 
31
 
32
+ [components.punctuation]
33
+ factory = "punctuation"
 
 
34
 
35
+ [components.spelling]
36
+ factory = "spelling"
 
 
 
37
 
38
+ [components.vocabulary]
39
+ factory = "vocabulary"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
 
41
  [corpora]
42
 
43
  [corpora.dev]
44
  @readers = "spacy.Corpus.v1"
45
+ path = ${paths.dev}
 
46
  gold_preproc = false
47
+ max_length = 0
48
  limit = 0
49
  augmenter = null
50
 
51
  [corpora.train]
52
  @readers = "spacy.Corpus.v1"
53
+ path = ${paths.train}
 
54
  gold_preproc = false
55
+ max_length = 0
56
  limit = 0
57
  augmenter = null
58
 
59
  [training]
60
+ seed = ${system.seed}
61
+ gpu_allocator = ${system.gpu_allocator}
 
 
 
62
  dropout = 0.1
63
+ accumulate_gradient = 1
64
  patience = 1600
65
  max_epochs = 0
66
  max_steps = 20000
67
  eval_frequency = 200
68
  frozen_components = []
69
  annotating_components = []
70
+ dev_corpus = "corpora.dev"
71
+ train_corpus = "corpora.train"
72
  before_to_disk = null
73
 
74
  [training.batcher]
75
+ @batchers = "spacy.batch_by_words.v1"
76
+ discard_oversize = false
77
+ tolerance = 0.2
 
78
  get_length = null
79
 
80
+ [training.batcher.size]
81
+ @schedules = "compounding.v1"
82
+ start = 100
83
+ stop = 1000
84
+ compound = 1.001
85
+ t = 0.0
86
+
87
  [training.logger]
88
+ @loggers = "spacy.ConsoleLogger.v1"
89
+ progress_bar = false
 
 
 
 
 
90
 
91
  [training.optimizer]
92
  @optimizers = "Adam.v1"
97
  grad_clip = 1.0
98
  use_averages = false
99
  eps = 0.00000001
100
+ learn_rate = 0.001
 
 
 
 
 
101
 
102
  [training.score_weights]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
103
 
104
  [pretraining]
105
 
106
  [initialize]
107
+ vectors = ${paths.vectors}
108
+ init_tok2vec = ${paths.init_tok2vec}
109
  vocab_data = null
110
  lookups = null
111
  before_init = null
custom_factories.py CHANGED
@@ -9,8 +9,33 @@ meta = get_model_meta(model_path)
9
  data_dir = f"{meta['lang']}_{meta['name']}-{meta['version']}"
10
  components_path = model_path / data_dir / "training"
11
 
12
- @Language.component("errors")
13
- def errors(doc):
14
- nlp_vocabulary = spacy.load(components_path)
15
- print("Loaded component")
16
- return nlp_vocabulary(doc)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  data_dir = f"{meta['lang']}_{meta['name']}-{meta['version']}"
10
  components_path = model_path / data_dir / "training"
11
 
12
+ @Language.component("vocabulary")
13
+ def vocabulary(doc):
14
+ nlp_vocabulary = spacy.load(components_path/"vocabulary")
15
+ return nlp_vocabulary(doc)
16
+
17
+ @Language.component("articles")
18
+ def articles(doc):
19
+ nlp_articles = spacy.load(components_path/"articles")
20
+ return nlp_articles(doc)
21
+
22
+ @Language.component("punctuation")
23
+ def punctuation(doc):
24
+ nlp_punctuation = spacy.load(components_path/"punctuation")
25
+ return nlp_punctuation(doc)
26
+
27
+ @Language.component("spelling")
28
+ def spelling(doc):
29
+ nlp_spelling = spacy.load(components_path/"spelling")
30
+ return nlp_spelling(doc)
31
+
32
+ @Language.component("grammar_major")
33
+ def grammar_major(doc):
34
+ nlp_grammar_major = spacy.load(components_path/"grammar_major")
35
+ return nlp_grammar_major(doc)
36
+
37
+ @Language.component("grammar_minor")
38
+ def grammar_minor(doc):
39
+ nlp_grammar_minor = spacy.load(components_path/"grammar_minor")
40
+ return nlp_grammar_minor(doc)
41
+
en_grammar_checker-any-py3-none-any.whl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bce34b4aa2e04c41a3fada9ada855e4745a8ca7e8858e63b672f569c2255fdc6
3
- size 406893070
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0a808972cae10f9b445a2435a70ee3664f8433086d3b3862febe7679cfb321e
3
+ size 2441454414
meta.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "lang":"en",
3
  "name":"grammar_checker",
4
- "version":"1.0.1",
5
  "description":"Essay Grammar Checker",
6
  "author":"Irina Proskurina",
7
  "email":"",
8
  "url":"",
9
  "license":"CC BY-SA 3.0",
10
- "spacy_version":">=3.4.4,<3.5.0",
11
  "parent_package":"spacy",
12
  "requirements":[
13
  "spacy-transformers>=1.0.0,<1.1.0"
@@ -24,77 +24,128 @@
24
  "name":null
25
  },
26
  "pipeline":[
27
- "errors"
 
 
 
 
 
28
  ],
29
  "components":[
30
- "errors"
 
 
 
 
 
31
  ],
32
  "labels":{
33
 
34
  },
35
  "performance":{
36
- "spans_errors_p":0.7937892339,
37
- "spans_errors_r":0.4476503759,
38
- "spans_errors_f":0.5724644939,
39
- "spans_errors_per_type":{
40
- "Numerals":{
41
- "p":0.7313328681,
42
- "r":0.577092511,
43
- "f":0.6451215759
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
  },
45
  "lex_item_choice":{
46
- "p":0.7750791975,
47
- "r":0.1950571353,
48
- "f":0.3116772824
49
  },
50
- "Articles":{
51
- "p":0.785046729,
52
- "r":0.4552258065,
53
- "f":0.5762822607
54
  },
55
- "Punctuation":{
56
- "p":0.6955835962,
57
- "r":0.2376077586,
58
- "f":0.3542168675
 
 
 
 
 
 
 
 
 
 
59
  },
60
- "Prepositions":{
61
- "p":0.8163471241,
62
- "r":0.3766294227,
63
- "f":0.5154507805
64
  },
65
- "Formational_affixes":{
66
- "p":0.7269700333,
67
- "r":0.6031307551,
68
- "f":0.6592853548
 
 
 
 
 
69
  },
 
 
 
 
 
 
 
 
 
 
70
  "Agreement_errors":{
71
- "p":0.7909018356,
72
- "r":0.5164147994,
73
- "f":0.6248423707
74
  },
75
- "Capitalisation":{
76
- "p":0.8034148593,
77
- "r":0.7899274047,
78
- "f":0.7966140471
79
  },
80
- "Noun_number":{
81
- "p":0.8251445087,
82
- "r":0.5558079169,
83
- "f":0.6642109345
84
  },
85
  "Tense_choice":{
86
- "p":0.7827648115,
87
- "r":0.5369458128,
88
- "f":0.6369612856
89
- },
90
- "Spelling":{
91
- "p":0.886746988,
92
- "r":0.4357608052,
93
- "f":0.5843588726
94
  }
95
  }
96
  },
97
- "speed":2779.5295317788,
98
  "spacy_git_version":"61dfdd9fb",
99
  "disabled":[
100
 
1
  {
2
  "lang":"en",
3
  "name":"grammar_checker",
4
+ "version":"1.0.0",
5
  "description":"Essay Grammar Checker",
6
  "author":"Irina Proskurina",
7
  "email":"",
8
  "url":"",
9
  "license":"CC BY-SA 3.0",
10
+ "spacy_version":">=3.3.0,<3.4.0",
11
  "parent_package":"spacy",
12
  "requirements":[
13
  "spacy-transformers>=1.0.0,<1.1.0"
24
  "name":null
25
  },
26
  "pipeline":[
27
+ "punctuation",
28
+ "spelling",
29
+ "articles",
30
+ "grammar_major",
31
+ "grammar_minor",
32
+ "vocabulary"
33
  ],
34
  "components":[
35
+ "punctuation",
36
+ "spelling",
37
+ "articles",
38
+ "grammar_major",
39
+ "grammar_minor",
40
+ "vocabulary"
41
  ],
42
  "labels":{
43
 
44
  },
45
  "performance":{
46
+ "spans_articles_p":0.8274481176,
47
+ "spans_articles_r":0.8197202306,
48
+ "spans_articles_f":0.823566046,
49
+ "spans_punctuation_p":0.8723066455,
50
+ "spans_punctuation_r":0.6734372264,
51
+ "spans_punctuation_f":0.7600790514,
52
+ "spans_spelling_p":0.9037660231,
53
+ "spans_spelling_r":0.9159588014,
54
+ "spans_spelling_f":0.9098215644,
55
+ "spans_spelling_per_type":{
56
+ "Spelling":{
57
+ "p":0.9044968253,
58
+ "r":0.9227970539,
59
+ "f":0.9135553018
60
+ },
61
+ "Capitalisation":{
62
+ "p":0.8959762655,
63
+ "r":0.8483146067,
64
+ "f":0.8714942736
65
+ }
66
+ },
67
+ "spans_vocabulary_p":0.8326700274,
68
+ "spans_vocabulary_r":0.4697166234,
69
+ "spans_vocabulary_f":0.6006188,
70
+ "spans_vocabulary_per_type":{
71
+ "lex_part_choice":{
72
+ "p":0.8783783784,
73
+ "r":0.0741444867,
74
+ "f":0.1367461431
75
  },
76
  "lex_item_choice":{
77
+ "p":0.8219627873,
78
+ "r":0.5290802348,
79
+ "f":0.6437755977
80
  },
81
+ "Category_confusion":{
82
+ "p":0.84472708,
83
+ "r":0.5153572583,
84
+ "f":0.6401606426
85
  },
86
+ "Formational_affixes":{
87
+ "p":0.8892307692,
88
+ "r":0.5608020699,
89
+ "f":0.6878222927
90
+ }
91
+ },
92
+ "spans_grammar_minor_p":0.8690835299,
93
+ "spans_grammar_minor_r":0.7119557892,
94
+ "spans_grammar_minor_f":0.7827117807,
95
+ "spans_grammar_minor_per_type":{
96
+ "Verb_pattern":{
97
+ "p":0.7399617591,
98
+ "r":0.2491951062,
99
+ "f":0.3728323699
100
  },
101
+ "Noun_number":{
102
+ "p":0.8889171489,
103
+ "r":0.9032758286,
104
+ "f":0.8960389694
105
  },
106
+ "Word_order":{
107
+ "p":0.8179384203,
108
+ "r":0.2189179506,
109
+ "f":0.3453928773
110
+ },
111
+ "Numerals":{
112
+ "p":0.7062761506,
113
+ "r":0.602426838,
114
+ "f":0.6502311248
115
  },
116
+ "Determiners":{
117
+ "p":0.5853658537,
118
+ "r":0.0231884058,
119
+ "f":0.0446096654
120
+ }
121
+ },
122
+ "spans_grammar_major_p":0.8694734373,
123
+ "spans_grammar_major_r":0.6123730074,
124
+ "spans_grammar_major_f":0.7186197538,
125
+ "spans_grammar_major_per_type":{
126
  "Agreement_errors":{
127
+ "p":0.8953090397,
128
+ "r":0.7739296902,
129
+ "f":0.8302062914
130
  },
131
+ "Prepositions":{
132
+ "p":0.881741712,
133
+ "r":0.5715202053,
134
+ "f":0.6935201401
135
  },
136
+ "Redundant_comp":{
137
+ "p":0.8028656126,
138
+ "r":0.314191802,
139
+ "f":0.4516397999
140
  },
141
  "Tense_choice":{
142
+ "p":0.86231523,
143
+ "r":0.7354107649,
144
+ "f":0.7938231022
 
 
 
 
 
145
  }
146
  }
147
  },
148
+ "speed":5613.7287946692,
149
  "spacy_git_version":"61dfdd9fb",
150
  "disabled":[
151
 
tokenizer CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24ad4ddf9a27837484c7fe2ae8a454167cfc9604ef2e7740a2e74a39cc2c1bc3
3
- size 76990
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/articles/config.cfg ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "./realec/articles_train.spacy"
3
+ dev = "./realec/articles_dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ gpu_allocator = "pytorch"
9
+ seed = 0
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["transformer","spancat"]
14
+ batch_size = 128
15
+ disabled = []
16
+ before_creation = null
17
+ after_creation = null
18
+ after_pipeline_creation = null
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.spancat]
24
+ factory = "spancat"
25
+ max_positive = null
26
+ scorer = {"@scorers":"spacy.spancat_scorer.v1"}
27
+ spans_key = "articles"
28
+ threshold = 0.5
29
+
30
+ [components.spancat.model]
31
+ @architectures = "spacy.SpanCategorizer.v1"
32
+
33
+ [components.spancat.model.reducer]
34
+ @layers = "spacy.mean_max_reducer.v1"
35
+ hidden_size = 128
36
+
37
+ [components.spancat.model.scorer]
38
+ @layers = "spacy.LinearLogistic.v1"
39
+ nO = null
40
+ nI = null
41
+
42
+ [components.spancat.model.tok2vec]
43
+ @architectures = "spacy-transformers.TransformerListener.v1"
44
+ grad_factor = 1.0
45
+ pooling = {"@layers":"reduce_mean.v1"}
46
+ upstream = "*"
47
+
48
+ [components.spancat.suggester]
49
+ @misc = "spacy.ngram_suggester.v1"
50
+ sizes = [1,2,3]
51
+
52
+ [components.transformer]
53
+ factory = "transformer"
54
+ max_batch_items = 4096
55
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
56
+
57
+ [components.transformer.model]
58
+ @architectures = "spacy-transformers.TransformerModel.v3"
59
+ name = "bert-base-cased"
60
+ mixed_precision = false
61
+
62
+ [components.transformer.model.get_spans]
63
+ @span_getters = "spacy-transformers.strided_spans.v1"
64
+ window = 128
65
+ stride = 96
66
+
67
+ [components.transformer.model.grad_scaler_config]
68
+
69
+ [components.transformer.model.tokenizer_config]
70
+ use_fast = true
71
+
72
+ [components.transformer.model.transformer_config]
73
+
74
+ [corpora]
75
+
76
+ [corpora.dev]
77
+ @readers = "spacy.Corpus.v1"
78
+ path = "./realec/articles_dev.spacy"
79
+ max_length = 0
80
+ gold_preproc = false
81
+ limit = 0
82
+ augmenter = null
83
+
84
+ [corpora.train]
85
+ @readers = "spacy.Corpus.v1"
86
+ path = "./realec/articles_train.spacy"
87
+ max_length = 0
88
+ gold_preproc = false
89
+ limit = 0
90
+ augmenter = null
91
+
92
+ [training]
93
+ accumulate_gradient = 3
94
+ dev_corpus = "corpora.dev"
95
+ train_corpus = "corpora.train"
96
+ frozen_components = []
97
+ seed = 0
98
+ gpu_allocator = "pytorch"
99
+ dropout = 0.1
100
+ patience = 1600
101
+ max_epochs = 0
102
+ max_steps = 20000
103
+ eval_frequency = 200
104
+ annotating_components = []
105
+ before_to_disk = null
106
+
107
+ [training.batcher]
108
+ @batchers = "spacy.batch_by_padded.v1"
109
+ discard_oversize = true
110
+ size = 2000
111
+ buffer = 256
112
+ get_length = null
113
+
114
+ [training.logger]
115
+ @loggers = "spacy.WandbLogger.v3"
116
+ project_name = "adwiser"
117
+ remove_config_values = ["paths.train","paths.dev","corpora.train.path","corpora.dev.path"]
118
+ model_log_interval = 1000
119
+ log_dataset_dir = null
120
+ entity = null
121
+ run_name = "articles"
122
+
123
+ [training.optimizer]
124
+ @optimizers = "Adam.v1"
125
+ beta1 = 0.9
126
+ beta2 = 0.999
127
+ L2_is_weight_decay = true
128
+ L2 = 0.01
129
+ grad_clip = 1.0
130
+ use_averages = false
131
+ eps = 0.00000001
132
+
133
+ [training.optimizer.learn_rate]
134
+ @schedules = "warmup_linear.v1"
135
+ warmup_steps = 250
136
+ total_steps = 20000
137
+ initial_rate = 0.00005
138
+
139
+ [training.score_weights]
140
+ spans_sc_f = 0.5
141
+ spans_sc_p = 0.0
142
+ spans_sc_r = 0.0
143
+ spans_Articles_f = 0.5
144
+
145
+ [pretraining]
146
+
147
+ [initialize]
148
+ vectors = null
149
+ init_tok2vec = null
150
+ vocab_data = null
151
+ lookups = null
152
+ before_init = null
153
+ after_init = null
154
+
155
+ [initialize.components]
156
+
157
+ [initialize.tokenizer]
training/articles/meta.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lang":"en",
3
+ "name":"pipeline",
4
+ "version":"0.0.0",
5
+ "spacy_version":">=3.3.0,<3.4.0",
6
+ "description":"",
7
+ "author":"",
8
+ "email":"",
9
+ "url":"",
10
+ "license":"",
11
+ "spacy_git_version":"Unknown",
12
+ "vectors":{
13
+ "width":0,
14
+ "vectors":0,
15
+ "keys":0,
16
+ "name":null,
17
+ "mode":"default"
18
+ },
19
+ "labels":{
20
+ "transformer":[
21
+
22
+ ],
23
+ "spancat":[
24
+ "Articles"
25
+ ]
26
+ },
27
+ "pipeline":[
28
+ "transformer",
29
+ "spancat"
30
+ ],
31
+ "components":[
32
+ "transformer",
33
+ "spancat"
34
+ ],
35
+ "disabled":[
36
+
37
+ ],
38
+ "performance":{
39
+ "spans_sc_f":0.0,
40
+ "spans_sc_p":0.0,
41
+ "spans_sc_r":0.0,
42
+ "spans_Articles_f":0.0,
43
+ "transformer_loss":100.8329219741,
44
+ "spancat_loss":2964.9278720927
45
+ }
46
+ }
training/articles/spancat/cfg ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels":[
3
+ "Articles"
4
+ ],
5
+ "spans_key":"articles",
6
+ "threshold":0.5,
7
+ "max_positive":null
8
+ }
training/articles/spancat/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bcd34ab7d5ff9adb0214059514deb98fb9c131f29d0a0a7a4c520fcafd3e875
3
+ size 4723491
training/articles/tokenizer ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/articles/transformer/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "max_batch_items":4096
3
+ }
training/articles/transformer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2036ce8a5df6924d046d42dd67dc03f670b4d1d25c2d8073de95822165e9a5ed
3
+ size 434208023
training/articles/vocab/key2row ADDED
@@ -0,0 +1 @@
 
1
+
training/articles/vocab/lookups.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76be8b528d0075f7aae98d6fa57a6d3c83ae480a8469e668d7b0af968995ac71
3
+ size 1
training/articles/vocab/strings.json ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6c403d7808ec8eeeafe465538d7df7f45c1202362b29c9c627a7ed8302dc210
3
+ size 287029
training/articles/vocab/vectors ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14772b683e726436d5948ad3fff2b43d036ef2ebbe3458aafed6004e05a40706
3
+ size 128
training/articles/vocab/vectors.cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "mode":"default"
3
+ }
training/grammar_major/config.cfg ADDED
@@ -0,0 +1,160 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "./realec/grammar_major_train.spacy"
3
+ dev = "./realec/grammar_major_dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ gpu_allocator = "pytorch"
9
+ seed = 0
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["transformer","spancat"]
14
+ batch_size = 128
15
+ disabled = []
16
+ before_creation = null
17
+ after_creation = null
18
+ after_pipeline_creation = null
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.spancat]
24
+ factory = "spancat"
25
+ max_positive = null
26
+ scorer = {"@scorers":"spacy.spancat_scorer.v1"}
27
+ spans_key = "grammar_major"
28
+ threshold = 0.5
29
+
30
+ [components.spancat.model]
31
+ @architectures = "spacy.SpanCategorizer.v1"
32
+
33
+ [components.spancat.model.reducer]
34
+ @layers = "spacy.mean_max_reducer.v1"
35
+ hidden_size = 128
36
+
37
+ [components.spancat.model.scorer]
38
+ @layers = "spacy.LinearLogistic.v1"
39
+ nO = null
40
+ nI = null
41
+
42
+ [components.spancat.model.tok2vec]
43
+ @architectures = "spacy-transformers.TransformerListener.v1"
44
+ grad_factor = 1.0
45
+ pooling = {"@layers":"reduce_mean.v1"}
46
+ upstream = "*"
47
+
48
+ [components.spancat.suggester]
49
+ @misc = "spacy.ngram_suggester.v1"
50
+ sizes = [1,2,3]
51
+
52
+ [components.transformer]
53
+ factory = "transformer"
54
+ max_batch_items = 4096
55
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
56
+
57
+ [components.transformer.model]
58
+ @architectures = "spacy-transformers.TransformerModel.v3"
59
+ name = "bert-base-cased"
60
+ mixed_precision = false
61
+
62
+ [components.transformer.model.get_spans]
63
+ @span_getters = "spacy-transformers.strided_spans.v1"
64
+ window = 128
65
+ stride = 96
66
+
67
+ [components.transformer.model.grad_scaler_config]
68
+
69
+ [components.transformer.model.tokenizer_config]
70
+ use_fast = true
71
+
72
+ [components.transformer.model.transformer_config]
73
+
74
+ [corpora]
75
+
76
+ [corpora.dev]
77
+ @readers = "spacy.Corpus.v1"
78
+ path = "./realec/grammar_major_dev.spacy"
79
+ max_length = 0
80
+ gold_preproc = false
81
+ limit = 0
82
+ augmenter = null
83
+
84
+ [corpora.train]
85
+ @readers = "spacy.Corpus.v1"
86
+ path = "./realec/grammar_major_train.spacy"
87
+ max_length = 0
88
+ gold_preproc = false
89
+ limit = 0
90
+ augmenter = null
91
+
92
+ [training]
93
+ accumulate_gradient = 3
94
+ dev_corpus = "corpora.dev"
95
+ train_corpus = "corpora.train"
96
+ frozen_components = []
97
+ seed = 0
98
+ gpu_allocator = "pytorch"
99
+ dropout = 0.1
100
+ patience = 1600
101
+ max_epochs = 0
102
+ max_steps = 20000
103
+ eval_frequency = 200
104
+ annotating_components = []
105
+ before_to_disk = null
106
+
107
+ [training.batcher]
108
+ @batchers = "spacy.batch_by_padded.v1"
109
+ discard_oversize = true
110
+ size = 2000
111
+ buffer = 256
112
+ get_length = null
113
+
114
+ [training.logger]
115
+ @loggers = "spacy.WandbLogger.v3"
116
+ project_name = "adwiser"
117
+ remove_config_values = ["paths.train","paths.dev","corpora.train.path","corpora.dev.path"]
118
+ model_log_interval = 1000
119
+ log_dataset_dir = null
120
+ entity = null
121
+ run_name = "grammar_major"
122
+
123
+ [training.optimizer]
124
+ @optimizers = "Adam.v1"
125
+ beta1 = 0.9
126
+ beta2 = 0.999
127
+ L2_is_weight_decay = true
128
+ L2 = 0.01
129
+ grad_clip = 1.0
130
+ use_averages = false
131
+ eps = 0.00000001
132
+
133
+ [training.optimizer.learn_rate]
134
+ @schedules = "warmup_linear.v1"
135
+ warmup_steps = 250
136
+ total_steps = 20000
137
+ initial_rate = 0.00005
138
+
139
+ [training.score_weights]
140
+ spans_sc_f = 0.5
141
+ spans_sc_p = 0.0
142
+ spans_sc_r = 0.0
143
+ spans_Agreement_errors_f = 0.13
144
+ spans_Prepositions_f = 0.12
145
+ spans_Redundant_comp_f = 0.14
146
+ spans_Tense_choice_f = 0.11
147
+
148
+ [pretraining]
149
+
150
+ [initialize]
151
+ vectors = null
152
+ init_tok2vec = null
153
+ vocab_data = null
154
+ lookups = null
155
+ before_init = null
156
+ after_init = null
157
+
158
+ [initialize.components]
159
+
160
+ [initialize.tokenizer]
training/grammar_major/meta.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lang":"en",
3
+ "name":"pipeline",
4
+ "version":"0.0.0",
5
+ "spacy_version":">=3.3.0,<3.4.0",
6
+ "description":"",
7
+ "author":"",
8
+ "email":"",
9
+ "url":"",
10
+ "license":"",
11
+ "spacy_git_version":"Unknown",
12
+ "vectors":{
13
+ "width":0,
14
+ "vectors":0,
15
+ "keys":0,
16
+ "name":null,
17
+ "mode":"default"
18
+ },
19
+ "labels":{
20
+ "transformer":[
21
+
22
+ ],
23
+ "spancat":[
24
+ "Agreement_errors",
25
+ "Prepositions",
26
+ "Tense_choice",
27
+ "Redundant_comp"
28
+ ]
29
+ },
30
+ "pipeline":[
31
+ "transformer",
32
+ "spancat"
33
+ ],
34
+ "components":[
35
+ "transformer",
36
+ "spancat"
37
+ ],
38
+ "disabled":[
39
+
40
+ ],
41
+ "performance":{
42
+ "spans_sc_f":0.0,
43
+ "spans_sc_p":0.0,
44
+ "spans_sc_r":0.0,
45
+ "spans_Agreement_errors_f":0.0,
46
+ "spans_Prepositions_f":0.0,
47
+ "spans_Redundant_comp_f":0.0,
48
+ "spans_Tense_choice_f":0.0,
49
+ "transformer_loss":155.2905472092,
50
+ "spancat_loss":4539.9674681051
51
+ }
52
+ }
training/grammar_major/spancat/cfg ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels":[
3
+ "Agreement_errors",
4
+ "Prepositions",
5
+ "Tense_choice",
6
+ "Redundant_comp"
7
+ ],
8
+ "spans_key":"grammar_major",
9
+ "threshold":0.5,
10
+ "max_positive":null
11
+ }
training/grammar_major/spancat/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e9fbf9b48ca97b5fe06e25c549763cea2b182321819a81a2ea9b61eede20cfa
3
+ size 4725039
training/grammar_major/tokenizer ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/grammar_major/transformer/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "max_batch_items":4096
3
+ }
training/grammar_major/transformer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:745dee6d568bdaf5acde6d3cb6fd6cccfed39741771b1f36ea1d94968d43b2ff
3
+ size 434208023
training/grammar_major/vocab/key2row ADDED
@@ -0,0 +1 @@
 
1
+
training/grammar_major/vocab/lookups.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76be8b528d0075f7aae98d6fa57a6d3c83ae480a8469e668d7b0af968995ac71
3
+ size 1
training/grammar_major/vocab/strings.json ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef59bf11f594ff3dbb5802fd852d888513130a338f18d441d0738d16cfd1d9f
3
+ size 173293
training/grammar_major/vocab/vectors ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14772b683e726436d5948ad3fff2b43d036ef2ebbe3458aafed6004e05a40706
3
+ size 128
training/grammar_major/vocab/vectors.cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "mode":"default"
3
+ }
training/grammar_minor/config.cfg ADDED
@@ -0,0 +1,161 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "./realec/grammar_minor_train.spacy"
3
+ dev = "./realec/grammar_minor_dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ gpu_allocator = "pytorch"
9
+ seed = 0
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["transformer","spancat"]
14
+ batch_size = 128
15
+ disabled = []
16
+ before_creation = null
17
+ after_creation = null
18
+ after_pipeline_creation = null
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.spancat]
24
+ factory = "spancat"
25
+ max_positive = null
26
+ scorer = {"@scorers":"spacy.spancat_scorer.v1"}
27
+ spans_key = "grammar_minor"
28
+ threshold = 0.5
29
+
30
+ [components.spancat.model]
31
+ @architectures = "spacy.SpanCategorizer.v1"
32
+
33
+ [components.spancat.model.reducer]
34
+ @layers = "spacy.mean_max_reducer.v1"
35
+ hidden_size = 128
36
+
37
+ [components.spancat.model.scorer]
38
+ @layers = "spacy.LinearLogistic.v1"
39
+ nO = null
40
+ nI = null
41
+
42
+ [components.spancat.model.tok2vec]
43
+ @architectures = "spacy-transformers.TransformerListener.v1"
44
+ grad_factor = 1.0
45
+ pooling = {"@layers":"reduce_mean.v1"}
46
+ upstream = "*"
47
+
48
+ [components.spancat.suggester]
49
+ @misc = "spacy.ngram_suggester.v1"
50
+ sizes = [1,2,3]
51
+
52
+ [components.transformer]
53
+ factory = "transformer"
54
+ max_batch_items = 4096
55
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
56
+
57
+ [components.transformer.model]
58
+ @architectures = "spacy-transformers.TransformerModel.v3"
59
+ name = "bert-base-cased"
60
+ mixed_precision = false
61
+
62
+ [components.transformer.model.get_spans]
63
+ @span_getters = "spacy-transformers.strided_spans.v1"
64
+ window = 128
65
+ stride = 96
66
+
67
+ [components.transformer.model.grad_scaler_config]
68
+
69
+ [components.transformer.model.tokenizer_config]
70
+ use_fast = true
71
+
72
+ [components.transformer.model.transformer_config]
73
+
74
+ [corpora]
75
+
76
+ [corpora.dev]
77
+ @readers = "spacy.Corpus.v1"
78
+ path = "./realec/grammar_minor_dev.spacy"
79
+ max_length = 0
80
+ gold_preproc = false
81
+ limit = 0
82
+ augmenter = null
83
+
84
+ [corpora.train]
85
+ @readers = "spacy.Corpus.v1"
86
+ path = "./realec/grammar_minor_train.spacy"
87
+ max_length = 0
88
+ gold_preproc = false
89
+ limit = 0
90
+ augmenter = null
91
+
92
+ [training]
93
+ accumulate_gradient = 3
94
+ dev_corpus = "corpora.dev"
95
+ train_corpus = "corpora.train"
96
+ frozen_components = []
97
+ seed = 0
98
+ gpu_allocator = "pytorch"
99
+ dropout = 0.1
100
+ patience = 1600
101
+ max_epochs = 0
102
+ max_steps = 20000
103
+ eval_frequency = 200
104
+ annotating_components = []
105
+ before_to_disk = null
106
+
107
+ [training.batcher]
108
+ @batchers = "spacy.batch_by_padded.v1"
109
+ discard_oversize = true
110
+ size = 2000
111
+ buffer = 256
112
+ get_length = null
113
+
114
+ [training.logger]
115
+ @loggers = "spacy.WandbLogger.v3"
116
+ project_name = "adwiser"
117
+ remove_config_values = ["paths.train","paths.dev","corpora.train.path","corpora.dev.path"]
118
+ model_log_interval = 1000
119
+ log_dataset_dir = null
120
+ entity = null
121
+ run_name = "grammar_minor"
122
+
123
+ [training.optimizer]
124
+ @optimizers = "Adam.v1"
125
+ beta1 = 0.9
126
+ beta2 = 0.999
127
+ L2_is_weight_decay = true
128
+ L2 = 0.01
129
+ grad_clip = 1.0
130
+ use_averages = false
131
+ eps = 0.00000001
132
+
133
+ [training.optimizer.learn_rate]
134
+ @schedules = "warmup_linear.v1"
135
+ warmup_steps = 250
136
+ total_steps = 20000
137
+ initial_rate = 0.00005
138
+
139
+ [training.score_weights]
140
+ spans_sc_f = 0.5
141
+ spans_sc_p = 0.0
142
+ spans_sc_r = 0.0
143
+ spans_Determiners_f = 0.18
144
+ spans_Noun_number_f = 0.01
145
+ spans_Numerals_f = 0.13
146
+ spans_Verb_pattern_f = 0.12
147
+ spans_Word_order_f = 0.07
148
+
149
+ [pretraining]
150
+
151
+ [initialize]
152
+ vectors = null
153
+ init_tok2vec = null
154
+ vocab_data = null
155
+ lookups = null
156
+ before_init = null
157
+ after_init = null
158
+
159
+ [initialize.components]
160
+
161
+ [initialize.tokenizer]
training/grammar_minor/meta.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lang":"en",
3
+ "name":"pipeline",
4
+ "version":"0.0.0",
5
+ "spacy_version":">=3.3.0,<3.4.0",
6
+ "description":"",
7
+ "author":"",
8
+ "email":"",
9
+ "url":"",
10
+ "license":"",
11
+ "spacy_git_version":"Unknown",
12
+ "vectors":{
13
+ "width":0,
14
+ "vectors":0,
15
+ "keys":0,
16
+ "name":null,
17
+ "mode":"default"
18
+ },
19
+ "labels":{
20
+ "transformer":[
21
+
22
+ ],
23
+ "spancat":[
24
+ "Verb_pattern",
25
+ "Determiners",
26
+ "Numerals",
27
+ "Word_order",
28
+ "Noun_number"
29
+ ]
30
+ },
31
+ "pipeline":[
32
+ "transformer",
33
+ "spancat"
34
+ ],
35
+ "components":[
36
+ "transformer",
37
+ "spancat"
38
+ ],
39
+ "disabled":[
40
+
41
+ ],
42
+ "performance":{
43
+ "spans_sc_f":0.0,
44
+ "spans_sc_p":0.0,
45
+ "spans_sc_r":0.0,
46
+ "spans_Determiners_f":0.0,
47
+ "spans_Noun_number_f":0.0,
48
+ "spans_Numerals_f":0.0,
49
+ "spans_Verb_pattern_f":0.0,
50
+ "spans_Word_order_f":0.0,
51
+ "transformer_loss":74.2849839284,
52
+ "spancat_loss":3224.212622695
53
+ }
54
+ }
training/grammar_minor/spancat/cfg ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels":[
3
+ "Verb_pattern",
4
+ "Determiners",
5
+ "Numerals",
6
+ "Word_order",
7
+ "Noun_number"
8
+ ],
9
+ "spans_key":"grammar_minor",
10
+ "threshold":0.5,
11
+ "max_positive":null
12
+ }
training/grammar_minor/spancat/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e0283722a11d5a844cd733d76a12574c63e598f946c2d5843a6f81d80ec9dbe
3
+ size 4725555
training/grammar_minor/tokenizer ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/grammar_minor/transformer/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "max_batch_items":4096
3
+ }
training/grammar_minor/transformer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a214501cb8a98fc0d7934072baa83b46d144b0f255f60afd65c190c2ed876d61
3
+ size 434208023
training/grammar_minor/vocab/key2row ADDED
@@ -0,0 +1 @@
 
1
+
training/grammar_minor/vocab/lookups.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76be8b528d0075f7aae98d6fa57a6d3c83ae480a8469e668d7b0af968995ac71
3
+ size 1
training/grammar_minor/vocab/strings.json ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:169ca15549fd804567ae7d8222e75f320265558fb39b5b90add91046eb7a3723
3
+ size 201573
training/grammar_minor/vocab/vectors ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14772b683e726436d5948ad3fff2b43d036ef2ebbe3458aafed6004e05a40706
3
+ size 128
training/grammar_minor/vocab/vectors.cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "mode":"default"
3
+ }
training/punctuation/config.cfg ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "./realec/punctuation_train.spacy"
3
+ dev = "./realec/punctuation_dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ gpu_allocator = "pytorch"
9
+ seed = 0
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["transformer","spancat"]
14
+ batch_size = 128
15
+ disabled = []
16
+ before_creation = null
17
+ after_creation = null
18
+ after_pipeline_creation = null
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.spancat]
24
+ factory = "spancat"
25
+ max_positive = null
26
+ scorer = {"@scorers":"spacy.spancat_scorer.v1"}
27
+ spans_key = "punctuation"
28
+ threshold = 0.5
29
+
30
+ [components.spancat.model]
31
+ @architectures = "spacy.SpanCategorizer.v1"
32
+
33
+ [components.spancat.model.reducer]
34
+ @layers = "spacy.mean_max_reducer.v1"
35
+ hidden_size = 128
36
+
37
+ [components.spancat.model.scorer]
38
+ @layers = "spacy.LinearLogistic.v1"
39
+ nO = null
40
+ nI = null
41
+
42
+ [components.spancat.model.tok2vec]
43
+ @architectures = "spacy-transformers.TransformerListener.v1"
44
+ grad_factor = 1.0
45
+ pooling = {"@layers":"reduce_mean.v1"}
46
+ upstream = "*"
47
+
48
+ [components.spancat.suggester]
49
+ @misc = "spacy.ngram_suggester.v1"
50
+ sizes = [1,2,3]
51
+
52
+ [components.transformer]
53
+ factory = "transformer"
54
+ max_batch_items = 4096
55
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
56
+
57
+ [components.transformer.model]
58
+ @architectures = "spacy-transformers.TransformerModel.v3"
59
+ name = "bert-base-cased"
60
+ mixed_precision = false
61
+
62
+ [components.transformer.model.get_spans]
63
+ @span_getters = "spacy-transformers.strided_spans.v1"
64
+ window = 128
65
+ stride = 96
66
+
67
+ [components.transformer.model.grad_scaler_config]
68
+
69
+ [components.transformer.model.tokenizer_config]
70
+ use_fast = true
71
+
72
+ [components.transformer.model.transformer_config]
73
+
74
+ [corpora]
75
+
76
+ [corpora.dev]
77
+ @readers = "spacy.Corpus.v1"
78
+ path = "./realec/punctuation_dev.spacy"
79
+ max_length = 0
80
+ gold_preproc = false
81
+ limit = 0
82
+ augmenter = null
83
+
84
+ [corpora.train]
85
+ @readers = "spacy.Corpus.v1"
86
+ path = "./realec/punctuation_train.spacy"
87
+ max_length = 0
88
+ gold_preproc = false
89
+ limit = 0
90
+ augmenter = null
91
+
92
+ [training]
93
+ accumulate_gradient = 3
94
+ dev_corpus = "corpora.dev"
95
+ train_corpus = "corpora.train"
96
+ frozen_components = []
97
+ seed = 0
98
+ gpu_allocator = "pytorch"
99
+ dropout = 0.1
100
+ patience = 1600
101
+ max_epochs = 0
102
+ max_steps = 20000
103
+ eval_frequency = 200
104
+ annotating_components = []
105
+ before_to_disk = null
106
+
107
+ [training.batcher]
108
+ @batchers = "spacy.batch_by_padded.v1"
109
+ discard_oversize = true
110
+ size = 2000
111
+ buffer = 256
112
+ get_length = null
113
+
114
+ [training.logger]
115
+ @loggers = "spacy.WandbLogger.v3"
116
+ project_name = "adwiser"
117
+ remove_config_values = ["paths.train","paths.dev","corpora.train.path","corpora.dev.path"]
118
+ model_log_interval = 1000
119
+ log_dataset_dir = null
120
+ entity = null
121
+ run_name = "punctuation"
122
+
123
+ [training.optimizer]
124
+ @optimizers = "Adam.v1"
125
+ beta1 = 0.9
126
+ beta2 = 0.999
127
+ L2_is_weight_decay = true
128
+ L2 = 0.01
129
+ grad_clip = 1.0
130
+ use_averages = false
131
+ eps = 0.00000001
132
+
133
+ [training.optimizer.learn_rate]
134
+ @schedules = "warmup_linear.v1"
135
+ warmup_steps = 250
136
+ total_steps = 20000
137
+ initial_rate = 0.00005
138
+
139
+ [training.score_weights]
140
+ spans_sc_f = 0.5
141
+ spans_sc_p = 0.0
142
+ spans_sc_r = 0.0
143
+ spans_Punctuation_f = 0.5
144
+
145
+ [pretraining]
146
+
147
+ [initialize]
148
+ vectors = null
149
+ init_tok2vec = null
150
+ vocab_data = null
151
+ lookups = null
152
+ before_init = null
153
+ after_init = null
154
+
155
+ [initialize.components]
156
+
157
+ [initialize.tokenizer]
training/punctuation/meta.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "lang":"en",
3
+ "name":"pipeline",
4
+ "version":"0.0.0",
5
+ "spacy_version":">=3.3.0,<3.4.0",
6
+ "description":"",
7
+ "author":"",
8
+ "email":"",
9
+ "url":"",
10
+ "license":"",
11
+ "spacy_git_version":"Unknown",
12
+ "vectors":{
13
+ "width":0,
14
+ "vectors":0,
15
+ "keys":0,
16
+ "name":null,
17
+ "mode":"default"
18
+ },
19
+ "labels":{
20
+ "transformer":[
21
+
22
+ ],
23
+ "spancat":[
24
+ "Punctuation"
25
+ ]
26
+ },
27
+ "pipeline":[
28
+ "transformer",
29
+ "spancat"
30
+ ],
31
+ "components":[
32
+ "transformer",
33
+ "spancat"
34
+ ],
35
+ "disabled":[
36
+
37
+ ],
38
+ "performance":{
39
+ "spans_sc_f":0.0,
40
+ "spans_sc_p":0.0,
41
+ "spans_sc_r":0.0,
42
+ "spans_Punctuation_f":0.0,
43
+ "transformer_loss":155.9655457946,
44
+ "spancat_loss":2200.9864988542
45
+ }
46
+ }
training/punctuation/spancat/cfg ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "labels":[
3
+ "Punctuation"
4
+ ],
5
+ "spans_key":"punctuation",
6
+ "threshold":0.5,
7
+ "max_positive":null
8
+ }
training/punctuation/spancat/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80d8d230b00ce3dfdf04de2e52fc4d14b2ac8e98e0c785e889c5e795940c06b2
3
+ size 4723491
training/punctuation/tokenizer ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0f3f5297dc14c5fd94488174149bd87b5f24dff28e6be993eb3ea7a92e53417
3
+ size 78126
training/punctuation/transformer/cfg ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ {
2
+ "max_batch_items":4096
3
+ }
training/punctuation/transformer/model ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a3cab4830ec3d78cf951043924a896eb0146f4647f4adf3160ac035127f6fd6
3
+ size 434208023
training/punctuation/vocab/key2row ADDED
@@ -0,0 +1 @@
 
1
+