adrianeboyd commited on
Commit
ae0cb2f
1 Parent(s): d9c0885

Update spaCy pipeline

Browse files
README.md CHANGED
@@ -14,59 +14,59 @@ model-index:
14
  metrics:
15
  - name: TAG (XPOS) Accuracy
16
  type: accuracy
17
- value: 0.9905775076
18
  - task:
19
  name: POS
20
  type: token-classification
21
  metrics:
22
  - name: POS (UPOS) Accuracy
23
  type: accuracy
24
- value: 0.9914628183
25
  - task:
26
  name: MORPH
27
  type: token-classification
28
  metrics:
29
  - name: Morph (UFeats) Accuracy
30
  type: accuracy
31
- value: 0.9700866279
32
  - task:
33
  name: LEMMA
34
  type: token-classification
35
  metrics:
36
  - name: Lemma Accuracy
37
  type: accuracy
38
- value: 0.9872611465
39
  - task:
40
  name: UNLABELED_DEPENDENCIES
41
  type: token-classification
42
  metrics:
43
  - name: Unlabeled Attachment Score (UAS)
44
  type: f_score
45
- value: 0.9592566441
46
  - task:
47
  name: LABELED_DEPENDENCIES
48
  type: token-classification
49
  metrics:
50
  - name: Labeled Attachment Score (LAS)
51
  type: f_score
52
- value: 0.9484326996
53
  - task:
54
  name: SENTS
55
  type: token-classification
56
  metrics:
57
  - name: Sentences F-Score
58
  type: f_score
59
- value: 0.9797854785
60
  ---
61
  ### Details: https://spacy.io/models/de#de_dep_news_trf
62
 
63
- German transformer pipeline (bert-base-german-cased). Components: transformer, tagger, morphologizer, parser, lemmatizer (trainable_lemmatizer).
64
 
65
  | Feature | Description |
66
  | --- | --- |
67
  | **Name** | `de_dep_news_trf` |
68
- | **Version** | `3.6.1` |
69
- | **spaCy** | `>=3.6.0,<3.7.0` |
70
  | **Default Pipeline** | `transformer`, `tagger`, `morphologizer`, `parser`, `lemmatizer`, `attribute_ruler` |
71
  | **Components** | `transformer`, `tagger`, `morphologizer`, `parser`, `lemmatizer`, `attribute_ruler` |
72
  | **Vectors** | 0 keys, 0 unique vectors (0 dimensions) |
@@ -96,15 +96,15 @@ German transformer pipeline (bert-base-german-cased). Components: transformer, t
96
  | `TOKEN_P` | 99.92 |
97
  | `TOKEN_R` | 99.90 |
98
  | `TOKEN_F` | 99.91 |
99
- | `TAG_ACC` | 99.06 |
100
- | `POS_ACC` | 99.15 |
101
  | `MORPH_ACC` | 97.01 |
102
- | `MORPH_MICRO_P` | 98.86 |
103
- | `MORPH_MICRO_R` | 98.84 |
104
- | `MORPH_MICRO_F` | 98.85 |
105
- | `SENTS_P` | 97.80 |
106
- | `SENTS_R` | 98.16 |
107
- | `SENTS_F` | 97.98 |
108
- | `DEP_UAS` | 95.93 |
109
- | `DEP_LAS` | 94.84 |
110
- | `LEMMA_ACC` | 98.73 |
14
  metrics:
15
  - name: TAG (XPOS) Accuracy
16
  type: accuracy
17
+ value: 0.9907095376
18
  - task:
19
  name: POS
20
  type: token-classification
21
  metrics:
22
  - name: POS (UPOS) Accuracy
23
  type: accuracy
24
+ value: 0.9917734358
25
  - task:
26
  name: MORPH
27
  type: token-classification
28
  metrics:
29
  - name: Morph (UFeats) Accuracy
30
  type: accuracy
31
+ value: 0.9700601968
32
  - task:
33
  name: LEMMA
34
  type: token-classification
35
  metrics:
36
  - name: Lemma Accuracy
37
  type: accuracy
38
+ value: 0.9872479319
39
  - task:
40
  name: UNLABELED_DEPENDENCIES
41
  type: token-classification
42
  metrics:
43
  - name: Unlabeled Attachment Score (UAS)
44
  type: f_score
45
+ value: 0.958309546
46
  - task:
47
  name: LABELED_DEPENDENCIES
48
  type: token-classification
49
  metrics:
50
  - name: Labeled Attachment Score (LAS)
51
  type: f_score
52
+ value: 0.9473953052
53
  - task:
54
  name: SENTS
55
  type: token-classification
56
  metrics:
57
  - name: Sentences F-Score
58
  type: f_score
59
+ value: 0.9789517128
60
  ---
61
  ### Details: https://spacy.io/models/de#de_dep_news_trf
62
 
63
+ German transformer pipeline (Transformer(name='bert-base-german-cased', piece_encoder='bert-wordpiece', stride=136, type='bert', width=768, window=176, vocab_size=30000)). Components: transformer, tagger, morphologizer, parser, lemmatizer (trainable_lemmatizer).
64
 
65
  | Feature | Description |
66
  | --- | --- |
67
  | **Name** | `de_dep_news_trf` |
68
+ | **Version** | `3.7.2` |
69
+ | **spaCy** | `>=3.7.0,<3.8.0` |
70
  | **Default Pipeline** | `transformer`, `tagger`, `morphologizer`, `parser`, `lemmatizer`, `attribute_ruler` |
71
  | **Components** | `transformer`, `tagger`, `morphologizer`, `parser`, `lemmatizer`, `attribute_ruler` |
72
  | **Vectors** | 0 keys, 0 unique vectors (0 dimensions) |
96
  | `TOKEN_P` | 99.92 |
97
  | `TOKEN_R` | 99.90 |
98
  | `TOKEN_F` | 99.91 |
99
+ | `TAG_ACC` | 99.07 |
100
+ | `POS_ACC` | 99.18 |
101
  | `MORPH_ACC` | 97.01 |
102
+ | `MORPH_MICRO_P` | 98.85 |
103
+ | `MORPH_MICRO_R` | 98.87 |
104
+ | `MORPH_MICRO_F` | 98.86 |
105
+ | `SENTS_P` | 97.75 |
106
+ | `SENTS_R` | 98.04 |
107
+ | `SENTS_F` | 97.90 |
108
+ | `DEP_UAS` | 95.83 |
109
+ | `DEP_LAS` | 94.74 |
110
+ | `LEMMA_ACC` | 98.72 |
accuracy.json CHANGED
@@ -3,67 +3,67 @@
3
  "token_p": 0.9991808044,
4
  "token_r": 0.9990356034,
5
  "token_f": 0.9991081986,
6
- "tag_acc": 0.9905775076,
7
- "pos_acc": 0.9914628183,
8
- "morph_acc": 0.9700866279,
9
- "morph_micro_p": 0.9886266605,
10
- "morph_micro_r": 0.9884199706,
11
- "morph_micro_f": 0.9885233048,
12
  "morph_per_feat": {
13
  "Case": {
14
- "p": 0.9763482571,
15
- "r": 0.9779552629,
16
- "f": 0.9771510993
17
  },
18
  "Definite": {
19
- "p": 0.9996579637,
20
- "r": 1.0,
21
- "f": 0.9998289526
22
  },
23
  "Gender": {
24
- "p": 0.9848792598,
25
- "r": 0.9853240009,
26
- "f": 0.9851015801
27
  },
28
  "Number": {
29
- "p": 0.9918825162,
30
- "r": 0.9934908501,
31
- "f": 0.9926860317
32
  },
33
  "PronType": {
34
- "p": 0.9982142857,
35
  "r": 0.9971620855,
36
- "f": 0.9976879082
37
  },
38
  "Degree": {
39
- "p": 0.9956001408,
40
- "r": 0.9874323617,
41
- "f": 0.9914994304
42
  },
43
  "VerbForm": {
44
- "p": 0.9950717806,
45
- "r": 0.9973155804,
46
- "f": 0.996192417
47
  },
48
  "Mood": {
49
- "p": 0.9897197811,
50
- "r": 0.9897197811,
51
- "f": 0.9897197811
52
  },
53
  "Person": {
54
- "p": 0.9991712707,
55
- "r": 0.9991712707,
56
- "f": 0.9991712707
57
  },
58
  "Tense": {
59
- "p": 0.9973382133,
60
- "r": 0.9973382133,
61
- "f": 0.9973382133
62
  },
63
  "Reflex": {
64
- "p": 0.9976359338,
65
- "r": 1.0,
66
- "f": 0.998816568
67
  },
68
  "Poss": {
69
  "p": 1.0,
@@ -71,66 +71,66 @@
71
  "f": 1.0
72
  },
73
  "Foreign": {
74
- "p": 0.7647058824,
75
- "r": 0.1959798995,
76
- "f": 0.312
77
  }
78
  },
79
- "sents_p": 0.9779699403,
80
- "sents_r": 0.9816077702,
81
- "sents_f": 0.9797854785,
82
- "dep_uas": 0.9592566441,
83
- "dep_las": 0.9484326996,
84
  "dep_las_per_type": {
85
  "cp": {
86
- "p": 0.9908256881,
87
- "r": 0.987202925,
88
- "f": 0.989010989
89
  },
90
  "nk": {
91
- "p": 0.9891317621,
92
- "r": 0.9895748493,
93
- "f": 0.9893532561
94
  },
95
  "sb": {
96
- "p": 0.9767034507,
97
- "r": 0.9780740221,
98
- "f": 0.9773882559
99
  },
100
  "ag": {
101
- "p": 0.9764125069,
102
- "r": 0.9839690437,
103
- "f": 0.9801762115
104
  },
105
  "mnr": {
106
- "p": 0.8712352684,
107
- "r": 0.8390079865,
108
- "f": 0.8548179872
109
  },
110
  "mo": {
111
- "p": 0.86918429,
112
- "r": 0.8730528019,
113
- "f": 0.8711142511
114
  },
115
  "oc": {
116
- "p": 0.9640071986,
117
- "r": 0.9674894642,
118
- "f": 0.9657451923
119
  },
120
  "root": {
121
- "p": 0.9857936998,
122
- "r": 0.9894606324,
123
- "f": 0.9876237624
124
  },
125
  "rc": {
126
- "p": 0.8392857143,
127
- "r": 0.8430493274,
128
- "f": 0.841163311
129
  },
130
  "op": {
131
- "p": 0.8259629101,
132
- "r": 0.8628912072,
133
- "f": 0.8440233236
134
  },
135
  "pm": {
136
  "p": 1.0,
@@ -138,84 +138,89 @@
138
  "f": 1.0
139
  },
140
  "nmc": {
141
- "p": 0.9888178914,
142
- "r": 0.9778830964,
143
- "f": 0.9833200953
144
  },
145
  "oa": {
146
- "p": 0.9628411707,
147
  "r": 0.9656992084,
148
- "f": 0.9642680718
149
  },
150
  "pd": {
151
- "p": 0.9203125,
152
- "r": 0.8661764706,
153
- "f": 0.8924242424
154
  },
155
  "ep": {
156
- "p": 0.9453125,
157
- "r": 0.9097744361,
158
- "f": 0.9272030651
159
  },
160
  "cd": {
161
- "p": 0.9356499356,
162
- "r": 0.9392764858,
163
- "f": 0.9374597034
164
  },
165
  "cj": {
166
- "p": 0.9112627986,
167
  "r": 0.9215976331,
168
- "f": 0.9164010787
169
  },
170
  "par": {
171
- "p": 0.5279503106,
172
- "r": 0.4166666667,
173
- "f": 0.4657534247
174
  },
175
  "svp": {
176
- "p": 0.9766990291,
177
- "r": 0.9673076923,
178
- "f": 0.9719806763
179
  },
180
  "pg": {
181
- "p": 0.9220779221,
182
- "r": 0.9044585987,
183
- "f": 0.9131832797
184
  },
185
  "da": {
186
- "p": 0.9153439153,
187
- "r": 0.8804071247,
188
- "f": 0.897535668
189
  },
190
  "app": {
191
- "p": 0.81,
192
- "r": 0.8181818182,
193
- "f": 0.8140703518
194
  },
195
  "ac": {
196
- "p": 0.9565217391,
197
- "r": 0.9401709402,
198
- "f": 0.9482758621
 
 
 
 
 
199
  },
200
  "pnc": {
201
- "p": 0.9025710419,
202
- "r": 0.8662337662,
203
- "f": 0.8840291584
204
  },
205
  "ju": {
206
- "p": 0.93,
207
- "r": 0.9393939394,
208
- "f": 0.9346733668
209
  },
210
  "ng": {
211
- "p": 0.8839050132,
212
- "r": 0.8909574468,
213
- "f": 0.8874172185
214
  },
215
  "re": {
216
- "p": 0.8224852071,
217
- "r": 0.8273809524,
218
- "f": 0.824925816
219
  },
220
  "dep": {
221
  "p": 0.0,
@@ -223,44 +228,39 @@
223
  "f": 0.0
224
  },
225
  "cm": {
226
- "p": 0.9661835749,
227
- "r": 0.9615384615,
228
- "f": 0.9638554217
229
  },
230
  "cc": {
231
- "p": 0.7759562842,
232
- "r": 0.7135678392,
233
- "f": 0.7434554974
234
  },
235
  "sbp": {
236
- "p": 0.9734513274,
237
- "r": 0.9482758621,
238
- "f": 0.96069869
239
  },
240
  "uc": {
241
- "p": 0.7619047619,
242
- "r": 0.5333333333,
243
- "f": 0.6274509804
244
  },
245
  "cvc": {
246
- "p": 0.8157894737,
247
  "r": 0.7848101266,
248
- "f": 0.8
249
  },
250
  "ams": {
251
- "p": 0.75,
252
- "r": 0.8823529412,
253
- "f": 0.8108108108
254
- },
255
- "rs": {
256
- "p": 0.7777777778,
257
- "r": 0.4516129032,
258
- "f": 0.5714285714
259
  },
260
  "ph": {
261
- "p": 0.8333333333,
262
- "r": 0.7142857143,
263
- "f": 0.7692307692
264
  },
265
  "dm": {
266
  "p": 1.0,
@@ -278,21 +278,16 @@
278
  "f": 0.0
279
  },
280
  "og": {
281
- "p": 0.4166666667,
282
  "r": 0.7142857143,
283
- "f": 0.5263157895
284
- },
285
- "avc": {
286
- "p": 0.0,
287
- "r": 0.0,
288
- "f": 0.0
289
  },
290
  "adc": {
291
- "p": 0.75,
292
  "r": 1.0,
293
- "f": 0.8571428571
294
  }
295
  },
296
- "lemma_acc": 0.9872611465,
297
- "speed": 2933.9454548672
298
  }
3
  "token_p": 0.9991808044,
4
  "token_r": 0.9990356034,
5
  "token_f": 0.9991081986,
6
+ "tag_acc": 0.9907095376,
7
+ "pos_acc": 0.9917734358,
8
+ "morph_acc": 0.9700601968,
9
+ "morph_micro_p": 0.9885202304,
10
+ "morph_micro_r": 0.9886522681,
11
+ "morph_micro_f": 0.9885862448,
12
  "morph_per_feat": {
13
  "Case": {
14
+ "p": 0.9765625,
15
+ "r": 0.9781171582,
16
+ "f": 0.9773392108
17
  },
18
  "Definite": {
19
+ "p": 0.9997719238,
20
+ "r": 0.9998859489,
21
+ "f": 0.9998289331
22
  },
23
  "Gender": {
24
+ "p": 0.9843639949,
25
+ "r": 0.986114247,
26
+ "f": 0.9852383436
27
  },
28
  "Number": {
29
+ "p": 0.9916774552,
30
+ "r": 0.9936529998,
31
+ "f": 0.9926642446
32
  },
33
  "PronType": {
34
+ "p": 0.9983763598,
35
  "r": 0.9971620855,
36
+ "f": 0.9977688532
37
  },
38
  "Degree": {
39
+ "p": 0.9952523299,
40
+ "r": 0.9879560133,
41
+ "f": 0.9915907498
42
  },
43
  "VerbForm": {
44
+ "p": 0.9949646454,
45
+ "r": 0.9972082036,
46
+ "f": 0.9960851612
47
  },
48
  "Mood": {
49
+ "p": 0.9903862092,
50
+ "r": 0.990714641,
51
+ "f": 0.9905503979
52
  },
53
  "Person": {
54
+ "p": 0.998619166,
55
+ "r": 0.9988950276,
56
+ "f": 0.9987570778
57
  },
58
  "Tense": {
59
+ "p": 0.9966744263,
60
+ "r": 0.9971718516,
61
+ "f": 0.9969230769
62
  },
63
  "Reflex": {
64
+ "p": 0.9976303318,
65
+ "r": 0.9976303318,
66
+ "f": 0.9976303318
67
  },
68
  "Poss": {
69
  "p": 1.0,
71
  "f": 1.0
72
  },
73
  "Foreign": {
74
+ "p": 0.8333333333,
75
+ "r": 0.175879397,
76
+ "f": 0.2904564315
77
  }
78
  },
79
+ "sents_p": 0.9775396662,
80
+ "sents_r": 0.9803678446,
81
+ "sents_f": 0.9789517128,
82
+ "dep_uas": 0.958309546,
83
+ "dep_las": 0.9473953052,
84
  "dep_las_per_type": {
85
  "cp": {
86
+ "p": 0.9853479853,
87
+ "r": 0.9835466179,
88
+ "f": 0.9844464776
89
  },
90
  "nk": {
91
+ "p": 0.9890901689,
92
+ "r": 0.9894526796,
93
+ "f": 0.9892713911
94
  },
95
  "sb": {
96
+ "p": 0.9754299754,
97
+ "r": 0.9749166813,
98
+ "f": 0.9751732608
99
  },
100
  "ag": {
101
+ "p": 0.9763346175,
102
+ "r": 0.9806522941,
103
+ "f": 0.9784886928
104
  },
105
  "mnr": {
106
+ "p": 0.8643410853,
107
+ "r": 0.8436317781,
108
+ "f": 0.8538608807
109
  },
110
  "mo": {
111
+ "p": 0.8667875857,
112
+ "r": 0.8701193607,
113
+ "f": 0.8684502776
114
  },
115
  "oc": {
116
+ "p": 0.9654965497,
117
+ "r": 0.9686935581,
118
+ "f": 0.9670924117
119
  },
120
  "root": {
121
+ "p": 0.9857790602,
122
+ "r": 0.988427361,
123
+ "f": 0.9871014343
124
  },
125
  "rc": {
126
+ "p": 0.8258928571,
127
+ "r": 0.8295964126,
128
+ "f": 0.8277404922
129
  },
130
  "op": {
131
+ "p": 0.8216833096,
132
+ "r": 0.8584202683,
133
+ "f": 0.8396501458
134
  },
135
  "pm": {
136
  "p": 1.0,
138
  "f": 1.0
139
  },
140
  "nmc": {
141
+ "p": 0.9857369255,
142
+ "r": 0.9826224329,
143
+ "f": 0.9841772152
144
  },
145
  "oa": {
146
+ "p": 0.9581151832,
147
  "r": 0.9656992084,
148
+ "f": 0.961892247
149
  },
150
  "pd": {
151
+ "p": 0.9106317411,
152
+ "r": 0.8691176471,
153
+ "f": 0.8893905192
154
  },
155
  "ep": {
156
+ "p": 0.9323308271,
157
+ "r": 0.9323308271,
158
+ "f": 0.9323308271
159
  },
160
  "cd": {
161
+ "p": 0.9385908209,
162
+ "r": 0.9379844961,
163
+ "f": 0.9382875606
164
  },
165
  "cj": {
166
+ "p": 0.9175257732,
167
  "r": 0.9215976331,
168
+ "f": 0.9195571956
169
  },
170
  "par": {
171
+ "p": 0.5632911392,
172
+ "r": 0.4362745098,
173
+ "f": 0.4917127072
174
  },
175
  "svp": {
176
+ "p": 0.9763779528,
177
+ "r": 0.9538461538,
178
+ "f": 0.9649805447
179
  },
180
  "pg": {
181
+ "p": 0.9185667752,
182
+ "r": 0.898089172,
183
+ "f": 0.9082125604
184
  },
185
  "da": {
186
+ "p": 0.9189189189,
187
+ "r": 0.8651399491,
188
+ "f": 0.8912188729
189
  },
190
  "app": {
191
+ "p": 0.8361204013,
192
+ "r": 0.8417508418,
193
+ "f": 0.8389261745
194
  },
195
  "ac": {
196
+ "p": 0.9732142857,
197
+ "r": 0.9316239316,
198
+ "f": 0.9519650655
199
+ },
200
+ "rs": {
201
+ "p": 0.6842105263,
202
+ "r": 0.4193548387,
203
+ "f": 0.52
204
  },
205
  "pnc": {
206
+ "p": 0.8959459459,
207
+ "r": 0.861038961,
208
+ "f": 0.8781456954
209
  },
210
  "ju": {
211
+ "p": 0.9215686275,
212
+ "r": 0.9494949495,
213
+ "f": 0.9353233831
214
  },
215
  "ng": {
216
+ "p": 0.8759894459,
217
+ "r": 0.8829787234,
218
+ "f": 0.8794701987
219
  },
220
  "re": {
221
+ "p": 0.8303030303,
222
+ "r": 0.8154761905,
223
+ "f": 0.8228228228
224
  },
225
  "dep": {
226
  "p": 0.0,
228
  "f": 0.0
229
  },
230
  "cm": {
231
+ "p": 0.9428571429,
232
+ "r": 0.9519230769,
233
+ "f": 0.9473684211
234
  },
235
  "cc": {
236
+ "p": 0.7700534759,
237
+ "r": 0.7236180905,
238
+ "f": 0.7461139896
239
  },
240
  "sbp": {
241
+ "p": 0.9380530973,
242
+ "r": 0.9137931034,
243
+ "f": 0.9257641921
244
  },
245
  "uc": {
246
+ "p": 0.8666666667,
247
+ "r": 0.4333333333,
248
+ "f": 0.5777777778
249
  },
250
  "cvc": {
251
+ "p": 0.8378378378,
252
  "r": 0.7848101266,
253
+ "f": 0.8104575163
254
  },
255
  "ams": {
256
+ "p": 0.7625,
257
+ "r": 0.8970588235,
258
+ "f": 0.8243243243
 
 
 
 
 
259
  },
260
  "ph": {
261
+ "p": 0.8571428571,
262
+ "r": 0.8571428571,
263
+ "f": 0.8571428571
264
  },
265
  "dm": {
266
  "p": 1.0,
278
  "f": 0.0
279
  },
280
  "og": {
281
+ "p": 0.4545454545,
282
  "r": 0.7142857143,
283
+ "f": 0.5555555556
 
 
 
 
 
284
  },
285
  "adc": {
286
+ "p": 0.6,
287
  "r": 1.0,
288
+ "f": 0.75
289
  }
290
  },
291
+ "lemma_acc": 0.9872479319,
292
+ "speed": 2451.8342899914
293
  }
config.cfg CHANGED
@@ -17,6 +17,7 @@ after_creation = null
17
  after_pipeline_creation = null
18
  batch_size = 64
19
  tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
 
20
 
21
  [components]
22
 
@@ -39,10 +40,11 @@ nO = null
39
  normalize = false
40
 
41
  [components.lemmatizer.model.tok2vec]
42
- @architectures = "spacy-transformers.TransformerListener.v1"
43
- grad_factor = 1.0
44
  upstream = "transformer"
45
  pooling = {"@layers":"reduce_mean.v1"}
 
46
 
47
  [components.morphologizer]
48
  factory = "morphologizer"
@@ -57,10 +59,11 @@ nO = null
57
  normalize = false
58
 
59
  [components.morphologizer.model.tok2vec]
60
- @architectures = "spacy-transformers.TransformerListener.v1"
61
- grad_factor = 1.0
62
  upstream = "transformer"
63
  pooling = {"@layers":"reduce_mean.v1"}
 
64
 
65
  [components.parser]
66
  factory = "parser"
@@ -80,10 +83,11 @@ use_upper = false
80
  nO = null
81
 
82
  [components.parser.model.tok2vec]
83
- @architectures = "spacy-transformers.TransformerListener.v1"
84
- grad_factor = 1.0
85
  upstream = "transformer"
86
  pooling = {"@layers":"reduce_mean.v1"}
 
87
 
88
  [components.tagger]
89
  factory = "tagger"
@@ -98,32 +102,44 @@ nO = null
98
  normalize = false
99
 
100
  [components.tagger.model.tok2vec]
101
- @architectures = "spacy-transformers.TransformerListener.v1"
102
- grad_factor = 1.0
103
  upstream = "transformer"
104
  pooling = {"@layers":"reduce_mean.v1"}
 
105
 
106
  [components.transformer]
107
- factory = "transformer"
108
- max_batch_items = 4096
109
- set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
110
 
111
  [components.transformer.model]
112
- name = "bert-base-german-cased"
113
- @architectures = "spacy-transformers.TransformerModel.v3"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
114
  mixed_precision = false
115
-
116
- [components.transformer.model.get_spans]
117
- @span_getters = "spacy-transformers.strided_spans.v1"
118
- window = 128
119
- stride = 96
120
 
121
  [components.transformer.model.grad_scaler_config]
122
 
123
- [components.transformer.model.tokenizer_config]
124
- use_fast = true
125
-
126
- [components.transformer.model.transformer_config]
 
127
 
128
  [corpora]
129
 
@@ -160,11 +176,11 @@ annotating_components = []
160
  before_update = null
161
 
162
  [training.batcher]
163
- @batchers = "spacy.batch_by_padded.v1"
164
- discard_oversize = true
165
- get_length = null
166
  size = 2000
167
- buffer = 256
 
168
 
169
  [training.logger]
170
  @loggers = "spacy.ConsoleLogger.v1"
@@ -242,6 +258,18 @@ require = false
242
  path = "corpus/labels/tagger.json"
243
  require = false
244
 
 
 
 
 
 
 
 
 
 
 
 
 
245
  [initialize.lookups]
246
  @misc = "spacy.LookupsDataLoader.v1"
247
  lang = ${nlp.lang}
17
  after_pipeline_creation = null
18
  batch_size = 64
19
  tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+ vectors = {"@vectors":"spacy.Vectors.v1"}
21
 
22
  [components]
23
 
40
  normalize = false
41
 
42
  [components.lemmatizer.model.tok2vec]
43
+ @architectures = "spacy-curated-transformers.LastTransformerLayerListener.v1"
44
+ width = ${components.transformer.model.hidden_width}
45
  upstream = "transformer"
46
  pooling = {"@layers":"reduce_mean.v1"}
47
+ grad_factor = 1.0
48
 
49
  [components.morphologizer]
50
  factory = "morphologizer"
59
  normalize = false
60
 
61
  [components.morphologizer.model.tok2vec]
62
+ @architectures = "spacy-curated-transformers.LastTransformerLayerListener.v1"
63
+ width = ${components.transformer.model.hidden_width}
64
  upstream = "transformer"
65
  pooling = {"@layers":"reduce_mean.v1"}
66
+ grad_factor = 1.0
67
 
68
  [components.parser]
69
  factory = "parser"
83
  nO = null
84
 
85
  [components.parser.model.tok2vec]
86
+ @architectures = "spacy-curated-transformers.LastTransformerLayerListener.v1"
87
+ width = ${components.transformer.model.hidden_width}
88
  upstream = "transformer"
89
  pooling = {"@layers":"reduce_mean.v1"}
90
+ grad_factor = 1.0
91
 
92
  [components.tagger]
93
  factory = "tagger"
102
  normalize = false
103
 
104
  [components.tagger.model.tok2vec]
105
+ @architectures = "spacy-curated-transformers.LastTransformerLayerListener.v1"
106
+ width = ${components.transformer.model.hidden_width}
107
  upstream = "transformer"
108
  pooling = {"@layers":"reduce_mean.v1"}
109
+ grad_factor = 1.0
110
 
111
  [components.transformer]
112
+ factory = "curated_transformer"
113
+ all_layer_outputs = false
114
+ frozen = false
115
 
116
  [components.transformer.model]
117
+ @architectures = "spacy-curated-transformers.BertTransformer.v1"
118
+ vocab_size = 30000
119
+ hidden_width = 768
120
+ piece_encoder = {"@architectures":"spacy-curated-transformers.BertWordpieceEncoder.v1"}
121
+ attention_probs_dropout_prob = 0.1
122
+ hidden_act = "gelu"
123
+ hidden_dropout_prob = 0.1
124
+ intermediate_width = 3072
125
+ layer_norm_eps = 0.0
126
+ max_position_embeddings = 512
127
+ model_max_length = 512
128
+ num_attention_heads = 12
129
+ num_hidden_layers = 12
130
+ padding_idx = 0
131
+ type_vocab_size = 2
132
+ torchscript = false
133
  mixed_precision = false
134
+ wrapped_listener = null
 
 
 
 
135
 
136
  [components.transformer.model.grad_scaler_config]
137
 
138
+ [components.transformer.model.with_spans]
139
+ @architectures = "spacy-curated-transformers.WithStridedSpans.v1"
140
+ stride = 136
141
+ window = 176
142
+ batch_size = 384
143
 
144
  [corpora]
145
 
176
  before_update = null
177
 
178
  [training.batcher]
179
+ @batchers = "spacy.batch_by_words.v1"
180
+ discard_oversize = false
 
181
  size = 2000
182
+ tolerance = 0.2
183
+ get_length = null
184
 
185
  [training.logger]
186
  @loggers = "spacy.ConsoleLogger.v1"
258
  path = "corpus/labels/tagger.json"
259
  require = false
260
 
261
+ [initialize.components.transformer]
262
+
263
+ [initialize.components.transformer.encoder_loader]
264
+ @model_loaders = "spacy-curated-transformers.HFTransformerEncoderLoader.v1"
265
+ name = "bert-base-german-cased"
266
+ revision = "main"
267
+
268
+ [initialize.components.transformer.piecer_loader]
269
+ @model_loaders = "spacy-curated-transformers.HFPieceEncoderLoader.v1"
270
+ name = "bert-base-german-cased"
271
+ revision = "main"
272
+
273
  [initialize.lookups]
274
  @misc = "spacy.LookupsDataLoader.v1"
275
  lang = ${nlp.lang}
de_dep_news_trf-any-py3-none-any.whl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7a018eb23ae715bbb44ae42f533d31a8f5083c13762899ba2d4b7ef424ad5d0
3
- size 412822128
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fb898b58c3647aa157e8b3ceb7acb43499f27e6d2f8a23f4c130f9c3de32cd4
3
+ size 410451513
lemmatizer/cfg CHANGED
@@ -56,203 +56,204 @@
56
  113,
57
  115,
58
  117,
59
- 118,
60
  119,
61
  121,
62
- 124,
 
63
  126,
64
- 127,
65
  128,
66
  130,
 
67
  133,
 
68
  135,
69
- 138,
 
70
  140,
71
- 142,
72
- 144,
73
  145,
74
- 146,
75
- 147,
76
- 149,
77
- 151,
78
- 153,
79
  155,
80
- 157,
81
- 159,
 
82
  161,
83
  163,
84
- 165,
85
  166,
 
86
  170,
87
  171,
88
  173,
 
89
  175,
 
90
  178,
91
  180,
92
  182,
93
  184,
94
  186,
95
- 187,
96
- 188,
97
  190,
 
98
  192,
99
- 193,
100
- 195,
101
  197,
102
- 199,
103
  200,
104
- 201,
105
- 203,
 
106
  207,
107
  208,
108
- 210,
109
- 212,
110
  215,
111
  216,
112
- 217,
113
  218,
114
  220,
115
- 222,
116
  224,
 
117
  227,
118
- 230,
119
- 234,
120
- 237,
121
  239,
122
- 240,
123
  242,
124
  244,
125
  245,
126
- 247,
127
- 249,
128
  250,
 
129
  252,
130
- 254,
131
- 255,
132
  257,
133
  260,
134
- 263,
135
  264,
 
136
  266,
137
- 267,
138
- 269,
139
  271,
140
  272,
141
  273,
142
  275,
143
- 276,
144
  278,
 
145
  281,
146
  282,
147
  283,
148
- 285,
149
- 287,
150
  288,
 
151
  290,
152
  291,
153
- 292,
154
  293,
 
 
155
  296,
156
  297,
157
- 298,
158
  299,
159
- 301,
160
- 302,
161
  303,
162
  304,
163
- 305,
164
  307,
165
  308,
166
- 311,
 
 
167
  314,
168
- 316,
169
  317,
 
170
  319,
171
  320,
172
  321,
173
- 322,
174
  323,
175
- 324,
176
  326,
 
177
  329,
178
- 334,
179
- 335,
 
180
  339,
181
- 341,
182
- 342,
183
- 344,
184
  345,
185
- 347,
 
 
 
186
  352,
187
  354,
188
- 355,
189
- 357,
190
  359,
 
191
  362,
192
  363,
193
- 364,
194
  365,
195
- 366,
196
  367,
197
  370,
198
- 372,
199
- 375,
200
  377,
201
- 378,
202
  379,
203
- 380,
204
- 381,
205
  383,
206
  384,
207
  385,
208
- 386,
209
  388,
210
- 389,
211
  391,
 
212
  393,
213
- 394,
214
  395,
215
  396,
216
- 397,
217
  400,
 
218
  402,
 
219
  404,
220
- 405,
221
  409,
222
- 410,
223
  411,
224
  412,
225
- 415,
 
226
  418,
227
- 420,
228
- 421,
229
  422,
230
- 423,
231
- 424,
232
  427,
 
 
233
  430,
234
  431,
235
- 432,
236
- 433,
237
- 435,
238
  437,
 
239
  439,
240
  440,
241
- 441,
242
  442,
243
  443,
244
  444,
245
  445,
246
- 448,
247
  449,
248
- 451,
249
  452,
250
  453,
251
- 456,
252
  457,
253
  458,
254
- 463,
255
- 465,
256
  466,
257
  468,
258
  471,
@@ -262,25 +263,24 @@
262
  477,
263
  478,
264
  479,
265
- 481,
266
  482,
267
  483,
268
  484,
269
- 485,
270
  486,
271
- 487,
272
- 488,
273
  489,
274
  490,
275
- 492,
 
276
  495,
277
- 496,
278
- 499,
279
- 500,
280
  501,
 
281
  504,
 
 
282
  507,
283
- 509,
284
  510,
285
  511,
286
  512,
@@ -288,89 +288,88 @@
288
  514,
289
  515,
290
  516,
291
- 517,
292
  519,
293
- 520,
294
  521,
295
- 524,
 
296
  526,
297
  527,
 
298
  530,
299
- 532,
300
  533,
301
- 535,
302
- 537,
 
303
  543,
 
304
  546,
305
- 547,
306
- 549,
307
- 552,
308
- 555,
309
- 559,
310
  560,
311
- 561,
312
- 562,
313
  564,
314
- 565,
315
- 566,
316
- 569,
317
- 570,
318
  571,
319
- 572,
320
- 573,
321
  574,
322
- 575,
323
- 578,
324
- 579,
325
  580,
326
- 581,
327
- 584,
328
  587,
329
- 589,
330
  590,
331
  592,
 
332
  594,
333
  595,
334
  596,
335
  597,
336
- 598,
337
  599,
338
  600,
 
339
  602,
340
- 603,
341
  607,
342
  608,
343
- 609,
344
- 612,
345
- 614,
346
  616,
347
  617,
348
- 619,
349
  620,
350
  621,
351
- 623,
352
  624,
353
  625,
354
  626,
355
  627,
356
- 629,
357
  630,
358
  631,
 
359
  633,
360
  634,
361
- 637,
 
362
  638,
 
363
  640,
 
364
  644,
365
  645,
 
366
  647,
367
- 648,
368
- 650,
369
  652,
370
  654,
 
371
  656,
372
  657,
373
- 658,
374
  659,
375
  660,
376
  661,
@@ -380,49 +379,49 @@
380
  667,
381
  671,
382
  672,
383
- 677,
384
  678,
385
  679,
386
- 682,
387
- 684,
388
  685,
389
- 687,
390
- 689,
391
- 691,
392
  692,
393
- 694,
394
  695,
395
  696,
396
  697,
397
  698,
398
- 700,
399
  701,
400
  702,
401
- 703,
402
  704,
 
403
  706,
404
- 707,
405
- 708,
406
  709,
407
  710,
408
  711,
409
- 712,
410
  714,
411
- 534,
412
- 719,
413
- 720,
414
  721,
415
  722,
416
- 726,
417
- 727,
418
- 728,
 
419
  731,
420
  732,
421
  733,
422
  734,
423
  735,
 
424
  737,
425
- 738,
426
  741,
427
  742,
428
  743,
@@ -430,127 +429,121 @@
430
  745,
431
  746,
432
  747,
433
- 748,
434
  749,
435
- 750,
436
  753,
437
  754,
438
- 756,
439
- 757,
440
  758,
441
- 759,
442
  761,
443
- 764,
444
- 766,
 
445
  768,
446
- 769,
447
  770,
448
  772,
 
 
449
  775,
450
  777,
451
- 778,
452
- 779,
453
  780,
454
- 781,
455
  782,
 
456
  784,
457
- 787,
458
- 789,
459
- 792,
460
- 795,
 
461
  796,
462
- 798,
463
  799,
464
  800,
 
465
  804,
466
- 805,
467
- 807,
468
  808,
469
- 810,
 
470
  812,
 
471
  816,
472
  820,
473
- 823,
474
- 826,
475
  827,
 
476
  832,
477
- 833,
478
- 836,
479
  837,
480
- 839,
481
  841,
482
  842,
483
- 843,
484
  844,
485
- 845,
486
  846,
487
- 851,
488
  852,
489
  853,
490
  854,
491
  858,
492
  859,
493
  860,
494
- 861,
495
  863,
496
- 864,
497
- 867,
498
  870,
499
  871,
500
- 872,
501
  873,
502
- 876,
503
- 878,
504
  879,
505
  880,
506
  881,
 
507
  883,
508
- 884,
509
  885,
510
- 889,
511
- 890,
512
  891,
513
  892,
514
  893,
515
  894,
 
 
 
 
516
  899,
517
- 901,
518
- 902,
519
- 903,
520
  904,
521
- 905,
522
  906,
523
- 909,
524
  910,
525
  912,
526
- 913,
527
  916,
528
  917,
529
- 918,
530
- 920,
531
  922,
532
  923,
533
  924,
534
- 925,
535
  927,
536
  928,
537
- 929,
538
  930,
539
- 931,
540
- 933,
541
  935,
 
542
  937,
543
- 938,
544
- 939,
545
  942,
546
  943,
 
547
  945,
548
  946,
549
- 774,
550
  947,
551
  948,
552
  951,
553
- 952,
554
  953,
555
  954,
556
  955,
@@ -559,224 +552,226 @@
559
  958,
560
  959,
561
  960,
 
562
  962,
563
- 964,
564
  965,
565
  966,
566
  967,
567
- 968,
568
- 969,
569
  972,
570
  973,
571
  974,
572
- 975,
573
  978,
574
  979,
575
- 980,
576
- 982,
577
  984,
578
  985,
579
- 986,
580
  987,
581
  989,
582
- 991,
583
  992,
 
584
  994,
585
- 995,
586
- 997,
587
  998,
 
588
  1000,
589
  1002,
590
  1004,
591
- 1005,
592
  1006,
593
  1007,
594
  1008,
 
595
  1010,
596
- 1011,
597
  1012,
598
  1013,
599
  1014,
600
- 1016,
601
  1017,
602
  1018,
603
  1019,
604
  1020,
605
- 1023,
606
- 1024,
 
607
  1026,
608
- 1027,
609
- 1030,
610
- 1031,
611
  1032,
612
- 1034,
613
- 1036,
614
  1037,
615
  1038,
616
  1039,
617
  1040,
618
  1041,
619
- 1042,
620
  1043,
621
- 1045,
622
- 1048,
623
- 1051,
624
  1052,
 
625
  1054,
626
  1055,
627
  1056,
628
- 1057,
629
- 1060,
630
- 1062,
631
  1064,
632
- 1065,
633
- 1068,
634
  1070,
635
- 1071,
636
  1073,
637
- 1074,
638
  1075,
639
  1076,
640
  1078,
641
  1079,
642
  1080,
643
- 1082,
644
  1083,
645
  1084,
 
646
  1086,
647
  1087,
648
  1088,
649
  1089,
650
  1090,
651
- 1092,
652
  1093,
653
  1094,
654
- 1095,
655
  1096,
656
  1097,
657
- 1098,
658
  1099,
659
  1100,
 
 
660
  1103,
661
  1104,
 
662
  1106,
663
  1107,
664
- 1108,
665
  1109,
666
  1110,
667
  1111,
 
 
668
  1114,
669
- 1116,
670
  1118,
 
 
671
  1122,
672
  1123,
673
- 1124,
674
- 1125,
675
  1126,
676
- 1129,
677
- 177,
678
  1130,
679
- 1131,
680
- 1132,
681
  1134,
 
682
  1136,
 
683
  1138,
 
684
  1140,
685
- 1141,
686
  1142,
687
- 1144,
688
  1145,
 
 
 
689
  1148,
690
  1149,
691
  1151,
692
- 1152,
693
  1154,
694
- 1155,
695
  1156,
696
- 50,
697
- 1157,
698
  1158,
699
  1159,
700
  1160,
701
  1161,
702
  1162,
 
703
  1165,
704
- 1166,
705
- 1167,
706
  1168,
 
707
  1170,
708
  1171,
709
- 1172,
710
  1174,
711
- 1175,
712
  1176,
713
  1177,
 
 
714
  1179,
715
  1180,
716
- 1183,
717
- 1184,
718
  1185,
719
  1186,
 
720
  1188,
721
- 1189,
722
  1190,
 
723
  1192,
724
  1194,
725
  1195,
 
 
726
  1199,
 
727
  1201,
728
- 1202,
729
- 1204,
730
- 1205,
731
  1206,
 
732
  1208,
733
- 1211,
 
 
 
 
734
  1216,
735
- 1217,
736
  1218,
737
  1219,
738
- 1220,
739
- 1221,
740
  1223,
741
- 1224,
742
- 1226,
743
- 1229,
744
  1230,
745
- 1231,
746
  1233,
747
- 1234,
748
- 900,
749
- 1235,
750
- 1236,
751
  1241,
752
- 1242,
753
  1243,
754
  1244,
755
- 1245,
756
- 1248,
757
  1249,
758
  1250,
759
  1251,
760
- 1252,
761
  1254,
 
762
  1255,
 
763
  1256,
764
  1257,
765
- 1259,
766
- 1260,
767
  1262,
 
768
  1264,
769
- 1266,
770
  1267,
771
  1268,
772
  1269,
773
- 1270,
774
  1271,
775
  1272,
776
- 1273,
777
  1274,
778
  1276,
 
779
  1278,
 
780
  1280,
781
  1281,
782
  1283,
@@ -784,107 +779,107 @@
784
  1285,
785
  1286,
786
  1288,
 
787
  1291,
788
- 1293,
789
  1294,
790
- 1295,
791
  1297,
792
  1298,
 
793
  1300,
794
- 1301,
795
- 1303,
796
- 1307,
797
- 1309,
798
- 1077,
799
  1310,
800
- 1311,
801
  1312,
802
- 1314,
803
- 1315,
804
  1316,
805
- 1317,
806
  1318,
807
- 1319,
808
- 462,
809
  1320,
 
810
  1322,
811
  1323,
812
  1324,
813
- 1326,
814
  1327,
815
  1328,
 
 
816
  1330,
817
  1331,
818
- 1334,
 
 
819
  1336,
820
- 1338,
821
  1340,
 
822
  1342,
 
823
  1345,
 
824
  1347,
825
  1348,
826
  1349,
827
  1352,
828
- 1353,
829
  1354,
830
- 1355,
831
  1356,
832
- 1358,
833
  1359,
834
  1360,
835
- 1361,
836
  1362,
837
  1363,
838
- 1364,
839
  1365,
840
- 1366,
841
  1367,
 
842
  1371,
843
  1372,
 
844
  1374,
845
  1375,
846
  1376,
847
- 1377,
848
- 1378,
849
  1379,
 
 
850
  1382,
 
851
  1384,
852
  1385,
 
853
  1387,
854
  1388,
 
855
  1390,
 
856
  1392,
857
  1393,
858
- 1394,
859
  1395,
860
- 1397,
861
- 1398,
862
  1401,
863
- 1402,
864
  1403,
865
  1404,
866
- 1240,
867
  1406,
868
- 1408,
869
- 1410,
870
  1411,
871
- 1412,
872
- 1415,
873
  1416,
874
- 1417,
875
  1418,
876
  1419,
877
  1420,
878
- 1421,
879
  1423,
880
  1424,
 
881
  1425,
882
- 274,
883
  1426,
884
  1428,
 
885
  1430,
886
- 1431,
887
- 1432,
888
  1433,
889
  1434,
890
  1435,
@@ -892,14 +887,15 @@
892
  1437,
893
  1438,
894
  1439,
895
- 1440,
896
  1441,
 
897
  1442,
898
  1444,
899
  1446,
900
  1447,
901
  1448,
902
  1449,
 
903
  1451,
904
  1452,
905
  1453,
@@ -907,19 +903,18 @@
907
  1455,
908
  1456,
909
  1457,
910
- 1458,
911
  1459,
912
- 1460,
913
  1461,
914
  1462,
915
- 1464,
916
  1465,
917
  1466,
918
  1467,
919
  1468,
920
- 1469,
921
  1470,
 
922
  1472,
 
923
  1474,
924
  1475,
925
  1476,
@@ -927,336 +922,339 @@
927
  1478,
928
  1479,
929
  1480,
930
- 797,
931
  1481,
932
  1482,
933
- 1483,
934
  1484,
935
- 1485,
936
  1486,
 
937
  1488,
938
  1489,
939
  1490,
 
940
  1492,
941
  1493,
 
 
 
942
  1496,
943
  1497,
944
  1498,
945
  1499,
946
  1500,
947
  1502,
948
- 1504,
949
  1505,
950
  1506,
951
- 1507,
952
  1508,
953
  1509,
954
- 1137,
955
  1510,
956
  1511,
957
- 1512,
958
  1514,
959
- 1515,
960
  1516,
961
  1517,
962
  1518,
963
  1519,
 
 
 
964
  1522,
965
  1523,
 
966
  1526,
 
 
 
967
  1530,
968
  1531,
969
  1532,
970
- 1534,
971
- 1535,
972
  1536,
973
  1537,
974
  1538,
975
- 1539,
976
  1541,
977
- 1543,
978
- 1544,
979
  1546,
980
  1547,
981
  1548,
982
  1549,
983
  1550,
 
984
  1552,
985
  1553,
986
  1554,
987
  1556,
988
  1557,
989
- 1558,
990
  1559,
991
- 1562,
 
992
  1563,
993
  1564,
 
994
  1566,
995
  1567,
996
- 1568,
997
  1570,
998
  1572,
999
  1573,
 
1000
  1575,
1001
- 1577,
1002
  1579,
 
1003
  1582,
1004
- 1585,
 
1005
  1586,
1006
- 1587,
1007
- 1590,
1008
  1591,
1009
- 1592,
1010
  1593,
1011
  1596,
1012
- 1598,
1013
  1599,
1014
  1600,
1015
  1601,
1016
- 1602,
1017
  1603,
1018
  1604,
1019
- 1606,
1020
  1607,
1021
  1608,
1022
- 1610,
1023
  1611,
1024
- 1612,
 
1025
  1616,
1026
  1617,
1027
  1618,
1028
  1619,
 
1029
  1621,
1030
- 1622,
1031
  1623,
 
1032
  1625,
1033
- 1626,
1034
  1627,
 
1035
  1629,
1036
- 1630,
1037
- 1631,
1038
  1633,
 
1039
  1635,
 
1040
  1637,
 
1041
  1640,
1042
  1641,
1043
  1642,
1044
- 1494,
1045
  1643,
1046
  1644,
1047
  1646,
1048
- 1647,
1049
  1648,
1050
- 1649,
1051
  1650,
1052
  1651,
1053
- 1652,
1054
- 1653,
1055
  1654,
 
1056
  1656,
1057
  1658,
 
1058
  1660,
1059
  1661,
1060
- 1663,
1061
  1664,
1062
  1665,
1063
  1666,
 
1064
  1668,
 
1065
  1670,
1066
- 1671,
1067
- 1673,
1068
  1674,
1069
- 1675,
1070
  1676,
1071
- 1677,
1072
  1678,
1073
- 1662,
1074
  1679,
1075
  1680,
1076
  1681,
1077
  1682,
1078
- 1683,
1079
  1684,
1080
- 1685,
1081
- 1688,
1082
  1689,
1083
  1690,
 
1084
  1692,
 
 
1085
  1694,
 
1086
  1696,
1087
  1697,
1088
  1698,
1089
  1699,
1090
- 1700,
1091
- 1701,
1092
  1702,
1093
  1703,
1094
- 1705,
1095
  1706,
1096
- 1707,
1097
  1710,
1098
  1711,
1099
  1712,
1100
  1713,
 
1101
  1715,
1102
- 1717,
 
1103
  1719,
1104
  1720,
1105
  1721,
1106
  1722,
1107
  1723,
1108
- 1724,
1109
- 1727,
1110
- 1728,
1111
  1729,
1112
- 1732,
1113
- 1734,
 
 
1114
  1736,
 
1115
  1738,
1116
  1739,
 
 
1117
  1742,
1118
  1743,
1119
- 1745,
1120
  1746,
1121
  1747,
1122
- 1748,
1123
- 1749,
1124
  1750,
1125
- 1751,
1126
  1752,
1127
  1754,
1128
- 1755,
1129
  1756,
1130
  1757,
1131
- 1758,
1132
- 1759,
1133
  1761,
1134
- 1762,
1135
  1763,
 
1136
  1765,
1137
  1766,
 
 
1138
  1769,
 
 
 
 
1139
  1775,
 
 
1140
  1778,
 
1141
  1781,
 
1142
  1783,
1143
  1784,
 
1144
  1786,
1145
- 1787,
1146
- 1788,
1147
- 1789,
1148
- 1791,
1149
  1792,
1150
- 1793,
1151
  1795,
1152
- 1796,
1153
- 1797,
1154
  1798,
1155
- 1799,
1156
  1800,
1157
  1801,
1158
- 1802,
1159
  1804,
1160
- 1805,
1161
  1806,
1162
  1807,
1163
  1808,
1164
- 1809,
1165
  1810,
1166
  1811,
1167
- 1812,
1168
  1813,
1169
  1814,
 
1170
  1816,
1171
  1817,
1172
  1818,
1173
  1819,
1174
- 1820,
1175
  1822,
1176
  1823,
1177
  1824,
1178
  1825,
 
1179
  1827,
1180
  1828,
1181
- 1829,
1182
  1830,
1183
- 932,
1184
  1832,
 
1185
  1834,
1186
  1835,
 
1187
  1837,
1188
- 1838,
1189
  1839,
1190
  1840,
1191
- 1843,
 
 
1192
  1845,
1193
  1846,
1194
- 1847,
1195
  1848,
1196
  1849,
 
1197
  1850,
1198
- 1851,
1199
  1853,
1200
- 1854,
1201
- 1856,
1202
- 1857,
1203
  1860,
 
1204
  1862,
1205
  1863,
1206
  1864,
 
1207
  1867,
1208
  1868,
1209
  1869,
1210
- 1870,
1211
  1871,
1212
- 1873,
1213
  1874,
1214
  1875,
1215
- 1876,
1216
- 1877,
1217
  1878,
1218
- 1879,
1219
  1880,
1220
  1881,
 
1221
  1883,
1222
  1884,
1223
- 1885,
1224
  1887,
1225
  1888,
 
1226
  1890,
1227
  1891,
1228
- 1892,
1229
  1893,
1230
  1894,
 
1231
  1896,
 
1232
  1898,
1233
  1899,
1234
  1901,
1235
- 1902,
1236
  1903,
1237
  1904,
1238
  1905,
 
1239
  1907,
1240
  1909,
1241
  1910,
1242
  1911,
1243
- 1912,
1244
  1913,
1245
- 1914,
1246
  1915,
1247
- 1917,
1248
  1918,
1249
  1919,
 
1250
  1921,
1251
  1922,
1252
- 1923,
1253
  1924,
1254
  1925,
1255
  1926,
 
1256
  1928,
1257
- 1930,
1258
  1931,
1259
- 1933,
 
1260
  1935,
1261
  1936,
1262
  1938,
@@ -1265,33 +1263,53 @@
1265
  1942,
1266
  1943,
1267
  1945,
 
 
1268
  1948,
1269
- 1565,
1270
  1949,
1271
- 1950,
1272
  1951,
1273
  1952,
1274
  1953,
1275
  1954,
1276
- 1955,
1277
  1956,
1278
- 1957,
1279
  1958,
1280
  1959,
1281
  1960,
1282
  1961,
1283
  1962,
1284
  1963,
1285
- 1964,
1286
  1965,
1287
- 1966,
1288
  1967,
1289
  1968,
1290
- 1969,
1291
  1972,
1292
- 1973,
1293
  1974,
1294
- 154,
1295
- 1976
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1296
  ]
1297
  }
56
  113,
57
  115,
58
  117,
 
59
  119,
60
  121,
61
+ 123,
62
+ 125,
63
  126,
 
64
  128,
65
  130,
66
+ 131,
67
  133,
68
+ 134,
69
  135,
70
+ 137,
71
+ 139,
72
  140,
73
+ 141,
74
+ 143,
75
  145,
76
+ 148,
77
+ 150,
78
+ 152,
79
+ 154,
 
80
  155,
81
+ 156,
82
+ 158,
83
+ 160,
84
  161,
85
  163,
 
86
  166,
87
+ 168,
88
  170,
89
  171,
90
  173,
91
+ 174,
92
  175,
93
+ 176,
94
  178,
95
  180,
96
  182,
97
  184,
98
  186,
 
 
99
  190,
100
+ 191,
101
  192,
102
+ 194,
 
103
  197,
104
+ 198,
105
  200,
106
+ 202,
107
+ 204,
108
+ 205,
109
  207,
110
  208,
111
+ 209,
112
+ 211,
113
  215,
114
  216,
 
115
  218,
116
  220,
117
+ 223,
118
  224,
119
+ 225,
120
  227,
121
+ 229,
122
+ 232,
123
+ 235,
124
  239,
 
125
  242,
126
  244,
127
  245,
128
+ 246,
129
+ 248,
130
  250,
131
+ 251,
132
  252,
133
+ 253,
134
+ 256,
135
  257,
136
  260,
137
+ 262,
138
  264,
139
+ 265,
140
  266,
141
+ 268,
 
142
  271,
143
  272,
144
  273,
145
  275,
 
146
  278,
147
+ 279,
148
  281,
149
  282,
150
  283,
151
+ 286,
 
152
  288,
153
+ 289,
154
  290,
155
  291,
 
156
  293,
157
+ 294,
158
+ 295,
159
  296,
160
  297,
 
161
  299,
162
+ 300,
 
163
  303,
164
  304,
 
165
  307,
166
  308,
167
+ 310,
168
+ 312,
169
+ 313,
170
  314,
171
+ 315,
172
  317,
173
+ 318,
174
  319,
175
  320,
176
  321,
 
177
  323,
178
+ 325,
179
  326,
180
+ 327,
181
  329,
182
+ 330,
183
+ 333,
184
+ 338,
185
  339,
186
+ 343,
 
 
187
  345,
188
+ 346,
189
+ 348,
190
+ 349,
191
+ 350,
192
  352,
193
  354,
 
 
194
  359,
195
+ 361,
196
  362,
197
  363,
 
198
  365,
 
199
  367,
200
  370,
201
+ 371,
202
+ 374,
203
  377,
 
204
  379,
205
+ 382,
 
206
  383,
207
  384,
208
  385,
209
+ 387,
210
  388,
211
+ 390,
212
  391,
213
+ 392,
214
  393,
 
215
  395,
216
  396,
217
+ 398,
218
  400,
219
+ 401,
220
  402,
221
+ 403,
222
  404,
223
+ 407,
224
  409,
 
225
  411,
226
  412,
227
+ 413,
228
+ 417,
229
  418,
230
+ 419,
 
231
  422,
232
+ 425,
 
233
  427,
234
+ 428,
235
+ 429,
236
  430,
237
  431,
238
+ 434,
 
 
239
  437,
240
+ 438,
241
  439,
242
  440,
 
243
  442,
244
  443,
245
  444,
246
  445,
247
+ 446,
248
  449,
249
+ 450,
250
  452,
251
  453,
252
+ 454,
253
  457,
254
  458,
255
+ 459,
256
+ 464,
257
  466,
258
  468,
259
  471,
263
  477,
264
  478,
265
  479,
266
+ 480,
267
  482,
268
  483,
269
  484,
 
270
  486,
 
 
271
  489,
272
  490,
273
+ 493,
274
+ 494,
275
  495,
276
+ 498,
 
 
277
  501,
278
+ 503,
279
  504,
280
+ 505,
281
+ 506,
282
  507,
283
+ 508,
284
  510,
285
  511,
286
  512,
288
  514,
289
  515,
290
  516,
291
+ 518,
292
  519,
 
293
  521,
294
+ 522,
295
+ 525,
296
  526,
297
  527,
298
+ 528,
299
  530,
300
+ 531,
301
  533,
302
+ 536,
303
+ 538,
304
+ 541,
305
  543,
306
+ 544,
307
  546,
308
+ 548,
309
+ 554,
310
+ 557,
311
+ 558,
 
312
  560,
313
+ 563,
 
314
  564,
315
+ 567,
 
 
 
316
  571,
 
 
317
  574,
318
+ 577,
 
 
319
  580,
320
+ 583,
321
+ 585,
322
  587,
323
+ 588,
324
  590,
325
  592,
326
+ 593,
327
  594,
328
  595,
329
  596,
330
  597,
 
331
  599,
332
  600,
333
+ 601,
334
  602,
335
+ 606,
336
  607,
337
  608,
338
+ 611,
339
+ 613,
340
+ 615,
341
  616,
342
  617,
343
+ 618,
344
  620,
345
  621,
346
+ 622,
347
  624,
348
  625,
349
  626,
350
  627,
351
+ 628,
352
  630,
353
  631,
354
+ 632,
355
  633,
356
  634,
357
+ 635,
358
+ 636,
359
  638,
360
+ 639,
361
  640,
362
+ 643,
363
  644,
364
  645,
365
+ 646,
366
  647,
367
+ 651,
 
368
  652,
369
  654,
370
+ 655,
371
  656,
372
  657,
 
373
  659,
374
  660,
375
  661,
379
  667,
380
  671,
381
  672,
382
+ 673,
383
  678,
384
  679,
385
+ 680,
386
+ 683,
387
  685,
388
+ 686,
389
+ 688,
390
+ 690,
391
  692,
392
+ 693,
393
  695,
394
  696,
395
  697,
396
  698,
397
+ 699,
398
  701,
399
  702,
 
400
  704,
401
+ 705,
402
  706,
 
 
403
  709,
404
  710,
405
  711,
406
+ 713,
407
  714,
408
+ 545,
409
+ 715,
410
+ 716,
411
  721,
412
  722,
413
+ 723,
414
+ 725,
415
+ 729,
416
+ 730,
417
  731,
418
  732,
419
  733,
420
  734,
421
  735,
422
+ 736,
423
  737,
424
+ 740,
425
  741,
426
  742,
427
  743,
429
  745,
430
  746,
431
  747,
 
432
  749,
433
+ 752,
434
  753,
435
  754,
436
+ 755,
 
437
  758,
438
+ 760,
439
  761,
440
+ 762,
441
+ 763,
442
+ 765,
443
  768,
 
444
  770,
445
  772,
446
+ 773,
447
+ 774,
448
  775,
449
  777,
 
 
450
  780,
 
451
  782,
452
+ 783,
453
  784,
454
+ 785,
455
+ 786,
456
+ 788,
457
+ 791,
458
+ 793,
459
  796,
 
460
  799,
461
  800,
462
+ 802,
463
  804,
 
 
464
  808,
465
+ 809,
466
+ 811,
467
  812,
468
+ 814,
469
  816,
470
  820,
471
+ 824,
 
472
  827,
473
+ 830,
474
  832,
 
 
475
  837,
476
+ 838,
477
  841,
478
  842,
 
479
  844,
 
480
  846,
481
+ 847,
482
  852,
483
  853,
484
  854,
485
  858,
486
  859,
487
  860,
488
+ 862,
489
  863,
490
+ 866,
491
+ 869,
492
  870,
493
  871,
 
494
  873,
495
+ 874,
496
+ 877,
497
  879,
498
  880,
499
  881,
500
+ 882,
501
  883,
 
502
  885,
503
+ 886,
504
+ 887,
505
  891,
506
  892,
507
  893,
508
  894,
509
+ 895,
510
+ 896,
511
+ 897,
512
+ 898,
513
  899,
 
 
 
514
  904,
 
515
  906,
516
+ 907,
517
  910,
518
  912,
519
+ 915,
520
  916,
521
  917,
522
+ 919,
523
+ 921,
524
  922,
525
  923,
526
  924,
527
+ 926,
528
  927,
529
  928,
 
530
  930,
531
+ 932,
532
+ 934,
533
  935,
534
+ 936,
535
  937,
536
+ 940,
537
+ 941,
538
  942,
539
  943,
540
+ 944,
541
  945,
542
  946,
543
+ 779,
544
  947,
545
  948,
546
  951,
 
547
  953,
548
  954,
549
  955,
552
  958,
553
  959,
554
  960,
555
+ 961,
556
  962,
557
+ 963,
558
  965,
559
  966,
560
  967,
561
+ 970,
562
+ 971,
563
  972,
564
  973,
565
  974,
566
+ 977,
567
  978,
568
  979,
569
+ 981,
570
+ 983,
571
  984,
572
  985,
 
573
  987,
574
  989,
575
+ 990,
576
  992,
577
+ 993,
578
  994,
579
+ 996,
 
580
  998,
581
+ 999,
582
  1000,
583
  1002,
584
  1004,
 
585
  1006,
586
  1007,
587
  1008,
588
+ 1009,
589
  1010,
 
590
  1012,
591
  1013,
592
  1014,
593
+ 1015,
594
  1017,
595
  1018,
596
  1019,
597
  1020,
598
+ 1021,
599
+ 1022,
600
+ 1025,
601
  1026,
602
+ 1028,
603
+ 1029,
 
604
  1032,
605
+ 1033,
606
+ 1035,
607
  1037,
608
  1038,
609
  1039,
610
  1040,
611
  1041,
 
612
  1043,
613
+ 1046,
614
+ 1049,
615
+ 1050,
616
  1052,
617
+ 1053,
618
  1054,
619
  1055,
620
  1056,
621
+ 1059,
622
+ 1061,
623
+ 1063,
624
  1064,
625
+ 1067,
626
+ 1069,
627
  1070,
628
+ 1072,
629
  1073,
 
630
  1075,
631
  1076,
632
  1078,
633
  1079,
634
  1080,
 
635
  1083,
636
  1084,
637
+ 1085,
638
  1086,
639
  1087,
640
  1088,
641
  1089,
642
  1090,
643
+ 1091,
644
  1093,
645
  1094,
 
646
  1096,
647
  1097,
 
648
  1099,
649
  1100,
650
+ 1101,
651
+ 1102,
652
  1103,
653
  1104,
654
+ 1105,
655
  1106,
656
  1107,
 
657
  1109,
658
  1110,
659
  1111,
660
+ 1112,
661
+ 1113,
662
  1114,
663
+ 1117,
664
  1118,
665
+ 1120,
666
+ 1121,
667
  1122,
668
  1123,
 
 
669
  1126,
670
+ 1128,
 
671
  1130,
 
 
672
  1134,
673
+ 1135,
674
  1136,
675
+ 1137,
676
  1138,
677
+ 1139,
678
  1140,
679
+ 523,
680
  1142,
 
681
  1145,
682
+ 196,
683
+ 1146,
684
+ 1147,
685
  1148,
686
  1149,
687
  1151,
688
+ 1153,
689
  1154,
 
690
  1156,
 
 
691
  1158,
692
  1159,
693
  1160,
694
  1161,
695
  1162,
696
+ 1164,
697
  1165,
 
 
698
  1168,
699
+ 1169,
700
  1170,
701
  1171,
702
+ 1173,
703
  1174,
 
704
  1176,
705
  1177,
706
+ 1178,
707
+ 50,
708
  1179,
709
  1180,
710
+ 1181,
711
+ 1182,
712
  1185,
713
  1186,
714
+ 1187,
715
  1188,
 
716
  1190,
717
+ 1191,
718
  1192,
719
  1194,
720
  1195,
721
+ 1197,
722
+ 1198,
723
  1199,
724
+ 1200,
725
  1201,
726
+ 1203,
 
 
727
  1206,
728
+ 1207,
729
  1208,
730
+ 1209,
731
+ 1210,
732
+ 1212,
733
+ 1213,
734
+ 1214,
735
  1216,
 
736
  1218,
737
  1219,
 
 
738
  1223,
739
+ 1225,
740
+ 1227,
741
+ 1228,
742
  1230,
 
743
  1233,
744
+ 1238,
745
+ 1239,
746
+ 1240,
 
747
  1241,
 
748
  1243,
749
  1244,
750
+ 1246,
 
751
  1249,
752
  1250,
753
  1251,
754
+ 1253,
755
  1254,
756
+ 707,
757
  1255,
758
+ 905,
759
  1256,
760
  1257,
 
 
761
  1262,
762
+ 1263,
763
  1264,
764
+ 1265,
765
  1267,
766
  1268,
767
  1269,
 
768
  1271,
769
  1272,
 
770
  1274,
771
  1276,
772
+ 1277,
773
  1278,
774
+ 1279,
775
  1280,
776
  1281,
777
  1283,
779
  1285,
780
  1286,
781
  1288,
782
+ 1289,
783
  1291,
 
784
  1294,
785
+ 1296,
786
  1297,
787
  1298,
788
+ 1299,
789
  1300,
790
+ 1302,
791
+ 1304,
792
+ 1305,
793
+ 1306,
794
+ 1308,
795
  1310,
 
796
  1312,
 
 
797
  1316,
 
798
  1318,
799
+ 1074,
 
800
  1320,
801
+ 1321,
802
  1322,
803
  1323,
804
  1324,
805
+ 1325,
806
  1327,
807
  1328,
808
+ 463,
809
+ 1329,
810
  1330,
811
  1331,
812
+ 1332,
813
+ 1333,
814
+ 1335,
815
  1336,
816
+ 1337,
817
  1340,
818
+ 1341,
819
  1342,
820
+ 1343,
821
  1345,
822
+ 1346,
823
  1347,
824
  1348,
825
  1349,
826
  1352,
 
827
  1354,
 
828
  1356,
829
+ 1357,
830
  1359,
831
  1360,
 
832
  1362,
833
  1363,
 
834
  1365,
 
835
  1367,
836
+ 1368,
837
  1371,
838
  1372,
839
+ 1373,
840
  1374,
841
  1375,
842
  1376,
 
 
843
  1379,
844
+ 1380,
845
+ 1381,
846
  1382,
847
+ 1383,
848
  1384,
849
  1385,
850
+ 1386,
851
  1387,
852
  1388,
853
+ 1389,
854
  1390,
855
+ 1391,
856
  1392,
857
  1393,
 
858
  1395,
859
+ 1396,
860
+ 1400,
861
  1401,
 
862
  1403,
863
  1404,
864
+ 1405,
865
  1406,
866
+ 1407,
867
+ 1409,
868
  1411,
869
+ 1413,
870
+ 1414,
871
  1416,
 
872
  1418,
873
  1419,
874
  1420,
 
875
  1423,
876
  1424,
877
+ 1261,
878
  1425,
 
879
  1426,
880
  1428,
881
+ 1429,
882
  1430,
 
 
883
  1433,
884
  1434,
885
  1435,
887
  1437,
888
  1438,
889
  1439,
 
890
  1441,
891
+ 267,
892
  1442,
893
  1444,
894
  1446,
895
  1447,
896
  1448,
897
  1449,
898
+ 1450,
899
  1451,
900
  1452,
901
  1453,
903
  1455,
904
  1456,
905
  1457,
 
906
  1459,
 
907
  1461,
908
  1462,
909
+ 1463,
910
  1465,
911
  1466,
912
  1467,
913
  1468,
 
914
  1470,
915
+ 1471,
916
  1472,
917
+ 1473,
918
  1474,
919
  1475,
920
  1476,
922
  1478,
923
  1479,
924
  1480,
 
925
  1481,
926
  1482,
 
927
  1484,
 
928
  1486,
929
+ 1487,
930
  1488,
931
  1489,
932
  1490,
933
+ 1491,
934
  1492,
935
  1493,
936
+ 1494,
937
+ 1495,
938
+ 801,
939
  1496,
940
  1497,
941
  1498,
942
  1499,
943
  1500,
944
  1502,
945
+ 1503,
946
  1505,
947
  1506,
 
948
  1508,
949
  1509,
 
950
  1510,
951
  1511,
952
+ 1513,
953
  1514,
 
954
  1516,
955
  1517,
956
  1518,
957
  1519,
958
+ 1520,
959
+ 1521,
960
+ 1155,
961
  1522,
962
  1523,
963
+ 1525,
964
  1526,
965
+ 1527,
966
+ 1528,
967
+ 1529,
968
  1530,
969
  1531,
970
  1532,
971
+ 1533,
 
972
  1536,
973
  1537,
974
  1538,
 
975
  1541,
976
+ 1545,
 
977
  1546,
978
  1547,
979
  1548,
980
  1549,
981
  1550,
982
+ 1551,
983
  1552,
984
  1553,
985
  1554,
986
  1556,
987
  1557,
 
988
  1559,
989
+ 1560,
990
+ 1561,
991
  1563,
992
  1564,
993
+ 1565,
994
  1566,
995
  1567,
996
+ 1569,
997
  1570,
998
  1572,
999
  1573,
1000
+ 1574,
1001
  1575,
1002
+ 1578,
1003
  1579,
1004
+ 1580,
1005
  1582,
1006
+ 1583,
1007
+ 1584,
1008
  1586,
1009
+ 1588,
1010
+ 1589,
1011
  1591,
 
1012
  1593,
1013
  1596,
 
1014
  1599,
1015
  1600,
1016
  1601,
 
1017
  1603,
1018
  1604,
 
1019
  1607,
1020
  1608,
 
1021
  1611,
1022
+ 1613,
1023
+ 1614,
1024
  1616,
1025
  1617,
1026
  1618,
1027
  1619,
1028
+ 1620,
1029
  1621,
 
1030
  1623,
1031
+ 1624,
1032
  1625,
 
1033
  1627,
1034
+ 1628,
1035
  1629,
 
 
1036
  1633,
1037
+ 1634,
1038
  1635,
1039
+ 1636,
1040
  1637,
1041
+ 1639,
1042
  1640,
1043
  1641,
1044
  1642,
 
1045
  1643,
1046
  1644,
1047
  1646,
 
1048
  1648,
 
1049
  1650,
1050
  1651,
 
 
1051
  1654,
1052
+ 1655,
1053
  1656,
1054
  1658,
1055
+ 1659,
1056
  1660,
1057
  1661,
1058
+ 1662,
1059
  1664,
1060
  1665,
1061
  1666,
1062
+ 1667,
1063
  1668,
1064
+ 1669,
1065
  1670,
1066
+ 1672,
 
1067
  1674,
 
1068
  1676,
 
1069
  1678,
 
1070
  1679,
1071
  1680,
1072
  1681,
1073
  1682,
 
1074
  1684,
1075
+ 1686,
1076
+ 1687,
1077
  1689,
1078
  1690,
1079
+ 1691,
1080
  1692,
1081
+ 1693,
1082
+ 1677,
1083
  1694,
1084
+ 1695,
1085
  1696,
1086
  1697,
1087
  1698,
1088
  1699,
 
 
1089
  1702,
1090
  1703,
1091
+ 1704,
1092
  1706,
1093
+ 1708,
1094
  1710,
1095
  1711,
1096
  1712,
1097
  1713,
1098
+ 1714,
1099
  1715,
1100
+ 1716,
1101
+ 1718,
1102
  1719,
1103
  1720,
1104
  1721,
1105
  1722,
1106
  1723,
1107
+ 1725,
1108
+ 1726,
 
1109
  1729,
1110
+ 1730,
1111
+ 1731,
1112
+ 1733,
1113
+ 1735,
1114
  1736,
1115
+ 1737,
1116
  1738,
1117
  1739,
1118
+ 1740,
1119
+ 1741,
1120
  1742,
1121
  1743,
 
1122
  1746,
1123
  1747,
 
 
1124
  1750,
 
1125
  1752,
1126
  1754,
 
1127
  1756,
1128
  1757,
1129
+ 1760,
 
1130
  1761,
 
1131
  1763,
1132
+ 1764,
1133
  1765,
1134
  1766,
1135
+ 1767,
1136
+ 1768,
1137
  1769,
1138
+ 1770,
1139
+ 1771,
1140
+ 1773,
1141
+ 1774,
1142
  1775,
1143
+ 1776,
1144
+ 1777,
1145
  1778,
1146
+ 1780,
1147
  1781,
1148
+ 1782,
1149
  1783,
1150
  1784,
1151
+ 1785,
1152
  1786,
 
 
 
 
1153
  1792,
 
1154
  1795,
 
 
1155
  1798,
 
1156
  1800,
1157
  1801,
 
1158
  1804,
 
1159
  1806,
1160
  1807,
1161
  1808,
 
1162
  1810,
1163
  1811,
 
1164
  1813,
1165
  1814,
1166
+ 1815,
1167
  1816,
1168
  1817,
1169
  1818,
1170
  1819,
1171
+ 1821,
1172
  1822,
1173
  1823,
1174
  1824,
1175
  1825,
1176
+ 1826,
1177
  1827,
1178
  1828,
 
1179
  1830,
1180
+ 1831,
1181
  1832,
1182
+ 1833,
1183
  1834,
1184
  1835,
1185
+ 1836,
1186
  1837,
 
1187
  1839,
1188
  1840,
1189
+ 1841,
1190
+ 1842,
1191
+ 1844,
1192
  1845,
1193
  1846,
 
1194
  1848,
1195
  1849,
1196
+ 929,
1197
  1850,
1198
+ 1852,
1199
  1853,
1200
+ 1855,
1201
+ 1858,
 
1202
  1860,
1203
+ 1861,
1204
  1862,
1205
  1863,
1206
  1864,
1207
+ 1865,
1208
  1867,
1209
  1868,
1210
  1869,
 
1211
  1871,
1212
+ 1872,
1213
  1874,
1214
  1875,
 
 
1215
  1878,
 
1216
  1880,
1217
  1881,
1218
+ 1882,
1219
  1883,
1220
  1884,
 
1221
  1887,
1222
  1888,
1223
+ 1889,
1224
  1890,
1225
  1891,
 
1226
  1893,
1227
  1894,
1228
+ 1895,
1229
  1896,
1230
+ 1897,
1231
  1898,
1232
  1899,
1233
  1901,
 
1234
  1903,
1235
  1904,
1236
  1905,
1237
+ 1906,
1238
  1907,
1239
  1909,
1240
  1910,
1241
  1911,
 
1242
  1913,
 
1243
  1915,
1244
+ 1916,
1245
  1918,
1246
  1919,
1247
+ 1920,
1248
  1921,
1249
  1922,
 
1250
  1924,
1251
  1925,
1252
  1926,
1253
+ 1927,
1254
  1928,
 
1255
  1931,
1256
+ 1932,
1257
+ 1934,
1258
  1935,
1259
  1936,
1260
  1938,
1263
  1942,
1264
  1943,
1265
  1945,
1266
+ 1946,
1267
+ 1947,
1268
  1948,
 
1269
  1949,
 
1270
  1951,
1271
  1952,
1272
  1953,
1273
  1954,
 
1274
  1956,
 
1275
  1958,
1276
  1959,
1277
  1960,
1278
  1961,
1279
  1962,
1280
  1963,
 
1281
  1965,
 
1282
  1967,
1283
  1968,
1284
+ 1970,
1285
  1972,
 
1286
  1974,
1287
+ 1977,
1288
+ 1978,
1289
+ 1980,
1290
+ 1581,
1291
+ 1981,
1292
+ 1982,
1293
+ 1983,
1294
+ 1984,
1295
+ 1986,
1296
+ 1987,
1297
+ 1988,
1298
+ 1989,
1299
+ 1990,
1300
+ 1991,
1301
+ 1992,
1302
+ 1993,
1303
+ 1994,
1304
+ 1995,
1305
+ 1996,
1306
+ 1997,
1307
+ 1998,
1308
+ 1999,
1309
+ 2002,
1310
+ 2003,
1311
+ 2004,
1312
+ 2005,
1313
+ 2006
1314
  ]
1315
  }
lemmatizer/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9635c87e56e769c00e07ac0ec8518e913ba6a21fc4b9bbd06b49c89132289692
3
- size 3977921
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04d0f793829d450f4e8924baa767e4c54a255d69ecf277016757897c2f569bb9
3
+ size 4033373
lemmatizer/trees CHANGED
Binary files a/lemmatizer/trees and b/lemmatizer/trees differ
meta.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
  "lang":"de",
3
  "name":"dep_news_trf",
4
- "version":"3.6.1",
5
- "description":"German transformer pipeline (bert-base-german-cased). Components: transformer, tagger, morphologizer, parser, lemmatizer (trainable_lemmatizer).",
6
  "author":"Explosion",
7
  "email":"contact@explosion.ai",
8
  "url":"https://explosion.ai",
9
  "license":"MIT",
10
- "spacy_version":">=3.6.0,<3.7.0",
11
- "spacy_git_version":"c067b5264",
12
  "vectors":{
13
  "width":0,
14
  "vectors":0,
@@ -819,67 +819,67 @@
819
  "token_p":0.9991808044,
820
  "token_r":0.9990356034,
821
  "token_f":0.9991081986,
822
- "tag_acc":0.9905775076,
823
- "pos_acc":0.9914628183,
824
- "morph_acc":0.9700866279,
825
- "morph_micro_p":0.9886266605,
826
- "morph_micro_r":0.9884199706,
827
- "morph_micro_f":0.9885233048,
828
  "morph_per_feat":{
829
  "Case":{
830
- "p":0.9763482571,
831
- "r":0.9779552629,
832
- "f":0.9771510993
833
  },
834
  "Definite":{
835
- "p":0.9996579637,
836
- "r":1.0,
837
- "f":0.9998289526
838
  },
839
  "Gender":{
840
- "p":0.9848792598,
841
- "r":0.9853240009,
842
- "f":0.9851015801
843
  },
844
  "Number":{
845
- "p":0.9918825162,
846
- "r":0.9934908501,
847
- "f":0.9926860317
848
  },
849
  "PronType":{
850
- "p":0.9982142857,
851
  "r":0.9971620855,
852
- "f":0.9976879082
853
  },
854
  "Degree":{
855
- "p":0.9956001408,
856
- "r":0.9874323617,
857
- "f":0.9914994304
858
  },
859
  "VerbForm":{
860
- "p":0.9950717806,
861
- "r":0.9973155804,
862
- "f":0.996192417
863
  },
864
  "Mood":{
865
- "p":0.9897197811,
866
- "r":0.9897197811,
867
- "f":0.9897197811
868
  },
869
  "Person":{
870
- "p":0.9991712707,
871
- "r":0.9991712707,
872
- "f":0.9991712707
873
  },
874
  "Tense":{
875
- "p":0.9973382133,
876
- "r":0.9973382133,
877
- "f":0.9973382133
878
  },
879
  "Reflex":{
880
- "p":0.9976359338,
881
- "r":1.0,
882
- "f":0.998816568
883
  },
884
  "Poss":{
885
  "p":1.0,
@@ -887,66 +887,66 @@
887
  "f":1.0
888
  },
889
  "Foreign":{
890
- "p":0.7647058824,
891
- "r":0.1959798995,
892
- "f":0.312
893
  }
894
  },
895
- "sents_p":0.9779699403,
896
- "sents_r":0.9816077702,
897
- "sents_f":0.9797854785,
898
- "dep_uas":0.9592566441,
899
- "dep_las":0.9484326996,
900
  "dep_las_per_type":{
901
  "cp":{
902
- "p":0.9908256881,
903
- "r":0.987202925,
904
- "f":0.989010989
905
  },
906
  "nk":{
907
- "p":0.9891317621,
908
- "r":0.9895748493,
909
- "f":0.9893532561
910
  },
911
  "sb":{
912
- "p":0.9767034507,
913
- "r":0.9780740221,
914
- "f":0.9773882559
915
  },
916
  "ag":{
917
- "p":0.9764125069,
918
- "r":0.9839690437,
919
- "f":0.9801762115
920
  },
921
  "mnr":{
922
- "p":0.8712352684,
923
- "r":0.8390079865,
924
- "f":0.8548179872
925
  },
926
  "mo":{
927
- "p":0.86918429,
928
- "r":0.8730528019,
929
- "f":0.8711142511
930
  },
931
  "oc":{
932
- "p":0.9640071986,
933
- "r":0.9674894642,
934
- "f":0.9657451923
935
  },
936
  "root":{
937
- "p":0.9857936998,
938
- "r":0.9894606324,
939
- "f":0.9876237624
940
  },
941
  "rc":{
942
- "p":0.8392857143,
943
- "r":0.8430493274,
944
- "f":0.841163311
945
  },
946
  "op":{
947
- "p":0.8259629101,
948
- "r":0.8628912072,
949
- "f":0.8440233236
950
  },
951
  "pm":{
952
  "p":1.0,
@@ -954,84 +954,89 @@
954
  "f":1.0
955
  },
956
  "nmc":{
957
- "p":0.9888178914,
958
- "r":0.9778830964,
959
- "f":0.9833200953
960
  },
961
  "oa":{
962
- "p":0.9628411707,
963
  "r":0.9656992084,
964
- "f":0.9642680718
965
  },
966
  "pd":{
967
- "p":0.9203125,
968
- "r":0.8661764706,
969
- "f":0.8924242424
970
  },
971
  "ep":{
972
- "p":0.9453125,
973
- "r":0.9097744361,
974
- "f":0.9272030651
975
  },
976
  "cd":{
977
- "p":0.9356499356,
978
- "r":0.9392764858,
979
- "f":0.9374597034
980
  },
981
  "cj":{
982
- "p":0.9112627986,
983
  "r":0.9215976331,
984
- "f":0.9164010787
985
  },
986
  "par":{
987
- "p":0.5279503106,
988
- "r":0.4166666667,
989
- "f":0.4657534247
990
  },
991
  "svp":{
992
- "p":0.9766990291,
993
- "r":0.9673076923,
994
- "f":0.9719806763
995
  },
996
  "pg":{
997
- "p":0.9220779221,
998
- "r":0.9044585987,
999
- "f":0.9131832797
1000
  },
1001
  "da":{
1002
- "p":0.9153439153,
1003
- "r":0.8804071247,
1004
- "f":0.897535668
1005
  },
1006
  "app":{
1007
- "p":0.81,
1008
- "r":0.8181818182,
1009
- "f":0.8140703518
1010
  },
1011
  "ac":{
1012
- "p":0.9565217391,
1013
- "r":0.9401709402,
1014
- "f":0.9482758621
 
 
 
 
 
1015
  },
1016
  "pnc":{
1017
- "p":0.9025710419,
1018
- "r":0.8662337662,
1019
- "f":0.8840291584
1020
  },
1021
  "ju":{
1022
- "p":0.93,
1023
- "r":0.9393939394,
1024
- "f":0.9346733668
1025
  },
1026
  "ng":{
1027
- "p":0.8839050132,
1028
- "r":0.8909574468,
1029
- "f":0.8874172185
1030
  },
1031
  "re":{
1032
- "p":0.8224852071,
1033
- "r":0.8273809524,
1034
- "f":0.824925816
1035
  },
1036
  "dep":{
1037
  "p":0.0,
@@ -1039,44 +1044,39 @@
1039
  "f":0.0
1040
  },
1041
  "cm":{
1042
- "p":0.9661835749,
1043
- "r":0.9615384615,
1044
- "f":0.9638554217
1045
  },
1046
  "cc":{
1047
- "p":0.7759562842,
1048
- "r":0.7135678392,
1049
- "f":0.7434554974
1050
  },
1051
  "sbp":{
1052
- "p":0.9734513274,
1053
- "r":0.9482758621,
1054
- "f":0.96069869
1055
  },
1056
  "uc":{
1057
- "p":0.7619047619,
1058
- "r":0.5333333333,
1059
- "f":0.6274509804
1060
  },
1061
  "cvc":{
1062
- "p":0.8157894737,
1063
  "r":0.7848101266,
1064
- "f":0.8
1065
  },
1066
  "ams":{
1067
- "p":0.75,
1068
- "r":0.8823529412,
1069
- "f":0.8108108108
1070
- },
1071
- "rs":{
1072
- "p":0.7777777778,
1073
- "r":0.4516129032,
1074
- "f":0.5714285714
1075
  },
1076
  "ph":{
1077
- "p":0.8333333333,
1078
- "r":0.7142857143,
1079
- "f":0.7692307692
1080
  },
1081
  "dm":{
1082
  "p":1.0,
@@ -1094,23 +1094,18 @@
1094
  "f":0.0
1095
  },
1096
  "og":{
1097
- "p":0.4166666667,
1098
  "r":0.7142857143,
1099
- "f":0.5263157895
1100
- },
1101
- "avc":{
1102
- "p":0.0,
1103
- "r":0.0,
1104
- "f":0.0
1105
  },
1106
  "adc":{
1107
- "p":0.75,
1108
  "r":1.0,
1109
- "f":0.8571428571
1110
  }
1111
  },
1112
- "lemma_acc":0.9872611465,
1113
- "speed":2933.9454548672
1114
  },
1115
  "sources":[
1116
  {
@@ -1133,6 +1128,6 @@
1133
  }
1134
  ],
1135
  "requirements":[
1136
- "spacy-transformers>=1.2.2,<1.3.0"
1137
  ]
1138
  }
1
  {
2
  "lang":"de",
3
  "name":"dep_news_trf",
4
+ "version":"3.7.2",
5
+ "description":"German transformer pipeline (Transformer(name='bert-base-german-cased', piece_encoder='bert-wordpiece', stride=136, type='bert', width=768, window=176, vocab_size=30000)). Components: transformer, tagger, morphologizer, parser, lemmatizer (trainable_lemmatizer).",
6
  "author":"Explosion",
7
  "email":"contact@explosion.ai",
8
  "url":"https://explosion.ai",
9
  "license":"MIT",
10
+ "spacy_version":">=3.7.0,<3.8.0",
11
+ "spacy_git_version":"6b4f77441",
12
  "vectors":{
13
  "width":0,
14
  "vectors":0,
819
  "token_p":0.9991808044,
820
  "token_r":0.9990356034,
821
  "token_f":0.9991081986,
822
+ "tag_acc":0.9907095376,
823
+ "pos_acc":0.9917734358,
824
+ "morph_acc":0.9700601968,
825
+ "morph_micro_p":0.9885202304,
826
+ "morph_micro_r":0.9886522681,
827
+ "morph_micro_f":0.9885862448,
828
  "morph_per_feat":{
829
  "Case":{
830
+ "p":0.9765625,
831
+ "r":0.9781171582,
832
+ "f":0.9773392108
833
  },
834
  "Definite":{
835
+ "p":0.9997719238,
836
+ "r":0.9998859489,
837
+ "f":0.9998289331
838
  },
839
  "Gender":{
840
+ "p":0.9843639949,
841
+ "r":0.986114247,
842
+ "f":0.9852383436
843
  },
844
  "Number":{
845
+ "p":0.9916774552,
846
+ "r":0.9936529998,
847
+ "f":0.9926642446
848
  },
849
  "PronType":{
850
+ "p":0.9983763598,
851
  "r":0.9971620855,
852
+ "f":0.9977688532
853
  },
854
  "Degree":{
855
+ "p":0.9952523299,
856
+ "r":0.9879560133,
857
+ "f":0.9915907498
858
  },
859
  "VerbForm":{
860
+ "p":0.9949646454,
861
+ "r":0.9972082036,
862
+ "f":0.9960851612
863
  },
864
  "Mood":{
865
+ "p":0.9903862092,
866
+ "r":0.990714641,
867
+ "f":0.9905503979
868
  },
869
  "Person":{
870
+ "p":0.998619166,
871
+ "r":0.9988950276,
872
+ "f":0.9987570778
873
  },
874
  "Tense":{
875
+ "p":0.9966744263,
876
+ "r":0.9971718516,
877
+ "f":0.9969230769
878
  },
879
  "Reflex":{
880
+ "p":0.9976303318,
881
+ "r":0.9976303318,
882
+ "f":0.9976303318
883
  },
884
  "Poss":{
885
  "p":1.0,
887
  "f":1.0
888
  },
889
  "Foreign":{
890
+ "p":0.8333333333,
891
+ "r":0.175879397,
892
+ "f":0.2904564315
893
  }
894
  },
895
+ "sents_p":0.9775396662,
896
+ "sents_r":0.9803678446,
897
+ "sents_f":0.9789517128,
898
+ "dep_uas":0.958309546,
899
+ "dep_las":0.9473953052,
900
  "dep_las_per_type":{
901
  "cp":{
902
+ "p":0.9853479853,
903
+ "r":0.9835466179,
904
+ "f":0.9844464776
905
  },
906
  "nk":{
907
+ "p":0.9890901689,
908
+ "r":0.9894526796,
909
+ "f":0.9892713911
910
  },
911
  "sb":{
912
+ "p":0.9754299754,
913
+ "r":0.9749166813,
914
+ "f":0.9751732608
915
  },
916
  "ag":{
917
+ "p":0.9763346175,
918
+ "r":0.9806522941,
919
+ "f":0.9784886928
920
  },
921
  "mnr":{
922
+ "p":0.8643410853,
923
+ "r":0.8436317781,
924
+ "f":0.8538608807
925
  },
926
  "mo":{
927
+ "p":0.8667875857,
928
+ "r":0.8701193607,
929
+ "f":0.8684502776
930
  },
931
  "oc":{
932
+ "p":0.9654965497,
933
+ "r":0.9686935581,
934
+ "f":0.9670924117
935
  },
936
  "root":{
937
+ "p":0.9857790602,
938
+ "r":0.988427361,
939
+ "f":0.9871014343
940
  },
941
  "rc":{
942
+ "p":0.8258928571,
943
+ "r":0.8295964126,
944
+ "f":0.8277404922
945
  },
946
  "op":{
947
+ "p":0.8216833096,
948
+ "r":0.8584202683,
949
+ "f":0.8396501458
950
  },
951
  "pm":{
952
  "p":1.0,
954
  "f":1.0
955
  },
956
  "nmc":{
957
+ "p":0.9857369255,
958
+ "r":0.9826224329,
959
+ "f":0.9841772152
960
  },
961
  "oa":{
962
+ "p":0.9581151832,
963
  "r":0.9656992084,
964
+ "f":0.961892247
965
  },
966
  "pd":{
967
+ "p":0.9106317411,
968
+ "r":0.8691176471,
969
+ "f":0.8893905192
970
  },
971
  "ep":{
972
+ "p":0.9323308271,
973
+ "r":0.9323308271,
974
+ "f":0.9323308271
975
  },
976
  "cd":{
977
+ "p":0.9385908209,
978
+ "r":0.9379844961,
979
+ "f":0.9382875606
980
  },
981
  "cj":{
982
+ "p":0.9175257732,
983
  "r":0.9215976331,
984
+ "f":0.9195571956
985
  },
986
  "par":{
987
+ "p":0.5632911392,
988
+ "r":0.4362745098,
989
+ "f":0.4917127072
990
  },
991
  "svp":{
992
+ "p":0.9763779528,
993
+ "r":0.9538461538,
994
+ "f":0.9649805447
995
  },
996
  "pg":{
997
+ "p":0.9185667752,
998
+ "r":0.898089172,
999
+ "f":0.9082125604
1000
  },
1001
  "da":{
1002
+ "p":0.9189189189,
1003
+ "r":0.8651399491,
1004
+ "f":0.8912188729
1005
  },
1006
  "app":{
1007
+ "p":0.8361204013,
1008
+ "r":0.8417508418,
1009
+ "f":0.8389261745
1010
  },
1011
  "ac":{
1012
+ "p":0.9732142857,
1013
+ "r":0.9316239316,
1014
+ "f":0.9519650655
1015
+ },
1016
+ "rs":{
1017
+ "p":0.6842105263,
1018
+ "r":0.4193548387,
1019
+ "f":0.52
1020
  },
1021
  "pnc":{
1022
+ "p":0.8959459459,
1023
+ "r":0.861038961,
1024
+ "f":0.8781456954
1025
  },
1026
  "ju":{
1027
+ "p":0.9215686275,
1028
+ "r":0.9494949495,
1029
+ "f":0.9353233831
1030
  },
1031
  "ng":{
1032
+ "p":0.8759894459,
1033
+ "r":0.8829787234,
1034
+ "f":0.8794701987
1035
  },
1036
  "re":{
1037
+ "p":0.8303030303,
1038
+ "r":0.8154761905,
1039
+ "f":0.8228228228
1040
  },
1041
  "dep":{
1042
  "p":0.0,
1044
  "f":0.0
1045
  },
1046
  "cm":{
1047
+ "p":0.9428571429,
1048
+ "r":0.9519230769,
1049
+ "f":0.9473684211
1050
  },
1051
  "cc":{
1052
+ "p":0.7700534759,
1053
+ "r":0.7236180905,
1054
+ "f":0.7461139896
1055
  },
1056
  "sbp":{
1057
+ "p":0.9380530973,
1058
+ "r":0.9137931034,
1059
+ "f":0.9257641921
1060
  },
1061
  "uc":{
1062
+ "p":0.8666666667,
1063
+ "r":0.4333333333,
1064
+ "f":0.5777777778
1065
  },
1066
  "cvc":{
1067
+ "p":0.8378378378,
1068
  "r":0.7848101266,
1069
+ "f":0.8104575163
1070
  },
1071
  "ams":{
1072
+ "p":0.7625,
1073
+ "r":0.8970588235,
1074
+ "f":0.8243243243
 
 
 
 
 
1075
  },
1076
  "ph":{
1077
+ "p":0.8571428571,
1078
+ "r":0.8571428571,
1079
+ "f":0.8571428571
1080
  },
1081
  "dm":{
1082
  "p":1.0,
1094
  "f":0.0
1095
  },
1096
  "og":{
1097
+ "p":0.4545454545,
1098
  "r":0.7142857143,
1099
+ "f":0.5555555556
 
 
 
 
 
1100
  },
1101
  "adc":{
1102
+ "p":0.6,
1103
  "r":1.0,
1104
+ "f":0.75
1105
  }
1106
  },
1107
+ "lemma_acc":0.9872479319,
1108
+ "speed":2451.8342899914
1109
  },
1110
  "sources":[
1111
  {
1128
  }
1129
  ],
1130
  "requirements":[
1131
+ "spacy-curated-transformers>=0.2.0,<0.3.0"
1132
  ]
1133
  }
morphologizer/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dafaa8f2141aba369eb18c3a6727746f13ecc43c975715c65fa0be05f6d65e82
3
- size 2061573
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42b7605d03ad031bc76ccd2ee122e163ac921f99128bf8cfd65ec15d367d57fc
3
+ size 2061657
parser/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8eec7f88b6a5e5244ba4ea9f103c1e7ca70c45bbb6406e5d0b12d9d2c4c06dfe
3
- size 768761
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cc78ddff0819cad730c1f9459a225e1f4b7a6b5ac154d4931ee505559930441
3
+ size 768845
tagger/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbbb191c59d892fc8135e553b55f152dd944d3f8d57f678b05e8f2a4d018063b
3
- size 166746
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f005994f1730473bcfff0d156af0f34e0ded9f05a458e30fd536bb277e63b3b
3
+ size 166830
transformer/cfg CHANGED
@@ -1,3 +1,3 @@
1
  {
2
- "max_batch_items":4096
3
  }
1
  {
2
+
3
  }
transformer/model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef8c2744ef38035601218e95c13856630fccd542cd509f3bf474ed63e28dc44f
3
- size 437388252
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5fda10bea0dc2a2d913d77b6ed66002cd1c2b433093d6cf5a1a5589c5054ac3
3
+ size 434276235
vocab/strings.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6a2bd3d2d62b594c5a1859c8a4a86a3cc82b314b27c93e5859aad8503a45455
3
- size 3040776
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5ecec2dc56d0e6ad02d9535368dd71aceb6299417b90e6ea6621145c3da8166
3
+ size 3041501