|
IWPT 2020 Shared Task on Parsing into Enhanced Universal Dependencies |
|
--------------------------------------------------------------------- |
|
|
|
This package contains data used in the IWPT 2020 shared task. The package is available from |
|
http://hdl.handle.net/11234/1-3238 (permanent URL). |
|
|
|
For more information on the shared task, see the IWPT conference proceedings in the ACL anthology |
|
https://www.aclweb.org/anthology/sigs/sigparse/ |
|
as well as the shared task web site: |
|
https://universaldependencies.org/iwpt20/ |
|
|
|
The package contains training, development and test (evaluation) datasets as they were used |
|
during the shared task. The data is based on a subset of Universal Dependencies release 2.5 |
|
(http://hdl.handle.net/11234/1-3105) but some treebanks contain additional enhanced annotations. |
|
Moreover, not all of these additions became part of Universal Dependencies release 2.6 |
|
(http://hdl.handle.net/11234/1-3226), which makes the shared task data unique and worth |
|
a separate release to enable later comparison with new parsing algorithms. |
|
|
|
|
|
LICENSE |
|
------- |
|
|
|
The package is distributed under the same license as Universal Dependencies 2.5. This is |
|
a collective license, i.e., individual treebanks may use different license terms. See |
|
https://lindat.mff.cuni.cz/repository/xmlui/page/licence-UD-2.5?locale-attribute=en |
|
for details. |
|
|
|
|
|
FILES AND FOLDERS |
|
----------------- |
|
|
|
There are 17 languages and 28 treebanks. Some treebanks contain training, development and |
|
test data. Some treebanks contain only the test data. Each treebank has a folder named |
|
UD_Language-TreebankID: |
|
|
|
UD_Arabic-PADT |
|
UD_Bulgarian-BTB |
|
UD_Czech-CAC |
|
UD_Czech-FicTree |
|
UD_Czech-PDT |
|
UD_Czech-PUD |
|
UD_Dutch-Alpino |
|
UD_Dutch-LassySmall |
|
UD_English-EWT |
|
UD_English-PUD |
|
UD_Estonian-EDT |
|
UD_Estonian-EWT |
|
UD_Finnish-PUD |
|
UD_Finnish-TDT |
|
UD_French-FQB |
|
UD_French-Sequoia |
|
UD_Italian-ISDT |
|
UD_Latvian-LVTB |
|
UD_Lithuanian-ALKSNIS |
|
UD_Polish-LFG |
|
UD_Polish-PDB |
|
UD_Polish-PUD |
|
UD_Russian-SynTagRus |
|
UD_Slovak-SNK |
|
UD_Swedish-PUD |
|
UD_Swedish-Talbanken |
|
UD_Tamil-TTB |
|
UD_Ukrainian-IU |
|
|
|
Each folder contains one or more CoNLL-U files with gold-standard annotation following |
|
the UD guidelines (https://universaldependencies.org/guidelines.html), and one or |
|
more corresponding plain text files (the parsing systems get these files as input and |
|
produce parsed CoNLL-U files on output). |
|
|
|
During the shared task, only the training and development portions were distributed in this |
|
form. The test data was distributed blind, i.e., only the plain text input without annotation. |
|
Moreover, test sets from all treebanks of one language were merged in one file and the |
|
participants were not told which part of the text comes from which treebank. In this package, |
|
we add four folders: |
|
|
|
test-blind |
|
test-gold |
|
dev-blind |
|
dev-gold |
|
|
|
The *-blind folders contain 17 text files each, named using language codes: |
|
|
|
ar.txt (Arabic) |
|
bg.txt (Bulgarian) |
|
cs.txt (Czech) |
|
en.txt (English) |
|
et.txt (Estonian) |
|
fi.txt (Finnish) |
|
fr.txt (French) |
|
it.txt (Italian) |
|
lt.txt (Lithuanian) |
|
lv.txt (Latvian) |
|
nl.txt (Dutch) |
|
pl.txt (Polish) |
|
ru.txt (Russian) |
|
sk.txt (Slovak) |
|
sv.txt (Swedish) |
|
ta.txt (Tamil) |
|
uk.txt (Ukrainian) |
|
|
|
The *-gold folders contain the corresponding gold-standard annotated files in two flavors: |
|
the real UD-compliant annotation, e.g., ar.conllu, and a file where empty nodes have been |
|
collapsed to make evaluation possible, e.g. ar.nen.conllu (nen = no empty nodes). In addition |
|
to the test files, we also provide the development files in the same form (they were also |
|
available to the participants of the shared task). |
|
|
|
|
|
FRENCH |
|
------ |
|
|
|
In addition to enhancements described in the UD v2 guidelines |
|
(https://universaldependencies.org/u/overview/enhanced-syntax.html), |
|
the French data also show neutralized diathesis in the spirit of (Candito et al. 2017). |
|
It is not possible to squeeze this information into the dependency labels in a way that would |
|
be both human-readable and valid according to the UD guidelines. Therefore, the French CoNLL-U |
|
files are provided in two flavors: "fulldeps" and "xoxdeps". The former is the intended, human- |
|
readable format where final and canonical grammatical functions are separated by the "@" |
|
character; e.g., "obl:agent@nsubj" means that the dependent is an oblique agent phrase (the |
|
final function) but canonically it corresponds to the subject of the verb in active form (the |
|
canonical function). Such dependency labels do not comply with the current UD guidelines which |
|
do not allow the "@" character in dependency labels (also, the full labels sometimes contain |
|
more colons ":" than permitted). The labels thus have been transformed reducing the number of |
|
colons and replacing "@" with "xox", hence the xoxdeps.conllu files. The systems participating |
|
in the shared task worked with the xoxdeps files, as these pass the official validation. However, |
|
the cryptic xoxdeps labels can be easily converted back to the original format, even in the |
|
parser output (provided the parser predicted the label correctly; see the tools below). |
|
|
|
|
|
TOOLS |
|
----- |
|
|
|
The package also contains a number of Perl and Python scripts that have been used to process |
|
the data during preparation and during the shared task. They are in the "tools" folder. |
|
|
|
validate.py |
|
|
|
The official Universal Dependencies validator. It is a Python 3 script, and it needs the |
|
third-party module regex to be installed on the system (use pip to install it). The script |
|
recognizes several levels of validity; system output in the shared task must be valid on |
|
level 2: |
|
|
|
python validate.py --level 2 --lang ud file.conllu |
|
|
|
enhanced_collapse_empty_nodes.pl |
|
|
|
Removes empty nodes from a CoNLL-U file and transforms all paths traversing an empty node |
|
into a long edge with a combined label, e.g., "conj>nsubj". Note that the output is not |
|
valid according to UD guidelines, as the ">" character cannot occur in a normal CoNLL-U |
|
file. After passing validation, system outputs and gold-standard files are processed with |
|
this script and then they can be evaluated (the evaluator cannot work with empty nodes |
|
directly). |
|
|
|
perl enhanced_collapse_empty_nodes.pl file.conllu > file.nen.conllu |
|
|
|
iwpt20_xud_eval.py |
|
|
|
The official evaluation script in the shared task. It takes valid gold-standard and |
|
system-output file after these have been processed with enhanced_collapse_empty_nodes.pl. |
|
It also requires that the text of the gold standard and system output differ only in |
|
whitespace characters (tokenization) while all non-whitespace characters must be identical |
|
(no normalization is allowed). |
|
|
|
python iwpt20_xud_eval.py -v gold.nen.conllu system.nen.conllu |
|
|
|
The script can be told that certain types of enhancements should not be evaluated. This |
|
is done with treebanks where some enhancements are not annotated in the gold standard and |
|
we do not want to penalize the system for predicting the enhancement. For example, |
|
|
|
--enhancements 156 |
|
|
|
means that gapping (1), relative clauses (5), and case-enhanced deprels (6) should be |
|
ignored. |
|
|
|
conllu-quick-fix.pl |
|
|
|
A Perl script that tries to make an invalid file valid by filling out UPOS tags if they |
|
are empty, fixing the format of morphological features etc. It can also make sure that |
|
every node in the enhanced graph is reachable from the virtual root node with id 0; |
|
however, this function is optional, as it modifies the system-output dependency structure, |
|
and the algorithm it uses is not optimal in terms of evaluation score. |
|
|
|
perl conllu-quick-fix.pl input.conllu > output.conllu |
|
perl conllu-quick-fix.pl --connect-to-root input.conllu > output.conllu |
|
|
|
conllu_to_text.pl |
|
|
|
This script was used to generate the untokenized text files that serve as the input for |
|
the trained parsing system (the "blind" data). It takes the ISO 639-1 code of the language |
|
of the text (e.g., "--language en") because it works a bit differently with languages that |
|
use Chinese characters. The option was not needed for the present shared task, as there |
|
were no such languages. |
|
|
|
perl conllu_to_text.pl --language xx file.conllu > file.txt |
|
|
|
mergetest.pl |
|
|
|
This script was used to merge the test data (both gold and blind) from multiple treebanks |
|
of one language. The test sets are essentially concatenated, but the script makes sure that |
|
there is a new document mark at the beginning of each gold CoNLL-U file, and between two |
|
text files there is an empty line. This increases the chance that the systems will break |
|
a sentence at this position and it will be possible to separate data from individual treebanks |
|
in the system output. The first argument of the script is the name of the target file, the |
|
remaining arguments are names of the source files (any number, but at least one). By default, |
|
the files are processed as text files. If the target file name ends in ".conllu", the files are |
|
processed as CoNLL-U files. |
|
|
|
perl mergetest.pl nl.conllu UD_Dutch-Alpino/nl_alpino-ud-test.conllu UD_Dutch-LassySmall/nl_lassysmall-ud-test.conllu |
|
perl mergetest.pl nl.txt UD_Dutch-Alpino/nl_alpino-ud-test.txt UD_Dutch-LassySmall/nl_lassysmall-ud-test.txt |
|
|
|
match_and_split_conllu_by_input_text.pl |
|
|
|
This script reverses mergetest.pl, i.e., splits the system output for a language into smaller |
|
files corresponding to individual treebanks. This must be done if we want to evaluate each |
|
treebank with different settings, i.e., ignoring enhancements that are not gold-annotated |
|
in the treebank. The script takes the input text of one treebank and the CoNLL-U file that |
|
starts with annotation of the input text but possibly contains more annotation of other text. |
|
Two CoNLL-U files will be generated: the first one corresponds to the input text, the second |
|
one is the rest. Therefore, if the language consists of more than two treebanks, the script |
|
must be run multiple times. The script can handle sentences that cross the file boundary |
|
(nodes whose parents lie in the other file will be re-attached to some other parent). However, |
|
the script cannot handle situations where a token crosses the file boundary. |
|
|
|
perl match_and_split_conllu_by_input_text.pl UD_Dutch-Alpino/nl_alpino-ud-test.txt nl.conllu nl_alpino.conllu nl_rest.conllu |
|
|
|
evaluate_all.pl |
|
|
|
This script takes one shared task submission (tgz archive with CoNLL-U files for individual |
|
languages), unpacks it, checks whether the files are valid and evaluates them. The script is |
|
provided as-is and it is not ready to be run outside the shared task submission site. Various |
|
paths are hard-coded in the source code! However, the code shows how the evaluation was done |
|
using the other tools described above. |
|
|
|
perl evaluate_all.pl team_name submission_id dev|test |
|
|
|
html_evaluation.pl |
|
|
|
This script relies on the same fixed folder structure as evaluate_all.pl, and it, too, has |
|
a path hard-wired in the source code. It scans the evaluation logs produced by evaluate_all.pl |
|
and creates a HTML file with score tables for the submission. |
|
|
|
perl html_evaluation.pl team_name submission_id dev|test |
|
|
|
expand_edeps_in_french.pl |
|
|
|
Takes a CoNLL-U file with the cryptic-but-ud-valid encoding of the French double relations, |
|
("xoxdeps") and converts them back to the human-readable form ("fulldeps"). |
|
|
|
perl expand_edeps_in_french.pl fr.xoxdeps.conllu > fr.fulldeps.conllu |
|
|
|
enhanced_graph_properties.pl |
|
|
|
Reads a CoNLL-U file and collects statistics about the enhanced graphs found in the DEPS |
|
column. Some of the statistics target abstract graph-theoretical properties, others target |
|
properties specific to Enhanced Universal Dependencies. The script also tries to collect |
|
clues about individual enhancement types defined in UD (thus assessing whether the |
|
enhancement is annotated in the given dataset). |
|
|
|
perl enhanced_graph_properties.pl file.conllu |
|
|
|
|
|
SYSTEM OUTPUTS |
|
-------------- |
|
|
|
The folder "sysoutputs" contains the official primary submission of each team participating |
|
in the shared task. The folder names are the lowercased team names as used at submission |
|
time; they slightly differ from the spelling used in the shared task overview paper: |
|
|
|
TurkuNLP ....... turkunlp |
|
Orange ......... orange_deskin |
|
Emory NLP ...... emorynlp |
|
FASTPARSE ...... fastparse |
|
UNIPI .......... unipi |
|
ShanghaiTech ... shanghaitech_alibaba |
|
CLASP .......... clasp |
|
ADAPT .......... adapt |
|
Køpsala ........ koebsala |
|
RobertNLP ...... robertnlp |
|
|
|
In addition, there are three baseline submissions generated by the shared task organizers |
|
and described on the shared task website and in the overview paper: |
|
|
|
baseline1 ... gold basic trees copied as enhanced graphs |
|
baseline2 ... UDPipe-predicted basic trees copied as enhanced graphs |
|
baseline3 ... UDPipe predicted basic trees, then Stanford Enhancer created enhanced graphs from them |
|
|
|
|
|
REFERENCES |
|
---------- |
|
|
|
Gosse Bouma, Djamé Seddah, Daniel Zeman (2020). |
|
Overview of the IWPT 2020 Shared Task on Parsing into Enhanced Universal Dependencies. |
|
In Proceedings of the 16th International Conference on Parsing Technologies and the IWPT 2020 |
|
Shared Task on Parsing into Enhanced Universal Dependencies, Seattle, WA, USA, |
|
ISBN 978-1-952148-11-8 |
|
|
|
Marie Candito, Bruno Guillaume, Guy Perrier, Djamé Seddah (2017). |
|
Enhanced UD Dependencies with Neutralized Diathesis Alternation. |
|
In Proceedings of the Fourth International Conference on Dependency Linguistics (Depling 2017), |
|
pages 42–53, Pisa, Italy, September 18–20 2017 |
|
|