Datasets:
Update files from the datasets library (from 1.3.0)
Browse filesRelease notes: https://github.com/huggingface/datasets/releases/tag/1.3.0
README.md
CHANGED
@@ -44,6 +44,7 @@ task_ids:
|
|
44 |
- [Dataset Curators](#dataset-curators)
|
45 |
- [Licensing Information](#licensing-information)
|
46 |
- [Citation Information](#citation-information)
|
|
|
47 |
|
48 |
## Dataset Description
|
49 |
|
@@ -155,3 +156,7 @@ The LOCNESS corpus (Granger, 1998) consists of essays written by native English
|
|
155 |
abstract = "This paper reports on the BEA-2019 Shared Task on Grammatical Error Correction (GEC). As with the CoNLL-2014 shared task, participants are required to correct all types of errors in test data. One of the main contributions of the BEA-2019 shared task is the introduction of a new dataset, the Write{\&}Improve+LOCNESS corpus, which represents a wider range of native and learner English levels and abilities. Another contribution is the introduction of tracks, which control the amount of annotated data available to participants. Systems are evaluated in terms of ERRANT F{\_}0.5, which allows us to report a much wider range of performance statistics. The competition was hosted on Codalab and remains open for further submissions on the blind test set.",
|
156 |
}
|
157 |
```
|
|
|
|
|
|
|
|
|
|
44 |
- [Dataset Curators](#dataset-curators)
|
45 |
- [Licensing Information](#licensing-information)
|
46 |
- [Citation Information](#citation-information)
|
47 |
+
- [Contributions](#contributions)
|
48 |
|
49 |
## Dataset Description
|
50 |
|
|
|
156 |
abstract = "This paper reports on the BEA-2019 Shared Task on Grammatical Error Correction (GEC). As with the CoNLL-2014 shared task, participants are required to correct all types of errors in test data. One of the main contributions of the BEA-2019 shared task is the introduction of a new dataset, the Write{\&}Improve+LOCNESS corpus, which represents a wider range of native and learner English levels and abilities. Another contribution is the introduction of tracks, which control the amount of annotated data available to participants. Systems are evaluated in terms of ERRANT F{\_}0.5, which allows us to report a much wider range of performance statistics. The competition was hosted on Codalab and remains open for further submissions on the blind test set.",
|
157 |
}
|
158 |
```
|
159 |
+
|
160 |
+
### Contributions
|
161 |
+
|
162 |
+
Thanks to [@aseifert](https://github.com/aseifert) for adding this dataset.
|