Datasets:

The viewer is disabled because this dataset repo requires arbitrary Python code execution. Please consider removing the loading script and relying on automated data support (you can use convert_to_parquet from the datasets library). If this is not possible, please open a discussion for direct help.

YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/datasets-cards)

ParaDocs

Data availability limits the scope of any given task. In machine translation, historical models were incapable of handling longer contexts, so the lack of document-level datasets was less noticeable. Now, despite the emergence of long-sequence methods, we remain within a sentence-level paradigm and without data to adequately approach context-aware machine translation. Most large-scale datasets have been processed through a pipeline that discards document-level metadata.

ParaDocs is a publicly available dataset that produces parallel annotations for the document-level metadata of three large publicly available corpora (ParaCrawl, Europal, and News Commentary) in many languages. Using this data and the following scripts, you can download parallel document contexts for the purpose of training context-aware machine translation systems.

If you have questions about this data or use of the following scripts, please do not hesitate to contact the maintainer at rewicks@jhu.edu.

Quick Start

The scripts to download and process the data can be found here:

Clone these scripts:

git clone https://github.com/rewicks/ParaDocs.git

From this directory, you can stream a specific language and split from huggingface with:

paradocs/paradocs-hf --name en-de-strict --minimum_size 2 --frequency_cutoff 100 --lid_cutoff 0.5

It may alternatively be faster to download the *.gz files of your desired split and then pipe them through the paradocs/paradocs file for filtering.

zcat data/en-de/strict* | paradocs/paradocs --minimum_size 2 --frequency_cutoff 100 --lid_cutoff 0.5

The filtering commandline arguments are explained in more detial in Section 3.2.

The Paper

If you use this dataset in your research. Please cite our paper.


license: apache-2.0

Downloads last month
226