|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import json |
|
|
|
import datasets |
|
from dataclasses import dataclass |
|
|
|
_CITATION = ''' |
|
@article{mrtydi, |
|
title={{Mr. TyDi}: A Multi-lingual Benchmark for Dense Retrieval}, |
|
author={Xinyu Zhang and Xueguang Ma and Peng Shi and Jimmy Lin}, |
|
year={2021}, |
|
journal={arXiv:2108.08787}, |
|
} |
|
''' |
|
|
|
fields = [ |
|
'title', 'desc', 'desc_title' |
|
] |
|
|
|
_DESCRIPTION = 'dataset load script for Mr. TyDi' |
|
|
|
_DATASET_URLS = { |
|
field: { |
|
'test': f'https://huggingface.co/datasets/crystina-z/neuclir/resolve/main/data/topics.neuclir22.en.{field}.tsv', |
|
} for field in fields |
|
} |
|
|
|
|
|
class NeuCLIR(datasets.GeneratorBasedBuilder): |
|
BUILDER_CONFIGS = [datasets.BuilderConfig( |
|
version=datasets.Version('1.1.0'), |
|
name=field, description=f'NeuCLIR dataset in language {field}.' |
|
) for field in fields |
|
] |
|
|
|
def _info(self): |
|
features = datasets.Features({ |
|
'query_id': datasets.Value('string'), |
|
'query': datasets.Value('string'), |
|
|
|
'positive_passages': [{ |
|
'docid': datasets.Value('string'), |
|
'text': datasets.Value('string'), 'title': datasets.Value('string') |
|
}], |
|
'negative_passages': [{ |
|
'docid': datasets.Value('string'), |
|
'text': datasets.Value('string'), 'title': datasets.Value('string'), |
|
}], |
|
}) |
|
|
|
return datasets.DatasetInfo( |
|
|
|
description=_DESCRIPTION, |
|
|
|
features=features, |
|
supervised_keys=None, |
|
|
|
homepage='https://github.com/castorini/mr.tydi', |
|
|
|
license='', |
|
|
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
lang = self.config.name |
|
downloaded_files = dl_manager.download_and_extract(_DATASET_URLS[lang]) |
|
|
|
splits = [ |
|
datasets.SplitGenerator( |
|
name='test', |
|
gen_kwargs={ |
|
'filepath': downloaded_files['test'], |
|
}, |
|
), |
|
] |
|
return splits |
|
|
|
def _generate_examples(self, filepath): |
|
lang = self.config.name |
|
with open(filepath, encoding="utf-8") as f: |
|
for i, line in enumerate(f): |
|
qid, query = line.strip().split('\t') |
|
data = {'query_id': qid, 'query': query} |
|
|
|
for feature in ['negative_passages', 'positive_passages']: |
|
data[feature] = [] |
|
|
|
yield qid, data |
|
|