Datasets:
rcds
/

Size Categories:
1K<n<10K
ArXiv:
Tags:
legal
License:
MultiLegalNeg / MultiLegalNeg.py
ramonachristen's picture
Update MultiLegalNeg.py
7dee5d6
import json
import datasets
import pandas as pd
from huggingface_hub.file_download import hf_hub_url
from collections import OrderedDict
try:
import lzma as xz
except ImportError:
import pylzma as xz
datasets.logging.set_verbosity_info()
logger = datasets.logging.get_logger(__name__)
_DESCRIPTION ="""\
"""
_HOMEPAGE = ""
_LICENSE = ""
_CITATION = ""
_URL = {
'data/'
}
_LANGUAGES = [
"de", "fr", "it", "swiss", "en"
]
_SUBSETS = [
"_sherlock", "_sfu", "_bioscope", "_dalloux", ""
]
_BUILDS = ['de', 'fr', 'it', 'swiss', 'fr_dalloux', 'fr_all', 'en_bioscope', 'en_sherlock', 'en_sfu', 'en_all', 'all_all']
class MultiLegalNegConfig(datasets.BuilderConfig):
def __init__(self, name:str, **kwargs):
super( MultiLegalNegConfig, self).__init__(**kwargs)
self.name = name
self.language = name.split("_")[0]
self.subset = f'_{name.split("_")[1]}' if len(name.split("_"))==2 else ""
class MultiLegalNeg(datasets.GeneratorBasedBuilder):
BUILDER_CONFIG_CLASS = MultiLegalNegConfig
BUILDER_CONFIGS = [
MultiLegalNegConfig(f"{build}") for build in _BUILDS
]
def _info(self):
features = datasets.Features(
{
"text": datasets.Value("string"),
"spans": [
{
"start": datasets.Value("int64"),
"end": datasets.Value("int64"),
"token_start": datasets.Value("int64"),
"token_end": datasets.Value("int64"),
"label": datasets.Value("string")
}
],
"tokens": [
{
"text": datasets.Value("string"),
"start": datasets.Value("int64"),
"end": datasets.Value("int64"),
"id": datasets.Value("int64"),
"ws": datasets.Value("bool")
}
]
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features = features,
homepage = _HOMEPAGE,
citation=_CITATION
)
def _split_generators(self, dl_manager):
languages = _LANGUAGES if self.config.language == "all" else [self.config.language]
subsets = _SUBSETS if self.config.subset == "_all" else [self.config.subset]
split_generators = []
for split in [datasets.Split.TRAIN, datasets.Split.TEST, datasets.Split.VALIDATION]:
filepaths = []
for language in languages:
for subset in subsets:
try:
filepaths.append(dl_manager.download((f'data/{split}/{language}{subset}_{split}.jsonl.xz')))
except:
break
split_generators.append(datasets.SplitGenerator(name=split, gen_kwargs={'filepaths': filepaths}))
return split_generators
def _generate_examples(self, filepaths):
id_ = 0
for filepath in filepaths:
if filepath:
logger.info("Generating examples from = %s", filepath)
try:
with xz.open(open(filepath,'rb'), 'rt', encoding='utf-8') as f:
json_list = list(f)
for json_str in json_list:
example = json.loads(json_str)
if example is not None and isinstance(example, dict):
yield id_, example
id_ +=1
except Exception:
logger.exception("Error while processing file %s", filepath)