Datasets:

Language Creators:
found
Annotations Creators:
found
machine-generated
Source Datasets:
extended
ArXiv:
License:
fairlex / fairlex.py
kiddothe2b's picture
initial commit
63943fb
# coding=utf-8
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Fairlex: A multilingual benchmark for evaluating fairness in legal text processing."""
import json
import os
import textwrap
import datasets
MAIN_CITATION = """\
@inproceedings{chalkidis-etal-2022-fairlex,
author={Chalkidis, Ilias and Passini, Tommaso and Zhang, Sheng and
Tomada, Letizia and Schwemer, Sebastian Felix and Søgaard, Anders},
title={FairLex: A Multilingual Benchmark for Evaluating Fairness in Legal Text Processing},
booktitle={Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics},
year={2022},
address={Dublin, Ireland}
}
"""
_DESCRIPTION = """\
Fairlex: A multilingual benchmark for evaluating fairness in legal text processing.
"""
ECTHR_ARTICLES = ["2", "3", "5", "6", "8", "9", "10", "11", "14", "P1-1"]
SCDB_ISSUE_AREAS = [
"Criminal Procedure",
"Civil Rights",
"First Amendment",
"Due Process",
"Privacy",
"Attorneys",
"Unions",
"Economic Activity",
"Judicial Power",
"Federalism",
"Federal Taxation",
]
FSCS_LABELS = ["dismissal", "approval"]
CAIL_LABELS = ["0", "<=12", "<=36", "<=60", "<=120", ">120"]
class FairlexConfig(datasets.BuilderConfig):
"""BuilderConfig for Fairlex."""
def __init__(
self,
label_column,
url,
data_url,
citation,
label_classes=None,
multi_label=None,
attributes=None,
**kwargs,
):
"""BuilderConfig for Fairlex.
Args:
label_column: `string`, name of the column in the jsonl file corresponding
to the label
url: `string`, url for the original project
data_url: `string`, url to download the zip file from
data_file: `string`, filename for data set
citation: `string`, citation for the data set
url: `string`, url for information about the data set
label_classes: `list[string]`, the list of classes if the label is
categorical. If not provided, then the label will be of type
`datasets.Value('float32')`.
multi_label: `boolean`, True if the task is multi-label
attributes: `List<string>`, names of the protected attributes
**kwargs: keyword arguments forwarded to super.
"""
super(FairlexConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs)
self.label_column = label_column
self.label_classes = label_classes
self.multi_label = multi_label
self.attributes = attributes
self.url = url
self.data_url = data_url
self.citation = citation
class Fairlex(datasets.GeneratorBasedBuilder):
"""Fairlex: A multilingual benchmark for evaluating fairness in legal text processing. Version 1.0"""
BUILDER_CONFIGS = [
FairlexConfig(
name="ecthr",
description=textwrap.dedent(
"""\
The European Court of Human Rights (ECtHR) hears allegations that a state has breached human rights
provisions of the European Convention of Human Rights (ECHR). We use the dataset of Chalkidis et al.
(2021), which contains 11K cases from ECtHR's public database. Each case is mapped to articles of the ECHR
that were violated (if any). This is a multi-label text classification task. Given the facts of a case,
the goal is to predict the ECHR articles that were violated, if any, as decided (ruled) by the court."""
),
label_column="labels",
label_classes=ECTHR_ARTICLES,
multi_label=True,
attributes=[
("applicant_age", ["n/a", "<=35", "<=65", ">65"]),
("applicant_gender", ["n/a", "male", "female"]),
("defendant_state", ["C.E. European", "Rest of Europe"]),
],
data_url="https://zenodo.org/record/6322643/files/ecthr.zip",
url="https://huggingface.co/datasets/ecthr_cases",
citation=textwrap.dedent(
"""\
@inproceedings{chalkidis-etal-2021-paragraph,
title = "Paragraph-level Rationale Extraction through Regularization: A case study on {E}uropean Court of Human Rights Cases",
author = "Chalkidis, Ilias and
Fergadiotis, Manos and
Tsarapatsanis, Dimitrios and
Aletras, Nikolaos and
Androutsopoulos, Ion and
Malakasiotis, Prodromos",
booktitle = "Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
month = jun,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.naacl-main.22",
doi = "10.18653/v1/2021.naacl-main.22",
pages = "226--241",
}
}"""
),
),
FairlexConfig(
name="scotus",
description=textwrap.dedent(
"""\
The US Supreme Court (SCOTUS) is the highest federal court in the United States of America and generally
hears only the most controversial or otherwise complex cases which have not been sufficiently well solved
by lower courts. We combine information from SCOTUS opinions with the Supreme Court DataBase (SCDB)
(Spaeth, 2020). SCDB provides metadata (e.g., date of publication, decisions, issues, decision directions
and many more) for all cases. We consider the available 14 thematic issue areas (e.g, Criminal Procedure,
Civil Rights, Economic Activity, etc.). This is a single-label multi-class document classification task.
Given the court opinion, the goal is to predict the issue area whose focus is on the subject matter
of the controversy (dispute). """
),
label_column="label",
label_classes=SCDB_ISSUE_AREAS,
multi_label=False,
attributes=[
("decision_direction", ["conservative", "liberal"]),
("respondent_type", ["other", "person", "organization", "public entity", "facility"]),
],
url="http://scdb.wustl.edu/data.php",
data_url="https://zenodo.org/record/6322643/files/scotus.zip",
citation=textwrap.dedent(
"""\
@misc{spaeth2020,
author = {Harold J. Spaeth and Lee Epstein and Andrew D. Martin, Jeffrey A. Segal
and Theodore J. Ruger and Sara C. Benesh},
year = {2020},
title ={{Supreme Court Database, Version 2020 Release 01}},
url= {http://Supremecourtdatabase.org},
howpublished={Washington University Law}
}"""
),
),
FairlexConfig(
name="fscs",
description=textwrap.dedent(
"""\
The Federal Supreme Court of Switzerland (FSCS) is the last level of appeal in Switzerland and similarly
to SCOTUS, the court generally hears only the most controversial or otherwise complex cases which have
not been sufficiently well solved by lower courts. The court often focus only on small parts of previous
decision, where they discuss possible wrong reasoning by the lower court. The Swiss-Judgment-Predict
dataset (Niklaus et al., 2021) contains more than 85K decisions from the FSCS written in one of three
languages (50K German, 31K French, 4K Italian) from the years 2000 to 2020. The dataset is not parallel,
i.e., all cases are unique and decisions are written only in a single language. The dataset provides labels
for a simplified binary (approval, dismissal) classification task. Given the facts of the case, the goal
is to predict if the plaintiff's request is valid or partially valid."""
),
label_column="label",
label_classes=FSCS_LABELS,
multi_label=False,
attributes=[
("decision_language", ["de", "fr", "it"]),
("legal_area", ["other", "public law", "penal law", "civil law", "social law", "insurance law"]),
(
"court_region",
[
"n/a",
"Région lémanique",
"Zürich",
"Espace Mittelland",
"Northwestern Switzerland",
"Eastern Switzerland",
"Central Switzerland",
"Ticino",
"Federation",
],
),
],
url="https://github.com/JoelNiklaus/SwissCourtRulingCorpus",
data_url="https://zenodo.org/record/6322643/files/fscs.zip",
citation=textwrap.dedent(
"""\
@InProceedings{niklaus-etal-2021-swiss,
author = {Niklaus, Joel
and Chalkidis, Ilias
and Stürmer, Matthias},
title = {Swiss-Court-Predict: A Multilingual Legal Judgment Prediction Benchmark},
booktitle = {Proceedings of the 2021 Natural Legal Language Processing Workshop},
year = {2021},
location = {Punta Cana, Dominican Republic},
}"""
),
),
FairlexConfig(
name="cail",
description=textwrap.dedent(
"""\
The Supreme People's Court of China (CAIL) is the last level of appeal in China and considers cases that
originated from the high people's courts concerning matters of national importance. The Chinese AI and Law
challenge (CAIL) dataset (Xiao et al., 2018) is a Chinese legal NLP dataset for judgment prediction and
contains over 1m criminal cases. The dataset provides labels for relevant article of criminal code
prediction, charge (type of crime) prediction, imprisonment term (period) prediction, and monetary penalty
prediction. The updated (soft) version of the CAIL dataset has 104K criminal court cases. The tasks is
crime severity prediction task, a multi-class classification task, where given the facts of a case,
the goal is to predict how severe was the committed crime with respect to the imprisonment term.
We approximate crime severity by the length of imprisonment term, split in 6 clusters
(0, >=12, >=36, >=60, >=120, >120 months)."""
),
label_column="label",
label_classes=CAIL_LABELS,
multi_label=False,
attributes=[
("defendant_gender", ["male", "female"]),
("court_region", ["Beijing", "Liaoning", "Hunan", "Guangdong", "Sichuan", "Guangxi", "Zhejiang"]),
],
url="https://github.com/thunlp/LegalPLMs",
data_url="https://zenodo.org/record/6322643/files/cail.zip",
citation=textwrap.dedent(
"""\
@article{wang-etal-2021-equality,
title={Equality before the Law: Legal Judgment Consistency Analysis for Fairness},
author={Yuzhong Wang and Chaojun Xiao and Shirong Ma and Haoxi Zhong and Cunchao Tu and Tianyang Zhang and Zhiyuan Liu and Maosong Sun},
year={2021},
journal={Science China - Information Sciences},
url={https://arxiv.org/abs/2103.13868}
}"""
),
),
]
def _info(self):
features = {"text": datasets.Value("string")}
if self.config.multi_label:
features["labels"] = datasets.features.Sequence(datasets.ClassLabel(names=self.config.label_classes))
else:
features["label"] = datasets.ClassLabel(names=self.config.label_classes)
for attribute_name, attribute_groups in self.config.attributes:
features[attribute_name] = datasets.ClassLabel(names=attribute_groups)
return datasets.DatasetInfo(
description=self.config.description,
features=datasets.Features(features),
homepage=self.config.url,
citation=self.config.citation + "\n" + MAIN_CITATION,
)
def _split_generators(self, dl_manager):
data_dir = dl_manager.download_and_extract(self.config.data_url)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
# These kwargs will be passed to _generate_examples
gen_kwargs={
"filepath": os.path.join(data_dir, "train.jsonl"),
"split": "train",
},
),
datasets.SplitGenerator(
name=datasets.Split.TEST,
# These kwargs will be passed to _generate_examples
gen_kwargs={
"filepath": os.path.join(data_dir, "test.jsonl"),
"split": "test",
},
),
datasets.SplitGenerator(
name=datasets.Split.VALIDATION,
# These kwargs will be passed to _generate_examples
gen_kwargs={
"filepath": os.path.join(data_dir, "val.jsonl"),
"split": "val",
},
),
]
def _generate_examples(self, filepath, split):
"""This function returns the examples in the raw (text) form."""
with open(filepath, encoding="utf-8") as f:
for id_, row in enumerate(f):
data = json.loads(row)
example = {
"text": data["text"],
self.config.label_column: data[self.config.label_column],
}
for attribute_name, _ in self.config.attributes:
example[attribute_name] = data["attributes"][attribute_name]
yield id_, example