File size: 12,953 Bytes
99384c1 d31b0d2 99384c1 776122e 99384c1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 |
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""FactChecksbr dataset"""
import textwrap
import csv
import os
import datasets
_CITATION = """\
@misc{FactChecksbr,
author = {R. S. Gomes, Juliana},
title = {FactChecks.br},
url = {https://github.com/fake-news-UFG/FactChecks.br},
doi = { 10.57967/hf/1016 },
}
"""
_DESCRIPTION = """\
Collection of Portuguese Fact-Checking Benchmarks.
"""
_HOMEPAGE = "https://github.com/fake-news-UFG/FactChecks.br"
_LICENSE = "https://raw.githubusercontent.com/fake-news-UFG/FactChecks.br/main/LICENSE"
_URL = "https://github.com/fake-news-UFG/FactChecks.br/releases/download/v0.1/FactChecksbr.zip"
# from https://huggingface.co/datasets/glue/blob/main/glue.py
class GlueConfig(datasets.BuilderConfig):
"""BuilderConfig for GLUE."""
def __init__(
self,
citation,
**kwargs,
):
super(GlueConfig, self).__init__(
version=datasets.Version("0.1.0", ""), **kwargs
)
self.citation = citation
class FactChecksbr(datasets.GeneratorBasedBuilder):
"""Collection of Portuguese Fact-Checking Benchmarks."""
VERSION = datasets.Version("0.1.0")
BUILDER_CONFIGS = [
GlueConfig(
name="fact_check_tweet_pt",
description="",
citation=textwrap.dedent(
"""\
@misc{kazemi2022matching,
title={Matching Tweets With Applicable Fact-Checks Across Languages},
author={Ashkan Kazemi and Zehua Li and Verónica Pérez-Rosas and Scott A. Hale and Rada Mihalcea},
year={2022},
eprint={2202.07094},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
"""
),
),
GlueConfig(
name="central_de_fatos",
description=textwrap.dedent(
"""\
In recent times, the interest for research dissecting the dissemination
and prevention of misinformation in the online environment has spiked dramatically.
Given that scenario, a recurring obstacle is the unavailability of public datasets
containing fact-checked instances."""
),
citation=textwrap.dedent(
"""\
@inproceedings{dsw,
author = {João Couto and Breno Pimenta and Igor M. de Araújo and Samuel Assis and Julio C. S. Reis and Ana Paula da Silva and Jussara Almeida and Fabrício Benevenuto},
title = {Central de Fatos: Um Repositório de Checagens de Fatos},
booktitle = {Anais do III Dataset Showcase Workshop},
location = {Rio de Janeiro},
year = {2021},
keywords = {},
issn = {0000-0000},
pages = {128--137},
publisher = {SBC},
address = {Porto Alegre, RS, Brasil},
doi = {10.5753/dsw.2021.17421},
url = {https://sol.sbc.org.br/index.php/dsw/article/view/17421}
}
"""
),
),
GlueConfig(
name="FakeNewsSet",
description="",
citation=textwrap.dedent(
"""\
@inproceedings{10.1145/3428658.3430965,
author = {da Silva, Fl\'{a}vio Roberto Matias and Freire,
Paulo M\'{a}rcio Souza and de Souza, Marcelo Pereira and de A. B. Plenamente,
Gustavo and Goldschmidt, Ronaldo Ribeiro},
title = {FakeNewsSetGen: A Process to Build Datasets That
Support Comparison Among Fake News Detection Methods},
year = {2020},
isbn = {9781450381963},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3428658.3430965},
doi = {10.1145/3428658.3430965},
abstract = {Due to easy access and low cost, social media online
news consumption has increased significantly for the last decade.
Despite their benefits, some social media allow anyone to post news
with intense spreading power, which amplifies an old problem: the dissemination of Fake News.
In the face of this scenario, several machine learning-based methods
to automatically detect Fake News (MLFN) have been proposed.
All of them require datasets to train and evaluate their detection models.
Although recent MLFN were designed to consider data regarding the news propagation
on social media, most of the few available datasets do not contain this kind of data.
Hence, comparing the performances amid those recent MLFN and the others is restricted
to a very limited number of datasets. Moreover, all existing datasets with propagation
data do not contain news in Portuguese, which impairs the evaluation of the MLFN in this language.
Thus, this work proposes FakeNewsSetGen, a process that builds Fake News datasets that contain news
propagation data and support comparison amid the state-of-the-art MLFN. FakeNewsSetGen's software
engineering process was guided to include all kind of data required by the existing MLFN. In order
to illustrate FakeNewsSetGen's viability and adequacy, a case study was carried out. It encompassed
the implementation of a FakeNewsSetGen prototype and the application of this prototype to create a
dataset called FakeNewsSet, with news in Portuguese. Five MLFN with different kind of data requirements
(two of them demanding news propagation data) were applied to FakeNewsSet and compared, demonstrating the
potential use of both the proposed process and the created dataset.},
booktitle = {Proceedings of the Brazilian Symposium on Multimedia and the Web},
pages = {241–248},
numpages = {8},
keywords = {Fake News detection, Dataset building process, social media},
location = {S\~{a}o Lu\'{\i}s, Brazil},
series = {WebMedia '20}
}
"""
),
),
GlueConfig(
name="FakeRecogna",
description=textwrap.dedent(
"""\
FakeRecogna is a dataset comprised of real and fake news.
The real news is not directly linked to fake news and vice-versa,
which could lead to a biased classification.
The news collection was performed by crawlers developed for mining
pages of well-known and of great national importance agency news."""
),
citation=textwrap.dedent(
"""\
@inproceedings{10.1007/978-3-030-98305-5_6,
author = {Garcia, Gabriel L. and Afonso, Luis C. S. and Papa, Jo\~{a}o P.},
title = {FakeRecogna: A New Brazilian Corpus for Fake News Detection},
year = {2022},
isbn = {978-3-030-98304-8},
publisher = {Springer-Verlag},
address = {Berlin, Heidelberg},
url = {https://doi.org/10.1007/978-3-030-98305-5_6},
doi = {10.1007/978-3-030-98305-5_6},
abstract = {Fake news has become a research topic of great importance in Natural
Language Processing due to its negative impact on our society. Although its pertinence,
there are few datasets available in Brazilian Portuguese and mostly comprise few samples.
Therefore, this paper proposes creating a new fake news dataset named FakeRecogna that
contains a greater number of samples, more up-to-date news, and covering a few of the
most important categories. We perform a toy evaluation over the created dataset using traditional
classifiers such as Naive Bayes, Optimum-Path Forest, and Support Vector Machines.
A Convolutional Neural Network is also evaluated in the context of fake news detection
in the proposed dataset.},
booktitle = {Computational Processing of the Portuguese Language: 15th International Conference, PROPOR 2022, Fortaleza, Brazil, March 21–23, 2022, Proceedings},
pages = {57–67},
numpages = {11},
keywords = {Fake news, Corpus, Portuguese},
location = {Fortaleza, Brazil}
}
"""
),
),
GlueConfig(
name="fakebr",
description="Fake.Br Corpus is composed of aligned true and fake news written in Brazilian Portuguese.",
citation=textwrap.dedent(
"""\
@article{silva:20,
title = "Towards automatically filtering fake news in Portuguese",
journal = "Expert Systems with Applications",
volume = "146",
pages = "113199",
year = "2020",
issn = "0957-4174",
doi = "https://doi.org/10.1016/j.eswa.2020.113199",
url = "http://www.sciencedirect.com/science/article/pii/S0957417420300257",
author = "Renato M. Silva and Roney L.S. Santos and Tiago A. Almeida and Thiago A.S. Pardo",
}
"""
),
),
]
DEFAULT_CONFIG_NAME = "fakebr"
def _info(self):
if self.config.name == "fakebr":
features = datasets.Features(
{
"claim_text": datasets.Value("string"),
"claim_author": datasets.Value("string"),
"claim_url": datasets.Value("string"),
"claim_date": datasets.Value("string"),
"category": datasets.Value("string"),
"is_fake": datasets.ClassLabel(num_classes=3, names=[1, 0, -1]),
}
)
elif self.config.name in ["central_de_fatos", "FakeRecogna"]:
features = datasets.Features(
{
"review_id": datasets.Value("string"),
"review_text": datasets.Value("string"),
"review_author": datasets.Value("string"),
"review_url": datasets.Value("string"),
"review_domain": datasets.Value("string"),
"review_date": datasets.Value("string"),
"category": datasets.Value("string"),
"is_fake": datasets.ClassLabel(num_classes=3, names=[1, 0, -1]),
}
)
elif self.config.name in ["fact_check_tweet_pt", "FakeNewsSet"]:
features = datasets.Features(
{
"review_id": datasets.Value("string"),
"review_url": datasets.Value("string"),
"review_domain": datasets.Value("string"),
"claim_ids": datasets.Sequence(
feature=datasets.Value(dtype="string", id=None)
),
"is_fake": datasets.ClassLabel(num_classes=3, names=[1, 0, -1]),
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
homepage=_HOMEPAGE,
license=_LICENSE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
url = _URL
data_dir = dl_manager.download_and_extract(url)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={
"filepath": os.path.join(
data_dir, "data", f"{self.config.name}.tsv"
),
},
),
]
def _generate_examples(self, filepath):
with open(filepath, encoding="utf-8") as f:
reader = csv.reader(f, delimiter="\t")
names = next(reader)
for idx, row in enumerate(reader):
row = dict(zip(names, row))
if "claim_ids" in row.keys():
row["claim_ids"] = eval(row["claim_ids"])
yield idx, row
|