Datasets:
Languages:
Arabic
Multilinguality:
monolingual
Size Categories:
10K<n<100K
Language Creators:
found
Annotations Creators:
no-annotation
Source Datasets:
original
License:
# coding=utf-8 | |
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor. | |
# | |
# Licensed under the Apache License, Version 2.0 (the "License"); | |
# you may not use this file except in compliance with the License. | |
# You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
"""The dataset contains a set of 31,030 Arabic newspaper articles alongwith metadata, extracted from various online Saudi newspapers.""" | |
import json | |
import os | |
import datasets | |
_CITATION = """\ | |
@misc{hagrima2015, | |
author = "M. Alhagri", | |
title = "Saudi Newspapers Arabic Corpus (SaudiNewsNet)", | |
year = 2015, | |
url = "http://github.com/ParallelMazen/SaudiNewsNet" | |
} | |
""" | |
_DESCRIPTION = """The dataset contains a set of 31,030 Arabic newspaper articles alongwith metadata, \ | |
extracted from various online Saudi newspapers and written in MSA.""" | |
_HOMEPAGE = "https://github.com/parallelfold/SaudiNewsNet" | |
_LICENSE = "Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License." | |
_URLs = [ | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-07-21.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-07-22.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-07-23.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-07-24.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-07-25.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-07-26.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-07-27.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-07-31.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-01.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-02.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-03.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-04.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-06.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-07.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-08.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-09.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-10.zip", | |
"https://github.com/parallelfold/SaudiNewsNet/raw/master/dataset/2015-08-11.zip", | |
] | |
_dirs = [ | |
"2015-07-21.json", | |
"2015-07-22.json", | |
"2015-07-23.json", | |
"2015-07-24.json", | |
"2015-07-25.json", | |
"2015-07-26.json", | |
"2015-07-27.json", | |
"2015-07-31.json", | |
"2015-08-01.json", | |
"2015-08-02.json", | |
"2015-08-03.json", | |
"2015-08-04.json", | |
"2015-08-06.json", | |
"2015-08-07.json", | |
"2015-08-08.json", | |
"2015-08-09.json", | |
"2015-08-10.json", | |
"2015-08-11.json", | |
] | |
class Saudinewsnet(datasets.GeneratorBasedBuilder): | |
"""a set of 31,030 Arabic newspaper articles along with metadata""" | |
def _info(self): | |
return datasets.DatasetInfo( | |
description=_DESCRIPTION, | |
features=datasets.Features( | |
{ | |
"source": datasets.Value( | |
"string" | |
), # A string identifief of the newspaper from which the article was extracted. | |
"url": datasets.Value("string"), # The full URL from which the article was extracted. | |
"date_extracted": datasets.Value( | |
"string" | |
), # The timestamp of the date on which the article was extracted. | |
"title": datasets.Value("string"), # The title of the article. Can be empty. | |
"author": datasets.Value("string"), # The author of the article. Can be empty. | |
"content": datasets.Value("string"), # The content of the article. | |
} | |
), | |
homepage=_HOMEPAGE, | |
citation=_CITATION, | |
supervised_keys=None, | |
) | |
def _split_generators(self, dl_manager): | |
"""Returns SplitGenerators.""" | |
datadir = dl_manager.download_and_extract(_URLs) | |
pt = [] | |
for dd, d in zip(datadir, _dirs): | |
pt.append(os.path.join(dd, d)) | |
return [ | |
datasets.SplitGenerator( | |
name=datasets.Split.TRAIN, | |
gen_kwargs={"filepath": pt, "split": "train"}, | |
) | |
] | |
def _generate_examples(self, filepath, split): | |
"""Generates examples""" | |
for file_idx, path in enumerate(filepath): | |
with open(path, encoding="utf-8") as f: | |
articles = json.load(f) | |
for _id, article in enumerate(articles): | |
title = article.get("title", "") | |
source = article["source"] | |
dt = article["date_extracted"] | |
link = article["url"] | |
author = article.get("author", "").strip(" ") | |
content = article["content"].strip("/n") | |
yield f"{file_idx}_{_id}", { | |
"title": title, | |
"source": source, | |
"date_extracted": dt, | |
"url": link, | |
"author": author, | |
"content": content, | |
} | |