Datasets:
Languages:
Chinese
Multilinguality:
zh
Size Categories:
100K<n<20M
Language Creators:
shibing624
Annotations Creators:
shibing624
Source Datasets:
https://www.biendata.xyz/competition/sohu_2021/data/
Tags:
License:
# -*- coding: utf-8 -*- | |
""" | |
@author:XuMing(xuming624@qq.com) | |
@description: | |
2021搜狐校园文本匹配算法大赛数据集 | |
upload: https://github.com/shibing624 | |
""" | |
import csv | |
import os | |
import json | |
import datasets | |
_CITATION = """https://github.com/shibing624/text2vec""" | |
_DESCRIPTION = """\ | |
2021搜狐校园文本匹配算法大赛数据集 | |
""" | |
_DATA_URL = "https://huggingface.co/datasets/shibing624/sts-sohu2021/resolve/main" | |
class Sohu(datasets.GeneratorBasedBuilder): | |
"""The Chinese Natural Language Inference (sts-sohu) Corpus.""" | |
BUILDER_CONFIGS = [ | |
datasets.BuilderConfig( | |
name="dda", | |
version=datasets.Version("1.0.0", ""), | |
description="Plain text import of sts-sohu2021", | |
), | |
datasets.BuilderConfig( | |
name="ddb", | |
version=datasets.Version("1.0.0", ""), | |
description="Plain text import of sts-sohu2021", | |
), | |
datasets.BuilderConfig( | |
name="dca", | |
version=datasets.Version("1.0.0", ""), | |
description="Plain text import of sts-sohu2021", | |
), | |
datasets.BuilderConfig( | |
name="dcb", | |
version=datasets.Version("1.0.0", ""), | |
description="Plain text import of sts-sohu2021", | |
), | |
datasets.BuilderConfig( | |
name="cca", | |
version=datasets.Version("1.0.0", ""), | |
description="Plain text import of sts-sohu2021", | |
), | |
datasets.BuilderConfig( | |
name="ccb", | |
version=datasets.Version("1.0.0", ""), | |
description="Plain text import of sts-sohu2021", | |
), | |
] | |
def _info(self): | |
return datasets.DatasetInfo( | |
description=_DESCRIPTION, | |
features=datasets.Features( | |
{ | |
"sentence1": datasets.Value("string"), | |
"sentence2": datasets.Value("string"), | |
"label": datasets.Value("int32"), | |
} | |
), | |
supervised_keys=None, | |
homepage="https://github.com/shibing624/text2vec", | |
citation=_CITATION, | |
) | |
def _split_generators(self, dl_manager): | |
# https://huggingface.co/datasets/shibing624/sts-sohu2021/resolve/main/dda.jsonl | |
return [ | |
datasets.SplitGenerator( | |
name=datasets.Split.TRAIN, | |
gen_kwargs={ | |
"filepath": dl_manager.download_and_extract(f"{_DATA_URL}/{self.config.name}-train.jsonl") | |
} | |
), | |
datasets.SplitGenerator( | |
name=datasets.Split.TEST, | |
gen_kwargs={ | |
"filepath": dl_manager.download_and_extract(f"{_DATA_URL}/{self.config.name}-test.jsonl") | |
} | |
), | |
] | |
def _generate_examples(self, filepath): | |
"""This function returns the examples in the raw (jsonl) form.""" | |
id = 0 | |
if isinstance(filepath, str): | |
filepath = [filepath] | |
for file in filepath: | |
with open(file, encoding="utf-8") as f: | |
for key, row in enumerate(f): | |
if row and len(row) > 1: | |
data = json.loads(row) | |
yield id, { | |
"sentence1": data["sentence1"], | |
"sentence2": data["sentence2"], | |
"label": int(data["label"]) | |
} | |
id += 1 |