Datasets:
Sub-tasks:
named-entity-recognition
Multilinguality:
multilingual
Size Categories:
unknown
Language Creators:
machine-generated
Annotations Creators:
expert-generated
Source Datasets:
original
ArXiv:
Tags:
word-segmentation
License:
"""HashSet dataset.""" | |
import datasets | |
import pandas as pd | |
import json | |
_CITATION = """ | |
@article{kodali2022hashset, | |
title={HashSet--A Dataset For Hashtag Segmentation}, | |
author={Kodali, Prashant and Bhatnagar, Akshala and Ahuja, Naman and Shrivastava, Manish and Kumaraguru, Ponnurangam}, | |
journal={arXiv preprint arXiv:2201.06741}, | |
year={2022} | |
} | |
""" | |
_DESCRIPTION = """ | |
Hashset is a new dataset consisiting on 1.9k manually annotated and 3.3M loosely supervised tweets for testing the | |
efficiency of hashtag segmentation models. We compare State of The Art Hashtag Segmentation models on Hashset and other | |
baseline datasets (STAN and BOUN). We compare and analyse the results across the datasets to argue that HashSet can act | |
as a good benchmark for hashtag segmentation tasks. | |
HashSet Manual: contains 1.9k manually annotated hashtags. Each row consists of the hashtag, segmented | |
hashtag ,named entity annotations, a list storing whether the hashtag contains mix of hindi and english | |
tokens and/or contains non-english tokens. | |
""" | |
_URL = "https://raw.githubusercontent.com/prashantkodali/HashSet/master/datasets/hashset/HashSet-Manual.csv" | |
class HashSetManual(datasets.GeneratorBasedBuilder): | |
VERSION = datasets.Version("1.0.0") | |
def _info(self): | |
return datasets.DatasetInfo( | |
description=_DESCRIPTION, | |
features=datasets.Features( | |
{ | |
"index": datasets.Value("int32"), | |
"hashtag": datasets.Value("string"), | |
"segmentation": datasets.Value("string"), | |
"spans": datasets.Sequence( | |
{ | |
"start": datasets.Value("int32"), | |
"end": datasets.Value("int32"), | |
"text": datasets.Value("string") | |
} | |
), | |
"source": datasets.Value("string"), | |
"gold_position": datasets.Value("int32"), | |
"mix": datasets.Value("bool"), | |
"other": datasets.Value("bool"), | |
"ner": datasets.Value("bool"), | |
"annotator_id": datasets.Value("int32"), | |
"annotation_id": datasets.Value("int32"), | |
"created_at": datasets.Value("timestamp[us]"), | |
"updated_at": datasets.Value("timestamp[us]"), | |
"lead_time": datasets.Value("float64"), | |
"rank": datasets.Sequence( | |
{ | |
"position": datasets.Value("int32"), | |
"candidate": datasets.Value("string") | |
} | |
) | |
} | |
), | |
supervised_keys=None, | |
homepage="https://github.com/prashantkodali/HashSet/", | |
citation=_CITATION, | |
) | |
def _split_generators(self, dl_manager): | |
downloaded_files = dl_manager.download(_URL) | |
return [ | |
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": downloaded_files }), | |
] | |
def _generate_examples(self, filepath): | |
def read_language_labels(field): | |
mix_label = "Hashtag has a mix of english and hindi tokens" | |
other_label = "Hashtag has non english token " | |
ner_label = "Hashtag has named entities" | |
try: | |
record = json.loads(field) | |
except json.decoder.JSONDecodeError: | |
record = {"choices": [field]} | |
mix = False | |
other = False | |
ner = False | |
if mix_label in record["choices"]: | |
mix = True | |
if other_label in record["choices"]: | |
other = True | |
if ner_label in record["choices"]: | |
ner = True | |
return mix, other, ner | |
def read_entities(field): | |
try: | |
record = json.loads(field) | |
except json.decoder.JSONDecodeError: | |
return [] | |
output = [] | |
for row in record: | |
output.append({ | |
"start": row.get("start", None), | |
"end": row.get("end", None), | |
"text": row.get("text", None) | |
}) | |
return output | |
def read_rank(row): | |
output = [] | |
for i in range(10): | |
output.append({ | |
"position": str(i+1), | |
"candidate": row[str(i+1)] | |
}) | |
return output | |
def get_gold_position(field): | |
output = field.strip("$") | |
try: | |
return int(output) | |
except ValueError: | |
return None | |
records = pd.read_csv(filepath).to_dict("records") | |
for idx, row in enumerate(records): | |
mix, other, ner = read_language_labels(row["mutlitoken"]) | |
yield idx, { | |
"index": row["Unnamed: 0"], | |
"hashtag": row["Hashtag"], | |
"segmentation": row["Final Segmentation"], | |
"spans": read_entities(row["charner"]), | |
"source": row["Source"], | |
"gold_position": get_gold_position(row["topk"]), | |
"mix": mix, | |
"other": other, | |
"ner": ner, | |
"annotator_id": int(row["annotator"]), | |
"annotation_id": int(row["annotation_id"]), | |
"created_at": row["created_at"], | |
"updated_at": row["updated_at"], | |
"lead_time": row["lead_time"], | |
"rank": read_rank(row) | |
} | |