"""Binkley""" import datasets import pandas as pd from collections import deque _CITATION = """ @inproceedings{inproceedings, author = {Lawrie, Dawn and Binkley, David and Morrell, Christopher}, year = {2010}, month = {11}, pages = {3 - 12}, title = {Normalizing Source Code Vocabulary}, journal = {Proceedings - Working Conference on Reverse Engineering, WCRE}, doi = {10.1109/WCRE.2010.10} } """ _DESCRIPTION = """ In programming languages, identifiers are tokens (also called symbols) which name language entities. Some of the kinds of entities an identifier might denote include variables, types, labels, subroutines, and packages. Binkley is a dataset for identifier segmentation, i.e. the task of adding spaces between the words on a identifier. """ _URL = "https://raw.githubusercontent.com/ruanchaves/hashformers/master/datasets/binkley.csv" class Binkley(datasets.GeneratorBasedBuilder): VERSION = datasets.Version("1.0.0") def _info(self): return datasets.DatasetInfo( description=_DESCRIPTION, features=datasets.Features( { "index": datasets.Value("int32"), "identifier": datasets.Value("string"), "segmentation": datasets.Value("string") } ), supervised_keys=None, homepage="", citation=_CITATION, ) def _split_generators(self, dl_manager): downloaded_files = dl_manager.download(_URL) return [ datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": downloaded_files}), ] def _generate_examples(self, filepath): def get_segmentation(needle, haystack, sep="-"): output = haystack needle = needle.lower() haystack = haystack.lower() counter = 0 pos = deque() iterator = iter(haystack) for char in needle: if char == sep: pos.appendleft(counter) continue while True: try: next_char = next(iterator) counter += 1 if next_char == char: break except StopIteration: break while pos: next_pos = pos.popleft() output = output[:next_pos] + " " + output[next_pos:] return output df = pd.read_csv(filepath, header=None)[[0,1]] df = df.dropna() records = df.to_dict("records") for idx, item in enumerate(records): yield idx, { "index": idx, "identifier": item[0], "segmentation": get_segmentation(item[1], item[0]) }