|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import json |
|
|
|
import datasets |
|
|
|
_CITATION = ''' |
|
@inproceedings{Ammanabrolu2020AAAI, |
|
title={Story Realization: Expanding Plot Events into Sentences}, |
|
author={Prithviraj Ammanabrolu and Ethan Tien and Wesley Cheung and Zhaochen Luo and William Ma and Lara J. Martin and Mark O. Riedl}, |
|
journal={Proceedings of the AAAI Conference on Artificial Intelligence (AAAI)}, |
|
year={2020}, |
|
volume={34}, |
|
number={05}, |
|
url={https://ojs.aaai.org//index.php/AAAI/article/view/6232} |
|
} |
|
''' |
|
|
|
_DESCRIPTION = 'Loading script for the science fiction TV show plot dataset.' |
|
|
|
URL = 'https://huggingface.co/datasets/lara-martin/Scifi_TV_Shows/blob/main/' |
|
_URLS = { |
|
'test':'Test-Train-Val/all-sci-fi-data-test.txt', |
|
'train':'Test-Train-Val/all-sci-fi-data-train.txt', |
|
'val':'Test-Train-Val/all-sci-fi-data-val.txt', |
|
'all':'all-sci-fi-data.txt', |
|
} |
|
|
|
_INPUT_OUTPUT = ["all-sci-fi-data-test_input.txt", "all-sci-fi-data-test_output.txt", "all-sci-fi-data-train_input.txt", "all-sci-fi-data-train_output.txt", "all-sci-fi-data-val_input.txt", "all-sci-fi-data-val_output.txt"] |
|
|
|
|
|
class ScifiTV(datasets.GeneratorBasedBuilder): |
|
BUILDER_CONFIGS = [ |
|
datasets.BuilderConfig( |
|
version=datasets.Version('1.1.0'), |
|
name=k, |
|
description=f'Science fiction TV show plot summaries.' |
|
) for k in _URLS.keys() |
|
] |
|
|
|
def _info(self): |
|
features = datasets.Features({ |
|
'event': datasets.Value('string'), |
|
'gen_event': datasets.Value('string'), |
|
'sent': datasets.Value('string'), |
|
'gen_sent': datasets.Value('string'), |
|
'story_num': datasets.Value('int16'), |
|
'entities': datasets.Value('string') |
|
}) |
|
|
|
return datasets.DatasetInfo( |
|
|
|
description=_DESCRIPTION, |
|
|
|
features=features, |
|
supervised_keys=None, |
|
|
|
homepage='https://github.com/rajammanabrolu/StoryRealization', |
|
|
|
license='The Creative Commons Attribution 4.0 International License. https://creativecommons.org/licenses/by/4.0/', |
|
|
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
downloaded_files = dl_manager.download(_URLS) |
|
return[ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={ |
|
'filepath': downloaded_files['train'], |
|
"split": "train", |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
gen_kwargs={ |
|
'filepath': downloaded_files['test'], |
|
"split": "test", |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.VALIDATION, |
|
gen_kwargs={ |
|
'filepath': downloaded_files['val'], |
|
"split": "val", |
|
}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, filepath): |
|
story_count = 0 |
|
with open(filepath, encoding="utf-8") as f: |
|
story = [] |
|
for id_, line in enumerate(f.readlines()): |
|
line = line.strip() |
|
if "%%%%%%" in line: |
|
for l in story: |
|
event, gen_event, sent, gen_sent = l.split("|||") |
|
line = line.replace("%%%%%%%%%%%%%%%%%", "") |
|
entities = line.replace("defaultdict(<type 'list'>, ", "")[:-1] |
|
yield id_, { |
|
'event': event, |
|
'gen_event': gen_event, |
|
'sent': sent, |
|
'gen_sent': gen_sent, |
|
'story_num': story_count, |
|
'entities': entities, |
|
} |
|
story = [] |
|
story_count+=1 |
|
elif "<EOS>" in line: |
|
continue |
|
else: |
|
story.append(line) |
|
|
|
|
|
|