The dataset viewer is not available for this subset.
Cannot get the split names for the config 'default' of the dataset.
Exception:    SplitsNotFoundError
Message:      The split names could not be parsed from the dataset config.
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/split_names.py", line 159, in compute
                  compute_split_names_from_info_response(
                File "/src/services/worker/src/worker/job_runners/config/split_names.py", line 131, in compute_split_names_from_info_response
                  config_info_response = get_previous_step_or_raise(kind="config-info", dataset=dataset, config=config)
                File "/src/libs/libcommon/src/libcommon/simple_cache.py", line 567, in get_previous_step_or_raise
                  raise CachedArtifactError(
              libcommon.simple_cache.CachedArtifactError: The previous step failed.
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/usr/local/lib/python3.9/tarfile.py", line 190, in nti
                  s = nts(s, "ascii", "strict")
                File "/usr/local/lib/python3.9/tarfile.py", line 174, in nts
                  return s.decode(encoding, errors)
              UnicodeDecodeError: 'ascii' codec can't decode byte 0xbb in position 1: ordinal not in range(128)
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/usr/local/lib/python3.9/tarfile.py", line 2588, in next
                  tarinfo = self.tarinfo.fromtarfile(self)
                File "/usr/local/lib/python3.9/tarfile.py", line 1292, in fromtarfile
                  obj = cls.frombuf(buf, tarfile.encoding, tarfile.errors)
                File "/usr/local/lib/python3.9/tarfile.py", line 1234, in frombuf
                  chksum = nti(buf[148:156])
                File "/usr/local/lib/python3.9/tarfile.py", line 193, in nti
                  raise InvalidHeaderError("invalid header")
              tarfile.InvalidHeaderError: invalid header
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 499, in get_dataset_config_info
                  for split_generator in builder._split_generators(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/webdataset/webdataset.py", line 86, in _split_generators
                  first_examples = list(islice(pipeline, self.NUM_EXAMPLES_FOR_FEATURES_INFERENCE))
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/packaged_modules/webdataset/webdataset.py", line 30, in _get_pipeline_from_tar
                  for filename, f in tar_iterator:
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/file_utils.py", line 1577, in __iter__
                  for x in self.generator(*self.args, **self.kwargs):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/file_utils.py", line 1637, in _iter_from_urlpath
                  yield from cls._iter_tar(f)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/utils/file_utils.py", line 1588, in _iter_tar
                  stream = tarfile.open(fileobj=f, mode="r|*")
                File "/usr/local/lib/python3.9/tarfile.py", line 1822, in open
                  t = cls(name, filemode, stream, **kwargs)
                File "/usr/local/lib/python3.9/tarfile.py", line 1703, in __init__
                  self.firstmember = self.next()
                File "/usr/local/lib/python3.9/tarfile.py", line 2600, in next
                  raise ReadError(str(e))
              tarfile.ReadError: invalid header
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/split_names.py", line 75, in compute_split_names_from_streaming_response
                  for split in get_dataset_split_names(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 572, in get_dataset_split_names
                  info = get_dataset_config_info(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/inspect.py", line 504, in get_dataset_config_info
                  raise SplitsNotFoundError("The split names could not be parsed from the dataset config.") from err
              datasets.inspect.SplitsNotFoundError: The split names could not be parsed from the dataset config.

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/datasets-cards)

data-intermediate

If you are looking for our test ready version, please refer to mango-ttic/data

Find more about us at mango.ttic.edu

Folder Structure

Each folder inside data-intermediate contains all intermediate files we used during data annotation and generation. Here is the tree structure from game data-intermediate/night .

data-intermediate/night/
├── night.all2all.json      # all simple paths between any 2 nodes
├── night.all_pairs.json    # all connectivity between any 2 nodes 
├── night.anno2code.json    # annotation to codename mapping
├── night.code2anno.json    # codename to annotation mapping
├── night.edges.json        # list of all edges
├── night.map.human         # human map derived from human annotation
├── night.map.machine       # machine map derived from exported action sequences
├── night.map.reversed      # reverse map derived from human annotation map
├── night.moves             # list of mentioned actions
├── night.nodes.json        # list of all nodes
├── night.valid_moves.csv   # human annotation
├── night.walkthrough       # enriched walkthrough exported from Jericho simulator
└── night.walkthrough_acts  # action sequences exported from Jericho simulator

Variations

70-step vs all-step version

In our paper, we benchmark using the first 70 steps of the walkthrough from each game. We also provide all-step versions of both data and data-intermediate collection.

  • 70-step data-intermediate-70steps.tar.zst: contains the first 70 steps of each walkthrough. If the complete walkthrough is shorter than 70 steps, then all steps are used.

  • All-step data-intermediate.tar.zst: contains all steps of each walkthrough.

Word-only & Word+ID

  • Word-only data-intermediate.tar.zst: Nodes are annotated by additional descriptive text to distinguish different locations with similar names.

  • Word + Object ID data-intermediate-objid.tar.zst: variation of the word-only version, where nodes are labeled using minimaly fixed names with object id from Jericho simulator.

  • Word + Random ID data-intermediate-randid.tar.zst: variation of the Jericho ID version, where the Jericho object id replaced with randomly generated integer.

We primarily rely on the word-only version as benchmark, yet providing word+ID version for diverse benchmark settings.

How to use

We use data-intermediate.tar.zst as an example here.

1. download from Huggingface

by directly download

You can selectively download certain variation of your choice.

by git

Make sure you have git-lfs installed

git lfs install
git clone https://huggingface.co/datasets/mango-ttic/data-intermediate

# or, use hf-mirror if your connection to huggingface.co is slow
# git clone https://hf-mirror.com/datasets/mango-ttic/data-intermediate

If you want to clone without large files - just their pointers

GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/datasets/mango-ttic/data-intermediate

# or, use hf-mirror if your connection to huggingface.co is slow
# GIT_LFS_SKIP_SMUDGE=1 git clone https://hf-mirror.com/datasets/mango-ttic/data-intermediate

2. decompress

Because some json files are huge, we use tar.zst to package the data efficiently.

silently decompress

tar -I 'zstd -d' -xf data-intermediate.tar.zst

or, verbosely decompress

zstd -d -c data-intermediate.tar.zst | tar -xvf -
Downloads last month
44