File size: 3,195 Bytes
56d97b3
 
c59bed5
56d97b3
 
 
 
 
 
 
 
 
 
 
 
a9c7ae3
 
 
 
 
 
56d97b3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c59bed5
56d97b3
 
c59bed5
56d97b3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a9c7ae3
56d97b3
c59bed5
56d97b3
c59bed5
56d97b3
 
c59bed5
56d97b3
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
import gzip
import json
import logging
import os
from argparse import ArgumentParser


parser = ArgumentParser()
parser.add_argument(
    "-d",
    "--data_path",
    type=str,
    required=True,
    help="Data path containing the CompGuessWhat?! datasets (files with 'jsonl.gz' extension)",
)
parser.add_argument(
    "--examples",
    type=int,
    default=5,
    help="Number of games to consider in the dummy dataset",
)
original_data_files = {
    "train": "compguesswhat.train.jsonl.gz",
    "valid": "compguesswhat.valid.jsonl.gz",
    "test": "compguesswhat.test.jsonl.gz",
}

zs_data_files = {
    "nd_valid": "compguesswhat.nd_valid.jsonl.gz",
    "od_valid": "compguesswhat.od_valid.jsonl.gz",
    "nd_test": "compguesswhat.nd_test.jsonl.gz",
    "od_test": "compguesswhat.od_test.jsonl.gz",
}

COMPGUESSWHAT_ROOT = "datasets/compguesswhat/"


def create_dummy_data_for_split(data_path, dataset_name, dataset_version, data_files):
    full_dataset_name = "-".join(["compguesswhat", dataset_name])
    dummy_data_path = os.path.join(
        COMPGUESSWHAT_ROOT,
        "dummy",
        full_dataset_name,
        dataset_version,
        "dummy_data",
        full_dataset_name,
        dataset_version,
    )

    if not os.path.exists(dummy_data_path):
        os.makedirs(dummy_data_path)

    for split_name, split_file in data_files.items():
        logging.info(f"Generating dummy data for split {split_name} (num. examples = {args.examples})")

        split_filepath = os.path.join(data_path, full_dataset_name, dataset_version, split_file)
        logging.info(f"Reading split file {split_filepath}")
        with gzip.open(split_filepath) as in_file:
            dummy_filepath = os.path.join(dummy_data_path, split_file)
            with gzip.open(dummy_filepath, mode="w") as out_file:
                for i, line in enumerate(in_file):
                    if i > args.examples:
                        break

                    data = json.loads(line.strip())
                    out_file.write(json.dumps(data).encode("utf-8"))
                    out_file.write(b"\n")


def main(args):
    # args.data_path is the directory containing the already downloaded dataset files
    # we assume that the dataset test was successful and we have the file dataset_info.json
    dataset_info_path = os.path.join(COMPGUESSWHAT_ROOT, "dataset_infos.json")

    if not os.path.exists(dataset_info_path):
        raise ValueError(
            "The file 'dataset_info.json' doesn't exists. Make sure that you run the dataset tests via datasets-cli."
        )

    with open(dataset_info_path, encoding="utf-8") as in_file:
        dataset_info = json.load(in_file)

    dataset_version = dataset_info["compguesswhat-original"]["version"]["version_str"]

    logging.info(f"Creating dummy data for CompGuessWhat?! {dataset_version}")

    logging.info("Original dataset...")
    create_dummy_data_for_split(args.data_path, "original", dataset_version, original_data_files)

    logging.info("Zero-shot dataset...")
    create_dummy_data_for_split(args.data_path, "zero_shot", dataset_version, zs_data_files)


if __name__ == "__main__":
    args = parser.parse_args()
    main(args)