import glob import json import os from collections import defaultdict _DATA_PATH = "data" data_split_sizes = defaultdict(dict) for stats in glob.glob(f"{_DATA_PATH}/*/stats.*.json"): folder_path = os.path.dirname(stats) task_name = folder_path.split("/")[-1] split_name = os.path.basename(stats).split(".")[1] if not os.path.exists(f"{folder_path}/COMPLETED"): continue with open(stats, "r") as f: split_stats = json.load(f) nb_examples = split_stats["examples"] if nb_examples > 0: data_split_sizes[task_name][split_name] = nb_examples with open("data_split_sizes.csv", "w", encoding="utf=8") as f: # The file is now merged into `tasks_splits_and_features.py` f.write("Data(sub)set|Number of examples per splits\n") for task_name in sorted(list(data_split_sizes.keys())): split_sizes_dict = json.dumps(data_split_sizes[task_name]) f.write(f"{task_name}|{split_sizes_dict}\n")