import argparse from datasets import load_dataset import os def save_dataset_to_parquet(dataset, output_path, dataset_name): os.makedirs(output_path, exist_ok=True) for split in dataset: split_dataset = dataset[split] # Save the entire split as a single Parquet file split_path = os.path.join(output_path, dataset_name, f"{split}.parquet") os.makedirs(os.path.dirname(split_path), exist_ok=True) # Create directories if needed split_dataset.to_parquet(split_path) def main(dataset_name): num_proc = os.cpu_count() dataset_paths = { "speech_vqav2": "./hf_generator_vqav2.py", "speech_vg": "./hf_generator_vg.py", "speech_asr": "./hf_generator_asr.py", "speech_tts": "./hf_generator_tts.py", "laion": "./hf_generator_laion.py", } if dataset_name not in dataset_paths: raise ValueError(f"Dataset name '{dataset_name}' not recognized. Available options: {list(dataset_paths.keys())}") ds = load_dataset( dataset_paths[dataset_name], trust_remote_code=True, cache_dir=f"~/.cache/{dataset_name}", num_proc=num_proc ) save_dataset_to_parquet(ds, "./hf_datasets/", dataset_name) if __name__ == "__main__": parser = argparse.ArgumentParser(description="Save specified Hugging Face dataset to Parquet format.") parser.add_argument("dataset_name", type=str, help="Name of the dataset to load and save to Parquet. Options: speech_vqav2, speech_vg, speech_asr, speech_tts, laion") args = parser.parse_args() main(args.dataset_name) # from datasets import load_dataset # import os # def save_dataset_to_parquet(dataset, output_path, dataset_name): # os.makedirs(output_path, exist_ok=True) # for split in dataset: # split_dataset = dataset[split] # # Save the entire split as a single Parquet file # split_path = os.path.join(output_path, dataset_name, f"{split}.parquet") # os.makedirs(os.path.dirname(split_path), exist_ok=True) # Create directories if needed # split_dataset.to_parquet(split_path) # num_proc = os.cpu_count() # ds = load_dataset("./hf_generator_vqav2.py", trust_remote_code=True, cache_dir="~/.cache/speech_vqav2", num_proc=num_proc) # save_dataset_to_parquet(ds, "./hf_datasets/", "speech_vqav2") # ds = load_dataset("./hf_generator_vg.py", trust_remote_code=True, cache_dir="~/.cache/speech_vg", num_proc=num_proc) # save_dataset_to_parquet(ds, "./hf_datasets/", "speech_vg") # ds = load_dataset("./hf_generator_asr.py", trust_remote_code=True, cache_dir="~/.cache/speech_asr", num_proc=64) # save_dataset_to_parquet(ds, "./hf_datasets/", "speech_asr") # ds = load_dataset("./hf_generator_tts.py", trust_remote_code=True, cache_dir="~/.cache/speech_tts", num_proc=64) # save_dataset_to_parquet(ds, "./hf_datasets/", "speech_tts") # ds = load_dataset("./hf_generator_laion.py", trust_remote_code=True, cache_dir="~/.cache/speech_laion", num_proc=64) # save_dataset_to_parquet(ds, "./hf_datasets/", "laion")