| from omegaconf import OmegaConf |
| from torch.utils.data import DataLoader, default_collate, ConcatDataset |
| from fvcore.common.registry import Registry |
|
|
| from .datasets.dataset_wrapper import DATASETWRAPPER_REGISTRY |
|
|
| DATASET_REGISTRY = Registry("dataset") |
| DATASET_REGISTRY.__doc__ = """ |
| Registry for datasets, which takes a list of dataset names and returns a dataset object. |
| Currently it performs similar as registering dataset loading functions, but remains in a |
| form of object class for future purposes. |
| """ |
|
|
| def get_dataset(cfg, split): |
| assert cfg.data.get(split), f"No valid dataset name in {split}." |
| dataset_list = [] |
| print(split, ': ', ', '.join(cfg.data.get(split))) |
| for dataset_name in cfg.data.get(split): |
| _dataset = DATASET_REGISTRY.get(dataset_name)(cfg, split) |
| assert len(_dataset), f"Dataset '{dataset_name}' is empty!" |
| wrapper = cfg.data_wrapper.get(split, cfg.data_wrapper) if not isinstance(cfg.data_wrapper, str) else cfg.data_wrapper |
| _dataset = DATASETWRAPPER_REGISTRY.get(wrapper)(cfg, _dataset, split=split) |
| |
| |
| if cfg.data.get('use_voxel', None): |
| _dataset = DATASETWRAPPER_REGISTRY.get('VoxelDatasetWrapper')(cfg, _dataset) |
| dataset_list.append(_dataset) |
|
|
| print('='*50) |
| print('Dataset\t\t\tSize') |
| total = sum([len(dataset) for dataset in dataset_list]) |
| for dataset_name, dataset in zip(cfg.data.get(split), dataset_list): |
| print(f'{dataset_name:<20} {len(dataset):>6} ({len(dataset) / total * 100:.1f}%)') |
| print(f'Total\t\t\t{total}') |
| print('='*50) |
| if split in ['warmup', 'pretrain', 'train']: |
| dataset_list = ConcatDataset(dataset_list) |
| return dataset_list |
|
|
|
|
| def build_dataloader(cfg, split='train'): |
| """_summary_ |
| Unittest: |
| dataloader_train = build_dataloader(default_cfg, split='train') |
| for _item in dataloader_train: |
| print(_item.keys()) |
| |
| Args: |
| cfg (_type_): _description_ |
| split (str, optional): _description_. Defaults to 'train'. |
| |
| Returns: |
| _type_: _description_ |
| """ |
| |
| if split in ['warmup','pretrain']: |
| dataset= get_dataset(cfg, split) |
| collate_fn = getattr(dataset.datasets[0], 'collate_fn', default_collate) |
| return DataLoader(dataset, |
| batch_size=cfg.dataloader.batchsize, |
| num_workers=cfg.dataloader.num_workers, |
| collate_fn = collate_fn, |
| pin_memory=True, |
| persistent_workers = False, |
| shuffle=True, |
| drop_last=True) |
| else: |
| loader_list = [] |
| collate_fn = default_collate |
| if split == 'train': |
| dataset = get_dataset(cfg, split) |
| return DataLoader(dataset, |
| batch_size=cfg.dataloader.get('batchsize_eval', cfg.dataloader.batchsize), |
| num_workers=8, |
| collate_fn = collate_fn, |
| pin_memory=True, |
| persistent_workers = True, |
| drop_last=True, |
| prefetch_factor=4, |
| shuffle=True) |
| else: |
| for dataset in get_dataset(cfg, split): |
| loader_list.append( |
| DataLoader(dataset, |
| batch_size=cfg.dataloader.get('batchsize_eval', cfg.dataloader.batchsize), |
| num_workers=8, |
| collate_fn = collate_fn, |
| pin_memory=True, |
| shuffle=False, |
| prefetch_factor=4, |
| persistent_workers = True)) |
| |
| |
| if len(loader_list) == 1: |
| return loader_list[0] |
| else: |
| return loader_list |
|
|
| if __name__ == '__main__': |
| pass |
|
|