|
|
|
# Dataloader |
|
|
|
Dataloader is the component that provides data to models. |
|
A dataloader usually (but not necessarily) takes raw information from [datasets](./datasets.md), |
|
and process them into a format needed by the model. |
|
|
|
## How the Existing Dataloader Works |
|
|
|
Detectron2 contains a builtin data loading pipeline. |
|
It's good to understand how it works, in case you need to write a custom one. |
|
|
|
Detectron2 provides two functions |
|
[build_detection_{train,test}_loader](../modules/data.html#detectron2.data.build_detection_train_loader) |
|
that create a default data loader from a given config. |
|
Here is how `build_detection_{train,test}_loader` work: |
|
|
|
1. It takes the name of a registered dataset (e.g., "coco_2017_train") and loads a `list[dict]` representing the dataset items |
|
in a lightweight format. These dataset items are not yet ready to be used by the model (e.g., images are |
|
not loaded into memory, random augmentations have not been applied, etc.). |
|
Details about the dataset format and dataset registration can be found in |
|
[datasets](./datasets.md). |
|
2. Each dict in this list is mapped by a function ("mapper"): |
|
* Users can customize this mapping function by specifying the "mapper" argument in |
|
`build_detection_{train,test}_loader`. The default mapper is [DatasetMapper](../modules/data.html#detectron2.data.DatasetMapper). |
|
* The output format of the mapper can be arbitrary, as long as it is accepted by the consumer of this data loader (usually the model). |
|
The outputs of the default mapper, after batching, follow the default model input format documented in |
|
[Use Models](./models.html#model-input-format). |
|
* The role of the mapper is to transform the lightweight representation of a dataset item into a format |
|
that is ready for the model to consume (including, e.g., read images, perform random data augmentation and convert to torch Tensors). |
|
If you would like to perform custom transformations to data, you often want a custom mapper. |
|
3. The outputs of the mapper are batched (simply into a list). |
|
4. This batched data is the output of the data loader. Typically, it's also the input of |
|
`model.forward()`. |
|
|
|
|
|
## Write a Custom Dataloader |
|
|
|
Using a different "mapper" with `build_detection_{train,test}_loader(mapper=)` works for most use cases |
|
of custom data loading. |
|
For example, if you want to resize all images to a fixed size for training, use: |
|
|
|
```python |
|
import detectron2.data.transforms as T |
|
from detectron2.data import DatasetMapper # the default mapper |
|
dataloader = build_detection_train_loader(cfg, |
|
mapper=DatasetMapper(cfg, is_train=True, augmentations=[ |
|
T.Resize((800, 800)) |
|
])) |
|
# use this dataloader instead of the default |
|
``` |
|
If the arguments of the default [DatasetMapper](../modules/data.html#detectron2.data.DatasetMapper) |
|
does not provide what you need, you may write a custom mapper function and use it instead, e.g.: |
|
|
|
```python |
|
from detectron2.data import detection_utils as utils |
|
# Show how to implement a minimal mapper, similar to the default DatasetMapper |
|
def mapper(dataset_dict): |
|
dataset_dict = copy.deepcopy(dataset_dict) # it will be modified by code below |
|
# can use other ways to read image |
|
image = utils.read_image(dataset_dict["file_name"], format="BGR") |
|
# See "Data Augmentation" tutorial for details usage |
|
auginput = T.AugInput(image) |
|
transform = T.Resize((800, 800))(auginput) |
|
image = torch.from_numpy(auginput.image.transpose(2, 0, 1)) |
|
annos = [ |
|
utils.transform_instance_annotations(annotation, [transform], image.shape[1:]) |
|
for annotation in dataset_dict.pop("annotations") |
|
] |
|
return { |
|
# create the format that the model expects |
|
"image": image, |
|
"instances": utils.annotations_to_instances(annos, image.shape[1:]) |
|
} |
|
dataloader = build_detection_train_loader(cfg, mapper=mapper) |
|
``` |
|
|
|
If you want to change not only the mapper (e.g., in order to implement different sampling or batching logic), |
|
`build_detection_train_loader` won't work and you will need to write a different data loader. |
|
The data loader is simply a |
|
python iterator that produces [the format](./models.md) that the model accepts. |
|
You can implement it using any tools you like. |
|
|
|
No matter what to implement, it's recommended to |
|
check out [API documentation of detectron2.data](../modules/data) to learn more about the APIs of |
|
these functions. |
|
|
|
## Use a Custom Dataloader |
|
|
|
If you use [DefaultTrainer](../modules/engine.html#detectron2.engine.defaults.DefaultTrainer), |
|
you can overwrite its `build_{train,test}_loader` method to use your own dataloader. |
|
See the [deeplab dataloader](../../projects/DeepLab/train_net.py) |
|
for an example. |
|
|
|
If you write your own training loop, you can plug in your data loader easily. |
|
|