Spaces:
Running
on
Zero
Running
on
Zero
# Dataloader | |
Dataloader is the component that provides data to models. | |
A dataloader usually (but not necessarily) takes raw information from [datasets](./datasets.md), | |
and process them into a format needed by the model. | |
## How the Existing Dataloader Works | |
Detectron2 contains a builtin data loading pipeline. | |
It's good to understand how it works, in case you need to write a custom one. | |
Detectron2 provides two functions | |
[build_detection_{train,test}_loader](../modules/data.html#detectron2.data.build_detection_train_loader) | |
that create a default data loader from a given config. | |
Here is how `build_detection_{train,test}_loader` work: | |
1. It takes the name of a registered dataset (e.g., "coco_2017_train") and loads a `list[dict]` representing the dataset items | |
in a lightweight format. These dataset items are not yet ready to be used by the model (e.g., images are | |
not loaded into memory, random augmentations have not been applied, etc.). | |
Details about the dataset format and dataset registration can be found in | |
[datasets](./datasets.md). | |
2. Each dict in this list is mapped by a function ("mapper"): | |
* Users can customize this mapping function by specifying the "mapper" argument in | |
`build_detection_{train,test}_loader`. The default mapper is [DatasetMapper](../modules/data.html#detectron2.data.DatasetMapper). | |
* The output format of the mapper can be arbitrary, as long as it is accepted by the consumer of this data loader (usually the model). | |
The outputs of the default mapper, after batching, follow the default model input format documented in | |
[Use Models](./models.html#model-input-format). | |
* The role of the mapper is to transform the lightweight representation of a dataset item into a format | |
that is ready for the model to consume (including, e.g., read images, perform random data augmentation and convert to torch Tensors). | |
If you would like to perform custom transformations to data, you often want a custom mapper. | |
3. The outputs of the mapper are batched (simply into a list). | |
4. This batched data is the output of the data loader. Typically, it's also the input of | |
`model.forward()`. | |
## Write a Custom Dataloader | |
Using a different "mapper" with `build_detection_{train,test}_loader(mapper=)` works for most use cases | |
of custom data loading. | |
For example, if you want to resize all images to a fixed size for training, use: | |
```python | |
import detectron2.data.transforms as T | |
from detectron2.data import DatasetMapper # the default mapper | |
dataloader = build_detection_train_loader(cfg, | |
mapper=DatasetMapper(cfg, is_train=True, augmentations=[ | |
T.Resize((800, 800)) | |
])) | |
# use this dataloader instead of the default | |
``` | |
If the arguments of the default [DatasetMapper](../modules/data.html#detectron2.data.DatasetMapper) | |
does not provide what you need, you may write a custom mapper function and use it instead, e.g.: | |
```python | |
from detectron2.data import detection_utils as utils | |
# Show how to implement a minimal mapper, similar to the default DatasetMapper | |
def mapper(dataset_dict): | |
dataset_dict = copy.deepcopy(dataset_dict) # it will be modified by code below | |
# can use other ways to read image | |
image = utils.read_image(dataset_dict["file_name"], format="BGR") | |
# See "Data Augmentation" tutorial for details usage | |
auginput = T.AugInput(image) | |
transform = T.Resize((800, 800))(auginput) | |
image = torch.from_numpy(auginput.image.transpose(2, 0, 1)) | |
annos = [ | |
utils.transform_instance_annotations(annotation, [transform], image.shape[1:]) | |
for annotation in dataset_dict.pop("annotations") | |
] | |
return { | |
# create the format that the model expects | |
"image": image, | |
"instances": utils.annotations_to_instances(annos, image.shape[1:]) | |
} | |
dataloader = build_detection_train_loader(cfg, mapper=mapper) | |
``` | |
If you want to change not only the mapper (e.g., in order to implement different sampling or batching logic), | |
`build_detection_train_loader` won't work and you will need to write a different data loader. | |
The data loader is simply a | |
python iterator that produces [the format](./models.md) that the model accepts. | |
You can implement it using any tools you like. | |
No matter what to implement, it's recommended to | |
check out [API documentation of detectron2.data](../modules/data) to learn more about the APIs of | |
these functions. | |
## Use a Custom Dataloader | |
If you use [DefaultTrainer](../modules/engine.html#detectron2.engine.defaults.DefaultTrainer), | |
you can overwrite its `build_{train,test}_loader` method to use your own dataloader. | |
See the [deeplab dataloader](../../projects/DeepLab/train_net.py) | |
for an example. | |
If you write your own training loop, you can plug in your data loader easily. | |