Transformers documentation

Backbones

You are viewing v4.37.0 version. A newer version v4.40.2 is available.
Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

Backbones

Backbones are models used for feature extraction for computer vision tasks. One can use a model as backbone in two ways:

  • initializing AutoBackbone class with a pretrained model,
  • initializing a supported backbone configuration and passing it to the model architecture.

Using AutoBackbone

You can use AutoBackbone class to initialize a model as a backbone and get the feature maps for any stage. You can define out_indices to indicate the index of the layers which you would like to get the feature maps from. You can also use out_features if you know the name of the layers. You can use them interchangeably. If you are using both out_indices and out_features, ensure they are consistent. Not passing any of the feature map arguments will make the backbone yield the feature maps of the last layer. To visualize how stages look like, let’s take the Swin model. Each stage is responsible from feature extraction, outputting feature maps.

Illustrating feature maps of the first stage looks like below.

Let’s see with an example. Note that out_indices=(0,) results in yielding the stem of the model. Stem refers to the stage before the first feature extraction stage. In above diagram, it refers to patch partition. We would like to have the feature maps from stem, first, and second stage of the model.

>>> from transformers import AutoImageProcessor, AutoBackbone
>>> import torch
>>> from PIL import Image
>>> import requests

>>> processor = AutoImageProcessor.from_pretrained("microsoft/swin-tiny-patch4-window7-224")
>>> model = AutoBackbone.from_pretrained("microsoft/swin-tiny-patch4-window7-224", out_indices=(0,1,2))
>>> url = "http://images.cocodataset.org/val2017/000000039769.jpg"
>>> image = Image.open(requests.get(url, stream=True).raw)

>>> inputs = processor(image, return_tensors="pt")
>>> outputs = model(**inputs)
>>> feature_maps = outputs.feature_maps

feature_maps object now has three feature maps, each can be accessed like below. Say we would like to get the feature map of the stem.

>>> list(feature_maps[0].shape)
[1, 96, 56, 56]

We can get the feature maps of first and second stages like below.

>>> list(feature_maps[1].shape)
[1, 96, 56, 56]
>>> list(feature_maps[2].shape)
[1, 192, 28, 28]

Initializing Backbone Configuration

In computer vision, models consist of backbone, neck, and a head. Backbone extracts the features, neck transforms the output of the backbone and head is used for the main task (e.g. object detection). You can initialize neck and head with model backbones by passing a model configuration to backbone_config. For example, below you can see how to initialize the MaskFormer model with instance segmentation head with ResNet backbone.

from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, ResNetConfig

backbone_config = ResNetConfig.from_pretrained("microsoft/resnet-50")
config = MaskFormerConfig(backbone_config=backbone_config)
model = MaskFormerForInstanceSegmentation(config)

You can also initialize a backbone with random weights to initialize the model neck with it.

backbone_config = ResNetConfig()
config = MaskFormerConfig(backbone_config=backbone_config)
model = MaskFormerForInstanceSegmentation(config)

timm models are also supported in transformers through TimmBackbone and TimmBackboneConfig.

from transformers import TimmBackboneConfig, TimmBackbone

backbone_config = TimmBackboneConfig("resnet50")
model = TimmBackbone(config=backbone_config)