|
|
|
""" |
|
MIT License |
|
Copyright (c) 2019 Microsoft |
|
Permission is hereby granted, free of charge, to any person obtaining a copy |
|
of this software and associated documentation files (the "Software"), to deal |
|
in the Software without restriction, including without limitation the rights |
|
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell |
|
copies of the Software, and to permit persons to whom the Software is |
|
furnished to do so, subject to the following conditions: |
|
The above copyright notice and this permission notice shall be included in all |
|
copies or substantial portions of the Software. |
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
|
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
|
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
|
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
|
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
|
SOFTWARE. |
|
""" |
|
|
|
import torch |
|
import torch.nn as nn |
|
import torch.nn.functional as F |
|
|
|
from detectron2.layers import ShapeSpec |
|
from detectron2.modeling.backbone import BACKBONE_REGISTRY |
|
from detectron2.modeling.backbone.backbone import Backbone |
|
|
|
from .hrnet import build_pose_hrnet_backbone |
|
|
|
|
|
class HRFPN(Backbone): |
|
"""HRFPN (High Resolution Feature Pyramids) |
|
Transforms outputs of HRNet backbone so they are suitable for the ROI_heads |
|
arXiv: https://arxiv.org/abs/1904.04514 |
|
Adapted from https://github.com/open-mmlab/mmdetection/blob/master/mmdet/models/necks/hrfpn.py |
|
Args: |
|
bottom_up: (list) output of HRNet |
|
in_features (list): names of the input features (output of HRNet) |
|
in_channels (list): number of channels for each branch |
|
out_channels (int): output channels of feature pyramids |
|
n_out_features (int): number of output stages |
|
pooling (str): pooling for generating feature pyramids (from {MAX, AVG}) |
|
share_conv (bool): Have one conv per output, or share one with all the outputs |
|
""" |
|
|
|
def __init__( |
|
self, |
|
bottom_up, |
|
in_features, |
|
n_out_features, |
|
in_channels, |
|
out_channels, |
|
pooling="AVG", |
|
share_conv=False, |
|
): |
|
super(HRFPN, self).__init__() |
|
assert isinstance(in_channels, list) |
|
self.bottom_up = bottom_up |
|
self.in_features = in_features |
|
self.n_out_features = n_out_features |
|
self.in_channels = in_channels |
|
self.out_channels = out_channels |
|
self.num_ins = len(in_channels) |
|
self.share_conv = share_conv |
|
|
|
if self.share_conv: |
|
self.fpn_conv = nn.Conv2d( |
|
in_channels=out_channels, out_channels=out_channels, kernel_size=3, padding=1 |
|
) |
|
else: |
|
self.fpn_conv = nn.ModuleList() |
|
for _ in range(self.n_out_features): |
|
self.fpn_conv.append( |
|
nn.Conv2d( |
|
in_channels=out_channels, |
|
out_channels=out_channels, |
|
kernel_size=3, |
|
padding=1, |
|
) |
|
) |
|
|
|
|
|
self.interp_conv = nn.ModuleList() |
|
for i in range(len(self.in_features)): |
|
self.interp_conv.append( |
|
nn.Sequential( |
|
nn.ConvTranspose2d( |
|
in_channels=in_channels[i], |
|
out_channels=in_channels[i], |
|
kernel_size=4, |
|
stride=2**i, |
|
padding=0, |
|
output_padding=0, |
|
bias=False, |
|
), |
|
nn.BatchNorm2d(in_channels[i], momentum=0.1), |
|
nn.ReLU(inplace=True), |
|
) |
|
) |
|
|
|
|
|
self.reduction_pooling_conv = nn.ModuleList() |
|
for i in range(self.n_out_features): |
|
self.reduction_pooling_conv.append( |
|
nn.Sequential( |
|
nn.Conv2d(sum(in_channels), out_channels, kernel_size=2**i, stride=2**i), |
|
nn.BatchNorm2d(out_channels, momentum=0.1), |
|
nn.ReLU(inplace=True), |
|
) |
|
) |
|
|
|
if pooling == "MAX": |
|
self.pooling = F.max_pool2d |
|
else: |
|
self.pooling = F.avg_pool2d |
|
|
|
self._out_features = [] |
|
self._out_feature_channels = {} |
|
self._out_feature_strides = {} |
|
|
|
for i in range(self.n_out_features): |
|
self._out_features.append("p%d" % (i + 1)) |
|
self._out_feature_channels.update({self._out_features[-1]: self.out_channels}) |
|
self._out_feature_strides.update({self._out_features[-1]: 2 ** (i + 2)}) |
|
|
|
|
|
def init_weights(self): |
|
for m in self.modules(): |
|
if isinstance(m, nn.Conv2d): |
|
nn.init.kaiming_normal_(m.weight, a=1) |
|
nn.init.constant_(m.bias, 0) |
|
|
|
def forward(self, inputs): |
|
bottom_up_features = self.bottom_up(inputs) |
|
assert len(bottom_up_features) == len(self.in_features) |
|
inputs = [bottom_up_features[f] for f in self.in_features] |
|
|
|
outs = [] |
|
for i in range(len(inputs)): |
|
outs.append(self.interp_conv[i](inputs[i])) |
|
shape_2 = min(o.shape[2] for o in outs) |
|
shape_3 = min(o.shape[3] for o in outs) |
|
out = torch.cat([o[:, :, :shape_2, :shape_3] for o in outs], dim=1) |
|
outs = [] |
|
for i in range(self.n_out_features): |
|
outs.append(self.reduction_pooling_conv[i](out)) |
|
for i in range(len(outs)): |
|
outs[-1 - i] = outs[-1 - i][ |
|
:, :, : outs[-1].shape[2] * 2**i, : outs[-1].shape[3] * 2**i |
|
] |
|
outputs = [] |
|
for i in range(len(outs)): |
|
if self.share_conv: |
|
outputs.append(self.fpn_conv(outs[i])) |
|
else: |
|
outputs.append(self.fpn_conv[i](outs[i])) |
|
|
|
assert len(self._out_features) == len(outputs) |
|
return dict(zip(self._out_features, outputs)) |
|
|
|
|
|
@BACKBONE_REGISTRY.register() |
|
def build_hrfpn_backbone(cfg, input_shape: ShapeSpec) -> HRFPN: |
|
|
|
in_channels = cfg.MODEL.HRNET.STAGE4.NUM_CHANNELS |
|
in_features = ["p%d" % (i + 1) for i in range(cfg.MODEL.HRNET.STAGE4.NUM_BRANCHES)] |
|
n_out_features = len(cfg.MODEL.ROI_HEADS.IN_FEATURES) |
|
out_channels = cfg.MODEL.HRNET.HRFPN.OUT_CHANNELS |
|
hrnet = build_pose_hrnet_backbone(cfg, input_shape) |
|
hrfpn = HRFPN( |
|
hrnet, |
|
in_features, |
|
n_out_features, |
|
in_channels, |
|
out_channels, |
|
pooling="AVG", |
|
share_conv=False, |
|
) |
|
|
|
return hrfpn |
|
|