Leffa / detectron2 /layers /csrc /deformable /deform_conv_cuda.cu
franciszzj's picture
init code
b213d84
// Copyright (c) Facebook, Inc. and its affiliates.
// modified from
// https://github.com/open-mmlab/mmdetection/blob/master/mmdet/ops/dcn/src/deform_conv_cuda.cpp
// Original license: Apache 2.0
// modify from
// https://github.com/chengdazhi/Deformable-Convolution-V2-PyTorch/blob/mmdetection/mmdet/ops/dcn/src/deform_conv_cuda.c
// Original license: Apache 2.0
#include <torch/types.h>
#include "deform_conv.h"
#include <cmath>
#include <vector>
namespace detectron2 {
void deformable_im2col(
const at::Tensor data_im,
const at::Tensor data_offset,
const int channels,
const int height,
const int width,
const int ksize_h,
const int ksize_w,
const int pad_h,
const int pad_w,
const int stride_h,
const int stride_w,
const int dilation_h,
const int dilation_w,
const int parallel_imgs,
const int deformable_group,
at::Tensor data_col);
void deformable_col2im(
const at::Tensor data_col,
const at::Tensor data_offset,
const int channels,
const int height,
const int width,
const int ksize_h,
const int ksize_w,
const int pad_h,
const int pad_w,
const int stride_h,
const int stride_w,
const int dilation_h,
const int dilation_w,
const int parallel_imgs,
const int deformable_group,
at::Tensor grad_im);
void deformable_col2im_coord(
const at::Tensor data_col,
const at::Tensor data_im,
const at::Tensor data_offset,
const int channels,
const int height,
const int width,
const int ksize_h,
const int ksize_w,
const int pad_h,
const int pad_w,
const int stride_h,
const int stride_w,
const int dilation_h,
const int dilation_w,
const int parallel_imgs,
const int deformable_group,
at::Tensor grad_offset);
void modulated_deformable_im2col_cuda(
const at::Tensor data_im,
const at::Tensor data_offset,
const at::Tensor data_mask,
const int batch_size,
const int channels,
const int height_im,
const int width_im,
const int height_col,
const int width_col,
const int kernel_h,
const int kenerl_w,
const int pad_h,
const int pad_w,
const int stride_h,
const int stride_w,
const int dilation_h,
const int dilation_w,
const int deformable_group,
at::Tensor data_col);
void modulated_deformable_col2im_cuda(
const at::Tensor data_col,
const at::Tensor data_offset,
const at::Tensor data_mask,
const int batch_size,
const int channels,
const int height_im,
const int width_im,
const int height_col,
const int width_col,
const int kernel_h,
const int kenerl_w,
const int pad_h,
const int pad_w,
const int stride_h,
const int stride_w,
const int dilation_h,
const int dilation_w,
const int deformable_group,
at::Tensor grad_im);
void modulated_deformable_col2im_coord_cuda(
const at::Tensor data_col,
const at::Tensor data_im,
const at::Tensor data_offset,
const at::Tensor data_mask,
const int batch_size,
const int channels,
const int height_im,
const int width_im,
const int height_col,
const int width_col,
const int kernel_h,
const int kenerl_w,
const int pad_h,
const int pad_w,
const int stride_h,
const int stride_w,
const int dilation_h,
const int dilation_w,
const int deformable_group,
at::Tensor grad_offset,
at::Tensor grad_mask);
void shape_check(
at::Tensor input,
at::Tensor offset,
at::Tensor* gradOutput,
at::Tensor weight,
int kH,
int kW,
int dH,
int dW,
int padH,
int padW,
int dilationH,
int dilationW,
int group,
int deformable_group) {
TORCH_CHECK(
weight.ndimension() == 4,
"4D weight tensor (nOutputPlane,nInputPlane,kH,kW) expected, "
"but got: %s",
weight.ndimension());
TORCH_CHECK(weight.is_contiguous(), "weight tensor has to be contiguous");
TORCH_CHECK(
kW > 0 && kH > 0,
"kernel size should be greater than zero, but got kH: %d kW: %d",
kH,
kW);
TORCH_CHECK(
(weight.size(2) == kH && weight.size(3) == kW),
"kernel size should be consistent with weight, ",
"but got kH: %d kW: %d weight.size(2): %d, weight.size(3): %d",
kH,
kW,
weight.size(2),
weight.size(3));
TORCH_CHECK(
dW > 0 && dH > 0,
"stride should be greater than zero, but got dH: %d dW: %d",
dH,
dW);
TORCH_CHECK(
dilationW > 0 && dilationH > 0,
"dilation should be greater than 0, but got dilationH: %d dilationW: %d",
dilationH,
dilationW);
int ndim = input.ndimension();
int dimf = 0;
int dimh = 1;
int dimw = 2;
if (ndim == 4) {
dimf++;
dimh++;
dimw++;
}
TORCH_CHECK(
ndim == 3 || ndim == 4,
"3D or 4D input tensor expected but got: %s",
ndim);
long nInputPlane = weight.size(1) * group;
long inputHeight = input.size(dimh);
long inputWidth = input.size(dimw);
long nOutputPlane = weight.size(0);
long outputHeight =
(inputHeight + 2 * padH - (dilationH * (kH - 1) + 1)) / dH + 1;
long outputWidth =
(inputWidth + 2 * padW - (dilationW * (kW - 1) + 1)) / dW + 1;
TORCH_CHECK(
nInputPlane % deformable_group == 0,
"input channels must divide deformable group size");
if (outputWidth < 1 || outputHeight < 1)
AT_ERROR(
"Given input size: (%ld x %ld x %ld). "
"Calculated output size: (%ld x %ld x %ld). Output size is too small",
nInputPlane,
inputHeight,
inputWidth,
nOutputPlane,
outputHeight,
outputWidth);
TORCH_CHECK(
input.size(1) == nInputPlane,
"invalid number of input planes, expected: %d, but got: %d",
nInputPlane,
input.size(1));
TORCH_CHECK(
(inputHeight + 2 * padH >= kH && inputWidth + 2 * padW >= kW),
"input image is smaller than kernel");
TORCH_CHECK(
(offset.size(2) == outputHeight && offset.size(3) == outputWidth),
"invalid spatial size of offset, expected height: %d width: %d, but "
"got height: %d width: %d",
outputHeight,
outputWidth,
offset.size(2),
offset.size(3));
TORCH_CHECK(
(offset.size(1) == deformable_group * 2 * kH * kW),
"invalid number of channels of offset");
if (gradOutput != NULL) {
TORCH_CHECK(
gradOutput->size(dimf) == nOutputPlane,
"invalid number of gradOutput planes, expected: %d, but got: %d",
nOutputPlane,
gradOutput->size(dimf));
TORCH_CHECK(
(gradOutput->size(dimh) == outputHeight &&
gradOutput->size(dimw) == outputWidth),
"invalid size of gradOutput, expected height: %d width: %d , but "
"got height: %d width: %d",
outputHeight,
outputWidth,
gradOutput->size(dimh),
gradOutput->size(dimw));
}
}
int deform_conv_forward_cuda(
at::Tensor input,
at::Tensor weight,
at::Tensor offset,
at::Tensor output,
at::Tensor columns,
at::Tensor ones,
int kW,
int kH,
int dW,
int dH,
int padW,
int padH,
int dilationW,
int dilationH,
int group,
int deformable_group,
int im2col_step) {
// todo: resize columns to include im2col: done
// todo: add im2col_step as input
// todo: add new output buffer and transpose it to output (or directly
// transpose output) todo: possibly change data indexing because of
// parallel_imgs
shape_check(
input,
offset,
NULL,
weight,
kH,
kW,
dH,
dW,
padH,
padW,
dilationH,
dilationW,
group,
deformable_group);
input = input.contiguous();
offset = offset.contiguous();
weight = weight.contiguous();
int batch = 1;
if (input.ndimension() == 3) {
// Force batch
batch = 0;
input.unsqueeze_(0);
offset.unsqueeze_(0);
}
// todo: assert batchsize dividable by im2col_step
long batchSize = input.size(0);
long nInputPlane = input.size(1);
long inputHeight = input.size(2);
long inputWidth = input.size(3);
long nOutputPlane = weight.size(0);
long outputWidth =
(inputWidth + 2 * padW - (dilationW * (kW - 1) + 1)) / dW + 1;
long outputHeight =
(inputHeight + 2 * padH - (dilationH * (kH - 1) + 1)) / dH + 1;
TORCH_CHECK((offset.size(0) == batchSize), "invalid batch size of offset");
output = output.view(
{batchSize / im2col_step,
im2col_step,
nOutputPlane,
outputHeight,
outputWidth});
columns = at::zeros(
{nInputPlane * kW * kH, im2col_step * outputHeight * outputWidth},
input.options());
if (ones.ndimension() != 2 ||
ones.size(0) * ones.size(1) < outputHeight * outputWidth) {
ones = at::ones({outputHeight, outputWidth}, input.options());
}
input = input.view(
{batchSize / im2col_step,
im2col_step,
nInputPlane,
inputHeight,
inputWidth});
offset = offset.view(
{batchSize / im2col_step,
im2col_step,
deformable_group * 2 * kH * kW,
outputHeight,
outputWidth});
at::Tensor output_buffer = at::zeros(
{batchSize / im2col_step,
nOutputPlane,
im2col_step * outputHeight,
outputWidth},
output.options());
output_buffer = output_buffer.view(
{output_buffer.size(0),
group,
output_buffer.size(1) / group,
output_buffer.size(2),
output_buffer.size(3)});
for (int elt = 0; elt < batchSize / im2col_step; elt++) {
deformable_im2col(
input[elt],
offset[elt],
nInputPlane,
inputHeight,
inputWidth,
kH,
kW,
padH,
padW,
dH,
dW,
dilationH,
dilationW,
im2col_step,
deformable_group,
columns);
columns = columns.view({group, columns.size(0) / group, columns.size(1)});
weight = weight.view(
{group,
weight.size(0) / group,
weight.size(1),
weight.size(2),
weight.size(3)});
for (int g = 0; g < group; g++) {
output_buffer[elt][g] = output_buffer[elt][g]
.flatten(1)
.addmm_(weight[g].flatten(1), columns[g])
.view_as(output_buffer[elt][g]);
}
}
output_buffer = output_buffer.view(
{output_buffer.size(0),
output_buffer.size(1) * output_buffer.size(2),
output_buffer.size(3),
output_buffer.size(4)});
output_buffer = output_buffer.view(
{batchSize / im2col_step,
nOutputPlane,
im2col_step,
outputHeight,
outputWidth});
output_buffer.transpose_(1, 2);
output.copy_(output_buffer);
output = output.view({batchSize, nOutputPlane, outputHeight, outputWidth});
input = input.view({batchSize, nInputPlane, inputHeight, inputWidth});
offset = offset.view(
{batchSize, deformable_group * 2 * kH * kW, outputHeight, outputWidth});
if (batch == 0) {
output = output.view({nOutputPlane, outputHeight, outputWidth});
input = input.view({nInputPlane, inputHeight, inputWidth});
offset = offset.view({offset.size(1), offset.size(2), offset.size(3)});
}
return 1;
}
int deform_conv_backward_input_cuda(
at::Tensor input,
at::Tensor offset,
at::Tensor gradOutput,
at::Tensor gradInput,
at::Tensor gradOffset,
at::Tensor weight,
at::Tensor columns,
int kW,
int kH,
int dW,
int dH,
int padW,
int padH,
int dilationW,
int dilationH,
int group,
int deformable_group,
int im2col_step) {
shape_check(
input,
offset,
&gradOutput,
weight,
kH,
kW,
dH,
dW,
padH,
padW,
dilationH,
dilationW,
group,
deformable_group);
input = input.contiguous();
offset = offset.contiguous();
gradOutput = gradOutput.contiguous();
weight = weight.contiguous();
int batch = 1;
if (input.ndimension() == 3) {
// Force batch
batch = 0;
input = input.view({1, input.size(0), input.size(1), input.size(2)});
offset = offset.view({1, offset.size(0), offset.size(1), offset.size(2)});
gradOutput = gradOutput.view(
{1, gradOutput.size(0), gradOutput.size(1), gradOutput.size(2)});
}
long batchSize = input.size(0);
long nInputPlane = input.size(1);
long inputHeight = input.size(2);
long inputWidth = input.size(3);
long nOutputPlane = weight.size(0);
long outputWidth =
(inputWidth + 2 * padW - (dilationW * (kW - 1) + 1)) / dW + 1;
long outputHeight =
(inputHeight + 2 * padH - (dilationH * (kH - 1) + 1)) / dH + 1;
TORCH_CHECK((offset.size(0) == batchSize), 3, "invalid batch size of offset");
gradInput = gradInput.view({batchSize, nInputPlane, inputHeight, inputWidth});
columns = at::zeros(
{nInputPlane * kW * kH, im2col_step * outputHeight * outputWidth},
input.options());
// change order of grad output
gradOutput = gradOutput.view(
{batchSize / im2col_step,
im2col_step,
nOutputPlane,
outputHeight,
outputWidth});
gradOutput.transpose_(1, 2);
gradInput = gradInput.view(
{batchSize / im2col_step,
im2col_step,
nInputPlane,
inputHeight,
inputWidth});
input = input.view(
{batchSize / im2col_step,
im2col_step,
nInputPlane,
inputHeight,
inputWidth});
gradOffset = gradOffset.view(
{batchSize / im2col_step,
im2col_step,
deformable_group * 2 * kH * kW,
outputHeight,
outputWidth});
offset = offset.view(
{batchSize / im2col_step,
im2col_step,
deformable_group * 2 * kH * kW,
outputHeight,
outputWidth});
for (int elt = 0; elt < batchSize / im2col_step; elt++) {
// divide into groups
columns = columns.view({group, columns.size(0) / group, columns.size(1)});
weight = weight.view(
{group,
weight.size(0) / group,
weight.size(1),
weight.size(2),
weight.size(3)});
gradOutput = gradOutput.view(
{gradOutput.size(0),
group,
gradOutput.size(1) / group,
gradOutput.size(2),
gradOutput.size(3),
gradOutput.size(4)});
for (int g = 0; g < group; g++) {
columns[g] = columns[g].addmm_(
weight[g].flatten(1).transpose(0, 1),
gradOutput[elt][g].flatten(1),
0.0f,
1.0f);
}
columns =
columns.view({columns.size(0) * columns.size(1), columns.size(2)});
gradOutput = gradOutput.view(
{gradOutput.size(0),
gradOutput.size(1) * gradOutput.size(2),
gradOutput.size(3),
gradOutput.size(4),
gradOutput.size(5)});
deformable_col2im_coord(
columns,
input[elt],
offset[elt],
nInputPlane,
inputHeight,
inputWidth,
kH,
kW,
padH,
padW,
dH,
dW,
dilationH,
dilationW,
im2col_step,
deformable_group,
gradOffset[elt]);
deformable_col2im(
columns,
offset[elt],
nInputPlane,
inputHeight,
inputWidth,
kH,
kW,
padH,
padW,
dH,
dW,
dilationH,
dilationW,
im2col_step,
deformable_group,
gradInput[elt]);
}
gradOutput.transpose_(1, 2);
gradOutput =
gradOutput.view({batchSize, nOutputPlane, outputHeight, outputWidth});
gradInput = gradInput.view({batchSize, nInputPlane, inputHeight, inputWidth});
input = input.view({batchSize, nInputPlane, inputHeight, inputWidth});
gradOffset = gradOffset.view(
{batchSize, deformable_group * 2 * kH * kW, outputHeight, outputWidth});
offset = offset.view(
{batchSize, deformable_group * 2 * kH * kW, outputHeight, outputWidth});
if (batch == 0) {
gradOutput = gradOutput.view({nOutputPlane, outputHeight, outputWidth});
input = input.view({nInputPlane, inputHeight, inputWidth});
gradInput = gradInput.view({nInputPlane, inputHeight, inputWidth});
offset = offset.view({offset.size(1), offset.size(2), offset.size(3)});
gradOffset =
gradOffset.view({offset.size(1), offset.size(2), offset.size(3)});
}
return 1;
}
int deform_conv_backward_parameters_cuda(
at::Tensor input,
at::Tensor offset,
at::Tensor gradOutput,
at::Tensor gradWeight, // at::Tensor gradBias,
at::Tensor columns,
at::Tensor ones,
int kW,
int kH,
int dW,
int dH,
int padW,
int padH,
int dilationW,
int dilationH,
int group,
int deformable_group,
float scale,
int im2col_step) {
// todo: transpose and reshape outGrad
// todo: reshape columns
// todo: add im2col_step as input
shape_check(
input,
offset,
&gradOutput,
gradWeight,
kH,
kW,
dH,
dW,
padH,
padW,
dilationH,
dilationW,
group,
deformable_group);
input = input.contiguous();
offset = offset.contiguous();
gradOutput = gradOutput.contiguous();
int batch = 1;
if (input.ndimension() == 3) {
// Force batch
batch = 0;
input = input.view(
at::IntList({1, input.size(0), input.size(1), input.size(2)}));
gradOutput = gradOutput.view(
{1, gradOutput.size(0), gradOutput.size(1), gradOutput.size(2)});
}
long batchSize = input.size(0);
long nInputPlane = input.size(1);
long inputHeight = input.size(2);
long inputWidth = input.size(3);
long nOutputPlane = gradWeight.size(0);
long outputWidth =
(inputWidth + 2 * padW - (dilationW * (kW - 1) + 1)) / dW + 1;
long outputHeight =
(inputHeight + 2 * padH - (dilationH * (kH - 1) + 1)) / dH + 1;
TORCH_CHECK((offset.size(0) == batchSize), "invalid batch size of offset");
columns = at::zeros(
{nInputPlane * kW * kH, im2col_step * outputHeight * outputWidth},
input.options());
gradOutput = gradOutput.view(
{batchSize / im2col_step,
im2col_step,
nOutputPlane,
outputHeight,
outputWidth});
gradOutput.transpose_(1, 2);
at::Tensor gradOutputBuffer = at::zeros_like(gradOutput);
gradOutputBuffer = gradOutputBuffer.view(
{batchSize / im2col_step,
nOutputPlane,
im2col_step,
outputHeight,
outputWidth});
gradOutputBuffer.copy_(gradOutput);
// gradOutput is not contiguous, so we do reshape (instead of view) next
gradOutputBuffer = gradOutputBuffer.reshape(
{batchSize / im2col_step,
nOutputPlane,
im2col_step * outputHeight,
outputWidth});
gradOutput.transpose_(1, 2);
gradOutput =
gradOutput.view({batchSize, nOutputPlane, outputHeight, outputWidth});
input = input.view(
{batchSize / im2col_step,
im2col_step,
nInputPlane,
inputHeight,
inputWidth});
offset = offset.view(
{batchSize / im2col_step,
im2col_step,
deformable_group * 2 * kH * kW,
outputHeight,
outputWidth});
for (int elt = 0; elt < batchSize / im2col_step; elt++) {
deformable_im2col(
input[elt],
offset[elt],
nInputPlane,
inputHeight,
inputWidth,
kH,
kW,
padH,
padW,
dH,
dW,
dilationH,
dilationW,
im2col_step,
deformable_group,
columns);
// divide into group
gradOutputBuffer = gradOutputBuffer.view(
{gradOutputBuffer.size(0),
group,
gradOutputBuffer.size(1) / group,
gradOutputBuffer.size(2),
gradOutputBuffer.size(3)});
columns = columns.view({group, columns.size(0) / group, columns.size(1)});
gradWeight = gradWeight.view(
{group,
gradWeight.size(0) / group,
gradWeight.size(1),
gradWeight.size(2),
gradWeight.size(3)});
for (int g = 0; g < group; g++) {
gradWeight[g] = gradWeight[g]
.flatten(1)
.addmm_(
gradOutputBuffer[elt][g].flatten(1),
columns[g].transpose(1, 0),
1.0,
scale)
.view_as(gradWeight[g]);
}
gradOutputBuffer = gradOutputBuffer.view(
{gradOutputBuffer.size(0),
gradOutputBuffer.size(1) * gradOutputBuffer.size(2),
gradOutputBuffer.size(3),
gradOutputBuffer.size(4)});
columns =
columns.view({columns.size(0) * columns.size(1), columns.size(2)});
gradWeight = gradWeight.view(
{gradWeight.size(0) * gradWeight.size(1),
gradWeight.size(2),
gradWeight.size(3),
gradWeight.size(4)});
}
input = input.view({batchSize, nInputPlane, inputHeight, inputWidth});
offset = offset.view(
{batchSize, deformable_group * 2 * kH * kW, outputHeight, outputWidth});
if (batch == 0) {
gradOutput = gradOutput.view({nOutputPlane, outputHeight, outputWidth});
input = input.view({nInputPlane, inputHeight, inputWidth});
}
return 1;
}
void modulated_deform_conv_cuda_forward(
at::Tensor input,
at::Tensor weight,
at::Tensor bias,
at::Tensor ones,
at::Tensor offset,
at::Tensor mask,
at::Tensor output,
at::Tensor columns,
int kernel_h,
int kernel_w,
const int stride_h,
const int stride_w,
const int pad_h,
const int pad_w,
const int dilation_h,
const int dilation_w,
const int group,
const int deformable_group,
const bool with_bias) {
shape_check(
input,
offset,
NULL,
weight,
kernel_h,
kernel_w,
stride_h,
stride_w,
pad_h,
pad_w,
dilation_h,
dilation_w,
group,
deformable_group);
TORCH_CHECK(input.is_contiguous(), "input tensor has to be contiguous");
TORCH_CHECK(weight.is_contiguous(), "weight tensor has to be contiguous");
const int batch = input.size(0);
const int channels = input.size(1);
const int height = input.size(2);
const int width = input.size(3);
const int channels_out = weight.size(0);
const int channels_kernel = weight.size(1);
const int kernel_h_ = weight.size(2);
const int kernel_w_ = weight.size(3);
if (kernel_h_ != kernel_h || kernel_w_ != kernel_w)
AT_ERROR(
"Input shape and kernel shape wont match: (%d x %d vs %d x %d).",
kernel_h_,
kernel_w,
kernel_h_,
kernel_w_);
if (channels != channels_kernel * group)
AT_ERROR(
"Input shape and kernel channels wont match: (%d vs %d).",
channels,
channels_kernel * group);
const int height_out =
(height + 2 * pad_h - (dilation_h * (kernel_h - 1) + 1)) / stride_h + 1;
const int width_out =
(width + 2 * pad_w - (dilation_w * (kernel_w - 1) + 1)) / stride_w + 1;
// mask shape check
TORCH_CHECK(
(mask.size(2) == height_out && mask.size(3) == width_out),
"invalid spatial size of mask, expected height: %d width: %d, but "
"got height: %d width: %d",
height_out,
width_out,
mask.size(2),
mask.size(3));
TORCH_CHECK(
(mask.size(1) == deformable_group * kernel_h * kernel_w),
"invalid number of channels of mask");
if (ones.ndimension() != 2 ||
ones.size(0) * ones.size(1) < height_out * width_out) {
// Resize plane and fill with ones...
ones = at::ones({height_out, width_out}, input.options());
}
// resize output
output = output.view({batch, channels_out, height_out, width_out}).zero_();
// resize temporary columns
columns = at::zeros(
{channels * kernel_h * kernel_w, 1 * height_out * width_out},
input.options());
output = output.view(
{output.size(0),
group,
output.size(1) / group,
output.size(2),
output.size(3)});
for (int b = 0; b < batch; b++) {
modulated_deformable_im2col_cuda(
input[b],
offset[b],
mask[b],
1,
channels,
height,
width,
height_out,
width_out,
kernel_h,
kernel_w,
pad_h,
pad_w,
stride_h,
stride_w,
dilation_h,
dilation_w,
deformable_group,
columns);
// divide into group
weight = weight.view(
{group,
weight.size(0) / group,
weight.size(1),
weight.size(2),
weight.size(3)});
columns = columns.view({group, columns.size(0) / group, columns.size(1)});
for (int g = 0; g < group; g++) {
output[b][g] = output[b][g]
.flatten(1)
.addmm_(weight[g].flatten(1), columns[g])
.view_as(output[b][g]);
}
weight = weight.view(
{weight.size(0) * weight.size(1),
weight.size(2),
weight.size(3),
weight.size(4)});
columns =
columns.view({columns.size(0) * columns.size(1), columns.size(2)});
}
output = output.view(
{output.size(0),
output.size(1) * output.size(2),
output.size(3),
output.size(4)});
if (with_bias) {
output += bias.view({1, bias.size(0), 1, 1});
}
}
void modulated_deform_conv_cuda_backward(
at::Tensor input,
at::Tensor weight,
at::Tensor bias,
at::Tensor ones,
at::Tensor offset,
at::Tensor mask,
at::Tensor columns,
at::Tensor grad_input,
at::Tensor grad_weight,
at::Tensor grad_bias,
at::Tensor grad_offset,
at::Tensor grad_mask,
at::Tensor grad_output,
int kernel_h,
int kernel_w,
int stride_h,
int stride_w,
int pad_h,
int pad_w,
int dilation_h,
int dilation_w,
int group,
int deformable_group,
const bool with_bias) {
shape_check(
input,
offset,
&grad_output,
weight,
kernel_h,
kernel_w,
stride_h,
stride_w,
pad_h,
pad_w,
dilation_h,
dilation_w,
group,
deformable_group);
TORCH_CHECK(input.is_contiguous(), "input tensor has to be contiguous");
TORCH_CHECK(weight.is_contiguous(), "weight tensor has to be contiguous");
const int batch = input.size(0);
const int channels = input.size(1);
const int height = input.size(2);
const int width = input.size(3);
const int channels_kernel = weight.size(1);
const int kernel_h_ = weight.size(2);
const int kernel_w_ = weight.size(3);
if (kernel_h_ != kernel_h || kernel_w_ != kernel_w)
AT_ERROR(
"Input shape and kernel shape wont match: (%d x %d vs %d x %d).",
kernel_h_,
kernel_w,
kernel_h_,
kernel_w_);
if (channels != channels_kernel * group)
AT_ERROR(
"Input shape and kernel channels wont match: (%d vs %d).",
channels,
channels_kernel * group);
const int height_out =
(height + 2 * pad_h - (dilation_h * (kernel_h - 1) + 1)) / stride_h + 1;
const int width_out =
(width + 2 * pad_w - (dilation_w * (kernel_w - 1) + 1)) / stride_w + 1;
// mask shape check
TORCH_CHECK(
(mask.size(2) == height_out && mask.size(3) == width_out),
"invalid spatial size of mask, expected height: %d width: %d, but "
"got height: %d width: %d",
height_out,
width_out,
mask.size(2),
mask.size(3));
TORCH_CHECK(
(mask.size(1) == deformable_group * kernel_h * kernel_w),
"invalid number of channels of mask");
if (ones.ndimension() != 2 ||
ones.size(0) * ones.size(1) < height_out * width_out) {
// Resize plane and fill with ones...
ones = at::ones({height_out, width_out}, input.options());
}
grad_input = grad_input.view({batch, channels, height, width});
columns = at::zeros(
{channels * kernel_h * kernel_w, height_out * width_out},
input.options());
grad_output = grad_output.view(
{grad_output.size(0),
group,
grad_output.size(1) / group,
grad_output.size(2),
grad_output.size(3)});
for (int b = 0; b < batch; b++) {
// divide int group
columns = columns.view({group, columns.size(0) / group, columns.size(1)});
weight = weight.view(
{group,
weight.size(0) / group,
weight.size(1),
weight.size(2),
weight.size(3)});
for (int g = 0; g < group; g++) {
columns[g].addmm_(
weight[g].flatten(1).transpose(0, 1),
grad_output[b][g].flatten(1),
0.0f,
1.0f);
}
columns =
columns.view({columns.size(0) * columns.size(1), columns.size(2)});
weight = weight.view(
{weight.size(0) * weight.size(1),
weight.size(2),
weight.size(3),
weight.size(4)});
// gradient w.r.t. input coordinate data
modulated_deformable_col2im_coord_cuda(
columns,
input[b],
offset[b],
mask[b],
1,
channels,
height,
width,
height_out,
width_out,
kernel_h,
kernel_w,
pad_h,
pad_w,
stride_h,
stride_w,
dilation_h,
dilation_w,
deformable_group,
grad_offset[b],
grad_mask[b]);
// gradient w.r.t. input data
modulated_deformable_col2im_cuda(
columns,
offset[b],
mask[b],
1,
channels,
height,
width,
height_out,
width_out,
kernel_h,
kernel_w,
pad_h,
pad_w,
stride_h,
stride_w,
dilation_h,
dilation_w,
deformable_group,
grad_input[b]);
// gradient w.r.t. weight, dWeight should accumulate across the batch and
// group
modulated_deformable_im2col_cuda(
input[b],
offset[b],
mask[b],
1,
channels,
height,
width,
height_out,
width_out,
kernel_h,
kernel_w,
pad_h,
pad_w,
stride_h,
stride_w,
dilation_h,
dilation_w,
deformable_group,
columns);
columns = columns.view({group, columns.size(0) / group, columns.size(1)});
grad_weight = grad_weight.view(
{group,
grad_weight.size(0) / group,
grad_weight.size(1),
grad_weight.size(2),
grad_weight.size(3)});
if (with_bias)
grad_bias = grad_bias.view({group, grad_bias.size(0) / group});
for (int g = 0; g < group; g++) {
grad_weight[g] =
grad_weight[g]
.flatten(1)
.addmm_(grad_output[b][g].flatten(1), columns[g].transpose(0, 1))
.view_as(grad_weight[g]);
if (with_bias) {
grad_bias[g] =
grad_bias[g]
.view({-1, 1})
.addmm_(grad_output[b][g].flatten(1), ones.view({-1, 1}))
.view(-1);
}
}
columns =
columns.view({columns.size(0) * columns.size(1), columns.size(2)});
grad_weight = grad_weight.view(
{grad_weight.size(0) * grad_weight.size(1),
grad_weight.size(2),
grad_weight.size(3),
grad_weight.size(4)});
if (with_bias)
grad_bias = grad_bias.view({grad_bias.size(0) * grad_bias.size(1)});
}
grad_output = grad_output.view(
{grad_output.size(0) * grad_output.size(1),
grad_output.size(2),
grad_output.size(3),
grad_output.size(4)});
}
} // namespace detectron2