Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Harmonic Networks: Deep Translation and Rotation Equivariance

About

Translating or rotating an input image should not affect the results of many computer vision tasks. Convolutional neural networks (CNNs) are already translation equivariant: input image translations produce proportionate feature map translations. This is not the case for rotations. Global rotation equivariance is typically sought through data augmentation, but patch-wise equivariance is more difficult. We present Harmonic Networks or H-Nets, a CNN exhibiting equivariance to patch-wise translation and 360-rotation. We achieve this by replacing regular CNN filters with circular harmonics, returning a maximal response and orientation for every receptive field patch. H-Nets use a rich, parameter-efficient and low computational complexity representation, and we show that deep feature maps within the network encode complicated rotational invariants. We demonstrate that our layers are general enough to be used in conjunction with the latest architectures and techniques, such as deep supervision and batch normalization. We also achieve state-of-the-art classification on rotated-MNIST, and competitive results on other benchmark challenges.

Daniel E. Worrall, Stephan J. Garbin, Daniyar Turmukhambetov, Gabriel J. Brostow• 2016

Related benchmarks

TaskDatasetResultRank
Image ClassificationMNIST rotated (test)
Test Error (%)1.69
105
ClassificationRotMNIST (test)
Classification Accuracy92.44
32
Image ClassificationMNIST original (test)--
20
Tile-level classificationPCam (test)
AUC0.939
19
Image ClassificationSIM2MNIST (test)
Error Rate9.28
5
Showing 5 of 5 rows

Other info

Follow for update