Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Making Convolutional Networks Shift-Invariant Again

About

Modern convolutional networks are not shift-invariant, as small input shifts or translations can cause drastic changes in the output. Commonly used downsampling methods, such as max-pooling, strided-convolution, and average-pooling, ignore the sampling theorem. The well-known signal processing fix is anti-aliasing by low-pass filtering before downsampling. However, simply inserting this module into deep networks degrades performance; as a result, it is seldomly used today. We show that when integrated correctly, it is compatible with existing architectural components, such as max-pooling and strided-convolution. We observe \textit{increased accuracy} in ImageNet classification, across several commonly-used architectures, such as ResNet, DenseNet, and MobileNet, indicating effective regularization. Furthermore, we observe \textit{better generalization}, in terms of stability and robustness to input corruptions. Our results demonstrate that this classical signal processing technique has been undeservingly overlooked in modern deep networks. Code and anti-aliased versions of popular networks are available at https://richzhang.github.io/antialiased-cnns/ .

Richard Zhang• 2019

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet (val)--
1206
Image ClassificationImageNet 1k (test)
Top-1 Accuracy79.3
798
Image ClassificationImageNet-A (test)
Top-1 Acc8.2
154
Image ClassificationImageNet-C (test)
mCE (Mean Corruption Error)62.6
110
Image ClassificationImageNet-R (test)--
105
Image ClassificationCIFAR-10-C (test)--
61
Robustness to CorruptionsImageNet-C (test)
mCE68.1
56
Image ClassificationImageNet-SK (test)
Top-1 Accuracy29.6
34
Adversarial RobustnessImageNet 1k (test)
FGSM Robustness32.9
34
RobustnessImageNet-C
mCE73.4
11
Showing 10 of 12 rows

Other info

Follow for update