Truly shift-equivariant convolutional neural networks with adaptive
polyphase upsampling
- URL: http://arxiv.org/abs/2105.04040v1
- Date: Sun, 9 May 2021 22:33:53 GMT
- Title: Truly shift-equivariant convolutional neural networks with adaptive
polyphase upsampling
- Authors: Anadi Chaman and Ivan Dokmani\'c
- Abstract summary: In image classification, adaptive polyphase downsampling (APS-D) was recently proposed to make CNNs perfectly shift invariant.
We propose adaptive polyphase upsampling (APS-U), a non-linear extension of conventional upsampling, which allows CNNs to exhibit perfect shift equivariance.
- Score: 28.153820129486025
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Convolutional neural networks lack shift equivariance due to the presence of
downsampling layers. In image classification, adaptive polyphase downsampling
(APS-D) was recently proposed to make CNNs perfectly shift invariant. However,
in networks used for image reconstruction tasks, it can not by itself restore
shift equivariance. We address this problem by proposing adaptive polyphase
upsampling (APS-U), a non-linear extension of conventional upsampling, which
allows CNNs to exhibit perfect shift equivariance. With MRI and CT
reconstruction experiments, we show that networks containing APS-D/U layers
exhibit state of the art equivariance performance without sacrificing on image
reconstruction quality. In addition, unlike prior methods like data
augmentation and anti-aliasing, the gains in equivariance obtained from APS-D/U
also extend to images outside the training distribution.
Related papers
- Improving Shift Invariance in Convolutional Neural Networks with Translation Invariant Polyphase Sampling [14.731788603429774]
Downsampling operators break the shift invariance of convolutional neural networks (CNNs)
We propose a learnable pooling operator called Translation Invariant Polyphase Sampling (TIPS)
TIPS results in consistent performance gains in terms of accuracy, shift consistency, and shift fidelity.
arXiv Detail & Related papers (2024-04-11T00:49:38Z) - Deep Neural Networks with Efficient Guaranteed Invariances [77.99182201815763]
We address the problem of improving the performance and in particular the sample complexity of deep neural networks.
Group-equivariant convolutions are a popular approach to obtain equivariant representations.
We propose a multi-stream architecture, where each stream is invariant to a different transformation.
arXiv Detail & Related papers (2023-03-02T20:44:45Z) - f-DM: A Multi-stage Diffusion Model via Progressive Signal
Transformation [56.04628143914542]
Diffusion models (DMs) have recently emerged as SoTA tools for generative modeling in various domains.
We propose f-DM, a generalized family of DMs which allows progressive signal transformation.
We apply f-DM in image generation tasks with a range of functions, including down-sampling, blurring, and learned transformations.
arXiv Detail & Related papers (2022-10-10T18:49:25Z) - Learning Invariant Representations for Equivariant Neural Networks Using
Orthogonal Moments [9.680414207552722]
The convolutional layers of standard convolutional neural networks (CNNs) are equivariant to translation.
Recently, a new class of CNNs is proposed in which the conventional layers of CNNs are replaced with equivariant convolution, pooling, and batch-normalization layers.
arXiv Detail & Related papers (2022-09-22T11:48:39Z) - Improving the Sample-Complexity of Deep Classification Networks with
Invariant Integration [77.99182201815763]
Leveraging prior knowledge on intraclass variance due to transformations is a powerful method to improve the sample complexity of deep neural networks.
We propose a novel monomial selection algorithm based on pruning methods to allow an application to more complex problems.
We demonstrate the improved sample complexity on the Rotated-MNIST, SVHN and CIFAR-10 datasets.
arXiv Detail & Related papers (2022-02-08T16:16:11Z) - Sampling Equivariant Self-attention Networks for Object Detection in
Aerial Images [36.9958603490323]
Objects in aerial images have greater variations in scale and orientation than in typical images, so detection is more difficult.
We propose sampling equivariant self-attention networks which consider self-attention restricted to a local image patch.
We also use a novel randomized normalization module to tackle overfitting due to limited aerial image data.
arXiv Detail & Related papers (2021-11-05T11:48:04Z) - Truly shift-invariant convolutional neural networks [0.0]
Recent works have shown that the output of a CNN can change significantly with small shifts in input.
We propose adaptive polyphase sampling (APS), a simple sub-sampling scheme that allows convolutional neural networks to achieve 100% consistency in classification performance under shifts.
arXiv Detail & Related papers (2020-11-28T20:57:35Z) - Invariant Deep Compressible Covariance Pooling for Aerial Scene
Categorization [80.55951673479237]
We propose a novel invariant deep compressible covariance pooling (IDCCP) to solve nuisance variations in aerial scene categorization.
We conduct extensive experiments on the publicly released aerial scene image data sets and demonstrate the superiority of this method compared with state-of-the-art methods.
arXiv Detail & Related papers (2020-11-11T11:13:07Z) - Learning Invariances in Neural Networks [51.20867785006147]
We show how to parameterize a distribution over augmentations and optimize the training loss simultaneously with respect to the network parameters and augmentation parameters.
We can recover the correct set and extent of invariances on image classification, regression, segmentation, and molecular property prediction from a large space of augmentations.
arXiv Detail & Related papers (2020-10-22T17:18:48Z) - Encoding Robustness to Image Style via Adversarial Feature Perturbations [72.81911076841408]
We adapt adversarial training by directly perturbing feature statistics, rather than image pixels, to produce robust models.
Our proposed method, Adversarial Batch Normalization (AdvBN), is a single network layer that generates worst-case feature perturbations during training.
arXiv Detail & Related papers (2020-09-18T17:52:34Z) - Group Equivariant Generative Adversarial Networks [7.734726150561089]
In this work, we explicitly incorporate inductive symmetry priors into the network architectures via group-equivariant convolutional networks.
Group-convariants have higher expressive power with fewer samples and lead to better gradient feedback between generator and discriminator.
arXiv Detail & Related papers (2020-05-04T17:38:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.