Group Equivariant Subsampling
- URL: http://arxiv.org/abs/2106.05886v1
- Date: Thu, 10 Jun 2021 16:14:00 GMT
- Title: Group Equivariant Subsampling
- Authors: Jin Xu, Hyunjik Kim, Tom Rainforth, Yee Whye Teh
- Abstract summary: Subsampling is used in convolutional neural networks (CNNs) in the form of pooling or strided convolutions.
We first introduce translation equivariant subsampling/upsampling layers that can be used to construct exact translation equivariant CNNs.
We then generalise these layers beyond translations to general groups, thus proposing group equivariant subsampling/upsampling.
- Score: 60.53371517247382
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Subsampling is used in convolutional neural networks (CNNs) in the form of
pooling or strided convolutions, to reduce the spatial dimensions of feature
maps and to allow the receptive fields to grow exponentially with depth.
However, it is known that such subsampling operations are not translation
equivariant, unlike convolutions that are translation equivariant. Here, we
first introduce translation equivariant subsampling/upsampling layers that can
be used to construct exact translation equivariant CNNs. We then generalise
these layers beyond translations to general groups, thus proposing group
equivariant subsampling/upsampling. We use these layers to construct group
equivariant autoencoders (GAEs) that allow us to learn low-dimensional
equivariant representations. We empirically verify on images that the
representations are indeed equivariant to input translations and rotations, and
thus generalise well to unseen positions and orientations. We further use GAEs
in models that learn object-centric representations on multi-object datasets,
and show improved data efficiency and decomposition compared to non-equivariant
baselines.
Related papers
- Restore Translation Using Equivariant Neural Networks [7.78895108256899]
In this paper, we propose a pre-classifier restorer to recover translated (or even rotated) inputs to a convolutional neural network.
The restorer is based on a theoretical result which gives a sufficient and necessary condition for an affine operator to be translational equivariant on a tensor space.
arXiv Detail & Related papers (2023-06-29T13:34:35Z) - Investigating how ReLU-networks encode symmetries [13.935148870831396]
We investigate whether equivariance of a network implies that all layers are equivariant.
We conjecture that CNNs trained to be equivariant will exhibit layerwise equivariance.
We show that it is typically easier to merge a network with a group-transformed version of itself than merging two different networks.
arXiv Detail & Related papers (2023-05-26T15:23:20Z) - Deep Neural Networks with Efficient Guaranteed Invariances [77.99182201815763]
We address the problem of improving the performance and in particular the sample complexity of deep neural networks.
Group-equivariant convolutions are a popular approach to obtain equivariant representations.
We propose a multi-stream architecture, where each stream is invariant to a different transformation.
arXiv Detail & Related papers (2023-03-02T20:44:45Z) - Scale Equivariant U-Net [0.0]
This paper introduces the Scale Equivariant U-Net (SEU-Net), a U-Net that is made approximately equivariant to a semigroup of scales and translations.
The proposed SEU-Net is trained for semantic segmentation of the Oxford Pet IIIT and the DIC-C2DH-HeLa dataset for cell segmentation.
The generalization metric to unseen scales is dramatically improved in comparison to the U-Net, even when the U-Net is trained with scale jittering.
arXiv Detail & Related papers (2022-10-10T09:19:40Z) - Learning Invariant Representations for Equivariant Neural Networks Using
Orthogonal Moments [9.680414207552722]
The convolutional layers of standard convolutional neural networks (CNNs) are equivariant to translation.
Recently, a new class of CNNs is proposed in which the conventional layers of CNNs are replaced with equivariant convolution, pooling, and batch-normalization layers.
arXiv Detail & Related papers (2022-09-22T11:48:39Z) - Improving the Sample-Complexity of Deep Classification Networks with
Invariant Integration [77.99182201815763]
Leveraging prior knowledge on intraclass variance due to transformations is a powerful method to improve the sample complexity of deep neural networks.
We propose a novel monomial selection algorithm based on pruning methods to allow an application to more complex problems.
We demonstrate the improved sample complexity on the Rotated-MNIST, SVHN and CIFAR-10 datasets.
arXiv Detail & Related papers (2022-02-08T16:16:11Z) - Group Equivariant Conditional Neural Processes [30.134634059773703]
We present the group equivariant conditional neural process (EquivCNP)
We show that EquivCNP achieves comparable performance to conventional conditional neural processes in a 1D regression task.
arXiv Detail & Related papers (2021-02-17T13:50:07Z) - LieTransformer: Equivariant self-attention for Lie Groups [49.9625160479096]
Group equivariant neural networks are used as building blocks of group invariant neural networks.
We extend the scope of the literature to self-attention, that is emerging as a prominent building block of deep learning models.
We propose the LieTransformer, an architecture composed of LieSelfAttention layers that are equivariant to arbitrary Lie groups and their discrete subgroups.
arXiv Detail & Related papers (2020-12-20T11:02:49Z) - Invariant Deep Compressible Covariance Pooling for Aerial Scene
Categorization [80.55951673479237]
We propose a novel invariant deep compressible covariance pooling (IDCCP) to solve nuisance variations in aerial scene categorization.
We conduct extensive experiments on the publicly released aerial scene image data sets and demonstrate the superiority of this method compared with state-of-the-art methods.
arXiv Detail & Related papers (2020-11-11T11:13:07Z) - Generalizing Convolutional Neural Networks for Equivariance to Lie
Groups on Arbitrary Continuous Data [52.78581260260455]
We propose a general method to construct a convolutional layer that is equivariant to transformations from any specified Lie group.
We apply the same model architecture to images, ball-and-stick molecular data, and Hamiltonian dynamical systems.
arXiv Detail & Related papers (2020-02-25T17:40:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.