Quantised Transforming Auto-Encoders: Achieving Equivariance to
Arbitrary Transformations in Deep Networks
- URL: http://arxiv.org/abs/2111.12873v1
- Date: Thu, 25 Nov 2021 02:26:38 GMT
- Title: Quantised Transforming Auto-Encoders: Achieving Equivariance to
Arbitrary Transformations in Deep Networks
- Authors: Jianbo Jiao and Jo\~ao F. Henriques
- Abstract summary: Convolutional Neural Networks (CNNs) are equivariant to image translation.
We propose an auto-encoder architecture whose embedding obeys an arbitrary set of equivariance relations simultaneously.
We demonstrate results of successful re-rendering of transformed versions of input images on several datasets.
- Score: 23.673155102696338
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work we investigate how to achieve equivariance to input
transformations in deep networks, purely from data, without being given a model
of those transformations. Convolutional Neural Networks (CNNs), for example,
are equivariant to image translation, a transformation that can be easily
modelled (by shifting the pixels vertically or horizontally). Other
transformations, such as out-of-plane rotations, do not admit a simple analytic
model. We propose an auto-encoder architecture whose embedding obeys an
arbitrary set of equivariance relations simultaneously, such as translation,
rotation, colour changes, and many others. This means that it can take an input
image, and produce versions transformed by a given amount that were not
observed before (e.g. a different point of view of the same object, or a colour
variation). Despite extending to many (even non-geometric) transformations, our
model reduces exactly to a CNN in the special case of translation-equivariance.
Equivariances are important for the interpretability and robustness of deep
networks, and we demonstrate results of successful re-rendering of transformed
versions of input images on several synthetic and real datasets, as well as
results on object pose estimation.
Related papers
- Revisiting Data Augmentation for Rotational Invariance in Convolutional
Neural Networks [0.29127054707887967]
We investigate how best to include rotational invariance in a CNN for image classification.
Our experiments show that networks trained with data augmentation alone can classify rotated images nearly as well as in the normal unrotated case.
arXiv Detail & Related papers (2023-10-12T15:53:24Z) - Self-Supervised Learning for Group Equivariant Neural Networks [75.62232699377877]
Group equivariant neural networks are the models whose structure is restricted to commute with the transformations on the input.
We propose two concepts for self-supervised tasks: equivariant pretext labels and invariant contrastive loss.
Experiments on standard image recognition benchmarks demonstrate that the equivariant neural networks exploit the proposed self-supervised tasks.
arXiv Detail & Related papers (2023-03-08T08:11:26Z) - Deep Neural Networks with Efficient Guaranteed Invariances [77.99182201815763]
We address the problem of improving the performance and in particular the sample complexity of deep neural networks.
Group-equivariant convolutions are a popular approach to obtain equivariant representations.
We propose a multi-stream architecture, where each stream is invariant to a different transformation.
arXiv Detail & Related papers (2023-03-02T20:44:45Z) - The Lie Derivative for Measuring Learned Equivariance [84.29366874540217]
We study the equivariance properties of hundreds of pretrained models, spanning CNNs, transformers, and Mixer architectures.
We find that many violations of equivariance can be linked to spatial aliasing in ubiquitous network layers, such as pointwise non-linearities.
For example, transformers can be more equivariant than convolutional neural networks after training.
arXiv Detail & Related papers (2022-10-06T15:20:55Z) - Roto-Translation Equivariant Super-Resolution of Two-Dimensional Flows
Using Convolutional Neural Networks [0.15229257192293202]
Convolutional neural networks (CNNs) often process vectors as quantities having no direction like colors in images.
This study investigates the effect of treating vectors as geometrical objects in terms of super-resolution of velocity on two-dimensional fluids.
arXiv Detail & Related papers (2022-02-22T07:07:07Z) - Revisiting Transformation Invariant Geometric Deep Learning: Are Initial
Representations All You Need? [80.86819657126041]
We show that transformation-invariant and distance-preserving initial representations are sufficient to achieve transformation invariance.
Specifically, we realize transformation-invariant and distance-preserving initial point representations by modifying multi-dimensional scaling.
We prove that TinvNN can strictly guarantee transformation invariance, being general and flexible enough to be combined with the existing neural networks.
arXiv Detail & Related papers (2021-12-23T03:52:33Z) - Training or Architecture? How to Incorporate Invariance in Neural
Networks [14.162739081163444]
We propose a method for provably invariant network architectures with respect to group actions.
In a nutshell, we intend to 'undo' any possible transformation before feeding the data into the actual network.
We analyze properties of such approaches, extend them to equivariant networks, and demonstrate their advantages in terms of robustness as well as computational efficiency in several numerical examples.
arXiv Detail & Related papers (2021-06-18T10:31:00Z) - Grounding inductive biases in natural images:invariance stems from
variations in data [20.432568247732206]
We study the factors of variation in a real dataset, ImageNet.
We show standard augmentation relies on a precise combination of translation and scale.
We find that the main factors of variation in ImageNet mostly relate to appearance.
arXiv Detail & Related papers (2021-06-09T14:58:57Z) - What Does CNN Shift Invariance Look Like? A Visualization Study [87.79405274610681]
Feature extraction with convolutional neural networks (CNNs) is a popular method to represent images for machine learning tasks.
We focus on measuring and visualizing the shift invariance of extracted features from popular off-the-shelf CNN models.
We conclude that features extracted from popular networks are not globally invariant, and that biases and artifacts exist within this variance.
arXiv Detail & Related papers (2020-11-09T01:16:30Z) - Generalizing Convolutional Neural Networks for Equivariance to Lie
Groups on Arbitrary Continuous Data [52.78581260260455]
We propose a general method to construct a convolutional layer that is equivariant to transformations from any specified Lie group.
We apply the same model architecture to images, ball-and-stick molecular data, and Hamiltonian dynamical systems.
arXiv Detail & Related papers (2020-02-25T17:40:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.