Disentangling images with Lie group transformations and sparse coding
- URL: http://arxiv.org/abs/2012.12071v1
- Date: Fri, 11 Dec 2020 19:11:32 GMT
- Title: Disentangling images with Lie group transformations and sparse coding
- Authors: Ho Yin Chau, Frank Qiu, Yubei Chen, Bruno Olshausen
- Abstract summary: We train a model that learns to disentangle spatial patterns and their continuous transformations in a completely unsupervised manner.
Training the model on a dataset consisting of controlled geometric transformations of specific MNIST digits shows that it can recover these transformations along with the digits.
- Score: 3.3454373538792552
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Discrete spatial patterns and their continuous transformations are two
important regularities contained in natural signals. Lie groups and
representation theory are mathematical tools that have been used in previous
works to model continuous image transformations. On the other hand, sparse
coding is an important tool for learning dictionaries of patterns in natural
signals. In this paper, we combine these ideas in a Bayesian generative model
that learns to disentangle spatial patterns and their continuous
transformations in a completely unsupervised manner. Images are modeled as a
sparse superposition of shape components followed by a transformation that is
parameterized by n continuous variables. The shape components and
transformations are not predefined, but are instead adapted to learn the
symmetries in the data, with the constraint that the transformations form a
representation of an n-dimensional torus. Training the model on a dataset
consisting of controlled geometric transformations of specific MNIST digits
shows that it can recover these transformations along with the digits. Training
on the full MNIST dataset shows that it can learn both the basic digit shapes
and the natural transformations such as shearing and stretching that are
contained in this data.
Related papers
- Unsupervised Representation Learning from Sparse Transformation Analysis [79.94858534887801]
We propose to learn representations from sequence data by factorizing the transformations of the latent variables into sparse components.
Input data are first encoded as distributions of latent activations and subsequently transformed using a probability flow model.
arXiv Detail & Related papers (2024-10-07T23:53:25Z) - Current Symmetry Group Equivariant Convolution Frameworks for Representation Learning [5.802794302956837]
Euclidean deep learning is often inadequate for addressing real-world signals where the representation space is irregular and curved with complex topologies.
We focus on the importance of symmetry group equivariant deep learning models and their realization of convolution-like operations on graphs, 3D shapes, and non-Euclidean spaces.
arXiv Detail & Related papers (2024-09-11T15:07:18Z) - Oracle-Preserving Latent Flows [58.720142291102135]
We develop a methodology for the simultaneous discovery of multiple nontrivial continuous symmetries across an entire labelled dataset.
The symmetry transformations and the corresponding generators are modeled with fully connected neural networks trained with a specially constructed loss function.
The two new elements in this work are the use of a reduced-dimensionality latent space and the generalization to transformations invariant with respect to high-dimensional oracles.
arXiv Detail & Related papers (2023-02-02T00:13:32Z) - Learning Symmetric Embeddings for Equivariant World Models [9.781637768189158]
We propose learning symmetric embedding networks (SENs) that encode an input space (e.g. images)
This network can be trained end-to-end with an equivariant task network to learn an explicitly symmetric representation.
Our experiments demonstrate that SENs facilitate the application of equivariant networks to data with complex symmetry representations.
arXiv Detail & Related papers (2022-04-24T22:31:52Z) - Disentangling Patterns and Transformations from One Sequence of Images
with Shape-invariant Lie Group Transformer [0.0]
We take a novel approach for representation learning based on a simpler and more intuitive formulation that the observed world is the combination of multiple independent patterns and transformations.
We propose a model that disentangles the scenes into the minimum number of basic components of patterns and Lie transformations from only one sequence of images.
arXiv Detail & Related papers (2022-03-21T11:55:13Z) - Revisiting Transformation Invariant Geometric Deep Learning: Are Initial
Representations All You Need? [80.86819657126041]
We show that transformation-invariant and distance-preserving initial representations are sufficient to achieve transformation invariance.
Specifically, we realize transformation-invariant and distance-preserving initial point representations by modifying multi-dimensional scaling.
We prove that TinvNN can strictly guarantee transformation invariance, being general and flexible enough to be combined with the existing neural networks.
arXiv Detail & Related papers (2021-12-23T03:52:33Z) - Frame Averaging for Equivariant Shape Space Learning [85.42901997467754]
A natural way to incorporate symmetries in shape space learning is to ask that the mapping to the shape space (encoder) and mapping from the shape space (decoder) are equivariant to the relevant symmetries.
We present a framework for incorporating equivariance in encoders and decoders by introducing two contributions.
arXiv Detail & Related papers (2021-12-03T06:41:19Z) - Topographic VAEs learn Equivariant Capsules [84.33745072274942]
We introduce the Topographic VAE: a novel method for efficiently training deep generative models with topographically organized latent variables.
We show that such a model indeed learns to organize its activations according to salient characteristics such as digit class, width, and style on MNIST.
We demonstrate approximate equivariance to complex transformations, expanding upon the capabilities of existing group equivariant neural networks.
arXiv Detail & Related papers (2021-09-03T09:25:57Z) - Augmenting Implicit Neural Shape Representations with Explicit
Deformation Fields [95.39603371087921]
Implicit neural representation is a recent approach to learn shape collections as zero level-sets of neural networks.
We advocate deformation-aware regularization for implicit neural representations, aiming at producing plausible deformations as latent code changes.
arXiv Detail & Related papers (2021-08-19T22:07:08Z) - Robust Training Using Natural Transformation [19.455666609149567]
We present NaTra, an adversarial training scheme to improve robustness of image classification algorithms.
We target attributes of the input images that are independent of the class identification, and manipulate those attributes to mimic real-world natural transformations.
We demonstrate the efficacy of our scheme by utilizing the disentangled latent representations derived from well-trained GANs.
arXiv Detail & Related papers (2021-05-10T01:56:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.