Addressing the Topological Defects of Disentanglement via Distributed
Operators
- URL: http://arxiv.org/abs/2102.05623v1
- Date: Wed, 10 Feb 2021 18:34:55 GMT
- Title: Addressing the Topological Defects of Disentanglement via Distributed
Operators
- Authors: Diane Bouchacourt, Mark Ibrahim, St\'ephane Deny
- Abstract summary: A popular approach to disentanglement consists in learning to map each of these factors to distinct subspaces of a model's latent representation.
Here, we show that for a broad family of transformations acting on images, this approach introduces topological defects.
Motivated by classical results from group representation theory, we study an alternative, more flexible approach to disentanglement.
- Score: 10.29148285032989
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A core challenge in Machine Learning is to learn to disentangle natural
factors of variation in data (e.g. object shape vs. pose). A popular approach
to disentanglement consists in learning to map each of these factors to
distinct subspaces of a model's latent representation. However, this approach
has shown limited empirical success to date. Here, we show that, for a broad
family of transformations acting on images--encompassing simple affine
transformations such as rotations and translations--this approach to
disentanglement introduces topological defects (i.e. discontinuities in the
encoder). Motivated by classical results from group representation theory, we
study an alternative, more flexible approach to disentanglement which relies on
distributed latent operators, potentially acting on the entire latent space. We
theoretically and empirically demonstrate the effectiveness of this approach to
disentangle affine transformations. Our work lays a theoretical foundation for
the recent success of a new generation of models using distributed operators
for disentanglement.
Related papers
- Relative Representations: Topological and Geometric Perspectives [53.88896255693922]
Relative representations are an established approach to zero-shot model stitching.
We introduce a normalization procedure in the relative transformation, resulting in invariance to non-isotropic rescalings and permutations.
Second, we propose to deploy topological densification when fine-tuning relative representations, a topological regularization loss encouraging clustering within classes.
arXiv Detail & Related papers (2024-09-17T08:09:22Z) - Nonparametric Partial Disentanglement via Mechanism Sparsity: Sparse
Actions, Interventions and Sparse Temporal Dependencies [58.179981892921056]
This work introduces a novel principle for disentanglement we call mechanism sparsity regularization.
We propose a representation learning method that induces disentanglement by simultaneously learning the latent factors.
We show that the latent factors can be recovered by regularizing the learned causal graph to be sparse.
arXiv Detail & Related papers (2024-01-10T02:38:21Z) - Flow Factorized Representation Learning [109.51947536586677]
We introduce a generative model which specifies a distinct set of latent probability paths that define different input transformations.
We show that our model achieves higher likelihoods on standard representation learning benchmarks while simultaneously being closer to approximately equivariant models.
arXiv Detail & Related papers (2023-09-22T20:15:37Z) - Transfer learning with affine model transformation [18.13383101189326]
This paper presents a general class of transfer learning regression called affine model transfer.
It is shown that the affine model transfer broadly encompasses various existing methods, including the most common procedure based on neural feature extractors.
arXiv Detail & Related papers (2022-10-18T10:50:24Z) - Towards Robust and Adaptive Motion Forecasting: A Causal Representation
Perspective [72.55093886515824]
We introduce a causal formalism of motion forecasting, which casts the problem as a dynamic process with three groups of latent variables.
We devise a modular architecture that factorizes the representations of invariant mechanisms and style confounders to approximate a causal graph.
Experiment results on synthetic and real datasets show that our three proposed components significantly improve the robustness and reusability of the learned motion representations.
arXiv Detail & Related papers (2021-11-29T18:59:09Z) - Tensor Component Analysis for Interpreting the Latent Space of GANs [41.020230946351816]
This paper addresses the problem of finding interpretable directions in the latent space of pre-trained Generative Adversarial Networks (GANs)
Our scheme allows for both linear edits corresponding to the individual modes of the tensor, and non-linear ones that model the multiplicative interactions between them.
We show experimentally that we can utilise the former to better separate style- from geometry-based transformations, and the latter to generate an extended set of possible transformations.
arXiv Detail & Related papers (2021-11-23T09:14:39Z) - Topographic VAEs learn Equivariant Capsules [84.33745072274942]
We introduce the Topographic VAE: a novel method for efficiently training deep generative models with topographically organized latent variables.
We show that such a model indeed learns to organize its activations according to salient characteristics such as digit class, width, and style on MNIST.
We demonstrate approximate equivariance to complex transformations, expanding upon the capabilities of existing group equivariant neural networks.
arXiv Detail & Related papers (2021-09-03T09:25:57Z) - LieTransformer: Equivariant self-attention for Lie Groups [49.9625160479096]
Group equivariant neural networks are used as building blocks of group invariant neural networks.
We extend the scope of the literature to self-attention, that is emerging as a prominent building block of deep learning models.
We propose the LieTransformer, an architecture composed of LieSelfAttention layers that are equivariant to arbitrary Lie groups and their discrete subgroups.
arXiv Detail & Related papers (2020-12-20T11:02:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.