Function Classes for Identifiable Nonlinear Independent Component
Analysis
- URL: http://arxiv.org/abs/2208.06406v1
- Date: Fri, 12 Aug 2022 17:58:31 GMT
- Title: Function Classes for Identifiable Nonlinear Independent Component
Analysis
- Authors: Simon Buchholz, Michel Besserve, Bernhard Sch\"olkopf
- Abstract summary: Unsupervised learning of latent variable models (LVMs) is widely used to represent data in machine learning.
Recent work suggests that constraining the function class of such models may promote identifiability.
We prove that a subclass of these transformations, conformal maps, is identifiable and provide novel theoretical results.
- Score: 10.828616610785524
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised learning of latent variable models (LVMs) is widely used to
represent data in machine learning. When such models reflect the ground truth
factors and the mechanisms mapping them to observations, there is reason to
expect that they allow generalization in downstream tasks. It is however well
known that such identifiability guaranties are typically not achievable without
putting constraints on the model class. This is notably the case for nonlinear
Independent Component Analysis, in which the LVM maps statistically independent
variables to observations via a deterministic nonlinear function. Several
families of spurious solutions fitting perfectly the data, but that do not
correspond to the ground truth factors can be constructed in generic settings.
However, recent work suggests that constraining the function class of such
models may promote identifiability. Specifically, function classes with
constraints on their partial derivatives, gathered in the Jacobian matrix, have
been proposed, such as orthogonal coordinate transformations (OCT), which
impose orthogonality of the Jacobian columns. In the present work, we prove
that a subclass of these transformations, conformal maps, is identifiable and
provide novel theoretical results suggesting that OCTs have properties that
prevent families of spurious solutions to spoil identifiability in a generic
setting.
Related papers
- Induced Covariance for Causal Discovery in Linear Sparse Structures [55.2480439325792]
Causal models seek to unravel the cause-effect relationships among variables from observed data.
This paper introduces a novel causal discovery algorithm designed for settings in which variables exhibit linearly sparse relationships.
arXiv Detail & Related papers (2024-10-02T04:01:38Z) - Representation Disentaglement via Regularization by Causal
Identification [3.9160947065896803]
We propose the use of a causal collider structured model to describe the underlying data generative process assumptions in disentangled representation learning.
For this, we propose regularization by identification (ReI), a modular regularization engine designed to align the behavior of large scale generative models with the disentanglement constraints imposed by causal identification.
arXiv Detail & Related papers (2023-02-28T23:18:54Z) - Identifiability of latent-variable and structural-equation models: from
linear to nonlinear [2.159277717031637]
In factor analysis, non-Gaussianity of the (latent) variables has been shown to provide identifiability.
More recently, we have shown how even general non nonlinear versions of such models can be estimated.
arXiv Detail & Related papers (2023-02-06T10:21:21Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Learning from few examples with nonlinear feature maps [68.8204255655161]
We explore the phenomenon and reveal key relationships between dimensionality of AI model's feature space, non-degeneracy of data distributions, and the model's generalisation capabilities.
The main thrust of our present analysis is on the influence of nonlinear feature transformations mapping original data into higher- and possibly infinite-dimensional spaces on the resulting model's generalisation capabilities.
arXiv Detail & Related papers (2022-03-31T10:36:50Z) - Discovering Latent Causal Variables via Mechanism Sparsity: A New
Principle for Nonlinear ICA [81.4991350761909]
Independent component analysis (ICA) refers to an ensemble of methods which formalize this goal and provide estimation procedure for practical application.
We show that the latent variables can be recovered up to a permutation if one regularizes the latent mechanisms to be sparse.
arXiv Detail & Related papers (2021-07-21T14:22:14Z) - I Don't Need $\mathbf{u}$: Identifiable Non-Linear ICA Without Side
Information [13.936583337756883]
We introduce a new approach for identifiable non-linear ICA models.
In particular, we focus on generative models which perform clustering in their latent space.
arXiv Detail & Related papers (2021-06-09T17:22:08Z) - Hessian Eigenspectra of More Realistic Nonlinear Models [73.31363313577941]
We make a emphprecise characterization of the Hessian eigenspectra for a broad family of nonlinear models.
Our analysis takes a step forward to identify the origin of many striking features observed in more complex machine learning models.
arXiv Detail & Related papers (2021-03-02T06:59:52Z) - Nonlinear Invariant Risk Minimization: A Causal Approach [5.63479133344366]
We propose a learning paradigm that enables out-of-distribution generalization in the nonlinear setting.
We show identifiability of the data representation up to very simple transformations.
Extensive experiments on both synthetic and real-world datasets show that our approach significantly outperforms a variety of baseline methods.
arXiv Detail & Related papers (2021-02-24T15:38:41Z) - ICE-BeeM: Identifiable Conditional Energy-Based Deep Models Based on
Nonlinear ICA [11.919315372249802]
We consider the identifiability theory of probabilistic models.
We show that our model can be used for the estimation of the components in the framework of Independently Modulated Component Analysis.
arXiv Detail & Related papers (2020-02-26T14:43:30Z) - Invariant Feature Coding using Tensor Product Representation [75.62232699377877]
We prove that the group-invariant feature vector contains sufficient discriminative information when learning a linear classifier.
A novel feature model that explicitly consider group action is proposed for principal component analysis and k-means clustering.
arXiv Detail & Related papers (2019-06-05T07:15:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.