Probing the Robustness of Independent Mechanism Analysis for
Representation Learning
- URL: http://arxiv.org/abs/2207.06137v1
- Date: Wed, 13 Jul 2022 11:51:16 GMT
- Title: Probing the Robustness of Independent Mechanism Analysis for
Representation Learning
- Authors: Joanna Sliwa, Shubhangi Ghosh, Vincent Stimper, Luigi Gresele,
Bernhard Sch\"olkopf
- Abstract summary: A recently proposed approach termed Independent Mechanism Analysis (IMA) postulates that each latent source should influence the observed mixtures independently.
We show that IMA-based regularization for recovering the true sources extend to mixing functions with various degrees of violation of the IMA principle.
- Score: 9.595894646594274
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One aim of representation learning is to recover the original latent code
that generated the data, a task which requires additional information or
inductive biases. A recently proposed approach termed Independent Mechanism
Analysis (IMA) postulates that each latent source should influence the observed
mixtures independently, complementing standard nonlinear independent component
analysis, and taking inspiration from the principle of independent causal
mechanisms. While it was shown in theory and experiments that IMA helps
recovering the true latents, the method's performance was so far only
characterized when the modeling assumptions are exactly satisfied. Here, we
test the method's robustness to violations of the underlying assumptions. We
find that the benefits of IMA-based regularization for recovering the true
sources extend to mixing functions with various degrees of violation of the IMA
principle, while standard regularizers do not provide the same merits.
Moreover, we show that unregularized maximum likelihood recovers mixing
functions which systematically deviate from the IMA principle, and provide an
argument elucidating the benefits of IMA-based regularization.
Related papers
- Nonparametric Partial Disentanglement via Mechanism Sparsity: Sparse
Actions, Interventions and Sparse Temporal Dependencies [58.179981892921056]
This work introduces a novel principle for disentanglement we call mechanism sparsity regularization.
We propose a representation learning method that induces disentanglement by simultaneously learning the latent factors.
We show that the latent factors can be recovered by regularizing the learned causal graph to be sparse.
arXiv Detail & Related papers (2024-01-10T02:38:21Z) - Independent Mechanism Analysis and the Manifold Hypothesis [18.84905847863953]
Independent Mechanism Analysis (IMA) seeks to address non-identifiability in nonlinear Independent Component Analysis (ICA)
We extend IMA to settings with a larger number of mixtures that reside on a manifold embedded in a higher-dimensional than the latent space.
We prove that the IMA principle is approximately satisfied with high probability when the directions along which the latent components influence the observations are chosen independently at random.
arXiv Detail & Related papers (2023-12-20T21:29:00Z) - Unsupervised representation learning with recognition-parametrised
probabilistic models [12.865596223775649]
We introduce a new approach to probabilistic unsupervised learning based on the recognition-parametrised model ( RPM)
Under the key assumption that observations are conditionally independent given latents, the RPM combines parametric prior observation-conditioned latent distributions with non-parametric observationfactors.
The RPM provides a powerful framework to discover meaningful latent structure underlying observational data, a function critical to both animal and artificial intelligence.
arXiv Detail & Related papers (2022-09-13T00:33:21Z) - Trustworthy Multimodal Regression with Mixture of Normal-inverse Gamma
Distributions [91.63716984911278]
We introduce a novel Mixture of Normal-Inverse Gamma distributions (MoNIG) algorithm, which efficiently estimates uncertainty in principle for adaptive integration of different modalities and produces a trustworthy regression result.
Experimental results on both synthetic and different real-world data demonstrate the effectiveness and trustworthiness of our method on various multimodal regression tasks.
arXiv Detail & Related papers (2021-11-11T14:28:12Z) - Discovering Latent Causal Variables via Mechanism Sparsity: A New
Principle for Nonlinear ICA [81.4991350761909]
Independent component analysis (ICA) refers to an ensemble of methods which formalize this goal and provide estimation procedure for practical application.
We show that the latent variables can be recovered up to a permutation if one regularizes the latent mechanisms to be sparse.
arXiv Detail & Related papers (2021-07-21T14:22:14Z) - Independent mechanism analysis, a new concept? [3.2548794659022393]
Identifiability can be recovered in settings where additional, typically observed variables are included in the generative process.
We provide theoretical and empirical evidence that our approach circumvents a number of nonidentifiability issues arising in nonlinear blind source separation.
arXiv Detail & Related papers (2021-06-09T16:45:00Z) - Bayesian Uncertainty Estimation of Learned Variational MRI
Reconstruction [63.202627467245584]
We introduce a Bayesian variational framework to quantify the model-immanent (epistemic) uncertainty.
We demonstrate that our approach yields competitive results for undersampled MRI reconstruction.
arXiv Detail & Related papers (2021-02-12T18:08:14Z) - Nonlinear Independent Component Analysis for Continuous-Time Signals [85.59763606620938]
We study the classical problem of recovering a multidimensional source process from observations of mixtures of this process.
We show that this recovery is possible for many popular models of processes (up to order and monotone scaling of their coordinates) if the mixture is given by a sufficiently differentiable, invertible function.
arXiv Detail & Related papers (2021-02-04T20:28:44Z) - Learning Robust Models Using The Principle of Independent Causal
Mechanisms [26.79262903241044]
We propose a new gradient-based learning framework whose objective function is derived from the ICM principle.
We show theoretically and experimentally that neural networks trained in this framework focus on relations remaining invariant across environments.
arXiv Detail & Related papers (2020-10-14T15:38:01Z) - Distributional Robustness and Regularization in Reinforcement Learning [62.23012916708608]
We introduce a new regularizer for empirical value functions and show that it lower bounds the Wasserstein distributionally robust value function.
It suggests using regularization as a practical tool for dealing with $textitexternal uncertainty$ in reinforcement learning.
arXiv Detail & Related papers (2020-03-05T19:56:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.