Identifiability-Guaranteed Simplex-Structured Post-Nonlinear Mixture
Learning via Autoencoder
- URL: http://arxiv.org/abs/2106.09070v1
- Date: Wed, 16 Jun 2021 18:20:58 GMT
- Title: Identifiability-Guaranteed Simplex-Structured Post-Nonlinear Mixture
Learning via Autoencoder
- Authors: Qi Lyu and Xiao Fu
- Abstract summary: This work focuses on the problem of unraveling nonlinearly mixed latent components in an unsupervised manner.
The latent components are assumed to reside in the probability simplex, and are transformed by an unknown post-nonlinear mixing system.
This problem finds various applications in signal and data analytics, e.g., nonlinear hyperspectral unmixing, image embedding, and nonlinear clustering.
- Score: 9.769870656657522
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This work focuses on the problem of unraveling nonlinearly mixed latent
components in an unsupervised manner. The latent components are assumed to
reside in the probability simplex, and are transformed by an unknown
post-nonlinear mixing system. This problem finds various applications in signal
and data analytics, e.g., nonlinear hyperspectral unmixing, image embedding,
and nonlinear clustering. Linear mixture learning problems are already
ill-posed, as identifiability of the target latent components is hard to
establish in general. With unknown nonlinearity involved, the problem is even
more challenging. Prior work offered a function equation-based formulation for
provable latent component identification. However, the identifiability
conditions are somewhat stringent and unrealistic. In addition, the
identifiability analysis is based on the infinite sample (i.e., population)
case, while the understanding for practical finite sample cases has been
elusive. Moreover, the algorithm in the prior work trades model expressiveness
with computational convenience, which often hinders the learning performance.
Our contribution is threefold. First, new identifiability conditions are
derived under largely relaxed assumptions. Second, comprehensive sample
complexity results are presented -- which are the first of the kind. Third, a
constrained autoencoder-based algorithmic framework is proposed for
implementation, which effectively circumvents the challenges in the existing
algorithm. Synthetic and real experiments corroborate our theoretical analyses.
Related papers
- Limits and Powers of Koopman Learning [0.0]
Dynamical systems provide a comprehensive way to study complex and changing behaviors across various sciences.
Koopman operators have emerged as a dominant approach because they allow the study of nonlinear dynamics using linear techniques.
This paper addresses a fundamental open question: textitWhen can we robustly learn the spectral properties of Koopman operators from trajectory data of dynamical systems, and when can we not?
arXiv Detail & Related papers (2024-07-08T18:24:48Z) - Learning Linear Causal Representations from Interventions under General
Nonlinear Mixing [52.66151568785088]
We prove strong identifiability results given unknown single-node interventions without access to the intervention targets.
This is the first instance of causal identifiability from non-paired interventions for deep neural network embeddings.
arXiv Detail & Related papers (2023-06-04T02:32:12Z) - Provable Subspace Identification Under Post-Nonlinear Mixtures [11.012445089716016]
Untrivial mixture learning aims at identifying linearly or nonlinearly mixed latent components in a blind manner.
This work shows that under a carefully designed criterion, a null space associated with the underlying mixing system suffices to guarantee identification/removal of the unknown nonlinearity.
arXiv Detail & Related papers (2022-10-14T05:26:40Z) - On Finite-Sample Identifiability of Contrastive Learning-Based Nonlinear
Independent Component Analysis [11.012445089716016]
This work puts forth a finite-sample identifiability analysis of GCL-based nICA.
Our framework judiciously combines the properties of the GCL loss function, statistical analysis, and numerical differentiation.
arXiv Detail & Related papers (2022-06-14T04:59:08Z) - Generalization of Neural Combinatorial Solvers Through the Lens of
Adversarial Robustness [68.97830259849086]
Most datasets only capture a simpler subproblem and likely suffer from spurious features.
We study adversarial robustness - a local generalization property - to reveal hard, model-specific instances and spurious features.
Unlike in other applications, where perturbation models are designed around subjective notions of imperceptibility, our perturbation models are efficient and sound.
Surprisingly, with such perturbations, a sufficiently expressive neural solver does not suffer from the limitations of the accuracy-robustness trade-off common in supervised learning.
arXiv Detail & Related papers (2021-10-21T07:28:11Z) - Probabilistic Simplex Component Analysis [66.30587591100566]
PRISM is a probabilistic simplex component analysis approach to identifying the vertices of a data-circumscribing simplex from data.
The problem has a rich variety of applications, the most notable being hyperspectral unmixing in remote sensing and non-negative matrix factorization in machine learning.
arXiv Detail & Related papers (2021-03-18T05:39:00Z) - Comparative Analysis of Extreme Verification Latency Learning Algorithms [3.3439097577935213]
This paper is a comprehensive survey and comparative analysis of some of the EVL algorithms to point out the weaknesses and strengths of different approaches.
This work is a very first effort to provide a review of some of the existing algorithms in this field to the research community.
arXiv Detail & Related papers (2020-11-26T16:34:56Z) - Learning Fast Approximations of Sparse Nonlinear Regression [50.00693981886832]
In this work, we bridge the gap by introducing the Threshold Learned Iterative Shrinkage Algorithming (NLISTA)
Experiments on synthetic data corroborate our theoretical results and show our method outperforms state-of-the-art methods.
arXiv Detail & Related papers (2020-10-26T11:31:08Z) - Learning Mixtures of Low-Rank Models [89.39877968115833]
We study the problem of learning computational mixtures of low-rank models.
We develop an algorithm that is guaranteed to recover the unknown matrices with near-optimal sample.
In addition, the proposed algorithm is provably stable against random noise.
arXiv Detail & Related papers (2020-09-23T17:53:48Z) - Learning while Respecting Privacy and Robustness to Distributional
Uncertainties and Adversarial Data [66.78671826743884]
The distributionally robust optimization framework is considered for training a parametric model.
The objective is to endow the trained model with robustness against adversarially manipulated input data.
Proposed algorithms offer robustness with little overhead.
arXiv Detail & Related papers (2020-07-07T18:25:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.