Identifiable Feature Learning for Spatial Data with Nonlinear ICA
- URL: http://arxiv.org/abs/2311.16849v1
- Date: Tue, 28 Nov 2023 15:00:11 GMT
- Title: Identifiable Feature Learning for Spatial Data with Nonlinear ICA
- Authors: Hermanni H\"alv\"a and Jonathan So and Richard E. Turner and Aapo
Hyv\"arinen
- Abstract summary: We introduce a new nonlinear ICA framework that employs latent components which apply naturally to data with higher-dimensional dependency structures.
In particular, we develop a new learning and algorithm that extends variational methods to handle the combination of a deep neural network mixing function with the TP prior inducing computational efficacy.
- Score: 18.480534062833673
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, nonlinear ICA has surfaced as a popular alternative to the many
heuristic models used in deep representation learning and disentanglement. An
advantage of nonlinear ICA is that a sophisticated identifiability theory has
been developed; in particular, it has been proven that the original components
can be recovered under sufficiently strong latent dependencies. Despite this
general theory, practical nonlinear ICA algorithms have so far been mainly
limited to data with one-dimensional latent dependencies, especially
time-series data. In this paper, we introduce a new nonlinear ICA framework
that employs $t$-process (TP) latent components which apply naturally to data
with higher-dimensional dependency structures, such as spatial and
spatio-temporal data. In particular, we develop a new learning and inference
algorithm that extends variational inference methods to handle the combination
of a deep neural network mixing function with the TP prior, and employs the
method of inducing points for computational efficacy. On the theoretical side,
we show that such TP independent components are identifiable under very general
conditions. Further, Gaussian Process (GP) nonlinear ICA is established as a
limit of the TP Nonlinear ICA model, and we prove that the identifiability of
the latent components at this GP limit is more restricted. Namely, those
components are identifiable if and only if they have distinctly different
covariance kernels. Our algorithm and identifiability theorems are explored on
simulated spatial data and real world spatio-temporal data.
Related papers
- TS-CausalNN: Learning Temporal Causal Relations from Non-linear Non-stationary Time Series Data [0.42156176975445486]
We propose a Time-Series Causal Neural Network (TS-CausalNN) to discover contemporaneous and lagged causal relations simultaneously.
In addition to the simple parallel design, an advantage of the proposed model is that it naturally handles the non-stationarity and non-linearity of the data.
arXiv Detail & Related papers (2024-04-01T20:33:29Z) - Nonlinear Independent Component Analysis for Principled Disentanglement
in Unsupervised Deep Learning [2.2329417756084093]
A central problem in unsupervised deep learning is how to find useful representations of high-dimensional data, sometimes called "disentanglement"
This paper reviews the state-of-the-art of nonlinear ICA theory and algorithms.
arXiv Detail & Related papers (2023-03-29T08:51:28Z) - Learning Low Dimensional State Spaces with Overparameterized Recurrent
Neural Nets [57.06026574261203]
We provide theoretical evidence for learning low-dimensional state spaces, which can also model long-term memory.
Experiments corroborate our theory, demonstrating extrapolation via learning low-dimensional state spaces with both linear and non-linear RNNs.
arXiv Detail & Related papers (2022-10-25T14:45:15Z) - NeuralSI: Structural Parameter Identification in Nonlinear Dynamical
Systems [9.77270939559057]
This paper explores a new framework, dubbed NeuralSI, for structural identification.
Our approach seeks to estimate nonlinear parameters from governing equations.
The trained model can also be extrapolated under both standard and extreme conditions.
arXiv Detail & Related papers (2022-08-26T16:32:51Z) - A Causality-Based Learning Approach for Discovering the Underlying
Dynamics of Complex Systems from Partial Observations with Stochastic
Parameterization [1.2882319878552302]
This paper develops a new iterative learning algorithm for complex turbulent systems with partial observations.
It alternates between identifying model structures, recovering unobserved variables, and estimating parameters.
Numerical experiments show that the new algorithm succeeds in identifying the model structure and providing suitable parameterizations for many complex nonlinear systems.
arXiv Detail & Related papers (2022-08-19T00:35:03Z) - On the Identifiability of Nonlinear ICA: Sparsity and Beyond [20.644375143901488]
How to make the nonlinear ICA model identifiable up to certain trivial indeterminacies is a long-standing problem in unsupervised learning.
Recent breakthroughs reformulate the standard independence assumption of sources as conditional independence given some auxiliary variables.
We show that under specific instantiations of such constraints, the independent latent sources can be identified from their nonlinear mixtures up to a permutation.
arXiv Detail & Related papers (2022-06-15T18:24:22Z) - Consistency of mechanistic causal discovery in continuous-time using
Neural ODEs [85.7910042199734]
We consider causal discovery in continuous-time for the study of dynamical systems.
We propose a causal discovery algorithm based on penalized Neural ODEs.
arXiv Detail & Related papers (2021-05-06T08:48:02Z) - Neural ODE Processes [64.10282200111983]
We introduce Neural ODE Processes (NDPs), a new class of processes determined by a distribution over Neural ODEs.
We show that our model can successfully capture the dynamics of low-dimensional systems from just a few data-points.
arXiv Detail & Related papers (2021-03-23T09:32:06Z) - Learning Fast Approximations of Sparse Nonlinear Regression [50.00693981886832]
In this work, we bridge the gap by introducing the Threshold Learned Iterative Shrinkage Algorithming (NLISTA)
Experiments on synthetic data corroborate our theoretical results and show our method outperforms state-of-the-art methods.
arXiv Detail & Related papers (2020-10-26T11:31:08Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - WICA: nonlinear weighted ICA [72.02008296553318]
Independent Component Analysis (ICA) aims to find a coordinate system in which the components of the data are independent.
We construct a new nonlinear ICA model, called WICA, which obtains better and more stable results than other algorithms.
arXiv Detail & Related papers (2020-01-13T10:38:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.