Binary Independent Component Analysis via Non-stationarity
- URL: http://arxiv.org/abs/2111.15431v1
- Date: Tue, 30 Nov 2021 14:23:53 GMT
- Title: Binary Independent Component Analysis via Non-stationarity
- Authors: Antti Hyttinen, Vit\'oria Barin-Pacela, Aapo Hyv\"arinen
- Abstract summary: We consider independent component analysis of binary data.
We start by assuming a linear mixing model in a continuous-valued latent space, followed by a binary observation model.
In stark contrast to the continuous-valued case, we prove non-identifiability of the model with few observed variables.
- Score: 7.283533791778359
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider independent component analysis of binary data. While fundamental
in practice, this case has been much less developed than ICA for continuous
data. We start by assuming a linear mixing model in a continuous-valued latent
space, followed by a binary observation model. Importantly, we assume that the
sources are non-stationary; this is necessary since any non-Gaussianity would
essentially be destroyed by the binarization. Interestingly, the model allows
for closed-form likelihood by employing the cumulative distribution function of
the multivariate Gaussian distribution. In stark contrast to the
continuous-valued case, we prove non-identifiability of the model with few
observed variables; our empirical results imply identifiability when the number
of observed variables is higher. We present a practical method for binary ICA
that uses only pairwise marginals, which are faster to compute than the full
multivariate likelihood.
Related papers
- Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis [56.442307356162864]
We study the theoretical aspects of score-based discrete diffusion models under the Continuous Time Markov Chain (CTMC) framework.
We introduce a discrete-time sampling algorithm in the general state space $[S]d$ that utilizes score estimators at predefined time points.
Our convergence analysis employs a Girsanov-based method and establishes key properties of the discrete score function.
arXiv Detail & Related papers (2024-10-03T09:07:13Z) - Detection of Unobserved Common Causes based on NML Code in Discrete,
Mixed, and Continuous Variables [1.5039745292757667]
We categorize all possible causal relationships between two random variables into the following four categories.
We show that CLOUD is more effective than existing methods in inferring causal relationships by extensive experiments on both synthetic and real-world data.
arXiv Detail & Related papers (2024-03-11T08:11:52Z) - Towards Faster Non-Asymptotic Convergence for Diffusion-Based Generative
Models [49.81937966106691]
We develop a suite of non-asymptotic theory towards understanding the data generation process of diffusion models.
In contrast to prior works, our theory is developed based on an elementary yet versatile non-asymptotic approach.
arXiv Detail & Related papers (2023-06-15T16:30:08Z) - Score-based Continuous-time Discrete Diffusion Models [102.65769839899315]
We extend diffusion models to discrete variables by introducing a Markov jump process where the reverse process denoises via a continuous-time Markov chain.
We show that an unbiased estimator can be obtained via simple matching the conditional marginal distributions.
We demonstrate the effectiveness of the proposed method on a set of synthetic and real-world music and image benchmarks.
arXiv Detail & Related papers (2022-11-30T05:33:29Z) - Bayesian Nonlocal Operator Regression (BNOR): A Data-Driven Learning
Framework of Nonlocal Models with Uncertainty Quantification [4.705624984585247]
We consider the problem of modeling heterogeneous materials where micro-scale dynamics and interactions affect global behavior.
We develop a Bayesian framework for uncertainty (UQ) in material response prediction when using nonlocal models.
This work is a first step towards statistical characterization of nonlocal model discrepancy in the context of homogenization.
arXiv Detail & Related papers (2022-10-06T22:37:59Z) - On the detrimental effect of invariances in the likelihood for
variational inference [21.912271882110986]
Variational Bayesian posterior inference often requires simplifying approximations such as mean-field parametrisation to ensure tractability.
Prior work has associated the variational mean-field approximation for Bayesian neural networks with underfitting in the case of small datasets or large model sizes.
arXiv Detail & Related papers (2022-09-15T09:13:30Z) - On the Strong Correlation Between Model Invariance and Generalization [54.812786542023325]
Generalization captures a model's ability to classify unseen data.
Invariance measures consistency of model predictions on transformations of the data.
From a dataset-centric view, we find a certain model's accuracy and invariance linearly correlated on different test sets.
arXiv Detail & Related papers (2022-07-14T17:08:25Z) - Nonparametric Conditional Local Independence Testing [69.31200003384122]
Conditional local independence is an independence relation among continuous time processes.
No nonparametric test of conditional local independence has been available.
We propose such a nonparametric test based on double machine learning.
arXiv Detail & Related papers (2022-03-25T10:31:02Z) - Exploiting Independent Instruments: Identification and Distribution
Generalization [3.701112941066256]
We exploit the independence for distribution generalization by taking into account higher moments.
We prove that the proposed estimator is invariant to distributional shifts on the instruments.
These results hold even in the under-identified case where the instruments are not sufficiently rich to identify the causal function.
arXiv Detail & Related papers (2022-02-03T21:49:04Z) - Nonlinear Independent Component Analysis for Continuous-Time Signals [85.59763606620938]
We study the classical problem of recovering a multidimensional source process from observations of mixtures of this process.
We show that this recovery is possible for many popular models of processes (up to order and monotone scaling of their coordinates) if the mixture is given by a sufficiently differentiable, invertible function.
arXiv Detail & Related papers (2021-02-04T20:28:44Z) - Binary Classification of Gaussian Mixtures: Abundance of Support
Vectors, Benign Overfitting and Regularization [39.35822033674126]
We study binary linear classification under a generative Gaussian mixture model.
We derive novel non-asymptotic bounds on the classification error of the latter.
Our results extend to a noisy model with constant probability noise flips.
arXiv Detail & Related papers (2020-11-18T07:59:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.