AI Giving Back to Statistics? Discovery of the Coordinate System of
Univariate Distributions by Beta Variational Autoencoder
- URL: http://arxiv.org/abs/2004.02687v1
- Date: Mon, 6 Apr 2020 14:11:13 GMT
- Title: AI Giving Back to Statistics? Discovery of the Coordinate System of
Univariate Distributions by Beta Variational Autoencoder
- Authors: Alex Glushkovsky
- Abstract summary: The article discusses experiences of training neural networks to classify univariate empirical distributions and to represent them on the two-dimensional latent space forcing disentanglement based on the inputs of cumulative distribution functions (CDF)
The representation on the latent two-dimensional coordinate system can be seen as an additional metadata of the real-world data that disentangles important distribution characteristics, such as shape of the CDF, classification probabilities of underlying theoretical distributions and their parameters, information entropy, and skewness.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Distributions are fundamental statistical elements that play essential
theoretical and practical roles. The article discusses experiences of training
neural networks to classify univariate empirical distributions and to represent
them on the two-dimensional latent space forcing disentanglement based on the
inputs of cumulative distribution functions (CDF). The latent space
representation has been performed using an unsupervised beta variational
autoencoder (beta-VAE). It separates distributions of different shapes while
overlapping similar ones and empirically realises relationships between
distributions that are known theoretically. The synthetic experiment of
generated univariate continuous and discrete (Bernoulli) distributions with
varying sample sizes and parameters has been performed to support the study.
The representation on the latent two-dimensional coordinate system can be seen
as an additional metadata of the real-world data that disentangles important
distribution characteristics, such as shape of the CDF, classification
probabilities of underlying theoretical distributions and their parameters,
information entropy, and skewness. Entropy changes, providing an "arrow of
time", determine dynamic trajectories along representations of distributions on
the latent space. In addition, post beta-VAE unsupervised segmentation of the
latent space based on weight-of-evidence (WOE) of posterior versus standard
isotopic two-dimensional normal densities has been applied detecting the
presence of assignable causes that distinguish exceptional CDF inputs.
Related papers
- Hessian-Informed Flow Matching [4.542719108171107]
Hessian-Informed Flow Matching is a novel approach that integrates the Hessian of an energy function into conditional flows.
This integration allows HI-FM to account for local curvature and anisotropic covariance structures.
Empirical evaluations on the MNIST and Lennard-Jones particles datasets demonstrate that HI-FM improves the likelihood of test samples.
arXiv Detail & Related papers (2024-10-15T09:34:52Z) - Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis [56.442307356162864]
We study the theoretical aspects of score-based discrete diffusion models under the Continuous Time Markov Chain (CTMC) framework.
We introduce a discrete-time sampling algorithm in the general state space $[S]d$ that utilizes score estimators at predefined time points.
Our convergence analysis employs a Girsanov-based method and establishes key properties of the discrete score function.
arXiv Detail & Related papers (2024-10-03T09:07:13Z) - Theoretical Insights for Diffusion Guidance: A Case Study for Gaussian
Mixture Models [59.331993845831946]
Diffusion models benefit from instillation of task-specific information into the score function to steer the sample generation towards desired properties.
This paper provides the first theoretical study towards understanding the influence of guidance on diffusion models in the context of Gaussian mixture models.
arXiv Detail & Related papers (2024-03-03T23:15:48Z) - Score Approximation, Estimation and Distribution Recovery of Diffusion
Models on Low-Dimensional Data [68.62134204367668]
This paper studies score approximation, estimation, and distribution recovery of diffusion models, when data are supported on an unknown low-dimensional linear subspace.
We show that with a properly chosen neural network architecture, the score function can be both accurately approximated and efficiently estimated.
The generated distribution based on the estimated score function captures the data geometric structures and converges to a close vicinity of the data distribution.
arXiv Detail & Related papers (2023-02-14T17:02:35Z) - Flexible Amortized Variational Inference in qBOLD MRI [56.4324135502282]
Oxygen extraction fraction (OEF) and deoxygenated blood volume (DBV) are more ambiguously determined from the data.
Existing inference methods tend to yield very noisy and underestimated OEF maps, while overestimating DBV.
This work describes a novel probabilistic machine learning approach that can infer plausible distributions of OEF and DBV.
arXiv Detail & Related papers (2022-03-11T10:47:16Z) - Structure-preserving GANs [6.438897276587413]
We introduce structure-preserving GANs as a data-efficient framework for learning distributions.
We show that we can reduce the discriminator space to its projection on the invariant discriminator space.
We contextualize our framework by building symmetry-preserving GANs for distributions with intrinsic group symmetry.
arXiv Detail & Related papers (2022-02-02T16:40:04Z) - Cycle Consistent Probability Divergences Across Different Spaces [38.43511529063335]
Discrepancy measures between probability distributions are at the core of statistical inference and machine learning.
This work proposes a novel unbalanced Monge optimal transport formulation for matching, up to isometries, distributions on different spaces.
arXiv Detail & Related papers (2021-11-22T16:35:58Z) - Instrumental Variable-Driven Domain Generalization with Unobserved
Confounders [53.735614014067394]
Domain generalization (DG) aims to learn from multiple source domains a model that can generalize well on unseen target domains.
We propose an instrumental variable-driven DG method (IV-DG) by removing the bias of the unobserved confounders with two-stage learning.
In the first stage, it learns the conditional distribution of the input features of one domain given input features of another domain.
In the second stage, it estimates the relationship by predicting labels with the learned conditional distribution.
arXiv Detail & Related papers (2021-10-04T13:32:57Z) - Inverse Estimation of Elastic Modulus Using Physics-Informed Generative
Adversarial Networks [0.0]
Agenerative adversarial networks (GANs) encode physical laws in the form of partial differential equations (PDEs)
In this work, PI-GANs are demonstrated for the application of elastic modulus estimation in mechanical testing.
Two feed-forward deep neural network generators used to model the deformation and material stiffness across a two dimensional domain.
arXiv Detail & Related papers (2020-05-20T20:14:10Z) - Few-shot Domain Adaptation by Causal Mechanism Transfer [107.08605582020866]
We study few-shot supervised domain adaptation (DA) for regression problems, where only a few labeled target domain data and many labeled source domain data are available.
Many of the current DA methods base their transfer assumptions on either parametrized distribution shift or apparent distribution similarities.
We propose mechanism transfer, a meta-distributional scenario in which a data generating mechanism is invariant among domains.
arXiv Detail & Related papers (2020-02-10T02:16:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.