Adversarial sampling of unknown and high-dimensional conditional
distributions
- URL: http://arxiv.org/abs/2111.05962v1
- Date: Mon, 8 Nov 2021 12:23:38 GMT
- Title: Adversarial sampling of unknown and high-dimensional conditional
distributions
- Authors: Malik Hassanaly, Andrew Glaws, Karen Stengel, Ryan N. King
- Abstract summary: In this paper the sampling method, as well as the inference of the underlying distribution, are handled with a data-driven method known as generative adversarial networks (GAN)
GAN trains two competing neural networks to produce a network that can effectively generate samples from the training set distribution.
It is shown that all the versions of the proposed algorithm effectively sample the target conditional distribution with minimal impact on the quality of the samples.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many engineering problems require the prediction of
realization-to-realization variability or a refined description of modeled
quantities. In that case, it is necessary to sample elements from unknown
high-dimensional spaces with possibly millions of degrees of freedom. While
there exist methods able to sample elements from probability density functions
(PDF) with known shapes, several approximations need to be made when the
distribution is unknown. In this paper the sampling method, as well as the
inference of the underlying distribution, are both handled with a data-driven
method known as generative adversarial networks (GAN), which trains two
competing neural networks to produce a network that can effectively generate
samples from the training set distribution. In practice, it is often necessary
to draw samples from conditional distributions. When the conditional variables
are continuous, only one (if any) data point corresponding to a particular
value of a conditioning variable may be available, which is not sufficient to
estimate the conditional distribution. This work handles this problem using an
a priori estimation of the conditional moments of a PDF. Two approaches,
stochastic estimation, and an external neural network are compared here for
computing these moments; however, any preferred method can be used. The
algorithm is demonstrated in the case of the deconvolution of a filtered
turbulent flow field. It is shown that all the versions of the proposed
algorithm effectively sample the target conditional distribution with minimal
impact on the quality of the samples compared to state-of-the-art methods.
Additionally, the procedure can be used as a metric for the diversity of
samples generated by a conditional GAN (cGAN) conditioned with continuous
variables.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Schr\"odinger bridge based deep conditional generative learning [0.0]
We introduce a novel Schr"odinger bridge based deep generative method for learning conditional distributions.
We apply our method to both low-dimensional and high-dimensional conditional generation problems.
arXiv Detail & Related papers (2024-09-25T19:08:13Z) - Generative Conditional Distributions by Neural (Entropic) Optimal Transport [12.152228552335798]
We introduce a novel neural entropic optimal transport method designed to learn generative models of conditional distributions.
Our method relies on the minimax training of two neural networks.
Our experiments on real-world datasets show the effectiveness of our algorithm compared to state-of-the-art conditional distribution learning techniques.
arXiv Detail & Related papers (2024-06-04T13:45:35Z) - User-defined Event Sampling and Uncertainty Quantification in Diffusion
Models for Physical Dynamical Systems [49.75149094527068]
We show that diffusion models can be adapted to make predictions and provide uncertainty quantification for chaotic dynamical systems.
We develop a probabilistic approximation scheme for the conditional score function which converges to the true distribution as the noise level decreases.
We are able to sample conditionally on nonlinear userdefined events at inference time, and matches data statistics even when sampling from the tails of the distribution.
arXiv Detail & Related papers (2023-06-13T03:42:03Z) - Unsupervised Learning of Sampling Distributions for Particle Filters [80.6716888175925]
We put forward four methods for learning sampling distributions from observed measurements.
Experiments demonstrate that learned sampling distributions exhibit better performance than designed, minimum-degeneracy sampling distributions.
arXiv Detail & Related papers (2023-02-02T15:50:21Z) - DensePure: Understanding Diffusion Models towards Adversarial Robustness [110.84015494617528]
We analyze the properties of diffusion models and establish the conditions under which they can enhance certified robustness.
We propose a new method DensePure, designed to improve the certified robustness of a pretrained model (i.e. a classifier)
We show that this robust region is a union of multiple convex sets, and is potentially much larger than the robust regions identified in previous works.
arXiv Detail & Related papers (2022-11-01T08:18:07Z) - Learning Multivariate CDFs and Copulas using Tensor Factorization [39.24470798045442]
Learning the multivariate distribution of data is a core challenge in statistics and machine learning.
In this work, we aim to learn multivariate cumulative distribution functions (CDFs), as they can handle mixed random variables.
We show that any grid sampled version of a joint CDF of mixed random variables admits a universal representation as a naive Bayes model.
We demonstrate the superior performance of the proposed model in several synthetic and real datasets and applications including regression, sampling and data imputation.
arXiv Detail & Related papers (2022-10-13T16:18:46Z) - Sampling from Arbitrary Functions via PSD Models [55.41644538483948]
We take a two-step approach by first modeling the probability distribution and then sampling from that model.
We show that these models can approximate a large class of densities concisely using few evaluations, and present a simple algorithm to effectively sample from these models.
arXiv Detail & Related papers (2021-10-20T12:25:22Z) - Unrolling Particles: Unsupervised Learning of Sampling Distributions [102.72972137287728]
Particle filtering is used to compute good nonlinear estimates of complex systems.
We show in simulations that the resulting particle filter yields good estimates in a wide range of scenarios.
arXiv Detail & Related papers (2021-10-06T16:58:34Z) - An Embedded Model Estimator for Non-Stationary Random Functions using
Multiple Secondary Variables [0.0]
This paper introduces the method and shows that it has consistency results that are similar in nature to those applying to geostatistical modelling and to Quantile Random Forests.
The algorithm works by estimating a conditional distribution for the target variable at each target location.
arXiv Detail & Related papers (2020-11-09T00:14:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.