Generative Learning With Euler Particle Transport
- URL: http://arxiv.org/abs/2012.06094v1
- Date: Fri, 11 Dec 2020 03:10:53 GMT
- Title: Generative Learning With Euler Particle Transport
- Authors: Yuan Gao, Jian Huang, Yuling Jiao, Jin Liu, Xiliang Lu and Zhijian
Yang
- Abstract summary: We propose an Euler particle transport (EPT) approach for generative learning.
The proposed approach is motivated by the problem of finding an optimal transport map from a reference distribution to a target distribution.
We show that the proposed density-ratio (difference) estimators do not suffer from the "curse of dimensionality" if data is supported on a lower-dimensional manifold.
- Score: 14.557451744544592
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose an Euler particle transport (EPT) approach for generative
learning. The proposed approach is motivated by the problem of finding an
optimal transport map from a reference distribution to a target distribution
characterized by the Monge-Ampere equation. Interpreting the infinitesimal
linearization of the Monge-Ampere equation from the perspective of gradient
flows in measure spaces leads to a stochastic McKean-Vlasov equation. We use
the forward Euler method to solve this equation. The resulting forward Euler
map pushes forward a reference distribution to the target. This map is the
composition of a sequence of simple residual maps, which are computationally
stable and easy to train. The key task in training is the estimation of the
density ratios or differences that determine the residual maps. We estimate the
density ratios (differences) based on the Bregman divergence with a gradient
penalty using deep density-ratio (difference) fitting. We show that the
proposed density-ratio (difference) estimators do not suffer from the "curse of
dimensionality" if data is supported on a lower-dimensional manifold. Numerical
experiments with multi-mode synthetic datasets and comparisons with the
existing methods on real benchmark datasets support our theoretical results and
demonstrate the effectiveness of the proposed method.
Related papers
- A Stein Gradient Descent Approach for Doubly Intractable Distributions [5.63014864822787]
We propose a novel Monte Carlo Stein variational gradient descent (MC-SVGD) approach for inference for doubly intractable distributions.
The proposed method achieves substantial computational gains over existing algorithms, while providing comparable inferential performance for the posterior distributions.
arXiv Detail & Related papers (2024-10-28T13:42:27Z) - Total Uncertainty Quantification in Inverse PDE Solutions Obtained with Reduced-Order Deep Learning Surrogate Models [50.90868087591973]
We propose an approximate Bayesian method for quantifying the total uncertainty in inverse PDE solutions obtained with machine learning surrogate models.
We test the proposed framework by comparing it with the iterative ensemble smoother and deep ensembling methods for a non-linear diffusion equation.
arXiv Detail & Related papers (2024-08-20T19:06:02Z) - Sequential transport maps using SoS density estimation and $α$-divergences [0.5999777817331317]
Transport-based density estimation methods are receiving growing interest because of their ability to efficiently generate samples from the approximated density.
We provide a new convergence analyses of the sequential transport maps based on information geometric properties of $alpha$-divergences.
We numerically demonstrate our methods on Bayesian inference problems and unsupervised learning tasks.
arXiv Detail & Related papers (2024-02-27T23:52:58Z) - Distributed Markov Chain Monte Carlo Sampling based on the Alternating
Direction Method of Multipliers [143.6249073384419]
In this paper, we propose a distributed sampling scheme based on the alternating direction method of multipliers.
We provide both theoretical guarantees of our algorithm's convergence and experimental evidence of its superiority to the state-of-the-art.
In simulation, we deploy our algorithm on linear and logistic regression tasks and illustrate its fast convergence compared to existing gradient-based methods.
arXiv Detail & Related papers (2024-01-29T02:08:40Z) - A Geometric Perspective on Diffusion Models [57.27857591493788]
We inspect the ODE-based sampling of a popular variance-exploding SDE.
We establish a theoretical relationship between the optimal ODE-based sampling and the classic mean-shift (mode-seeking) algorithm.
arXiv Detail & Related papers (2023-05-31T15:33:16Z) - Efficient Training of Energy-Based Models Using Jarzynski Equality [13.636994997309307]
Energy-based models (EBMs) are generative models inspired by statistical physics.
The computation of its gradient with respect to the model parameters requires sampling the model distribution.
Here we show how results for nonequilibrium thermodynamics based on Jarzynski equality can be used to perform this computation efficiently.
arXiv Detail & Related papers (2023-05-30T21:07:52Z) - Gaussian process regression and conditional Karhunen-Lo\'{e}ve models
for data assimilation in inverse problems [68.8204255655161]
We present a model inversion algorithm, CKLEMAP, for data assimilation and parameter estimation in partial differential equation models.
The CKLEMAP method provides better scalability compared to the standard MAP method.
arXiv Detail & Related papers (2023-01-26T18:14:12Z) - Near-optimal estimation of smooth transport maps with kernel
sums-of-squares [81.02564078640275]
Under smoothness conditions, the squared Wasserstein distance between two distributions could be efficiently computed with appealing statistical error upper bounds.
The object of interest for applications such as generative modeling is the underlying optimal transport map.
We propose the first tractable algorithm for which the statistical $L2$ error on the maps nearly matches the existing minimax lower-bounds for smooth map estimation.
arXiv Detail & Related papers (2021-12-03T13:45:36Z) - Density Ratio Estimation via Infinitesimal Classification [85.08255198145304]
We propose DRE-infty, a divide-and-conquer approach to reduce Density ratio estimation (DRE) to a series of easier subproblems.
Inspired by Monte Carlo methods, we smoothly interpolate between the two distributions via an infinite continuum of intermediate bridge distributions.
We show that our approach performs well on downstream tasks such as mutual information estimation and energy-based modeling on complex, high-dimensional datasets.
arXiv Detail & Related papers (2021-11-22T06:26:29Z) - Deep Generative Learning via Schr\"{o}dinger Bridge [14.138796631423954]
We learn a generative model via entropy with a Schr"odinger Bridge.
We show that the generative model via Schr"odinger Bridge is comparable with state-of-the-art GANs.
arXiv Detail & Related papers (2021-06-19T03:35:42Z) - Learning Implicit Generative Models with Theoretical Guarantees [12.761710596142109]
We propose a textbfunified textbfframework for textbfimplicit textbfmodeling (UnifiGem)
UnifiGem integrates approaches from optimal transport, numerical ODE, density-ratio (density-difference) estimation and deep neural networks.
Experimental results on both synthetic datasets and real benchmark datasets support our theoretical findings and demonstrate the effectiveness of UnifiGem.
arXiv Detail & Related papers (2020-02-07T15:55:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.