Convergence of the Inexact Langevin Algorithm and Score-based Generative
Models in KL Divergence
- URL: http://arxiv.org/abs/2211.01512v2
- Date: Fri, 2 Jun 2023 14:57:31 GMT
- Title: Convergence of the Inexact Langevin Algorithm and Score-based Generative
Models in KL Divergence
- Authors: Kaylee Yingxi Yang, Andre Wibisono
- Abstract summary: We study the Inexact Langevin Dynamics (ILD), Inexact Langevin Algorithm (ILA), and Score-based Generative Modeling (SGM) when utilizing estimated score functions for sampling.
- Score: 4.974890682815778
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the Inexact Langevin Dynamics (ILD), Inexact Langevin Algorithm
(ILA), and Score-based Generative Modeling (SGM) when utilizing estimated score
functions for sampling. Our focus lies in establishing stable biased
convergence guarantees in terms of the Kullback-Leibler (KL) divergence. To
achieve these guarantees, we impose two key assumptions: 1) the target
distribution satisfies the log-Sobolev inequality (LSI), and 2) the score
estimator exhibits a bounded Moment Generating Function (MGF) error. Notably,
the MGF error assumption we adopt is more lenient compared to the $L^\infty$
error assumption used in existing literature. However, it is stronger than the
$L^2$ error assumption utilized in recent works, which often leads to unstable
bounds. We explore the question of how to obtain a provably accurate score
estimator that satisfies the MGF error assumption. Specifically, we demonstrate
that a simple estimator based on kernel density estimation fulfills the MGF
error assumption for sub-Gaussian target distribution, at the population level.
Related papers
- Score-based generative models are provably robust: an uncertainty quantification perspective [4.396860522241307]
We show that score-based generative models (SGMs) are provably robust to the multiple sources of error in practical implementation.
Our primary tool is the Wasserstein uncertainty propagation (WUP) theorem.
We show how errors due to (a) finite sample approximation, (b) early stopping, (c) score-matching objective choice, (d) score function parametrization, and (e) reference distribution choice, impact the quality of the generative model.
arXiv Detail & Related papers (2024-05-24T17:50:17Z) - Multi-hypothesis 3D human pose estimation metrics favor miscalibrated
distributions [3.8575800313102806]
Well-calibrated distributions can make ambiguities explicit and preserve uncertainty for downstream tasks.
We identify that miscalibration can be attributed to the use of sample-based metrics such as minMPJPE.
To mitigate this problem, we propose an accurate and well-calibrated model called Conditional Graph Normalizing Flow (cGNF)
arXiv Detail & Related papers (2022-10-20T11:47:07Z) - Sampling is as easy as learning the score: theory for diffusion models
with minimal data assumptions [45.04514545004051]
We provide convergence guarantees for score-based generative models (SGMs)
We also examine SGMs based on the critically damped Langevin diffusion (CLD)
arXiv Detail & Related papers (2022-09-22T17:55:01Z) - Utilising the CLT Structure in Stochastic Gradient based Sampling :
Improved Analysis and Faster Algorithms [14.174806471635403]
We consider approximations of sampling algorithms, such as Gradient Langevin Dynamics (SGLD) and the Random Batch Method (RBM) for Interacting Particle Dynamcs (IPD)
We observe that the noise introduced by the approximation is nearly Gaussian due to the Central Limit Theorem (CLT) while the driving Brownian motion is exactly Gaussian.
We harness this structure to absorb the approximation error inside the diffusion process, and obtain improved convergence guarantees for these algorithms.
arXiv Detail & Related papers (2022-06-08T10:17:40Z) - Robust Estimation for Nonparametric Families via Generative Adversarial
Networks [92.64483100338724]
We provide a framework for designing Generative Adversarial Networks (GANs) to solve high dimensional robust statistics problems.
Our work extend these to robust mean estimation, second moment estimation, and robust linear regression.
In terms of techniques, our proposed GAN losses can be viewed as a smoothed and generalized Kolmogorov-Smirnov distance.
arXiv Detail & Related papers (2022-02-02T20:11:33Z) - On the Double Descent of Random Features Models Trained with SGD [78.0918823643911]
We study properties of random features (RF) regression in high dimensions optimized by gradient descent (SGD)
We derive precise non-asymptotic error bounds of RF regression under both constant and adaptive step-size SGD setting.
We observe the double descent phenomenon both theoretically and empirically.
arXiv Detail & Related papers (2021-10-13T17:47:39Z) - Optimal policy evaluation using kernel-based temporal difference methods [78.83926562536791]
We use kernel Hilbert spaces for estimating the value function of an infinite-horizon discounted Markov reward process.
We derive a non-asymptotic upper bound on the error with explicit dependence on the eigenvalues of the associated kernel operator.
We prove minimax lower bounds over sub-classes of MRPs.
arXiv Detail & Related papers (2021-09-24T14:48:20Z) - On the Practicality of Differential Privacy in Federated Learning by
Tuning Iteration Times [51.61278695776151]
Federated Learning (FL) is well known for its privacy protection when training machine learning models among distributed clients collaboratively.
Recent studies have pointed out that the naive FL is susceptible to gradient leakage attacks.
Differential Privacy (DP) emerges as a promising countermeasure to defend against gradient leakage attacks.
arXiv Detail & Related papers (2021-01-11T19:43:12Z) - Generative Modeling with Denoising Auto-Encoders and Langevin Sampling [88.83704353627554]
We show that both DAE and DSM provide estimates of the score of the smoothed population density.
We then apply our results to the homotopy method of arXiv:1907.05600 and provide theoretical justification for its empirical success.
arXiv Detail & Related papers (2020-01-31T23:50:03Z) - Distributionally Robust Bayesian Quadrature Optimization [60.383252534861136]
We study BQO under distributional uncertainty in which the underlying probability distribution is unknown except for a limited set of its i.i.d. samples.
A standard BQO approach maximizes the Monte Carlo estimate of the true expected objective given the fixed sample set.
We propose a novel posterior sampling based algorithm, namely distributionally robust BQO (DRBQO) for this purpose.
arXiv Detail & Related papers (2020-01-19T12:00:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.