Convergence guarantee for consistency models
- URL: http://arxiv.org/abs/2308.11449v1
- Date: Tue, 22 Aug 2023 13:57:35 GMT
- Title: Convergence guarantee for consistency models
- Authors: Junlong Lyu, Zhitang Chen, Shoubo Feng
- Abstract summary: We provide the first convergence guarantees for the Consistency Models (CMs), a newly emerging type of one-step generative models.
Under the basic assumptions on score-matching errors, consistency errors and smoothness of the data distribution, CMs can efficiently sample from any realistic data distribution in one step with small $W$ error.
- Score: 9.893455771918793
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We provide the first convergence guarantees for the Consistency Models (CMs),
a newly emerging type of one-step generative models that can generate
comparable samples to those generated by Diffusion Models. Our main result is
that, under the basic assumptions on score-matching errors, consistency errors
and smoothness of the data distribution, CMs can efficiently sample from any
realistic data distribution in one step with small $W_2$ error. Our results (1)
hold for $L^2$-accurate score and consistency assumption (rather than
$L^\infty$-accurate); (2) do note require strong assumptions on the data
distribution such as log-Sobelev inequality; (3) scale polynomially in all
parameters; and (4) match the state-of-the-art convergence guarantee for
score-based generative models (SGMs). We also provide the result that the
Multistep Consistency Sampling procedure can further reduce the error comparing
to one step sampling, which support the original statement of "Consistency
Models, Yang Song 2023". Our result further imply a TV error guarantee when
take some Langevin-based modifications to the output distributions.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Informed Correctors for Discrete Diffusion Models [32.87362154118195]
We propose a family of informed correctors that more reliably counteracts discretization error by leveraging information learned by the model.
We also propose $k$-Gillespie's, a sampling algorithm that better utilizes each model evaluation, while still enjoying the speed and flexibility of $tau$-leaping.
Across several real and synthetic datasets, we show that $k$-Gillespie's with informed correctors reliably produces higher quality samples at lower computational cost.
arXiv Detail & Related papers (2024-07-30T23:29:29Z) - Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution [67.9215891673174]
We propose score entropy as a novel loss that naturally extends score matching to discrete spaces.
We test our Score Entropy Discrete Diffusion models on standard language modeling tasks.
arXiv Detail & Related papers (2023-10-25T17:59:12Z) - Towards Faster Non-Asymptotic Convergence for Diffusion-Based Generative
Models [49.81937966106691]
We develop a suite of non-asymptotic theory towards understanding the data generation process of diffusion models.
In contrast to prior works, our theory is developed based on an elementary yet versatile non-asymptotic approach.
arXiv Detail & Related papers (2023-06-15T16:30:08Z) - Tailoring Language Generation Models under Total Variation Distance [55.89964205594829]
The standard paradigm of neural language generation adopts maximum likelihood estimation (MLE) as the optimizing method.
We develop practical bounds to apply it to language generation.
We introduce the TaiLr objective that balances the tradeoff of estimating TVD.
arXiv Detail & Related papers (2023-02-26T16:32:52Z) - Learning Multivariate CDFs and Copulas using Tensor Factorization [39.24470798045442]
Learning the multivariate distribution of data is a core challenge in statistics and machine learning.
In this work, we aim to learn multivariate cumulative distribution functions (CDFs), as they can handle mixed random variables.
We show that any grid sampled version of a joint CDF of mixed random variables admits a universal representation as a naive Bayes model.
We demonstrate the superior performance of the proposed model in several synthetic and real datasets and applications including regression, sampling and data imputation.
arXiv Detail & Related papers (2022-10-13T16:18:46Z) - Sampling is as easy as learning the score: theory for diffusion models
with minimal data assumptions [45.04514545004051]
We provide convergence guarantees for score-based generative models (SGMs)
We also examine SGMs based on the critically damped Langevin diffusion (CLD)
arXiv Detail & Related papers (2022-09-22T17:55:01Z) - Convergence for score-based generative modeling with polynomial
complexity [9.953088581242845]
We prove the first convergence guarantees for the core mechanic behind Score-based generative modeling.
Compared to previous works, we do not incur error that grows exponentially in time or that suffers from a curse of dimensionality.
We show that a predictor-corrector gives better convergence than using either portion alone.
arXiv Detail & Related papers (2022-06-13T14:57:35Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.