Restart Sampling for Improving Generative Processes
- URL: http://arxiv.org/abs/2306.14878v2
- Date: Wed, 1 Nov 2023 04:17:43 GMT
- Title: Restart Sampling for Improving Generative Processes
- Authors: Yilun Xu, Mingyang Deng, Xiang Cheng, Yonglong Tian, Ziming Liu, Tommi
Jaakkola
- Abstract summary: ODE-based samplers are fast but plateau in performance while SDE-based samplers deliver higher sample quality at the cost of increased sampling time.
We propose a novel sampling algorithm called Restart in order to better balance discretization errors and contraction.
- Score: 30.745245429072735
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generative processes that involve solving differential equations, such as
diffusion models, frequently necessitate balancing speed and quality. ODE-based
samplers are fast but plateau in performance while SDE-based samplers deliver
higher sample quality at the cost of increased sampling time. We attribute this
difference to sampling errors: ODE-samplers involve smaller discretization
errors while stochasticity in SDE contracts accumulated errors. Based on these
findings, we propose a novel sampling algorithm called Restart in order to
better balance discretization errors and contraction. The sampling method
alternates between adding substantial noise in additional forward steps and
strictly following a backward ODE. Empirically, Restart sampler surpasses
previous SDE and ODE samplers in both speed and accuracy. Restart not only
outperforms the previous best SDE results, but also accelerates the sampling
speed by 10-fold / 2-fold on CIFAR-10 / ImageNet $64 \times 64$. In addition,
it attains significantly better sample quality than ODE samplers within
comparable sampling times. Moreover, Restart better balances text-image
alignment/visual quality versus diversity than previous samplers in the
large-scale text-to-image Stable Diffusion model pre-trained on LAION $512
\times 512$. Code is available at
https://github.com/Newbeeer/diffusion_restart_sampling
Related papers
- DC-Solver: Improving Predictor-Corrector Diffusion Sampler via Dynamic Compensation [68.55191764622525]
Diffusion models (DPMs) have shown remarkable performance in visual synthesis but are computationally expensive due to the need for multiple evaluations during the sampling.
Recent predictor synthesis-or diffusion samplers have significantly reduced the required number of evaluations, but inherently suffer from a misalignment issue.
We introduce a new fast DPM sampler called DC-CPRr, which leverages dynamic compensation to mitigate the misalignment.
arXiv Detail & Related papers (2024-09-05T17:59:46Z) - Fast ODE-based Sampling for Diffusion Models in Around 5 Steps [17.500594480727617]
We propose Approximate MEan-Direction solver (AMED-r) that eliminates truncation errors by directly learning the mean direction for fast sampling.
Our method can be easily used as a plugin to further improve existing ODE-based samplers.
arXiv Detail & Related papers (2023-11-30T13:07:19Z) - Sampler Scheduler for Diffusion Models [0.0]
Diffusion modeling (DM) has high-quality generative performance.
Currently, there is a contradiction in samplers for diffusion-based generative models.
We propose the feasibility of using different samplers (ODE/SDE) on different sampling steps of the same sampling process.
arXiv Detail & Related papers (2023-11-12T13:35:25Z) - Parallel Sampling of Diffusion Models [76.3124029406809]
Diffusion models are powerful generative models but suffer from slow sampling.
We present ParaDiGMS, a novel method to accelerate the sampling of pretrained diffusion models by denoising multiple steps in parallel.
arXiv Detail & Related papers (2023-05-25T17:59:42Z) - Fast Diffusion Probabilistic Model Sampling through the lens of Backward
Error Analysis [26.907301901503835]
Denoising diffusion probabilistic models (DDPMs) are a class of powerful generative models.
DDPMs generally need hundreds or thousands of sequential function evaluations (steps) of neural networks to generate a sample.
This paper aims to develop a fast sampling method for DDPMs requiring much fewer steps while retaining high sample quality.
arXiv Detail & Related papers (2023-04-22T16:58:47Z) - Preconditioned Score-based Generative Models [49.88840603798831]
An intuitive acceleration method is to reduce the sampling iterations which however causes severe performance degradation.
We propose a model-agnostic bfem preconditioned diffusion sampling (PDS) method that leverages matrix preconditioning to alleviate the aforementioned problem.
PDS alters the sampling process of a vanilla SGM at marginal extra computation cost, and without model retraining.
arXiv Detail & Related papers (2023-02-13T16:30:53Z) - DPM-Solver++: Fast Solver for Guided Sampling of Diffusion Probabilistic
Models [45.612477740555406]
We propose DPM-r++, a high-order solver for guided sampling of DPMs.
We show that DPM-r++ can generate high-quality samples within only 15 to 20 steps for guided sampling by pixel-space and latent-space DPMs.
arXiv Detail & Related papers (2022-11-02T13:14:30Z) - Adaptive Sketches for Robust Regression with Importance Sampling [64.75899469557272]
We introduce data structures for solving robust regression through gradient descent (SGD)
Our algorithm effectively runs $T$ steps of SGD with importance sampling while using sublinear space and just making a single pass over the data.
arXiv Detail & Related papers (2022-07-16T03:09:30Z) - ProDiff: Progressive Fast Diffusion Model For High-Quality
Text-to-Speech [63.780196620966905]
We propose ProDiff, on progressive fast diffusion model for high-quality text-to-speech.
ProDiff parameterizes the denoising model by directly predicting clean data to avoid distinct quality degradation in accelerating sampling.
Our evaluation demonstrates that ProDiff needs only 2 iterations to synthesize high-fidelity mel-spectrograms.
ProDiff enables a sampling speed of 24x faster than real-time on a single NVIDIA 2080Ti GPU.
arXiv Detail & Related papers (2022-07-13T17:45:43Z) - Score-Based Generative Modeling through Stochastic Differential
Equations [114.39209003111723]
We present a differential equation that transforms a complex data distribution to a known prior distribution by injecting noise.
A corresponding reverse-time SDE transforms the prior distribution back into the data distribution by slowly removing the noise.
By leveraging advances in score-based generative modeling, we can accurately estimate these scores with neural networks.
We demonstrate high fidelity generation of 1024 x 1024 images for the first time from a score-based generative model.
arXiv Detail & Related papers (2020-11-26T19:39:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.