On Analyzing Generative and Denoising Capabilities of Diffusion-based
Deep Generative Models
- URL: http://arxiv.org/abs/2206.00070v1
- Date: Tue, 31 May 2022 19:29:27 GMT
- Title: On Analyzing Generative and Denoising Capabilities of Diffusion-based
Deep Generative Models
- Authors: Kamil Deja, Anna Kuzina, Tomasz Trzci\'nski, Jakub M. Tomczak
- Abstract summary: Diffusion-based Deep Generative Models (DDGMs) offer state-of-the-art performance in generative modeling.
We study how the small amount of noise is transformed during the backward diffusion process.
- Score: 18.018935233383935
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Diffusion-based Deep Generative Models (DDGMs) offer state-of-the-art
performance in generative modeling. Their main strength comes from their unique
setup in which a model (the backward diffusion process) is trained to reverse
the forward diffusion process, which gradually adds noise to the input signal.
Although DDGMs are well studied, it is still unclear how the small amount of
noise is transformed during the backward diffusion process. Here, we focus on
analyzing this problem to gain more insight into the behavior of DDGMs and
their denoising and generative capabilities. We observe a fluid transition
point that changes the functionality of the backward diffusion process from
generating a (corrupted) image from noise to denoising the corrupted image to
the final sample. Based on this observation, we postulate to divide a DDGM into
two parts: a denoiser and a generator. The denoiser could be parameterized by a
denoising auto-encoder, while the generator is a diffusion-based model with its
own set of parameters. We experimentally validate our proposition, showing its
pros and cons.
Related papers
- Frequency Regulation for Exposure Bias Mitigation in Diffusion Models [13.095683155232281]
We make a key observation: the energy of predicted noisy samples in the reverse process continuously declines compared to perturbed samples in the forward process.<n>We introduce a dynamic frequency regulation mechanism utilizing wavelet transforms, which separately adjusts the low- and high-frequency subbands.<n>We derive the rigorous mathematical form of exposure bias.
arXiv Detail & Related papers (2025-07-14T08:58:38Z) - ReDDiT: Rehashing Noise for Discrete Visual Generation [53.813067778912]
We propose a rehashing noise approach for discrete diffusion transformer (termed ReDDiT)<n>We aim to extend absorbing states and improve expressive capacity of discrete diffusion models.<n>Experiments show that ReDDiT significantly outperforms the baseline model.
arXiv Detail & Related papers (2025-05-26T08:17:20Z) - Remasking Discrete Diffusion Models with Inference-Time Scaling [12.593164604625384]
We introduce the remasking diffusion model (ReMDM) sampler, a method that can be applied to pretrained masked diffusion models in a principled way.
Most interestingly, ReMDM endows discrete diffusion with a form of inference-time compute scaling.
arXiv Detail & Related papers (2025-03-01T02:37:51Z) - There and Back Again: On the relation between Noise and Image Inversions in Diffusion Models [3.5707423185282665]
Inversion-based techniques reverse the denoising process and map images back to their approximated starting noise.
We show that latents exhibit structural patterns in the form of less diverse noise predicted for smooth image regions.
This leads to a low diversity of generated editions based on the DDIM inversion procedure and ill-defined latent-to-image mapping.
arXiv Detail & Related papers (2024-10-31T00:30:35Z) - Diffusion Priors for Variational Likelihood Estimation and Image Denoising [10.548018200066858]
We propose adaptive likelihood estimation and MAP inference during the reverse diffusion process to tackle real-world noise.
Experiments and analyses on diverse real-world datasets demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2024-10-23T02:52:53Z) - Data Augmentation for Seizure Prediction with Generative Diffusion Model [26.967247641926814]
Seizure prediction is of great importance to improve the life of patients.
The severe imbalance problem between preictal and interictal data still poses a great challenge.
Data augmentation is an intuitive way to solve this problem.
We propose a novel data augmentation method with diffusion model called DiffEEG.
arXiv Detail & Related papers (2023-06-14T05:44:53Z) - UDPM: Upsampling Diffusion Probabilistic Models [33.51145642279836]
Denoising Diffusion Probabilistic Models (DDPM) have recently gained significant attention.
DDPMs generate high-quality samples from complex data distributions by defining an inverse process.
Unlike generative adversarial networks (GANs), the latent space of diffusion models is less interpretable.
In this work, we propose to generalize the denoising diffusion process into an Upsampling Diffusion Probabilistic Model (UDPM)
arXiv Detail & Related papers (2023-05-25T17:25:14Z) - DiffusionAD: Norm-guided One-step Denoising Diffusion for Anomaly Detection [80.20339155618612]
DiffusionAD is a novel anomaly detection pipeline comprising a reconstruction sub-network and a segmentation sub-network.<n>A rapid one-step denoising paradigm achieves hundreds of times acceleration while preserving comparable reconstruction quality.<n>Considering the diversity in the manifestation of anomalies, we propose a norm-guided paradigm to integrate the benefits of multiple noise scales.
arXiv Detail & Related papers (2023-03-15T16:14:06Z) - Diffusion Models in Vision: A Survey [80.82832715884597]
A diffusion model is a deep generative model that is based on two stages, a forward diffusion stage and a reverse diffusion stage.
Diffusion models are widely appreciated for the quality and diversity of the generated samples, despite their known computational burdens.
arXiv Detail & Related papers (2022-09-10T22:00:30Z) - Cold Diffusion: Inverting Arbitrary Image Transforms Without Noise [52.59444045853966]
We show that an entire family of generative models can be constructed by varying the choice of image degradation.
The success of fully deterministic models calls into question the community's understanding of diffusion models.
arXiv Detail & Related papers (2022-08-19T15:18:39Z) - Diffusion-GAN: Training GANs with Diffusion [135.24433011977874]
Generative adversarial networks (GANs) are challenging to train stably.
We propose Diffusion-GAN, a novel GAN framework that leverages a forward diffusion chain to generate instance noise.
We show that Diffusion-GAN can produce more realistic images with higher stability and data efficiency than state-of-the-art GANs.
arXiv Detail & Related papers (2022-06-05T20:45:01Z) - Truncated Diffusion Probabilistic Models and Diffusion-based Adversarial
Auto-Encoders [137.1060633388405]
Diffusion-based generative models learn how to generate the data by inferring a reverse diffusion chain.
We propose a faster and cheaper approach that adds noise not until the data become pure random noise.
We show that the proposed model can be cast as an adversarial auto-encoder empowered by both the diffusion process and a learnable implicit prior.
arXiv Detail & Related papers (2022-02-19T20:18:49Z) - Come-Closer-Diffuse-Faster: Accelerating Conditional Diffusion Models
for Inverse Problems through Stochastic Contraction [31.61199061999173]
Diffusion models have a critical downside - they are inherently slow to sample from, needing few thousand steps of iteration to generate images from pure Gaussian noise.
We show that starting from Gaussian noise is unnecessary. Instead, starting from a single forward diffusion with better initialization significantly reduces the number of sampling steps in the reverse conditional diffusion.
New sampling strategy, dubbed ComeCloser-DiffuseFaster (CCDF), also reveals a new insight on how the existing feedforward neural network approaches for inverse problems can be synergistically combined with the diffusion models.
arXiv Detail & Related papers (2021-12-09T04:28:41Z) - Denoising Diffusion Gamma Models [91.22679787578438]
We introduce the Denoising Diffusion Gamma Model (DDGM) and show that noise from Gamma distribution provides improved results for image and speech generation.
Our approach preserves the ability to efficiently sample state in the training diffusion process while using Gamma noise.
arXiv Detail & Related papers (2021-10-10T10:46:31Z) - Non Gaussian Denoising Diffusion Models [91.22679787578438]
We show that noise from Gamma distribution provides improved results for image and speech generation.
We also show that using a mixture of Gaussian noise variables in the diffusion process improves the performance over a diffusion process that is based on a single distribution.
arXiv Detail & Related papers (2021-06-14T16:42:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.