Non Gaussian Denoising Diffusion Models
- URL: http://arxiv.org/abs/2106.07582v1
- Date: Mon, 14 Jun 2021 16:42:43 GMT
- Title: Non Gaussian Denoising Diffusion Models
- Authors: Eliya Nachmani, Robin San Roman, Lior Wolf
- Abstract summary: We show that noise from Gamma distribution provides improved results for image and speech generation.
We also show that using a mixture of Gaussian noise variables in the diffusion process improves the performance over a diffusion process that is based on a single distribution.
- Score: 91.22679787578438
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generative diffusion processes are an emerging and effective tool for image
and speech generation. In the existing methods, the underline noise
distribution of the diffusion process is Gaussian noise. However, fitting
distributions with more degrees of freedom, could help the performance of such
generative models. In this work, we investigate other types of noise
distribution for the diffusion process. Specifically, we show that noise from
Gamma distribution provides improved results for image and speech generation.
Moreover, we show that using a mixture of Gaussian noise variables in the
diffusion process improves the performance over a diffusion process that is
based on a single distribution. Our approach preserves the ability to
efficiently sample state in the training diffusion process while using Gamma
noise and a mixture of noise.
Related papers
- Immiscible Diffusion: Accelerating Diffusion Training with Noise Assignment [56.609042046176555]
suboptimal noise-data mapping leads to slow training of diffusion models.
Drawing inspiration from the immiscibility phenomenon in physics, we propose Immiscible Diffusion.
Our approach is remarkably simple, requiring only one line of code to restrict the diffuse-able area for each image.
arXiv Detail & Related papers (2024-06-18T06:20:42Z) - Diffusion Gaussian Mixture Audio Denoise [23.760755498636943]
We propose a DiffGMM model, a denoising model based on the diffusion and Gaussian mixture models.
Given a noisy audio signal, we first apply a 1D-U-Net to extract features and train linear layers to estimate parameters for the Gaussian mixture model.
The noisy signal is continuously subtracted from the estimated noise to output clean audio signals.
arXiv Detail & Related papers (2024-06-13T14:18:10Z) - Blue noise for diffusion models [50.99852321110366]
We introduce a novel and general class of diffusion models taking correlated noise within and across images into account.
Our framework allows introducing correlation across images within a single mini-batch to improve gradient flow.
We perform both qualitative and quantitative evaluations on a variety of datasets using our method.
arXiv Detail & Related papers (2024-02-07T14:59:25Z) - An Analysis of the Variance of Diffusion-based Speech Enhancement [15.736484513462973]
We show that the scale of the variance is a dominant parameter for speech enhancement performance.
We show that a larger variance increases the noise attenuation and allows for reducing the computational footprint.
arXiv Detail & Related papers (2024-02-01T17:46:19Z) - Diffusion Models With Learned Adaptive Noise [12.530583016267768]
We propose a learned diffusion process that applies noise at different rates across an image.
MuLAN sets a new state-of-the-art in density estimation on CIFAR-10 and ImageNet.
arXiv Detail & Related papers (2023-12-20T18:00:16Z) - Diffusion-GAN: Training GANs with Diffusion [135.24433011977874]
Generative adversarial networks (GANs) are challenging to train stably.
We propose Diffusion-GAN, a novel GAN framework that leverages a forward diffusion chain to generate instance noise.
We show that Diffusion-GAN can produce more realistic images with higher stability and data efficiency than state-of-the-art GANs.
arXiv Detail & Related papers (2022-06-05T20:45:01Z) - Truncated Diffusion Probabilistic Models and Diffusion-based Adversarial
Auto-Encoders [137.1060633388405]
Diffusion-based generative models learn how to generate the data by inferring a reverse diffusion chain.
We propose a faster and cheaper approach that adds noise not until the data become pure random noise.
We show that the proposed model can be cast as an adversarial auto-encoder empowered by both the diffusion process and a learnable implicit prior.
arXiv Detail & Related papers (2022-02-19T20:18:49Z) - Denoising Diffusion Gamma Models [91.22679787578438]
We introduce the Denoising Diffusion Gamma Model (DDGM) and show that noise from Gamma distribution provides improved results for image and speech generation.
Our approach preserves the ability to efficiently sample state in the training diffusion process while using Gamma noise.
arXiv Detail & Related papers (2021-10-10T10:46:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.