DAG: Depth-Aware Guidance with Denoising Diffusion Probabilistic Models
- URL: http://arxiv.org/abs/2212.08861v1
- Date: Sat, 17 Dec 2022 12:47:19 GMT
- Title: DAG: Depth-Aware Guidance with Denoising Diffusion Probabilistic Models
- Authors: Gyeongnyeon Kim, Wooseok Jang, Gyuseong Lee, Susung Hong, Junyoung
Seo, Seungryong Kim
- Abstract summary: We propose a novel guidance approach for diffusion models that uses estimated depth information derived from the rich intermediate representations of diffusion models.
Experiments and extensive ablation studies demonstrate the effectiveness of our method in guiding the diffusion models toward geometrically plausible image generation.
- Score: 23.70476220346754
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, generative models have undergone significant advancement due
to the success of diffusion models. The success of these models is often
attributed to their use of guidance techniques, such as classifier and
classifier-free methods, which provides effective mechanisms to trade-off
between fidelity and diversity. However, these methods are not capable of
guiding a generated image to be aware of its geometric configuration, e.g.,
depth, which hinders the application of diffusion models to areas that require
a certain level of depth awareness. To address this limitation, we propose a
novel guidance approach for diffusion models that uses estimated depth
information derived from the rich intermediate representations of diffusion
models. To do this, we first present a label-efficient depth estimation
framework using the internal representations of diffusion models. At the
sampling phase, we utilize two guidance techniques to self-condition the
generated image using the estimated depth map, the first of which uses
pseudo-labeling, and the subsequent one uses a depth-domain diffusion prior.
Experiments and extensive ablation studies demonstrate the effectiveness of our
method in guiding the diffusion models toward geometrically plausible image
generation. Project page is available at https://ku-cvlab.github.io/DAG/.
Related papers
- Diffusion Models in Low-Level Vision: A Survey [82.77962165415153]
diffusion model-based solutions have emerged as widely acclaimed for their ability to produce samples of superior quality and diversity.
We present three generic diffusion modeling frameworks and explore their correlations with other deep generative models.
We summarize extended diffusion models applied in other tasks, including medical, remote sensing, and video scenarios.
arXiv Detail & Related papers (2024-06-17T01:49:27Z) - Guided Diffusion from Self-Supervised Diffusion Features [49.78673164423208]
Guidance serves as a key concept in diffusion models, yet its effectiveness is often limited by the need for extra data annotation or pretraining.
We propose a framework to extract guidance from, and specifically for, diffusion models.
arXiv Detail & Related papers (2023-12-14T11:19:11Z) - Fast Diffusion EM: a diffusion model for blind inverse problems with
application to deconvolution [0.0]
Current methods assume the degradation to be known and provide impressive results in terms of restoration and diversity.
In this work, we leverage the efficiency of those models to jointly estimate the restored image and unknown parameters of the kernel model.
Our method alternates between approximating the expected log-likelihood of the problem using samples drawn from a diffusion model and a step to estimate unknown model parameters.
arXiv Detail & Related papers (2023-09-01T06:47:13Z) - Directional diffusion models for graph representation learning [9.457273750874357]
We propose a new class of models called it directional diffusion models
These models incorporate data-dependent, anisotropic, and directional noises in the forward diffusion process.
We conduct extensive experiments on 12 publicly available datasets, focusing on two distinct graph representation learning tasks.
arXiv Detail & Related papers (2023-06-22T21:27:48Z) - Interpretable ODE-style Generative Diffusion Model via Force Field
Construction [0.0]
This paper aims to identify various physical models that are suitable for constructing ODE-style generative diffusion models accurately from a mathematical perspective.
We perform a case study where we use the theoretical model identified by our method to develop a range of new diffusion model methods.
arXiv Detail & Related papers (2023-03-14T16:58:11Z) - SinDiffusion: Learning a Diffusion Model from a Single Natural Image [159.4285444680301]
We present SinDiffusion, leveraging denoising diffusion models to capture internal distribution of patches from a single natural image.
It is based on two core designs. First, SinDiffusion is trained with a single model at a single scale instead of multiple models with progressive growing of scales.
Second, we identify that a patch-level receptive field of the diffusion network is crucial and effective for capturing the image's patch statistics.
arXiv Detail & Related papers (2022-11-22T18:00:03Z) - Unifying Diffusion Models' Latent Space, with Applications to
CycleDiffusion and Guidance [95.12230117950232]
We show that a common latent space emerges from two diffusion models trained independently on related domains.
Applying CycleDiffusion to text-to-image diffusion models, we show that large-scale text-to-image diffusion models can be used as zero-shot image-to-image editors.
arXiv Detail & Related papers (2022-10-11T15:53:52Z) - Diffusion Models in Vision: A Survey [80.82832715884597]
A diffusion model is a deep generative model that is based on two stages, a forward diffusion stage and a reverse diffusion stage.
Diffusion models are widely appreciated for the quality and diversity of the generated samples, despite their known computational burdens.
arXiv Detail & Related papers (2022-09-10T22:00:30Z) - A Survey on Generative Diffusion Model [75.93774014861978]
Diffusion models are an emerging class of deep generative models.
They have certain limitations, including a time-consuming iterative generation process and confinement to high-dimensional Euclidean space.
This survey presents a plethora of advanced techniques aimed at enhancing diffusion models.
arXiv Detail & Related papers (2022-09-06T16:56:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.