Elucidating The Design Space of Classifier-Guided Diffusion Generation
- URL: http://arxiv.org/abs/2310.11311v1
- Date: Tue, 17 Oct 2023 14:34:58 GMT
- Title: Elucidating The Design Space of Classifier-Guided Diffusion Generation
- Authors: Jiajun Ma, Tianyang Hu, Wenjia Wang and Jiacheng Sun
- Abstract summary: We show that it is possible to achieve significant performance improvements over existing guidance schemes by leveraging off-the-shelf classifiers in a training-free fashion.
Our proposed approach has great potential and can be readily scaled up to text-to-image generation tasks.
- Score: 17.704873767509557
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Guidance in conditional diffusion generation is of great importance for
sample quality and controllability. However, existing guidance schemes are to
be desired. On one hand, mainstream methods such as classifier guidance and
classifier-free guidance both require extra training with labeled data, which
is time-consuming and unable to adapt to new conditions. On the other hand,
training-free methods such as universal guidance, though more flexible, have
yet to demonstrate comparable performance. In this work, through a
comprehensive investigation into the design space, we show that it is possible
to achieve significant performance improvements over existing guidance schemes
by leveraging off-the-shelf classifiers in a training-free fashion, enjoying
the best of both worlds. Employing calibration as a general guideline, we
propose several pre-conditioning techniques to better exploit pretrained
off-the-shelf classifiers for guiding diffusion generation. Extensive
experiments on ImageNet validate our proposed method, showing that
state-of-the-art diffusion models (DDPM, EDM, DiT) can be further improved (up
to 20%) using off-the-shelf classifiers with barely any extra computational
cost. With the proliferation of publicly available pretrained classifiers, our
proposed approach has great potential and can be readily scaled up to
text-to-image generation tasks. The code is available at
https://github.com/AlexMaOLS/EluCD/tree/main.
Related papers
- Adding Conditional Control to Diffusion Models with Reinforcement Learning [59.295203871547336]
Diffusion models are powerful generative models that allow for precise control over the characteristics of the generated samples.
This work presents a novel method based on reinforcement learning (RL) to add additional controls, leveraging an offline dataset.
arXiv Detail & Related papers (2024-06-17T22:00:26Z) - Dreamguider: Improved Training free Diffusion-based Conditional Generation [31.68823843900196]
Dreamguider is a method that enables inference-time guidance without compute-heavy backpropagation through the diffusion network.
We present experiments using Dreamguider on multiple tasks across multiple datasets and models to show the effectiveness of the proposed modules.
arXiv Detail & Related papers (2024-06-04T17:59:32Z) - Manifold Preserving Guided Diffusion [121.97907811212123]
Conditional image generation still faces challenges of cost, generalizability, and the need for task-specific training.
We propose Manifold Preserving Guided Diffusion (MPGD), a training-free conditional generation framework.
arXiv Detail & Related papers (2023-11-28T02:08:06Z) - Guided Flows for Generative Modeling and Decision Making [55.42634941614435]
We show that Guided Flows significantly improves the sample quality in conditional image generation and zero-shot text synthesis-to-speech.
Notably, we are first to apply flow models for plan generation in the offline reinforcement learning setting ax speedup in compared to diffusion models.
arXiv Detail & Related papers (2023-11-22T15:07:59Z) - Consistency Regularization for Generalizable Source-free Domain
Adaptation [62.654883736925456]
Source-free domain adaptation (SFDA) aims to adapt a well-trained source model to an unlabelled target domain without accessing the source dataset.
Existing SFDA methods ONLY assess their adapted models on the target training set, neglecting the data from unseen but identically distributed testing sets.
We propose a consistency regularization framework to develop a more generalizable SFDA method.
arXiv Detail & Related papers (2023-08-03T07:45:53Z) - Guiding The Last Layer in Federated Learning with Pre-Trained Models [18.382057374270143]
Federated Learning (FL) is an emerging paradigm that allows a model to be trained across a number of participants without sharing data.
We show that fitting a classification head using the Nearest Class Means (NCM) can be done exactly and orders of magnitude more efficiently than existing proposals.
arXiv Detail & Related papers (2023-06-06T18:02:02Z) - End-to-End Diffusion Latent Optimization Improves Classifier Guidance [81.27364542975235]
Direct Optimization of Diffusion Latents (DOODL) is a novel guidance method.
It enables plug-and-play guidance by optimizing diffusion latents.
It outperforms one-step classifier guidance on computational and human evaluation metrics.
arXiv Detail & Related papers (2023-03-23T22:43:52Z) - Binary Classification from Multiple Unlabeled Datasets via Surrogate Set
Classification [94.55805516167369]
We propose a new approach for binary classification from m U-sets for $mge2$.
Our key idea is to consider an auxiliary classification task called surrogate set classification (SSC)
arXiv Detail & Related papers (2021-02-01T07:36:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.