Adversarial Score Distillation: When score distillation meets GAN
- URL: http://arxiv.org/abs/2312.00739v2
- Date: Tue, 10 Sep 2024 02:34:24 GMT
- Title: Adversarial Score Distillation: When score distillation meets GAN
- Authors: Min Wei, Jingkai Zhou, Junyao Sun, Xuesong Zhang,
- Abstract summary: We decipher existing score distillation with the Wasserstein Generative Adversarial Network (WGAN) paradigm.
With the WGAN paradigm, we find that existing score distillation either employs a fixed sub-optimal discriminator or conducts incomplete discriminator optimization.
We propose the Adversarial Score Distillation (ASD), which maintains an optimizable discriminator and updates it using the complete optimization objective.
- Score: 3.2794321281011394
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing score distillation methods are sensitive to classifier-free guidance (CFG) scale: manifested as over-smoothness or instability at small CFG scales, while over-saturation at large ones. To explain and analyze these issues, we revisit the derivation of Score Distillation Sampling (SDS) and decipher existing score distillation with the Wasserstein Generative Adversarial Network (WGAN) paradigm. With the WGAN paradigm, we find that existing score distillation either employs a fixed sub-optimal discriminator or conducts incomplete discriminator optimization, resulting in the scale-sensitive issue. We propose the Adversarial Score Distillation (ASD), which maintains an optimizable discriminator and updates it using the complete optimization objective. Experiments show that the proposed ASD performs favorably in 2D distillation and text-to-3D tasks against existing methods. Furthermore, to explore the generalization ability of our WGAN paradigm, we extend ASD to the image editing task, which achieves competitive results. The project page and code are at https://github.com/2y7c3/ASD.
Related papers
- Knowledge Distillation via Query Selection for Detection Transformer [25.512519971607237]
This paper addresses the challenge of compressing DETR by leveraging knowledge distillation.
A critical aspect of DETRs' performance is their reliance on queries to interpret object representations accurately.
Our visual analysis indicates that hard-negative queries, focusing on foreground elements, are crucial for enhancing distillation outcomes.
arXiv Detail & Related papers (2024-09-10T11:49:28Z) - VividDreamer: Invariant Score Distillation For Hyper-Realistic Text-to-3D Generation [33.05759961083337]
This paper presents Invariant Score Distillation (ISD), a novel method for high-fidelity text-to-3D generation.
ISD aims to tackle the over-saturation and over-smoothing problems in Score Distillation Sampling (SDS)
arXiv Detail & Related papers (2024-07-13T09:33:16Z) - SteinDreamer: Variance Reduction for Text-to-3D Score Distillation via Stein Identity [70.32101198891465]
We show that gradient estimation in score distillation is inherent to high variance.
We propose a more general solution to reduce variance for score distillation, termed Stein Score Distillation (SSD)
We demonstrate that SteinDreamer achieves faster convergence than existing methods due to more stable gradient updates.
arXiv Detail & Related papers (2023-12-31T23:04:25Z) - Taming Mode Collapse in Score Distillation for Text-to-3D Generation [70.32101198891465]
"Janus" artifact is a problem in text-to-3D generation where the generated objects fake each view with multiple front faces.
We propose a new update rule for 3D score distillation, dubbed Entropic Score Distillation ( ESD)
Although embarrassingly straightforward, our experiments successfully demonstrate that ESD can be an effective treatment for Janus artifacts in score distillation.
arXiv Detail & Related papers (2023-12-31T22:47:06Z) - Text-to-3D with Classifier Score Distillation [80.14832887529259]
Classifier-free guidance is considered an auxiliary trick rather than the most essential.
We name this method Score Distillation (CSD), which can be interpreted as using an implicit classification model for generation.
We validate the effectiveness of CSD across a variety of text-to-3D tasks including shape generation, texture synthesis, and shape editing.
arXiv Detail & Related papers (2023-10-30T10:25:40Z) - Noise-Free Score Distillation [78.79226724549456]
Noise-Free Score Distillation (NFSD) process requires minimal modifications to the original SDS framework.
We achieve more effective distillation of pre-trained text-to-image diffusion models while using a nominal CFG scale.
arXiv Detail & Related papers (2023-10-26T17:12:26Z) - StereoDistill: Pick the Cream from LiDAR for Distilling Stereo-based 3D
Object Detection [93.10989714186788]
We propose a cross-modal distillation method named StereoDistill to narrow the gap between the stereo and LiDAR-based approaches.
Key designs of StereoDistill are: the X-component Guided Distillation(XGD) for regression and the Cross-anchor Logit Distillation(CLD) for classification.
arXiv Detail & Related papers (2023-01-04T13:38:48Z) - SEA: Bridging the Gap Between One- and Two-stage Detector Distillation
via SEmantic-aware Alignment [76.80165589520385]
We name our method SEA (SEmantic-aware Alignment) distillation given the nature of abstracting dense fine-grained information.
It achieves new state-of-the-art results on the challenging object detection task on both one- and two-stage detectors.
arXiv Detail & Related papers (2022-03-02T04:24:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.