Score-based Conditional Generation with Fewer Labeled Data by
Self-calibrating Classifier Guidance
- URL: http://arxiv.org/abs/2307.04081v3
- Date: Wed, 7 Feb 2024 16:00:21 GMT
- Title: Score-based Conditional Generation with Fewer Labeled Data by
Self-calibrating Classifier Guidance
- Authors: Paul Kuo-Ming Huang, Si-An Chen, Hsuan-Tien Lin
- Abstract summary: Score-based generative models (SGMs) are a popular family of deep generative models that achieve leading image generation quality.
We argue that the problem is rooted in the classifier's tendency to overfit without coordinating with the underlying unconditional distribution.
We propose improving classifier-guided SGMs by letting the classifier regularize itself.
- Score: 16.025002076222002
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Score-based generative models (SGMs) are a popular family of deep generative
models that achieve leading image generation quality. Early studies extend SGMs
to tackle class-conditional generation by coupling an unconditional SGM with
the guidance of a trained classifier. Nevertheless, such classifier-guided SGMs
do not always achieve accurate conditional generation, especially when trained
with fewer labeled data. We argue that the problem is rooted in the
classifier's tendency to overfit without coordinating with the underlying
unconditional distribution. To make the classifier respect the unconditional
distribution, we propose improving classifier-guided SGMs by letting the
classifier regularize itself. The key idea of our proposed method is to use
principles from energy-based models to convert the classifier into another view
of the unconditional SGM. Existing losses for unconditional SGMs can then be
leveraged to achieve regularization by calibrating the classifier's internal
unconditional scores. The regularization scheme can be applied to not only the
labeled data but also unlabeled ones to further improve the classifier. Across
various percentages of fewer labeled data, empirical results show that the
proposed approach significantly enhances conditional generation quality. The
enhancements confirm the potential of the proposed self-calibration technique
for generative modeling with limited labeled data.
Related papers
- Exploring Beyond Logits: Hierarchical Dynamic Labeling Based on Embeddings for Semi-Supervised Classification [49.09505771145326]
We propose a Hierarchical Dynamic Labeling (HDL) algorithm that does not depend on model predictions and utilizes image embeddings to generate sample labels.
Our approach has the potential to change the paradigm of pseudo-label generation in semi-supervised learning.
arXiv Detail & Related papers (2024-04-26T06:00:27Z) - Generative Robust Classification [3.4773470589069477]
Training adversarially robust discriminative (i.e., softmax) classification has been the dominant approach to robust classification.
We investigate using adversarial training (AT)-based generative models.
We find it straightforward to apply advanced data augmentation to achieve better robustness in our approach.
arXiv Detail & Related papers (2022-12-14T15:33:11Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - NorMatch: Matching Normalizing Flows with Discriminative Classifiers for
Semi-Supervised Learning [8.749830466953584]
Semi-Supervised Learning (SSL) aims to learn a model using a tiny labeled set and massive amounts of unlabeled data.
In this work we introduce a new framework for SSL named NorMatch.
We demonstrate, through numerical and visual results, that NorMatch achieves state-of-the-art performance on several datasets.
arXiv Detail & Related papers (2022-11-17T15:39:18Z) - Zero-Shot Logit Adjustment [89.68803484284408]
Generalized Zero-Shot Learning (GZSL) is a semantic-descriptor-based learning technique.
In this paper, we propose a new generation-based technique to enhance the generator's effect while neglecting the improvement of the classifier.
Our experiments demonstrate that the proposed technique achieves state-of-the-art when combined with the basic generator, and it can improve various generative zero-shot learning frameworks.
arXiv Detail & Related papers (2022-04-25T17:54:55Z) - Dual Projection Generative Adversarial Networks for Conditional Image
Generation [26.563829113916942]
We propose a Dual Projection GAN (P2GAN) model that learns to balance between em data matching and em label matching.
We then propose an improved cGAN model with Auxiliary Classification that directly aligns the fake and real conditionals $P(textclass|textimage)$ by minimizing their $f$-divergence.
arXiv Detail & Related papers (2021-08-20T06:10:38Z) - Score-based Generative Modeling in Latent Space [93.8985523558869]
Score-based generative models (SGMs) have recently demonstrated impressive results in terms of both sample quality and distribution coverage.
Here, we propose the Latent Score-based Generative Model (LSGM), a novel approach that trains SGMs in a latent space.
Moving from data to latent space allows us to train more expressive generative models, apply SGMs to non-continuous data, and learn smoother SGMs in a smaller space.
arXiv Detail & Related papers (2021-06-10T17:26:35Z) - Classify and Generate Reciprocally: Simultaneous Positive-Unlabelled
Learning and Conditional Generation with Extra Data [77.31213472792088]
The scarcity of class-labeled data is a ubiquitous bottleneck in many machine learning problems.
We address this problem by leveraging Positive-Unlabeled(PU) classification and the conditional generation with extra unlabeled data.
We present a novel training framework to jointly target both PU classification and conditional generation when exposed to extra data.
arXiv Detail & Related papers (2020-06-14T08:27:40Z) - Pre-training Is (Almost) All You Need: An Application to Commonsense
Reasoning [61.32992639292889]
Fine-tuning of pre-trained transformer models has become the standard approach for solving common NLP tasks.
We introduce a new scoring method that casts a plausibility ranking task in a full-text format.
We show that our method provides a much more stable training phase across random restarts.
arXiv Detail & Related papers (2020-04-29T10:54:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.