Stable Attribute Group Editing for Reliable Few-shot Image Generation
- URL: http://arxiv.org/abs/2302.00179v1
- Date: Wed, 1 Feb 2023 01:51:47 GMT
- Title: Stable Attribute Group Editing for Reliable Few-shot Image Generation
- Authors: Guanqi Ding, Xinzhe Han, Shuhui Wang, Xin Jin, Dandan Tu, Qingming
Huang
- Abstract summary: We present an editing-based'' framework Attribute Group Editing (AGE) for reliable few-shot image generation.
We find that class inconsistency is a common problem in GAN-generated images for downstream classification.
We propose to boost the downstream classification performance of SAGE by enhancing the pixel and frequency components.
- Score: 88.59350889410794
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Few-shot image generation aims to generate data of an unseen category based
on only a few samples. Apart from basic content generation, a bunch of
downstream applications hopefully benefit from this task, such as low-data
detection and few-shot classification. To achieve this goal, the generated
images should guarantee category retention for classification beyond the visual
quality and diversity. In our preliminary work, we present an ``editing-based''
framework Attribute Group Editing (AGE) for reliable few-shot image generation,
which largely improves the generation performance. Nevertheless, AGE's
performance on downstream classification is not as satisfactory as expected.
This paper investigates the class inconsistency problem and proposes Stable
Attribute Group Editing (SAGE) for more stable class-relevant image generation.
SAGE takes use of all given few-shot images and estimates a class center
embedding based on the category-relevant attribute dictionary. Meanwhile,
according to the projection weights on the category-relevant attribute
dictionary, we can select category-irrelevant attributes from the similar seen
categories. Consequently, SAGE injects the whole distribution of the novel
class into StyleGAN's latent space, thus largely remains the category retention
and stability of the generated images. Going one step further, we find that
class inconsistency is a common problem in GAN-generated images for downstream
classification. Even though the generated images look photo-realistic and
requires no category-relevant editing, they are usually of limited help for
downstream classification. We systematically discuss this issue from both the
generative model and classification model perspectives, and propose to boost
the downstream classification performance of SAGE by enhancing the pixel and
frequency components.
Related papers
- Diversified in-domain synthesis with efficient fine-tuning for few-shot
classification [64.86872227580866]
Few-shot image classification aims to learn an image classifier using only a small set of labeled examples per class.
We propose DISEF, a novel approach which addresses the generalization challenge in few-shot learning using synthetic data.
We validate our method in ten different benchmarks, consistently outperforming baselines and establishing a new state-of-the-art for few-shot classification.
arXiv Detail & Related papers (2023-12-05T17:18:09Z) - CHIP: Contrastive Hierarchical Image Pretraining [1.6454089842548987]
We propose a one-shot/few-shot classification model that can classify an object of any unseen class into a relatively general category.
Our model uses a three-level hierarchical contrastive loss based ResNet152 for classifying an object based on its features extracted from Image embedding.
arXiv Detail & Related papers (2023-10-12T13:11:38Z) - Image-free Classifier Injection for Zero-Shot Classification [72.66409483088995]
Zero-shot learning models achieve remarkable results on image classification for samples from classes that were not seen during training.
We aim to equip pre-trained models with zero-shot classification capabilities without the use of image data.
We achieve this with our proposed Image-free Injection with Semantics (ICIS)
arXiv Detail & Related papers (2023-08-21T09:56:48Z) - Balanced Classification: A Unified Framework for Long-Tailed Object
Detection [74.94216414011326]
Conventional detectors suffer from performance degradation when dealing with long-tailed data due to a classification bias towards the majority head categories.
We introduce a unified framework called BAlanced CLassification (BACL), which enables adaptive rectification of inequalities caused by disparities in category distribution.
BACL consistently achieves performance improvements across various datasets with different backbones and architectures.
arXiv Detail & Related papers (2023-08-04T09:11:07Z) - Robust Image Ordinal Regression with Controllable Image Generation [12.787258917766005]
We propose a novel framework called CIG based on controllable image generation.
Our main idea is to generate extra training samples with specific labels near category boundaries.
We evaluate the effectiveness of our new CIG approach in three different image ordinal regression scenarios.
arXiv Detail & Related papers (2023-05-07T08:10:56Z) - Attribute Group Editing for Reliable Few-shot Image Generation [85.52840521454411]
We propose a new editing-based method, i.e., Attribute Group Editing (AGE), for few-shot image generation.
AGE examines the internal representation learned in GANs and identifies semantically meaningful directions.
arXiv Detail & Related papers (2022-03-16T06:54:09Z) - A Compositional Feature Embedding and Similarity Metric for
Ultra-Fine-Grained Visual Categorization [16.843126268445726]
Fine-grained visual categorization (FGVC) aims at classifying objects with small inter-class variances.
This paper proposes a novel compositional feature embedding and similarity metric ( CECS) for ultra-fine-grained visual categorization.
Experimental results on two ultra-FGVC datasets and one FGVC dataset with recent benchmark methods consistently demonstrate that the proposed CECS method achieves the state-the-art performance.
arXiv Detail & Related papers (2021-09-25T15:05:25Z) - Reconciliation of Statistical and Spatial Sparsity For Robust Image and
Image-Set Classification [27.319334479994787]
We propose a novel Joint Statistical and Spatial Sparse representation, dubbed textitJ3S, to model the image or image-set data for classification.
We propose to solve the joint sparse coding problem based on the J3S model, by coupling the local and global image representations using joint sparsity.
Experiments show that the proposed J3S-based image classification scheme outperforms the popular or state-of-the-art competing methods over FMD, UIUC, ETH-80 and YTC databases.
arXiv Detail & Related papers (2021-06-01T06:33:24Z) - I Am Going MAD: Maximum Discrepancy Competition for Comparing
Classifiers Adaptively [135.7695909882746]
We name the MAximum Discrepancy (MAD) competition.
We adaptively sample a small test set from an arbitrarily large corpus of unlabeled images.
Human labeling on the resulting model-dependent image sets reveals the relative performance of the competing classifiers.
arXiv Detail & Related papers (2020-02-25T03:32:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.