Feature Generation for Long-tail Classification
- URL: http://arxiv.org/abs/2111.05956v1
- Date: Wed, 10 Nov 2021 21:34:29 GMT
- Title: Feature Generation for Long-tail Classification
- Authors: Rahul Vigneswaran and Marc T. Law and Vineeth N. Balasubramanian and
Makarand Tapaswi
- Abstract summary: We show how to generate meaningful features by estimating the tail category's distribution.
We also present a qualitative analysis of generated features using t-SNE visualizations and analyze the nearest neighbors used to calibrate the tail class distributions.
- Score: 36.186909933006675
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The visual world naturally exhibits an imbalance in the number of object or
scene instances resulting in a \emph{long-tailed distribution}. This imbalance
poses significant challenges for classification models based on deep learning.
Oversampling instances of the tail classes attempts to solve this imbalance.
However, the limited visual diversity results in a network with poor
representation ability. A simple counter to this is decoupling the
representation and classifier networks and using oversampling only to train the
classifier. In this paper, instead of repeatedly re-sampling the same image
(and thereby features), we explore a direction that attempts to generate
meaningful features by estimating the tail category's distribution. Inspired by
ideas from recent work on few-shot learning, we create calibrated distributions
to sample additional features that are subsequently used to train the
classifier. Through several experiments on the CIFAR-100-LT (long-tail) dataset
with varying imbalance factors and on mini-ImageNet-LT (long-tail), we show the
efficacy of our approach and establish a new state-of-the-art. We also present
a qualitative analysis of generated features using t-SNE visualizations and
analyze the nearest neighbors used to calibrate the tail class distributions.
Our code is available at https://github.com/rahulvigneswaran/TailCalibX.
Related papers
- How Re-sampling Helps for Long-Tail Learning? [45.187004699024435]
Long-tail learning has received significant attention due to the challenge it poses with extremely imbalanced datasets.
Recent studies claim that re-sampling brings negligible performance improvements in modern long-tail learning tasks.
We propose a new context shift augmentation module that generates diverse training images for the tail class.
arXiv Detail & Related papers (2023-10-27T16:20:34Z) - Adjusting Logit in Gaussian Form for Long-Tailed Visual Recognition [37.62659619941791]
We study the problem of long-tailed visual recognition from the perspective of feature level.
Two novel logit adjustment methods are proposed to improve model performance at a modest computational overhead.
Experiments conducted on benchmark datasets demonstrate the superior performance of the proposed method over the state-of-the-art ones.
arXiv Detail & Related papers (2023-05-18T02:06:06Z) - Improving GANs for Long-Tailed Data through Group Spectral
Regularization [51.58250647277375]
We propose a novel group Spectral Regularizer (gSR) that prevents the spectral explosion alleviating mode collapse.
We find that gSR effectively combines with existing augmentation and regularization techniques, leading to state-of-the-art image generation performance on long-tailed data.
arXiv Detail & Related papers (2022-08-21T17:51:05Z) - Constructing Balance from Imbalance for Long-tailed Image Recognition [50.6210415377178]
The imbalance between majority (head) classes and minority (tail) classes severely skews the data-driven deep neural networks.
Previous methods tackle with data imbalance from the viewpoints of data distribution, feature space, and model design.
We propose a concise paradigm by progressively adjusting label space and dividing the head classes and tail classes.
Our proposed model also provides a feature evaluation method and paves the way for long-tailed feature learning.
arXiv Detail & Related papers (2022-08-04T10:22:24Z) - Generating Representative Samples for Few-Shot Classification [8.62483598990205]
Few-shot learning aims to learn new categories with a few visual samples per class.
Few-shot class representations are often biased due to data scarcity.
We generate visual samples based on semantic embeddings using a conditional variational autoencoder model.
arXiv Detail & Related papers (2022-05-05T20:58:33Z) - Improving Tail-Class Representation with Centroid Contrastive Learning [145.73991900239017]
We propose interpolative centroid contrastive learning (ICCL) to improve long-tailed representation learning.
ICCL interpolates two images from a class-agnostic sampler and a class-aware sampler, and trains the model such that the representation of the ICCL can be used to retrieve the centroids for both source classes.
Our result shows a significant accuracy gain of 2.8% on the iNaturalist 2018 dataset with a real-world long-tailed distribution.
arXiv Detail & Related papers (2021-10-19T15:24:48Z) - Improving Calibration for Long-Tailed Recognition [68.32848696795519]
We propose two methods to improve calibration and performance in such scenarios.
For dataset bias due to different samplers, we propose shifted batch normalization.
Our proposed methods set new records on multiple popular long-tailed recognition benchmark datasets.
arXiv Detail & Related papers (2021-04-01T13:55:21Z) - ResLT: Residual Learning for Long-tailed Recognition [64.19728932445523]
We propose a more fundamental perspective for long-tailed recognition, i.e., from the aspect of parameter space.
We design the effective residual fusion mechanism -- with one main branch optimized to recognize images from all classes, another two residual branches are gradually fused and optimized to enhance images from medium+tail classes and tail classes respectively.
We test our method on several benchmarks, i.e., long-tailed version of CIFAR-10, CIFAR-100, Places, ImageNet, and iNaturalist 2018.
arXiv Detail & Related papers (2021-01-26T08:43:50Z) - Free Lunch for Few-shot Learning: Distribution Calibration [10.474018806591397]
We show that a simple logistic regression classifier trained using the features sampled from our calibrated distribution can outperform the state-of-the-art accuracy on two datasets.
arXiv Detail & Related papers (2021-01-16T07:58:40Z) - Imbalanced Data Learning by Minority Class Augmentation using Capsule
Adversarial Networks [31.073558420480964]
We propose a method to restore the balance in imbalanced images, by coalescing two concurrent methods.
In our model, generative and discriminative networks play a novel competitive game.
The coalescing of capsule-GAN is effective at recognizing highly overlapping classes with much fewer parameters compared with the convolutional-GAN.
arXiv Detail & Related papers (2020-04-05T12:36:06Z) - Embedding Propagation: Smoother Manifold for Few-Shot Classification [131.81692677836202]
We propose to use embedding propagation as an unsupervised non-parametric regularizer for manifold smoothing in few-shot classification.
We empirically show that embedding propagation yields a smoother embedding manifold.
We show that embedding propagation consistently improves the accuracy of the models in multiple semi-supervised learning scenarios by up to 16% points.
arXiv Detail & Related papers (2020-03-09T13:51:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.