Deep Representation Learning on Long-tailed Data: A Learnable Embedding
Augmentation Perspective
- URL: http://arxiv.org/abs/2002.10826v3
- Date: Sun, 12 Apr 2020 16:32:33 GMT
- Title: Deep Representation Learning on Long-tailed Data: A Learnable Embedding
Augmentation Perspective
- Authors: Jialun Liu, Yifan Sun, Chuchu Han, Zhaopeng Dou, Wenhui Li
- Abstract summary: In the deep feature space, the head classes and the tail classes present different distribution patterns.
We propose to construct each feature into a "feature cloud"
It allows each tail sample to push the samples from other classes far away, recovering the intra-class diversity of tail classes.
- Score: 17.602607883721973
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper considers learning deep features from long-tailed data. We observe
that in the deep feature space, the head classes and the tail classes present
different distribution patterns. The head classes have a relatively large
spatial span, while the tail classes have significantly small spatial span, due
to the lack of intra-class diversity. This uneven distribution between head and
tail classes distorts the overall feature space, which compromises the
discriminative ability of the learned features. Intuitively, we seek to expand
the distribution of the tail classes by transferring from the head classes, so
as to alleviate the distortion of the feature space. To this end, we propose to
construct each feature into a "feature cloud". If a sample belongs to a tail
class, the corresponding feature cloud will have relatively large distribution
range, in compensation to its lack of diversity. It allows each tail sample to
push the samples from other classes far away, recovering the intra-class
diversity of tail classes. Extensive experimental evaluations on person
re-identification and face recognition tasks confirm the effectiveness of our
method.
Related papers
- Granularity Matters in Long-Tail Learning [62.30734737735273]
We offer a novel perspective on long-tail learning, inspired by an observation: datasets with finer granularity tend to be less affected by data imbalance.
We introduce open-set auxiliary classes that are visually similar to existing ones, aiming to enhance representation learning for both head and tail classes.
To prevent the overwhelming presence of auxiliary classes from disrupting training, we introduce a neighbor-silencing loss.
arXiv Detail & Related papers (2024-10-21T13:06:21Z) - Dual Compensation Residual Networks for Class Imbalanced Learning [98.35401757647749]
We propose Dual Compensation Residual Networks to better fit both tail and head classes.
An important factor causing overfitting is that there is severe feature drift between training and test data on tail classes.
We also propose a Residual Balanced Multi-Proxies classifier to alleviate the under-fitting issue.
arXiv Detail & Related papers (2023-08-25T04:06:30Z) - Feature Fusion from Head to Tail for Long-Tailed Visual Recognition [39.86973663532936]
The biased decision boundary caused by inadequate semantic information in tail classes is one of the key factors contributing to their low recognition accuracy.
We propose to augment tail classes by grafting the diverse semantic information from head classes, referred to as head-to-tail fusion (H2T)
Both theoretical analysis and practical experimentation demonstrate that H2T can contribute to a more optimized solution for the decision boundary.
arXiv Detail & Related papers (2023-06-12T08:50:46Z) - Balancing Logit Variation for Long-tailed Semantic Segmentation [28.92929059563813]
We introduce category-wise variation into the network predictions in the training phase.
We close the gap between the feature areas of different categories, resulting in a more balanced representation.
Our method manifests itself in strong generalizability to various datasets and task settings.
arXiv Detail & Related papers (2023-06-03T09:19:24Z) - Adjusting Logit in Gaussian Form for Long-Tailed Visual Recognition [37.62659619941791]
We study the problem of long-tailed visual recognition from the perspective of feature level.
Two novel logit adjustment methods are proposed to improve model performance at a modest computational overhead.
Experiments conducted on benchmark datasets demonstrate the superior performance of the proposed method over the state-of-the-art ones.
arXiv Detail & Related papers (2023-05-18T02:06:06Z) - Constructing Balance from Imbalance for Long-tailed Image Recognition [50.6210415377178]
The imbalance between majority (head) classes and minority (tail) classes severely skews the data-driven deep neural networks.
Previous methods tackle with data imbalance from the viewpoints of data distribution, feature space, and model design.
We propose a concise paradigm by progressively adjusting label space and dividing the head classes and tail classes.
Our proposed model also provides a feature evaluation method and paves the way for long-tailed feature learning.
arXiv Detail & Related papers (2022-08-04T10:22:24Z) - Learning Muti-expert Distribution Calibration for Long-tailed Video
Classification [88.12433458277168]
We propose an end-to-end multi-experts distribution calibration method based on two-level distribution information.
By modeling this two-level distribution information, the model can consider the head classes and the tail classes.
Our method achieves state-of-the-art performance on the long-tailed video classification task.
arXiv Detail & Related papers (2022-05-22T09:52:34Z) - Feature Generation for Long-tail Classification [36.186909933006675]
We show how to generate meaningful features by estimating the tail category's distribution.
We also present a qualitative analysis of generated features using t-SNE visualizations and analyze the nearest neighbors used to calibrate the tail class distributions.
arXiv Detail & Related papers (2021-11-10T21:34:29Z) - Learning Debiased and Disentangled Representations for Semantic
Segmentation [52.35766945827972]
We propose a model-agnostic and training scheme for semantic segmentation.
By randomly eliminating certain class information in each training iteration, we effectively reduce feature dependencies among classes.
Models trained with our approach demonstrate strong results on multiple semantic segmentation benchmarks.
arXiv Detail & Related papers (2021-10-31T16:15:09Z) - Improving Tail-Class Representation with Centroid Contrastive Learning [145.73991900239017]
We propose interpolative centroid contrastive learning (ICCL) to improve long-tailed representation learning.
ICCL interpolates two images from a class-agnostic sampler and a class-aware sampler, and trains the model such that the representation of the ICCL can be used to retrieve the centroids for both source classes.
Our result shows a significant accuracy gain of 2.8% on the iNaturalist 2018 dataset with a real-world long-tailed distribution.
arXiv Detail & Related papers (2021-10-19T15:24:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.