Long-tailed Recognition by Routing Diverse Distribution-Aware Experts
- URL: http://arxiv.org/abs/2010.01809v4
- Date: Sun, 1 May 2022 18:00:20 GMT
- Title: Long-tailed Recognition by Routing Diverse Distribution-Aware Experts
- Authors: Xudong Wang, Long Lian, Zhongqi Miao, Ziwei Liu, Stella X. Yu
- Abstract summary: We propose a new long-tailed classifier called RoutIng Diverse Experts (RIDE)
It reduces the model variance with multiple experts, reduces the model bias with a distribution-aware diversity loss, reduces the computational cost with a dynamic expert routing module.
RIDE outperforms the state-of-the-art by 5% to 7% on CIFAR100-LT, ImageNet-LT and iNaturalist 2018 benchmarks.
- Score: 64.71102030006422
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Natural data are often long-tail distributed over semantic classes. Existing
recognition methods tackle this imbalanced classification by placing more
emphasis on the tail data, through class re-balancing/re-weighting or
ensembling over different data groups, resulting in increased tail accuracies
but reduced head accuracies.
We take a dynamic view of the training data and provide a principled model
bias and variance analysis as the training data fluctuates: Existing long-tail
classifiers invariably increase the model variance and the head-tail model bias
gap remains large, due to more and larger confusion with hard negatives for the
tail.
We propose a new long-tailed classifier called RoutIng Diverse Experts
(RIDE). It reduces the model variance with multiple experts, reduces the model
bias with a distribution-aware diversity loss, reduces the computational cost
with a dynamic expert routing module. RIDE outperforms the state-of-the-art by
5% to 7% on CIFAR100-LT, ImageNet-LT and iNaturalist 2018 benchmarks. It is
also a universal framework that is applicable to various backbone networks,
long-tailed algorithms, and training mechanisms for consistent performance
gains. Our code is available at:
https://github.com/frank-xwang/RIDE-LongTailRecognition.
Related papers
- A dual-branch model with inter- and intra-branch contrastive loss for
long-tailed recognition [7.225494453600985]
Models trained on long-tailed datasets have poor adaptability to tail classes and the decision boundaries are ambiguous.
We propose a simple yet effective model, named Dual-Branch Long-Tailed Recognition (DB-LTR), which includes an imbalanced learning branch and a Contrastive Learning Branch (CoLB)
CoLB can improve the capability of the model in adapting to tail classes and assist the imbalanced learning branch to learn a well-represented feature space and discriminative decision boundary.
arXiv Detail & Related papers (2023-09-28T03:31:11Z) - Class Instance Balanced Learning for Long-Tailed Classification [0.0]
Long-tailed image classification task deals with large imbalances in the class frequencies of the training data.
Previous approaches have shown that combining cross-entropy and contrastive learning can improve performance on the long-tailed task.
We propose a novel class instance balanced loss (CIBL), which reweights the relative contributions of a cross-entropy and a contrastive loss as a function of the frequency of class instances in the training batch.
arXiv Detail & Related papers (2023-07-11T15:09:10Z) - Long-Tailed Visual Recognition via Self-Heterogeneous Integration with
Knowledge Excavation [53.94265240561697]
We propose a novel MoE-based method called Self-Heterogeneous Integration with Knowledge Excavation (SHIKE)
SHIKE achieves the state-of-the-art performance of 56.3%, 60.3%, 75.4%, and 41.9% on CIFAR100-LT (IF100), ImageNet-LT, iNaturalist 2018, and Places-LT, respectively.
arXiv Detail & Related papers (2023-04-03T18:13:39Z) - Improving GANs for Long-Tailed Data through Group Spectral
Regularization [51.58250647277375]
We propose a novel group Spectral Regularizer (gSR) that prevents the spectral explosion alleviating mode collapse.
We find that gSR effectively combines with existing augmentation and regularization techniques, leading to state-of-the-art image generation performance on long-tailed data.
arXiv Detail & Related papers (2022-08-21T17:51:05Z) - Constructing Balance from Imbalance for Long-tailed Image Recognition [50.6210415377178]
The imbalance between majority (head) classes and minority (tail) classes severely skews the data-driven deep neural networks.
Previous methods tackle with data imbalance from the viewpoints of data distribution, feature space, and model design.
We propose a concise paradigm by progressively adjusting label space and dividing the head classes and tail classes.
Our proposed model also provides a feature evaluation method and paves the way for long-tailed feature learning.
arXiv Detail & Related papers (2022-08-04T10:22:24Z) - Learning Muti-expert Distribution Calibration for Long-tailed Video
Classification [88.12433458277168]
We propose an end-to-end multi-experts distribution calibration method based on two-level distribution information.
By modeling this two-level distribution information, the model can consider the head classes and the tail classes.
Our method achieves state-of-the-art performance on the long-tailed video classification task.
arXiv Detail & Related papers (2022-05-22T09:52:34Z) - Class Balancing GAN with a Classifier in the Loop [58.29090045399214]
We introduce a novel theoretically motivated Class Balancing regularizer for training GANs.
Our regularizer makes use of the knowledge from a pre-trained classifier to ensure balanced learning of all the classes in the dataset.
We demonstrate the utility of our regularizer in learning representations for long-tailed distributions via achieving better performance than existing approaches over multiple datasets.
arXiv Detail & Related papers (2021-06-17T11:41:30Z) - Distributional Robustness Loss for Long-tail Learning [20.800627115140465]
Real-world data is often unbalanced and long-tailed, but deep models struggle to recognize rare classes in the presence of frequent classes.
We show that the feature extractor part of deep networks suffers greatly from this bias.
We propose a new loss based on robustness theory, which encourages the model to learn high-quality representations for both head and tail classes.
arXiv Detail & Related papers (2021-04-07T11:34:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.