Global and Local Mixture Consistency Cumulative Learning for Long-tailed
Visual Recognitions
- URL: http://arxiv.org/abs/2305.08661v1
- Date: Mon, 15 May 2023 14:09:09 GMT
- Title: Global and Local Mixture Consistency Cumulative Learning for Long-tailed
Visual Recognitions
- Authors: Fei Du, Peng Yang and Qi Jia and Fengtao Nan and Xiaoting Chen and Yun
Yang
- Abstract summary: We propose an efficient one-stage training strategy for long-tailed visual recognition called Global and Local Mixture Consistency cumulative learning (GLMC)
Our approach achieves state-of-the-art accuracy on CIFAR10-LT, CIFAR100-LT, and ImageNet-LT datasets.
- Score: 8.925666400268502
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, our goal is to design a simple learning paradigm for long-tail
visual recognition, which not only improves the robustness of the feature
extractor but also alleviates the bias of the classifier towards head classes
while reducing the training skills and overhead. We propose an efficient
one-stage training strategy for long-tailed visual recognition called Global
and Local Mixture Consistency cumulative learning (GLMC). Our core ideas are
twofold: (1) a global and local mixture consistency loss improves the
robustness of the feature extractor. Specifically, we generate two augmented
batches by the global MixUp and local CutMix from the same batch data,
respectively, and then use cosine similarity to minimize the difference. (2) A
cumulative head tail soft label reweighted loss mitigates the head class bias
problem. We use empirical class frequencies to reweight the mixed label of the
head-tail class for long-tailed data and then balance the conventional loss and
the rebalanced loss with a coefficient accumulated by epochs. Our approach
achieves state-of-the-art accuracy on CIFAR10-LT, CIFAR100-LT, and ImageNet-LT
datasets. Additional experiments on balanced ImageNet and CIFAR demonstrate
that GLMC can significantly improve the generalization of backbones. Code is
made publicly available at https://github.com/ynu-yangpeng/GLMC.
Related papers
- Generalized Kullback-Leibler Divergence Loss [105.66549870868971]
We prove that the Kullback-Leibler (KL) Divergence loss is equivalent to the Decoupled Kullback-Leibler (DKL) Divergence loss.
Thanks to the decoupled structure of DKL loss, we have identified two areas for improvement.
arXiv Detail & Related papers (2025-03-11T04:43:33Z) - Class Instance Balanced Learning for Long-Tailed Classification [0.0]
Long-tailed image classification task deals with large imbalances in the class frequencies of the training data.
Previous approaches have shown that combining cross-entropy and contrastive learning can improve performance on the long-tailed task.
We propose a novel class instance balanced loss (CIBL), which reweights the relative contributions of a cross-entropy and a contrastive loss as a function of the frequency of class instances in the training batch.
arXiv Detail & Related papers (2023-07-11T15:09:10Z) - Integrating Local Real Data with Global Gradient Prototypes for
Classifier Re-Balancing in Federated Long-Tailed Learning [60.41501515192088]
Federated Learning (FL) has become a popular distributed learning paradigm that involves multiple clients training a global model collaboratively.
The data samples usually follow a long-tailed distribution in the real world, and FL on the decentralized and long-tailed data yields a poorly-behaved global model.
In this work, we integrate the local real data with the global gradient prototypes to form the local balanced datasets.
arXiv Detail & Related papers (2023-01-25T03:18:10Z) - Compound Batch Normalization for Long-tailed Image Classification [77.42829178064807]
We propose a compound batch normalization method based on a Gaussian mixture.
It can model the feature space more comprehensively and reduce the dominance of head classes.
The proposed method outperforms existing methods on long-tailed image classification.
arXiv Detail & Related papers (2022-12-02T07:31:39Z) - Towards Calibrated Hyper-Sphere Representation via Distribution Overlap
Coefficient for Long-tailed Learning [8.208237033120492]
Long-tailed learning aims to tackle the challenge that head classes dominate the training procedure under severe class imbalance in real-world scenarios.
Motivated by this, we generalize the cosine-based classifiers to a von Mises-Fisher (vMF) mixture model.
We measure representation quality upon the hyper-sphere space via calculating distribution overlap coefficient.
arXiv Detail & Related papers (2022-08-22T03:53:29Z) - Long-Tailed Classification with Gradual Balanced Loss and Adaptive
Feature Generation [19.17617301462919]
We propose a new method, Gradual Balanced Loss and Adaptive Feature Generator (GLAG) to alleviate imbalance.
State-of-the-art results have been achieved on long-tail datasets such as CIFAR100-LT, ImageNetLT, and iNaturalist.
arXiv Detail & Related papers (2022-02-28T01:20:35Z) - Improving Contrastive Learning on Imbalanced Seed Data via Open-World
Sampling [96.8742582581744]
We present an open-world unlabeled data sampling framework called Model-Aware K-center (MAK)
MAK follows three simple principles: tailness, proximity, and diversity.
We demonstrate that MAK can consistently improve both the overall representation quality and the class balancedness of the learned features.
arXiv Detail & Related papers (2021-11-01T15:09:41Z) - Calibrating Class Activation Maps for Long-Tailed Visual Recognition [60.77124328049557]
We present two effective modifications of CNNs to improve network learning from long-tailed distribution.
First, we present a Class Activation Map (CAMC) module to improve the learning and prediction of network classifiers.
Second, we investigate the use of normalized classifiers for representation learning in long-tailed problems.
arXiv Detail & Related papers (2021-08-29T05:45:03Z) - Distributional Robustness Loss for Long-tail Learning [20.800627115140465]
Real-world data is often unbalanced and long-tailed, but deep models struggle to recognize rare classes in the presence of frequent classes.
We show that the feature extractor part of deep networks suffers greatly from this bias.
We propose a new loss based on robustness theory, which encourages the model to learn high-quality representations for both head and tail classes.
arXiv Detail & Related papers (2021-04-07T11:34:04Z) - ResLT: Residual Learning for Long-tailed Recognition [64.19728932445523]
We propose a more fundamental perspective for long-tailed recognition, i.e., from the aspect of parameter space.
We design the effective residual fusion mechanism -- with one main branch optimized to recognize images from all classes, another two residual branches are gradually fused and optimized to enhance images from medium+tail classes and tail classes respectively.
We test our method on several benchmarks, i.e., long-tailed version of CIFAR-10, CIFAR-100, Places, ImageNet, and iNaturalist 2018.
arXiv Detail & Related papers (2021-01-26T08:43:50Z) - Class-incremental Learning with Rectified Feature-Graph Preservation [24.098892115785066]
A central theme of this paper is to learn new classes that arrive in sequential phases over time.
We propose a weighted-Euclidean regularization for old knowledge preservation.
We show how it can work with binary cross-entropy to increase class separation for effective learning of new classes.
arXiv Detail & Related papers (2020-12-15T07:26:04Z) - Long-tailed Recognition by Routing Diverse Distribution-Aware Experts [64.71102030006422]
We propose a new long-tailed classifier called RoutIng Diverse Experts (RIDE)
It reduces the model variance with multiple experts, reduces the model bias with a distribution-aware diversity loss, reduces the computational cost with a dynamic expert routing module.
RIDE outperforms the state-of-the-art by 5% to 7% on CIFAR100-LT, ImageNet-LT and iNaturalist 2018 benchmarks.
arXiv Detail & Related papers (2020-10-05T06:53:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.