Towards Calibrated Hyper-Sphere Representation via Distribution Overlap
Coefficient for Long-tailed Learning
- URL: http://arxiv.org/abs/2208.10043v1
- Date: Mon, 22 Aug 2022 03:53:29 GMT
- Title: Towards Calibrated Hyper-Sphere Representation via Distribution Overlap
Coefficient for Long-tailed Learning
- Authors: Hualiang Wang, Siming Fu, Xiaoxuan He, Hangxiang Fang, Zuozhu Liu,
Haoji Hu
- Abstract summary: Long-tailed learning aims to tackle the challenge that head classes dominate the training procedure under severe class imbalance in real-world scenarios.
Motivated by this, we generalize the cosine-based classifiers to a von Mises-Fisher (vMF) mixture model.
We measure representation quality upon the hyper-sphere space via calculating distribution overlap coefficient.
- Score: 8.208237033120492
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Long-tailed learning aims to tackle the crucial challenge that head classes
dominate the training procedure under severe class imbalance in real-world
scenarios. However, little attention has been given to how to quantify the
dominance severity of head classes in the representation space. Motivated by
this, we generalize the cosine-based classifiers to a von Mises-Fisher (vMF)
mixture model, denoted as vMF classifier, which enables to quantitatively
measure representation quality upon the hyper-sphere space via calculating
distribution overlap coefficient. To our knowledge, this is the first work to
measure representation quality of classifiers and features from the perspective
of distribution overlap coefficient. On top of it, we formulate the inter-class
discrepancy and class-feature consistency loss terms to alleviate the
interference among the classifier weights and align features with classifier
weights. Furthermore, a novel post-training calibration algorithm is devised to
zero-costly boost the performance via inter-class overlap coefficients. Our
method outperforms previous work with a large margin and achieves
state-of-the-art performance on long-tailed image classification, semantic
segmentation, and instance segmentation tasks (e.g., we achieve 55.0\% overall
accuracy with ResNetXt-50 in ImageNet-LT). Our code is available at
https://github.com/VipaiLab/vMF\_OP.
Related papers
- SFC: Shared Feature Calibration in Weakly Supervised Semantic
Segmentation [28.846513129022803]
Image-level weakly supervised semantic segmentation has received increasing attention due to its low annotation cost.
Existing methods mainly rely on Class Mapping (CAM) to obtain pseudo-labels for training semantic segmentation models.
In this work, we are the first to demonstrate that long-tailed distribution in training data can cause the CAM calculated through weights over-activated for head classes and under-activated for tail classes due to the shared features among head- and tail- classes.
arXiv Detail & Related papers (2024-01-22T06:43:13Z) - Class-Imbalanced Semi-Supervised Learning for Large-Scale Point Cloud
Semantic Segmentation via Decoupling Optimization [64.36097398869774]
Semi-supervised learning (SSL) has been an active research topic for large-scale 3D scene understanding.
The existing SSL-based methods suffer from severe training bias due to class imbalance and long-tail distributions of the point cloud data.
We introduce a new decoupling optimization framework, which disentangles feature representation learning and classifier in an alternative optimization manner to shift the bias decision boundary effectively.
arXiv Detail & Related papers (2024-01-13T04:16:40Z) - Gramian Attention Heads are Strong yet Efficient Vision Learners [26.79263390835444]
We introduce a novel architecture design that enhances expressiveness by incorporating multiple head classifiers (ie, classification heads)
Our approach employs attention-based aggregation, utilizing pairwise feature similarity to enhance multiple lightweight heads with minimal resource overhead.
Our models eventually surpass state-of-the-art CNNs and ViTs regarding the accuracy-grained trade-off on ImageNet-1K.
arXiv Detail & Related papers (2023-10-25T09:08:58Z) - Class Instance Balanced Learning for Long-Tailed Classification [0.0]
Long-tailed image classification task deals with large imbalances in the class frequencies of the training data.
Previous approaches have shown that combining cross-entropy and contrastive learning can improve performance on the long-tailed task.
We propose a novel class instance balanced loss (CIBL), which reweights the relative contributions of a cross-entropy and a contrastive loss as a function of the frequency of class instances in the training batch.
arXiv Detail & Related papers (2023-07-11T15:09:10Z) - Maximally Compact and Separated Features with Regular Polytope Networks [22.376196701232388]
We show how to extract from CNNs features the properties of emphmaximum inter-class separability and emphmaximum intra-class compactness.
We obtain features similar to what can be obtained with the well-known citewen2016discriminative and other similar approaches.
arXiv Detail & Related papers (2023-01-15T15:20:57Z) - Leveraging Angular Information Between Feature and Classifier for
Long-tailed Learning: A Prediction Reformulation Approach [90.77858044524544]
We reformulate the recognition probabilities through included angles without re-balancing the classifier weights.
Inspired by the performance improvement of the predictive form reformulation, we explore the different properties of this angular prediction.
Our method is able to obtain the best performance among peer methods without pretraining on CIFAR10/100-LT and ImageNet-LT.
arXiv Detail & Related papers (2022-12-03T07:52:48Z) - Compound Batch Normalization for Long-tailed Image Classification [77.42829178064807]
We propose a compound batch normalization method based on a Gaussian mixture.
It can model the feature space more comprehensively and reduce the dominance of head classes.
The proposed method outperforms existing methods on long-tailed image classification.
arXiv Detail & Related papers (2022-12-02T07:31:39Z) - Prediction Calibration for Generalized Few-shot Semantic Segmentation [101.69940565204816]
Generalized Few-shot Semantic (GFSS) aims to segment each image pixel into either base classes with abundant training examples or novel classes with only a handful of (e.g., 1-5) training images per class.
We build a cross-attention module that guides the classifier's final prediction using the fused multi-level features.
Our PCN outperforms the state-the-art alternatives by large margins.
arXiv Detail & Related papers (2022-10-15T13:30:12Z) - You Only Need End-to-End Training for Long-Tailed Recognition [8.789819609485225]
Cross-entropy loss tends to produce highly correlated features on imbalanced data.
We propose two novel modules, Block-based Relatively Balanced Batch Sampler (B3RS) and Batch Embedded Training (BET)
Experimental results on the long-tailed classification benchmarks, CIFAR-LT and ImageNet-LT, demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2021-12-11T11:44:09Z) - Calibrating Class Activation Maps for Long-Tailed Visual Recognition [60.77124328049557]
We present two effective modifications of CNNs to improve network learning from long-tailed distribution.
First, we present a Class Activation Map (CAMC) module to improve the learning and prediction of network classifiers.
Second, we investigate the use of normalized classifiers for representation learning in long-tailed problems.
arXiv Detail & Related papers (2021-08-29T05:45:03Z) - Prior Guided Feature Enrichment Network for Few-Shot Segmentation [64.91560451900125]
State-of-the-art semantic segmentation methods require sufficient labeled data to achieve good results.
Few-shot segmentation is proposed to tackle this problem by learning a model that quickly adapts to new classes with a few labeled support samples.
Theses frameworks still face the challenge of generalization ability reduction on unseen classes due to inappropriate use of high-level semantic information.
arXiv Detail & Related papers (2020-08-04T10:41:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.