Stochastic smoothing of the top-K calibrated hinge loss for deep
imbalanced classification
- URL: http://arxiv.org/abs/2202.02193v1
- Date: Fri, 4 Feb 2022 15:39:32 GMT
- Title: Stochastic smoothing of the top-K calibrated hinge loss for deep
imbalanced classification
- Authors: Camille Garcin, Maximilien Servajean, Alexis Joly, Joseph Salmon
- Abstract summary: We introduce a top-K hinge loss inspired by recent developments on top-K losses.
Our proposal is based on the smoothing of the top-K operator building on the flexible "perturbed" framework.
We show that our loss function performs very well in the case of balanced datasets, while benefiting from a significantly lower computational time.
- Score: 8.189630642296416
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In modern classification tasks, the number of labels is getting larger and
larger, as is the size of the datasets encountered in practice. As the number
of classes increases, class ambiguity and class imbalance become more and more
problematic to achieve high top-1 accuracy. Meanwhile, Top-K metrics (metrics
allowing K guesses) have become popular, especially for performance reporting.
Yet, proposing top-K losses tailored for deep learning remains a challenge,
both theoretically and practically. In this paper we introduce a stochastic
top-K hinge loss inspired by recent developments on top-K calibrated losses.
Our proposal is based on the smoothing of the top-K operator building on the
flexible "perturbed optimizer" framework. We show that our loss function
performs very well in the case of balanced datasets, while benefiting from a
significantly lower computational time than the state-of-the-art top-K loss
function. In addition, we propose a simple variant of our loss for the
imbalanced case. Experiments on a heavy-tailed dataset show that our loss
function significantly outperforms other baseline loss functions.
Related papers
- LEARN: An Invex Loss for Outlier Oblivious Robust Online Optimization [56.67706781191521]
An adversary can introduce outliers by corrupting loss functions in an arbitrary number of k, unknown to the learner.
We present a robust online rounds optimization framework, where an adversary can introduce outliers by corrupting loss functions in an arbitrary number of k, unknown.
arXiv Detail & Related papers (2024-08-12T17:08:31Z) - OrCo: Towards Better Generalization via Orthogonality and Contrast for Few-Shot Class-Incremental Learning [57.43911113915546]
Few-Shot Class-Incremental Learning (FSCIL) introduces a paradigm in which the problem space expands with limited data.
FSCIL methods inherently face the challenge of catastrophic forgetting as data arrives incrementally.
We propose the OrCo framework built on two core principles: features' orthogonality in the representation space, and contrastive learning.
arXiv Detail & Related papers (2024-03-27T13:30:48Z) - Optimizing for ROC Curves on Class-Imbalanced Data by Training over a Family of Loss Functions [3.06506506650274]
Training reliable classifiers under severe class imbalance is a challenging problem in computer vision.
Recent work has proposed techniques that mitigate the effects of training under imbalance by modifying the loss functions or optimization methods.
We propose training over a family of loss functions, instead of a single loss function.
arXiv Detail & Related papers (2024-02-08T04:31:21Z) - Alternate Loss Functions for Classification and Robust Regression Can Improve the Accuracy of Artificial Neural Networks [6.452225158891343]
This paper shows that training speed and final accuracy of neural networks can significantly depend on the loss function used to train neural networks.
Two new classification loss functions that significantly improve performance on a wide variety of benchmark tasks are proposed.
arXiv Detail & Related papers (2023-03-17T12:52:06Z) - Optimizing Partial Area Under the Top-k Curve: Theory and Practice [151.5072746015253]
We develop a novel metric named partial Area Under the top-k Curve (AUTKC)
AUTKC has a better discrimination ability, and its Bayes optimal score function could give a correct top-K ranking with respect to the conditional probability.
We present an empirical surrogate risk minimization framework to optimize the proposed metric.
arXiv Detail & Related papers (2022-09-03T11:09:13Z) - Neural Collapse Inspired Attraction-Repulsion-Balanced Loss for
Imbalanced Learning [97.81549071978789]
We propose Attraction-Repulsion-Balanced Loss (ARB-Loss) to balance the different components of the gradients.
We perform experiments on the large-scale classification and segmentation datasets and our ARB-Loss can achieve state-of-the-art performance.
arXiv Detail & Related papers (2022-04-19T08:23:23Z) - Label Distributionally Robust Losses for Multi-class Classification:
Consistency, Robustness and Adaptivity [55.29408396918968]
We study a family of loss functions named label-distributionally robust (LDR) losses for multi-class classification.
Our contributions include both consistency and robustness by establishing top-$k$ consistency of LDR losses for multi-class classification.
We propose a new adaptive LDR loss that automatically adapts the individualized temperature parameter to the noise degree of class label of each instance.
arXiv Detail & Related papers (2021-12-30T00:27:30Z) - Striking the Right Balance: Recall Loss for Semantic Segmentation [24.047359482606307]
Class imbalance is a fundamental problem in computer vision applications such as semantic segmentation.
We propose a hard-class mining loss by reshaping the vanilla cross entropy loss.
We show that the novel recall loss changes gradually between the standard cross entropy loss and the inverse frequency weighted loss.
arXiv Detail & Related papers (2021-06-28T18:02:03Z) - Distributional Robustness Loss for Long-tail Learning [20.800627115140465]
Real-world data is often unbalanced and long-tailed, but deep models struggle to recognize rare classes in the presence of frequent classes.
We show that the feature extractor part of deep networks suffers greatly from this bias.
We propose a new loss based on robustness theory, which encourages the model to learn high-quality representations for both head and tail classes.
arXiv Detail & Related papers (2021-04-07T11:34:04Z) - PCPL: Predicate-Correlation Perception Learning for Unbiased Scene Graph
Generation [58.98802062945709]
We propose a novel Predicate-Correlation Perception Learning scheme to adaptively seek out appropriate loss weights.
Our PCPL framework is further equipped with a graph encoder module to better extract context features.
arXiv Detail & Related papers (2020-09-02T08:30:09Z) - Normalized Loss Functions for Deep Learning with Noisy Labels [39.32101898670049]
We show that the commonly used Cross Entropy (CE) loss is not robust to noisy labels.
We propose a framework to build robust loss functions called Active Passive Loss (APL)
arXiv Detail & Related papers (2020-06-24T08:25:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.