Long-Tailed Learning as Multi-Objective Optimization
- URL: http://arxiv.org/abs/2310.20490v2
- Date: Wed, 1 Nov 2023 11:28:55 GMT
- Title: Long-Tailed Learning as Multi-Objective Optimization
- Authors: Weiqi Li, Fan Lyu, Fanhua Shang, Liang Wan, Wei Feng
- Abstract summary: We argue that the seesaw dilemma is derived from gradient imbalance of different classes.
We propose a Gradient-Balancing Grouping (GBG) strategy to gather the classes with similar gradient directions.
- Score: 29.012779934262973
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Real-world data is extremely imbalanced and presents a long-tailed
distribution, resulting in models that are biased towards classes with
sufficient samples and perform poorly on rare classes. Recent methods propose
to rebalance classes but they undertake the seesaw dilemma (what is increasing
performance on tail classes may decrease that of head classes, and vice versa).
In this paper, we argue that the seesaw dilemma is derived from gradient
imbalance of different classes, in which gradients of inappropriate classes are
set to important for updating, thus are prone to overcompensation or
undercompensation on tail classes. To achieve ideal compensation, we formulate
the long-tailed recognition as an multi-objective optimization problem, which
fairly respects the contributions of head and tail classes simultaneously. For
efficiency, we propose a Gradient-Balancing Grouping (GBG) strategy to gather
the classes with similar gradient directions, thus approximately make every
update under a Pareto descent direction. Our GBG method drives classes with
similar gradient directions to form more representative gradient and provide
ideal compensation to the tail classes. Moreover, We conduct extensive
experiments on commonly used benchmarks in long-tailed learning and demonstrate
the superiority of our method over existing SOTA methods.
Related papers
- Gradient-Aware Logit Adjustment Loss for Long-tailed Classifier [30.931850375858573]
In the real-world setting, data often follows a long-tailed distribution, where head classes contain significantly more training samples than tail classes.
We propose the Gradient-Aware Logit Adjustment (GALA) loss, which adjusts the logits based on accumulated gradients to balance the optimization process.
Our approach achieves top-1 accuracy of 48.5%, 41.4%, and 73.3% on popular long-tailed recognition benchmark datasets.
arXiv Detail & Related papers (2024-03-14T02:21:01Z) - Dual Compensation Residual Networks for Class Imbalanced Learning [98.35401757647749]
We propose Dual Compensation Residual Networks to better fit both tail and head classes.
An important factor causing overfitting is that there is severe feature drift between training and test data on tail classes.
We also propose a Residual Balanced Multi-Proxies classifier to alleviate the under-fitting issue.
arXiv Detail & Related papers (2023-08-25T04:06:30Z) - The Equalization Losses: Gradient-Driven Training for Long-tailed Object
Recognition [84.51875325962061]
We propose a gradient-driven training mechanism to tackle the long-tail problem.
We introduce a new family of gradient-driven loss functions, namely equalization losses.
Our method consistently outperforms the baseline models.
arXiv Detail & Related papers (2022-10-11T16:00:36Z) - Improving GANs for Long-Tailed Data through Group Spectral
Regularization [51.58250647277375]
We propose a novel group Spectral Regularizer (gSR) that prevents the spectral explosion alleviating mode collapse.
We find that gSR effectively combines with existing augmentation and regularization techniques, leading to state-of-the-art image generation performance on long-tailed data.
arXiv Detail & Related papers (2022-08-21T17:51:05Z) - Constructing Balance from Imbalance for Long-tailed Image Recognition [50.6210415377178]
The imbalance between majority (head) classes and minority (tail) classes severely skews the data-driven deep neural networks.
Previous methods tackle with data imbalance from the viewpoints of data distribution, feature space, and model design.
We propose a concise paradigm by progressively adjusting label space and dividing the head classes and tail classes.
Our proposed model also provides a feature evaluation method and paves the way for long-tailed feature learning.
arXiv Detail & Related papers (2022-08-04T10:22:24Z) - A Theoretical Analysis of the Learning Dynamics under Class Imbalance [0.10231119246773925]
We show that the learning curves for minority and majority classes follow sub-optimal trajectories when training with a gradient-based trajectory.
This slowdown is related to the imbalance ratio and can be traced back to a competition between the optimization of different classes.
We find that GD is not guaranteed to decrease the loss for each class but that this problem can be addressed by performing a per-class normalization of the gradient.
arXiv Detail & Related papers (2022-07-01T12:54:38Z) - Neural Collapse Inspired Attraction-Repulsion-Balanced Loss for
Imbalanced Learning [97.81549071978789]
We propose Attraction-Repulsion-Balanced Loss (ARB-Loss) to balance the different components of the gradients.
We perform experiments on the large-scale classification and segmentation datasets and our ARB-Loss can achieve state-of-the-art performance.
arXiv Detail & Related papers (2022-04-19T08:23:23Z) - Relieving Long-tailed Instance Segmentation via Pairwise Class Balance [85.53585498649252]
Long-tailed instance segmentation is a challenging task due to the extreme imbalance of training samples among classes.
It causes severe biases of the head classes (with majority samples) against the tailed ones.
We propose a novel Pairwise Class Balance (PCB) method, built upon a confusion matrix which is updated during training to accumulate the ongoing prediction preferences.
arXiv Detail & Related papers (2022-01-08T07:48:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.