BalanceBenchmark: A Survey for Imbalanced Learning
- URL: http://arxiv.org/abs/2502.10816v2
- Date: Tue, 18 Feb 2025 04:47:55 GMT
- Title: BalanceBenchmark: A Survey for Imbalanced Learning
- Authors: Shaoxuan Xu, Menglu Cui, Chengxiang Huang, Hongfa Wang, DiHu,
- Abstract summary: Multimodal learning has gained attention for its capacity to integrate information from different modalities.
It is often hindered by the multimodal imbalance problem, where certain modality dominates while others remain underutilized.
We systematically categorize various mainstream multimodal imbalance algorithms into four groups based on the strategies they employ to mitigate imbalance.
- Score: 9.858467766666223
- License:
- Abstract: Multimodal learning has gained attention for its capacity to integrate information from different modalities. However, it is often hindered by the multimodal imbalance problem, where certain modality dominates while others remain underutilized. Although recent studies have proposed various methods to alleviate this problem, they lack comprehensive and fair comparisons. In this paper, we systematically categorize various mainstream multimodal imbalance algorithms into four groups based on the strategies they employ to mitigate imbalance. To facilitate a comprehensive evaluation of these methods, we introduce BalanceBenchmark, a benchmark including multiple widely used multidimensional datasets and evaluation metrics from three perspectives: performance, imbalance degree, and complexity. To ensure fair comparisons, we have developed a modular and extensible toolkit that standardizes the experimental workflow across different methods. Based on the experiments using BalanceBenchmark, we have identified several key insights into the characteristics and advantages of different method groups in terms of performance, balance degree and computational complexity. We expect such analysis could inspire more efficient approaches to address the imbalance problem in the future, as well as foundation models. The code of the toolkit is available at https://github.com/GeWu-Lab/BalanceBenchmark.
Related papers
- SeMi: When Imbalanced Semi-Supervised Learning Meets Mining Hard Examples [54.760757107700755]
Semi-Supervised Learning (SSL) can leverage abundant unlabeled data to boost model performance.
The class-imbalanced data distribution in real-world scenarios poses great challenges to SSL, resulting in performance degradation.
We propose a method that enhances the performance of Imbalanced Semi-Supervised Learning by Mining Hard Examples (SeMi)
arXiv Detail & Related papers (2025-01-10T14:35:16Z) - POGEMA: A Benchmark Platform for Cooperative Multi-Agent Navigation [76.67608003501479]
We introduce and specify an evaluation protocol defining a range of domain-related metrics computed on the basics of the primary evaluation indicators.
The results of such a comparison, which involves a variety of state-of-the-art MARL, search-based, and hybrid methods, are presented.
arXiv Detail & Related papers (2024-07-20T16:37:21Z) - Balanced Audiovisual Dataset for Imbalance Analysis [31.510912639133014]
The imbalance problem is widespread in the field of machine learning, which also exists in multimodal learning areas.
Recent works have attempted to solve the modality imbalance problem from algorithm perspective, however, they do not fully analyze the influence of modality bias in datasets.
arXiv Detail & Related papers (2023-02-14T15:35:17Z) - Revisiting Long-tailed Image Classification: Survey and Benchmarks with
New Evaluation Metrics [88.39382177059747]
A corpus of metrics is designed for measuring the accuracy, robustness, and bounds of algorithms for learning with long-tailed distribution.
Based on our benchmarks, we re-evaluate the performance of existing methods on CIFAR10 and CIFAR100 datasets.
arXiv Detail & Related papers (2023-02-03T02:40:54Z) - ASE: Anomaly Scoring Based Ensemble Learning for Imbalanced Datasets [3.214208422566496]
We come up with a bagging ensemble learning framework based on an anomaly detection scoring system.
We test out that our ensemble learning model can dramatically improve performance of base estimators.
arXiv Detail & Related papers (2022-03-21T07:20:41Z) - On Modality Bias Recognition and Reduction [70.69194431713825]
We study the modality bias problem in the context of multi-modal classification.
We propose a plug-and-play loss function method, whereby the feature space for each label is adaptively learned.
Our method yields remarkable performance improvements compared with the baselines.
arXiv Detail & Related papers (2022-02-25T13:47:09Z) - Towards Inter-class and Intra-class Imbalance in Class-imbalanced
Learning [24.01370257373491]
Imbalanced Learning (IL) is an important problem that widely exists in data mining applications.
We present Duple-Balanced Ensemble, namely DUBE, a versatile ensemble learning framework.
Unlike prevailing methods, DUBE directly performs inter-class and intra-class balancing without relying on heavy distance-based computation.
arXiv Detail & Related papers (2021-11-24T20:50:54Z) - Few-Shot Learning with Class Imbalance [13.60699610822265]
Few-shot learning aims to train models on a limited number of labeled samples given in a support set in order to generalize to unseen samples from a query set.
In the standard setup, the support set contains an equal amount of data points for each class.
We present a detailed study of few-shot class-imbalance along three axes: meta-dataset vs. task imbalance, effect of different imbalance distributions (linear, step, random), and effect of rebalancing techniques.
arXiv Detail & Related papers (2021-01-07T12:54:32Z) - Towards Model-Agnostic Post-Hoc Adjustment for Balancing Ranking
Fairness and Algorithm Utility [54.179859639868646]
Bipartite ranking aims to learn a scoring function that ranks positive individuals higher than negative ones from labeled data.
There have been rising concerns on whether the learned scoring function can cause systematic disparity across different protected groups.
We propose a model post-processing framework for balancing them in the bipartite ranking scenario.
arXiv Detail & Related papers (2020-06-15T10:08:39Z) - Long-Tailed Recognition Using Class-Balanced Experts [128.73438243408393]
We propose an ensemble of class-balanced experts that combines the strength of diverse classifiers.
Our ensemble of class-balanced experts reaches results close to state-of-the-art and an extended ensemble establishes a new state-of-the-art on two benchmarks for long-tailed recognition.
arXiv Detail & Related papers (2020-04-07T20:57:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.