Binary Classification: Counterbalancing Class Imbalance by Applying
Regression Models in Combination with One-Sided Label Shifts
- URL: http://arxiv.org/abs/2011.14764v1
- Date: Mon, 30 Nov 2020 13:24:47 GMT
- Title: Binary Classification: Counterbalancing Class Imbalance by Applying
Regression Models in Combination with One-Sided Label Shifts
- Authors: Peter Bellmann, Heinke Hihn, Daniel A. Braun, Friedhelm Schwenker
- Abstract summary: We introduce a novel method, which addresses the issues of class imbalance.
We generate a set of negative and positive target labels, such that the corresponding regression task becomes balanced.
We evaluate our approach on a number of publicly available data sets and compare our proposed method to one of the most popular oversampling techniques.
- Score: 0.4970364068620607
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many real-world pattern recognition scenarios, such as in medical
applications, the corresponding classification tasks can be of an imbalanced
nature. In the current study, we focus on binary, imbalanced classification
tasks, i.e.~binary classification tasks in which one of the two classes is
under-represented (minority class) in comparison to the other class (majority
class). In the literature, many different approaches have been proposed, such
as under- or oversampling, to counter class imbalance. In the current work, we
introduce a novel method, which addresses the issues of class imbalance. To
this end, we first transfer the binary classification task to an equivalent
regression task. Subsequently, we generate a set of negative and positive
target labels, such that the corresponding regression task becomes balanced,
with respect to the redefined target label set. We evaluate our approach on a
number of publicly available data sets in combination with Support Vector
Machines. Moreover, we compare our proposed method to one of the most popular
oversampling techniques (SMOTE). Based on the detailed discussion of the
presented outcomes of our experimental evaluation, we provide promising ideas
for future research directions.
Related papers
- Learning with Complementary Labels Revisited: The Selected-Completely-at-Random Setting Is More Practical [66.57396042747706]
Complementary-label learning is a weakly supervised learning problem.
We propose a consistent approach that does not rely on the uniform distribution assumption.
We find that complementary-label learning can be expressed as a set of negative-unlabeled binary classification problems.
arXiv Detail & Related papers (2023-11-27T02:59:17Z) - Deep Imbalanced Regression via Hierarchical Classification Adjustment [50.19438850112964]
Regression tasks in computer vision are often formulated into classification by quantizing the target space into classes.
The majority of training samples lie in a head range of target values, while a minority of samples span a usually larger tail range.
We propose to construct hierarchical classifiers for solving imbalanced regression tasks.
Our novel hierarchical classification adjustment (HCA) for imbalanced regression shows superior results on three diverse tasks.
arXiv Detail & Related papers (2023-10-26T04:54:39Z) - Mutual Exclusive Modulator for Long-Tailed Recognition [12.706961256329572]
Long-tailed recognition is the task of learning high-performance classifiers given extremely imbalanced training samples between categories.
We introduce a mutual exclusive modulator which can estimate the probability of an image belonging to each group.
Our method achieves competitive performance compared to the state-of-the-art benchmarks.
arXiv Detail & Related papers (2023-02-19T07:31:49Z) - Prototypical Classifier for Robust Class-Imbalanced Learning [64.96088324684683]
We propose textitPrototypical, which does not require fitting additional parameters given the embedding network.
Prototypical produces balanced and comparable predictions for all classes even though the training set is class-imbalanced.
We test our method on CIFAR-10LT, CIFAR-100LT and Webvision datasets, observing that Prototypical obtains substaintial improvements compared with state of the arts.
arXiv Detail & Related papers (2021-10-22T01:55:01Z) - Equity-Directed Bootstrapping: Examples and Analysis [3.007949058551534]
We show how an equity-directed bootstrap can bring test set sensitivities and specificities closer to satisfying the equal odds criterion.
In the context of na"ive Bayes and logistic regression, we analyze the equity-directed bootstrap, demonstrating that it works by bringing odds ratios close to one.
arXiv Detail & Related papers (2021-08-14T22:09:27Z) - Statistical Theory for Imbalanced Binary Classification [8.93993657323783]
We show that optimal classification performance depends on certain properties of class imbalance that have not previously been formalized.
Specifically, we propose a novel sub-type of class imbalance, which we call Uniform Class Imbalance.
These results provide some of the first meaningful finite-sample statistical theory for imbalanced binary classification.
arXiv Detail & Related papers (2021-07-05T03:55:43Z) - Disentangling Sampling and Labeling Bias for Learning in Large-Output
Spaces [64.23172847182109]
We show that different negative sampling schemes implicitly trade-off performance on dominant versus rare labels.
We provide a unified means to explicitly tackle both sampling bias, arising from working with a subset of all labels, and labeling bias, which is inherent to the data due to label imbalance.
arXiv Detail & Related papers (2021-05-12T15:40:13Z) - SetConv: A New Approach for Learning from Imbalanced Data [29.366843553056594]
We propose a set convolution operation and an episodic training strategy to extract a single representative for each class.
We prove that our proposed algorithm is permutation-invariant despite the order of inputs.
arXiv Detail & Related papers (2021-04-03T22:33:30Z) - Towards Model-Agnostic Post-Hoc Adjustment for Balancing Ranking
Fairness and Algorithm Utility [54.179859639868646]
Bipartite ranking aims to learn a scoring function that ranks positive individuals higher than negative ones from labeled data.
There have been rising concerns on whether the learned scoring function can cause systematic disparity across different protected groups.
We propose a model post-processing framework for balancing them in the bipartite ranking scenario.
arXiv Detail & Related papers (2020-06-15T10:08:39Z) - M2m: Imbalanced Classification via Major-to-minor Translation [79.09018382489506]
In most real-world scenarios, labeled training datasets are highly class-imbalanced, where deep neural networks suffer from generalizing to a balanced testing criterion.
In this paper, we explore a novel yet simple way to alleviate this issue by augmenting less-frequent classes via translating samples from more-frequent classes.
Our experimental results on a variety of class-imbalanced datasets show that the proposed method improves the generalization on minority classes significantly compared to other existing re-sampling or re-weighting methods.
arXiv Detail & Related papers (2020-04-01T13:21:17Z) - Imbalanced classification: a paradigm-based review [21.578692329486643]
Multiple resampling techniques have been proposed to address the class imbalance issues.
There is no general guidance on when to use each technique.
We provide a paradigm-based review of the common resampling techniques for binary classification under imbalanced class sizes.
arXiv Detail & Related papers (2020-02-11T18:34:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.