Review of Methods for Handling Class-Imbalanced in Classification
Problems
- URL: http://arxiv.org/abs/2211.05456v1
- Date: Thu, 10 Nov 2022 10:07:10 GMT
- Title: Review of Methods for Handling Class-Imbalanced in Classification
Problems
- Authors: Satyendra Singh Rawat (Amity University, Gwalior, India), Amit Kumar
Mishra (Amity University, Gwalior, India)
- Abstract summary: In some cases, one class contains the majority of examples while the other, which is frequently the more important class, is nevertheless represented by a smaller proportion of examples.
The article examines the most widely used methods for addressing the problem of learning with a class imbalance, including data-level, algorithm-level, hybrid, cost-sensitive learning, and deep learning.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learning classifiers using skewed or imbalanced datasets can occasionally
lead to classification issues; this is a serious issue. In some cases, one
class contains the majority of examples while the other, which is frequently
the more important class, is nevertheless represented by a smaller proportion
of examples. Using this kind of data could make many carefully designed
machine-learning systems ineffective. High training fidelity was a term used to
describe biases vs. all other instances of the class. The best approach to all
possible remedies to this issue is typically to gain from the minority class.
The article examines the most widely used methods for addressing the problem of
learning with a class imbalance, including data-level, algorithm-level, hybrid,
cost-sensitive learning, and deep learning, etc. including their advantages and
limitations. The efficiency and performance of the classifier are assessed
using a myriad of evaluation metrics.
Related papers
- Unsupervised Estimation of Ensemble Accuracy [0.0]
We present a method for estimating the joint power of several classifiers.
It differs from existing approaches which focus on "diversity" measures by not relying on labels.
We demonstrate the method on popular large-scale face recognition datasets.
arXiv Detail & Related papers (2023-11-18T02:31:36Z) - Class-Incremental Learning: A Survey [84.30083092434938]
Class-Incremental Learning (CIL) enables the learner to incorporate the knowledge of new classes incrementally.
CIL tends to catastrophically forget the characteristics of former ones, and its performance drastically degrades.
We provide a rigorous and unified evaluation of 17 methods in benchmark image classification tasks to find out the characteristics of different algorithms.
arXiv Detail & Related papers (2023-02-07T17:59:05Z) - Multi-Label Quantification [78.83284164605473]
Quantification, variously called "labelled prevalence estimation" or "learning to quantify", is the supervised learning task of generating predictors of the relative frequencies of the classes of interest in unsupervised data samples.
We propose methods for inferring estimators of class prevalence values that strive to leverage the dependencies among the classes of interest in order to predict their relative frequencies more accurately.
arXiv Detail & Related papers (2022-11-15T11:29:59Z) - Class-Imbalanced Complementary-Label Learning via Weighted Loss [8.934943507699131]
Complementary-label learning (CLL) is widely used in weakly supervised classification.
It faces a significant challenge in real-world datasets when confronted with class-imbalanced training samples.
We propose a novel problem setting that enables learning from class-imbalanced complementary labels for multi-class classification.
arXiv Detail & Related papers (2022-09-28T16:02:42Z) - Latent Preserving Generative Adversarial Network for Imbalance
classification [17.992830267031877]
We present an end-to-end deep generative classifier.
In this paper, we propose a domain-constraint autoencoder to preserve the latent-space as prior for a generator.
Experiments are carried out on three different multi-class imbalanced problems and a comparison with state-of-the-art methods.
arXiv Detail & Related papers (2022-09-04T07:49:27Z) - Learning from Multiple Unlabeled Datasets with Partial Risk
Regularization [80.54710259664698]
In this paper, we aim to learn an accurate classifier without any class labels.
We first derive an unbiased estimator of the classification risk that can be estimated from the given unlabeled sets.
We then find that the classifier obtained as such tends to cause overfitting as its empirical risks go negative during training.
Experiments demonstrate that our method effectively mitigates overfitting and outperforms state-of-the-art methods for learning from multiple unlabeled sets.
arXiv Detail & Related papers (2022-07-04T16:22:44Z) - Multi-Class Classification from Single-Class Data with Confidences [90.48669386745361]
We propose an empirical risk minimization framework that is loss-/model-/optimizer-independent.
We show that our method can be Bayes-consistent with a simple modification even if the provided confidences are highly noisy.
arXiv Detail & Related papers (2021-06-16T15:38:13Z) - Theoretical Insights Into Multiclass Classification: A High-dimensional
Asymptotic View [82.80085730891126]
We provide the first modernally precise analysis of linear multiclass classification.
Our analysis reveals that the classification accuracy is highly distribution-dependent.
The insights gained may pave the way for a precise understanding of other classification algorithms.
arXiv Detail & Related papers (2020-11-16T05:17:29Z) - Evaluating Nonlinear Decision Trees for Binary Classification Tasks with
Other Existing Methods [8.870380386952993]
Classification of datasets into two or more distinct classes is an important machine learning task.
Many methods are able to classify binary classification tasks with a very high accuracy on test data, but cannot provide any easily interpretable explanation.
We highlight and evaluate a recently proposed nonlinear decision tree approach with a number of commonly used classification methods on a number of datasets.
arXiv Detail & Related papers (2020-08-25T00:00:23Z) - M2m: Imbalanced Classification via Major-to-minor Translation [79.09018382489506]
In most real-world scenarios, labeled training datasets are highly class-imbalanced, where deep neural networks suffer from generalizing to a balanced testing criterion.
In this paper, we explore a novel yet simple way to alleviate this issue by augmenting less-frequent classes via translating samples from more-frequent classes.
Our experimental results on a variety of class-imbalanced datasets show that the proposed method improves the generalization on minority classes significantly compared to other existing re-sampling or re-weighting methods.
arXiv Detail & Related papers (2020-04-01T13:21:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.