Automated Imbalanced Classification via Layered Learning
- URL: http://arxiv.org/abs/2205.02553v1
- Date: Thu, 5 May 2022 10:32:24 GMT
- Title: Automated Imbalanced Classification via Layered Learning
- Authors: Vitor Cerqueira, Luis Torgo, Paula Brance, Colin Bellinger
- Abstract summary: Applying resampling strategies to balance the class distribution of training instances is a common approach to tackle these problems.
Many state-of-the-art methods find instances of interest close to the decision boundary to drive the resampling process.
Over-sampling may increase the chance of overfitting by propagating the information contained in instances from the minority class.
- Score: 0.734084539365505
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper we address imbalanced binary classification (IBC) tasks.
Applying resampling strategies to balance the class distribution of training
instances is a common approach to tackle these problems. Many state-of-the-art
methods find instances of interest close to the decision boundary to drive the
resampling process. However, under-sampling the majority class may potentially
lead to important information loss. Over-sampling also may increase the chance
of overfitting by propagating the information contained in instances from the
minority class. The main contribution of our work is a new method called ICLL
for tackling IBC tasks which is not based on resampling training observations.
Instead, ICLL follows a layered learning paradigm to model the data in two
stages. In the first layer, ICLL learns to distinguish cases close to the
decision boundary from cases which are clearly from the majority class, where
this dichotomy is defined using a hierarchical clustering analysis. In the
subsequent layer, we use instances close to the decision boundary and instances
from the minority class to solve the original predictive task. A second
contribution of our work is the automatic definition of the layers which
comprise the layered learning strategy using a hierarchical clustering model.
This is a relevant discovery as this process is usually performed manually
according to domain knowledge. We carried out extensive experiments using 100
benchmark data sets. The results show that the proposed method leads to a
better performance relatively to several state-of-the-art methods for IBC.
Related papers
- MultIOD: Rehearsal-free Multihead Incremental Object Detector [17.236182938227163]
We propose MultIOD, a class-incremental object detector based on CenterNet.
We employ transfer learning between classes learned initially and those learned incrementally to tackle catastrophic forgetting.
Results show that our method outperforms state-of-the-art methods on two Pascal VOC datasets.
arXiv Detail & Related papers (2023-09-11T09:32:45Z) - SLCA: Slow Learner with Classifier Alignment for Continual Learning on a
Pre-trained Model [73.80068155830708]
We present an extensive analysis for continual learning on a pre-trained model (CLPM)
We propose a simple but extremely effective approach named Slow Learner with Alignment (SLCA)
Across a variety of scenarios, our proposal provides substantial improvements for CLPM.
arXiv Detail & Related papers (2023-03-09T08:57:01Z) - Enhancing Classification with Hierarchical Scalable Query on Fusion
Transformer [0.4129225533930965]
This paper proposes a method to boost fine-grained classification through a hierarchical approach via learnable independent query embeddings.
We exploit the idea of hierarchy to learn query embeddings that are scalable across all levels.
Our method is able to outperform the existing methods with an improvement of 11% at the fine-grained classification.
arXiv Detail & Related papers (2023-02-28T11:00:55Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - Local overlap reduction procedure for dynamic ensemble selection [13.304462985219237]
Class imbalance is a characteristic known for making learning more challenging for classification models.
We propose a DS technique which attempts to minimize the effects of the local class overlap during the classification procedure.
Experimental results show that the proposed technique can significantly outperform the baseline.
arXiv Detail & Related papers (2022-06-16T21:31:05Z) - Envelope Imbalance Learning Algorithm based on Multilayer Fuzzy C-means
Clustering and Minimum Interlayer discrepancy [14.339674126923903]
This paper proposes a deep instance envelope network-based imbalanced learning algorithm with the multilayer fuzzy c-means (MlFCM) and a minimum interlayer discrepancy mechanism based on the maximum mean discrepancy (MIDMD)
This algorithm can guarantee high quality balanced instances using a deep instance envelope network in the absence of prior knowledge.
arXiv Detail & Related papers (2021-11-02T04:59:57Z) - MCDAL: Maximum Classifier Discrepancy for Active Learning [74.73133545019877]
Recent state-of-the-art active learning methods have mostly leveraged Generative Adversarial Networks (GAN) for sample acquisition.
We propose in this paper a novel active learning framework that we call Maximum Discrepancy for Active Learning (MCDAL)
In particular, we utilize two auxiliary classification layers that learn tighter decision boundaries by maximizing the discrepancies among them.
arXiv Detail & Related papers (2021-07-23T06:57:08Z) - Partial Is Better Than All: Revisiting Fine-tuning Strategy for Few-shot
Learning [76.98364915566292]
A common practice is to train a model on the base set first and then transfer to novel classes through fine-tuning.
We propose to transfer partial knowledge by freezing or fine-tuning particular layer(s) in the base model.
We conduct extensive experiments on CUB and mini-ImageNet to demonstrate the effectiveness of our proposed method.
arXiv Detail & Related papers (2021-02-08T03:27:05Z) - Binary Classification from Multiple Unlabeled Datasets via Surrogate Set
Classification [94.55805516167369]
We propose a new approach for binary classification from m U-sets for $mge2$.
Our key idea is to consider an auxiliary classification task called surrogate set classification (SSC)
arXiv Detail & Related papers (2021-02-01T07:36:38Z) - Few-shot Action Recognition with Prototype-centered Attentive Learning [88.10852114988829]
Prototype-centered Attentive Learning (PAL) model composed of two novel components.
First, a prototype-centered contrastive learning loss is introduced to complement the conventional query-centered learning objective.
Second, PAL integrates a attentive hybrid learning mechanism that can minimize the negative impacts of outliers.
arXiv Detail & Related papers (2021-01-20T11:48:12Z) - Multi-Stage Transfer Learning with an Application to Selection Process [5.933303832684138]
In multi-stage processes, decisions happen in an ordered sequence of stages.
In this work, we proposed a textitMulti-StaGe Transfer Learning (MSGTL) approach that uses knowledge from simple classifiers trained in early stages.
We show that it is possible to control the trade-off between conserving knowledge and fine-tuning using a simple probabilistic map.
arXiv Detail & Related papers (2020-06-01T21:27:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.