Weighted Least Squares Twin Support Vector Machine with Fuzzy Rough Set
Theory for Imbalanced Data Classification
- URL: http://arxiv.org/abs/2105.01198v1
- Date: Mon, 3 May 2021 22:33:39 GMT
- Title: Weighted Least Squares Twin Support Vector Machine with Fuzzy Rough Set
Theory for Imbalanced Data Classification
- Authors: Maysam Behmanesh, Peyman Adibi, Hossein Karshenas
- Abstract summary: Support vector machines (SVMs) are powerful supervised learning tools developed to solve classification problems.
We propose an approach that efficiently used fuzzy rough set theory in weighted least squares twin support vector machine called FRLSTSVM for classification of imbalanced data.
- Score: 0.483420384410068
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Support vector machines (SVMs) are powerful supervised learning tools
developed to solve classification problems. However, SVMs are likely to perform
poorly in the classification of imbalanced data. The rough set theory presents
a mathematical tool for inference in nondeterministic cases that provides
methods for removing irrelevant information from data. In this work, we propose
an approach that efficiently used fuzzy rough set theory in weighted least
squares twin support vector machine called FRLSTSVM for classification of
imbalanced data. The first innovation is introducing a new fuzzy rough set
based under-sampling strategy to make the classifier robust in terms of
imbalanced data. For constructing the two proximal hyperplanes in FRLSTSVM,
data points from the minority class remain unchanged while a subset of data
points in the majority class are selected using a new method. In this model, we
embedded the weight biases in the LSTSVM formulations to overcome the bias
phenomenon in the original twin SVM for the classification of imbalanced data.
In order to determine these weights in this formulation, we introduced a new
strategy that uses fuzzy rough set theory as the second innovation.
Experimental results on famous imbalanced datasets, compared with related
traditional SVM-based methods, demonstrate the superiority of our proposed
FRLSTSVM model in imbalanced data classification.
Related papers
- Enhancing Imbalance Learning: A Novel Slack-Factor Fuzzy SVM Approach [0.0]
Support vector machines (FSVMs) address class imbalance by assigning varying fuzzy memberships to samples.
The recently developed slack-factor-based FSVM (SFFSVM) improves traditional FSVMs by using slack factors to adjust fuzzy memberships based on misclassification likelihood.
We propose an improved slack-factor-based FSVM (ISFFSVM) that introduces a novel location parameter.
arXiv Detail & Related papers (2024-11-26T05:47:01Z) - Intuitionistic Fuzzy Universum Twin Support Vector Machine for Imbalanced Data [0.0]
One of the major difficulties in machine learning methods is categorizing datasets that are imbalanced.
We propose intuitionistic fuzzy universum twin support vector machines for imbalanced data (IFUTSVM-ID)
We use an intuitionistic fuzzy membership scheme to mitigate the impact of noise and outliers.
arXiv Detail & Related papers (2024-10-27T04:25:42Z) - Projection based fuzzy least squares twin support vector machine for
class imbalance problems [0.9668407688201361]
We propose a novel fuzzy based approach to deal with class imbalanced as well noisy datasets.
The proposed algorithms are evaluated on several benchmark and synthetic datasets.
arXiv Detail & Related papers (2023-09-27T14:28:48Z) - Boosting Differentiable Causal Discovery via Adaptive Sample Reweighting [62.23057729112182]
Differentiable score-based causal discovery methods learn a directed acyclic graph from observational data.
We propose a model-agnostic framework to boost causal discovery performance by dynamically learning the adaptive weights for the Reweighted Score function, ReScore.
arXiv Detail & Related papers (2023-03-06T14:49:59Z) - Parametric Classification for Generalized Category Discovery: A Baseline
Study [70.73212959385387]
Generalized Category Discovery (GCD) aims to discover novel categories in unlabelled datasets using knowledge learned from labelled samples.
We investigate the failure of parametric classifiers, verify the effectiveness of previous design choices when high-quality supervision is available, and identify unreliable pseudo-labels as a key problem.
We propose a simple yet effective parametric classification method that benefits from entropy regularisation, achieves state-of-the-art performance on multiple GCD benchmarks and shows strong robustness to unknown class numbers.
arXiv Detail & Related papers (2022-11-21T18:47:11Z) - Adaptive Distribution Calibration for Few-Shot Learning with
Hierarchical Optimal Transport [78.9167477093745]
We propose a novel distribution calibration method by learning the adaptive weight matrix between novel samples and base classes.
Experimental results on standard benchmarks demonstrate that our proposed plug-and-play model outperforms competing approaches.
arXiv Detail & Related papers (2022-10-09T02:32:57Z) - Learning to Re-weight Examples with Optimal Transport for Imbalanced
Classification [74.62203971625173]
Imbalanced data pose challenges for deep learning based classification models.
One of the most widely-used approaches for tackling imbalanced data is re-weighting.
We propose a novel re-weighting method based on optimal transport (OT) from a distributional point of view.
arXiv Detail & Related papers (2022-08-05T01:23:54Z) - Handling Imbalanced Classification Problems With Support Vector Machines
via Evolutionary Bilevel Optimization [73.17488635491262]
Support vector machines (SVMs) are popular learning algorithms to deal with binary classification problems.
This article introduces EBCS-SVM: evolutionary bilevel cost-sensitive SVMs.
arXiv Detail & Related papers (2022-04-21T16:08:44Z) - SetConv: A New Approach for Learning from Imbalanced Data [29.366843553056594]
We propose a set convolution operation and an episodic training strategy to extract a single representative for each class.
We prove that our proposed algorithm is permutation-invariant despite the order of inputs.
arXiv Detail & Related papers (2021-04-03T22:33:30Z) - Label-Imbalanced and Group-Sensitive Classification under
Overparameterization [32.923780772605596]
Label-imbalanced and group-sensitive classification seeks to appropriately modify standard training algorithms to optimize relevant metrics.
We show that a logit-adjusted loss modification to standard empirical risk minimization might be ineffective in general.
We show that our results extend naturally to binary classification with sensitive groups, thus treating the two common types of imbalances (label/group) in a unifying way.
arXiv Detail & Related papers (2021-03-02T08:09:43Z) - Estimating Average Treatment Effects with Support Vector Machines [77.34726150561087]
Support vector machine (SVM) is one of the most popular classification algorithms in the machine learning literature.
We adapt SVM as a kernel-based weighting procedure that minimizes the maximum mean discrepancy between the treatment and control groups.
We characterize the bias of causal effect estimation arising from this trade-off, connecting the proposed SVM procedure to the existing kernel balancing methods.
arXiv Detail & Related papers (2021-02-23T20:22:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.