Enhancing Robustness and Efficiency of Least Square Twin SVM via Granular Computing
- URL: http://arxiv.org/abs/2410.17338v1
- Date: Tue, 22 Oct 2024 18:13:01 GMT
- Title: Enhancing Robustness and Efficiency of Least Square Twin SVM via Granular Computing
- Authors: M. Tanveer, R. K. Sharma, A. Quadir, M. Sajid,
- Abstract summary: In the domain of machine learning, least square twin support vector machine (LSTSVM) stands out as one of the state-of-the-art models.
LSTSVM suffers from sensitivity to noise and inversions, overlooking the principle and instability in resampling.
We propose the robust granular ball LSTSVM (GBLSTSVM), which is trained using granular balls instead of original data points.
- Score: 0.2999888908665658
- License:
- Abstract: In the domain of machine learning, least square twin support vector machine (LSTSVM) stands out as one of the state-of-the-art models. However, LSTSVM suffers from sensitivity to noise and outliers, overlooking the SRM principle and instability in resampling. Moreover, its computational complexity and reliance on matrix inversions hinder the efficient processing of large datasets. As a remedy to the aforementioned challenges, we propose the robust granular ball LSTSVM (GBLSTSVM). GBLSTSVM is trained using granular balls instead of original data points. The core of a granular ball is found at its center, where it encapsulates all the pertinent information of the data points within the ball of specified radius. To improve scalability and efficiency, we further introduce the large-scale GBLSTSVM (LS-GBLSTSVM), which incorporates the SRM principle through regularization terms. Experiments are performed on UCI, KEEL, and NDC benchmark datasets; both the proposed GBLSTSVM and LS-GBLSTSVM models consistently outperform the baseline models.
Related papers
- Enhanced Feature Based Granular Ball Twin Support Vector Machine [0.5492530316344587]
We propose enhanced feature based granular ball twin support vector machine (EF-GBTSVM)
The proposed model employs the coarse granularity of granular balls (GBs) as input rather than individual data samples.
We undertake a thorough evaluation of the proposed EF-GBTSVM model on benchmark UCI and KEEL datasets.
arXiv Detail & Related papers (2024-10-08T08:10:43Z) - Granular Ball Twin Support Vector Machine [0.0]
Nonparametric likelihood Estimator in MixtureTwin support vector machine (TSVM) is an emerging machine learning model with versatile applicability in classification and regression endeavors.
TSVM confronts formidable obstacles to its efficiency and applicability on large-scale datasets.
We propose the granular ball twin support vector machine (GBTSVM) and a novel large-scale granular ball twin support vector machine (LS-GBTSVM)
We conduct a comprehensive evaluation of GBTSVM and LS-GBTSVM models on benchmark datasets from UCI, KEEL, and NDC datasets.
arXiv Detail & Related papers (2024-10-07T06:20:36Z) - GL-TSVM: A robust and smooth twin support vector machine with guardian loss function [0.0]
We introduce the guardian loss (G-loss) a novel loss function distinguished by its asymmetric, bounded, and smooth characteristics.
To adhere to the structural risk minimization (SRM) principle, we incorporate a regularization term into the objective function of GL-TSVM.
The experimental analysis on UCI and KEEL datasets substantiates the effectiveness of the proposed GL-TSVM.
arXiv Detail & Related papers (2024-08-29T08:14:20Z) - Separability and Scatteredness (S&S) Ratio-Based Efficient SVM
Regularization Parameter, Kernel, and Kernel Parameter Selection [10.66048003460524]
Support Vector Machine (SVM) is a robust machine learning algorithm with broad applications in classification, regression, and outlier detection.
This work shows that the SVM performance can be modeled as a function of separability and scatteredness (S&S) of the data.
arXiv Detail & Related papers (2023-05-17T13:51:43Z) - GBSVM: Granular-ball Support Vector Machine [46.60182022640765]
GBSVM is a significant attempt to construct a classifier using the coarse-to-fine granularity of a granular-ball as input, rather than a single data point.
This paper has fixed the errors of the original model of the existing GBSVM, and derived its dual model.
The experimental results on the UCI benchmark datasets demonstrate that GBSVM has good robustness and efficiency.
arXiv Detail & Related papers (2022-10-06T09:11:44Z) - Parameter-Efficient Sparsity for Large Language Models Fine-Tuning [63.321205487234074]
We propose a.
sparse-efficient Sparse Training (PST) method to reduce the number of trainable parameters during sparse-aware training.
Experiments with diverse networks (i.e., BERT, RoBERTa and GPT-2) demonstrate PST performs on par or better than previous sparsity methods.
arXiv Detail & Related papers (2022-05-23T02:43:45Z) - Meta Clustering Learning for Large-scale Unsupervised Person
Re-identification [124.54749810371986]
We propose a "small data for big task" paradigm dubbed Meta Clustering Learning (MCL)
MCL only pseudo-labels a subset of the entire unlabeled data via clustering to save computing for the first-phase training.
Our method significantly saves computational cost while achieving a comparable or even better performance compared to prior works.
arXiv Detail & Related papers (2021-11-19T04:10:18Z) - Estimating Average Treatment Effects with Support Vector Machines [77.34726150561087]
Support vector machine (SVM) is one of the most popular classification algorithms in the machine learning literature.
We adapt SVM as a kernel-based weighting procedure that minimizes the maximum mean discrepancy between the treatment and control groups.
We characterize the bias of causal effect estimation arising from this trade-off, connecting the proposed SVM procedure to the existing kernel balancing methods.
arXiv Detail & Related papers (2021-02-23T20:22:56Z) - Coded Stochastic ADMM for Decentralized Consensus Optimization with Edge
Computing [113.52575069030192]
Big data, including applications with high security requirements, are often collected and stored on multiple heterogeneous devices, such as mobile devices, drones and vehicles.
Due to the limitations of communication costs and security requirements, it is of paramount importance to extract information in a decentralized manner instead of aggregating data to a fusion center.
We consider the problem of learning model parameters in a multi-agent system with data locally processed via distributed edge nodes.
A class of mini-batch alternating direction method of multipliers (ADMM) algorithms is explored to develop the distributed learning model.
arXiv Detail & Related papers (2020-10-02T10:41:59Z) - On Coresets for Support Vector Machines [61.928187390362176]
A coreset is a small, representative subset of the original data points.
We show that our algorithm can be used to extend the applicability of any off-the-shelf SVM solver to streaming, distributed, and dynamic data settings.
arXiv Detail & Related papers (2020-02-15T23:25:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.