Sparse Learning and Class Probability Estimation with Weighted Support
Vector Machines
- URL: http://arxiv.org/abs/2312.10618v1
- Date: Sun, 17 Dec 2023 06:12:33 GMT
- Title: Sparse Learning and Class Probability Estimation with Weighted Support
Vector Machines
- Authors: Liyun Zeng and Hao Helen Zhang
- Abstract summary: weighted Support Vector Machines (wSVMs) have shown great values in robustly predicting the class probability and classification for various problems with high accuracy.
We propose novel wSVMs frameworks that incorporate automatic variable selection with accurate probability estimation for sparse learning problems.
The proposed wSVMs-based sparse learning methods have wide applications and can be further extended to $K$-class problems through ensemble learning.
- Score: 1.3597551064547502
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Classification and probability estimation have broad applications in modern
machine learning and data science applications, including biology, medicine,
engineering, and computer science. The recent development of a class of
weighted Support Vector Machines (wSVMs) has shown great values in robustly
predicting the class probability and classification for various problems with
high accuracy. The current framework is based on the $\ell^2$-norm regularized
binary wSVMs optimization problem, which only works with dense features and has
poor performance at sparse features with redundant noise in most real
applications. The sparse learning process requires a prescreen of the important
variables for each binary wSVMs for accurately estimating pairwise conditional
probability. In this paper, we proposed novel wSVMs frameworks that incorporate
automatic variable selection with accurate probability estimation for sparse
learning problems. We developed efficient algorithms for effective variable
selection for solving either the $\ell^1$-norm or elastic net regularized
binary wSVMs optimization problems. The binary class probability is then
estimated either by the $\ell^2$-norm regularized wSVMs framework with selected
variables or by elastic net regularized wSVMs directly. The two-step approach
of $\ell^1$-norm followed by $\ell^2$-norm wSVMs show a great advantage in both
automatic variable selection and reliable probability estimators with the most
efficient time. The elastic net regularized wSVMs offer the best performance in
terms of variable selection and probability estimation with the additional
advantage of variable grouping in the compensation of more computation time for
high dimensional problems. The proposed wSVMs-based sparse learning methods
have wide applications and can be further extended to $K$-class problems
through ensemble learning.
Related papers
- Computation-Aware Gaussian Processes: Model Selection And Linear-Time Inference [55.150117654242706]
We show that model selection for computation-aware GPs trained on 1.8 million data points can be done within a few hours on a single GPU.
As a result of this work, Gaussian processes can be trained on large-scale datasets without significantly compromising their ability to quantify uncertainty.
arXiv Detail & Related papers (2024-11-01T21:11:48Z) - A Stochastic Approach to Bi-Level Optimization for Hyperparameter Optimization and Meta Learning [74.80956524812714]
We tackle the general differentiable meta learning problem that is ubiquitous in modern deep learning.
These problems are often formalized as Bi-Level optimizations (BLO)
We introduce a novel perspective by turning a given BLO problem into a ii optimization, where the inner loss function becomes a smooth distribution, and the outer loss becomes an expected loss over the inner distribution.
arXiv Detail & Related papers (2024-10-14T12:10:06Z) - Stochastic Q-learning for Large Discrete Action Spaces [79.1700188160944]
In complex environments with discrete action spaces, effective decision-making is critical in reinforcement learning (RL)
We present value-based RL approaches which, as opposed to optimizing over the entire set of $n$ actions, only consider a variable set of actions, possibly as small as $mathcalO(log(n)$)$.
The presented value-based RL methods include, among others, Q-learning, StochDQN, StochDDQN, all of which integrate this approach for both value-function updates and action selection.
arXiv Detail & Related papers (2024-05-16T17:58:44Z) - A Safe Screening Rule with Bi-level Optimization of $\nu$ Support Vector
Machine [15.096652880354199]
We propose a safe screening rule with bi-level optimization for $nu$-SVM.
Our SRBO-$nu$-SVM is strictly deduced by integrating the Karush-Kuhn-Tucker conditions.
We also develop an efficient dual coordinate descent method (DCDM) to further improve computational speed.
arXiv Detail & Related papers (2024-03-04T06:55:57Z) - Cost-sensitive probabilistic predictions for support vector machines [1.743685428161914]
Support vector machines (SVMs) are widely used and constitute one of the best examined and used machine learning models.
We propose a novel approach to generate probabilistic outputs for the SVM.
arXiv Detail & Related papers (2023-10-09T11:00:17Z) - Numerical Optimizations for Weighted Low-rank Estimation on Language
Model [73.12941276331316]
Singular value decomposition (SVD) is one of the most popular compression methods that approximates a target matrix with smaller matrices.
Standard SVD treats the parameters within the matrix with equal importance, which is a simple but unrealistic assumption.
We show that our method can perform better than current SOTA methods in neural-based language models.
arXiv Detail & Related papers (2022-11-02T00:58:02Z) - Linear Algorithms for Nonparametric Multiclass Probability Estimation [0.0]
Support Vector Machines (wSVMs) have been developed to estimate class probabilities through ensemble learning.
We propose two new learning schemes, the baseline learning and the One-vs. All (OVA) learning, to further improve wSVMs in terms of computational efficiency and estimation accuracy.
arXiv Detail & Related papers (2022-05-25T03:15:22Z) - Handling Imbalanced Classification Problems With Support Vector Machines
via Evolutionary Bilevel Optimization [73.17488635491262]
Support vector machines (SVMs) are popular learning algorithms to deal with binary classification problems.
This article introduces EBCS-SVM: evolutionary bilevel cost-sensitive SVMs.
arXiv Detail & Related papers (2022-04-21T16:08:44Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - AML-SVM: Adaptive Multilevel Learning with Support Vector Machines [0.0]
This paper proposes an adaptive multilevel learning framework for the nonlinear SVM.
It improves the classification quality across the refinement process, and leverages multi-threaded parallel processing for better performance.
arXiv Detail & Related papers (2020-11-05T00:17:02Z) - A quantum extension of SVM-perf for training nonlinear SVMs in almost
linear time [0.2855485723554975]
We propose a quantum algorithm for training nonlinear support vector machines (SVM) for feature space learning.
Based on the classical SVM-perf algorithm of Joachims, our algorithm has a running time which scales linearly in the number of training examples.
arXiv Detail & Related papers (2020-06-18T06:25:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.