Support Vector Machines with the Hard-Margin Loss: Optimal Training via
Combinatorial Benders' Cuts
- URL: http://arxiv.org/abs/2207.07690v1
- Date: Fri, 15 Jul 2022 18:21:51 GMT
- Title: Support Vector Machines with the Hard-Margin Loss: Optimal Training via
Combinatorial Benders' Cuts
- Authors: \'Italo Santana, Breno Serrano, Maximilian Schiffer, Thibaut Vidal
- Abstract summary: We show how to train the hard-margin SVM model to global optimality.
We introduce an iterative sampling and sub decomposition algorithm that solves the problem.
- Score: 8.281391209717105
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The classical hinge-loss support vector machines (SVMs) model is sensitive to
outlier observations due to the unboundedness of its loss function. To
circumvent this issue, recent studies have focused on non-convex loss
functions, such as the hard-margin loss, which associates a constant penalty to
any misclassified or within-margin sample. Applying this loss function yields
much-needed robustness for critical applications but it also leads to an
NP-hard model that makes training difficult, since current exact optimization
algorithms show limited scalability, whereas heuristics are not able to find
high-quality solutions consistently. Against this background, we propose new
integer programming strategies that significantly improve our ability to train
the hard-margin SVM model to global optimality. We introduce an iterative
sampling and decomposition approach, in which smaller subproblems are used to
separate combinatorial Benders' cuts. Those cuts, used within a branch-and-cut
algorithm, permit to converge much more quickly towards a global optimum.
Through extensive numerical analyses on classical benchmark data sets, our
solution algorithm solves, for the first time, 117 new data sets to optimality
and achieves a reduction of 50% in the average optimality gap for the hardest
datasets of the benchmark.
Related papers
- Unified Framework for Neural Network Compression via Decomposition and Optimal Rank Selection [3.3454373538792552]
We present a unified framework that applies decomposition and optimal rank selection, employing a composite compression loss within defined rank constraints.
Our approach includes an automatic rank search in a continuous space, efficiently identifying optimal rank configurations without the use of training data.
Using various benchmark datasets, we demonstrate the efficacy of our method through a comprehensive analysis.
arXiv Detail & Related papers (2024-09-05T14:15:54Z) - Enhancing GNNs Performance on Combinatorial Optimization by Recurrent Feature Update [0.09986418756990156]
We introduce a novel algorithm, denoted hereafter as QRF-GNN, leveraging the power of GNNs to efficiently solve Combinatorial optimization (CO) problems.
It relies on unsupervised learning by minimizing the loss function derived from QUBO relaxation.
Results of experiments show that QRF-GNN drastically surpasses existing learning-based approaches and is comparable to the state-of-the-art conventionals.
arXiv Detail & Related papers (2024-07-23T13:34:35Z) - Robust support vector machines via conic optimization [0.8594140167290099]
We consider the problem of convex computational machines robust to uncertainty.
We show that the proposed estimator is competitive with the SVMs with the loss in outlier-free and better in the presence of outliers.
arXiv Detail & Related papers (2024-02-02T05:42:50Z) - Optimal Rates for Random Order Online Optimization [60.011653053877126]
We study the citetgarber 2020online, where the loss functions may be chosen by an adversary, but are then presented online in a uniformly random order.
We show that citetgarber 2020online algorithms achieve the optimal bounds and significantly improve their stability.
arXiv Detail & Related papers (2021-06-29T09:48:46Z) - Solving weakly supervised regression problem using low-rank manifold
regularization [77.34726150561087]
We solve a weakly supervised regression problem.
Under "weakly" we understand that for some training points the labels are known, for some unknown, and for others uncertain due to the presence of random noise or other reasons such as lack of resources.
In the numerical section, we applied the suggested method to artificial and real datasets using Monte-Carlo modeling.
arXiv Detail & Related papers (2021-04-13T23:21:01Z) - Divide and Learn: A Divide and Conquer Approach for Predict+Optimize [50.03608569227359]
The predict+optimize problem combines machine learning ofproblem coefficients with a optimization prob-lem that uses the predicted coefficients.
We show how to directlyexpress the loss of the optimization problem in terms of thepredicted coefficients as a piece-wise linear function.
We propose a novel divide and algorithm to tackle optimization problems without this restriction and predict itscoefficients using the optimization loss.
arXiv Detail & Related papers (2020-12-04T00:26:56Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - Provably Convergent Working Set Algorithm for Non-Convex Regularized
Regression [0.0]
This paper proposes a working set algorithm for non-regular regularizers with convergence guarantees.
Our results demonstrate high gain compared to the full problem solver for both block-coordinates or a gradient solver.
arXiv Detail & Related papers (2020-06-24T07:40:31Z) - Least Squares Regression with Markovian Data: Fundamental Limits and
Algorithms [69.45237691598774]
We study the problem of least squares linear regression where the data-points are dependent and are sampled from a Markov chain.
We establish sharp information theoretic minimax lower bounds for this problem in terms of $tau_mathsfmix$.
We propose an algorithm based on experience replay--a popular reinforcement learning technique--that achieves a significantly better error rate.
arXiv Detail & Related papers (2020-06-16T04:26:50Z) - Quasi-Newton Solver for Robust Non-Rigid Registration [35.66014845211251]
We propose a formulation for robust non-rigid registration based on a globally smooth robust estimator for data fitting and regularization.
We apply the majorization-minimization algorithm to the problem, which reduces each iteration to solving a simple least-squares problem with L-BFGS.
arXiv Detail & Related papers (2020-04-09T01:45:05Z) - Second-Order Guarantees in Centralized, Federated and Decentralized
Nonconvex Optimization [64.26238893241322]
Simple algorithms have been shown to lead to good empirical results in many contexts.
Several works have pursued rigorous analytical justification for studying non optimization problems.
A key insight in these analyses is that perturbations play a critical role in allowing local descent algorithms.
arXiv Detail & Related papers (2020-03-31T16:54:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.