Hard Shape-Constrained Kernel Machines
- URL: http://arxiv.org/abs/2005.12636v2
- Date: Sat, 17 Oct 2020 09:59:26 GMT
- Title: Hard Shape-Constrained Kernel Machines
- Authors: Pierre-Cyril Aubin-Frankowski, Zoltan Szabo
- Abstract summary: We prove that hard affine shape constraints on function derivatives can be encoded in kernel machines.
We present a tightened second-order cone constrained reformulation, that can be readily implemented in convex solvers.
- Score: 4.94950858749529
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Shape constraints (such as non-negativity, monotonicity, convexity) play a
central role in a large number of applications, as they usually improve
performance for small sample size and help interpretability. However enforcing
these shape requirements in a hard fashion is an extremely challenging problem.
Classically, this task is tackled (i) in a soft way (without out-of-sample
guarantees), (ii) by specialized transformation of the variables on a
case-by-case basis, or (iii) by using highly restricted function classes, such
as polynomials or polynomial splines. In this paper, we prove that hard affine
shape constraints on function derivatives can be encoded in kernel machines
which represent one of the most flexible and powerful tools in machine learning
and statistics. Particularly, we present a tightened second-order cone
constrained reformulation, that can be readily implemented in convex solvers.
We prove performance guarantees on the solution, and demonstrate the efficiency
of the approach in joint quantile regression with applications to economics and
to the analysis of aircraft trajectories, among others.
Related papers
- Shape Arithmetic Expressions: Advancing Scientific Discovery Beyond Closed-Form Equations [56.78271181959529]
Generalized Additive Models (GAMs) can capture non-linear relationships between variables and targets, but they cannot capture intricate feature interactions.
We propose Shape Expressions Arithmetic ( SHAREs) that fuses GAM's flexible shape functions with the complex feature interactions found in mathematical expressions.
We also design a set of rules for constructing SHAREs that guarantee transparency of the found expressions beyond the standard constraints.
arXiv Detail & Related papers (2024-04-15T13:44:01Z) - Optimizing Solution-Samplers for Combinatorial Problems: The Landscape
of Policy-Gradient Methods [52.0617030129699]
We introduce a novel theoretical framework for analyzing the effectiveness of DeepMatching Networks and Reinforcement Learning methods.
Our main contribution holds for a broad class of problems including Max-and Min-Cut, Max-$k$-Bipartite-Bi, Maximum-Weight-Bipartite-Bi, and Traveling Salesman Problem.
As a byproduct of our analysis we introduce a novel regularization process over vanilla descent and provide theoretical and experimental evidence that it helps address vanishing-gradient issues and escape bad stationary points.
arXiv Detail & Related papers (2023-10-08T23:39:38Z) - Symmetric Tensor Networks for Generative Modeling and Constrained
Combinatorial Optimization [72.41480594026815]
Constrained optimization problems abound in industry, from portfolio optimization to logistics.
One of the major roadblocks in solving these problems is the presence of non-trivial hard constraints which limit the valid search space.
In this work, we encode arbitrary integer-valued equality constraints of the form Ax=b, directly into U(1) symmetric networks (TNs) and leverage their applicability as quantum-inspired generative models.
arXiv Detail & Related papers (2022-11-16T18:59:54Z) - Learning Globally Smooth Functions on Manifolds [94.22412028413102]
Learning smooth functions is generally challenging, except in simple cases such as learning linear or kernel models.
This work proposes to overcome these obstacles by combining techniques from semi-infinite constrained learning and manifold regularization.
We prove that, under mild conditions, this method estimates the Lipschitz constant of the solution, learning a globally smooth solution as a byproduct.
arXiv Detail & Related papers (2022-10-01T15:45:35Z) - Sparse Polynomial Optimization: Theory and Practice [5.27013884159732]
Book presents several efforts to tackle this challenge with important scientific implications.
It provides alternative optimization schemes that scale well in terms of computational complexity.
We present sparsity-exploiting hierarchies of relaxations, for either unconstrained or constrained problems.
arXiv Detail & Related papers (2022-08-23T18:56:05Z) - Approximate Latent Force Model Inference [1.3927943269211591]
latent force models offer an interpretable alternative to purely data driven tools for inference in dynamical systems.
We show that a neural operator approach can scale our model to thousands of instances, enabling fast, distributed computation.
arXiv Detail & Related papers (2021-09-24T09:55:00Z) - Handling Hard Affine SDP Shape Constraints in RKHSs [3.8073142980733]
We propose a unified and modular convex optimization framework to encode hard affine SDP constraints on function derivatives.
We prove the consistency of the proposed scheme and that of its adaptive variant, leveraging geometric properties of vRKHSs.
arXiv Detail & Related papers (2021-01-05T14:08:58Z) - Binary matrix factorization on special purpose hardware [5.926928436252818]
We focus on the important binary matrix factorization (BMF) problem which has many applications in data mining.
We propose two QUBO formulations for BMF and show how clustering constraints can easily be incorporated into these formulations.
We also propose a simple baseline algorithm which outperforms our more sophisticated methods in a few situations.
arXiv Detail & Related papers (2020-10-17T01:44:24Z) - Conditional gradient methods for stochastically constrained convex
minimization [54.53786593679331]
We propose two novel conditional gradient-based methods for solving structured convex optimization problems.
The most important feature of our framework is that only a subset of the constraints is processed at each iteration.
Our algorithms rely on variance reduction and smoothing used in conjunction with conditional gradient steps, and are accompanied by rigorous convergence guarantees.
arXiv Detail & Related papers (2020-07-07T21:26:35Z) - Scalable Differentiable Physics for Learning and Control [99.4302215142673]
Differentiable physics is a powerful approach to learning and control problems that involve physical objects and environments.
We develop a scalable framework for differentiable physics that can support a large number of objects and their interactions.
arXiv Detail & Related papers (2020-07-04T19:07:51Z) - Convergence of a Stochastic Gradient Method with Momentum for Non-Smooth
Non-Convex Optimization [25.680334940504405]
This paper establishes the convergence of the rate of a non-smooth subient method with momentum for constrained problems.
For problems, we show how the unconstrained case can be analyzed under weaker assumptions than the state-of-the-art.
arXiv Detail & Related papers (2020-02-13T12:10:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.