Learning Globally Smooth Functions on Manifolds
- URL: http://arxiv.org/abs/2210.00301v1
- Date: Sat, 1 Oct 2022 15:45:35 GMT
- Title: Learning Globally Smooth Functions on Manifolds
- Authors: Juan Cervino, Luiz Chamon, Benjamin D. Haeffele, Rene Vidal, and
Alejandro Ribeiro
- Abstract summary: Learning smooth functions is generally challenging, except in simple cases such as learning linear or kernel models.
This work proposes to overcome these obstacles by combining techniques from semi-infinite constrained learning and manifold regularization.
We prove that, under mild conditions, this method estimates the Lipschitz constant of the solution, learning a globally smooth solution as a byproduct.
- Score: 94.22412028413102
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Smoothness and low dimensional structures play central roles in improving
generalization and stability in learning and statistics. The combination of
these properties has led to many advances in semi-supervised learning,
generative modeling, and control of dynamical systems. However, learning smooth
functions is generally challenging, except in simple cases such as learning
linear or kernel models. Typical methods are either too conservative, relying
on crude upper bounds such as spectral normalization, too lax, penalizing
smoothness on average, or too computationally intensive, requiring the solution
of large-scale semi-definite programs. These issues are only exacerbated when
trying to simultaneously exploit low dimensionality using, e.g., manifolds.
This work proposes to overcome these obstacles by combining techniques from
semi-infinite constrained learning and manifold regularization. To do so, it
shows that, under typical conditions, the problem of learning a Lipschitz
continuous function on a manifold is equivalent to a dynamically weighted
manifold regularization problem. This observation leads to a practical
algorithm based on a weighted Laplacian penalty whose weights are adapted using
stochastic gradient techniques. We prove that, under mild conditions, this
method estimates the Lipschitz constant of the solution, learning a globally
smooth solution as a byproduct. Numerical examples illustrate the advantages of
using this method to impose global smoothness on manifolds as opposed to
imposing smoothness on average.
Related papers
- Riemannian stochastic optimization methods avoid strict saddle points [68.80251170757647]
We show that policies under study avoid strict saddle points / submanifolds with probability 1.
This result provides an important sanity check as it shows that, almost always, the limit state of an algorithm can only be a local minimizer.
arXiv Detail & Related papers (2023-11-04T11:12:24Z) - Infeasible Deterministic, Stochastic, and Variance-Reduction Algorithms for Optimization under Orthogonality Constraints [9.301728976515255]
This article provides new practical and theoretical developments for the landing algorithm.
First, the method is extended to the Stiefel manifold.
We also consider variance reduction algorithms when the cost function is an average of many functions.
arXiv Detail & Related papers (2023-03-29T07:36:54Z) - Hessian Based Smoothing Splines for Manifold Learning [0.228438857884398]
We propose a multidimensional smoothing spline algorithm in the context of manifold learning.
We generalize the bending energy penalty of thin-plate splines to a quadratic form on the Sobolev space of a flat manifold.
The existence and uniqueness of the solution is shown by applying the theory of reproducing Hilbert spaces.
arXiv Detail & Related papers (2023-02-10T02:49:05Z) - Smooth over-parameterized solvers for non-smooth structured optimization [3.756550107432323]
Non-smoothness encodes structural constraints on the solutions, such as sparsity, group sparsity, low-rank edges and sharp edges.
We operate a non-weighted but smooth overparametrization of the underlying nonsmooth optimization problems.
Our main contribution is to apply the Variable Projection (VarPro) which defines a new formulation by explicitly minimizing over part of the variables.
arXiv Detail & Related papers (2022-05-03T09:23:07Z) - Faster One-Sample Stochastic Conditional Gradient Method for Composite
Convex Minimization [61.26619639722804]
We propose a conditional gradient method (CGM) for minimizing convex finite-sum objectives formed as a sum of smooth and non-smooth terms.
The proposed method, equipped with an average gradient (SAG) estimator, requires only one sample per iteration. Nevertheless, it guarantees fast convergence rates on par with more sophisticated variance reduction techniques.
arXiv Detail & Related papers (2022-02-26T19:10:48Z) - Subgradient methods near active manifolds: saddle point avoidance, local
convergence, and asymptotic normality [4.709588811674973]
We show that aiming and subgradient approximation fully expose the smooth substructure of the problem.
We prove these properties hold for a wide class of problems, including cone reducible/decomposable functions and generic semialgebraic problems.
The normality results appear to be new even in the most classical setting.
arXiv Detail & Related papers (2021-08-26T15:02:16Z) - Optimization on manifolds: A symplectic approach [127.54402681305629]
We propose a dissipative extension of Dirac's theory of constrained Hamiltonian systems as a general framework for solving optimization problems.
Our class of (accelerated) algorithms are not only simple and efficient but also applicable to a broad range of contexts.
arXiv Detail & Related papers (2021-07-23T13:43:34Z) - Nonsmooth Implicit Differentiation for Machine Learning and Optimization [0.0]
In view of training increasingly complex learning architectures, we establish a nonsmooth implicit function theorem with an operational calculus.
Our result applies to most practical problems (i.e., definable problems) provided that a nonsmooth form of the classical invertibility condition is fulfilled.
This approach allows for formal subdifferentiation: for instance, replacing derivatives by Clarke Jacobians in the usual differentiation formulas is fully justified.
arXiv Detail & Related papers (2021-06-08T13:59:47Z) - Stability and Generalization of Stochastic Gradient Methods for Minimax
Problems [71.60601421935844]
Many machine learning problems can be formulated as minimax problems such as Generative Adversarial Networks (GANs)
We provide a comprehensive generalization analysis of examples from training gradient methods for minimax problems.
arXiv Detail & Related papers (2021-05-08T22:38:00Z) - Solving weakly supervised regression problem using low-rank manifold
regularization [77.34726150561087]
We solve a weakly supervised regression problem.
Under "weakly" we understand that for some training points the labels are known, for some unknown, and for others uncertain due to the presence of random noise or other reasons such as lack of resources.
In the numerical section, we applied the suggested method to artificial and real datasets using Monte-Carlo modeling.
arXiv Detail & Related papers (2021-04-13T23:21:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.