Faster Riemannian Newton-type Optimization by Subsampling and Cubic
Regularization
- URL: http://arxiv.org/abs/2302.11076v1
- Date: Wed, 22 Feb 2023 00:37:44 GMT
- Title: Faster Riemannian Newton-type Optimization by Subsampling and Cubic
Regularization
- Authors: Yian Deng, Tingting Mu
- Abstract summary: This work is on constrained large-scale non-constrained optimization where the constraint set implies a manifold structure.
We propose a new second-order saddleian optimization algorithm, aiming at improving convergence and reducing computational cost.
- Score: 3.867143522757309
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work is on constrained large-scale non-convex optimization where the
constraint set implies a manifold structure. Solving such problems is important
in a multitude of fundamental machine learning tasks. Recent advances on
Riemannian optimization have enabled the convenient recovery of solutions by
adapting unconstrained optimization algorithms over manifolds. However, it
remains challenging to scale up and meanwhile maintain stable convergence rates
and handle saddle points. We propose a new second-order Riemannian optimization
algorithm, aiming at improving convergence rate and reducing computational
cost. It enhances the Riemannian trust-region algorithm that explores curvature
information to escape saddle points through a mixture of subsampling and cubic
regularization techniques. We conduct rigorous analysis to study the
convergence behavior of the proposed algorithm. We also perform extensive
experiments to evaluate it based on two general machine learning tasks using
multiple datasets. The proposed algorithm exhibits improved computational speed
and convergence behavior compared to a large set of state-of-the-art Riemannian
optimization algorithms.
Related papers
- Nonconvex Federated Learning on Compact Smooth Submanifolds With Heterogeneous Data [23.661713049508375]
We propose an algorithm that learns over a submanifold in the setting of a client.
We show that our proposed algorithm converges sub-ly to a neighborhood of a first-order optimal solution by using a novel analysis.
arXiv Detail & Related papers (2024-06-12T17:53:28Z) - Riemannian Bilevel Optimization [35.42472057648458]
We focus in particular on batch and gradient-based methods, with the explicit goal of avoiding second-order information.
We propose and analyze $mathrmRF2SA$, a method that leverages first-order gradient information.
We provide explicit convergence rates for reaching $epsilon$-stationary points under various setups.
arXiv Detail & Related papers (2024-05-22T20:49:01Z) - Zeroth-order Riemannian Averaging Stochastic Approximation Algorithms [19.99781875916751]
We show that textttZo-RASA achieves optimal sample complexities for generating $epsilon$-approximation first-order stationary solutions.
We improve the algorithm's practicality by using geometrics and vector transport, instead of exponential mappings and parallel transports.
arXiv Detail & Related papers (2023-09-25T20:13:36Z) - Accelerating Cutting-Plane Algorithms via Reinforcement Learning
Surrogates [49.84541884653309]
A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms.
Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability.
We propose a method for accelerating cutting-plane algorithms via reinforcement learning.
arXiv Detail & Related papers (2023-07-17T20:11:56Z) - Linearization Algorithms for Fully Composite Optimization [61.20539085730636]
This paper studies first-order algorithms for solving fully composite optimization problems convex compact sets.
We leverage the structure of the objective by handling differentiable and non-differentiable separately, linearizing only the smooth parts.
arXiv Detail & Related papers (2023-02-24T18:41:48Z) - On a class of geodesically convex optimization problems solved via
Euclidean MM methods [50.428784381385164]
We show how a difference of Euclidean convexization functions can be written as a difference of different types of problems in statistics and machine learning.
Ultimately, we helps the broader broader the broader the broader the broader the work.
arXiv Detail & Related papers (2022-06-22T23:57:40Z) - First-Order Algorithms for Min-Max Optimization in Geodesic Metric
Spaces [93.35384756718868]
min-max algorithms have been analyzed in the Euclidean setting.
We prove that the extraiteient (RCEG) method corrected lastrate convergence at a linear rate.
arXiv Detail & Related papers (2022-06-04T18:53:44Z) - Unified Convergence Analysis for Adaptive Optimization with Moving Average Estimator [75.05106948314956]
We show that an increasing large momentum parameter for the first-order moment is sufficient for adaptive scaling.
We also give insights for increasing the momentum in a stagewise manner in accordance with stagewise decreasing step size.
arXiv Detail & Related papers (2021-04-30T08:50:24Z) - Accelerated Algorithms for Convex and Non-Convex Optimization on
Manifolds [9.632674803757475]
We propose a scheme for solving convex and non- optimization problems on distance.
Our proposed algorithm adapts to the level of complexity in the objective function.
arXiv Detail & Related papers (2020-10-18T02:48:22Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.