Unconstrained optimisation on Riemannian manifolds
- URL: http://arxiv.org/abs/2008.11091v2
- Date: Mon, 31 Aug 2020 11:00:51 GMT
- Title: Unconstrained optimisation on Riemannian manifolds
- Authors: Tuyen Trung Truong
- Abstract summary: We give explicit descriptions of versions of (Local-) Backtracking Gradient Descent and New Q-Newton's method.
We also do explicit calculations for calculating the smallest eigenvalue of a symmetric square matrix.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we give explicit descriptions of versions of (Local-)
Backtracking Gradient Descent and New Q-Newton's method to the Riemannian
setting.Here are some easy to state consequences of results in this paper,
where X is a general Riemannian manifold of finite dimension and
$f:X\rightarrow \mathbb{R}$ a $C^2$ function which is Morse (that is, all its
critical points are non-degenerate).
{\bf Theorem.} For random choices of the hyperparameters in the Riemanian
Local Backtracking Gradient Descent algorithm and for random choices of the
initial point $x_0$, the sequence $\{x_n\}$ constructed by the algorithm either
(i) converges to a local minimum of $f$ or (ii) eventually leaves every compact
subsets of $X$ (in other words, diverges to infinity on $X$). If $f$ has
compact sublevels, then only the former alternative happens. The convergence
rate is the same as in the classical paper by Armijo.
{\bf Theorem.} Assume that $f$ is $C^3$. For random choices of the
hyperparametes in the Riemannian New Q-Newton's method, if the sequence
constructed by the algorithm converges, then the limit is a critical point of
$f$. We have a local Stable-Center manifold theorem, near saddle points of $f$,
for the dynamical system associated to the algorithm. If the limit point is a
non-degenerate minimum point, then the rate of convergence is quadratic. If
moreover $X$ is an open subset of a Lie group and the initial point $x_0$ is
chosen randomly, then we can globally avoid saddle points.
As an application, we propose a general method using Riemannian Backtracking
GD to find minimum of a function on a bounded ball in a Euclidean space, and do
explicit calculations for calculating the smallest eigenvalue of a symmetric
square matrix.
Related papers
- A simple and improved algorithm for noisy, convex, zeroth-order optimisation [59.51990161522328]
We construct an algorithm that returns a point $hat xin barmathcal X$ such that $f(hat x)$ is as small as possible.
We prove that this method is such that the $f(hat x) - min_xin barmathcal X f(x)$ is of smaller order than $d2/sqrtn$ up to poly-logarithmic terms.
arXiv Detail & Related papers (2024-06-26T18:19:10Z) - An Oblivious Stochastic Composite Optimization Algorithm for Eigenvalue
Optimization Problems [76.2042837251496]
We introduce two oblivious mirror descent algorithms based on a complementary composite setting.
Remarkably, both algorithms work without prior knowledge of the Lipschitz constant or smoothness of the objective function.
We show how to extend our framework to scale and demonstrate the efficiency and robustness of our methods on large scale semidefinite programs.
arXiv Detail & Related papers (2023-06-30T08:34:29Z) - Accelerated Methods for Riemannian Min-Max Optimization Ensuring Bounded
Geometric Penalties [21.141544548229774]
We study the form $min_x max_y f(x, y) where $mathcalN$ are Hadamard.
We show global interest accelerated by reducing gradient convergence constants.
arXiv Detail & Related papers (2023-05-25T15:43:07Z) - Convergence of a Normal Map-based Prox-SGD Method under the KL
Inequality [0.0]
We present a novel map-based algorithm ($mathsfnorMtext-mathsfSGD$) for $symbol$k$ convergence problems.
arXiv Detail & Related papers (2023-05-10T01:12:11Z) - Randomized Block-Coordinate Optimistic Gradient Algorithms for
Root-Finding Problems [8.0153031008486]
We develop two new algorithms to approximate a solution of nonlinear equations in large-scale settings.
We apply our methods to a class of large-scale finite-sum inclusions, which covers prominent applications in machine learning.
arXiv Detail & Related papers (2023-01-08T21:46:27Z) - An Optimal Stochastic Algorithm for Decentralized Nonconvex Finite-sum
Optimization [25.21457349137344]
We show a proof to show DEAREST requires at most $mathcal O(+sqrtmnLvarepsilon-2)$ first-order oracle (IFO) calls and $mathcal O(Lvarepsilon-2/sqrt1-lambda_W)$ communication rounds.
arXiv Detail & Related papers (2022-10-25T11:37:11Z) - A first-order primal-dual method with adaptivity to local smoothness [64.62056765216386]
We consider the problem of finding a saddle point for the convex-concave objective $min_x max_y f(x) + langle Ax, yrangle - g*(y)$, where $f$ is a convex function with locally Lipschitz gradient and $g$ is convex and possibly non-smooth.
We propose an adaptive version of the Condat-Vu algorithm, which alternates between primal gradient steps and dual steps.
arXiv Detail & Related papers (2021-10-28T14:19:30Z) - Finding Global Minima via Kernel Approximations [90.42048080064849]
We consider the global minimization of smooth functions based solely on function evaluations.
In this paper, we consider an approach that jointly models the function to approximate and finds a global minimum.
arXiv Detail & Related papers (2020-12-22T12:59:30Z) - Small Covers for Near-Zero Sets of Polynomials and Learning Latent
Variable Models [56.98280399449707]
We show that there exists an $epsilon$-cover for $S$ of cardinality $M = (k/epsilon)O_d(k1/d)$.
Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models hidden variables.
arXiv Detail & Related papers (2020-12-14T18:14:08Z) - A fast and simple modification of Newton's method helping to avoid
saddle points [0.0]
This paper roughly says that if $f$ is $C3$ and a sequence $x_n$, then the limit point is a critical point and is not a saddle point, and the convergence rate is the same as that of Newton's method.
arXiv Detail & Related papers (2020-06-02T10:38:00Z) - Complexity of Finding Stationary Points of Nonsmooth Nonconvex Functions [84.49087114959872]
We provide the first non-asymptotic analysis for finding stationary points of nonsmooth, nonsmooth functions.
In particular, we study Hadamard semi-differentiable functions, perhaps the largest class of nonsmooth functions.
arXiv Detail & Related papers (2020-02-10T23:23:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.