Noisy Low-rank Matrix Optimization: Geometry of Local Minima and
Convergence Rate
- URL: http://arxiv.org/abs/2203.03899v1
- Date: Tue, 8 Mar 2022 07:44:47 GMT
- Title: Noisy Low-rank Matrix Optimization: Geometry of Local Minima and
Convergence Rate
- Authors: Ziye Ma, Somayeh Sojoudi
- Abstract summary: This paper is concerned with low-rank matrix optimization, which has found a wide range of applications in machine learning.
We develop a framework that can deal with random corruptions to general objective functions, where the noise model is arbitrary.
We characterize the geometry of the spurious local minima of the problem in a local region around ground truth in the case when the RIP constant is greater than $1/3$.
- Score: 14.191310794366075
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper is concerned with low-rank matrix optimization, which has found a
wide range of applications in machine learning. This problem in the special
case of matrix sense has been studied extensively through the notion of
Restricted Isometry Property (RIP), leading to a wealth of results on the
geometric landscape of the problem and the convergence rate of common
algorithms. However, the existing results are not able to handle the problem
with a general objective function subject to noisy data. In this paper, we
address this problem by developing a mathematical framework that can deal with
random corruptions to general objective functions, where the noise model is
arbitrary. We prove that as long as the RIP constant of the noiseless objective
is less than $1/3$, any spurious local solution of the noisy optimization
problem must be close to the ground truth solution. By working through the
strict saddle property, we also show that an approximate solution can be found
in polynomial time. We characterize the geometry of the spurious local minima
of the problem in a local region around the ground truth in the case when the
RIP constant is greater than $1/3$. This paper offers the first set of results
on the global and local optimization landscapes of general low-rank
optimization problems under arbitrary random corruptions.
Related papers
- Characterization of Locality in Spin States and Forced Moves for
Optimizations [0.36868085124383626]
In optimization problems, the existence of local minima in energy landscapes is problematic to use to seek the global minimum.
We develop an algorithm to get out of the local minima efficiently while it does not yield the exact samplings.
As the proposed algorithm is based on a rejection-free algorithm, the computational cost is low.
arXiv Detail & Related papers (2023-12-05T07:21:00Z) - Riemannian stochastic optimization methods avoid strict saddle points [68.80251170757647]
We show that policies under study avoid strict saddle points / submanifolds with probability 1.
This result provides an important sanity check as it shows that, almost always, the limit state of an algorithm can only be a local minimizer.
arXiv Detail & Related papers (2023-11-04T11:12:24Z) - Run Time Analysis for Random Local Search on Generalized Majority
Functions [1.0965065178451106]
We study how one of its properties -- neutrality -- influences the run time of random local search.
We provide a theorem, applicable to many optimization algorithms, that links the run time of MAJORITY with its symmetric version HASMAJORITY.
arXiv Detail & Related papers (2022-04-27T08:40:33Z) - Distributed stochastic proximal algorithm with random reshuffling for
non-smooth finite-sum optimization [28.862321453597918]
Non-smooth finite-sum minimization is a fundamental problem in machine learning.
This paper develops a distributed proximal-gradient algorithm with random reshuffling to solve the problem.
arXiv Detail & Related papers (2021-11-06T07:29:55Z) - Sharp Restricted Isometry Property Bounds for Low-rank Matrix Recovery
Problems with Corrupted Measurements [18.16299386229588]
We show a general low-rank matrix recovery problem with linear measurements corrupted by some noise.
We present a local guarantee for problems with an arbitrary RIP constant, which states that any local saddler is perturbed close to the ground or far from it.
arXiv Detail & Related papers (2021-05-18T02:17:59Z) - Why Do Local Methods Solve Nonconvex Problems? [54.284687261929115]
Non-used optimization is ubiquitous in modern machine learning.
We rigorously formalize it for instances of machine learning problems.
We hypothesize a unified explanation for this phenomenon.
arXiv Detail & Related papers (2021-03-24T19:34:11Z) - Community detection using fast low-cardinality semidefinite programming [94.4878715085334]
We propose a new low-cardinality algorithm that generalizes the local update to maximize a semidefinite relaxation derived from Leiden-k-cut.
This proposed algorithm is scalable, outperforms state-of-the-art algorithms, and outperforms in real-world time with little additional cost.
arXiv Detail & Related papers (2020-12-04T15:46:30Z) - A Two-Timescale Framework for Bilevel Optimization: Complexity Analysis
and Application to Actor-Critic [142.1492359556374]
Bilevel optimization is a class of problems which exhibit a two-level structure.
We propose a two-timescale approximation (TTSA) algorithm for tackling such a bilevel problem.
We show that a two-timescale natural actor-critic policy optimization algorithm can be viewed as a special case of our TTSA framework.
arXiv Detail & Related papers (2020-07-10T05:20:02Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z) - Quasi-Newton Solver for Robust Non-Rigid Registration [35.66014845211251]
We propose a formulation for robust non-rigid registration based on a globally smooth robust estimator for data fitting and regularization.
We apply the majorization-minimization algorithm to the problem, which reduces each iteration to solving a simple least-squares problem with L-BFGS.
arXiv Detail & Related papers (2020-04-09T01:45:05Z) - Second-Order Guarantees in Centralized, Federated and Decentralized
Nonconvex Optimization [64.26238893241322]
Simple algorithms have been shown to lead to good empirical results in many contexts.
Several works have pursued rigorous analytical justification for studying non optimization problems.
A key insight in these analyses is that perturbations play a critical role in allowing local descent algorithms.
arXiv Detail & Related papers (2020-03-31T16:54:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.