Optimal Randomized First-Order Methods for Least-Squares Problems
- URL: http://arxiv.org/abs/2002.09488v2
- Date: Wed, 26 Feb 2020 00:34:08 GMT
- Title: Optimal Randomized First-Order Methods for Least-Squares Problems
- Authors: Jonathan Lacotte, Mert Pilanci
- Abstract summary: This class of algorithms encompasses several randomized methods among the fastest solvers for least-squares problems.
We focus on two classical embeddings, namely, Gaussian projections and subsampled Hadamard transforms.
Our resulting algorithm yields the best complexity known for solving least-squares problems with no condition number dependence.
- Score: 56.05635751529922
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We provide an exact analysis of a class of randomized algorithms for solving
overdetermined least-squares problems. We consider first-order methods, where
the gradients are pre-conditioned by an approximation of the Hessian, based on
a subspace embedding of the data matrix. This class of algorithms encompasses
several randomized methods among the fastest solvers for least-squares
problems. We focus on two classical embeddings, namely, Gaussian projections
and subsampled randomized Hadamard transforms (SRHT). Our key technical
innovation is the derivation of the limiting spectral density of SRHT
embeddings. Leveraging this novel result, we derive the family of normalized
orthogonal polynomials of the SRHT density and we find the optimal
pre-conditioned first-order method along with its rate of convergence. Our
analysis of Gaussian embeddings proceeds similarly, and leverages classical
random matrix theory results. In particular, we show that for a given sketch
size, SRHT embeddings exhibits a faster rate of convergence than Gaussian
embeddings. Then, we propose a new algorithm by optimizing the computational
complexity over the choice of the sketching dimension. To our knowledge, our
resulting algorithm yields the best known complexity for solving least-squares
problems with no condition number dependence.
Related papers
- Sketching for Convex and Nonconvex Regularized Least Squares with Sharp
Guarantees [23.614074988517864]
In this paper, we propose a fast approximation for least square problems regularized by or convex non regularization functions.
Results are among the first to demonstrate minimax rates for estimation by solving sketched convex or non approximation problems.
arXiv Detail & Related papers (2023-11-03T09:35:01Z) - Accelerating Cutting-Plane Algorithms via Reinforcement Learning
Surrogates [49.84541884653309]
A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms.
Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability.
We propose a method for accelerating cutting-plane algorithms via reinforcement learning.
arXiv Detail & Related papers (2023-07-17T20:11:56Z) - Adaptive Stochastic Optimisation of Nonconvex Composite Objectives [2.1700203922407493]
We propose and analyse a family of generalised composite mirror descent algorithms.
With adaptive step sizes, the proposed algorithms converge without requiring prior knowledge of the problem.
We exploit the low-dimensional structure of the decision sets for high-dimensional problems.
arXiv Detail & Related papers (2022-11-21T18:31:43Z) - Accelerated Message Passing for Entropy-Regularized MAP Inference [89.15658822319928]
Maximum a posteriori (MAP) inference in discrete-valued random fields is a fundamental problem in machine learning.
Due to the difficulty of this problem, linear programming (LP) relaxations are commonly used to derive specialized message passing algorithms.
We present randomized methods for accelerating these algorithms by leveraging techniques that underlie classical accelerated gradient.
arXiv Detail & Related papers (2020-07-01T18:43:32Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z) - Effective Dimension Adaptive Sketching Methods for Faster Regularized
Least-Squares Optimization [56.05635751529922]
We propose a new randomized algorithm for solving L2-regularized least-squares problems based on sketching.
We consider two of the most popular random embeddings, namely, Gaussian embeddings and the Subsampled Randomized Hadamard Transform (SRHT)
arXiv Detail & Related papers (2020-06-10T15:00:09Z) - Optimal Iterative Sketching with the Subsampled Randomized Hadamard
Transform [64.90148466525754]
We study the performance of iterative sketching for least-squares problems.
We show that the convergence rate for Haar and randomized Hadamard matrices are identical, andally improve upon random projections.
These techniques may be applied to other algorithms that employ randomized dimension reduction.
arXiv Detail & Related papers (2020-02-03T16:17:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.