Sketching for Convex and Nonconvex Regularized Least Squares with Sharp
Guarantees
- URL: http://arxiv.org/abs/2311.01806v1
- Date: Fri, 3 Nov 2023 09:35:01 GMT
- Title: Sketching for Convex and Nonconvex Regularized Least Squares with Sharp
Guarantees
- Authors: Yingzhen Yang, Ping Li
- Abstract summary: In this paper, we propose a fast approximation for least square problems regularized by or convex non regularization functions.
Results are among the first to demonstrate minimax rates for estimation by solving sketched convex or non approximation problems.
- Score: 23.614074988517864
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Randomized algorithms are important for solving large-scale optimization
problems. In this paper, we propose a fast sketching algorithm for least square
problems regularized by convex or nonconvex regularization functions, Sketching
for Regularized Optimization (SRO). Our SRO algorithm first generates a sketch
of the original data matrix, then solves the sketched problem. Different from
existing randomized algorithms, our algorithm handles general Frechet
subdifferentiable regularization functions in an unified framework. We present
general theoretical result for the approximation error between the optimization
results of the original problem and the sketched problem for regularized least
square problems which can be convex or nonconvex. For arbitrary convex
regularizer, relative-error bound is proved for the approximation error.
Importantly, minimax rates for sparse signal estimation by solving the sketched
sparse convex or nonconvex learning problems are also obtained using our
general theoretical result under mild conditions. To the best of our knowledge,
our results are among the first to demonstrate minimax rates for convex or
nonconvex sparse learning problem by sketching under a unified theoretical
framework. We further propose an iterative sketching algorithm which reduces
the approximation error exponentially by iteratively invoking the sketching
algorithm. Experimental results demonstrate the effectiveness of the proposed
SRO and Iterative SRO algorithms.
Related papers
- Sketch-and-Solve: Optimized Overdetermined Least-Squares Using Randomized Numerical Linear Algebra [0.0]
This paper focuses on applying sketch-and-solve algorithms to efficiently solve the overdetermined least squares problem.
We introduce the Sketch-and-Apply (SAA-SAS) algorithm, which leverages randomized numerical linear algebra techniques to compute approximate solutions efficiently.
Our results highlight the potential of sketch-and-solve techniques in efficiently handling large-scale numerical linear algebra problems.
arXiv Detail & Related papers (2024-09-22T04:29:51Z) - Bregman-divergence-based Arimoto-Blahut algorithm [53.64687146666141]
We generalize the Arimoto-Blahut algorithm to a general function defined over Bregman-divergence system.
We propose a convex-optimization-free algorithm that can be applied to classical and quantum rate-distortion theory.
arXiv Detail & Related papers (2024-08-10T06:16:24Z) - Learning the Positions in CountSketch [49.57951567374372]
We consider sketching algorithms which first compress data by multiplication with a random sketch matrix, and then apply the sketch to quickly solve an optimization problem.
In this work, we propose the first learning-based algorithms that also optimize the locations of the non-zero entries.
arXiv Detail & Related papers (2023-06-11T07:28:35Z) - Sharp Analysis of Sketch-and-Project Methods via a Connection to
Randomized Singular Value Decomposition [14.453949553412821]
We develop a theoretical framework for obtaining sharp guarantees on the convergence rate of sketch-and-project methods.
We show that the convergence rate improves at least linearly with the sketch size, and even faster when the data matrix exhibits certain spectral decays.
Our experiments support the theory and demonstrate that even extremely sparse sketches exhibit the convergence properties predicted by our framework.
arXiv Detail & Related papers (2022-08-20T03:11:13Z) - A Parameter-free Algorithm for Convex-concave Min-max Problems [33.39558541452866]
cave-free optimization algorithms refer to algorithms whose convergence rate is optimal with respect to the initial point without any learning rate to tune.
As a by-product, we utilize the parameter-free algorithm to design a new algorithm, which obtains fast rates for min-max problems with a growth condition.
arXiv Detail & Related papers (2021-02-27T18:10:06Z) - Learning the Positions in CountSketch [51.15935547615698]
We consider sketching algorithms which first compress data by multiplication with a random sketch matrix, and then apply the sketch to quickly solve an optimization problem.
In this work we propose the first learning algorithm that also optimize the locations of the non-zero entries.
We show this algorithm gives better accuracy for low rank approximation than previous work, and apply it to other problems such as $k$-means clustering for the first time.
arXiv Detail & Related papers (2020-07-20T05:06:29Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z) - Effective Dimension Adaptive Sketching Methods for Faster Regularized
Least-Squares Optimization [56.05635751529922]
We propose a new randomized algorithm for solving L2-regularized least-squares problems based on sketching.
We consider two of the most popular random embeddings, namely, Gaussian embeddings and the Subsampled Randomized Hadamard Transform (SRHT)
arXiv Detail & Related papers (2020-06-10T15:00:09Z) - Optimal Randomized First-Order Methods for Least-Squares Problems [56.05635751529922]
This class of algorithms encompasses several randomized methods among the fastest solvers for least-squares problems.
We focus on two classical embeddings, namely, Gaussian projections and subsampled Hadamard transforms.
Our resulting algorithm yields the best complexity known for solving least-squares problems with no condition number dependence.
arXiv Detail & Related papers (2020-02-21T17:45:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.