Lower Bounds and a Near-Optimal Shrinkage Estimator for Least Squares
using Random Projections
- URL: http://arxiv.org/abs/2006.08160v1
- Date: Mon, 15 Jun 2020 06:42:18 GMT
- Title: Lower Bounds and a Near-Optimal Shrinkage Estimator for Least Squares
using Random Projections
- Authors: Srivatsan Sridhar, Mert Pilanci, Ayfer \"Ozg\"ur
- Abstract summary: We derive an improved estimator for the least squares solution using the Gaussian sketch.
Empirically, this estimator achieves smaller error on simulated and real datasets.
- Score: 37.27708297562079
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we consider the deterministic optimization using random
projections as a statistical estimation problem, where the squared distance
between the predictions from the estimator and the true solution is the error
metric. In approximately solving a large scale least squares problem using
Gaussian sketches, we show that the sketched solution has a conditional
Gaussian distribution with the true solution as its mean. Firstly, tight worst
case error lower bounds with explicit constants are derived for any estimator
using the Gaussian sketch, and the classical sketching is shown to be the
optimal unbiased estimator. For biased estimators, the lower bound also
incorporates prior knowledge about the true solution.
Secondly, we use the James-Stein estimator to derive an improved estimator
for the least squares solution using the Gaussian sketch. An upper bound on the
expected error of this estimator is derived, which is smaller than the error of
the classical Gaussian sketch solution for any given data. The upper and lower
bounds match when the SNR of the true solution is known to be small and the
data matrix is well conditioned. Empirically, this estimator achieves smaller
error on simulated and real datasets, and works for other common sketching
methods as well.
Related papers
- Distributed Least Squares in Small Space via Sketching and Bias Reduction [0.0]
Matrix sketching is a powerful tool for reducing the size of large data matrices.
We show that these limitations can be circumvented in the distributed setting by designing sketching methods that minimize the bias of the estimator, rather than its error.
In particular, we give a sparse sketching method running in optimal space and current matrix multiplication time, which recovers a nearly-unbiased least squares estimator using two passes over the data.
arXiv Detail & Related papers (2024-05-08T18:16:37Z) - Sketching for Convex and Nonconvex Regularized Least Squares with Sharp
Guarantees [23.614074988517864]
In this paper, we propose a fast approximation for least square problems regularized by or convex non regularization functions.
Results are among the first to demonstrate minimax rates for estimation by solving sketched convex or non approximation problems.
arXiv Detail & Related papers (2023-11-03T09:35:01Z) - Distributed Sketching for Randomized Optimization: Exact
Characterization, Concentration and Lower Bounds [54.51566432934556]
We consider distributed optimization methods for problems where forming the Hessian is computationally challenging.
We leverage randomized sketches for reducing the problem dimensions as well as preserving privacy and improving straggler resilience in asynchronous distributed systems.
arXiv Detail & Related papers (2022-03-18T05:49:13Z) - Near-optimal estimation of smooth transport maps with kernel
sums-of-squares [81.02564078640275]
Under smoothness conditions, the squared Wasserstein distance between two distributions could be efficiently computed with appealing statistical error upper bounds.
The object of interest for applications such as generative modeling is the underlying optimal transport map.
We propose the first tractable algorithm for which the statistical $L2$ error on the maps nearly matches the existing minimax lower-bounds for smooth map estimation.
arXiv Detail & Related papers (2021-12-03T13:45:36Z) - Near-optimal inference in adaptive linear regression [60.08422051718195]
Even simple methods like least squares can exhibit non-normal behavior when data is collected in an adaptive manner.
We propose a family of online debiasing estimators to correct these distributional anomalies in at least squares estimation.
We demonstrate the usefulness of our theory via applications to multi-armed bandit, autoregressive time series estimation, and active learning with exploration.
arXiv Detail & Related papers (2021-07-05T21:05:11Z) - One-Bit Compressed Sensing via One-Shot Hard Thresholding [7.594050968868919]
A problem of 1-bit compressed sensing is to estimate a sparse signal from a few binary measurements.
We present a novel and concise analysis that moves away from the widely used non-constrained notion of width.
arXiv Detail & Related papers (2020-07-07T17:28:03Z) - Learning Minimax Estimators via Online Learning [55.92459567732491]
We consider the problem of designing minimax estimators for estimating parameters of a probability distribution.
We construct an algorithm for finding a mixed-case Nash equilibrium.
arXiv Detail & Related papers (2020-06-19T22:49:42Z) - Effective Dimension Adaptive Sketching Methods for Faster Regularized
Least-Squares Optimization [56.05635751529922]
We propose a new randomized algorithm for solving L2-regularized least-squares problems based on sketching.
We consider two of the most popular random embeddings, namely, Gaussian embeddings and the Subsampled Randomized Hadamard Transform (SRHT)
arXiv Detail & Related papers (2020-06-10T15:00:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.