Faster Algorithms and Constant Lower Bounds for the Worst-Case Expected
Error
- URL: http://arxiv.org/abs/2112.13832v1
- Date: Mon, 27 Dec 2021 18:47:25 GMT
- Title: Faster Algorithms and Constant Lower Bounds for the Worst-Case Expected
Error
- Authors: Jonah Brown-Cohen
- Abstract summary: The goal is to design estimators that minimize the worst-case expected error.
Chen, Valiant and Valiant show that, when data values are $ell_infty$-normalized, there is a time algorithm to compute an estimator for the mean.
In this paper we design provably efficient algorithms for approximating the optimal semilinear estimator based on online convex optimization.
- Score: 0.3997680012976965
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The study of statistical estimation without distributional assumptions on
data values, but with knowledge of data collection methods was recently
introduced by Chen, Valiant and Valiant (NeurIPS 2020). In this framework, the
goal is to design estimators that minimize the worst-case expected error. Here
the expectation is over a known, randomized data collection process from some
population, and the data values corresponding to each element of the population
are assumed to be worst-case.
Chen, Valiant and Valiant show that, when data values are
$\ell_{\infty}$-normalized, there is a polynomial time algorithm to compute an
estimator for the mean with worst-case expected error that is within a factor
$\frac{\pi}{2}$ of the optimum within the natural class of semilinear
estimators. However, their algorithm is based on optimizing a somewhat complex
concave objective function over a constrained set of positive semidefinite
matrices, and thus does not come with explicit runtime guarantees beyond being
polynomial time in the input.
In this paper we design provably efficient algorithms for approximating the
optimal semilinear estimator based on online convex optimization. In the
setting where data values are $\ell_{\infty}$-normalized, our algorithm
achieves a $\frac{\pi}{2}$-approximation by iteratively solving a sequence of
standard SDPs. When data values are $\ell_2$-normalized, our algorithm
iteratively computes the top eigenvector of a sequence of matrices, and does
not lose any multiplicative approximation factor. We complement these positive
results by stating a simple combinatorial condition which, if satisfied by a
data collection process, implies that any (not necessarily semilinear)
estimator for the mean has constant worst-case expected error.
Related papers
- Obtaining Lower Query Complexities through Lightweight Zeroth-Order Proximal Gradient Algorithms [65.42376001308064]
We propose two variance reduced ZO estimators for complex gradient problems.
We improve the state-of-the-art function complexities from $mathcalOleft(minfracdn1/2epsilon2, fracdepsilon3right)$ to $tildecalOleft(fracdepsilon2right)$.
arXiv Detail & Related papers (2024-10-03T15:04:01Z) - Uncertainty quantification for iterative algorithms in linear models with application to early stopping [4.150180443030652]
This paper investigates the iterates $hbb1,dots,hbbT$ obtained from iterative algorithms in high-dimensional linear regression problems.
The analysis and proposed estimators are applicable to Gradient Descent (GD), GD and their accelerated variants such as Fast Iterative Soft-Thresholding (FISTA)
arXiv Detail & Related papers (2024-04-27T10:20:41Z) - Robust Sparse Estimation for Gaussians with Optimal Error under Huber Contamination [42.526664955704746]
We study sparse estimation tasks in Huber's contamination model with a focus on mean estimation, PCA, and linear regression.
For each of these tasks, we give the first sample and computationally efficient robust estimators with optimal error guarantees.
At the technical level, we develop a novel multidimensional filtering method in the sparse regime that may find other applications.
arXiv Detail & Related papers (2024-03-15T15:51:27Z) - Privately Estimating a Gaussian: Efficient, Robust and Optimal [6.901744415870126]
We give efficient algorithms for privately estimating a Gaussian distribution in both pure and approximate differential privacy (DP) models.
In the pure DP setting, we give an efficient algorithm that estimates an unknown $d$-dimensional Gaussian distribution up to an arbitrary tiny total variation error.
For the special case of mean estimation, our algorithm achieves the optimal sample complexity of $widetilde O(d)$, improving on a $widetilde O(d1.5)$ bound from prior work.
arXiv Detail & Related papers (2022-12-15T18:27:39Z) - Optimal Randomized Approximations for Matrix based Renyi's Entropy [16.651155375441796]
We develop random approximations for integer order $alpha$ cases and series approximations for non-integer $alpha$ cases.
Large-scale simulations and real-world applications validate the effectiveness of the developed approximations.
arXiv Detail & Related papers (2022-05-16T02:24:52Z) - Sparse PCA via $l_{2,p}$-Norm Regularization for Unsupervised Feature
Selection [138.97647716793333]
We propose a simple and efficient unsupervised feature selection method, by combining reconstruction error with $l_2,p$-norm regularization.
We present an efficient optimization algorithm to solve the proposed unsupervised model, and analyse the convergence and computational complexity of the algorithm theoretically.
arXiv Detail & Related papers (2020-12-29T04:08:38Z) - List-Decodable Mean Estimation in Nearly-PCA Time [50.79691056481693]
We study the fundamental task of list-decodable mean estimation in high dimensions.
Our algorithm runs in time $widetildeO(ndk)$ for all $k = O(sqrtd) cup Omega(d)$, where $n$ is the size of the dataset.
A variant of our algorithm has runtime $widetildeO(ndk)$ for all $k$, at the expense of an $O(sqrtlog k)$ factor in the recovery guarantee
arXiv Detail & Related papers (2020-11-19T17:21:37Z) - Online Model Selection for Reinforcement Learning with Function
Approximation [50.008542459050155]
We present a meta-algorithm that adapts to the optimal complexity with $tildeO(L5/6 T2/3)$ regret.
We also show that the meta-algorithm automatically admits significantly improved instance-dependent regret bounds.
arXiv Detail & Related papers (2020-11-19T10:00:54Z) - Debiasing Distributed Second Order Optimization with Surrogate Sketching
and Scaled Regularization [101.5159744660701]
In distributed second order optimization, a standard strategy is to average many local estimates, each of which is based on a small sketch or batch of the data.
Here, we introduce a new technique for debiasing the local estimates, which leads to both theoretical and empirical improvements in the convergence rate of distributed second order methods.
arXiv Detail & Related papers (2020-07-02T18:08:14Z) - Median Matrix Completion: from Embarrassment to Optimality [16.667260586938234]
We consider matrix completion with absolute deviation loss and obtain an estimator of the median matrix.
Despite several appealing properties of median, the non-smooth absolute deviation loss leads to computational challenge.
We propose a novel refinement step, which turns such inefficient estimators into a rate (near-optimal) matrix completion procedure.
arXiv Detail & Related papers (2020-06-18T10:01:22Z) - Private Stochastic Convex Optimization: Optimal Rates in Linear Time [74.47681868973598]
We study the problem of minimizing the population loss given i.i.d. samples from a distribution over convex loss functions.
A recent work of Bassily et al. has established the optimal bound on the excess population loss achievable given $n$ samples.
We describe two new techniques for deriving convex optimization algorithms both achieving the optimal bound on excess loss and using $O(minn, n2/d)$ gradient computations.
arXiv Detail & Related papers (2020-05-10T19:52:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.