Hardness and Algorithms for Robust and Sparse Optimization
- URL: http://arxiv.org/abs/2206.14354v1
- Date: Wed, 29 Jun 2022 01:40:38 GMT
- Title: Hardness and Algorithms for Robust and Sparse Optimization
- Authors: Eric Price, Sandeep Silwal, Samson Zhou
- Abstract summary: We explore algorithms and limitations for sparse optimization problems such as sparse linear regression and robust linear regression.
Specifically, the sparse linear regression problem seeks a $k$-sparse vector $xinmathbbRd$ to minimize $|Ax-b|$.
The robust linear regression problem seeks a set $S$ that ignores at most $k$ rows and a vector $x$ to minimize $|(Ax-b)_S|$.
- Score: 17.842787715567436
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We explore algorithms and limitations for sparse optimization problems such
as sparse linear regression and robust linear regression. The goal of the
sparse linear regression problem is to identify a small number of key features,
while the goal of the robust linear regression problem is to identify a small
number of erroneous measurements. Specifically, the sparse linear regression
problem seeks a $k$-sparse vector $x\in\mathbb{R}^d$ to minimize $\|Ax-b\|_2$,
given an input matrix $A\in\mathbb{R}^{n\times d}$ and a target vector
$b\in\mathbb{R}^n$, while the robust linear regression problem seeks a set $S$
that ignores at most $k$ rows and a vector $x$ to minimize $\|(Ax-b)_S\|_2$.
We first show bicriteria, NP-hardness of approximation for robust regression
building on the work of [OWZ15] which implies a similar result for sparse
regression. We further show fine-grained hardness of robust regression through
a reduction from the minimum-weight $k$-clique conjecture. On the positive
side, we give an algorithm for robust regression that achieves arbitrarily
accurate additive error and uses runtime that closely matches the lower bound
from the fine-grained hardness result, as well as an algorithm for sparse
regression with similar runtime. Both our upper and lower bounds rely on a
general reduction from robust linear regression to sparse regression that we
introduce. Our algorithms, inspired by the 3SUM problem, use approximate
nearest neighbor data structures and may be of independent interest for solving
sparse optimization problems. For instance, we demonstrate that our techniques
can also be used for the well-studied sparse PCA problem.
Related papers
- Turnstile $\ell_p$ leverage score sampling with applications [56.403488578703865]
We develop a novel algorithm for sampling rows $a_i$ of a matrix $AinmathbbRntimes d$, proportional to their $ell_p$ norm, when $A$ is presented in a turnstile data stream.
Our algorithm not only returns the set of sampled row indexes, it also returns slightly perturbed rows $tildea_i approx a_i$, and approximates their sampling probabilities up to $varepsilon$ relative error.
For logistic regression, our framework yields the first algorithm that achieves a $
arXiv Detail & Related papers (2024-06-01T07:33:41Z) - Computational-Statistical Gaps for Improper Learning in Sparse Linear Regression [4.396860522241307]
We show that an efficient learning algorithm for sparse linear regression can be used to solve sparse PCA problems with a negative spike.
We complement our reduction with low-degree and statistical query lower bounds for the sparse problems from which we reduce.
arXiv Detail & Related papers (2024-02-21T19:55:01Z) - Nearly Minimax Optimal Regret for Learning Linear Mixture Stochastic
Shortest Path [80.60592344361073]
We study the Shortest Path (SSP) problem with a linear mixture transition kernel.
An agent repeatedly interacts with a environment and seeks to reach certain goal state while minimizing the cumulative cost.
Existing works often assume a strictly positive lower bound of the iteration cost function or an upper bound of the expected length for the optimal policy.
arXiv Detail & Related papers (2024-02-14T07:52:00Z) - Best Policy Identification in Linear MDPs [70.57916977441262]
We investigate the problem of best identification in discounted linear Markov+Delta Decision in the fixed confidence setting under a generative model.
The lower bound as the solution of an intricate non- optimization program can be used as the starting point to devise such algorithms.
arXiv Detail & Related papers (2022-08-11T04:12:50Z) - Sparse Mixed Linear Regression with Guarantees: Taming an Intractable
Problem with Invex Relaxation [31.061339148448006]
In this paper, we study the problem of sparse mixed linear regression on an unlabeled dataset.
We provide a novel invex relaxation for this intractable problem which leads to a solution with provable theoretical guarantees.
arXiv Detail & Related papers (2022-06-02T17:38:11Z) - Statistical Query Lower Bounds for List-Decodable Linear Regression [55.06171096484622]
We study the problem of list-decodable linear regression, where an adversary can corrupt a majority of the examples.
Our main result is a Statistical Query (SQ) lower bound of $dmathrmpoly (1/alpha)$ for this problem.
arXiv Detail & Related papers (2021-06-17T17:45:21Z) - Coordinate Methods for Matrix Games [41.821881312775496]
We develop primal-dual coordinate methods for solving bilinear saddle-point problems of the form $min_x in mathcalX max_yinmathY ytop A x$.
Our methods push existing sublinear methods towards their limits in terms of per-iteration complexity and sample complexity.
arXiv Detail & Related papers (2020-09-17T17:55:03Z) - Conditional Uncorrelation and Efficient Non-approximate Subset Selection
in Sparse Regression [72.84177488527398]
We consider sparse regression from the view of correlation, and propose the formula of conditional uncorrelation.
By the proposed method, the computational complexity is reduced from $O(frac16k3+mk2+mkd)$ to $O(frac16k3+frac12mk2)$ for each candidate subset in sparse regression.
arXiv Detail & Related papers (2020-09-08T20:32:26Z) - Truncated Linear Regression in High Dimensions [26.41623833920794]
In truncated linear regression, $(A_i, y_i)_i$ whose dependent variable equals $y_i= A_irm T cdot x* + eta_i$ is some fixed unknown vector of interest.
The goal is to recover $x*$ under some favorable conditions on the $A_i$'s and the noise distribution.
We prove that there exists a computationally and statistically efficient method for recovering $k$-sparse $n$-dimensional vectors $x*$ from $m$ truncated samples.
arXiv Detail & Related papers (2020-07-29T00:31:34Z) - Optimal Robust Linear Regression in Nearly Linear Time [97.11565882347772]
We study the problem of high-dimensional robust linear regression where a learner is given access to $n$ samples from the generative model $Y = langle X,w* rangle + epsilon$
We propose estimators for this problem under two settings: (i) $X$ is L4-L2 hypercontractive, $mathbbE [XXtop]$ has bounded condition number and $epsilon$ has bounded variance and (ii) $X$ is sub-Gaussian with identity second moment and $epsilon$ is
arXiv Detail & Related papers (2020-07-16T06:44:44Z) - Fractional ridge regression: a fast, interpretable reparameterization of
ridge regression [0.0]
Ridge regression (RR) is a regularization technique that penalizes the L2-norm of the coefficients in linear regression.
We provide an algorithm to solve FRR, as well as open-source software implementations in Python.
arXiv Detail & Related papers (2020-05-07T03:12:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.