Average case analysis of Lasso under ultra-sparse conditions
- URL: http://arxiv.org/abs/2302.13093v1
- Date: Sat, 25 Feb 2023 14:50:32 GMT
- Title: Average case analysis of Lasso under ultra-sparse conditions
- Authors: Koki Okajima, Xiangming Meng, Takashi Takahashi, Yoshiyuki Kabashima
- Abstract summary: We analyze the performance of the least absolute shrinkage and selection operator (Lasso) for the linear model when the number of regressors grows larger.
The obtained bound for perfect support recovery is a generalization of that given in previous literature.
- Score: 4.568911586155097
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We analyze the performance of the least absolute shrinkage and selection
operator (Lasso) for the linear model when the number of regressors $N$ grows
larger keeping the true support size $d$ finite, i.e., the ultra-sparse case.
The result is based on a novel treatment of the non-rigorous replica method in
statistical physics, which has been applied only to problem settings where $N$
,$d$ and the number of observations $M$ tend to infinity at the same rate. Our
analysis makes it possible to assess the average performance of Lasso with
Gaussian sensing matrices without assumptions on the scaling of $N$ and $M$,
the noise distribution, and the profile of the true signal. Under mild
conditions on the noise distribution, the analysis also offers a lower bound on
the sample complexity necessary for partial and perfect support recovery when
$M$ diverges as $M = O(\log N)$. The obtained bound for perfect support
recovery is a generalization of that given in previous literature, which only
considers the case of Gaussian noise and diverging $d$. Extensive numerical
experiments strongly support our analysis.
Related papers
- Minimax-optimal estimation for sparse multi-reference alignment with
collision-free signals [1.3658544194443192]
We investigate minimax optimality for signal estimation under the MRA model for so-called collision-free signals.
We demonstrate that the minimax optimal rate of estimation in for the sparse MRA problem in this setting is $sigma2/sqrtn$, where $n$ is the sample size.
arXiv Detail & Related papers (2023-12-13T02:06:52Z) - KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal [70.15267479220691]
We consider and analyze the sample complexity of model reinforcement learning with a generative variance-free model.
Our analysis shows that it is nearly minimax-optimal for finding an $varepsilon$-optimal policy when $varepsilon$ is sufficiently small.
arXiv Detail & Related papers (2022-05-27T19:39:24Z) - Optimal Online Generalized Linear Regression with Stochastic Noise and
Its Application to Heteroscedastic Bandits [88.6139446295537]
We study the problem of online generalized linear regression in the setting of a generalized linear model with possibly unbounded additive noise.
We provide a sharp analysis of the classical follow-the-regularized-leader (FTRL) algorithm to cope with the label noise.
We propose an algorithm based on FTRL to achieve the first variance-aware regret bound.
arXiv Detail & Related papers (2022-02-28T08:25:26Z) - Sharp Analysis of Random Fourier Features in Classification [9.383533125404755]
We show for the first time that random Fourier features classification can achieve $O(sqrtn)$ learning rate with only $Omega(sqrtn log n)$ features.
arXiv Detail & Related papers (2021-09-22T09:49:27Z) - Heavy-tailed Streaming Statistical Estimation [58.70341336199497]
We consider the task of heavy-tailed statistical estimation given streaming $p$ samples.
We design a clipped gradient descent and provide an improved analysis under a more nuanced condition on the noise of gradients.
arXiv Detail & Related papers (2021-08-25T21:30:27Z) - Towards Sample-Optimal Compressive Phase Retrieval with Sparse and
Generative Priors [59.33977545294148]
We show that $O(k log L)$ samples suffice to guarantee that the signal is close to any vector that minimizes an amplitude-based empirical loss function.
We adapt this result to sparse phase retrieval, and show that $O(s log n)$ samples are sufficient for a similar guarantee when the underlying signal is $s$-sparse and $n$-dimensional.
arXiv Detail & Related papers (2021-06-29T12:49:54Z) - Convergence Rates of Stochastic Gradient Descent under Infinite Noise
Variance [14.06947898164194]
Heavy tails emerge in gradient descent (SGD) in various scenarios.
We provide convergence guarantees for SGD under a state-dependent and heavy-tailed noise with a potentially infinite variance.
Our results indicate that even under heavy-tailed noise with infinite variance, SGD can converge to the global optimum.
arXiv Detail & Related papers (2021-02-20T13:45:11Z) - Stochastic Approximation for Online Tensorial Independent Component
Analysis [98.34292831923335]
Independent component analysis (ICA) has been a popular dimension reduction tool in statistical machine learning and signal processing.
In this paper, we present a by-product online tensorial algorithm that estimates for each independent component.
arXiv Detail & Related papers (2020-12-28T18:52:37Z) - Faster Convergence of Stochastic Gradient Langevin Dynamics for
Non-Log-Concave Sampling [110.88857917726276]
We provide a new convergence analysis of gradient Langevin dynamics (SGLD) for sampling from a class of distributions that can be non-log-concave.
At the core of our approach is a novel conductance analysis of SGLD using an auxiliary time-reversible Markov Chain.
arXiv Detail & Related papers (2020-10-19T15:23:18Z) - Tight Nonparametric Convergence Rates for Stochastic Gradient Descent
under the Noiseless Linear Model [0.0]
We analyze the convergence of single-pass, fixed step-size gradient descent on the least-square risk under this model.
As a special case, we analyze an online algorithm for estimating a real function on the unit interval from the noiseless observation of its value at randomly sampled points.
arXiv Detail & Related papers (2020-06-15T08:25:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.