On Model Selection Consistency of Lasso for High-Dimensional Ising
Models on Tree-like Graphs
- URL: http://arxiv.org/abs/2110.08500v1
- Date: Sat, 16 Oct 2021 07:23:02 GMT
- Title: On Model Selection Consistency of Lasso for High-Dimensional Ising
Models on Tree-like Graphs
- Authors: Xiangming Meng and Tomoyuki Obuchi and Yoshiyuki Kabashima
- Abstract summary: We consider the problem of high-dimensional Ising model selection using neighborhood-based least absolute shrinkage and selection operator (Lasso)
It is rigorously proved that consistent model selection can be achieved with sample sizes $n=Omega(d3logp)$ for any tree-like graph in the paramagnetic phase.
Given the popularity and efficiency of Lasso, our rigorous analysis provides a theoretical backing for its practical use in Ising model selection.
- Score: 13.14903445595385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider the problem of high-dimensional Ising model selection using
neighborhood-based least absolute shrinkage and selection operator (Lasso). It
is rigorously proved that under some mild coherence conditions on the
population covariance matrix of the Ising model, consistent model selection can
be achieved with sample sizes $n=\Omega{(d^3\log{p})}$ for any tree-like graph
in the paramagnetic phase, where $p$ is the number of variables and $d$ is the
maximum node degree. When the same conditions are imposed directly on the
sample covariance matrices, it is shown that a reduced sample size
$n=\Omega{(d^2\log{p})}$ suffices. The obtained sufficient conditions for
consistent model selection with Lasso are the same in the scaling of the sample
complexity as that of $\ell_1$-regularized logistic regression. Given the
popularity and efficiency of Lasso, our rigorous analysis provides a
theoretical backing for its practical use in Ising model selection.
Related papers
- Computational-Statistical Gaps in Gaussian Single-Index Models [77.1473134227844]
Single-Index Models are high-dimensional regression problems with planted structure.
We show that computationally efficient algorithms, both within the Statistical Query (SQ) and the Low-Degree Polynomial (LDP) framework, necessarily require $Omega(dkstar/2)$ samples.
arXiv Detail & Related papers (2024-03-08T18:50:19Z) - KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal [70.15267479220691]
We consider and analyze the sample complexity of model reinforcement learning with a generative variance-free model.
Our analysis shows that it is nearly minimax-optimal for finding an $varepsilon$-optimal policy when $varepsilon$ is sufficiently small.
arXiv Detail & Related papers (2022-05-27T19:39:24Z) - What Makes A Good Fisherman? Linear Regression under Self-Selection Bias [32.6588421908864]
In the classical setting of self-selection, the goal is to learn $k$ models, simultaneously from observations $(x(i), y(i))$.
In this work, we present the first computationally and statistically efficient estimation algorithms for the most standard setting of this problem where the models are linear.
arXiv Detail & Related papers (2022-05-06T14:03:05Z) - Universal and data-adaptive algorithms for model selection in linear
contextual bandits [52.47796554359261]
We consider the simplest non-trivial instance of model-selection: distinguishing a simple multi-armed bandit problem from a linear contextual bandit problem.
We introduce new algorithms that explore in a data-adaptive manner and provide guarantees of the form $mathcalO(dalpha T1- alpha)$.
Our approach extends to model selection among nested linear contextual bandits under some additional assumptions.
arXiv Detail & Related papers (2021-11-08T18:05:35Z) - On the Power of Preconditioning in Sparse Linear Regression [24.140675945592704]
We show that a preconditioned Lasso can solve a large class of sparse linear regression problems nearly optimally.
For the first time, we construct random-design instances which are provably hard for an optimally preconditioned Lasso.
arXiv Detail & Related papers (2021-06-17T02:12:01Z) - The Lasso with general Gaussian designs with applications to hypothesis
testing [21.342900543543816]
The Lasso is a method for high-dimensional regression.
We show that the Lasso estimator can be precisely characterized in the regime in which both $n$ and $p$ are large.
arXiv Detail & Related papers (2020-07-27T17:48:54Z) - Model-Based Multi-Agent RL in Zero-Sum Markov Games with Near-Optimal
Sample Complexity [67.02490430380415]
We show that model-based MARL achieves a sample complexity of $tilde O(|S||B|(gamma)-3epsilon-2)$ for finding the Nash equilibrium (NE) value up to some $epsilon$ error.
We also show that such a sample bound is minimax-optimal (up to logarithmic factors) if the algorithm is reward-agnostic, where the algorithm queries state transition samples without reward knowledge.
arXiv Detail & Related papers (2020-07-15T03:25:24Z) - Breaking the Sample Size Barrier in Model-Based Reinforcement Learning
with a Generative Model [50.38446482252857]
This paper is concerned with the sample efficiency of reinforcement learning, assuming access to a generative model (or simulator)
We first consider $gamma$-discounted infinite-horizon Markov decision processes (MDPs) with state space $mathcalS$ and action space $mathcalA$.
We prove that a plain model-based planning algorithm suffices to achieve minimax-optimal sample complexity given any target accuracy level.
arXiv Detail & Related papers (2020-05-26T17:53:18Z) - Learning Gaussian Graphical Models via Multiplicative Weights [54.252053139374205]
We adapt an algorithm of Klivans and Meka based on the method of multiplicative weight updates.
The algorithm enjoys a sample complexity bound that is qualitatively similar to others in the literature.
It has a low runtime $O(mp2)$ in the case of $m$ samples and $p$ nodes, and can trivially be implemented in an online manner.
arXiv Detail & Related papers (2020-02-20T10:50:58Z) - Eigen-Stratified Models [0.0]
Stratified models depend in an arbitrary way on a selected categorical feature that takes $K$ values, and depend linearly on the other $n$ features.
Laplacian regularization with respect to a graph on the feature values can greatly improve the performance of a stratified model.
arXiv Detail & Related papers (2020-01-27T16:26:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.