Spatially relaxed inference on high-dimensional linear models
- URL: http://arxiv.org/abs/2106.02590v1
- Date: Fri, 4 Jun 2021 16:37:19 GMT
- Title: Spatially relaxed inference on high-dimensional linear models
- Authors: J\'er\^ome-Alexis Chevalier, Tuan-Binh Nguyen, Bertrand Thirion,
Joseph Salmon
- Abstract summary: We study the properties of ensembled clustered inference algorithms which combine spatially constrained clustering, statistical inference, and ensembling to aggregate several clustered inference solutions.
We show that ensembled clustered inference algorithms control the $delta$-FWER under standard assumptions for $delta$ equal to the largest cluster diameter.
- Score: 48.989769153211995
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the inference problem for high-dimensional linear models, when
covariates have an underlying spatial organization reflected in their
correlation. A typical example of such a setting is high-resolution imaging, in
which neighboring pixels are usually very similar. Accurate point and
confidence intervals estimation is not possible in this context with many more
covariates than samples, furthermore with high correlation between covariates.
This calls for a reformulation of the statistical inference problem, that takes
into account the underlying spatial structure: if covariates are locally
correlated, it is acceptable to detect them up to a given spatial uncertainty.
We thus propose to rely on the $\delta$-FWER, that is the probability of making
a false discovery at a distance greater than $\delta$ from any true positive.
With this target measure in mind, we study the properties of ensembled
clustered inference algorithms which combine three techniques: spatially
constrained clustering, statistical inference, and ensembling to aggregate
several clustered inference solutions. We show that ensembled clustered
inference algorithms control the $\delta$-FWER under standard assumptions for
$\delta$ equal to the largest cluster diameter. We complement the theoretical
analysis with empirical results, demonstrating accurate $\delta$-FWER control
and decent power achieved by such inference algorithms.
Related papers
- Adaptive $k$-nearest neighbor classifier based on the local estimation of the shape operator [49.87315310656657]
We introduce a new adaptive $k$-nearest neighbours ($kK$-NN) algorithm that explores the local curvature at a sample to adaptively defining the neighborhood size.
Results on many real-world datasets indicate that the new $kK$-NN algorithm yields superior balanced accuracy compared to the established $k$-NN method.
arXiv Detail & Related papers (2024-09-08T13:08:45Z) - Persistent Classification: A New Approach to Stability of Data and Adversarial Examples [6.469716438197741]
We study the differences between persistence metrics along interpolants of natural and adversarial points.
We show that adversarial examples have significantly lower persistence than natural examples for large neural networks.
We connect this lack of persistence with decision boundary geometry by measuring angles of interpolants with respect to decision boundaries.
arXiv Detail & Related papers (2024-04-11T18:13:42Z) - Best Arm Identification with Fixed Budget: A Large Deviation Perspective [54.305323903582845]
We present sred, a truly adaptive algorithm that can reject arms in it any round based on the observed empirical gaps between the rewards of various arms.
In particular, we present sred, a truly adaptive algorithm that can reject arms in it any round based on the observed empirical gaps between the rewards of various arms.
arXiv Detail & Related papers (2023-12-19T13:17:43Z) - On counterfactual inference with unobserved confounding [36.18241676876348]
Given an observational study with $n$ independent but heterogeneous units, our goal is to learn the counterfactual distribution for each unit.
We introduce a convex objective that pools all $n$ samples to jointly learn all $n$ parameter vectors.
We derive sufficient conditions for compactly supported distributions to satisfy the logarithmic Sobolev inequality.
arXiv Detail & Related papers (2022-11-14T04:14:37Z) - Subspace clustering in high-dimensions: Phase transitions \&
Statistical-to-Computational gap [24.073221004661427]
A simple model to study subspace clustering is the high-dimensional $k$-Gaussian mixture model.
We provide an exact characterization of the statistically optimal reconstruction error in this model in the high-dimensional regime with extensive sparsity.
arXiv Detail & Related papers (2022-05-26T17:47:35Z) - Lattice-Based Methods Surpass Sum-of-Squares in Clustering [98.46302040220395]
Clustering is a fundamental primitive in unsupervised learning.
Recent work has established lower bounds against the class of low-degree methods.
We show that, perhaps surprisingly, this particular clustering model textitdoes not exhibit a statistical-to-computational gap.
arXiv Detail & Related papers (2021-12-07T18:50:17Z) - Fuzzy Clustering with Similarity Queries [56.96625809888241]
The fuzzy or soft objective is a popular generalization of the well-known $k$-means problem.
We show that by making few queries, the problem becomes easier to solve.
arXiv Detail & Related papers (2021-06-04T02:32:26Z) - Computationally efficient sparse clustering [67.95910835079825]
We provide a finite sample analysis of a new clustering algorithm based on PCA.
We show that it achieves the minimax optimal misclustering rate in the regime $|theta infty$.
arXiv Detail & Related papers (2020-05-21T17:51:30Z) - A Precise High-Dimensional Asymptotic Theory for Boosting and
Minimum-$\ell_1$-Norm Interpolated Classifiers [3.167685495996986]
This paper establishes a precise high-dimensional theory for boosting on separable data.
Under a class of statistical models, we provide an exact analysis of the universality error of boosting.
We also explicitly pin down the relation between the boosting test error and the optimal Bayes error.
arXiv Detail & Related papers (2020-02-05T00:24:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.