Bounded P-values in Parametric Programming-based Selective Inference
- URL: http://arxiv.org/abs/2307.11351v2
- Date: Thu, 28 Dec 2023 08:13:45 GMT
- Title: Bounded P-values in Parametric Programming-based Selective Inference
- Authors: Tomohiro Shiraishi, Daiki Miwa, Vo Nguyen Le Duy, Ichiro Takeuchi
- Abstract summary: We introduce a procedure to reduce the computational cost while guaranteeing the desired precision, by proposing a method to compute the lower and upper bounds of p-values.
We demonstrate the effectiveness of the proposed method in hypothesis testing problems for feature selection in linear models and attention region identification in deep neural networks.
- Score: 23.35466397627952
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Selective inference (SI) has been actively studied as a promising framework
for statistical hypothesis testing for data-driven hypotheses. The basic idea
of SI is to make inferences conditional on an event that a hypothesis is
selected. In order to perform SI, this event must be characterized in a
traceable form. When selection event is too difficult to characterize,
additional conditions are introduced for tractability. This additional
conditions often causes the loss of power, and this issue is referred to as
over-conditioning in [Fithian et al., 2014]. Parametric programming-based SI
(PP-based SI) has been proposed as one way to address the over-conditioning
issue. The main problem of PP-based SI is its high computational cost due to
the need to exhaustively explore the data space. In this study, we introduce a
procedure to reduce the computational cost while guaranteeing the desired
precision, by proposing a method to compute the lower and upper bounds of
p-values. We also proposed three types of search strategies that efficiently
improve these bounds. We demonstrate the effectiveness of the proposed method
in hypothesis testing problems for feature selection in linear models and
attention region identification in deep neural networks.
Related papers
- Bayesian Estimation and Tuning-Free Rank Detection for Probability Mass Function Tensors [17.640500920466984]
This paper presents a novel framework for estimating the joint PMF and automatically inferring its rank from observed data.
We derive a deterministic solution based on variational inference (VI) to approximate the posterior distributions of various model parameters. Additionally, we develop a scalable version of the VI-based approach by leveraging variational inference (SVI)
Experiments involving both synthetic data and real movie recommendation data illustrate the advantages of our VI and SVI-based methods in terms of estimation accuracy, automatic rank detection, and computational efficiency.
arXiv Detail & Related papers (2024-10-08T20:07:49Z) - Optimal Multi-Distribution Learning [88.3008613028333]
Multi-distribution learning seeks to learn a shared model that minimizes the worst-case risk across $k$ distinct data distributions.
We propose a novel algorithm that yields an varepsilon-optimal randomized hypothesis with a sample complexity on the order of (d+k)/varepsilon2.
arXiv Detail & Related papers (2023-12-08T16:06:29Z) - Calibrating Neural Simulation-Based Inference with Differentiable
Coverage Probability [50.44439018155837]
We propose to include a calibration term directly into the training objective of the neural model.
By introducing a relaxation of the classical formulation of calibration error we enable end-to-end backpropagation.
It is directly applicable to existing computational pipelines allowing reliable black-box posterior inference.
arXiv Detail & Related papers (2023-10-20T10:20:45Z) - Sub-linear Regret in Adaptive Model Predictive Control [56.705978425244496]
We present STT-MPC (Self-Tuning Tube-based Model Predictive Control), an online oracle that combines the certainty-equivalence principle and polytopic tubes.
We analyze the regret of the algorithm, when compared to an algorithm initially aware of the system dynamics.
arXiv Detail & Related papers (2023-10-07T15:07:10Z) - Bounding Counterfactuals under Selection Bias [60.55840896782637]
We propose a first algorithm to address both identifiable and unidentifiable queries.
We prove that, in spite of the missingness induced by the selection bias, the likelihood of the available data is unimodal.
arXiv Detail & Related papers (2022-07-26T10:33:10Z) - Self-Certifying Classification by Linearized Deep Assignment [65.0100925582087]
We propose a novel class of deep predictors for classifying metric data on graphs within PAC-Bayes risk certification paradigm.
Building on the recent PAC-Bayes literature and data-dependent priors, this approach enables learning posterior distributions on the hypothesis space.
arXiv Detail & Related papers (2022-01-26T19:59:14Z) - More Powerful Conditional Selective Inference for Generalized Lasso by
Parametric Programming [20.309302270008146]
Conditional selective inference (SI) has been studied intensively as a new statistical inference framework for data-driven hypotheses.
We propose a more powerful and general conditional SI method for a class of problems that can be converted into quadratic parametric programming.
arXiv Detail & Related papers (2021-05-11T10:12:00Z) - More Powerful and General Selective Inference for Stepwise Feature
Selection using the Homotopy Continuation Approach [17.191349670354228]
Conditional selective inference (SI) has been actively studied as a new statistical inference framework for data-driven hypotheses.
The main limitation of the existing conditional SI methods is the loss of power due to over-conditioning.
We develop a more powerful and general conditional SI method for SFS using the homotopy method which enables us to overcome this limitation.
arXiv Detail & Related papers (2020-12-25T09:01:45Z) - Quantifying Statistical Significance of Neural Network-based Image
Segmentation by Selective Inference [23.97765106673937]
We use a conditional selective inference (SI) framework to compute exact (non-asymptotic) valid p-values for the segmentation results.
Our proposed method can successfully control the false positive rate, has good performance in terms of computational efficiency, and provides good results when applied to medical image data.
arXiv Detail & Related papers (2020-10-05T07:16:40Z) - Pre-training Is (Almost) All You Need: An Application to Commonsense
Reasoning [61.32992639292889]
Fine-tuning of pre-trained transformer models has become the standard approach for solving common NLP tasks.
We introduce a new scoring method that casts a plausibility ranking task in a full-text format.
We show that our method provides a much more stable training phase across random restarts.
arXiv Detail & Related papers (2020-04-29T10:54:40Z) - Parametric Programming Approach for More Powerful and General Lasso
Selective Inference [25.02674598600182]
Selective Inference (SI) has been actively studied in the past few years for conducting inference on the features of linear models.
The main limitation of the original SI approach for Lasso is that the inference is conducted not only conditional on the selected features but also on their signs.
We propose a parametric programming-based method that can conduct SI without conditioning on signs even when we have thousands of active features.
arXiv Detail & Related papers (2020-04-21T04:46:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.