More Powerful Conditional Selective Inference for Generalized Lasso by
Parametric Programming
- URL: http://arxiv.org/abs/2105.04920v1
- Date: Tue, 11 May 2021 10:12:00 GMT
- Title: More Powerful Conditional Selective Inference for Generalized Lasso by
Parametric Programming
- Authors: Vo Nguyen Le Duy, Ichiro Takeuchi
- Abstract summary: Conditional selective inference (SI) has been studied intensively as a new statistical inference framework for data-driven hypotheses.
We propose a more powerful and general conditional SI method for a class of problems that can be converted into quadratic parametric programming.
- Score: 20.309302270008146
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Conditional selective inference (SI) has been studied intensively as a new
statistical inference framework for data-driven hypotheses. The basic concept
of conditional SI is to make the inference conditional on the selection event,
which enables an exact and valid statistical inference to be conducted even
when the hypothesis is selected based on the data. Conditional SI has mainly
been studied in the context of model selection, such as vanilla lasso or
generalized lasso. The main limitation of existing approaches is the low
statistical power owing to over-conditioning, which is required for
computational tractability. In this study, we propose a more powerful and
general conditional SI method for a class of problems that can be converted
into quadratic parametric programming, which includes generalized lasso. The
key concept is to compute the continuum path of the optimal solution in the
direction of the selected test statistic and to identify the subset of the data
space that corresponds to the model selection event by following the solution
path. The proposed parametric programming-based method not only avoids the
aforementioned major drawback of over-conditioning, but also improves the
performance and practicality of SI in various respects. We conducted several
experiments to demonstrate the effectiveness and efficiency of our proposed
method.
Related papers
- Statistical Inference for Temporal Difference Learning with Linear Function Approximation [62.69448336714418]
Temporal Difference (TD) learning, arguably the most widely used for policy evaluation, serves as a natural framework for this purpose.
In this paper, we study the consistency properties of TD learning with Polyak-Ruppert averaging and linear function approximation, and obtain three significant improvements over existing results.
arXiv Detail & Related papers (2024-10-21T15:34:44Z) - An incremental preference elicitation-based approach to learning potentially non-monotonic preferences in multi-criteria sorting [53.36437745983783]
We first construct a max-margin optimization-based model to model potentially non-monotonic preferences.
We devise information amount measurement methods and question selection strategies to pinpoint the most informative alternative in each iteration.
Two incremental preference elicitation-based algorithms are developed to learn potentially non-monotonic preferences.
arXiv Detail & Related papers (2024-09-04T14:36:20Z) - OTClean: Data Cleaning for Conditional Independence Violations using
Optimal Transport [51.6416022358349]
sys is a framework that harnesses optimal transport theory for data repair under Conditional Independence (CI) constraints.
We develop an iterative algorithm inspired by Sinkhorn's matrix scaling algorithm, which efficiently addresses high-dimensional and large-scale data.
arXiv Detail & Related papers (2024-03-04T18:23:55Z) - Globally-Optimal Greedy Experiment Selection for Active Sequential
Estimation [1.1530723302736279]
We study the problem of active sequential estimation, which involves adaptively selecting experiments for sequentially collected data.
The goal is to design experiment selection rules for more accurate model estimation.
We propose a class of greedy experiment selection methods and provide statistical analysis for the maximum likelihood.
arXiv Detail & Related papers (2024-02-13T17:09:29Z) - Bounded P-values in Parametric Programming-based Selective Inference [23.35466397627952]
We introduce a procedure to reduce the computational cost while guaranteeing the desired precision, by proposing a method to compute the lower and upper bounds of p-values.
We demonstrate the effectiveness of the proposed method in hypothesis testing problems for feature selection in linear models and attention region identification in deep neural networks.
arXiv Detail & Related papers (2023-07-21T04:55:03Z) - Bounding Counterfactuals under Selection Bias [60.55840896782637]
We propose a first algorithm to address both identifiable and unidentifiable queries.
We prove that, in spite of the missingness induced by the selection bias, the likelihood of the available data is unimodal.
arXiv Detail & Related papers (2022-07-26T10:33:10Z) - Black-box Selective Inference via Bootstrapping [5.960626580825523]
Conditional selective inference requires an exact characterization of the selection event, which is often unavailable except for a few examples like the lasso.
This work addresses this challenge by introducing a generic approach to estimate the selection event, facilitating feasible inference conditioned on the selection event.
arXiv Detail & Related papers (2022-03-28T05:18:21Z) - More Powerful and General Selective Inference for Stepwise Feature
Selection using the Homotopy Continuation Approach [17.191349670354228]
Conditional selective inference (SI) has been actively studied as a new statistical inference framework for data-driven hypotheses.
The main limitation of the existing conditional SI methods is the loss of power due to over-conditioning.
We develop a more powerful and general conditional SI method for SFS using the homotopy method which enables us to overcome this limitation.
arXiv Detail & Related papers (2020-12-25T09:01:45Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z) - Marginal likelihood computation for model selection and hypothesis
testing: an extensive review [66.37504201165159]
This article provides a comprehensive study of the state-of-the-art of the topic.
We highlight limitations, benefits, connections and differences among the different techniques.
Problems and possible solutions with the use of improper priors are also described.
arXiv Detail & Related papers (2020-05-17T18:31:58Z) - Parametric Programming Approach for More Powerful and General Lasso
Selective Inference [25.02674598600182]
Selective Inference (SI) has been actively studied in the past few years for conducting inference on the features of linear models.
The main limitation of the original SI approach for Lasso is that the inference is conducted not only conditional on the selected features but also on their signs.
We propose a parametric programming-based method that can conduct SI without conditioning on signs even when we have thousands of active features.
arXiv Detail & Related papers (2020-04-21T04:46:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.