Computing Valid p-value for Optimal Changepoint by Selective Inference
using Dynamic Programming
- URL: http://arxiv.org/abs/2002.09132v2
- Date: Mon, 22 Feb 2021 13:06:56 GMT
- Title: Computing Valid p-value for Optimal Changepoint by Selective Inference
using Dynamic Programming
- Authors: Vo Nguyen Le Duy, Hiroki Toda, Ryota Sugiyama, Ichiro Takeuchi
- Abstract summary: We introduce a novel method to perform statistical inference on the significance of changepoints (CPs)
Based on the selective inference (SI) framework, we propose an exact (non-asymptotic) approach to compute valid p-values for testing the significance of the CPs.
We conduct experiments on both synthetic and real-world datasets, through which we offer evidence that our proposed method is more powerful than existing methods.
- Score: 21.361641617994714
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: There is a vast body of literature related to methods for detecting
changepoints (CP). However, less attention has been paid to assessing the
statistical reliability of the detected CPs. In this paper, we introduce a
novel method to perform statistical inference on the significance of the CPs,
estimated by a Dynamic Programming (DP)-based optimal CP detection algorithm.
Based on the selective inference (SI) framework, we propose an exact
(non-asymptotic) approach to compute valid p-values for testing the
significance of the CPs. Although it is well-known that SI has low statistical
power because of over-conditioning, we address this disadvantage by introducing
parametric programming techniques. Then, we propose an efficient method to
conduct SI with the minimum amount of conditioning, leading to high statistical
power. We conduct experiments on both synthetic and real-world datasets,
through which we offer evidence that our proposed method is more powerful than
existing methods, has decent performance in terms of computational efficiency,
and provides good results in many practical applications.
Related papers
- Statistical Inference for Temporal Difference Learning with Linear Function Approximation [62.69448336714418]
Temporal Difference (TD) learning, arguably the most widely used for policy evaluation, serves as a natural framework for this purpose.
In this paper, we study the consistency properties of TD learning with Polyak-Ruppert averaging and linear function approximation, and obtain three significant improvements over existing results.
arXiv Detail & Related papers (2024-10-21T15:34:44Z) - Stratified Prediction-Powered Inference for Hybrid Language Model Evaluation [62.2436697657307]
Prediction-powered inference (PPI) is a method that improves statistical estimates based on limited human-labeled data.
We propose a method called Stratified Prediction-Powered Inference (StratPPI)
We show that the basic PPI estimates can be considerably improved by employing simple data stratification strategies.
arXiv Detail & Related papers (2024-06-06T17:37:39Z) - Policy Gradient with Active Importance Sampling [55.112959067035916]
Policy gradient (PG) methods significantly benefit from IS, enabling the effective reuse of previously collected samples.
However, IS is employed in RL as a passive tool for re-weighting historical samples.
We look for the best behavioral policy from which to collect samples to reduce the policy gradient variance.
arXiv Detail & Related papers (2024-05-09T09:08:09Z) - Bounded P-values in Parametric Programming-based Selective Inference [23.35466397627952]
We introduce a procedure to reduce the computational cost while guaranteeing the desired precision, by proposing a method to compute the lower and upper bounds of p-values.
We demonstrate the effectiveness of the proposed method in hypothesis testing problems for feature selection in linear models and attention region identification in deep neural networks.
arXiv Detail & Related papers (2023-07-21T04:55:03Z) - Online Bootstrap Inference For Policy Evaluation in Reinforcement
Learning [90.59143158534849]
The recent emergence of reinforcement learning has created a demand for robust statistical inference methods.
Existing methods for statistical inference in online learning are restricted to settings involving independently sampled observations.
The online bootstrap is a flexible and efficient approach for statistical inference in linear approximation algorithms, but its efficacy in settings involving Markov noise has yet to be explored.
arXiv Detail & Related papers (2021-08-08T18:26:35Z) - Fast and Robust Online Inference with Stochastic Gradient Descent via
Random Scaling [0.9806910643086042]
We develop a new method of online inference for a vector of parameters estimated by the Polyak-Rtupper averaging procedure of gradient descent algorithms.
Our approach is fully operational with online data and is rigorously underpinned by a functional central limit theorem.
arXiv Detail & Related papers (2021-06-06T15:38:37Z) - Statistical Optimality and Computational Efficiency of Nystr\"om Kernel
PCA [0.913755431537592]
We study the trade-off between computational complexity and statistical accuracy in Nystr"om approximate kernel principal component analysis (KPCA)
We show that Nystr"om approximate KPCA outperforms the statistical behavior of another popular approximation scheme, the random feature approximation, when applied to KPCA.
arXiv Detail & Related papers (2021-05-19T01:49:35Z) - More Powerful Conditional Selective Inference for Generalized Lasso by
Parametric Programming [20.309302270008146]
Conditional selective inference (SI) has been studied intensively as a new statistical inference framework for data-driven hypotheses.
We propose a more powerful and general conditional SI method for a class of problems that can be converted into quadratic parametric programming.
arXiv Detail & Related papers (2021-05-11T10:12:00Z) - Average-Reward Off-Policy Policy Evaluation with Function Approximation [66.67075551933438]
We consider off-policy policy evaluation with function approximation in average-reward MDPs.
bootstrapping is necessary and, along with off-policy learning and FA, results in the deadly triad.
We propose two novel algorithms, reproducing the celebrated success of Gradient TD algorithms in the average-reward setting.
arXiv Detail & Related papers (2021-01-08T00:43:04Z) - Logistic Q-Learning [87.00813469969167]
We propose a new reinforcement learning algorithm derived from a regularized linear-programming formulation of optimal control in MDPs.
The main feature of our algorithm is a convex loss function for policy evaluation that serves as a theoretically sound alternative to the widely used squared Bellman error.
arXiv Detail & Related papers (2020-10-21T17:14:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.