Effectiveness of Constant Stepsize in Markovian LSA and Statistical
Inference
- URL: http://arxiv.org/abs/2312.10894v1
- Date: Mon, 18 Dec 2023 02:51:57 GMT
- Title: Effectiveness of Constant Stepsize in Markovian LSA and Statistical
Inference
- Authors: Dongyan Huo, Yudong Chen, Qiaomin Xie
- Abstract summary: We study the effectiveness of using a constant stepsize in statistical inference via linear approximation (LSA) algorithms with Markovian data.
Our results show that using a constant stepsize enjoys easy hyper parameter tuning, fast convergence, and consistently better CI coverage, especially when data is limited.
- Score: 9.689344942945652
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we study the effectiveness of using a constant stepsize in
statistical inference via linear stochastic approximation (LSA) algorithms with
Markovian data. After establishing a Central Limit Theorem (CLT), we outline an
inference procedure that uses averaged LSA iterates to construct confidence
intervals (CIs). Our procedure leverages the fast mixing property of
constant-stepsize LSA for better covariance estimation and employs
Richardson-Romberg (RR) extrapolation to reduce the bias induced by constant
stepsize and Markovian data. We develop theoretical results for guiding
stepsize selection in RR extrapolation, and identify several important settings
where the bias provably vanishes even without extrapolation. We conduct
extensive numerical experiments and compare against classical inference
approaches. Our results show that using a constant stepsize enjoys easy
hyperparameter tuning, fast convergence, and consistently better CI coverage,
especially when data is limited.
Related papers
- Gaussian Approximation and Multiplier Bootstrap for Polyak-Ruppert Averaged Linear Stochastic Approximation with Applications to TD Learning [15.041074872715752]
We prove the validity of the confidence intervals for parameter estimation with LSA based on multiplier bootstrap.
We illustrate our findings in the setting of temporal difference learning with linear function approximation.
arXiv Detail & Related papers (2024-05-26T17:43:30Z) - Low-rank extended Kalman filtering for online learning of neural
networks from streaming data [71.97861600347959]
We propose an efficient online approximate Bayesian inference algorithm for estimating the parameters of a nonlinear function from a potentially non-stationary data stream.
The method is based on the extended Kalman filter (EKF), but uses a novel low-rank plus diagonal decomposition of the posterior matrix.
In contrast to methods based on variational inference, our method is fully deterministic, and does not require step-size tuning.
arXiv Detail & Related papers (2023-05-31T03:48:49Z) - Online Kernel CUSUM for Change-Point Detection [12.383181837411469]
We present a computationally efficient online kernel Cumulative Sum (CUSUM) method for change-point detection.
Our approach exhibits increased sensitivity to small changes compared to existing kernel-based change-point detection methods.
arXiv Detail & Related papers (2022-11-28T05:08:30Z) - Asymptotically Unbiased Instance-wise Regularized Partial AUC
Optimization: Theory and Algorithm [101.44676036551537]
One-way Partial AUC (OPAUC) and Two-way Partial AUC (TPAUC) measures the average performance of a binary classifier.
Most of the existing methods could only optimize PAUC approximately, leading to inevitable biases that are not controllable.
We present a simpler reformulation of the PAUC problem via distributional robust optimization AUC.
arXiv Detail & Related papers (2022-10-08T08:26:22Z) - Bias and Extrapolation in Markovian Linear Stochastic Approximation with
Constant Stepsizes [9.689344942945652]
We consider Linear Approximation (LSA) with a constant stepsize and Markovian data.
We show that the bias vector of this limit admits an infinite series expansion with respect to the stepsize.
We show that the bias can be reduced using Richardson-Romberg extrapolation with $mge 2$ stepsizes.
arXiv Detail & Related papers (2022-10-03T14:11:03Z) - Accelerated and instance-optimal policy evaluation with linear function
approximation [17.995515643150657]
Existing algorithms fail to match at least one of these lower bounds.
We develop an accelerated, variance-reduced fast temporal difference algorithm that simultaneously matches both lower bounds and attains a strong notion of instance-optimality.
arXiv Detail & Related papers (2021-12-24T17:21:04Z) - Differentiable Annealed Importance Sampling and the Perils of Gradient
Noise [68.44523807580438]
Annealed importance sampling (AIS) and related algorithms are highly effective tools for marginal likelihood estimation.
Differentiability is a desirable property as it would admit the possibility of optimizing marginal likelihood as an objective.
We propose a differentiable algorithm by abandoning Metropolis-Hastings steps, which further unlocks mini-batch computation.
arXiv Detail & Related papers (2021-07-21T17:10:14Z) - Fast and Robust Online Inference with Stochastic Gradient Descent via
Random Scaling [0.9806910643086042]
We develop a new method of online inference for a vector of parameters estimated by the Polyak-Rtupper averaging procedure of gradient descent algorithms.
Our approach is fully operational with online data and is rigorously underpinned by a functional central limit theorem.
arXiv Detail & Related papers (2021-06-06T15:38:37Z) - Benign Overfitting of Constant-Stepsize SGD for Linear Regression [122.70478935214128]
inductive biases are central in preventing overfitting empirically.
This work considers this issue in arguably the most basic setting: constant-stepsize SGD for linear regression.
We reflect on a number of notable differences between the algorithmic regularization afforded by (unregularized) SGD in comparison to ordinary least squares.
arXiv Detail & Related papers (2021-03-23T17:15:53Z) - Doubly Robust Off-Policy Actor-Critic: Convergence and Optimality [131.45028999325797]
We develop a doubly robust off-policy AC (DR-Off-PAC) for discounted MDP.
DR-Off-PAC adopts a single timescale structure, in which both actor and critics are updated simultaneously with constant stepsize.
We study the finite-time convergence rate and characterize the sample complexity for DR-Off-PAC to attain an $epsilon$-accurate optimal policy.
arXiv Detail & Related papers (2021-02-23T18:56:13Z) - Least Squares Regression with Markovian Data: Fundamental Limits and
Algorithms [69.45237691598774]
We study the problem of least squares linear regression where the data-points are dependent and are sampled from a Markov chain.
We establish sharp information theoretic minimax lower bounds for this problem in terms of $tau_mathsfmix$.
We propose an algorithm based on experience replay--a popular reinforcement learning technique--that achieves a significantly better error rate.
arXiv Detail & Related papers (2020-06-16T04:26:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.