Unbiased Estimation using Underdamped Langevin Dynamics
- URL: http://arxiv.org/abs/2206.07202v2
- Date: Tue, 15 Aug 2023 19:12:23 GMT
- Title: Unbiased Estimation using Underdamped Langevin Dynamics
- Authors: Hamza Ruzayqat, Neil K. Chada, Ajay Jasra
- Abstract summary: We focus upon developing an unbiased method via the underdamped Langevin dynamics.
We prove, under standard assumptions, that our estimator is of finite variance and either has finite expected cost, or has finite cost with a high probability.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work we consider the unbiased estimation of expectations
w.r.t.~probability measures that have non-negative Lebesgue density, and which
are known point-wise up-to a normalizing constant. We focus upon developing an
unbiased method via the underdamped Langevin dynamics, which has proven to be
popular of late due to applications in statistics and machine learning.
Specifically in continuous-time, the dynamics can be constructed {so that as
the time goes to infinity they} admit the probability of interest as a
stationary measure. {In many cases, time-discretized versions of the
underdamped Langevin dynamics are used in practice which are run only with a
fixed number of iterations.} We develop a novel scheme based upon doubly
randomized estimation as in \cite{ub_grad,disc_model}, which requires access
only to time-discretized versions of the dynamics. {The proposed scheme aims to
remove the dicretization bias and the bias resulting from running the dynamics
for a finite number of iterations}. We prove, under standard assumptions, that
our estimator is of finite variance and either has finite expected cost, or has
finite cost with a high probability. To illustrate our theoretical findings we
provide numerical experiments which verify our theory, which include
challenging examples from Bayesian statistics and statistical physics.
Related papers
- Constrained Sampling with Primal-Dual Langevin Monte Carlo [15.634831573546041]
This work considers the problem of sampling from a probability distribution known up to a normalization constant.
It satisfies a set of statistical constraints specified by the expected values of general nonlinear functions.
We put forward a discrete-time primal-dual Langevin Monte Carlo algorithm (PD-LMC) that simultaneously constrains the target distribution and samples from it.
arXiv Detail & Related papers (2024-11-01T13:26:13Z) - Distributionally Robust Instrumental Variables Estimation [10.765695227417865]
We propose a distributionally robust framework for instrumental variables (IV) estimation.
We show that Wasserstein DRIVE could be preferable in practice, particularly when the practitioner is uncertain about model assumptions or distributional shifts in data.
arXiv Detail & Related papers (2024-10-21T04:33:38Z) - Learning Unstable Continuous-Time Stochastic Linear Control Systems [0.0]
We study the problem of system identification for continuous-time dynamics, based on a single finite-length state trajectory.
We present a method for estimating the possibly unstable open-loop matrix by employing properly randomized control inputs.
We establish theoretical performance guarantees showing that the estimation error decays with trajectory length, a measure of excitability, and the signal-to-noise ratio.
arXiv Detail & Related papers (2024-09-17T16:24:51Z) - Relaxed Quantile Regression: Prediction Intervals for Asymmetric Noise [51.87307904567702]
Quantile regression is a leading approach for obtaining such intervals via the empirical estimation of quantiles in the distribution of outputs.
We propose Relaxed Quantile Regression (RQR), a direct alternative to quantile regression based interval construction that removes this arbitrary constraint.
We demonstrate that this added flexibility results in intervals with an improvement in desirable qualities.
arXiv Detail & Related papers (2024-06-05T13:36:38Z) - Extending Path-Dependent NJ-ODEs to Noisy Observations and a Dependent
Observation Framework [6.404122934568861]
We introduce a new loss function, which allows us to deal with noisy observations and explain why the previously used loss function did not lead to a consistent estimator.
arXiv Detail & Related papers (2023-07-24T22:01:22Z) - Continuous-Time Modeling of Counterfactual Outcomes Using Neural
Controlled Differential Equations [84.42837346400151]
Estimating counterfactual outcomes over time has the potential to unlock personalized healthcare.
Existing causal inference approaches consider regular, discrete-time intervals between observations and treatment decisions.
We propose a controllable simulation environment based on a model of tumor growth for a range of scenarios.
arXiv Detail & Related papers (2022-06-16T17:15:15Z) - Uncertainty estimation of pedestrian future trajectory using Bayesian
approximation [137.00426219455116]
Under dynamic traffic scenarios, planning based on deterministic predictions is not trustworthy.
The authors propose to quantify uncertainty during forecasting using approximation which deterministic approaches fail to capture.
The effect of dropout weights and long-term prediction on future state uncertainty has been studied.
arXiv Detail & Related papers (2022-05-04T04:23:38Z) - Time varying regression with hidden linear dynamics [74.9914602730208]
We revisit a model for time-varying linear regression that assumes the unknown parameters evolve according to a linear dynamical system.
Counterintuitively, we show that when the underlying dynamics are stable the parameters of this model can be estimated from data by combining just two ordinary least squares estimates.
arXiv Detail & Related papers (2021-12-29T23:37:06Z) - Learning to Estimate Without Bias [57.82628598276623]
Gauss theorem states that the weighted least squares estimator is a linear minimum variance unbiased estimation (MVUE) in linear models.
In this paper, we take a first step towards extending this result to non linear settings via deep learning with bias constraints.
A second motivation to BCE is in applications where multiple estimates of the same unknown are averaged for improved performance.
arXiv Detail & Related papers (2021-10-24T10:23:51Z) - Optimization Variance: Exploring Generalization Properties of DNNs [83.78477167211315]
The test error of a deep neural network (DNN) often demonstrates double descent.
We propose a novel metric, optimization variance (OV), to measure the diversity of model updates.
arXiv Detail & Related papers (2021-06-03T09:34:17Z) - Removing the mini-batching error in Bayesian inference using Adaptive
Langevin dynamics [0.0]
We advocate using the so-called Adaptive Langevin dynamics, which is a modification of standard inertial Langevin dynamics with a dynamical friction.
We investigate the practical relevance of the assumptions underpinning Adaptive Langevin.
arXiv Detail & Related papers (2021-05-21T13:39:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.