Stochastic Gradient MCMC for Nonlinear State Space Models
- URL: http://arxiv.org/abs/1901.10568v3
- Date: Sun, 16 Jul 2023 16:04:24 GMT
- Title: Stochastic Gradient MCMC for Nonlinear State Space Models
- Authors: Christopher Aicher, Srshti Putcha, Christopher Nemeth, Paul Fearnhead,
and Emily B. Fox
- Abstract summary: Inference for nonlinear, non-Gaussian SSMs is often tackled with particle methods that do not scale well to long time series.
MCMC methods have been developed to scale inference for finite-state hidden Markov models and linear SSMs.
We present error bounds that account for both buffering error and particle error in the case of nonlinear SSMs that are log-concave in the latent process.
- Score: 4.583433328833251
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: State space models (SSMs) provide a flexible framework for modeling complex
time series via a latent stochastic process. Inference for nonlinear,
non-Gaussian SSMs is often tackled with particle methods that do not scale well
to long time series. The challenge is two-fold: not only do computations scale
linearly with time, as in the linear case, but particle filters additionally
suffer from increasing particle degeneracy with longer series. Stochastic
gradient MCMC methods have been developed to scale Bayesian inference for
finite-state hidden Markov models and linear SSMs using buffered stochastic
gradient estimates to account for temporal dependencies. We extend these
stochastic gradient estimators to nonlinear SSMs using particle methods. We
present error bounds that account for both buffering error and particle error
in the case of nonlinear SSMs that are log-concave in the latent process. We
evaluate our proposed particle buffered stochastic gradient using stochastic
gradient MCMC for inference on both long sequential synthetic and
minute-resolution financial returns data, demonstrating the importance of this
class of methods.
Related papers
- pDANSE: Particle-based Data-driven Nonlinear State Estimation from Nonlinear Measurements [55.95348868409957]
We consider the problem of designing a data-driven nonlinear state estimation (DANSE) method that uses (noisy) nonlinear measurements.<n>A recurrent neural network (RNN) provides parameters of a Gaussian prior that characterize the state of the model-free process.<n>The second-order statistics of the state posterior are computed using the nonlinear measurements observed at the time point.
arXiv Detail & Related papers (2025-10-31T14:26:48Z) - Nonparametric learning of stochastic differential equations from sparse and noisy data [2.389598109913754]
We learn the entire drift function directly from data without strong structural assumptions.<n>We develop an Expectation-Maximization (EM) algorithm that employs a novel Sequential Monte Carlo (SMC) method.<n>The resulting EM-SMC-RKHS procedure enables accurate estimation of the drift function of dynamical systems in low-data regimes.
arXiv Detail & Related papers (2025-08-15T17:01:59Z) - Inference-Time Scaling of Diffusion Language Models with Particle Gibbs Sampling [62.640128548633946]
We introduce a novel inference-time scaling approach based on particle Gibbs sampling for discrete diffusion models.<n>Our method consistently outperforms prior inference-time strategies on reward-guided text generation tasks.
arXiv Detail & Related papers (2025-07-11T08:00:47Z) - MultiPDENet: PDE-embedded Learning with Multi-time-stepping for Accelerated Flow Simulation [48.41289705783405]
We propose a PDE-embedded network with multiscale time stepping (MultiPDENet)
In particular, we design a convolutional filter based on the structure of finite difference with a small number of parameters to optimize.
A Physics Block with a 4th-order Runge-Kutta integrator at the fine time scale is established that embeds the structure of PDEs to guide the prediction.
arXiv Detail & Related papers (2025-01-27T12:15:51Z) - On the Trajectory Regularity of ODE-based Diffusion Sampling [79.17334230868693]
Diffusion-based generative models use differential equations to establish a smooth connection between a complex data distribution and a tractable prior distribution.
In this paper, we identify several intriguing trajectory properties in the ODE-based sampling process of diffusion models.
arXiv Detail & Related papers (2024-05-18T15:59:41Z) - Machine-Learned Closure of URANS for Stably Stratified Turbulence: Connecting Physical Timescales & Data Hyperparameters of Deep Time-Series Models [0.0]
We develop time-series machine learning (ML) methods for closure modeling of the Unsteady Reynolds Averaged Navier Stokes equations.
We consider decaying SST which are homogeneous and stably stratified by a uniform density gradient.
We find that the ratio of the timescales of the minimum information required by the ML models to accurately capture the dynamics of the SST corresponds to the Reynolds number of the flow.
arXiv Detail & Related papers (2024-04-24T18:58:00Z) - Differentiable and Stable Long-Range Tracking of Multiple Posterior Modes [1.534667887016089]
We leverage training data to discriminatively learn particle-based representations of uncertainty in latent object states.
Our approach achieves dramatic improvements in accuracy, while also showing much greater stability across multiple training runs.
arXiv Detail & Related papers (2024-04-12T19:33:52Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Efficient Exploration in Continuous-time Model-based Reinforcement
Learning [37.14026153342745]
Reinforcement learning algorithms typically consider discrete-time dynamics, even though the underlying systems are often continuous in time.
We introduce a model-based reinforcement learning algorithm that represents continuous-time dynamics.
arXiv Detail & Related papers (2023-10-30T15:04:40Z) - Monte Carlo Neural PDE Solver for Learning PDEs via Probabilistic Representation [59.45669299295436]
We propose a Monte Carlo PDE solver for training unsupervised neural solvers.
We use the PDEs' probabilistic representation, which regards macroscopic phenomena as ensembles of random particles.
Our experiments on convection-diffusion, Allen-Cahn, and Navier-Stokes equations demonstrate significant improvements in accuracy and efficiency.
arXiv Detail & Related papers (2023-02-10T08:05:19Z) - Score-based Continuous-time Discrete Diffusion Models [102.65769839899315]
We extend diffusion models to discrete variables by introducing a Markov jump process where the reverse process denoises via a continuous-time Markov chain.
We show that an unbiased estimator can be obtained via simple matching the conditional marginal distributions.
We demonstrate the effectiveness of the proposed method on a set of synthetic and real-world music and image benchmarks.
arXiv Detail & Related papers (2022-11-30T05:33:29Z) - Efficient Learning of the Parameters of Non-Linear Models using
Differentiable Resampling in Particle Filters [1.9499120576896227]
It has been widely documented that the sampling and resampling steps in particle filters be differentiated.
We consider two state-space models and show that NUTS improves the mixing of the Markov chain and can produce more accurate results in less computational time.
arXiv Detail & Related papers (2021-11-02T08:03:09Z) - The Connection between Discrete- and Continuous-Time Descriptions of
Gaussian Continuous Processes [60.35125735474386]
We show that discretizations yielding consistent estimators have the property of invariance under coarse-graining'
This result explains why combining differencing schemes for derivatives reconstruction and local-in-time inference approaches does not work for time series analysis of second or higher order differential equations.
arXiv Detail & Related papers (2021-01-16T17:11:02Z) - Convergence and sample complexity of gradient methods for the model-free
linear quadratic regulator problem [27.09339991866556]
We show that ODE searches for optimal control for an unknown computation system by directly searching over the corresponding space of controllers.
We take a step towards demystifying the performance and efficiency of such methods by focusing on the gradient-flow dynamics set of stabilizing feedback gains and a similar result holds for the forward disctization of the ODE.
arXiv Detail & Related papers (2019-12-26T16:56:59Z) - Targeted stochastic gradient Markov chain Monte Carlo for hidden Markov models with rare latent states [48.705095800341944]
Markov chain Monte Carlo (MCMC) algorithms for hidden Markov models often rely on the forward-backward sampler.
This makes them computationally slow as the length of the time series increases, motivating the development of sub-sampling-based approaches.
We propose a targeted sub-sampling approach that over-samples observations corresponding to rare latent states when calculating the gradient of parameters associated with them.
arXiv Detail & Related papers (2018-10-31T17:44:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.