De-Sequentialized Monte Carlo: a parallel-in-time particle smoother
- URL: http://arxiv.org/abs/2202.02264v1
- Date: Fri, 4 Feb 2022 17:46:32 GMT
- Title: De-Sequentialized Monte Carlo: a parallel-in-time particle smoother
- Authors: Adrien Corenflos and Nicolas Chopin and Simo S\"arkk\"a
- Abstract summary: We propose dSMC (de-Sequentialized Monte Carlo), a new particle smoother that is able to process $T$ observations in $mathcalO(log T)$ time.
This compares favourably with standard particle smoothers, the complexity of which is linear in $T$.
We also design a particle Gibbs sampler based on dSMC, which is able to perform parameter inference in a state-space model at a $mathcalO(log(T))$ cost on parallel hardware.
- Score: 3.97478982737167
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Particle smoothers are SMC (Sequential Monte Carlo) algorithms designed to
approximate the joint distribution of the states given observations from a
state-space model. We propose dSMC (de-Sequentialized Monte Carlo), a new
particle smoother that is able to process $T$ observations in $\mathcal{O}(\log
T)$ time on parallel architecture. This compares favourably with standard
particle smoothers, the complexity of which is linear in $T$. We derive
$\mathcal{L}_p$ convergence results for dSMC, with an explicit upper bound,
polynomial in $T$. We then discuss how to reduce the variance of the smoothing
estimates computed by dSMC by (i) designing good proposal distributions for
sampling the particles at the initialization of the algorithm, as well as by
(ii) using lazy resampling to increase the number of particles used in dSMC.
Finally, we design a particle Gibbs sampler based on dSMC, which is able to
perform parameter inference in a state-space model at a $\mathcal{O}(\log(T))$
cost on parallel hardware.
Related papers
- Enhanced SMC$^2$: Leveraging Gradient Information from Differentiable Particle Filters Within Langevin Proposals [41.95156859549931]
Sequential Monte Carlo Squared (SMC$2$) is a Bayesian method which can infer the states and parameters of non-linear, non-Gaussian state-space models.
Standard random-walk proposal in SMC$2$ faces challenges, particularly with high-dimensional parameter spaces.
This study outlines a novel approach by harnessing first-order gradients derived from a Common Random Numbers - Particle Filter (CRN-PF) using PyTorch.
arXiv Detail & Related papers (2024-07-24T14:05:44Z) - Convergence Bounds for Sequential Monte Carlo on Multimodal Distributions using Soft Decomposition [6.872242798058046]
We prove bounds on the variance of a function $f$ under the empirical measure of the samples obtained by the Sequential Monte Carlo (SMC) algorithm.
We show that bounds can be obtained in the truly multi-modal setting, with mixing times that depend on local MCMC dynamics.
arXiv Detail & Related papers (2024-05-29T22:43:45Z) - Projection by Convolution: Optimal Sample Complexity for Reinforcement Learning in Continuous-Space MDPs [56.237917407785545]
We consider the problem of learning an $varepsilon$-optimal policy in a general class of continuous-space Markov decision processes (MDPs) having smooth Bellman operators.
Key to our solution is a novel projection technique based on ideas from harmonic analysis.
Our result bridges the gap between two popular but conflicting perspectives on continuous-space MDPs.
arXiv Detail & Related papers (2024-05-10T09:58:47Z) - Closed-form Filtering for Non-linear Systems [83.91296397912218]
We propose a new class of filters based on Gaussian PSD Models, which offer several advantages in terms of density approximation and computational efficiency.
We show that filtering can be efficiently performed in closed form when transitions and observations are Gaussian PSD Models.
Our proposed estimator enjoys strong theoretical guarantees, with estimation error that depends on the quality of the approximation and is adaptive to the regularity of the transition probabilities.
arXiv Detail & Related papers (2024-02-15T08:51:49Z) - SMC Is All You Need: Parallel Strong Scaling [0.695967916921061]
We develop a fully parallel sequential Monte Carlo (pSMC) method which provably delivers parallel strong scaling.
The pSMC converges to infinitesimal accuracy MSE$=O(varepsilon2)$ with a fixed finite time-complexity Cost$=O(1)$ and with no efficiency leakage.
arXiv Detail & Related papers (2024-02-09T04:13:38Z) - Langevin Quasi-Monte Carlo [6.146093081175471]
Langevin Monte Carlo (LMC) and its gradient versions are powerful algorithms for sampling from complex high-dimensional distributions.
We show that the estimation error of LMC can also be reduced by using quasi-random samples.
arXiv Detail & Related papers (2023-09-22T07:15:18Z) - Stochastic Approximation Approaches to Group Distributionally Robust
Optimization [96.26317627118912]
Group distributionally robust optimization (GDRO)
Online learning techniques to reduce the number of samples required in each round from $m$ to $1$, keeping the same sample.
A novel formulation of weighted GDRO, which allows us to derive distribution-dependent convergence rates.
arXiv Detail & Related papers (2023-02-18T09:24:15Z) - Efficient Learning of the Parameters of Non-Linear Models using
Differentiable Resampling in Particle Filters [1.9499120576896227]
It has been widely documented that the sampling and resampling steps in particle filters be differentiated.
We consider two state-space models and show that NUTS improves the mixing of the Markov chain and can produce more accurate results in less computational time.
arXiv Detail & Related papers (2021-11-02T08:03:09Z) - Mean-Square Analysis with An Application to Optimal Dimension Dependence
of Langevin Monte Carlo [60.785586069299356]
This work provides a general framework for the non-asymotic analysis of sampling error in 2-Wasserstein distance.
Our theoretical analysis is further validated by numerical experiments.
arXiv Detail & Related papers (2021-09-08T18:00:05Z) - Sample Complexity of Asynchronous Q-Learning: Sharper Analysis and
Variance Reduction [63.41789556777387]
Asynchronous Q-learning aims to learn the optimal action-value function (or Q-function) of a Markov decision process (MDP)
We show that the number of samples needed to yield an entrywise $varepsilon$-accurate estimate of the Q-function is at most on the order of $frac1mu_min (1-gamma)5varepsilon2+ fract_mixmu_min (1-gamma)$ up to some logarithmic factor.
arXiv Detail & Related papers (2020-06-04T17:51:00Z) - Robustly Learning any Clusterable Mixture of Gaussians [55.41573600814391]
We study the efficient learnability of high-dimensional Gaussian mixtures in the adversarial-robust setting.
We provide an algorithm that learns the components of an $epsilon$-corrupted $k$-mixture within information theoretically near-optimal error proofs of $tildeO(epsilon)$.
Our main technical contribution is a new robust identifiability proof clusters from a Gaussian mixture, which can be captured by the constant-degree Sum of Squares proof system.
arXiv Detail & Related papers (2020-05-13T16:44:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.