Efficient Learning of the Parameters of Non-Linear Models using
Differentiable Resampling in Particle Filters
- URL: http://arxiv.org/abs/2111.01409v1
- Date: Tue, 2 Nov 2021 08:03:09 GMT
- Title: Efficient Learning of the Parameters of Non-Linear Models using
Differentiable Resampling in Particle Filters
- Authors: Conor Rosato, Paul Horridge, Thomas B. Sch\"on, Simon Maskell
- Abstract summary: It has been widely documented that the sampling and resampling steps in particle filters be differentiated.
We consider two state-space models and show that NUTS improves the mixing of the Markov chain and can produce more accurate results in less computational time.
- Score: 1.9499120576896227
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It has been widely documented that the sampling and resampling steps in
particle filters cannot be differentiated. The {\itshape reparameterisation
trick} was introduced to allow the sampling step to be reformulated into a
differentiable function. We extend the {\itshape reparameterisation trick} to
include the stochastic input to resampling therefore limiting the
discontinuities in the gradient calculation after this step. Knowing the
gradients of the prior and likelihood allows us to run particle Markov Chain
Monte Carlo (p-MCMC) and use the No-U-Turn Sampler (NUTS) as the proposal when
estimating parameters.
We compare the Metropolis-adjusted Langevin algorithm (MALA), Hamiltonian
Monte Carlo with different number of steps and NUTS. We consider two
state-space models and show that NUTS improves the mixing of the Markov chain
and can produce more accurate results in less computational time.
Related papers
- Differentiable Interacting Multiple Model Particle Filtering [24.26220422457388]
We propose a sequential Monte Carlo algorithm for parameter learning when the studied model exhibits random discontinuous jumps in behaviour.
We adopt the emerging framework of differentiable particle filtering, wherein parameters are trained by gradient descent.
We establish new theoretical results of the presented algorithms and demonstrate superior numerical performance compared to the previous state-of-the-art algorithms.
arXiv Detail & Related papers (2024-10-01T12:05:18Z) - Closed-form Filtering for Non-linear Systems [83.91296397912218]
We propose a new class of filters based on Gaussian PSD Models, which offer several advantages in terms of density approximation and computational efficiency.
We show that filtering can be efficiently performed in closed form when transitions and observations are Gaussian PSD Models.
Our proposed estimator enjoys strong theoretical guarantees, with estimation error that depends on the quality of the approximation and is adaptive to the regularity of the transition probabilities.
arXiv Detail & Related papers (2024-02-15T08:51:49Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Low-rank extended Kalman filtering for online learning of neural
networks from streaming data [71.97861600347959]
We propose an efficient online approximate Bayesian inference algorithm for estimating the parameters of a nonlinear function from a potentially non-stationary data stream.
The method is based on the extended Kalman filter (EKF), but uses a novel low-rank plus diagonal decomposition of the posterior matrix.
In contrast to methods based on variational inference, our method is fully deterministic, and does not require step-size tuning.
arXiv Detail & Related papers (2023-05-31T03:48:49Z) - Restoration-Degradation Beyond Linear Diffusions: A Non-Asymptotic
Analysis For DDIM-Type Samplers [90.45898746733397]
We develop a framework for non-asymptotic analysis of deterministic samplers used for diffusion generative modeling.
We show that one step along the probability flow ODE can be expressed as two steps: 1) a restoration step that runs ascent on the conditional log-likelihood at some infinitesimally previous time, and 2) a degradation step that runs the forward process using noise pointing back towards the current gradient.
arXiv Detail & Related papers (2023-03-06T18:59:19Z) - Differentiable Annealed Importance Sampling and the Perils of Gradient
Noise [68.44523807580438]
Annealed importance sampling (AIS) and related algorithms are highly effective tools for marginal likelihood estimation.
Differentiability is a desirable property as it would admit the possibility of optimizing marginal likelihood as an objective.
We propose a differentiable algorithm by abandoning Metropolis-Hastings steps, which further unlocks mini-batch computation.
arXiv Detail & Related papers (2021-07-21T17:10:14Z) - Bayesian multiscale deep generative model for the solution of
high-dimensional inverse problems [0.0]
A novel multiscale Bayesian inference approach is introduced based on deep probabilistic generative models.
The method allows high-dimensional parameter estimation while exhibiting stability, efficiency and accuracy.
arXiv Detail & Related papers (2021-02-04T11:47:21Z) - Sampling in Combinatorial Spaces with SurVAE Flow Augmented MCMC [83.48593305367523]
Hybrid Monte Carlo is a powerful Markov Chain Monte Carlo method for sampling from complex continuous distributions.
We introduce a new approach based on augmenting Monte Carlo methods with SurVAE Flows to sample from discrete distributions.
We demonstrate the efficacy of our algorithm on a range of examples from statistics, computational physics and machine learning, and observe improvements compared to alternative algorithms.
arXiv Detail & Related papers (2021-02-04T02:21:08Z) - Reversible Jump PDMP Samplers for Variable Selection [1.5469452301122175]
A new class of Monte Carlo algorithms based on piecewise deterministic Markov processes (PDMPs) have recently shown great promise.
PDMP samplers can only sample from posterior densities that are differentiable almost everywhere.
We show how to develop reversible jump PDMP samplers that can jointly explore the discrete space of models and the continuous space of parameters.
arXiv Detail & Related papers (2020-10-22T14:46:33Z) - Unbiased Gradient Estimation for Variational Auto-Encoders using Coupled
Markov Chains [34.77971292478243]
The variational auto-encoder (VAE) is a deep latent variable model that has two neural networks in an autoencoder-like architecture.
We develop a training scheme for VAEs by introducing unbiased estimators of the log-likelihood gradient.
We show experimentally that VAEs fitted with unbiased estimators exhibit better predictive performance.
arXiv Detail & Related papers (2020-10-05T08:11:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.