Differentiable Particle Filtering via Entropy-Regularized Optimal
Transport
- URL: http://arxiv.org/abs/2102.07850v1
- Date: Mon, 15 Feb 2021 21:05:33 GMT
- Title: Differentiable Particle Filtering via Entropy-Regularized Optimal
Transport
- Authors: Adrien Corenflos, James Thornton, Arnaud Doucet, George Deligiannidis
- Abstract summary: We introduce a principled differentiable particle filter and provide convergence results.
By leveraging optimal transport ideas, we introduce a principled differentiable particle filter and provide convergence results.
- Score: 19.556744028461004
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Particle Filtering (PF) methods are an established class of procedures for
performing inference in non-linear state-space models. Resampling is a key
ingredient of PF, necessary to obtain low variance likelihood and states
estimates. However, traditional resampling methods result in PF-based loss
functions being non-differentiable with respect to model and PF parameters. In
a variational inference context, resampling also yields high variance gradient
estimates of the PF-based evidence lower bound. By leveraging optimal transport
ideas, we introduce a principled differentiable particle filter and provide
convergence results. We demonstrate this novel method on a variety of
applications.
Related papers
- Adversarial Transform Particle Filters [11.330617592263744]
The particle filter (PF) and the ensemble Kalman filter (EnKF) are widely used for approximate inference in state-space models.
We propose the Adversarial Transform Particle Filter (ATPF), a novel filtering framework that combines the strengths of the PF and the EnKF through adversarial learning.
arXiv Detail & Related papers (2025-02-10T05:31:35Z) - Diffusing Differentiable Representations [60.72992910766525]
We introduce a novel, training-free method for sampling differentiable representations (diffreps) using pretrained diffusion models.
We identify an implicit constraint on the samples induced by the diffrep and demonstrate that addressing this constraint significantly improves the consistency and detail of the generated objects.
arXiv Detail & Related papers (2024-12-09T20:42:58Z) - Understanding Reinforcement Learning-Based Fine-Tuning of Diffusion Models: A Tutorial and Review [63.31328039424469]
This tutorial provides a comprehensive survey of methods for fine-tuning diffusion models to optimize downstream reward functions.
We explain the application of various RL algorithms, including PPO, differentiable optimization, reward-weighted MLE, value-weighted sampling, and path consistency learning.
arXiv Detail & Related papers (2024-07-18T17:35:32Z) - Differentiable Particle Filtering using Optimal Placement Resampling [0.0]
A good proposal distribution and a good resampling scheme are crucial to obtain low variance estimates.
This work proposes a differentiable resampling scheme by deterministic sampling from an empirical cumulative distribution function.
arXiv Detail & Related papers (2024-02-26T15:09:56Z) - Nonlinear Filtering with Brenier Optimal Transport Maps [4.745059103971596]
This paper is concerned with the problem of nonlinear filtering, i.e., computing the conditional distribution of the state of a dynamical system.
Conventional sequential importance resampling (SIR) particle filters suffer from fundamental limitations, in scenarios involving degenerate likelihoods or high-dimensional states.
In this paper, we explore an alternative method, which is based on estimating the Brenier optimal transport (OT) map from the current prior distribution of the state to the posterior distribution at the next time step.
arXiv Detail & Related papers (2023-10-21T01:34:30Z) - An overview of differentiable particle filters for data-adaptive
sequential Bayesian inference [19.09640071505051]
Particle filters (PFs) provide an efficient mechanism for solving non-linear sequential state estimation problems.
An emerging trend involves constructing components of particle filters using neural networks and optimising them by gradient descent.
Differentiable particle filters are a promising computational tool for performing inference on sequential data in complex, high-dimensional tasks.
arXiv Detail & Related papers (2023-02-19T18:03:53Z) - Sampling with Mollified Interaction Energy Descent [57.00583139477843]
We present a new optimization-based method for sampling called mollified interaction energy descent (MIED)
MIED minimizes a new class of energies on probability measures called mollified interaction energies (MIEs)
We show experimentally that for unconstrained sampling problems our algorithm performs on par with existing particle-based algorithms like SVGD.
arXiv Detail & Related papers (2022-10-24T16:54:18Z) - Matching Normalizing Flows and Probability Paths on Manifolds [57.95251557443005]
Continuous Normalizing Flows (CNFs) are generative models that transform a prior distribution to a model distribution by solving an ordinary differential equation (ODE)
We propose to train CNFs by minimizing probability path divergence (PPD), a novel family of divergences between the probability density path generated by the CNF and a target probability density path.
We show that CNFs learned by minimizing PPD achieve state-of-the-art results in likelihoods and sample quality on existing low-dimensional manifold benchmarks.
arXiv Detail & Related papers (2022-07-11T08:50:19Z) - Computational Doob's h-transforms for Online Filtering of Discretely
Observed Diffusions [65.74069050283998]
We propose a computational framework to approximate Doob's $h$-transforms.
The proposed approach can be orders of magnitude more efficient than state-of-the-art particle filters.
arXiv Detail & Related papers (2022-06-07T15:03:05Z) - Fourier Series Expansion Based Filter Parametrization for Equivariant
Convolutions [73.33133942934018]
2D filter parametrization technique plays an important role when designing equivariant convolutions.
New equivariant convolution method based on the proposed filter parametrization method, named F-Conv.
F-Conv evidently outperforms previous filter parametrization based method in image super-resolution task.
arXiv Detail & Related papers (2021-07-30T10:01:52Z) - Variational Rejection Particle Filtering [28.03831528555717]
Variational Rejection Particle Filtering (VRPF) leads to novel variational bounds on the marginal likelihood.
We present theoretical properties of the variational bound and demonstrate experiments on various models of sequential data.
arXiv Detail & Related papers (2021-03-29T05:29:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.