Deep Learning for the Benes Filter
- URL: http://arxiv.org/abs/2203.05561v1
- Date: Wed, 9 Mar 2022 14:08:38 GMT
- Title: Deep Learning for the Benes Filter
- Authors: Alexander Lobbe
- Abstract summary: We present a new numerical method based on the mesh-free neural network representation of the density of the solution of the Benes model.
We discuss the role of nonlinearity in the filtering model equations for the choice of the domain of the neural network.
- Score: 91.3755431537592
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Benes filter is a well-known continuous-time stochastic filtering model
in one dimension that has the advantage of being explicitly solvable. From an
evolution equation point of view, the Benes filter is also the solution of the
filtering equations given a particular set of coefficient functions. In
general, the filtering stochastic partial differential equations (SPDE) arise
as the evolution equations for the conditional distribution of an underlying
signal given partial, and possibly noisy, observations. Their numerical
approximation presents a central issue for theoreticians and practitioners
alike, who are actively seeking accurate and fast methods, especially for such
high-dimensional settings as numerical weather prediction, for example. In this
paper we present a brief study of a new numerical method based on the mesh-free
neural network representation of the density of the solution of the Benes model
achieved by deep learning. Based on the classical SPDE splitting method, our
algorithm includes a recursive normalisation procedure to recover the
normalised conditional distribution of the signal process. Within the
analytically tractable setting of the Benes filter, we discuss the role of
nonlinearity in the filtering model equations for the choice of the domain of
the neural network. Further we present the first study of the neural network
method with an adaptive domain for the Benes model.
Related papers
- Closed-form Filtering for Non-linear Systems [83.91296397912218]
We propose a new class of filters based on Gaussian PSD Models, which offer several advantages in terms of density approximation and computational efficiency.
We show that filtering can be efficiently performed in closed form when transitions and observations are Gaussian PSD Models.
Our proposed estimator enjoys strong theoretical guarantees, with estimation error that depends on the quality of the approximation and is adaptive to the regularity of the transition probabilities.
arXiv Detail & Related papers (2024-02-15T08:51:49Z) - Low-rank extended Kalman filtering for online learning of neural
networks from streaming data [71.97861600347959]
We propose an efficient online approximate Bayesian inference algorithm for estimating the parameters of a nonlinear function from a potentially non-stationary data stream.
The method is based on the extended Kalman filter (EKF), but uses a novel low-rank plus diagonal decomposition of the posterior matrix.
In contrast to methods based on variational inference, our method is fully deterministic, and does not require step-size tuning.
arXiv Detail & Related papers (2023-05-31T03:48:49Z) - Simple initialization and parametrization of sinusoidal networks via
their kernel bandwidth [92.25666446274188]
sinusoidal neural networks with activations have been proposed as an alternative to networks with traditional activation functions.
We first propose a simplified version of such sinusoidal neural networks, which allows both for easier practical implementation and simpler theoretical analysis.
We then analyze the behavior of these networks from the neural tangent kernel perspective and demonstrate that their kernel approximates a low-pass filter with an adjustable bandwidth.
arXiv Detail & Related papers (2022-11-26T07:41:48Z) - Computational Doob's h-transforms for Online Filtering of Discretely
Observed Diffusions [65.74069050283998]
We propose a computational framework to approximate Doob's $h$-transforms.
The proposed approach can be orders of magnitude more efficient than state-of-the-art particle filters.
arXiv Detail & Related papers (2022-06-07T15:03:05Z) - An energy-based deep splitting method for the nonlinear filtering
problem [0.0]
The main goal of this paper is to approximately solve the nonlinear filtering problem through deep learning.
This is achieved by solving the Zakai equation by a deep splitting method, previously developed for approximate solution of (stochastic) partial differential equations.
This is combined with an energy-based model for the approximation of functions by a deep neural network.
arXiv Detail & Related papers (2022-03-31T16:26:54Z) - An application of the splitting-up method for the computation of a
neural network representation for the solution for the filtering equations [68.8204255655161]
Filtering equations play a central role in many real-life applications, including numerical weather prediction, finance and engineering.
One of the classical approaches to approximate the solution of the filtering equations is to use a PDE inspired method, called the splitting-up method.
We combine this method with a neural network representation to produce an approximation of the unnormalised conditional distribution of the signal process.
arXiv Detail & Related papers (2022-01-10T11:01:36Z) - Non-parametric generalized linear model [7.936841911281107]
A fundamental problem in statistical neuroscience is to model how neurons encode information by analyzing electrophysiological recordings.
A popular and widely-used approach is to fit the spike trains with an autoregressive point process model.
In practice a sufficiently rich but small ensemble of temporal basis functions needs to be chosen to parameterize the filters.
arXiv Detail & Related papers (2020-09-02T21:54:53Z) - Parameterizing uncertainty by deep invertible networks, an application
to reservoir characterization [0.9176056742068814]
Uncertainty quantification for full-waveform inversion provides a probabilistic characterization of the ill-conditioning of the problem.
We propose an approach characterized by training a deep network that "pushes forward" Gaussian random inputs into the model space as if they were sampled from the actual posterior distribution.
arXiv Detail & Related papers (2020-04-16T18:37:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.