State and parameter learning with PaRIS particle Gibbs
- URL: http://arxiv.org/abs/2301.00900v1
- Date: Mon, 2 Jan 2023 23:27:33 GMT
- Title: State and parameter learning with PaRIS particle Gibbs
- Authors: Gabriel Cardoso, Yazid Janati El Idrissi, Sylvain Le Corff, Eric
Moulines, Jimmy Olsson
- Abstract summary: Non-linear state-space models are ubiquitous in statistical machine learning.
PaRIS is a sequential Monte Carlo technique allowing for efficient online approximation of expectations of additive functionals.
We design a novel additive smoothing algorithm, the Parisian particle Gibbs PPG sampler, which can be viewed as a PaRIS algorithm driven by conditional SMC moves.
- Score: 11.290331898505594
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Non-linear state-space models, also known as general hidden Markov models,
are ubiquitous in statistical machine learning, being the most classical
generative models for serial data and sequences in general. The particle-based,
rapid incremental smoother PaRIS is a sequential Monte Carlo (SMC) technique
allowing for efficient online approximation of expectations of additive
functionals under the smoothing distribution in these models. Such expectations
appear naturally in several learning contexts, such as likelihood estimation
(MLE) and Markov score climbing (MSC). PARIS has linear computational
complexity, limited memory requirements and comes with non-asymptotic bounds,
convergence results and stability guarantees. Still, being based on
self-normalised importance sampling, the PaRIS estimator is biased. Our first
contribution is to design a novel additive smoothing algorithm, the Parisian
particle Gibbs PPG sampler, which can be viewed as a PaRIS algorithm driven by
conditional SMC moves, resulting in bias-reduced estimates of the targeted
quantities. We substantiate the PPG algorithm with theoretical results,
including new bounds on bias and variance as well as deviation inequalities.
Our second contribution is to apply PPG in a learning framework, covering MLE
and MSC as special examples. In this context, we establish, under standard
assumptions, non-asymptotic bounds highlighting the value of bias reduction and
the implicit Rao--Blackwellization of PPG. These are the first non-asymptotic
results of this kind in this setting. We illustrate our theoretical results
with numerical experiments supporting our claims.
Related papers
- Polynomial Chaos Expanded Gaussian Process [2.287415292857564]
In complex and unknown processes, global models are initially generated over the entire experimental space.
This study addresses the need for models that effectively represent both global and local experimental spaces.
arXiv Detail & Related papers (2024-05-02T07:11:05Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - Algebraic and Statistical Properties of the Ordinary Least Squares Interpolator [3.4320157633663064]
We provide results for the minimum $ell$-norm OLS interpolator.
We present statistical results such as an extension of the Gauss-Markov theorem.
We conduct simulations that further explore the properties of the OLS interpolator.
arXiv Detail & Related papers (2023-09-27T16:41:10Z) - PAPAL: A Provable PArticle-based Primal-Dual ALgorithm for Mixed Nash
Equilibrium [62.51015395213579]
We consider the non-AL equilibrium nonconptotic objective function in two-player zero-sum continuous games.
The proposed algorithm employs the movements of particles to represent the updates of random strategies for the $ilon$-mixed Nash equilibrium.
arXiv Detail & Related papers (2023-03-02T05:08:15Z) - GEC: A Unified Framework for Interactive Decision Making in MDP, POMDP,
and Beyond [101.5329678997916]
We study sample efficient reinforcement learning (RL) under the general framework of interactive decision making.
We propose a novel complexity measure, generalized eluder coefficient (GEC), which characterizes the fundamental tradeoff between exploration and exploitation.
We show that RL problems with low GEC form a remarkably rich class, which subsumes low Bellman eluder dimension problems, bilinear class, low witness rank problems, PO-bilinear class, and generalized regular PSR.
arXiv Detail & Related papers (2022-11-03T16:42:40Z) - Learning Graphical Factor Models with Riemannian Optimization [70.13748170371889]
This paper proposes a flexible algorithmic framework for graph learning under low-rank structural constraints.
The problem is expressed as penalized maximum likelihood estimation of an elliptical distribution.
We leverage geometries of positive definite matrices and positive semi-definite matrices of fixed rank that are well suited to elliptical models.
arXiv Detail & Related papers (2022-10-21T13:19:45Z) - Optimal Estimation of Generic Dynamics by Path-Dependent Neural Jump ODEs [3.74142789780782]
This paper studies the problem of forecasting general processes using a path-dependent extension of the Neural Jump ODE (NJ-ODE) framework citepherreraneural.
We show that PD-NJ-ODE can be applied successfully to classical filtering problems and to limit order book (LOB) data.
arXiv Detail & Related papers (2022-06-28T20:50:14Z) - Optimal regularizations for data generation with probabilistic graphical
models [0.0]
Empirically, well-chosen regularization schemes dramatically improve the quality of the inferred models.
We consider the particular case of L 2 and L 1 regularizations in the Maximum A Posteriori (MAP) inference of generative pairwise graphical models.
arXiv Detail & Related papers (2021-12-02T14:45:16Z) - Learning to Estimate Without Bias [57.82628598276623]
Gauss theorem states that the weighted least squares estimator is a linear minimum variance unbiased estimation (MVUE) in linear models.
In this paper, we take a first step towards extending this result to non linear settings via deep learning with bias constraints.
A second motivation to BCE is in applications where multiple estimates of the same unknown are averaged for improved performance.
arXiv Detail & Related papers (2021-10-24T10:23:51Z) - Hierarchical Gaussian Process Models for Regression Discontinuity/Kink
under Sharp and Fuzzy Designs [0.0]
We propose nonparametric Bayesian estimators for causal inference exploiting Regression Discontinuity/Kink (RD/RK)
These estimators are extended to hierarchical GP models with an intermediate Bayesian neural network layer.
Monte Carlo simulations show that our estimators perform similarly and often better than competing estimators in terms of precision, coverage and interval length.
arXiv Detail & Related papers (2021-10-03T04:23:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.