Bayesian System ID: Optimal management of parameter, model, and
measurement uncertainty
- URL: http://arxiv.org/abs/2003.02359v1
- Date: Wed, 4 Mar 2020 22:48:30 GMT
- Title: Bayesian System ID: Optimal management of parameter, model, and
measurement uncertainty
- Authors: Nicholas Galioto and Alex Gorodetsky
- Abstract summary: We evaluate the robustness of a probabilistic formulation of system identification (ID) to sparse, noisy, and indirect data.
We show that the log posterior has improved geometric properties compared with the objective function surfaces of traditional methods.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We evaluate the robustness of a probabilistic formulation of system
identification (ID) to sparse, noisy, and indirect data. Specifically, we
compare estimators of future system behavior derived from the Bayesian
posterior of a learning problem to several commonly used least squares-based
optimization objectives used in system ID. Our comparisons indicate that the
log posterior has improved geometric properties compared with the objective
function surfaces of traditional methods that include differentially
constrained least squares and least squares reconstructions of discrete time
steppers like dynamic mode decomposition (DMD). These properties allow it to be
both more sensitive to new data and less affected by multiple minima ---
overall yielding a more robust approach. Our theoretical results indicate that
least squares and regularized least squares methods like dynamic mode
decomposition and sparse identification of nonlinear dynamics (SINDy) can be
derived from the probabilistic formulation by assuming noiseless measurements.
We also analyze the computational complexity of a Gaussian filter-based
approximate marginal Markov Chain Monte Carlo scheme that we use to obtain the
Bayesian posterior for both linear and nonlinear problems. We then empirically
demonstrate that obtaining the marginal posterior of the parameter dynamics and
making predictions by extracting optimal estimators (e.g., mean, median, mode)
yields orders of magnitude improvement over the aforementioned approaches. We
attribute this performance to the fact that the Bayesian approach captures
parameter, model, and measurement uncertainties, whereas the other methods
typically neglect at least one type of uncertainty.
Related papers
- Convergence of uncertainty estimates in Ensemble and Bayesian sparse
model discovery [4.446017969073817]
We show empirical success in terms of accuracy and robustness to noise with bootstrapping-based sequential thresholding least-squares estimator.
We show that this bootstrapping-based ensembling technique can perform a provably correct variable selection procedure with an exponential convergence rate of the error rate.
arXiv Detail & Related papers (2023-01-30T04:07:59Z) - Robust identification of non-autonomous dynamical systems using
stochastic dynamics models [0.0]
This paper considers the problem of system identification (ID) of linear and nonlinear non-autonomous systems from noisy and sparse data.
We propose and analyze an objective function derived from a Bayesian formulation for learning a hidden Markov model.
We show that our proposed approach has improved smoothness and inherent regularization that make it well-suited for system ID.
arXiv Detail & Related papers (2022-12-20T16:36:23Z) - Bayesian Spline Learning for Equation Discovery of Nonlinear Dynamics
with Quantified Uncertainty [8.815974147041048]
We develop a novel framework to identify parsimonious governing equations of nonlinear (spatiotemporal) dynamics from sparse, noisy data with quantified uncertainty.
The proposed algorithm is evaluated on multiple nonlinear dynamical systems governed by canonical ordinary and partial differential equations.
arXiv Detail & Related papers (2022-10-14T20:37:36Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - A Priori Denoising Strategies for Sparse Identification of Nonlinear
Dynamical Systems: A Comparative Study [68.8204255655161]
We investigate and compare the performance of several local and global smoothing techniques to a priori denoise the state measurements.
We show that, in general, global methods, which use the entire measurement data set, outperform local methods, which employ a neighboring data subset around a local point.
arXiv Detail & Related papers (2022-01-29T23:31:25Z) - Time varying regression with hidden linear dynamics [74.9914602730208]
We revisit a model for time-varying linear regression that assumes the unknown parameters evolve according to a linear dynamical system.
Counterintuitively, we show that when the underlying dynamics are stable the parameters of this model can be estimated from data by combining just two ordinary least squares estimates.
arXiv Detail & Related papers (2021-12-29T23:37:06Z) - Inverting brain grey matter models with likelihood-free inference: a
tool for trustable cytoarchitecture measurements [62.997667081978825]
characterisation of the brain grey matter cytoarchitecture with quantitative sensitivity to soma density and volume remains an unsolved challenge in dMRI.
We propose a new forward model, specifically a new system of equations, requiring a few relatively sparse b-shells.
We then apply modern tools from Bayesian analysis known as likelihood-free inference (LFI) to invert our proposed model.
arXiv Detail & Related papers (2021-11-15T09:08:27Z) - MINIMALIST: Mutual INformatIon Maximization for Amortized Likelihood
Inference from Sampled Trajectories [61.3299263929289]
Simulation-based inference enables learning the parameters of a model even when its likelihood cannot be computed in practice.
One class of methods uses data simulated with different parameters to infer an amortized estimator for the likelihood-to-evidence ratio.
We show that this approach can be formulated in terms of mutual information between model parameters and simulated data.
arXiv Detail & Related papers (2021-06-03T12:59:16Z) - Slice Sampling for General Completely Random Measures [74.24975039689893]
We present a novel Markov chain Monte Carlo algorithm for posterior inference that adaptively sets the truncation level using auxiliary slice variables.
The efficacy of the proposed algorithm is evaluated on several popular nonparametric models.
arXiv Detail & Related papers (2020-06-24T17:53:53Z) - Statistical Outlier Identification in Multi-robot Visual SLAM using
Expectation Maximization [18.259478519717426]
This paper introduces a novel and distributed method for detecting inter-map loop closure outliers in simultaneous localization and mapping (SLAM)
The proposed algorithm does not rely on a good initialization and can handle more than two maps at a time.
arXiv Detail & Related papers (2020-02-07T06:34:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.