Exact nonlinear state estimation
- URL: http://arxiv.org/abs/2310.10976v1
- Date: Tue, 17 Oct 2023 03:44:29 GMT
- Title: Exact nonlinear state estimation
- Authors: Hristo G. Chipilski
- Abstract summary: The majority of data assimilation methods in the geosciences are based on Gaussian assumptions.
Non-parametric, particle-based DA algorithms have superior accuracy, but their application to high-dimensional models still poses operational challenges.
This article introduces a new nonlinear estimation theory which attempts to bridge the existing gap in DA methodology.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The majority of data assimilation (DA) methods in the geosciences are based
on Gaussian assumptions. While these assumptions facilitate efficient
algorithms, they cause analysis biases and subsequent forecast degradations.
Non-parametric, particle-based DA algorithms have superior accuracy, but their
application to high-dimensional models still poses operational challenges.
Drawing inspiration from recent advances in the field of generative artificial
intelligence (AI), this article introduces a new nonlinear estimation theory
which attempts to bridge the existing gap in DA methodology. Specifically, a
Conjugate Transform Filter (CTF) is derived and shown to generalize the
celebrated Kalman filter to arbitrarily non-Gaussian distributions. The new
filter has several desirable properties, such as its ability to preserve
statistical relationships in the prior state and convergence to highly accurate
observations. An ensemble approximation of the new theory (ECTF) is also
presented and validated using idealized statistical experiments that feature
bounded quantities with non-Gaussian distributions, a prevalent challenge in
Earth system models. Results from these experiments indicate that the greatest
benefits from ECTF occur when observation errors are small relative to the
forecast uncertainty and when state variables exhibit strong nonlinear
dependencies. Ultimately, the new filtering theory offers exciting avenues for
improving conventional DA algorithms through their principled integration with
AI techniques.
Related papers
- State-observation augmented diffusion model for nonlinear assimilation [6.682908186025083]
We propose a novel data-driven assimilation algorithm based on generative models.
Our State-Observation Augmented Diffusion (SOAD) model is designed to handle nonlinear physical and observational models more effectively.
arXiv Detail & Related papers (2024-07-31T03:47:20Z) - Inflationary Flows: Calibrated Bayesian Inference with Diffusion-Based Models [0.0]
We show how diffusion-based models can be repurposed for performing principled, identifiable Bayesian inference.
We show how such maps can be learned via standard DBM training using a novel noise schedule.
The result is a class of highly expressive generative models, uniquely defined on a low-dimensional latent space.
arXiv Detail & Related papers (2024-07-11T19:58:19Z) - A PAC-Bayesian Perspective on the Interpolating Information Criterion [54.548058449535155]
We show how a PAC-Bayes bound is obtained for a general class of models, characterizing factors which influence performance in the interpolating regime.
We quantify how the test error for overparameterized models achieving effectively zero training error depends on the quality of the implicit regularization imposed by e.g. the combination of model, parameter-initialization scheme.
arXiv Detail & Related papers (2023-11-13T01:48:08Z) - A Unified Momentum-based Paradigm of Decentralized SGD for Non-Convex
Models and Heterogeneous Data [0.261072980439312]
We propose a unified paradigm called U.MP, D-MP and GT-D, which provides a convergence guarantee for non general objectives.
In theory we provide the convergence analysis objectives two approaches for these non-MP algorithms.
arXiv Detail & Related papers (2023-03-01T02:13:22Z) - Efficient CDF Approximations for Normalizing Flows [64.60846767084877]
We build upon the diffeomorphic properties of normalizing flows to estimate the cumulative distribution function (CDF) over a closed region.
Our experiments on popular flow architectures and UCI datasets show a marked improvement in sample efficiency as compared to traditional estimators.
arXiv Detail & Related papers (2022-02-23T06:11:49Z) - Learning to Estimate Without Bias [57.82628598276623]
Gauss theorem states that the weighted least squares estimator is a linear minimum variance unbiased estimation (MVUE) in linear models.
In this paper, we take a first step towards extending this result to non linear settings via deep learning with bias constraints.
A second motivation to BCE is in applications where multiple estimates of the same unknown are averaged for improved performance.
arXiv Detail & Related papers (2021-10-24T10:23:51Z) - On the Double Descent of Random Features Models Trained with SGD [78.0918823643911]
We study properties of random features (RF) regression in high dimensions optimized by gradient descent (SGD)
We derive precise non-asymptotic error bounds of RF regression under both constant and adaptive step-size SGD setting.
We observe the double descent phenomenon both theoretically and empirically.
arXiv Detail & Related papers (2021-10-13T17:47:39Z) - Low-rank statistical finite elements for scalable model-data synthesis [0.8602553195689513]
statFEM acknowledges a priori model misspecification, by embedding forcing within the governing equations.
The method reconstructs the observed data-generating processes with minimal loss of information.
This article overcomes this hurdle by embedding a low-rank approximation of the underlying dense covariance matrix.
arXiv Detail & Related papers (2021-09-10T09:51:43Z) - Loss function based second-order Jensen inequality and its application
to particle variational inference [112.58907653042317]
Particle variational inference (PVI) uses an ensemble of models as an empirical approximation for the posterior distribution.
PVI iteratively updates each model with a repulsion force to ensure the diversity of the optimized models.
We derive a novel generalization error bound and show that it can be reduced by enhancing the diversity of models.
arXiv Detail & Related papers (2021-06-09T12:13:51Z) - Latent Gaussian Model Boosting [0.0]
Tree-boosting shows excellent predictive accuracy on many data sets.
We obtain increased predictive accuracy compared to existing approaches in both simulated and real-world data experiments.
arXiv Detail & Related papers (2021-05-19T07:36:30Z) - CASTLE: Regularization via Auxiliary Causal Graph Discovery [89.74800176981842]
We introduce Causal Structure Learning (CASTLE) regularization and propose to regularize a neural network by jointly learning the causal relationships between variables.
CASTLE efficiently reconstructs only the features in the causal DAG that have a causal neighbor, whereas reconstruction-based regularizers suboptimally reconstruct all input features.
arXiv Detail & Related papers (2020-09-28T09:49:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.