Using machine learning to correct model error in data assimilation and
forecast applications
- URL: http://arxiv.org/abs/2010.12605v2
- Date: Mon, 10 May 2021 06:46:32 GMT
- Title: Using machine learning to correct model error in data assimilation and
forecast applications
- Authors: Alban Farchi and Patrick Laloyaux and Massimo Bonavita and Marc
Bocquet
- Abstract summary: We propose to use this method to correct the error of an existent, knowledge-based model.
The resulting surrogate model is an hybrid model between the original (knowledge-based) model and the ML model.
Using the hybrid surrogate models for DA yields a significantly better analysis than using the original model.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The idea of using machine learning (ML) methods to reconstruct the dynamics
of a system is the topic of recent studies in the geosciences, in which the key
output is a surrogate model meant to emulate the dynamical model. In order to
treat sparse and noisy observations in a rigorous way, ML can be combined to
data assimilation (DA). This yields a class of iterative methods in which, at
each iteration a DA step assimilates the observations, and alternates with a ML
step to learn the underlying dynamics of the DA analysis. In this article, we
propose to use this method to correct the error of an existent, knowledge-based
model. In practice, the resulting surrogate model is an hybrid model between
the original (knowledge-based) model and the ML model. We demonstrate
numerically the feasibility of the method using a two-layer, two-dimensional
quasi-geostrophic channel model. Model error is introduced by the means of
perturbed parameters. The DA step is performed using the strong-constraint
4D-Var algorithm, while the ML step is performed using deep learning tools. The
ML models are able to learn a substantial part of the model error and the
resulting hybrid surrogate models produce better short- to mid-range forecasts.
Furthermore, using the hybrid surrogate models for DA yields a significantly
better analysis than using the original model.
Related papers
- SMILE: Zero-Shot Sparse Mixture of Low-Rank Experts Construction From Pre-Trained Foundation Models [85.67096251281191]
We present an innovative approach to model fusion called zero-shot Sparse MIxture of Low-rank Experts (SMILE) construction.
SMILE allows for the upscaling of source models into an MoE model without extra data or further training.
We conduct extensive experiments across diverse scenarios, such as image classification and text generation tasks, using full fine-tuning and LoRA fine-tuning.
arXiv Detail & Related papers (2024-08-19T17:32:15Z) - Analytical results for uncertainty propagation through trained machine learning regression models [0.10878040851637999]
This paper addresses the challenge of uncertainty propagation through trained/fixed machine learning (ML) regression models.
We present numerical experiments in which we validate our methods and compare them with a Monte Carlo approach from a computational efficiency point of view.
arXiv Detail & Related papers (2024-04-17T10:16:20Z) - A Two-Scale Complexity Measure for Deep Learning Models [2.7446241148152257]
We introduce a novel capacity measure 2sED for statistical models based on the effective dimension.
The new quantity provably bounds the generalization error under mild assumptions on the model.
simulations on standard data sets and popular model architectures show that 2sED correlates well with the training error.
arXiv Detail & Related papers (2024-01-17T12:50:50Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Learning from aggregated data with a maximum entropy model [73.63512438583375]
We show how a new model, similar to a logistic regression, may be learned from aggregated data only by approximating the unobserved feature distribution with a maximum entropy hypothesis.
We present empirical evidence on several public datasets that the model learned this way can achieve performances comparable to those of a logistic model trained with the full unaggregated data.
arXiv Detail & Related papers (2022-10-05T09:17:27Z) - Real-to-Sim: Predicting Residual Errors of Robotic Systems with Sparse
Data using a Learning-based Unscented Kalman Filter [65.93205328894608]
We learn the residual errors between a dynamic and/or simulator model and the real robot.
We show that with the learned residual errors, we can further close the reality gap between dynamic models, simulations, and actual hardware.
arXiv Detail & Related papers (2022-09-07T15:15:12Z) - On the Influence of Enforcing Model Identifiability on Learning dynamics
of Gaussian Mixture Models [14.759688428864159]
We propose a technique for extracting submodels from singular models.
Our method enforces model identifiability during training.
We show how the method can be applied to more complex models like deep neural networks.
arXiv Detail & Related papers (2022-06-17T07:50:22Z) - Bayesian Active Learning for Discrete Latent Variable Models [19.852463786440122]
Active learning seeks to reduce the amount of data required to fit the parameters of a model.
latent variable models play a vital role in neuroscience, psychology, and a variety of other engineering and scientific disciplines.
arXiv Detail & Related papers (2022-02-27T19:07:12Z) - Learning continuous models for continuous physics [94.42705784823997]
We develop a test based on numerical analysis theory to validate machine learning models for science and engineering applications.
Our results illustrate how principled numerical analysis methods can be coupled with existing ML training/testing methodologies to validate models for science and engineering applications.
arXiv Detail & Related papers (2022-02-17T07:56:46Z) - Distilling Interpretable Models into Human-Readable Code [71.11328360614479]
Human-readability is an important and desirable standard for machine-learned model interpretability.
We propose to train interpretable models using conventional methods, and then distill them into concise, human-readable code.
We describe a piecewise-linear curve-fitting algorithm that produces high-quality results efficiently and reliably across a broad range of use cases.
arXiv Detail & Related papers (2021-01-21T01:46:36Z) - Combining data assimilation and machine learning to infer unresolved
scale parametrisation [0.0]
In recent years, machine learning has been proposed to devise data-driven parametrisations of unresolved processes in dynamical numerical models.
Our goal is to go beyond the use of high-resolution simulations and train ML-based parametrisation using direct data.
We show that in both cases the hybrid model yields forecasts with better skill than the truncated model.
arXiv Detail & Related papers (2020-09-09T14:12:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.