Active operator inference for learning low-dimensional dynamical-system
models from noisy data
- URL: http://arxiv.org/abs/2107.09256v1
- Date: Tue, 20 Jul 2021 04:30:07 GMT
- Title: Active operator inference for learning low-dimensional dynamical-system
models from noisy data
- Authors: Wayne Isaac Tan Uy, Yuepeng Wang, Yuxiao Wen, Benjamin Peherstorfer
- Abstract summary: Noise poses a challenge for learning dynamical-system models because already small variations can distort the dynamics described by trajectory data.
This work builds on operator inference from scientific machine learning to infer low-dimensional models from high-dimensional state trajectories polluted with noise.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Noise poses a challenge for learning dynamical-system models because already
small variations can distort the dynamics described by trajectory data. This
work builds on operator inference from scientific machine learning to infer
low-dimensional models from high-dimensional state trajectories polluted with
noise. The presented analysis shows that, under certain conditions, the
inferred operators are unbiased estimators of the well-studied projection-based
reduced operators from traditional model reduction. Furthermore, the connection
between operator inference and projection-based model reduction enables
bounding the mean-squared errors of predictions made with the learned models
with respect to traditional reduced models. The analysis also motivates an
active operator inference approach that judiciously samples high-dimensional
trajectories with the aim of achieving a low mean-squared error by reducing the
effect of noise. Numerical experiments with high-dimensional linear and
nonlinear state dynamics demonstrate that predictions obtained with active
operator inference have orders of magnitude lower mean-squared errors than
operator inference with traditional, equidistantly sampled trajectory data.
Related papers
- Deep Learning for Koopman Operator Estimation in Idealized Atmospheric Dynamics [2.2489531925874013]
Deep learning is revolutionizing weather forecasting, with new data-driven models achieving accuracy on par with operational physical models for medium-term predictions.
These models often lack interpretability, making their underlying dynamics difficult to understand and explain.
This paper proposes methodologies to estimate the Koopman operator, providing a linear representation of complex nonlinear dynamics to enhance the transparency of data-driven models.
arXiv Detail & Related papers (2024-09-10T13:56:54Z) - Low-rank finetuning for LLMs: A fairness perspective [54.13240282850982]
Low-rank approximation techniques have become the de facto standard for fine-tuning Large Language Models.
This paper investigates the effectiveness of these methods in capturing the shift of fine-tuning datasets from the initial pre-trained data distribution.
We show that low-rank fine-tuning inadvertently preserves undesirable biases and toxic behaviors.
arXiv Detail & Related papers (2024-05-28T20:43:53Z) - Learning with Noisy Foundation Models [95.50968225050012]
This paper is the first work to comprehensively understand and analyze the nature of noise in pre-training datasets.
We propose a tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization.
arXiv Detail & Related papers (2024-03-11T16:22:41Z) - Towards Theoretical Understandings of Self-Consuming Generative Models [56.84592466204185]
This paper tackles the emerging challenge of training generative models within a self-consuming loop.
We construct a theoretical framework to rigorously evaluate how this training procedure impacts the data distributions learned by future models.
We present results for kernel density estimation, delivering nuanced insights such as the impact of mixed data training on error propagation.
arXiv Detail & Related papers (2024-02-19T02:08:09Z) - Data-driven Nonlinear Model Reduction using Koopman Theory: Integrated
Control Form and NMPC Case Study [56.283944756315066]
We propose generic model structures combining delay-coordinate encoding of measurements and full-state decoding to integrate reduced Koopman modeling and state estimation.
A case study demonstrates that our approach provides accurate control models and enables real-time capable nonlinear model predictive control of a high-purity cryogenic distillation column.
arXiv Detail & Related papers (2024-01-09T11:54:54Z) - A PAC-Bayesian Perspective on the Interpolating Information Criterion [54.548058449535155]
We show how a PAC-Bayes bound is obtained for a general class of models, characterizing factors which influence performance in the interpolating regime.
We quantify how the test error for overparameterized models achieving effectively zero training error depends on the quality of the implicit regularization imposed by e.g. the combination of model, parameter-initialization scheme.
arXiv Detail & Related papers (2023-11-13T01:48:08Z) - Operator inference with roll outs for learning reduced models from
scarce and low-quality data [0.0]
We propose to combine data-driven modeling via operator inference with the dynamic training via roll outs of neural ordinary differential equations.
We show that operator inference with roll outs provides predictive models from training trajectories even if data are sampled sparsely in time and polluted with noise of up to 10%.
arXiv Detail & Related papers (2022-12-02T19:41:31Z) - Deep Kernel Learning of Dynamical Models from High-Dimensional Noisy
Data [1.3750624267664155]
The framework is composed of an encoder that compresses high-dimensional measurements into low-dimensional state variables.
The training of the proposed model is carried out in an unsupervised manner, not relying on labeled data.
Results show that the method can effectively denoise measurements, learn compact state representations and latent dynamical models.
arXiv Detail & Related papers (2022-08-27T09:47:44Z) - How robust are pre-trained models to distribution shift? [82.08946007821184]
We show how spurious correlations affect the performance of popular self-supervised learning (SSL) and auto-encoder based models (AE)
We develop a novel evaluation scheme with the linear head trained on out-of-distribution (OOD) data, to isolate the performance of the pre-trained models from a potential bias of the linear head used for evaluation.
arXiv Detail & Related papers (2022-06-17T16:18:28Z) - Operator inference of non-Markovian terms for learning reduced models
from partially observed state trajectories [0.0]
This work introduces a non-intrusive model reduction approach for learning reduced models from trajectories of high-dimensional dynamical systems.
The proposed approach compensates for the loss of information due to the partially observed states by constructing non-Markovian reduced models.
Numerical results demonstrate that the proposed approach leads to non-Markovian reduced models that are predictive far beyond the training regime.
arXiv Detail & Related papers (2021-03-01T23:55:52Z) - Operator Inference and Physics-Informed Learning of Low-Dimensional
Models for Incompressible Flows [5.756349331930218]
We suggest a new approach to learning structured low-order models for incompressible flow from data.
We show that learning dynamics of the velocity and pressure can be decoupled, thus leading to an efficient operator inference approach.
arXiv Detail & Related papers (2020-10-13T21:26:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.