Interpretable reduced-order modeling with time-scale separation
- URL: http://arxiv.org/abs/2303.02189v1
- Date: Fri, 3 Mar 2023 19:23:59 GMT
- Title: Interpretable reduced-order modeling with time-scale separation
- Authors: Sebastian Kaltenbach, Phaedon-Stelios Koutsourelakis and Petros
Koumoutsakos
- Abstract summary: Partial Differential Equations (PDEs) with high dimensionality are commonly encountered in computational physics and engineering.
We propose a data-driven scheme that automates the identification of the time-scales involved.
We show that this data-driven scheme can automatically learn the independent processes that decompose a system of linear ODEs.
- Score: 9.889399863931676
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Partial Differential Equations (PDEs) with high dimensionality are commonly
encountered in computational physics and engineering. However, finding
solutions for these PDEs can be computationally expensive, making model-order
reduction crucial. We propose such a data-driven scheme that automates the
identification of the time-scales involved and can produce stable predictions
forward in time as well as under different initial conditions not included in
the training data. To this end, we combine a non-linear autoencoder
architecture with a time-continuous model for the latent dynamics in the
complex space. It readily allows for the inclusion of sparse and irregularly
sampled training data. The learned, latent dynamics are interpretable and
reveal the different temporal scales involved. We show that this data-driven
scheme can automatically learn the independent processes that decompose a
system of linear ODEs along the eigenvectors of the system's matrix. Apart from
this, we demonstrate the applicability of the proposed framework in a hidden
Markov Model and the (discretized) Kuramoto-Shivashinsky (KS) equation.
Additionally, we propose a probabilistic version, which captures predictive
uncertainties and further improves upon the results of the deterministic
framework.
Related papers
- Deep Generative Modeling for Identification of Noisy, Non-Stationary Dynamical Systems [3.1484174280822845]
We focus on finding parsimonious ordinary differential equation (ODE) models for nonlinear, noisy, and non-autonomous dynamical systems.
Our method, dynamic SINDy, combines variational inference with SINDy (sparse identification of nonlinear dynamics) to model time-varying coefficients of sparse ODEs.
arXiv Detail & Related papers (2024-10-02T23:00:00Z) - On the Trajectory Regularity of ODE-based Diffusion Sampling [79.17334230868693]
Diffusion-based generative models use differential equations to establish a smooth connection between a complex data distribution and a tractable prior distribution.
In this paper, we identify several intriguing trajectory properties in the ODE-based sampling process of diffusion models.
arXiv Detail & Related papers (2024-05-18T15:59:41Z) - Probabilistic Forecasting with Stochastic Interpolants and Föllmer Processes [18.344934424278048]
We propose a framework for probabilistic forecasting of dynamical systems based on generative modeling.
We show that the drift and the diffusion coefficients of this SDE can be adjusted after training, and that a specific choice that minimizes the impact of the estimation error gives a F"ollmer process.
arXiv Detail & Related papers (2024-03-20T16:33:06Z) - InVAErt networks: a data-driven framework for model synthesis and
identifiability analysis [0.0]
inVAErt is a framework for data-driven analysis and synthesis of physical systems.
It uses a deterministic decoder to represent the forward and inverse maps, a normalizing flow to capture the probabilistic distribution of system outputs, and a variational encoder to learn a compact latent representation for the lack of bijectivity between inputs and outputs.
arXiv Detail & Related papers (2023-07-24T07:58:18Z) - Capturing dynamical correlations using implicit neural representations [85.66456606776552]
We develop an artificial intelligence framework which combines a neural network trained to mimic simulated data from a model Hamiltonian with automatic differentiation to recover unknown parameters from experimental data.
In doing so, we illustrate the ability to build and train a differentiable model only once, which then can be applied in real-time to multi-dimensional scattering data.
arXiv Detail & Related papers (2023-04-08T07:55:36Z) - $Φ$-DVAE: Physics-Informed Dynamical Variational Autoencoders for Unstructured Data Assimilation [3.2873782624127843]
We develop a physics-informed dynamical variational autoencoder ($Phi$-DVAE) to embed diverse data streams into time-evolving physical systems.
Our approach combines a standard, possibly nonlinear, filter for the latent state-space model and a VAE, to assimilate the unstructured data into the latent dynamical system.
A variational Bayesian framework is used for the joint estimation of the encoding, latent states, and unknown system parameters.
arXiv Detail & Related papers (2022-09-30T17:34:48Z) - Capturing Actionable Dynamics with Structured Latent Ordinary
Differential Equations [68.62843292346813]
We propose a structured latent ODE model that captures system input variations within its latent representation.
Building on a static variable specification, our model learns factors of variation for each input to the system, thus separating the effects of the system inputs in the latent space.
arXiv Detail & Related papers (2022-02-25T20:00:56Z) - Mixed Effects Neural ODE: A Variational Approximation for Analyzing the
Dynamics of Panel Data [50.23363975709122]
We propose a probabilistic model called ME-NODE to incorporate (fixed + random) mixed effects for analyzing panel data.
We show that our model can be derived using smooth approximations of SDEs provided by the Wong-Zakai theorem.
We then derive Evidence Based Lower Bounds for ME-NODE, and develop (efficient) training algorithms.
arXiv Detail & Related papers (2022-02-18T22:41:51Z) - Learning Time-Varying Graphs from Online Data [39.21234914444073]
This work proposes an algorithmic framework to learn time-varying graphs from online data.
It renders it model-independent, i.e., it can be theoretically analyzed in its abstract formulation.
We specialize the framework to three well-known graph learning models, namely, the Gaussian graphical model (GGM), the structural equation model (SEM), and the smoothness-based model (SBM)
arXiv Detail & Related papers (2021-10-21T09:46:44Z) - Using Data Assimilation to Train a Hybrid Forecast System that Combines
Machine-Learning and Knowledge-Based Components [52.77024349608834]
We consider the problem of data-assisted forecasting of chaotic dynamical systems when the available data is noisy partial measurements.
We show that by using partial measurements of the state of the dynamical system, we can train a machine learning model to improve predictions made by an imperfect knowledge-based model.
arXiv Detail & Related papers (2021-02-15T19:56:48Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.