Model Order Reduction based on Runge-Kutta Neural Network
- URL: http://arxiv.org/abs/2103.13805v1
- Date: Thu, 25 Mar 2021 13:02:16 GMT
- Title: Model Order Reduction based on Runge-Kutta Neural Network
- Authors: Qinyu Zhuang, Juan Manuel Lorenzi, Hans-Joachim Bungartz, Dirk
Hartmann
- Abstract summary: In this work, we apply some modifications for both steps respectively and investigate how they are impacted by testing with three simulation models.
For the model reconstruction step, two types of neural network architectures are compared: Multilayer Perceptron (MLP) and Runge-Kutta Neural Network (RKNN)
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model Order Reduction (MOR) methods enable the generation of
real-time-capable digital twins, which can enable various novel value streams
in industry. While traditional projection-based methods are robust and accurate
for linear problems, incorporating Machine Learning to deal with nonlinearity
becomes a new choice for reducing complex problems. Such methods usually
consist of two steps. The first step is dimension reduction by projection-based
method, and the second is the model reconstruction by Neural Network. In this
work, we apply some modifications for both steps respectively and investigate
how they are impacted by testing with three simulation models. In all cases
Proper Orthogonal Decomposition (POD) is used for dimension reduction. For this
step, the effects of generating the input snapshot database with constant input
parameters is compared with time-dependent input parameters. For the model
reconstruction step, two types of neural network architectures are compared:
Multilayer Perceptron (MLP) and Runge-Kutta Neural Network (RKNN). The MLP
learns the system state directly while RKNN learns the derivative of system
state and predicts the new state as a Runge-Kutta integrator.
Related papers
- IVP-VAE: Modeling EHR Time Series with Initial Value Problem Solvers [20.784780497613557]
We propose to model time series purely with continuous processes whose state evolution can be approximated directly by IVPs.
This eliminates the need for recurrent computation and enables multiple states to evolve in parallel.
Experiments on three real-world datasets show that the proposed method can systematically outperform its predecessors, achieve state-of-the-art results, and have significant advantages in terms of data efficiency.
arXiv Detail & Related papers (2023-05-11T11:53:31Z) - A predictive physics-aware hybrid reduced order model for reacting flows [65.73506571113623]
A new hybrid predictive Reduced Order Model (ROM) is proposed to solve reacting flow problems.
The number of degrees of freedom is reduced from thousands of temporal points to a few POD modes with their corresponding temporal coefficients.
Two different deep learning architectures have been tested to predict the temporal coefficients.
arXiv Detail & Related papers (2023-01-24T08:39:20Z) - Online model error correction with neural networks in the incremental
4D-Var framework [0.0]
We develop a new weak-constraint 4D-Var formulation which can be used to train a neural network for online model error correction.
The method is implemented in the ECMWF Object-Oriented Prediction System.
The results confirm that online learning is effective and yields a more accurate model error correction than offline learning.
arXiv Detail & Related papers (2022-10-25T07:45:33Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - Dynamically-Scaled Deep Canonical Correlation Analysis [77.34726150561087]
Canonical Correlation Analysis (CCA) is a method for feature extraction of two views by finding maximally correlated linear projections of them.
We introduce a novel dynamic scaling method for training an input-dependent canonical correlation model.
arXiv Detail & Related papers (2022-03-23T12:52:49Z) - On the adaptation of recurrent neural networks for system identification [2.5234156040689237]
This paper presents a transfer learning approach which enables fast and efficient adaptation of Recurrent Neural Network (RNN) models of dynamical systems.
The system dynamics are then assumed to change, leading to an unacceptable degradation of the nominal model performance on the perturbed system.
To cope with the mismatch, the model is augmented with an additive correction term trained on fresh data from the new dynamic regime.
arXiv Detail & Related papers (2022-01-21T12:04:17Z) - Mixed Precision Low-bit Quantization of Neural Network Language Models
for Speech Recognition [67.95996816744251]
State-of-the-art language models (LMs) represented by long-short term memory recurrent neural networks (LSTM-RNNs) and Transformers are becoming increasingly complex and expensive for practical applications.
Current quantization methods are based on uniform precision and fail to account for the varying performance sensitivity at different parts of LMs to quantization errors.
Novel mixed precision neural network LM quantization methods are proposed in this paper.
arXiv Detail & Related papers (2021-11-29T12:24:02Z) - Accelerating Neural ODEs Using Model Order Reduction [0.0]
We show that mathematical model order reduction methods can be used for compressing and accelerating Neural ODEs.
We implement our novel compression method by developing Neural ODEs that integrate the necessary subspace-projection and operations as layers of the neural network.
arXiv Detail & Related papers (2021-05-28T19:27:09Z) - Self-Learning for Received Signal Strength Map Reconstruction with
Neural Architecture Search [63.39818029362661]
We present a model based on Neural Architecture Search (NAS) and self-learning for received signal strength ( RSS) map reconstruction.
The approach first finds an optimal NN architecture and simultaneously train the deduced model over some ground-truth measurements of a given ( RSS) map.
Experimental results show that signal predictions of this second model outperforms non-learning based state-of-the-art techniques and NN models with no architecture search.
arXiv Detail & Related papers (2021-05-17T12:19:22Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.