Initialization Approach for Nonlinear State-Space Identification via the
Subspace Encoder Approach
- URL: http://arxiv.org/abs/2304.02119v2
- Date: Thu, 6 Apr 2023 21:57:48 GMT
- Title: Initialization Approach for Nonlinear State-Space Identification via the
Subspace Encoder Approach
- Authors: Rishi Ramkannan, Gerben I. Beintema, Roland T\'oth, Maarten Schoukens
- Abstract summary: SUBNET has been developed to identify nonlinear state-space models from input-output data.
State encoder function is introduced to reconstruct the current state from past input-output data.
This paper focuses on an initialisation of the subspace encoder approach using the Best Linear Approximation (BLA)
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The SUBNET neural network architecture has been developed to identify
nonlinear state-space models from input-output data. To achieve this, it
combines the rolled-out nonlinear state-space equations and a state encoder
function, both parameterised as neural networks The encoder function is
introduced to reconstruct the current state from past input-output data. Hence,
it enables the forward simulation of the rolled-out state-space model. While
this approach has shown to provide high-accuracy and consistent model
estimation, its convergence can be significantly improved by efficient
initialization of the training process. This paper focuses on such an
initialisation of the subspace encoder approach using the Best Linear
Approximation (BLA). Using the BLA provided state-space matrices and its
associated reconstructability map, both the state-transition part of the
network and the encoder are initialized. The performance of the improved
initialisation scheme is evaluated on a Wiener-Hammerstein simulation example
and a benchmark dataset. The results show that for a weakly nonlinear system,
the proposed initialisation based on the linear reconstructability map results
in a faster convergence and a better model quality.
Related papers
- The Convex Landscape of Neural Networks: Characterizing Global Optima
and Stationary Points via Lasso Models [75.33431791218302]
Deep Neural Network Network (DNN) models are used for programming purposes.
In this paper we examine the use of convex neural recovery models.
We show that all the stationary non-dimensional objective objective can be characterized as the standard a global subsampled convex solvers program.
We also show that all the stationary non-dimensional objective objective can be characterized as the standard a global subsampled convex solvers program.
arXiv Detail & Related papers (2023-12-19T23:04:56Z) - Predicting the Initial Conditions of the Universe using a Deterministic
Neural Network [10.158552381785078]
Finding the initial conditions that led to the current state of the universe is challenging because it involves searching over an intractable input space of initial conditions.
Deep learning has emerged as a surrogate for N-body simulations by directly learning the mapping between the linear input of an N-body simulation and the final nonlinear output from the simulation.
In this work, we pioneer the use of a deterministic convolutional neural network for learning the reverse mapping and show that it accurately recovers the initial linear displacement field over a wide range of scales.
arXiv Detail & Related papers (2023-03-23T06:04:36Z) - Learning Large-scale Neural Fields via Context Pruned Meta-Learning [60.93679437452872]
We introduce an efficient optimization-based meta-learning technique for large-scale neural field training.
We show how gradient re-scaling at meta-test time allows the learning of extremely high-quality neural fields.
Our framework is model-agnostic, intuitive, straightforward to implement, and shows significant reconstruction improvements for a wide range of signals.
arXiv Detail & Related papers (2023-02-01T17:32:16Z) - Learning Reduced Nonlinear State-Space Models: an Output-Error Based
Canonical Approach [8.029702645528412]
We investigate the effectiveness of deep learning in the modeling of dynamic systems with nonlinear behavior.
We show its ability to identify three different nonlinear systems.
The performances are evaluated in terms of open-loop prediction on test data generated in simulation as well as a real world data-set of unmanned aerial vehicle flight measurements.
arXiv Detail & Related papers (2022-04-19T06:33:23Z) - Deep Equilibrium Assisted Block Sparse Coding of Inter-dependent
Signals: Application to Hyperspectral Imaging [71.57324258813675]
A dataset of inter-dependent signals is defined as a matrix whose columns demonstrate strong dependencies.
A neural network is employed to act as structure prior and reveal the underlying signal interdependencies.
Deep unrolling and Deep equilibrium based algorithms are developed, forming highly interpretable and concise deep-learning-based architectures.
arXiv Detail & Related papers (2022-03-29T21:00:39Z) - Integrating Recurrent Neural Networks with Data Assimilation for
Scalable Data-Driven State Estimation [0.0]
Data assimilation (DA) is integrated with machine learning to perform entirely data-driven online state estimation.
recurrent neural networks (RNNs) are implemented as surrogate models to replace key components of the DA cycle in numerical weather prediction (NWP)
It is shown how these RNNs can be using DA methods to directly update the hidden/reservoir state with observations of the target system.
arXiv Detail & Related papers (2021-09-25T03:56:53Z) - Data-driven Weight Initialization with Sylvester Solvers [72.11163104763071]
We propose a data-driven scheme to initialize the parameters of a deep neural network.
We show that our proposed method is especially effective in few-shot and fine-tuning settings.
arXiv Detail & Related papers (2021-05-02T07:33:16Z) - Improved Initialization of State-Space Artificial Neural Networks [0.0]
The identification of black-box nonlinear state-space models requires a flexible representation of the state and output equation.
This paper introduces an improved approach for nonlinear state-space models represented as a recurrent artificial neural network.
arXiv Detail & Related papers (2021-03-26T15:16:08Z) - Nonlinear state-space identification using deep encoder networks [0.0]
This paper introduces a method that approximates the simulation loss by splitting the data set into independent sections to the multiple shooting method.
The main contribution is the introduction of an encoder function to estimate the initial state at the start each section.
arXiv Detail & Related papers (2020-12-14T16:49:06Z) - MetaSDF: Meta-learning Signed Distance Functions [85.81290552559817]
Generalizing across shapes with neural implicit representations amounts to learning priors over the respective function space.
We formalize learning of a shape space as a meta-learning problem and leverage gradient-based meta-learning algorithms to solve this task.
arXiv Detail & Related papers (2020-06-17T05:14:53Z) - MSE-Optimal Neural Network Initialization via Layer Fusion [68.72356718879428]
Deep neural networks achieve state-of-the-art performance for a range of classification and inference tasks.
The use of gradient combined nonvolutionity renders learning susceptible to novel problems.
We propose fusing neighboring layers of deeper networks that are trained with random variables.
arXiv Detail & Related papers (2020-01-28T18:25:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.