Nonlinear system identification with regularized Tensor Network
B-splines
- URL: http://arxiv.org/abs/2003.07594v1
- Date: Tue, 17 Mar 2020 09:22:20 GMT
- Title: Nonlinear system identification with regularized Tensor Network
B-splines
- Authors: Ridvan Karagoz, Kim Batselier
- Abstract summary: The TNBS-NARX model is validated through the identification of the cascaded watertank benchmark nonlinear system.
It achieves state-of-the-art performance while identifying a 16-dimensional B-spline surface in 4 seconds on a standard desktop computer.
- Score: 2.817412580574242
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This article introduces the Tensor Network B-spline model for the regularized
identification of nonlinear systems using a nonlinear autoregressive exogenous
(NARX) approach. Tensor network theory is used to alleviate the curse of
dimensionality of multivariate B-splines by representing the high-dimensional
weight tensor as a low-rank approximation. An iterative algorithm based on the
alternating linear scheme is developed to directly estimate the low-rank tensor
network approximation, removing the need to ever explicitly construct the
exponentially large weight tensor. This reduces the computational and storage
complexity significantly, allowing the identification of NARX systems with a
large number of inputs and lags. The proposed algorithm is numerically stable,
robust to noise, guaranteed to monotonically converge, and allows the
straightforward incorporation of regularization. The TNBS-NARX model is
validated through the identification of the cascaded watertank benchmark
nonlinear system, on which it achieves state-of-the-art performance while
identifying a 16-dimensional B-spline surface in 4 seconds on a standard
desktop computer. An open-source MATLAB implementation is available on GitHub.
Related papers
- Matrix Completion via Nonsmooth Regularization of Fully Connected Neural Networks [7.349727826230864]
It has been shown that enhanced performance could be attained by using nonlinear estimators such as deep neural networks.
In this paper, we control over-fitting by regularizing FCNN model in terms of norm intermediate representations.
Our simulations indicate the superiority of the proposed algorithm in comparison with existing linear and nonlinear algorithms.
arXiv Detail & Related papers (2024-03-15T12:00:37Z) - An L-BFGS-B approach for linear and nonlinear system identification under $\ell_1$- and group-Lasso regularization [0.0]
We propose a very efficient numerical method for identifying linear and nonlinear discrete-time state-space models.
A Python implementation of the proposed identification method is available in the package jax-sysid.
arXiv Detail & Related papers (2024-03-06T16:17:34Z) - Optimization Guarantees of Unfolded ISTA and ADMM Networks With Smooth
Soft-Thresholding [57.71603937699949]
We study optimization guarantees, i.e., achieving near-zero training loss with the increase in the number of learning epochs.
We show that the threshold on the number of training samples increases with the increase in the network width.
arXiv Detail & Related papers (2023-09-12T13:03:47Z) - Low-rank extended Kalman filtering for online learning of neural
networks from streaming data [71.97861600347959]
We propose an efficient online approximate Bayesian inference algorithm for estimating the parameters of a nonlinear function from a potentially non-stationary data stream.
The method is based on the extended Kalman filter (EKF), but uses a novel low-rank plus diagonal decomposition of the posterior matrix.
In contrast to methods based on variational inference, our method is fully deterministic, and does not require step-size tuning.
arXiv Detail & Related papers (2023-05-31T03:48:49Z) - Online Weak-form Sparse Identification of Partial Differential Equations [0.5156484100374058]
This paper presents an online algorithm for identification of partial differential equations (PDEs) based on the weak-form sparse identification of nonlinear dynamics algorithm (WSINDy)
The core of the method combines a weak-form discretization of candidate PDEs with an online proximal gradient descent approach to the sparse regression problem.
arXiv Detail & Related papers (2022-03-08T10:11:09Z) - Training Recurrent Neural Networks by Sequential Least Squares and the
Alternating Direction Method of Multipliers [0.20305676256390928]
We propose the use of convex and twice-differentiable loss and regularization terms for determining optimal hidden network parameters.
We combine sequential least squares with alternating direction multipliers.
The performance of the algorithm is tested in a nonlinear system identification benchmark.
arXiv Detail & Related papers (2021-12-31T08:43:04Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Unfolding Projection-free SDP Relaxation of Binary Graph Classifier via
GDPA Linearization [59.87663954467815]
Algorithm unfolding creates an interpretable and parsimonious neural network architecture by implementing each iteration of a model-based algorithm as a neural layer.
In this paper, leveraging a recent linear algebraic theorem called Gershgorin disc perfect alignment (GDPA), we unroll a projection-free algorithm for semi-definite programming relaxation (SDR) of a binary graph.
Experimental results show that our unrolled network outperformed pure model-based graph classifiers, and achieved comparable performance to pure data-driven networks but using far fewer parameters.
arXiv Detail & Related papers (2021-09-10T07:01:15Z) - A novel Deep Neural Network architecture for non-linear system
identification [78.69776924618505]
We present a novel Deep Neural Network (DNN) architecture for non-linear system identification.
Inspired by fading memory systems, we introduce inductive bias (on the architecture) and regularization (on the loss function)
This architecture allows for automatic complexity selection based solely on available data.
arXiv Detail & Related papers (2021-06-06T10:06:07Z) - Identification of Probability weighted ARX models with arbitrary domains [75.91002178647165]
PieceWise Affine models guarantees universal approximation, local linearity and equivalence to other classes of hybrid system.
In this work, we focus on the identification of PieceWise Auto Regressive with eXogenous input models with arbitrary regions (NPWARX)
The architecture is conceived following the Mixture of Expert concept, developed within the machine learning field.
arXiv Detail & Related papers (2020-09-29T12:50:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.