Unconstrained Parametrization of Dissipative and Contracting Neural
Ordinary Differential Equations
- URL: http://arxiv.org/abs/2304.02976v2
- Date: Fri, 15 Sep 2023 15:28:30 GMT
- Title: Unconstrained Parametrization of Dissipative and Contracting Neural
Ordinary Differential Equations
- Authors: Daniele Martinelli, Clara Luc\'ia Galimberti, Ian R. Manchester, Luca
Furieri, and Giancarlo Ferrari-Trecate
- Abstract summary: We introduce and study a class of Deep Neural Networks (DNNs) in continuous-time.
We show how to endow our proposed NodeRENs with contractivity and dissipativity -- crucial properties for robust learning and control.
- Score: 0.9437165725355698
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this work, we introduce and study a class of Deep Neural Networks (DNNs)
in continuous-time. The proposed architecture stems from the combination of
Neural Ordinary Differential Equations (Neural ODEs) with the model structure
of recently introduced Recurrent Equilibrium Networks (RENs). We show how to
endow our proposed NodeRENs with contractivity and dissipativity -- crucial
properties for robust learning and control. Most importantly, as for RENs, we
derive parametrizations of contractive and dissipative NodeRENs which are
unconstrained, hence enabling their learning for a large number of parameters.
We validate the properties of NodeRENs, including the possibility of handling
irregularly sampled data, in a case study in nonlinear system identification.
Related papers
- Neural Parameter Regression for Explicit Representations of PDE Solution Operators [22.355460388065964]
We introduce Neural Regression (NPR), a novel framework specifically developed for learning solution operators in Partial Differential Equations (PDEs)
NPR employs Physics-Informed Neural Network (PINN, Raissi et al., 2021) techniques to regress Neural Network (NN) parameters.
The framework shows remarkable adaptability to new initial and boundary conditions, allowing for rapid fine-tuning and inference.
arXiv Detail & Related papers (2024-03-19T14:30:56Z) - Regularizing with Pseudo-Negatives for Continual Self-Supervised Learning [62.40718385934608]
We introduce a novel Pseudo-Negative Regularization (PNR) framework for effective continual self-supervised learning (CSSL)
Our PNR leverages pseudo-negatives obtained through model-based augmentation in a way that newly learned representations may not contradict what has been learned in the past.
arXiv Detail & Related papers (2023-06-08T10:59:35Z) - Benign Overfitting in Deep Neural Networks under Lazy Training [72.28294823115502]
We show that when the data distribution is well-separated, DNNs can achieve Bayes-optimal test error for classification.
Our results indicate that interpolating with smoother functions leads to better generalization.
arXiv Detail & Related papers (2023-05-30T19:37:44Z) - Neural Delay Differential Equations: System Reconstruction and Image
Classification [14.59919398960571]
We propose a new class of continuous-depth neural networks with delay, named Neural Delay Differential Equations (NDDEs)
Compared to NODEs, NDDEs have a stronger capacity of nonlinear representations.
We achieve lower loss and higher accuracy not only for the data produced synthetically but also for the CIFAR10, a well-known image dataset.
arXiv Detail & Related papers (2023-04-11T16:09:28Z) - Learning Discretized Neural Networks under Ricci Flow [51.36292559262042]
We study Discretized Neural Networks (DNNs) composed of low-precision weights and activations.
DNNs suffer from either infinite or zero gradients due to the non-differentiable discrete function during training.
arXiv Detail & Related papers (2023-02-07T10:51:53Z) - Learning Low Dimensional State Spaces with Overparameterized Recurrent
Neural Nets [57.06026574261203]
We provide theoretical evidence for learning low-dimensional state spaces, which can also model long-term memory.
Experiments corroborate our theory, demonstrating extrapolation via learning low-dimensional state spaces with both linear and non-linear RNNs.
arXiv Detail & Related papers (2022-10-25T14:45:15Z) - Tunable Complexity Benchmarks for Evaluating Physics-Informed Neural
Networks on Coupled Ordinary Differential Equations [64.78260098263489]
In this work, we assess the ability of physics-informed neural networks (PINNs) to solve increasingly-complex coupled ordinary differential equations (ODEs)
We show that PINNs eventually fail to produce correct solutions to these benchmarks as their complexity increases.
We identify several reasons why this may be the case, including insufficient network capacity, poor conditioning of the ODEs, and high local curvature, as measured by the Laplacian of the PINN loss.
arXiv Detail & Related papers (2022-10-14T15:01:32Z) - Neural Ordinary Differential Equations for Nonlinear System
Identification [0.9864260997723973]
We present a study comparing NODE's performance against neural state-space models and classical linear system identification methods.
Experiments show that NODEs can consistently improve the prediction accuracy by an order of magnitude compared to benchmark methods.
arXiv Detail & Related papers (2022-02-28T22:25:53Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.