Neural Differential Equations for Learning to Program Neural Nets
Through Continuous Learning Rules
- URL: http://arxiv.org/abs/2206.01649v1
- Date: Fri, 3 Jun 2022 15:48:53 GMT
- Title: Neural Differential Equations for Learning to Program Neural Nets
Through Continuous Learning Rules
- Authors: Kazuki Irie, Francesco Faccio, J\"urgen Schmidhuber
- Abstract summary: We introduce a novel combination of learning rules and Neural ODEs to build continuous-time sequence processing nets.
This yields continuous-time counterparts of Fast Weight Programmers and linear Transformers.
- Score: 10.924226420146626
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural ordinary differential equations (ODEs) have attracted much attention
as continuous-time counterparts of deep residual neural networks (NNs), and
numerous extensions for recurrent NNs have been proposed. Since the 1980s, ODEs
have also been used to derive theoretical results for NN learning rules, e.g.,
the famous connection between Oja's rule and principal component analysis. Such
rules are typically expressed as additive iterative update processes which have
straightforward ODE counterparts. Here we introduce a novel combination of
learning rules and Neural ODEs to build continuous-time sequence processing
nets that learn to manipulate short-term memory in rapidly changing synaptic
connections of other nets. This yields continuous-time counterparts of Fast
Weight Programmers and linear Transformers. Our novel models outperform the
best existing Neural Controlled Differential Equation based models on various
time series classification tasks, while also addressing their scalability
limitations. Our code is public.
Related papers
- How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Learning Neural Constitutive Laws From Motion Observations for
Generalizable PDE Dynamics [97.38308257547186]
Many NN approaches learn an end-to-end model that implicitly models both the governing PDE and material models.
We argue that the governing PDEs are often well-known and should be explicitly enforced rather than learned.
We introduce a new framework termed "Neural Constitutive Laws" (NCLaw) which utilizes a network architecture that strictly guarantees standard priors.
arXiv Detail & Related papers (2023-04-27T17:42:24Z) - On the balance between the training time and interpretability of neural
ODE for time series modelling [77.34726150561087]
The paper shows that modern neural ODE cannot be reduced to simpler models for time-series modelling applications.
The complexity of neural ODE is compared to or exceeds the conventional time-series modelling tools.
We propose a new view on time-series modelling using combined neural networks and an ODE system approach.
arXiv Detail & Related papers (2022-06-07T13:49:40Z) - Neural Piecewise-Constant Delay Differential Equations [17.55759866368141]
In this article, we introduce a new sort of continuous-depth neural network, called the Neural Piecewise-Constant Delay Differential Equations (PCDDEs)
We show that the Neural PCDDEs do outperform the several existing continuous-depth neural frameworks on the one-dimensional piecewise-constant delay population dynamics and real-world datasets.
arXiv Detail & Related papers (2022-01-04T03:44:15Z) - Reinforcement Learning with External Knowledge by using Logical Neural
Networks [67.46162586940905]
A recent neuro-symbolic framework called the Logical Neural Networks (LNNs) can simultaneously provide key-properties of both neural networks and symbolic logic.
We propose an integrated method that enables model-free reinforcement learning from external knowledge sources.
arXiv Detail & Related papers (2021-03-03T12:34:59Z) - Hypersolvers: Toward Fast Continuous-Depth Models [16.43439140464003]
We introduce hypersolvers, neural networks designed to solve ODEs with low overhead and theoretical guarantees on accuracy.
The synergistic combination of hypersolvers and Neural ODEs allows for cheap inference and unlocks a new frontier for practical application of continuous-depth models.
arXiv Detail & Related papers (2020-07-19T06:31:31Z) - Liquid Time-constant Networks [117.57116214802504]
We introduce a new class of time-continuous recurrent neural network models.
Instead of declaring a learning system's dynamics by implicit nonlinearities, we construct networks of linear first-order dynamical systems.
These neural networks exhibit stable and bounded behavior, yield superior expressivity within the family of neural ordinary differential equations.
arXiv Detail & Related papers (2020-06-08T09:53:35Z) - Neural Ordinary Differential Equation based Recurrent Neural Network
Model [0.7233897166339269]
differential equations are a promising new member in the neural network family.
This paper explores the strength of the ordinary differential equation (ODE) is explored with a new extension.
Two new ODE-based RNN models (GRU-ODE model and LSTM-ODE) can compute the hidden state and cell state at any point of time using an ODE solver.
Experiments show that these new ODE based RNN models require less training time than Latent ODEs and conventional Neural ODEs.
arXiv Detail & Related papers (2020-05-20T01:02:29Z) - Time Dependence in Non-Autonomous Neural ODEs [74.78386661760662]
We propose a novel family of Neural ODEs with time-varying weights.
We outperform previous Neural ODE variants in both speed and representational capacity.
arXiv Detail & Related papers (2020-05-05T01:41:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.