Towards Understanding Normalization in Neural ODEs
- URL: http://arxiv.org/abs/2004.09222v2
- Date: Mon, 27 Apr 2020 19:43:36 GMT
- Title: Towards Understanding Normalization in Neural ODEs
- Authors: Julia Gusak, Larisa Markeeva, Talgat Daulbaev, Alexandr Katrutsa,
Andrzej Cichocki, Ivan Oseledets
- Abstract summary: We show that it is possible to achieve 93% accuracy in the CIFAR-10 classification task.
This is the highest reported accuracy among neural ODEs tested on this problem.
- Score: 71.26657499537366
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Normalization is an important and vastly investigated technique in deep
learning. However, its role for Ordinary Differential Equation based networks
(neural ODEs) is still poorly understood. This paper investigates how different
normalization techniques affect the performance of neural ODEs. Particularly,
we show that it is possible to achieve 93% accuracy in the CIFAR-10
classification task, and to the best of our knowledge, this is the highest
reported accuracy among neural ODEs tested on this problem.
Related papers
- Training Stiff Neural Ordinary Differential Equations with Implicit Single-Step Methods [3.941173292703699]
Stiff systems of ordinary differential equations (ODEs) are pervasive in many science and engineering fields.
Standard neural ODE approaches struggle to learn them.
This paper proposes an approach based on single-step implicit schemes to enable neural ODEs to handle stiffness.
arXiv Detail & Related papers (2024-10-08T01:08:17Z) - Adaptive Feedforward Gradient Estimation in Neural ODEs [0.0]
We propose a novel approach that leverages adaptive feedforward gradient estimation to improve the efficiency, consistency, and interpretability of Neural ODEs.
Our method eliminates the need for backpropagation and the adjoint method, reducing computational overhead and memory usage while maintaining accuracy.
arXiv Detail & Related papers (2024-09-22T18:21:01Z) - ODE Discovery for Longitudinal Heterogeneous Treatment Effects Inference [69.24516189971929]
In this paper, we introduce a new type of solution in the longitudinal setting: a closed-form ordinary differential equation (ODE)
While we still rely on continuous optimization to learn an ODE, the resulting inference machine is no longer a neural network.
arXiv Detail & Related papers (2024-03-16T02:07:45Z) - Improving Neural ODEs via Knowledge Distillation [35.92851907503015]
We propose a new training based on knowledge distillation to construct more powerful and robust Neural ODEs fitting image recognition tasks.
The experimental results show that the new training manner can improve the classification accuracy of Neural ODEs by 24% on CIFAR10 and 5% on SVHN.
arXiv Detail & Related papers (2022-03-10T01:23:41Z) - Dynamic Neural Diversification: Path to Computationally Sustainable
Neural Networks [68.8204255655161]
Small neural networks with a constrained number of trainable parameters, can be suitable resource-efficient candidates for many simple tasks.
We explore the diversity of the neurons within the hidden layer during the learning process.
We analyze how the diversity of the neurons affects predictions of the model.
arXiv Detail & Related papers (2021-09-20T15:12:16Z) - Meta-Solver for Neural Ordinary Differential Equations [77.8918415523446]
We investigate how the variability in solvers' space can improve neural ODEs performance.
We show that the right choice of solver parameterization can significantly affect neural ODEs models in terms of robustness to adversarial attacks.
arXiv Detail & Related papers (2021-03-15T17:26:34Z) - Optimization Theory for ReLU Neural Networks Trained with Normalization
Layers [82.61117235807606]
The success of deep neural networks in part due to the use of normalization layers.
Our analysis shows how the introduction of normalization changes the landscape and can enable faster activation.
arXiv Detail & Related papers (2020-06-11T23:55:54Z) - ODEN: A Framework to Solve Ordinary Differential Equations using
Artificial Neural Networks [0.0]
We prove a specific loss function, which does not require knowledge of the exact solution, to evaluate neural networks' performance.
Neural networks are shown to be proficient at approximating continuous solutions within their training domains.
A user-friendly and adaptable open-source code (ODE$mathcalN$) is provided on GitHub.
arXiv Detail & Related papers (2020-05-28T15:34:10Z) - Stochasticity in Neural ODEs: An Empirical Study [68.8204255655161]
Regularization of neural networks (e.g. dropout) is a widespread technique in deep learning that allows for better generalization.
We show that data augmentation during the training improves the performance of both deterministic and versions of the same model.
However, the improvements obtained by the data augmentation completely eliminate the empirical regularization gains, making the performance of neural ODE and neural SDE negligible.
arXiv Detail & Related papers (2020-02-22T22:12:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.