Transfer Learning using Neural Ordinary Differential Equations
- URL: http://arxiv.org/abs/2001.07342v1
- Date: Tue, 21 Jan 2020 04:59:08 GMT
- Title: Transfer Learning using Neural Ordinary Differential Equations
- Authors: Rajath S, Sumukh Aithal K, Natarajan Subramanyam
- Abstract summary: We use EfficientNets to explore transfer learning on CIFAR-10 dataset.
Using NODE for fine tuning provides more stability during training and validation.
- Score: 0.32228025627337864
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A concept of using Neural Ordinary Differential Equations(NODE) for Transfer
Learning has been introduced. In this paper we use the EfficientNets to explore
transfer learning on CIFAR-10 dataset. We use NODE for fine-tuning our model.
Using NODE for fine tuning provides more stability during training and
validation.These continuous depth blocks can also have a trade off between
numerical precision and speed .Using Neural ODEs for transfer learning has
resulted in much stable convergence of the loss function.
Related papers
- Component Fourier Neural Operator for Singularly Perturbed Differential Equations [3.9482103923304877]
Solving Singularly Perturbed Differential Equations (SPDEs) poses computational challenges arising from the rapid transitions in their solutions within thin regions.
In this manuscript, we introduce Component Fourier Neural Operator (ComFNO), an innovative operator learning method that builds upon Fourier Neural Operator (FNO)
Our approach is not limited to FNO and can be applied to other neural network frameworks, such as Deep Operator Network (DeepONet)
arXiv Detail & Related papers (2024-09-07T09:40:51Z) - Neural Differential Equations for Learning to Program Neural Nets
Through Continuous Learning Rules [10.924226420146626]
We introduce a novel combination of learning rules and Neural ODEs to build continuous-time sequence processing nets.
This yields continuous-time counterparts of Fast Weight Programmers and linear Transformers.
arXiv Detail & Related papers (2022-06-03T15:48:53Z) - Virtual Analog Modeling of Distortion Circuits Using Neural Ordinary
Differential Equations [1.8352113484137629]
Recent research in deep learning has shown that neural networks can learn differential equations governing dynamical systems.
In this paper, we adapt this concept to Virtual Analog (VA) modeling to learn the ordinary differential equations (ODEs) governing the first-order and the second-order diode clipper.
The proposed models achieve performance comparable to state-of-the-art recurrent neural networks (RNNs) albeit using fewer parameters.
arXiv Detail & Related papers (2022-05-04T05:19:46Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Improving Neural ODEs via Knowledge Distillation [35.92851907503015]
We propose a new training based on knowledge distillation to construct more powerful and robust Neural ODEs fitting image recognition tasks.
The experimental results show that the new training manner can improve the classification accuracy of Neural ODEs by 24% on CIFAR10 and 5% on SVHN.
arXiv Detail & Related papers (2022-03-10T01:23:41Z) - Incorporating NODE with Pre-trained Neural Differential Operator for
Learning Dynamics [73.77459272878025]
We propose to enhance the supervised signal in learning dynamics by pre-training a neural differential operator (NDO)
NDO is pre-trained on a class of symbolic functions, and it learns the mapping between the trajectory samples of these functions to their derivatives.
We provide theoretical guarantee on that the output of NDO can well approximate the ground truth derivatives by proper tuning the complexity of the library.
arXiv Detail & Related papers (2021-06-08T08:04:47Z) - Reinforcement Learning with External Knowledge by using Logical Neural
Networks [67.46162586940905]
A recent neuro-symbolic framework called the Logical Neural Networks (LNNs) can simultaneously provide key-properties of both neural networks and symbolic logic.
We propose an integrated method that enables model-free reinforcement learning from external knowledge sources.
arXiv Detail & Related papers (2021-03-03T12:34:59Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Time Dependence in Non-Autonomous Neural ODEs [74.78386661760662]
We propose a novel family of Neural ODEs with time-varying weights.
We outperform previous Neural ODE variants in both speed and representational capacity.
arXiv Detail & Related papers (2020-05-05T01:41:46Z) - Fractional Deep Neural Network via Constrained Optimization [0.0]
This paper introduces a novel algorithmic framework for a deep neural network (DNN)
Fractional-DNN can be viewed as a time-discretization of a fractional in time nonlinear ordinary differential equation (ODE)
arXiv Detail & Related papers (2020-04-01T21:58:21Z) - Stochasticity in Neural ODEs: An Empirical Study [68.8204255655161]
Regularization of neural networks (e.g. dropout) is a widespread technique in deep learning that allows for better generalization.
We show that data augmentation during the training improves the performance of both deterministic and versions of the same model.
However, the improvements obtained by the data augmentation completely eliminate the empirical regularization gains, making the performance of neural ODE and neural SDE negligible.
arXiv Detail & Related papers (2020-02-22T22:12:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.