Meta Learning Backpropagation And Improving It
- URL: http://arxiv.org/abs/2012.14905v2
- Date: Tue, 16 Feb 2021 17:28:31 GMT
- Title: Meta Learning Backpropagation And Improving It
- Authors: Louis Kirsch and J\"urgen Schmidhuber
- Abstract summary: We show that simple weight-sharing and sparsity in an NN is sufficient to express powerful learning algorithms (LAs) in a reusable fashion.
A simple implementation of VS-ML called VS-ML RNN allows for implementing the backpropagation LA solely by running an RNN in forward-mode.
It can even meta-learn new LAs that improve upon backpropagation and generalize to datasets outside of the meta training distribution.
- Score: 4.061135251278187
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many concepts have been proposed for meta learning with neural networks
(NNs), e.g., NNs that learn to control fast weights, hyper networks, learned
learning rules, and meta recurrent NNs. Our Variable Shared Meta Learning
(VS-ML) unifies the above and demonstrates that simple weight-sharing and
sparsity in an NN is sufficient to express powerful learning algorithms (LAs)
in a reusable fashion. A simple implementation of VS-ML called VS-ML RNN allows
for implementing the backpropagation LA solely by running an RNN in
forward-mode. It can even meta-learn new LAs that improve upon backpropagation
and generalize to datasets outside of the meta training distribution without
explicit gradient calculation. Introspection reveals that our meta-learned LAs
learn qualitatively different from gradient descent through fast association.
Related papers
- ConML: A Universal Meta-Learning Framework with Task-Level Contrastive Learning [49.447777286862994]
ConML is a universal meta-learning framework that can be applied to various meta-learning algorithms.
We demonstrate that ConML integrates seamlessly with optimization-based, metric-based, and amortization-based meta-learning algorithms.
arXiv Detail & Related papers (2024-10-08T12:22:10Z) - Meta-ticket: Finding optimal subnetworks for few-shot learning within
randomly initialized neural networks [16.146036509065247]
Few-shot learning for neural networks (NNs) is an important problem that aims to train NNs with a few data.
We propose a novel meta-learning approach, called Meta-ticket, to find optimal sparseworks for few-shot learning within randomly NNs.
arXiv Detail & Related papers (2022-05-31T09:03:57Z) - Meta-learning Spiking Neural Networks with Surrogate Gradient Descent [1.90365714903665]
Bi-level learning, such as meta-learning, is increasingly used in deep learning to overcome limitations.
We show that SNNs meta-trained using MAML match or exceed the performance of conventional ANNs meta-trained with MAML on event-based meta-datasets.
Our results emphasize how meta-learning techniques can become instrumental for deploying neuromorphic learning technologies on real-world problems.
arXiv Detail & Related papers (2022-01-26T06:53:46Z) - Low-bit Quantization of Recurrent Neural Network Language Models Using
Alternating Direction Methods of Multipliers [67.688697838109]
This paper presents a novel method to train quantized RNNLMs from scratch using alternating direction methods of multipliers (ADMM)
Experiments on two tasks suggest the proposed ADMM quantization achieved a model size compression factor of up to 31 times over the full precision baseline RNNLMs.
arXiv Detail & Related papers (2021-11-29T09:30:06Z) - Can we learn gradients by Hamiltonian Neural Networks? [68.8204255655161]
We propose a meta-learner based on ODE neural networks that learns gradients.
We demonstrate that our method outperforms a meta-learner based on LSTM for an artificial task and the MNIST dataset with ReLU activations in the optimizee.
arXiv Detail & Related papers (2021-10-31T18:35:10Z) - Reinforcement Learning with External Knowledge by using Logical Neural
Networks [67.46162586940905]
A recent neuro-symbolic framework called the Logical Neural Networks (LNNs) can simultaneously provide key-properties of both neural networks and symbolic logic.
We propose an integrated method that enables model-free reinforcement learning from external knowledge sources.
arXiv Detail & Related papers (2021-03-03T12:34:59Z) - Fast On-Device Adaptation for Spiking Neural Networks via
Online-Within-Online Meta-Learning [31.78005607111787]
Spiking Neural Networks (SNNs) have recently gained popularity as machine learning models for on-device edge intelligence.
In this paper, we propose an online-within-online meta-learning rule for SNNs termed OWOML-SNN, that enables lifelong learning on a stream of tasks.
arXiv Detail & Related papers (2021-02-21T04:28:49Z) - Meta-Learning with Neural Tangent Kernels [58.06951624702086]
We propose the first meta-learning paradigm in the Reproducing Kernel Hilbert Space (RKHS) induced by the meta-model's Neural Tangent Kernel (NTK)
Within this paradigm, we introduce two meta-learning algorithms, which no longer need a sub-optimal iterative inner-loop adaptation as in the MAML framework.
We achieve this goal by 1) replacing the adaptation with a fast-adaptive regularizer in the RKHS; and 2) solving the adaptation analytically based on the NTK theory.
arXiv Detail & Related papers (2021-02-07T20:53:23Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Is the Meta-Learning Idea Able to Improve the Generalization of Deep
Neural Networks on the Standard Supervised Learning? [34.00378876525579]
We propose a novel metalearning based training procedure (M) for deep neural networks (DNNs)
M simulates the meta-training process by considering a batch of training samples as a task.
The experimental results show a consistently improved performance on all the generalizations with different sizes.
arXiv Detail & Related papers (2020-02-27T21:29:54Z) - Deep Transform and Metric Learning Network: Wedding Deep Dictionary
Learning and Neural Networks [34.49034775978504]
We propose a novel DDL approach where each DL layer can be formulated as a combination of one linear layer and a Recurrent Neural Network (RNN)
Our proposed work unveils new insights into Neural Networks and DDL and provides a new, efficient and competitive approach to jointly learn a deep transform and a metric for inference applications.
arXiv Detail & Related papers (2020-02-18T22:04:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.