Continual Weight Updates and Convolutional Architectures for Equilibrium
Propagation
- URL: http://arxiv.org/abs/2005.04169v1
- Date: Wed, 29 Apr 2020 12:14:06 GMT
- Title: Continual Weight Updates and Convolutional Architectures for Equilibrium
Propagation
- Authors: Maxence Ernoult, Julie Grollier, Damien Querlioz, Yoshua Bengio,
Benjamin Scellier
- Abstract summary: Equilibrium Propagation (EP) is a biologically inspired alternative algorithm to backpropagation (BP) for training neural networks.
We propose a discrete-time formulation of EP which enables to simplify equations, speed up training and extend EP to CNNs.
Our CNN model achieves the best performance ever reported on MNIST with EP.
- Score: 69.87491240509485
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Equilibrium Propagation (EP) is a biologically inspired alternative algorithm
to backpropagation (BP) for training neural networks. It applies to RNNs fed by
a static input x that settle to a steady state, such as Hopfield networks. EP
is similar to BP in that in the second phase of training, an error signal
propagates backwards in the layers of the network, but contrary to BP, the
learning rule of EP is spatially local. Nonetheless, EP suffers from two major
limitations. On the one hand, due to its formulation in terms of real-time
dynamics, EP entails long simulation times, which limits its applicability to
practical tasks. On the other hand, the biological plausibility of EP is
limited by the fact that its learning rule is not local in time: the synapse
update is performed after the dynamics of the second phase have converged and
requires information of the first phase that is no longer available physically.
Our work addresses these two issues and aims at widening the spectrum of EP
from standard machine learning models to more bio-realistic neural networks.
First, we propose a discrete-time formulation of EP which enables to simplify
equations, speed up training and extend EP to CNNs. Our CNN model achieves the
best performance ever reported on MNIST with EP. Using the same discrete-time
formulation, we introduce Continual Equilibrium Propagation (C-EP): the weights
of the network are adjusted continually in the second phase of training using
local information in space and time. We show that in the limit of slow changes
of synaptic strengths and small nudging, C-EP is equivalent to BPTT (Theorem
1). We numerically demonstrate Theorem 1 and C-EP training on MNIST and
generalize it to the bio-realistic situation of a neural network with
asymmetric connections between neurons.
Related papers
- Scaling SNNs Trained Using Equilibrium Propagation to Convolutional Architectures [2.2146860305758485]
Equilibrium Propagation (EP) is a biologically plausible local learning algorithm initially developed for convergent recurrent neural networks (RNNs)
EP is a powerful candidate for training Spiking Neural Networks (SNNs), which are commonly trained by BPTT.
We provide a formulation for training convolutional spiking convergent RNNs using EP, bridging the gap between spiking and non-spiking convergent RNNs.
arXiv Detail & Related papers (2024-05-04T03:06:14Z) - ETLP: Event-based Three-factor Local Plasticity for online learning with
neuromorphic hardware [105.54048699217668]
We show a competitive performance in accuracy with a clear advantage in the computational complexity for Event-Based Three-factor Local Plasticity (ETLP)
We also show that when using local plasticity, threshold adaptation in spiking neurons and a recurrent topology are necessary to learntemporal patterns with a rich temporal structure.
arXiv Detail & Related papers (2023-01-19T19:45:42Z) - Sequence Learning Using Equilibrium Propagation [2.3361887733755897]
Equilibrium Propagation (EP) is a powerful and more bio-plausible alternative to conventional learning frameworks such as backpropagation.
We leverage recent developments in modern hopfield networks to further understand energy based models and develop solutions for complex sequence classification tasks using EP.
arXiv Detail & Related papers (2022-09-14T20:01:22Z) - Holomorphic Equilibrium Propagation Computes Exact Gradients Through
Finite Size Oscillations [5.279475826661643]
Equilibrium propagation (EP) is an alternative to backpropagation (BP) that allows the training of deep neural networks with local learning rules.
We show analytically that this extension naturally leads to exact gradients even for finite-amplitude teaching signals.
We establish the first benchmark for EP on the ImageNet 32x32 dataset and show that it matches the performance of an equivalent network trained with BP.
arXiv Detail & Related papers (2022-09-01T15:23:49Z) - A Theoretical Framework for Inference and Learning in Predictive Coding
Networks [41.58529335439799]
Predictive coding (PC) is an influential theory in computational neuroscience.
We provide a comprehensive theoretical analysis of the properties of PCNs trained with prospective configuration.
arXiv Detail & Related papers (2022-07-21T04:17:55Z) - On the Convergence of Certified Robust Training with Interval Bound
Propagation [147.77638840942447]
We present a theoretical analysis on the convergence of IBP training.
We show that when using IBP training to train a randomly two-layer ReLU neural network with logistic loss, gradient descent can linearly converge to zero robust training error.
arXiv Detail & Related papers (2022-03-16T21:49:13Z) - Scaling Equilibrium Propagation to Deep ConvNets by Drastically Reducing
its Gradient Estimator Bias [62.43908463620527]
In practice, EP does not scale to visual tasks harder than MNIST.
We show that a bias in the gradient estimate of EP, inherent in the use of finite nudging, is responsible for this phenomenon.
These results highlight EP as a scalable approach to compute error gradients in deep neural networks, thereby motivating its hardware implementation.
arXiv Detail & Related papers (2021-01-14T10:23:40Z) - Scaling Equilibrium Propagation to Deep ConvNets by Drastically Reducing
its Gradient Estimator Bias [65.13042449121411]
In practice, training a network with the gradient estimates provided by EP does not scale to visual tasks harder than MNIST.
We show that a bias in the gradient estimate of EP, inherent in the use of finite nudging, is responsible for this phenomenon.
We apply these techniques to train an architecture with asymmetric forward and backward connections, yielding a 13.2% test error.
arXiv Detail & Related papers (2020-06-06T09:36:07Z) - Equilibrium Propagation with Continual Weight Updates [69.87491240509485]
We propose a learning algorithm that bridges Machine Learning and Neuroscience, by computing gradients closely matching those of Backpropagation Through Time (BPTT)
We prove theoretically that, provided the learning rates are sufficiently small, at each time step of the second phase the dynamics of neurons and synapses follow the gradients of the loss given by BPTT.
These results bring EP a step closer to biology by better complying with hardware constraints while maintaining its intimate link with backpropagation.
arXiv Detail & Related papers (2020-04-29T14:54:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.