Towards Scaling Difference Target Propagation by Learning Backprop
Targets
- URL: http://arxiv.org/abs/2201.13415v1
- Date: Mon, 31 Jan 2022 18:20:43 GMT
- Title: Towards Scaling Difference Target Propagation by Learning Backprop
Targets
- Authors: Maxence Ernoult, Fabrice Normandin, Abhinav Moudgil, Sean Spinney,
Eugene Belilovsky, Irina Rish, Blake Richards, Yoshua Bengio
- Abstract summary: Difference Target Propagation is a biologically-plausible learning algorithm with close relation with Gauss-Newton (GN) optimization.
We propose a novel feedback weight training scheme that ensures both that DTP approximates BP and that layer-wise feedback weight training can be restored.
We report the best performance ever achieved by DTP on CIFAR-10 and ImageNet.
- Score: 64.90165892557776
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The development of biologically-plausible learning algorithms is important
for understanding learning in the brain, but most of them fail to scale-up to
real-world tasks, limiting their potential as explanations for learning by real
brains. As such, it is important to explore learning algorithms that come with
strong theoretical guarantees and can match the performance of backpropagation
(BP) on complex tasks. One such algorithm is Difference Target Propagation
(DTP), a biologically-plausible learning algorithm whose close relation with
Gauss-Newton (GN) optimization has been recently established. However, the
conditions under which this connection rigorously holds preclude layer-wise
training of the feedback pathway synaptic weights (which is more biologically
plausible). Moreover, good alignment between DTP weight updates and loss
gradients is only loosely guaranteed and under very specific conditions for the
architecture being trained. In this paper, we propose a novel feedback weight
training scheme that ensures both that DTP approximates BP and that layer-wise
feedback weight training can be restored without sacrificing any theoretical
guarantees. Our theory is corroborated by experimental results and we report
the best performance ever achieved by DTP on CIFAR-10 and ImageNet 32$\times$32
Related papers
- Layer-wise Feedback Propagation [53.00944147633484]
We present Layer-wise Feedback Propagation (LFP), a novel training approach for neural-network-like predictors.
LFP assigns rewards to individual connections based on their respective contributions to solving a given task.
We demonstrate its effectiveness in achieving comparable performance to gradient descent on various models and datasets.
arXiv Detail & Related papers (2023-08-23T10:48:28Z) - Provable Reward-Agnostic Preference-Based Reinforcement Learning [61.39541986848391]
Preference-based Reinforcement Learning (PbRL) is a paradigm in which an RL agent learns to optimize a task using pair-wise preference-based feedback over trajectories.
We propose a theoretical reward-agnostic PbRL framework where exploratory trajectories that enable accurate learning of hidden reward functions are acquired.
arXiv Detail & Related papers (2023-05-29T15:00:09Z) - The Cascaded Forward Algorithm for Neural Network Training [61.06444586991505]
We propose a new learning framework for neural networks, namely Cascaded Forward (CaFo) algorithm, which does not rely on BP optimization as that in FF.
Unlike FF, our framework directly outputs label distributions at each cascaded block, which does not require generation of additional negative samples.
In our framework each block can be trained independently, so it can be easily deployed into parallel acceleration systems.
arXiv Detail & Related papers (2023-03-17T02:01:11Z) - Hebbian Deep Learning Without Feedback [0.0]
We present SoftHebb, an algorithm that trains deep neural networks without any feedback, target, or error signals.
Its increased efficiency and biological compatibility do not trade off accuracy compared to state-of-the-art bio-plausible learning.
arXiv Detail & Related papers (2022-09-23T23:12:59Z) - A Theoretical Framework for Inference Learning [1.433758865948252]
Backpropagation (BP) is the most successful and widely used algorithm in deep learning.
Inference learning algorithm (IL) has close connections to neurobiological models of cortical function.
IL has equal performance to BP on supervised learning and auto-associative tasks.
arXiv Detail & Related papers (2022-06-01T00:38:55Z) - Credit Assignment in Neural Networks through Deep Feedback Control [59.14935871979047]
Deep Feedback Control (DFC) is a new learning method that uses a feedback controller to drive a deep neural network to match a desired output target and whose control signal can be used for credit assignment.
The resulting learning rule is fully local in space and time and approximates Gauss-Newton optimization for a wide range of connectivity patterns.
To further underline its biological plausibility, we relate DFC to a multi-compartment model of cortical pyramidal neurons with a local voltage-dependent synaptic plasticity rule, consistent with recent theories of dendritic processing.
arXiv Detail & Related papers (2021-06-15T05:30:17Z) - Predictive Coding Can Do Exact Backpropagation on Any Neural Network [40.51949948934705]
We generalize (IL and) Z-IL by directly defining them on computational graphs.
This is the first biologically plausible algorithm that is shown to be equivalent to BP in the way of updating parameters on any neural network.
arXiv Detail & Related papers (2021-03-08T11:52:51Z) - A Theoretical Framework for Target Propagation [75.52598682467817]
We analyze target propagation (TP), a popular but not yet fully understood alternative to backpropagation (BP)
Our theory shows that TP is closely related to Gauss-Newton optimization and thus substantially differs from BP.
We provide a first solution to this problem through a novel reconstruction loss that improves feedback weight training.
arXiv Detail & Related papers (2020-06-25T12:07:06Z) - Regularized Evolutionary Population-Based Training [11.624954122221562]
This paper presents an algorithm called Population-Based Training (EPBT) that interleaves the training of a DNN's weights with the metalearning of loss functions.
EPBT results in faster, more accurate learning on image classification benchmarks.
arXiv Detail & Related papers (2020-02-11T06:28:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.