Predictive Coding Can Do Exact Backpropagation on Convolutional and
Recurrent Neural Networks
- URL: http://arxiv.org/abs/2103.03725v1
- Date: Fri, 5 Mar 2021 14:57:01 GMT
- Title: Predictive Coding Can Do Exact Backpropagation on Convolutional and
Recurrent Neural Networks
- Authors: Tommaso Salvatori, Yuhang Song, Thomas Lukasiewicz, Rafal Bogacz,
Zhenghua Xu
- Abstract summary: Predictive coding networks (PCNs) are an influential model for information processing in the brain.
BP is commonly regarded to be the most successful learning method in modern machine learning.
We show that a biologically plausible algorithm is able to exactly replicate the accuracy of BP on complex architectures.
- Score: 40.51949948934705
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Predictive coding networks (PCNs) are an influential model for information
processing in the brain. They have appealing theoretical interpretations and
offer a single mechanism that accounts for diverse perceptual phenomena of the
brain. On the other hand, backpropagation (BP) is commonly regarded to be the
most successful learning method in modern machine learning. Thus, it is
exciting that recent work formulates inference learning (IL) that trains PCNs
to approximate BP. However, there are several remaining critical issues: (i) IL
is an approximation to BP with unrealistic/non-trivial requirements, (ii) IL
approximates BP in single-step weight updates; whether it leads to the same
point as BP after the weight updates are conducted for more steps is unknown,
and (iii) IL is computationally significantly more costly than BP. To solve
these issues, a variant of IL that is strictly equivalent to BP in fully
connected networks has been proposed. In this work, we build on this result by
showing that it also holds for more complex architectures, namely,
convolutional neural networks and (many-to-one) recurrent neural networks. To
our knowledge, we are the first to show that a biologically plausible algorithm
is able to exactly replicate the accuracy of BP on such complex architectures,
bridging the existing gap between IL and BP, and setting an unprecedented
performance for PCNs, which can now be considered as efficient alternatives to
BP.
Related papers
- Predictive Coding Networks and Inference Learning: Tutorial and Survey [0.7510165488300368]
Predictive coding networks (PCNs) are based on the neuroscientific framework of predictive coding.
Unlike traditional neural networks trained with backpropagation (BP), PCNs utilize inference learning (IL), a more biologically plausible algorithm.
As inherently probabilistic (graphical) latent variable models, PCNs provide a versatile framework for both supervised learning and unsupervised (generative) modeling.
arXiv Detail & Related papers (2024-07-04T18:39:20Z) - A Theoretical Framework for Inference and Learning in Predictive Coding
Networks [41.58529335439799]
Predictive coding (PC) is an influential theory in computational neuroscience.
We provide a comprehensive theoretical analysis of the properties of PCNs trained with prospective configuration.
arXiv Detail & Related papers (2022-07-21T04:17:55Z) - BP-Net: Cuff-less, Calibration-free, and Non-invasive Blood Pressure
Estimation via a Generic Deep Convolutional Architecture [16.36324484557899]
The paper focuses on development of robust and accurate processing solutions for continuous and computation-less blood pressure (BP) monitoring.
The proposed framework is a novel convolutional architecture that provides longer effective memory.
The proposed BP-Net architecture is more accurate than canonical recurrent networks and enhances the long-term robustness of the BP estimation task.
arXiv Detail & Related papers (2021-12-31T02:34:39Z) - A Theoretical View of Linear Backpropagation and Its Convergence [55.69505060636719]
Backpropagation (BP) is widely used for calculating gradients in deep neural networks (DNNs)
Recently, a linear variant of BP named LinBP was introduced for generating more transferable adversarial examples for performing black-box attacks.
We provide theoretical analyses on LinBP in neural-network-involved learning tasks, including adversarial attack and model training.
arXiv Detail & Related papers (2021-12-21T07:18:00Z) - Towards Evaluating and Training Verifiably Robust Neural Networks [81.39994285743555]
We study the relationship between IBP and CROWN, and prove that CROWN is always tighter than IBP when choosing appropriate bounding lines.
We propose a relaxed version of CROWN, linear bound propagation (LBP), that can be used to verify large networks to obtain lower verified errors.
arXiv Detail & Related papers (2021-04-01T13:03:48Z) - Predictive Coding Can Do Exact Backpropagation on Any Neural Network [40.51949948934705]
We generalize (IL and) Z-IL by directly defining them on computational graphs.
This is the first biologically plausible algorithm that is shown to be equivalent to BP in the way of updating parameters on any neural network.
arXiv Detail & Related papers (2021-03-08T11:52:51Z) - Belief Propagation Neural Networks [103.97004780313105]
We introduce belief propagation neural networks (BPNNs)
BPNNs operate on factor graphs and generalize Belief propagation (BP)
We show that BPNNs converges 1.7x faster on Ising models while providing tighter bounds.
On challenging model counting problems, BPNNs compute estimates 100's of times faster than state-of-the-art handcrafted methods.
arXiv Detail & Related papers (2020-07-01T07:39:51Z) - A Theoretical Framework for Target Propagation [75.52598682467817]
We analyze target propagation (TP), a popular but not yet fully understood alternative to backpropagation (BP)
Our theory shows that TP is closely related to Gauss-Newton optimization and thus substantially differs from BP.
We provide a first solution to this problem through a novel reconstruction loss that improves feedback weight training.
arXiv Detail & Related papers (2020-06-25T12:07:06Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.