Understanding Predictive Coding as an Adaptive Trust-Region Method
- URL: http://arxiv.org/abs/2305.18188v1
- Date: Mon, 29 May 2023 16:25:55 GMT
- Title: Understanding Predictive Coding as an Adaptive Trust-Region Method
- Authors: Francesco Innocenti, Ryan Singh, Christopher L. Buckley
- Abstract summary: We develop a theory of PC as an adaptive trust-region (TR) algorithm that uses second-order information.
We show that the learning dynamics of PC can be interpreted as interpolating between BP's loss gradient direction and a TR direction found by the PC inference dynamics.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Predictive coding (PC) is a brain-inspired local learning algorithm that has
recently been suggested to provide advantages over backpropagation (BP) in
biologically relevant scenarios. While theoretical work has mainly focused on
showing how PC can approximate BP in various limits, the putative benefits of
"natural" PC are less understood. Here we develop a theory of PC as an adaptive
trust-region (TR) algorithm that uses second-order information. We show that
the learning dynamics of PC can be interpreted as interpolating between BP's
loss gradient direction and a TR direction found by the PC inference dynamics.
Our theory suggests that PC should escape saddle points faster than BP, a
prediction which we prove in a shallow linear model and support with
experiments on deeper networks. This work lays a foundation for understanding
PC in deep and wide networks.
Related papers
- On the Infinite Width and Depth Limits of Predictive Coding Networks [8.779034498638826]
Predictive coding (PC) is a biologically plausible alternative to standard backpropagation (BP)<n>Recent work has improved the training stability of deep PC networks.<n>We study the infinite width and depth limits of PCNs.
arXiv Detail & Related papers (2026-02-07T20:47:32Z) - Towards Scaling Deep Neural Networks with Predictive Coding: Theory and Practice [1.2691047660244335]
Backpropagation (BP) is the standard algorithm for training the deep neural networks that power modern artificial intelligence.<n>This thesis studies an alternative, potentially more efficient brain-inspired algorithm called predictive coding (PC)
arXiv Detail & Related papers (2025-10-24T14:47:49Z) - Error Optimization: Overcoming Exponential Signal Decay in Deep Predictive Coding Networks [11.970327820917761]
Predictive Coding (PC) offers a biologically plausible alternative to backpropagation for neural network training, yet struggles with deeper architectures.<n>This paper identifies the root cause: an inherent signal decay problem where gradients attenuate exponentially with depth, becoming computationally negligible due to numerical precision constraints.<n>To address this fundamental limitation, we introduce Error Optimization (EO), a novel re parameterization that preserves PC's theoretical properties while eliminating signal decay.
arXiv Detail & Related papers (2025-05-26T15:39:16Z) - Tight Stability, Convergence, and Robustness Bounds for Predictive Coding Networks [60.3634789164648]
Energy-based learning algorithms, such as predictive coding (PC), have garnered significant attention in the machine learning community.
We rigorously analyze the stability, robustness, and convergence of PC through the lens of dynamical systems theory.
arXiv Detail & Related papers (2024-10-07T02:57:26Z) - Predictive Coding Networks and Inference Learning: Tutorial and Survey [0.7510165488300368]
Predictive coding networks (PCNs) are based on the neuroscientific framework of predictive coding.
Unlike traditional neural networks trained with backpropagation (BP), PCNs utilize inference learning (IL), a more biologically plausible algorithm.
As inherently probabilistic (graphical) latent variable models, PCNs provide a versatile framework for both supervised learning and unsupervised (generative) modeling.
arXiv Detail & Related papers (2024-07-04T18:39:20Z) - Deep Predictive Coding with Bi-directional Propagation for
Classification and Reconstruction [1.4480964546077346]
This paper presents a new learning algorithm termed Deep Bi-directional Predictive Coding (DBPC)
DBPC allows developing networks to simultaneously perform classification and reconstruction tasks using the same weights.
The performance of DBPC has been evaluated on both, classification and reconstruction tasks using the MNIST and FashionMNIST datasets.
arXiv Detail & Related papers (2023-05-29T10:17:13Z) - A Theoretical Framework for Inference and Learning in Predictive Coding
Networks [41.58529335439799]
Predictive coding (PC) is an influential theory in computational neuroscience.
We provide a comprehensive theoretical analysis of the properties of PCNs trained with prospective configuration.
arXiv Detail & Related papers (2022-07-21T04:17:55Z) - Towards Scaling Difference Target Propagation by Learning Backprop
Targets [64.90165892557776]
Difference Target Propagation is a biologically-plausible learning algorithm with close relation with Gauss-Newton (GN) optimization.
We propose a novel feedback weight training scheme that ensures both that DTP approximates BP and that layer-wise feedback weight training can be restored.
We report the best performance ever achieved by DTP on CIFAR-10 and ImageNet.
arXiv Detail & Related papers (2022-01-31T18:20:43Z) - A Theoretical View of Linear Backpropagation and Its Convergence [55.69505060636719]
Backpropagation (BP) is widely used for calculating gradients in deep neural networks (DNNs)
Recently, a linear variant of BP named LinBP was introduced for generating more transferable adversarial examples for performing black-box attacks.
We provide theoretical analyses on LinBP in neural-network-involved learning tasks, including adversarial attack and model training.
arXiv Detail & Related papers (2021-12-21T07:18:00Z) - Credit Assignment in Neural Networks through Deep Feedback Control [59.14935871979047]
Deep Feedback Control (DFC) is a new learning method that uses a feedback controller to drive a deep neural network to match a desired output target and whose control signal can be used for credit assignment.
The resulting learning rule is fully local in space and time and approximates Gauss-Newton optimization for a wide range of connectivity patterns.
To further underline its biological plausibility, we relate DFC to a multi-compartment model of cortical pyramidal neurons with a local voltage-dependent synaptic plasticity rule, consistent with recent theories of dendritic processing.
arXiv Detail & Related papers (2021-06-15T05:30:17Z) - Predictive Coding Can Do Exact Backpropagation on Convolutional and
Recurrent Neural Networks [40.51949948934705]
Predictive coding networks (PCNs) are an influential model for information processing in the brain.
BP is commonly regarded to be the most successful learning method in modern machine learning.
We show that a biologically plausible algorithm is able to exactly replicate the accuracy of BP on complex architectures.
arXiv Detail & Related papers (2021-03-05T14:57:01Z) - A Theoretical Framework for Target Propagation [75.52598682467817]
We analyze target propagation (TP), a popular but not yet fully understood alternative to backpropagation (BP)
Our theory shows that TP is closely related to Gauss-Newton optimization and thus substantially differs from BP.
We provide a first solution to this problem through a novel reconstruction loss that improves feedback weight training.
arXiv Detail & Related papers (2020-06-25T12:07:06Z) - Predictive Coding Approximates Backprop along Arbitrary Computation
Graphs [68.8204255655161]
We develop a strategy to translate core machine learning architectures into their predictive coding equivalents.
Our models perform equivalently to backprop on challenging machine learning benchmarks.
Our method raises the potential that standard machine learning algorithms could in principle be directly implemented in neural circuitry.
arXiv Detail & Related papers (2020-06-07T15:35:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.