Biological credit assignment through dynamic inversion of feedforward
networks
- URL: http://arxiv.org/abs/2007.05112v2
- Date: Mon, 4 Jan 2021 00:31:24 GMT
- Title: Biological credit assignment through dynamic inversion of feedforward
networks
- Authors: William F. Podlaski, Christian K. Machens
- Abstract summary: We show that feedforward network transformations can be effectively inverted through dynamics.
We map these dynamics onto generic feedforward networks, and show that the resulting algorithm performs well on supervised and unsupervised datasets.
- Score: 11.345796608258434
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning depends on changes in synaptic connections deep inside the brain. In
multilayer networks, these changes are triggered by error signals fed back from
the output, generally through a stepwise inversion of the feedforward
processing steps. The gold standard for this process -- backpropagation --
works well in artificial neural networks, but is biologically implausible.
Several recent proposals have emerged to address this problem, but many of
these biologically-plausible schemes are based on learning an independent set
of feedback connections. This complicates the assignment of errors to each
synapse by making it dependent upon a second learning problem, and by fitting
inversions rather than guaranteeing them. Here, we show that feedforward
network transformations can be effectively inverted through dynamics. We derive
this dynamic inversion from the perspective of feedback control, where the
forward transformation is reused and dynamically interacts with fixed or random
feedback to propagate error signals during the backward pass. Importantly, this
scheme does not rely upon a second learning problem for feedback because
accurate inversion is guaranteed through the network dynamics. We map these
dynamics onto generic feedforward networks, and show that the resulting
algorithm performs well on several supervised and unsupervised datasets.
Finally, we discuss potential links between dynamic inversion and second-order
optimization. Overall, our work introduces an alternative perspective on credit
assignment in the brain, and proposes a special role for temporal dynamics and
feedback control during learning.
Related papers
- Dynamical stability and chaos in artificial neural network trajectories along training [3.379574469735166]
We study the dynamical properties of this process by analyzing through this lens the network trajectories of a shallow neural network.
We find hints of regular and chaotic behavior depending on the learning rate regime.
This work also contributes to the cross-fertilization of ideas between dynamical systems theory, network theory and machine learning.
arXiv Detail & Related papers (2024-04-08T17:33:11Z) - Leveraging Low-Rank and Sparse Recurrent Connectivity for Robust
Closed-Loop Control [63.310780486820796]
We show how a parameterization of recurrent connectivity influences robustness in closed-loop settings.
We find that closed-form continuous-time neural networks (CfCs) with fewer parameters can outperform their full-rank, fully-connected counterparts.
arXiv Detail & Related papers (2023-10-05T21:44:18Z) - Centered Self-Attention Layers [89.21791761168032]
The self-attention mechanism in transformers and the message-passing mechanism in graph neural networks are repeatedly applied.
We show that this application inevitably leads to oversmoothing, i.e., to similar representations at the deeper layers.
We present a correction term to the aggregating operator of these mechanisms.
arXiv Detail & Related papers (2023-06-02T15:19:08Z) - Dynamics-aware Adversarial Attack of Adaptive Neural Networks [75.50214601278455]
We investigate the dynamics-aware adversarial attack problem of adaptive neural networks.
We propose a Leaded Gradient Method (LGM) and show the significant effects of the lagged gradient.
Our LGM achieves impressive adversarial attack performance compared with the dynamic-unaware attack methods.
arXiv Detail & Related papers (2022-10-15T01:32:08Z) - Learning Dynamics and Generalization in Reinforcement Learning [59.530058000689884]
We show theoretically that temporal difference learning encourages agents to fit non-smooth components of the value function early in training.
We show that neural networks trained using temporal difference algorithms on dense reward tasks exhibit weaker generalization between states than randomly networks and gradient networks trained with policy methods.
arXiv Detail & Related papers (2022-06-05T08:49:16Z) - Minimizing Control for Credit Assignment with Strong Feedback [65.59995261310529]
Current methods for gradient-based credit assignment in deep neural networks need infinitesimally small feedback signals.
We combine strong feedback influences on neural activity with gradient-based learning and show that this naturally leads to a novel view on neural network optimization.
We show that the use of strong feedback in DFC allows learning forward and feedback connections simultaneously, using a learning rule fully local in space and time.
arXiv Detail & Related papers (2022-04-14T22:06:21Z) - Data-driven emergence of convolutional structure in neural networks [83.4920717252233]
We show how fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs.
By carefully designing data models, we show that the emergence of this pattern is triggered by the non-Gaussian, higher-order local structure of the inputs.
arXiv Detail & Related papers (2022-02-01T17:11:13Z) - Error-driven Input Modulation: Solving the Credit Assignment Problem
without a Backward Pass [8.39059551023011]
Supervised learning in artificial neural networks typically relies on backpropagation.
We show that this approach lacks biological plausibility in many regards.
We propose to replace the backward pass with a second forward pass in which the input signal is modulated based on the error of the network.
arXiv Detail & Related papers (2022-01-27T17:12:05Z) - On the role of feedback in visual processing: a predictive coding
perspective [0.6193838300896449]
We consider deep convolutional networks (CNNs) as models of feed-forward visual processing and implement Predictive Coding (PC) dynamics.
We find that the network increasingly relies on top-down predictions as the noise level increases.
In addition, the accuracy of the network implementing PC dynamics significantly increases over time-steps, compared to its equivalent forward network.
arXiv Detail & Related papers (2021-06-08T10:07:23Z) - Predify: Augmenting deep neural networks with brain-inspired predictive
coding dynamics [0.5284812806199193]
We take inspiration from a popular framework in neuroscience: 'predictive coding'
We show that implementing this strategy into two popular networks, VGG16 and EfficientNetB0, improves their robustness against various corruptions.
arXiv Detail & Related papers (2021-06-04T22:48:13Z) - Implicit recurrent networks: A novel approach to stationary input
processing with recurrent neural networks in deep learning [0.0]
In this work, we introduce and test a novel implementation of recurrent neural networks into deep learning.
We provide an algorithm which implements the backpropagation algorithm on a implicit implementation of recurrent networks.
A single-layer implicit recurrent network is able to solve the XOR problem, while a feed-forward network with monotonically increasing activation function fails at this task.
arXiv Detail & Related papers (2020-10-20T18:55:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.