Deep Predictive Coding with Bi-directional Propagation for
Classification and Reconstruction
- URL: http://arxiv.org/abs/2305.18472v1
- Date: Mon, 29 May 2023 10:17:13 GMT
- Title: Deep Predictive Coding with Bi-directional Propagation for
Classification and Reconstruction
- Authors: Senhui Qiu, Saugat Bhattacharyya, Damien Coyle, Shirin Dora
- Abstract summary: This paper presents a new learning algorithm termed Deep Bi-directional Predictive Coding (DBPC)
DBPC allows developing networks to simultaneously perform classification and reconstruction tasks using the same weights.
The performance of DBPC has been evaluated on both, classification and reconstruction tasks using the MNIST and FashionMNIST datasets.
- Score: 1.4480964546077346
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents a new learning algorithm, termed Deep Bi-directional
Predictive Coding (DBPC) that allows developing networks to simultaneously
perform classification and reconstruction tasks using the same weights.
Predictive Coding (PC) has emerged as a prominent theory underlying information
processing in the brain. The general concept for learning in PC is that each
layer learns to predict the activities of neurons in the previous layer which
enables local computation of error and in-parallel learning across layers. In
this paper, we extend existing PC approaches by developing a network which
supports both feedforward and feedback propagation of information. Each layer
in the networks trained using DBPC learn to predict the activities of neurons
in the previous and next layer which allows the network to simultaneously
perform classification and reconstruction tasks using feedforward and feedback
propagation, respectively. DBPC also relies on locally available information
for learning, thus enabling in-parallel learning across all layers in the
network. The proposed approach has been developed for training both, fully
connected networks and convolutional neural networks. The performance of DBPC
has been evaluated on both, classification and reconstruction tasks using the
MNIST and FashionMNIST datasets. The classification and the reconstruction
performance of networks trained using DBPC is similar to other approaches used
for comparison but DBPC uses a significantly smaller network. Further, the
significant benefit of DBPC is its ability to achieve this performance using
locally available information and in-parallel learning mechanisms which results
in an efficient training protocol. This results clearly indicate that DBPC is a
much more efficient approach for developing networks that can simultaneously
perform both classification and reconstruction.
Related papers
- Information-Theoretic Greedy Layer-wise Training for Traffic Sign Recognition [0.5024983453990065]
layer-wise training eliminates the need for cross-entropy loss and backpropagation.<n>Most existing layer-wise training approaches have been evaluated only on relatively small datasets.<n>We propose a novel layer-wise training approach based on the recently developed deterministic information bottleneck (DIB) and the matrix-based R'enyi's $alpha$-order entropy functional.
arXiv Detail & Related papers (2025-10-31T17:24:58Z) - Predictive Coding Networks and Inference Learning: Tutorial and Survey [0.7510165488300368]
Predictive coding networks (PCNs) are based on the neuroscientific framework of predictive coding.
Unlike traditional neural networks trained with backpropagation (BP), PCNs utilize inference learning (IL), a more biologically plausible algorithm.
As inherently probabilistic (graphical) latent variable models, PCNs provide a versatile framework for both supervised learning and unsupervised (generative) modeling.
arXiv Detail & Related papers (2024-07-04T18:39:20Z) - An intertwined neural network model for EEG classification in
brain-computer interfaces [0.6696153817334769]
The brain computer interface (BCI) is a nonstimulatory direct and occasionally bidirectional communication link between the brain and a computer or an external device.
We present a deep neural network architecture specifically engineered to provide state-of-the-art performance in multiclass motor imagery classification.
arXiv Detail & Related papers (2022-08-04T09:00:34Z) - Biologically Plausible Training of Deep Neural Networks Using a Top-down
Credit Assignment Network [32.575847142016585]
Top-Down Credit Assignment Network (TDCA-network) is designed to train a bottom-up network using a Top-Down Credit Assignment Network (TDCA-network)
TDCA-network serves as a substitute for the conventional loss function and the back-propagation algorithm, widely used in neural network training.
The results indicate TDCA-network holds promising potential to train neural networks across diverse datasets.
arXiv Detail & Related papers (2022-08-01T07:14:37Z) - Learning on Arbitrary Graph Topologies via Predictive Coding [38.761663028090204]
We show how predictive coding can be used to perform inference and learning on arbitrary graph topologies.
We experimentally show how this formulation, called PC graphs, can be used to flexibly perform different tasks with the same network.
arXiv Detail & Related papers (2022-01-31T12:43:22Z) - Distributed Learning for Time-varying Networks: A Scalable Design [13.657740129012804]
We propose a distributed learning framework based on a scalable deep neural network (DNN) design.
By exploiting the permutation equivalence and invariance properties of the learning tasks, the DNNs with different scales for different clients can be built up.
Model aggregation can also be conducted based on these two sub-matrices to improve the learning convergence and performance.
arXiv Detail & Related papers (2021-07-31T12:44:28Z) - Learning Structures for Deep Neural Networks [99.8331363309895]
We propose to adopt the efficient coding principle, rooted in information theory and developed in computational neuroscience.
We show that sparse coding can effectively maximize the entropy of the output signals.
Our experiments on a public image classification dataset demonstrate that using the structure learned from scratch by our proposed algorithm, one can achieve a classification accuracy comparable to the best expert-designed structure.
arXiv Detail & Related papers (2021-05-27T12:27:24Z) - Joint Learning of Neural Transfer and Architecture Adaptation for Image
Recognition [77.95361323613147]
Current state-of-the-art visual recognition systems rely on pretraining a neural network on a large-scale dataset and finetuning the network weights on a smaller dataset.
In this work, we prove that dynamically adapting network architectures tailored for each domain task along with weight finetuning benefits in both efficiency and effectiveness.
Our method can be easily generalized to an unsupervised paradigm by replacing supernet training with self-supervised learning in the source domain tasks and performing linear evaluation in the downstream tasks.
arXiv Detail & Related papers (2021-03-31T08:15:17Z) - PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive
Learning [109.84770951839289]
We present PredRNN, a new recurrent network for learning visual dynamics from historical context.
We show that our approach obtains highly competitive results on three standard datasets.
arXiv Detail & Related papers (2021-03-17T08:28:30Z) - Local Critic Training for Model-Parallel Learning of Deep Neural
Networks [94.69202357137452]
We propose a novel model-parallel learning method, called local critic training.
We show that the proposed approach successfully decouples the update process of the layer groups for both convolutional neural networks (CNNs) and recurrent neural networks (RNNs)
We also show that trained networks by the proposed method can be used for structural optimization.
arXiv Detail & Related papers (2021-02-03T09:30:45Z) - Understanding the Effects of Data Parallelism and Sparsity on Neural
Network Training [126.49572353148262]
We study two factors in neural network training: data parallelism and sparsity.
Despite their promising benefits, understanding of their effects on neural network training remains elusive.
arXiv Detail & Related papers (2020-03-25T10:49:22Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.