Extension of Direct Feedback Alignment to Convolutional and Recurrent
Neural Network for Bio-plausible Deep Learning
- URL: http://arxiv.org/abs/2006.12830v1
- Date: Tue, 23 Jun 2020 08:42:22 GMT
- Title: Extension of Direct Feedback Alignment to Convolutional and Recurrent
Neural Network for Bio-plausible Deep Learning
- Authors: Donghyeon Han and Gwangtae Park and Junha Ryu and Hoi-jun Yoo
- Abstract summary: We focus on the improvement of the direct feedback alignment (DFA) algorithm.
We extend the usage of the DFA to convolutional and recurrent neural networks (CNNs and RNNs)
We propose a new DFA algorithm for BP-level accurate CNN and RNN training.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Throughout this paper, we focus on the improvement of the direct feedback
alignment (DFA) algorithm and extend the usage of the DFA to convolutional and
recurrent neural networks (CNNs and RNNs). Even though the DFA algorithm is
biologically plausible and has a potential of high-speed training, it has not
been considered as the substitute for back-propagation (BP) due to the low
accuracy in the CNN and RNN training. In this work, we propose a new DFA
algorithm for BP-level accurate CNN and RNN training. Firstly, we divide the
network into several modules and apply the DFA algorithm within the module.
Second, the DFA with the sparse backward weight is applied. It comes with a
form of dilated convolution in the CNN case, and in a form of sparse matrix
multiplication in the RNN case. Additionally, the error propagation method of
CNN becomes simpler through the group convolution. Finally, hybrid DFA
increases the accuracy of the CNN and RNN training to the BP-level while taking
advantage of the parallelism and hardware efficiency of the DFA algorithm.
Related papers
- Model Parallel Training and Transfer Learning for Convolutional Neural Networks by Domain Decomposition [0.0]
Deep convolutional neural networks (CNNs) have been shown to be very successful in a wide range of image processing applications.
Due to their increasing number of model parameters and an increasing availability of large amounts of training data, parallelization strategies to efficiently train complex CNNs are necessary.
arXiv Detail & Related papers (2024-08-26T17:35:01Z) - Training Convolutional Neural Networks with the Forward-Forward
algorithm [1.74440662023704]
Forward Forward (FF) algorithm has up to now only been used in fully connected networks.
We show how the FF paradigm can be extended to CNNs.
Our FF-trained CNN, featuring a novel spatially-extended labeling technique, achieves a classification accuracy of 99.16% on the MNIST hand-written digits dataset.
arXiv Detail & Related papers (2023-12-22T18:56:35Z) - Recurrent Bilinear Optimization for Binary Neural Networks [58.972212365275595]
BNNs neglect the intrinsic bilinear relationship of real-valued weights and scale factors.
Our work is the first attempt to optimize BNNs from the bilinear perspective.
We obtain robust RBONNs, which show impressive performance over state-of-the-art BNNs on various models and datasets.
arXiv Detail & Related papers (2022-09-04T06:45:33Z) - Recursive Least Squares for Training and Pruning Convolutional Neural
Networks [27.089496826735672]
Convolutional neural networks (CNNs) have succeeded in many practical applications.
High computation and storage requirements make them difficult to deploy on resource-constrained devices.
We propose a novel algorithm for training and pruning CNNs.
arXiv Detail & Related papers (2022-01-13T07:14:08Z) - Credit Assignment Through Broadcasting a Global Error Vector [4.683806391173103]
Backpropagation (BP) uses detailed, unit-specific feedback to train deep neural networks (DNNs) with remarkable success.
Here, we explore the extent to which a globally broadcast learning signal, coupled with local weight updates, enables training of DNNs.
arXiv Detail & Related papers (2021-06-08T04:08:46Z) - BreakingBED -- Breaking Binary and Efficient Deep Neural Networks by
Adversarial Attacks [65.2021953284622]
We study robustness of CNNs against white-box and black-box adversarial attacks.
Results are shown for distilled CNNs, agent-based state-of-the-art pruned models, and binarized neural networks.
arXiv Detail & Related papers (2021-03-14T20:43:19Z) - A Meta-Learning Approach to the Optimal Power Flow Problem Under
Topology Reconfigurations [69.73803123972297]
We propose a DNN-based OPF predictor that is trained using a meta-learning (MTL) approach.
The developed OPF-predictor is validated through simulations using benchmark IEEE bus systems.
arXiv Detail & Related papers (2020-12-21T17:39:51Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Computational optimization of convolutional neural networks using
separated filters architecture [69.73393478582027]
We consider a convolutional neural network transformation that reduces computation complexity and thus speedups neural network processing.
Use of convolutional neural networks (CNN) is the standard approach to image recognition despite the fact they can be too computationally demanding.
arXiv Detail & Related papers (2020-02-18T17:42:13Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.