Gradient-Free Neural Network Training via Synaptic-Level Reinforcement
Learning
- URL: http://arxiv.org/abs/2105.14383v1
- Date: Sat, 29 May 2021 22:26:18 GMT
- Title: Gradient-Free Neural Network Training via Synaptic-Level Reinforcement
Learning
- Authors: Aman Bhargava, Mohammad R. Rezaei, Milad Lankarany
- Abstract summary: It is widely believed that there is a consistent, synaptic-level learning mechanism in specific brain regions that actualizes learning.
Here we propose an algorithm based on reinforcement learning to generate and apply a simple synaptic-level learning policy.
The robustness and lack of reliance on gradient opens the door for new techniques for training difficult-to-differentiate neural networks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: An ongoing challenge in neural information processing is: how do neurons
adjust their connectivity to improve task performance over time (i.e.,
actualize learning)? It is widely believed that there is a consistent,
synaptic-level learning mechanism in specific brain regions that actualizes
learning. However, the exact nature of this mechanism remains unclear. Here we
propose an algorithm based on reinforcement learning (RL) to generate and apply
a simple synaptic-level learning policy for multi-layer perceptron (MLP)
models. In this algorithm, the action space for each MLP synapse consists of a
small increase, decrease, or null action on the synapse weight, and the state
for each synapse consists of the last two actions and reward signals. A binary
reward signal indicates improvement or deterioration in task performance. The
static policy produces superior training relative to the adaptive policy and is
agnostic to activation function, network shape, and task. Trained MLPs yield
character recognition performance comparable to identically shaped networks
trained with gradient descent. 0 hidden unit character recognition tests
yielded an average validation accuracy of 88.28%, 1.86$\pm$0.47% higher than
the same MLP trained with gradient descent. 32 hidden unit character
recognition tests yielded an average validation accuracy of 88.45%,
1.11$\pm$0.79% lower than the same MLP trained with gradient descent. The
robustness and lack of reliance on gradient computations opens the door for new
techniques for training difficult-to-differentiate artificial neural networks
such as spiking neural networks (SNNs) and recurrent neural networks (RNNs).
Further, the method's simplicity provides a unique opportunity for further
development of local rule-driven multi-agent connectionist models for machine
intelligence analogous to cellular automata.
Related papers
- A Contrastive Symmetric Forward-Forward Algorithm (SFFA) for Continual Learning Tasks [7.345136916791223]
Forward-Forward Algorithm (FFA) has recently gained momentum as an alternative to the conventional back-propagation algorithm for neural network learning.
This work proposes the Symmetric Forward-Forward Algorithm (SFFA), a novel modification of the original FFA which partitions each layer into positive and negative neurons.
arXiv Detail & Related papers (2024-09-11T16:21:44Z) - Grad-Instructor: Universal Backpropagation with Explainable Evaluation Neural Networks for Meta-learning and AutoML [0.0]
An Evaluation Neural Network (ENN) is trained via deep reinforcement learning to predict the performance of the target network.
The ENN then works as an additional evaluation function during backpropagation.
arXiv Detail & Related papers (2024-06-15T08:37:51Z) - Graph Neural Networks Provably Benefit from Structural Information: A
Feature Learning Perspective [53.999128831324576]
Graph neural networks (GNNs) have pioneered advancements in graph representation learning.
This study investigates the role of graph convolution within the context of feature learning theory.
arXiv Detail & Related papers (2023-06-24T10:21:11Z) - SPIDE: A Purely Spike-based Method for Training Feedback Spiking Neural
Networks [56.35403810762512]
Spiking neural networks (SNNs) with event-based computation are promising brain-inspired models for energy-efficient applications on neuromorphic hardware.
We study spike-based implicit differentiation on the equilibrium state (SPIDE) that extends the recently proposed training method.
arXiv Detail & Related papers (2023-02-01T04:22:59Z) - Desire Backpropagation: A Lightweight Training Algorithm for Multi-Layer
Spiking Neural Networks based on Spike-Timing-Dependent Plasticity [13.384228628766236]
Spiking neural networks (SNNs) are a viable alternative to conventional artificial neural networks.
We present desire backpropagation, a method to derive the desired spike activity of all neurons, including the hidden ones.
We trained three-layer networks to classify MNIST and Fashion-MNIST images and reached an accuracy of 98.41% and 87.56%, respectively.
arXiv Detail & Related papers (2022-11-10T08:32:13Z) - An STDP-Based Supervised Learning Algorithm for Spiking Neural Networks [20.309112286222238]
Spiking Neural Networks (SNN) provide a more biological plausible model for the brain.
We propose a supervised learning algorithm based on Spike-Timing Dependent Plasticity (STDP) for a hierarchical SNN consisting of Leaky Integrate-and-fire neurons.
arXiv Detail & Related papers (2022-03-07T13:40:09Z) - Why Lottery Ticket Wins? A Theoretical Perspective of Sample Complexity
on Pruned Neural Networks [79.74580058178594]
We analyze the performance of training a pruned neural network by analyzing the geometric structure of the objective function.
We show that the convex region near a desirable model with guaranteed generalization enlarges as the neural network model is pruned.
arXiv Detail & Related papers (2021-10-12T01:11:07Z) - Training Feedback Spiking Neural Networks by Implicit Differentiation on
the Equilibrium State [66.2457134675891]
Spiking neural networks (SNNs) are brain-inspired models that enable energy-efficient implementation on neuromorphic hardware.
Most existing methods imitate the backpropagation framework and feedforward architectures for artificial neural networks.
We propose a novel training method that does not rely on the exact reverse of the forward computation.
arXiv Detail & Related papers (2021-09-29T07:46:54Z) - A proof of convergence for stochastic gradient descent in the training
of artificial neural networks with ReLU activation for constant target
functions [3.198144010381572]
We study the gradient descent (SGD) optimization method in the training of fully-connected feedforward artificial neural networks with ReLU activation.
The main result of this work proves that the risk of the SGD process converges to zero if the target function under consideration is constant.
arXiv Detail & Related papers (2021-04-01T06:28:30Z) - How Neural Networks Extrapolate: From Feedforward to Graph Neural
Networks [80.55378250013496]
We study how neural networks trained by gradient descent extrapolate what they learn outside the support of the training distribution.
Graph Neural Networks (GNNs) have shown some success in more complex tasks.
arXiv Detail & Related papers (2020-09-24T17:48:59Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.