Human-Level Control through Directly-Trained Deep Spiking Q-Networks
- URL: http://arxiv.org/abs/2201.07211v3
- Date: Tue, 11 Apr 2023 01:41:12 GMT
- Title: Human-Level Control through Directly-Trained Deep Spiking Q-Networks
- Authors: Guisong Liu, Wenjie Deng, Xiurui Xie, Li Huang, Huajin Tang
- Abstract summary: Spiking Neural Networks (SNNs) have great potential on neuromorphic hardware because of their high energy-efficiency.
We propose a directly-trained deep spiking reinforcement learning architecture based on the Leaky Integrate-and-Fire neurons and Deep Q-Network.
Our work is the first one to achieve state-of-the-art performance on multiple Atari games with the directly-trained SNN.
- Score: 16.268397551693862
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As the third-generation neural networks, Spiking Neural Networks (SNNs) have
great potential on neuromorphic hardware because of their high
energy-efficiency. However, Deep Spiking Reinforcement Learning (DSRL), i.e.,
the Reinforcement Learning (RL) based on SNNs, is still in its preliminary
stage due to the binary output and the non-differentiable property of the
spiking function. To address these issues, we propose a Deep Spiking Q-Network
(DSQN) in this paper. Specifically, we propose a directly-trained deep spiking
reinforcement learning architecture based on the Leaky Integrate-and-Fire (LIF)
neurons and Deep Q-Network (DQN). Then, we adapt a direct spiking learning
algorithm for the Deep Spiking Q-Network. We further demonstrate the advantages
of using LIF neurons in DSQN theoretically. Comprehensive experiments have been
conducted on 17 top-performing Atari games to compare our method with the
state-of-the-art conversion method. The experimental results demonstrate the
superiority of our method in terms of performance, stability, robustness and
energy-efficiency. To the best of our knowledge, our work is the first one to
achieve state-of-the-art performance on multiple Atari games with the
directly-trained SNN.
Related papers
- Fully Spiking Actor Network with Intra-layer Connections for
Reinforcement Learning [51.386945803485084]
We focus on the task where the agent needs to learn multi-dimensional deterministic policies to control.
Most existing spike-based RL methods take the firing rate as the output of SNNs, and convert it to represent continuous action space (i.e., the deterministic policy) through a fully-connected layer.
To develop a fully spiking actor network without any floating-point matrix operations, we draw inspiration from the non-spiking interneurons found in insects.
arXiv Detail & Related papers (2024-01-09T07:31:34Z) - SpikingJelly: An open-source machine learning infrastructure platform
for spike-based intelligence [51.6943465041708]
Spiking neural networks (SNNs) aim to realize brain-inspired intelligence on neuromorphic chips with high energy efficiency.
We contribute a full-stack toolkit for pre-processing neuromorphic datasets, building deep SNNs, optimizing their parameters, and deploying SNNs on neuromorphic chips.
arXiv Detail & Related papers (2023-10-25T13:15:17Z) - Deep Multi-Threshold Spiking-UNet for Image Processing [51.88730892920031]
This paper introduces the novel concept of Spiking-UNet for image processing, which combines the power of Spiking Neural Networks (SNNs) with the U-Net architecture.
To achieve an efficient Spiking-UNet, we face two primary challenges: ensuring high-fidelity information propagation through the network via spikes and formulating an effective training strategy.
Experimental results show that, on image segmentation and denoising, our Spiking-UNet achieves comparable performance to its non-spiking counterpart.
arXiv Detail & Related papers (2023-07-20T16:00:19Z) - Joint A-SNN: Joint Training of Artificial and Spiking Neural Networks
via Self-Distillation and Weight Factorization [12.1610509770913]
Spiking Neural Networks (SNNs) mimic the spiking nature of brain neurons.
We propose a joint training framework of ANN and SNN, in which the ANN can guide the SNN's optimization.
Our method consistently outperforms many other state-of-the-art training methods.
arXiv Detail & Related papers (2023-05-03T13:12:17Z) - SPIDE: A Purely Spike-based Method for Training Feedback Spiking Neural
Networks [56.35403810762512]
Spiking neural networks (SNNs) with event-based computation are promising brain-inspired models for energy-efficient applications on neuromorphic hardware.
We study spike-based implicit differentiation on the equilibrium state (SPIDE) that extends the recently proposed training method.
arXiv Detail & Related papers (2023-02-01T04:22:59Z) - Solving the Spike Feature Information Vanishing Problem in Spiking Deep
Q Network with Potential Based Normalization [7.796499799525251]
We propose a potential based layer normalization(pbLN) method to directly train spiking deep Q networks.
Experiment shows that compared with state-of-art ANN-SNN conversion method and other SDQN works, the proposed pbLN spiking deep Q networks (PL-SDQN) achieved better performance on Atari game tasks.
arXiv Detail & Related papers (2022-06-08T02:45:18Z) - Deep Reinforcement Learning with Spiking Q-learning [51.386945803485084]
spiking neural networks (SNNs) are expected to realize artificial intelligence (AI) with less energy consumption.
It provides a promising energy-efficient way for realistic control tasks by combining SNNs with deep reinforcement learning (RL)
arXiv Detail & Related papers (2022-01-21T16:42:11Z) - Keys to Accurate Feature Extraction Using Residual Spiking Neural
Networks [1.101002667958165]
Spiking neural networks (SNNs) have become an interesting alternative to conventional artificial neural networks (ANNs)
We present a study on the key components of modern spiking architectures.
We design a spiking version of the successful residual network (ResNet) architecture and test different components and training strategies on it.
arXiv Detail & Related papers (2021-11-10T21:29:19Z) - Effective and Efficient Computation with Multiple-timescale Spiking
Recurrent Neural Networks [0.9790524827475205]
We show how a novel type of adaptive spiking recurrent neural network (SRNN) is able to achieve state-of-the-art performance.
We calculate a $>$100x energy improvement for our SRNNs over classical RNNs on the harder tasks.
arXiv Detail & Related papers (2020-05-24T01:04:53Z) - Rectified Linear Postsynaptic Potential Function for Backpropagation in
Deep Spiking Neural Networks [55.0627904986664]
Spiking Neural Networks (SNNs) usetemporal spike patterns to represent and transmit information, which is not only biologically realistic but also suitable for ultra-low-power event-driven neuromorphic implementation.
This paper investigates the contribution of spike timing dynamics to information encoding, synaptic plasticity and decision making, providing a new perspective to design of future DeepSNNs and neuromorphic hardware systems.
arXiv Detail & Related papers (2020-03-26T11:13:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.