Spiker+: a framework for the generation of efficient Spiking Neural
Networks FPGA accelerators for inference at the edge
- URL: http://arxiv.org/abs/2401.01141v1
- Date: Tue, 2 Jan 2024 10:42:42 GMT
- Title: Spiker+: a framework for the generation of efficient Spiking Neural
Networks FPGA accelerators for inference at the edge
- Authors: Alessio Carpegna, Alessandro Savino, Stefano Di Carlo
- Abstract summary: Spiker+ is a framework for generating efficient, low-power, and low-area customized Spiking Neural Networks (SNN) accelerators on FPGA for inference at the edge.
Spiker+ is tested on two benchmark datasets, the MNIST and the Spiking Heidelberg Digits (SHD)
- Score: 49.42371633618761
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Including Artificial Neural Networks in embedded systems at the edge allows
applications to exploit Artificial Intelligence capabilities directly within
devices operating at the network periphery. This paper introduces Spiker+, a
comprehensive framework for generating efficient, low-power, and low-area
customized Spiking Neural Networks (SNN) accelerators on FPGA for inference at
the edge. Spiker+ presents a configurable multi-layer hardware SNN, a library
of highly efficient neuron architectures, and a design framework, enabling the
development of complex neural network accelerators with few lines of Python
code. Spiker+ is tested on two benchmark datasets, the MNIST and the Spiking
Heidelberg Digits (SHD). On the MNIST, it demonstrates competitive performance
compared to state-of-the-art SNN accelerators. It outperforms them in terms of
resource allocation, with a requirement of 7,612 logic cells and 18 Block RAMs
(BRAMs), which makes it fit in very small FPGA, and power consumption, draining
only 180mW for a complete inference on an input image. The latency is
comparable to the ones observed in the state-of-the-art, with 780us/img. To the
authors' knowledge, Spiker+ is the first SNN accelerator tested on the SHD. In
this case, the accelerator requires 18,268 logic cells and 51 BRAM, with an
overall power consumption of 430mW and a latency of 54 us for a complete
inference on input data. This underscores the significance of Spiker+ in the
hardware-accelerated SNN landscape, making it an excellent solution to deploy
configurable and tunable SNN architectures in resource and power-constrained
edge applications.
Related papers
- RNC: Efficient RRAM-aware NAS and Compilation for DNNs on Resource-Constrained Edge Devices [0.30458577208819987]
We aim to develop edge-friendly deep neural networks (DNNs) for accelerators based on resistive random-access memory (RRAM)
We propose an edge compilation and resource-constrained RRAM-aware neural architecture search (NAS) framework to search for optimized neural networks meeting specific hardware constraints.
The resulting model from NAS optimized for speed achieved 5x-30x speedup.
arXiv Detail & Related papers (2024-09-27T15:35:36Z) - Embedded event based object detection with spiking neural network [0.18641315013048293]
This research introduces an embedded neuromorphic testbench that utilizes the SPiking Low-power Event-based ArchiTecture accelerator.
We use this testbench to load a state-of-the-art SNN solution, estimate the performance loss associated with deploying the network on dedicated hardware, and run real-world event-based OD on neuromorphic hardware.
arXiv Detail & Related papers (2024-06-25T15:02:01Z) - SpikeExplorer: hardware-oriented Design Space Exploration for Spiking Neural Networks on FPGA [42.170149806080204]
SpikExplorer is a Python tool for hardware-oriented Automatic Design Space Exploration.
It searches the optimal network architecture, neuron model, and internal and training parameters.
It reaches 95.8% accuracy on the MNIST dataset, with a power consumption of 180mW/image and a latency of 0.12 ms/image.
arXiv Detail & Related papers (2024-04-04T17:53:08Z) - RAMAN: A Re-configurable and Sparse tinyML Accelerator for Inference on
Edge [1.8293684411977293]
Deep Neural Network (DNN) based inference at the edge is challenging as these compute and data-intensive algorithms need to be implemented at low cost and low power.
We present RAMAN, a Re-configurable and spArse tinyML Accelerator for infereNce on edge, architected to exploit the sparsity to reduce area (storage), power as well as latency.
arXiv Detail & Related papers (2023-06-10T17:25:58Z) - Training High-Performance Low-Latency Spiking Neural Networks by
Differentiation on Spike Representation [70.75043144299168]
Spiking Neural Network (SNN) is a promising energy-efficient AI model when implemented on neuromorphic hardware.
It is a challenge to efficiently train SNNs due to their non-differentiability.
We propose the Differentiation on Spike Representation (DSR) method, which could achieve high performance.
arXiv Detail & Related papers (2022-05-01T12:44:49Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - An Adaptive Device-Edge Co-Inference Framework Based on Soft
Actor-Critic [72.35307086274912]
High-dimension parameter model and large-scale mathematical calculation restrict execution efficiency, especially for Internet of Things (IoT) devices.
We propose a new Deep Reinforcement Learning (DRL)-Soft Actor Critic for discrete (SAC-d), which generates the emphexit point, emphexit point, and emphcompressing bits by soft policy iterations.
Based on the latency and accuracy aware reward design, such an computation can well adapt to the complex environment like dynamic wireless channel and arbitrary processing, and is capable of supporting the 5G URL
arXiv Detail & Related papers (2022-01-09T09:31:50Z) - E3NE: An End-to-End Framework for Accelerating Spiking Neural Networks
with Emerging Neural Encoding on FPGAs [6.047137174639418]
End-to-end framework E3NE automates the generation of efficient SNN inference logic for FPGAs.
E3NE uses less than 50% of hardware resources and 20% less power, while reducing the latency by an order of magnitude.
arXiv Detail & Related papers (2021-11-19T04:01:19Z) - AdderNet and its Minimalist Hardware Design for Energy-Efficient
Artificial Intelligence [111.09105910265154]
We present a novel minimalist hardware architecture using adder convolutional neural network (AdderNet)
The whole AdderNet can practically achieve 16% enhancement in speed.
We conclude the AdderNet is able to surpass all the other competitors.
arXiv Detail & Related papers (2021-01-25T11:31:52Z) - PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with
Pattern-based Weight Pruning [57.20262984116752]
We introduce a new dimension, fine-grained pruning patterns inside the coarse-grained structures, revealing a previously unknown point in design space.
With the higher accuracy enabled by fine-grained pruning patterns, the unique insight is to use the compiler to re-gain and guarantee high hardware efficiency.
arXiv Detail & Related papers (2020-01-01T04:52:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.