SpikePoint: An Efficient Point-based Spiking Neural Network for Event
Cameras Action Recognition
- URL: http://arxiv.org/abs/2310.07189v2
- Date: Tue, 23 Jan 2024 08:20:05 GMT
- Title: SpikePoint: An Efficient Point-based Spiking Neural Network for Event
Cameras Action Recognition
- Authors: Hongwei Ren, Yue Zhou, Yulong Huang, Haotian Fu, Xiaopeng Lin, Jie
Song, Bojun Cheng
- Abstract summary: Spiking Neural Networks (SNNs) have gained significant attention due to their remarkable efficiency and fault tolerance.
We propose SpikePoint, a novel end-to-end point-based SNN architecture.
SpikePoint excels at processing sparse event cloud data, effectively extracting both global and local features.
- Score: 11.178792888084692
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Event cameras are bio-inspired sensors that respond to local changes in light
intensity and feature low latency, high energy efficiency, and high dynamic
range. Meanwhile, Spiking Neural Networks (SNNs) have gained significant
attention due to their remarkable efficiency and fault tolerance. By
synergistically harnessing the energy efficiency inherent in event cameras and
the spike-based processing capabilities of SNNs, their integration could enable
ultra-low-power application scenarios, such as action recognition tasks.
However, existing approaches often entail converting asynchronous events into
conventional frames, leading to additional data mapping efforts and a loss of
sparsity, contradicting the design concept of SNNs and event cameras. To
address this challenge, we propose SpikePoint, a novel end-to-end point-based
SNN architecture. SpikePoint excels at processing sparse event cloud data,
effectively extracting both global and local features through a singular-stage
structure. Leveraging the surrogate training method, SpikePoint achieves high
accuracy with few parameters and maintains low power consumption, specifically
employing the identity mapping feature extractor on diverse datasets.
SpikePoint achieves state-of-the-art (SOTA) performance on four event-based
action recognition datasets using only 16 timesteps, surpassing other SNN
methods. Moreover, it also achieves SOTA performance across all methods on
three datasets, utilizing approximately 0.3\% of the parameters and 0.5\% of
power consumption employed by artificial neural networks (ANNs). These results
emphasize the significance of Point Cloud and pave the way for many
ultra-low-power event-based data processing applications.
Related papers
- Point Cloud Understanding via Attention-Driven Contrastive Learning [64.65145700121442]
Transformer-based models have advanced point cloud understanding by leveraging self-attention mechanisms.
PointACL is an attention-driven contrastive learning framework designed to address these limitations.
Our method employs an attention-driven dynamic masking strategy that guides the model to focus on under-attended regions.
arXiv Detail & Related papers (2024-11-22T05:41:00Z) - Efficient Spatio-Temporal Signal Recognition on Edge Devices Using PointLCA-Net [0.45609532372046985]
This paper presents an approach that combines PointNet's feature extraction with the in-memory computing capabilities and energy efficiency of neuromorphic systems fortemporal signal recognition.
PointNet achieves high accuracy and significantly lower energy burden during both inference and training than comparable approaches.
arXiv Detail & Related papers (2024-11-21T20:48:40Z) - Spiking Neural Network as Adaptive Event Stream Slicer [10.279359105384334]
Event-based cameras provide rich edge information, high dynamic range, and high temporal resolution.
Many state-of-the-art event-based algorithms rely on splitting the events into fixed groups, resulting in the omission of crucial temporal information.
SpikeSlicer is a novel-designed plug-and-play event processing method capable of splitting events stream adaptively.
arXiv Detail & Related papers (2024-10-03T06:41:10Z) - PVAFN: Point-Voxel Attention Fusion Network with Multi-Pooling Enhancing for 3D Object Detection [59.355022416218624]
integration of point and voxel representations is becoming more common in LiDAR-based 3D object detection.
We propose a novel two-stage 3D object detector, called Point-Voxel Attention Fusion Network (PVAFN)
PVAFN uses a multi-pooling strategy to integrate both multi-scale and region-specific information effectively.
arXiv Detail & Related papers (2024-08-26T19:43:01Z) - A Novel Spike Transformer Network for Depth Estimation from Event Cameras via Cross-modality Knowledge Distillation [3.355813093377501]
Event cameras operate differently from traditional digital cameras, continuously capturing data and generating binary spikes that encode time, location, and light intensity.
This necessitates the development of innovative, spike-aware algorithms tailored for event cameras.
We propose a purely spike-driven spike transformer network for depth estimation from spiking camera data.
arXiv Detail & Related papers (2024-04-26T11:32:53Z) - SFOD: Spiking Fusion Object Detector [10.888008544975662]
Spiking Fusion Object Detector (SFOD) is a simple and efficient approach to SNN-based object detection.
We design a Spiking Fusion Module, achieving the first-time fusion of feature maps from different scales in SNNs applied to event cameras.
We establish state-of-the-art classification results based on SNNs, achieving 93.7% accuracy on the NCAR dataset.
arXiv Detail & Related papers (2024-03-22T13:24:50Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - Hybrid SNN-ANN: Energy-Efficient Classification and Object Detection for
Event-Based Vision [64.71260357476602]
Event-based vision sensors encode local pixel-wise brightness changes in streams of events rather than image frames.
Recent progress in object recognition from event-based sensors has come from conversions of deep neural networks.
We propose a hybrid architecture for end-to-end training of deep neural networks for event-based pattern recognition and object detection.
arXiv Detail & Related papers (2021-12-06T23:45:58Z) - Energy-Efficient Model Compression and Splitting for Collaborative
Inference Over Time-Varying Channels [52.60092598312894]
We propose a technique to reduce the total energy bill at the edge device by utilizing model compression and time-varying model split between the edge and remote nodes.
Our proposed solution results in minimal energy consumption and $CO$ emission compared to the considered baselines.
arXiv Detail & Related papers (2021-06-02T07:36:27Z) - Spike-FlowNet: Event-based Optical Flow Estimation with Energy-Efficient
Hybrid Neural Networks [40.44712305614071]
We present Spike-FlowNet, a deep hybrid neural network architecture integrating SNNs and ANNs for efficiently estimating optical flow from sparse event camera outputs.
The network is end-to-end trained with self-supervised learning on Multi-Vehicle Stereo Event Camera (MVSEC) dataset.
arXiv Detail & Related papers (2020-03-14T20:37:21Z) - Real-Time High-Performance Semantic Image Segmentation of Urban Street
Scenes [98.65457534223539]
We propose a real-time high-performance DCNN-based method for robust semantic segmentation of urban street scenes.
The proposed method achieves the accuracy of 73.6% and 68.0% mean Intersection over Union (mIoU) with the inference speed of 51.0 fps and 39.3 fps.
arXiv Detail & Related papers (2020-03-11T08:45:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.