S3TC: Spiking Separated Spatial and Temporal Convolutions with
Unsupervised STDP-based Learning for Action Recognition
- URL: http://arxiv.org/abs/2309.12761v1
- Date: Fri, 22 Sep 2023 10:05:35 GMT
- Title: S3TC: Spiking Separated Spatial and Temporal Convolutions with
Unsupervised STDP-based Learning for Action Recognition
- Authors: Mireille El-Assal and Pierre Tirilly and Ioan Marius Bilasco
- Abstract summary: Spiking Neural Networks (SNNs) have significantly lower computational costs (thousands of times) than regular non-spiking networks when implemented on neuromorphic hardware.
We introduce, for the first time, Spiking Separated Spatial and Temporal Convolutions (S3TCs) for the sake of reducing the number of parameters required for video analysis.
- Score: 1.2123876307427106
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video analysis is a major computer vision task that has received a lot of
attention in recent years. The current state-of-the-art performance for video
analysis is achieved with Deep Neural Networks (DNNs) that have high
computational costs and need large amounts of labeled data for training.
Spiking Neural Networks (SNNs) have significantly lower computational costs
(thousands of times) than regular non-spiking networks when implemented on
neuromorphic hardware. They have been used for video analysis with methods like
3D Convolutional Spiking Neural Networks (3D CSNNs). However, these networks
have a significantly larger number of parameters compared with spiking 2D CSNN.
This, not only increases the computational costs, but also makes these networks
more difficult to implement with neuromorphic hardware. In this work, we use
CSNNs trained in an unsupervised manner with the Spike Timing-Dependent
Plasticity (STDP) rule, and we introduce, for the first time, Spiking Separated
Spatial and Temporal Convolutions (S3TCs) for the sake of reducing the number
of parameters required for video analysis. This unsupervised learning has the
advantage of not needing large amounts of labeled data for training.
Factorizing a single spatio-temporal spiking convolution into a spatial and a
temporal spiking convolution decreases the number of parameters of the network.
We test our network with the KTH, Weizmann, and IXMAS datasets, and we show
that S3TCs successfully extract spatio-temporal information from videos, while
increasing the output spiking activity, and outperforming spiking 3D
convolutions.
Related papers
- Transferability of Convolutional Neural Networks in Stationary Learning
Tasks [96.00428692404354]
We introduce a novel framework for efficient training of convolutional neural networks (CNNs) for large-scale spatial problems.
We show that a CNN trained on small windows of such signals achieves a nearly performance on much larger windows without retraining.
Our results show that the CNN is able to tackle problems with many hundreds of agents after being trained with fewer than ten.
arXiv Detail & Related papers (2023-07-21T13:51:45Z) - Learning Delays in Spiking Neural Networks using Dilated Convolutions
with Learnable Spacings [1.534667887016089]
Spiking Neural Networks (SNNs) are promising research direction for building power-efficient information processing systems.
In SNNs, delays refer to the time needed for one spike to travel from one neuron to another.
It has been shown theoretically that plastic delays greatly increase the expressivity in SNNs.
We propose a new discrete-time algorithm that addresses this issue in deep feedforward SNNs using backpropagation.
arXiv Detail & Related papers (2023-06-30T14:01:53Z) - Spiking Two-Stream Methods with Unsupervised STDP-based Learning for
Action Recognition [1.9981375888949475]
Deep Convolutional Neural Networks (CNNs) are currently the state-of-the-art methods for video analysis.
We use Convolutional Spiking Neural Networks (CSNNs) trained with the unsupervised Spike Timing-Dependent Plasticity (STDP) rule for action classification.
We show that two-stream CSNNs can successfully extract information from videos despite using limited training data.
arXiv Detail & Related papers (2023-06-23T20:54:44Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - NAF: Neural Attenuation Fields for Sparse-View CBCT Reconstruction [79.13750275141139]
This paper proposes a novel and fast self-supervised solution for sparse-view CBCT reconstruction.
The desired attenuation coefficients are represented as a continuous function of 3D spatial coordinates, parameterized by a fully-connected deep neural network.
A learning-based encoder entailing hash coding is adopted to help the network capture high-frequency details.
arXiv Detail & Related papers (2022-09-29T04:06:00Z) - 2D versus 3D Convolutional Spiking Neural Networks Trained with
Unsupervised STDP for Human Action Recognition [1.9981375888949475]
Spiking neural networks (SNNs) are third generation biologically plausible models that process the information in the form of spikes.
Unsupervised learning with SNNs using the spike timing dependent plasticity (STDP) rule has the potential to overcome some bottlenecks.
We show that STDP-based convolutional SNNs can learn motion patterns using 3D kernels, thus enabling motion-based recognition from videos.
arXiv Detail & Related papers (2022-05-26T16:34:22Z) - SpikeMS: Deep Spiking Neural Network for Motion Segmentation [7.491944503744111]
textitSpikeMS is the first deep encoder-decoder SNN architecture for the real-world large-scale problem of motion segmentation.
We show that textitSpikeMS is capable of textitincremental predictions, or predictions from smaller amounts of test data than it is trained on.
arXiv Detail & Related papers (2021-05-13T21:34:55Z) - RANP: Resource Aware Neuron Pruning at Initialization for 3D CNNs [32.054160078692036]
We introduce a Resource Aware Neuron Pruning (RANP) algorithm that prunes 3D CNNs to high sparsity levels.
Our algorithm leads to roughly 50%-95% reduction in FLOPs and 35%-80% reduction in memory with negligible loss in accuracy compared to the unpruned networks.
arXiv Detail & Related papers (2021-02-09T04:35:29Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Tensor train decompositions on recurrent networks [60.334946204107446]
Matrix product state (MPS) tensor trains have more attractive features than MPOs, in terms of storage reduction and computing time at inference.
We show that MPS tensor trains should be at the forefront of LSTM network compression through a theoretical analysis and practical experiments on NLP task.
arXiv Detail & Related papers (2020-06-09T18:25:39Z) - Event-Based Angular Velocity Regression with Spiking Networks [51.145071093099396]
Spiking Neural Networks (SNNs) process information conveyed as temporal spikes rather than numeric values.
We propose, for the first time, a temporal regression problem of numerical values given events from an event camera.
We show that we can successfully train an SNN to perform angular velocity regression.
arXiv Detail & Related papers (2020-03-05T17:37:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.