Spiking Two-Stream Methods with Unsupervised STDP-based Learning for
Action Recognition
- URL: http://arxiv.org/abs/2306.13783v1
- Date: Fri, 23 Jun 2023 20:54:44 GMT
- Title: Spiking Two-Stream Methods with Unsupervised STDP-based Learning for
Action Recognition
- Authors: Mireille El-Assal and Pierre Tirilly and Ioan Marius Bilasco
- Abstract summary: Deep Convolutional Neural Networks (CNNs) are currently the state-of-the-art methods for video analysis.
We use Convolutional Spiking Neural Networks (CSNNs) trained with the unsupervised Spike Timing-Dependent Plasticity (STDP) rule for action classification.
We show that two-stream CSNNs can successfully extract information from videos despite using limited training data.
- Score: 1.9981375888949475
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Video analysis is a computer vision task that is useful for many applications
like surveillance, human-machine interaction, and autonomous vehicles. Deep
Convolutional Neural Networks (CNNs) are currently the state-of-the-art methods
for video analysis. However they have high computational costs, and need a
large amount of labeled data for training. In this paper, we use Convolutional
Spiking Neural Networks (CSNNs) trained with the unsupervised Spike
Timing-Dependent Plasticity (STDP) learning rule for action classification.
These networks represent the information using asynchronous low-energy spikes.
This allows the network to be more energy efficient and neuromorphic
hardware-friendly. However, the behaviour of CSNNs is not studied enough with
spatio-temporal computer vision models. Therefore, we explore transposing
two-stream neural networks into the spiking domain. Implementing this model
with unsupervised STDP-based CSNNs allows us to further study the performance
of these networks with video analysis. In this work, we show that two-stream
CSNNs can successfully extract spatio-temporal information from videos despite
using limited training data, and that the spiking spatial and temporal streams
are complementary. We also show that using a spatio-temporal stream within a
spiking STDP-based two-stream architecture leads to information redundancy and
does not improve the performance.
Related papers
- CNN2GNN: How to Bridge CNN with GNN [59.42117676779735]
We propose a novel CNN2GNN framework to unify CNN and GNN together via distillation.
The performance of distilled boosted'' two-layer GNN on Mini-ImageNet is much higher than CNN containing dozens of layers such as ResNet152.
arXiv Detail & Related papers (2024-04-23T08:19:08Z) - TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - S3TC: Spiking Separated Spatial and Temporal Convolutions with
Unsupervised STDP-based Learning for Action Recognition [1.2123876307427106]
Spiking Neural Networks (SNNs) have significantly lower computational costs (thousands of times) than regular non-spiking networks when implemented on neuromorphic hardware.
We introduce, for the first time, Spiking Separated Spatial and Temporal Convolutions (S3TCs) for the sake of reducing the number of parameters required for video analysis.
arXiv Detail & Related papers (2023-09-22T10:05:35Z) - Transferability of Convolutional Neural Networks in Stationary Learning
Tasks [96.00428692404354]
We introduce a novel framework for efficient training of convolutional neural networks (CNNs) for large-scale spatial problems.
We show that a CNN trained on small windows of such signals achieves a nearly performance on much larger windows without retraining.
Our results show that the CNN is able to tackle problems with many hundreds of agents after being trained with fewer than ten.
arXiv Detail & Related papers (2023-07-21T13:51:45Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Backdoor Attack Detection in Computer Vision by Applying Matrix
Factorization on the Weights of Deep Networks [6.44397009982949]
We introduce a novel method for backdoor detection that extracts features from pre-trained DNN's weights.
In comparison to other detection techniques, this has a number of benefits, such as not requiring any training data.
Our method outperforms the competing algorithms in terms of efficiency and is more accurate, helping to ensure the safe application of deep learning and AI.
arXiv Detail & Related papers (2022-12-15T20:20:18Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - 2D versus 3D Convolutional Spiking Neural Networks Trained with
Unsupervised STDP for Human Action Recognition [1.9981375888949475]
Spiking neural networks (SNNs) are third generation biologically plausible models that process the information in the form of spikes.
Unsupervised learning with SNNs using the spike timing dependent plasticity (STDP) rule has the potential to overcome some bottlenecks.
We show that STDP-based convolutional SNNs can learn motion patterns using 3D kernels, thus enabling motion-based recognition from videos.
arXiv Detail & Related papers (2022-05-26T16:34:22Z) - A Study On the Effects of Pre-processing On Spatio-temporal Action
Recognition Using Spiking Neural Networks Trained with STDP [0.0]
It is important to study the behavior of SNNs trained with unsupervised learning methods on video classification tasks.
This paper presents methods of transposing temporal information into a static format, and then transforming the visual information into spikes using latency coding.
We show the effect of the similarity in the shape and speed of certain actions on action recognition with spiking neural networks.
arXiv Detail & Related papers (2021-05-31T07:07:48Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z) - 1D CNN Based Network Intrusion Detection with Normalization on
Imbalanced Data [0.19336815376402716]
Intrusion detection system (IDS) plays an essential role in computer networks protecting computing resources and data from outside attacks.
Recent IDS faces challenges improving flexibility and efficiency of the IDS for unexpected and unpredictable attacks.
We propose a deep learning approach for developing the efficient and flexible IDS using one-dimensional Convolutional Neural Network (1D-CNN)
arXiv Detail & Related papers (2020-03-01T12:23:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.