Spiking-Fer: Spiking Neural Network for Facial Expression Recognition
With Event Cameras
- URL: http://arxiv.org/abs/2304.10211v1
- Date: Thu, 20 Apr 2023 10:59:56 GMT
- Title: Spiking-Fer: Spiking Neural Network for Facial Expression Recognition
With Event Cameras
- Authors: Sami Barchid, Benjamin Allaert, Amel Aissaoui, Jos\'e Mennesson,
Chaabane Dj\'eraba
- Abstract summary: "Spiking-FER" is a deep convolutional SNN model, and compare it against a similar Artificial Neural Network (ANN)
Experiments show that the proposed approach achieves comparable performance to the ANN architecture, while consuming less energy by orders of magnitude (up to 65.39x)
- Score: 2.9398911304923447
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Facial Expression Recognition (FER) is an active research domain that has
shown great progress recently, notably thanks to the use of large deep learning
models. However, such approaches are particularly energy intensive, which makes
their deployment difficult for edge devices. To address this issue, Spiking
Neural Networks (SNNs) coupled with event cameras are a promising alternative,
capable of processing sparse and asynchronous events with lower energy
consumption. In this paper, we establish the first use of event cameras for
FER, named "Event-based FER", and propose the first related benchmarks by
converting popular video FER datasets to event streams. To deal with this new
task, we propose "Spiking-FER", a deep convolutional SNN model, and compare it
against a similar Artificial Neural Network (ANN). Experiments show that the
proposed approach achieves comparable performance to the ANN architecture,
while consuming less energy by orders of magnitude (up to 65.39x). In addition,
an experimental study of various event-based data augmentation techniques is
performed to provide insights into the efficient transformations specific to
event-based FER.
Related papers
- Spiking Neural Network as Adaptive Event Stream Slicer [10.279359105384334]
Event-based cameras provide rich edge information, high dynamic range, and high temporal resolution.
Many state-of-the-art event-based algorithms rely on splitting the events into fixed groups, resulting in the omission of crucial temporal information.
SpikeSlicer is a novel-designed plug-and-play event processing method capable of splitting events stream adaptively.
arXiv Detail & Related papers (2024-10-03T06:41:10Z) - Spike-EVPR: Deep Spiking Residual Network with Cross-Representation
Aggregation for Event-Based Visual Place Recognition [4.357768397230497]
Event cameras have been successfully applied to visual place recognition (VPR) tasks by using deep artificial neural networks (ANNs)
We propose a novel deep spiking network architecture called Spike-EVPR for event-based VPR tasks.
To address the aforementioned issues, we propose a novel deep spiking network architecture called Spike-EVPR for event-based VPR tasks.
arXiv Detail & Related papers (2024-02-16T06:45:25Z) - EventTransAct: A video transformer-based framework for Event-camera
based action recognition [52.537021302246664]
Event cameras offer new opportunities compared to standard action recognition in RGB videos.
In this study, we employ a computationally efficient model, namely the video transformer network (VTN), which initially acquires spatial embeddings per event-frame.
In order to better adopt the VTN for the sparse and fine-grained nature of event data, we design Event-Contrastive Loss ($mathcalL_EC$) and event-specific augmentations.
arXiv Detail & Related papers (2023-08-25T23:51:07Z) - Deep Multi-Threshold Spiking-UNet for Image Processing [51.88730892920031]
This paper introduces the novel concept of Spiking-UNet for image processing, which combines the power of Spiking Neural Networks (SNNs) with the U-Net architecture.
To achieve an efficient Spiking-UNet, we face two primary challenges: ensuring high-fidelity information propagation through the network via spikes and formulating an effective training strategy.
Experimental results show that, on image segmentation and denoising, our Spiking-UNet achieves comparable performance to its non-spiking counterpart.
arXiv Detail & Related papers (2023-07-20T16:00:19Z) - Best of Both Worlds: Hybrid SNN-ANN Architecture for Event-based Optical Flow Estimation [12.611797572621398]
Spiking Neural Networks (SNNs) with their asynchronous event-driven compute show great potential for extracting features from event streams.
We propose a novel SNN-ANN hybrid architecture that combines the strengths of both.
arXiv Detail & Related papers (2023-06-05T15:26:02Z) - MSS-DepthNet: Depth Prediction with Multi-Step Spiking Neural Network [8.53512216864715]
Spiking neural network is a novel event-based computational paradigm that is considered to be well suited for processing event camera tasks.
This work proposes a spiking neural network architecture with a novel residual block designed and multi-dimension attention modules combined.
This model outperforms previous ANN networks of the same size on the MVSEC dataset and shows great computational efficiency.
arXiv Detail & Related papers (2022-11-22T10:35:36Z) - Hybrid SNN-ANN: Energy-Efficient Classification and Object Detection for
Event-Based Vision [64.71260357476602]
Event-based vision sensors encode local pixel-wise brightness changes in streams of events rather than image frames.
Recent progress in object recognition from event-based sensors has come from conversions of deep neural networks.
We propose a hybrid architecture for end-to-end training of deep neural networks for event-based pattern recognition and object detection.
arXiv Detail & Related papers (2021-12-06T23:45:58Z) - Video-based Facial Expression Recognition using Graph Convolutional
Networks [57.980827038988735]
We introduce a Graph Convolutional Network (GCN) layer into a common CNN-RNN based model for video-based facial expression recognition.
We evaluate our method on three widely-used datasets, CK+, Oulu-CASIA and MMI, and also one challenging wild dataset AFEW8.0.
arXiv Detail & Related papers (2020-10-26T07:31:51Z) - The FaceChannel: A Fast & Furious Deep Neural Network for Facial
Expression Recognition [71.24825724518847]
Current state-of-the-art models for automatic Facial Expression Recognition (FER) are based on very deep neural networks that are effective but rather expensive to train.
We formalize the FaceChannel, a light-weight neural network that has much fewer parameters than common deep neural networks.
We demonstrate how our model achieves a comparable, if not better, performance to the current state-of-the-art in FER.
arXiv Detail & Related papers (2020-09-15T09:25:37Z) - The FaceChannel: A Light-weight Deep Neural Network for Facial
Expression Recognition [71.24825724518847]
Current state-of-the-art models for automatic FER are based on very deep neural networks that are difficult to train.
We formalize the FaceChannel, a light-weight neural network that has much fewer parameters than common deep neural networks.
We demonstrate how the FaceChannel achieves a comparable, if not better, performance, as compared to the current state-of-the-art in FER.
arXiv Detail & Related papers (2020-04-17T12:03:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.