Hand Gesture Recognition Using Temporal Convolutions and Attention
Mechanism
- URL: http://arxiv.org/abs/2110.08717v1
- Date: Sun, 17 Oct 2021 04:23:59 GMT
- Title: Hand Gesture Recognition Using Temporal Convolutions and Attention
Mechanism
- Authors: Elahe Rahimian, Soheil Zabihi, Amir Asif, Dario Farina, S. Farokh
Atashzar, Arash Mohammadi
- Abstract summary: We propose the novel Temporal Convolutions-based Hand Gesture Recognition architecture (TC-HGR) to reduce this computational burden.
We classified 17 hand gestures via surface Electromyogram (sEMG) signals by the adoption of attention mechanisms and temporal convolutions.
The proposed method led to 81.65% and 80.72% classification accuracy for window sizes of 300ms and 200ms, respectively.
- Score: 16.399230849853915
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Advances in biosignal signal processing and machine learning, in particular
Deep Neural Networks (DNNs), have paved the way for the development of
innovative Human-Machine Interfaces for decoding the human intent and
controlling artificial limbs. DNN models have shown promising results with
respect to other algorithms for decoding muscle electrical activity, especially
for recognition of hand gestures. Such data-driven models, however, have been
challenged by their need for a large number of trainable parameters and their
structural complexity. Here we propose the novel Temporal Convolutions-based
Hand Gesture Recognition architecture (TC-HGR) to reduce this computational
burden. With this approach, we classified 17 hand gestures via surface
Electromyogram (sEMG) signals by the adoption of attention mechanisms and
temporal convolutions. The proposed method led to 81.65% and 80.72%
classification accuracy for window sizes of 300ms and 200ms, respectively. The
number of parameters to train the proposed TC-HGR architecture is 11.9 times
less than that of its state-of-the-art counterpart.
Related papers
- The Role of Functional Muscle Networks in Improving Hand Gesture Perception for Human-Machine Interfaces [2.367412330421982]
Surface electromyography (sEMG) has been explored for its rich informational context and accessibility.
This paper proposes the decoding of muscle synchronization rather than individual muscle activation.
It achieves an accuracy of 85.1%, demonstrating improved performance compared to existing methods.
arXiv Detail & Related papers (2024-08-05T15:17:34Z) - A Deep Learning Sequential Decoder for Transient High-Density
Electromyography in Hand Gesture Recognition Using Subject-Embedded Transfer
Learning [11.170031300110315]
Hand gesture recognition (HGR) has gained significant attention due to the increasing use of AI-powered human-computers.
These interfaces have a range of applications, including the control of extended reality, agile prosthetics, and exoskeletons.
These interfaces have a range of applications, including the control of extended reality, agile prosthetics, and exoskeletons.
arXiv Detail & Related papers (2023-09-23T05:32:33Z) - Agile gesture recognition for capacitive sensing devices: adapting
on-the-job [55.40855017016652]
We demonstrate a hand gesture recognition system that uses signals from capacitive sensors embedded into the etee hand controller.
The controller generates real-time signals from each of the wearer five fingers.
We use a machine learning technique to analyse the time series signals and identify three features that can represent 5 fingers within 500 ms.
arXiv Detail & Related papers (2023-05-12T17:24:02Z) - Light-weighted CNN-Attention based architecture for Hand Gesture
Recognition via ElectroMyography [19.51045409936039]
We propose a light-weighted hybrid architecture (HDCAM) based on Convolutional Neural Network (CNN) and attention mechanism.
The proposed HDCAM model with 58,441 parameters reached a new state-of-the-art (SOTA) performance with 82.91% and 81.28% accuracy on window sizes of 300 ms and 200 ms for classifying 17 hand gestures.
arXiv Detail & Related papers (2022-10-27T02:12:07Z) - ViT-HGR: Vision Transformer-based Hand Gesture Recognition from High
Density Surface EMG Signals [14.419091034872682]
We investigate and design a Vision Transformer (ViT) based architecture to perform hand gesture recognition from High Density (HD-sEMG) signals.
The proposed ViT-HGR framework can overcome the training time problems and can accurately classify a large number of hand gestures from scratch.
Our experiments with 64-sample (31.25 ms) window size yield average test accuracy of 84.62 +/- 3.07%, where only 78, 210 number of parameters is utilized.
arXiv Detail & Related papers (2022-01-25T02:42:50Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - DriPP: Driven Point Processes to Model Stimuli Induced Patterns in M/EEG
Signals [62.997667081978825]
We develop a novel statistical point process model-called driven temporal point processes (DriPP)
We derive a fast and principled expectation-maximization (EM) algorithm to estimate the parameters of this model.
Results on standard MEG datasets demonstrate that our methodology reveals event-related neural responses.
arXiv Detail & Related papers (2021-12-08T13:07:21Z) - TEMGNet: Deep Transformer-based Decoding of Upperlimb sEMG for Hand
Gestures Recognition [16.399230849853915]
We develop a framework based on the Transformer architecture for processing sEMG signals.
We propose a novel Vision Transformer (ViT)-based neural network architecture (referred to as the TEMGNet) to classify and recognize upperlimb hand gestures.
arXiv Detail & Related papers (2021-09-25T15:03:22Z) - Domain Adaptive Robotic Gesture Recognition with Unsupervised
Kinematic-Visual Data Alignment [60.31418655784291]
We propose a novel unsupervised domain adaptation framework which can simultaneously transfer multi-modality knowledge, i.e., both kinematic and visual data, from simulator to real robot.
It remedies the domain gap with enhanced transferable features by using temporal cues in videos, and inherent correlations in multi-modal towards recognizing gesture.
Results show that our approach recovers the performance with great improvement gains, up to 12.91% in ACC and 20.16% in F1score without using any annotations in real robot.
arXiv Detail & Related papers (2021-03-06T09:10:03Z) - Progressive Spatio-Temporal Graph Convolutional Network for
Skeleton-Based Human Action Recognition [97.14064057840089]
We propose a method to automatically find a compact and problem-specific network for graph convolutional networks in a progressive manner.
Experimental results on two datasets for skeleton-based human action recognition indicate that the proposed method has competitive or even better classification performance.
arXiv Detail & Related papers (2020-11-11T09:57:49Z) - Temporal Attention-Augmented Graph Convolutional Network for Efficient
Skeleton-Based Human Action Recognition [97.14064057840089]
Graphal networks (GCNs) have been very successful in modeling non-Euclidean data structures.
Most GCN-based action recognition methods use deep feed-forward networks with high computational complexity to process all skeletons in an action.
We propose a temporal attention module (TAM) for increasing the efficiency in skeleton-based action recognition.
arXiv Detail & Related papers (2020-10-23T08:01:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.