Gesture Recognition from Skeleton Data for Intuitive Human-Machine
Interaction
- URL: http://arxiv.org/abs/2008.11497v1
- Date: Wed, 26 Aug 2020 11:28:50 GMT
- Title: Gesture Recognition from Skeleton Data for Intuitive Human-Machine
Interaction
- Authors: Andr\'e Br\'as, Miguel Sim\~ao, Pedro Neto
- Abstract summary: We propose an approach for segmentation and classification of dynamic gestures based on a set of handcrafted features.
The method for gesture recognition applies a sliding window, which extracts information from both the spatial and temporal dimensions.
At the end, the recognized gestures are used to interact with a collaborative robot.
- Score: 0.6875312133832077
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human gesture recognition has assumed a capital role in industrial
applications, such as Human-Machine Interaction. We propose an approach for
segmentation and classification of dynamic gestures based on a set of
handcrafted features, which are drawn from the skeleton data provided by the
Kinect sensor. The module for gesture detection relies on a feedforward neural
network which performs framewise binary classification. The method for gesture
recognition applies a sliding window, which extracts information from both the
spatial and temporal dimensions. Then we combine windows of varying durations
to get a multi-temporal scale approach and an additional gain in performance.
Encouraged by the recent success of Recurrent Neural Networks for time series
domains, we also propose a method for simultaneous gesture segmentation and
classification based on the bidirectional Long Short-Term Memory cells, which
have shown ability for learning the temporal relationships on long temporal
scales. We evaluate all the different approaches on the dataset published for
the ChaLearn Looking at People Challenge 2014. The most effective method
achieves a Jaccard index of 0.75, which suggests a performance almost on pair
with that presented by the state-of-the-art techniques. At the end, the
recognized gestures are used to interact with a collaborative robot.
Related papers
- Long-Distance Gesture Recognition using Dynamic Neural Networks [14.106548659369716]
We propose a novel, accurate and efficient method for the recognition of gestures from longer distances.
It uses a dynamic neural network to select features from gesture-containing spatial regions of the input sensor data for further processing.
We demonstrate the performance of our method on the LD-ConGR long-distance dataset.
arXiv Detail & Related papers (2023-08-09T00:56:38Z) - Learning from Temporal Spatial Cubism for Cross-Dataset Skeleton-based
Action Recognition [88.34182299496074]
Action labels are only available on a source dataset, but unavailable on a target dataset in the training stage.
We utilize a self-supervision scheme to reduce the domain shift between two skeleton-based action datasets.
By segmenting and permuting temporal segments or human body parts, we design two self-supervised learning classification tasks.
arXiv Detail & Related papers (2022-07-17T07:05:39Z) - A Spatio-Temporal Multilayer Perceptron for Gesture Recognition [70.34489104710366]
We propose a multilayer state-weighted perceptron for gesture recognition in the context of autonomous vehicles.
An evaluation of TCG and Drive&Act datasets is provided to showcase the promising performance of our approach.
We deploy our model to our autonomous vehicle to show its real-time capability and stable execution.
arXiv Detail & Related papers (2022-04-25T08:42:47Z) - Joint-bone Fusion Graph Convolutional Network for Semi-supervised
Skeleton Action Recognition [65.78703941973183]
We propose a novel correlation-driven joint-bone fusion graph convolutional network (CD-JBF-GCN) as an encoder and use a pose prediction head as a decoder.
Specifically, the CD-JBF-GC can explore the motion transmission between the joint stream and the bone stream.
The pose prediction based auto-encoder in the self-supervised training stage allows the network to learn motion representation from unlabeled data.
arXiv Detail & Related papers (2022-02-08T16:03:15Z) - Metric-based multimodal meta-learning for human movement identification
via footstep recognition [3.300376360949452]
We describe a novel metric-based learning approach that introduces a multimodal framework.
We learn general-purpose representations from low multisensory data obtained from omnipresent sensing systems.
Our results employ a metric-based contrastive learning approach for multi-sensor data to mitigate the impact of data scarcity.
arXiv Detail & Related papers (2021-11-15T18:46:14Z) - Efficient Modelling Across Time of Human Actions and Interactions [92.39082696657874]
We argue that current fixed-sized-temporal kernels in 3 convolutional neural networks (CNNDs) can be improved to better deal with temporal variations in the input.
We study how we can better handle between classes of actions, by enhancing their feature differences over different layers of the architecture.
The proposed approaches are evaluated on several benchmark action recognition datasets and show competitive results.
arXiv Detail & Related papers (2021-10-05T15:39:11Z) - Learning Multi-Granular Spatio-Temporal Graph Network for Skeleton-based
Action Recognition [49.163326827954656]
We propose a novel multi-granular-temporal graph network for skeleton-based action classification.
We develop a dual-head graph network consisting of two inter-leaved branches, which enables us to extract at least two-temporal resolutions.
We conduct extensive experiments on three large-scale datasets.
arXiv Detail & Related papers (2021-08-10T09:25:07Z) - SHREC 2021: Track on Skeleton-based Hand Gesture Recognition in the Wild [62.450907796261646]
Recognition of hand gestures can be performed directly from the stream of hand skeletons estimated by software.
Despite the recent advancements in gesture and action recognition from skeletons, it is unclear how well the current state-of-the-art techniques can perform in a real-world scenario.
This paper presents the results of the SHREC 2021: Track on Skeleton-based Hand Gesture Recognition in the Wild contest.
arXiv Detail & Related papers (2021-06-21T10:57:49Z) - Memory Group Sampling Based Online Action Recognition Using Kinetic
Skeleton Features [4.674689979981502]
We propose two core ideas to handle the online action recognition problem.
First, we combine the spatial and temporal skeleton features to depict the actions.
Second, we propose a memory group sampling method to combine the previous action frames and current action frames.
Third, an improved 1D CNN network is employed for training and testing using the features from sampled frames.
arXiv Detail & Related papers (2020-11-01T16:43:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.