Towards Domain-Independent and Real-Time Gesture Recognition Using
mmWave Signal
- URL: http://arxiv.org/abs/2111.06195v1
- Date: Thu, 11 Nov 2021 13:28:28 GMT
- Title: Towards Domain-Independent and Real-Time Gesture Recognition Using
mmWave Signal
- Authors: Yadong Li, Dongheng Zhang, Jinbo Chen, Jinwei Wan, Dong Zhang, Yang
Hu, Qibin Sun, Yan Chen
- Abstract summary: DI-Gesture is a domain-independent and real-time mmWave gesture recognition system.
In real-time scenario, the accuracy of DI-Gesutre reaches over 97% with average inference time of 2.87ms.
- Score: 11.76969975145963
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human gesture recognition using millimeter wave (mmWave) signals provides
attractive applications including smart home and in-car interface. While
existing works achieve promising performance under controlled settings,
practical applications are still limited due to the need of intensive data
collection, extra training efforts when adapting to new domains (i.e.
environments, persons and locations) and poor performance for real-time
recognition. In this paper, we propose DI-Gesture, a domain-independent and
real-time mmWave gesture recognition system. Specifically, we first derive the
signal variation corresponding to human gestures with spatial-temporal
processing. To enhance the robustness of the system and reduce data collecting
efforts, we design a data augmentation framework based on the correlation
between signal patterns and gesture variations. Furthermore, we propose a
dynamic window mechanism to perform gesture segmentation automatically and
accurately, thus enable real-time recognition. Finally, we build a lightweight
neural network to extract spatial-temporal information from the data for
gesture classification. Extensive experimental results show DI-Gesture achieves
an average accuracy of 97.92%, 99.18% and 98.76% for new users, environments
and locations, respectively. In real-time scenario, the accuracy of DI-Gesutre
reaches over 97% with average inference time of 2.87ms, which demonstrates the
superior robustness and effectiveness of our system.
Related papers
- An LSTM Feature Imitation Network for Hand Movement Recognition from sEMG Signals [2.632402517354116]
We propose utilizing a feature-imitating network (FIN) for closed-form temporal feature learning over a 300ms signal window on Ninapro DB2.
We then explore transfer learning capabilities by applying the pre-trained LSTM-FIN for tuning to a downstream hand movement recognition task.
arXiv Detail & Related papers (2024-05-23T21:45:15Z) - Robust Collaborative Perception without External Localization and Clock Devices [52.32342059286222]
A consistent spatial-temporal coordination across multiple agents is fundamental for collaborative perception.
Traditional methods depend on external devices to provide localization and clock signals.
We propose a novel approach: aligning by recognizing the inherent geometric patterns within the perceptual data of various agents.
arXiv Detail & Related papers (2024-05-05T15:20:36Z) - Long-Distance Gesture Recognition using Dynamic Neural Networks [14.106548659369716]
We propose a novel, accurate and efficient method for the recognition of gestures from longer distances.
It uses a dynamic neural network to select features from gesture-containing spatial regions of the input sensor data for further processing.
We demonstrate the performance of our method on the LD-ConGR long-distance dataset.
arXiv Detail & Related papers (2023-08-09T00:56:38Z) - Hand gesture recognition using 802.11ad mmWave sensor in the mobile
device [2.5476515662939563]
We explore the feasibility of AI assisted hand-gesture recognition using 802.11ad 60GHz (mmWave) technology in smartphones.
We built a prototype system, where radar sensing and communication waveform can coexist by time-division duplex (TDD)
It can gather sensing data and predict gestures within 100 milliseconds.
arXiv Detail & Related papers (2022-11-14T03:36:17Z) - Adaptive Local-Component-aware Graph Convolutional Network for One-shot
Skeleton-based Action Recognition [54.23513799338309]
We present an Adaptive Local-Component-aware Graph Convolutional Network for skeleton-based action recognition.
Our method provides a stronger representation than the global embedding and helps our model reach state-of-the-art.
arXiv Detail & Related papers (2022-09-21T02:33:07Z) - Deep Impulse Responses: Estimating and Parameterizing Filters with Deep
Networks [76.830358429947]
Impulse response estimation in high noise and in-the-wild settings is a challenging problem.
We propose a novel framework for parameterizing and estimating impulse responses based on recent advances in neural representation learning.
arXiv Detail & Related papers (2022-02-07T18:57:23Z) - Improved Speech Emotion Recognition using Transfer Learning and
Spectrogram Augmentation [56.264157127549446]
Speech emotion recognition (SER) is a challenging task that plays a crucial role in natural human-computer interaction.
One of the main challenges in SER is data scarcity.
We propose a transfer learning strategy combined with spectrogram augmentation.
arXiv Detail & Related papers (2021-08-05T10:39:39Z) - Real-Time Activity Recognition and Intention Recognition Using a
Vision-based Embedded System [4.060731229044571]
We introduce a real-time activity recognition to recognize people's intentions to pass or not pass a door.
This system, if applied in elevators and automatic doors will save energy and increase efficiency.
Our embedded system was implemented with an accuracy of 98.78% on our Intention Recognition dataset.
arXiv Detail & Related papers (2021-07-27T11:38:44Z) - Domain Adaptive Robotic Gesture Recognition with Unsupervised
Kinematic-Visual Data Alignment [60.31418655784291]
We propose a novel unsupervised domain adaptation framework which can simultaneously transfer multi-modality knowledge, i.e., both kinematic and visual data, from simulator to real robot.
It remedies the domain gap with enhanced transferable features by using temporal cues in videos, and inherent correlations in multi-modal towards recognizing gesture.
Results show that our approach recovers the performance with great improvement gains, up to 12.91% in ACC and 20.16% in F1score without using any annotations in real robot.
arXiv Detail & Related papers (2021-03-06T09:10:03Z) - Gesture Recognition from Skeleton Data for Intuitive Human-Machine
Interaction [0.6875312133832077]
We propose an approach for segmentation and classification of dynamic gestures based on a set of handcrafted features.
The method for gesture recognition applies a sliding window, which extracts information from both the spatial and temporal dimensions.
At the end, the recognized gestures are used to interact with a collaborative robot.
arXiv Detail & Related papers (2020-08-26T11:28:50Z) - A Generative Learning Approach for Spatio-temporal Modeling in Connected
Vehicular Network [55.852401381113786]
This paper proposes LaMI (Latency Model Inpainting), a novel framework to generate a comprehensive-temporal quality framework for wireless access latency of connected vehicles.
LaMI adopts the idea from image inpainting and synthesizing and can reconstruct the missing latency samples by a two-step procedure.
In particular, it first discovers the spatial correlation between samples collected in various regions using a patching-based approach and then feeds the original and highly correlated samples into a Varienational Autocoder (VAE)
arXiv Detail & Related papers (2020-03-16T03:43:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.