GaitPT: Skeletons Are All You Need For Gait Recognition
- URL: http://arxiv.org/abs/2308.10623v2
- Date: Wed, 24 Jan 2024 12:06:49 GMT
- Title: GaitPT: Skeletons Are All You Need For Gait Recognition
- Authors: Andy Catruna, Adrian Cosma and Emilian Radoi
- Abstract summary: We propose a novel gait recognition architecture called Gait Pyramid Transformer (GaitPT)
GaitPT uses pose estimation skeletons to capture unique walking patterns, without relying on appearance information.
Our results show that GaitPT achieves state-of-the-art performance compared to other skeleton-based gait recognition works.
- Score: 4.089889918897877
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The analysis of patterns of walking is an important area of research that has
numerous applications in security, healthcare, sports and human-computer
interaction. Lately, walking patterns have been regarded as a unique
fingerprinting method for automatic person identification at a distance. In
this work, we propose a novel gait recognition architecture called Gait Pyramid
Transformer (GaitPT) that leverages pose estimation skeletons to capture unique
walking patterns, without relying on appearance information. GaitPT adopts a
hierarchical transformer architecture that effectively extracts both spatial
and temporal features of movement in an anatomically consistent manner, guided
by the structure of the human skeleton. Our results show that GaitPT achieves
state-of-the-art performance compared to other skeleton-based gait recognition
works, in both controlled and in-the-wild scenarios. GaitPT obtains 82.6%
average accuracy on CASIA-B, surpassing other works by a margin of 6%.
Moreover, it obtains 52.16% Rank-1 accuracy on GREW, outperforming both
skeleton-based and appearance-based approaches.
Related papers
- GaitFormer: Learning Gait Representations with Noisy Multi-Task Learning [4.831663144935878]
We propose DenseGait, the largest dataset for pretraining gait analysis systems containing 217K anonymized tracklets.
We also propose GaitFormer, a transformer-based model that achieves 92.5% accuracy on CASIA-B and 85.33% on FVG.
arXiv Detail & Related papers (2023-10-30T10:28:44Z) - SkeleTR: Towrads Skeleton-based Action Recognition in the Wild [86.03082891242698]
SkeleTR is a new framework for skeleton-based action recognition.
It first models the intra-person skeleton dynamics for each skeleton sequence with graph convolutions.
It then uses stacked Transformer encoders to capture person interactions that are important for action recognition in general scenarios.
arXiv Detail & Related papers (2023-09-20T16:22:33Z) - Distillation-guided Representation Learning for Unconstrained Gait Recognition [50.0533243584942]
We propose a framework, termed GAit DEtection and Recognition (GADER), for human authentication in challenging outdoor scenarios.
GADER builds discriminative features through a novel gait recognition method, where only frames containing gait information are used.
We evaluate our method on multiple State-of-The-Arts(SoTA) gait baselines and demonstrate consistent improvements on indoor and outdoor datasets.
arXiv Detail & Related papers (2023-07-27T01:53:57Z) - One-Shot Action Recognition via Multi-Scale Spatial-Temporal Skeleton
Matching [77.6989219290789]
One-shot skeleton action recognition aims to learn a skeleton action recognition model with a single training sample.
This paper presents a novel one-shot skeleton action recognition technique that handles skeleton action recognition via multi-scale spatial-temporal feature matching.
arXiv Detail & Related papers (2023-07-14T11:52:10Z) - DyGait: Exploiting Dynamic Representations for High-performance Gait
Recognition [35.642868929840034]
Gait recognition is a biometric technology that recognizes the identity of humans through their walking patterns.
We propose a novel and high-performance framework named DyGait to focus on the extraction of dynamic features.
Our network achieves an average Rank-1 accuracy of 71.4% on the GREW dataset, 66.3% on the Gait3D dataset, 98.4% on the CASIA-B dataset and 98.3% on the OU-M dataset.
arXiv Detail & Related papers (2023-03-27T07:36:47Z) - Multi-Modal Human Authentication Using Silhouettes, Gait and RGB [59.46083527510924]
Whole-body-based human authentication is a promising approach for remote biometrics scenarios.
We propose Dual-Modal Ensemble (DME), which combines both RGB and silhouette data to achieve more robust performances for indoor and outdoor whole-body based recognition.
Within DME, we propose GaitPattern, which is inspired by the double helical gait pattern used in traditional gait analysis.
arXiv Detail & Related papers (2022-10-08T15:17:32Z) - Towards a Deeper Understanding of Skeleton-based Gait Recognition [4.812321790984493]
In recent years, most gait recognition methods used the person's silhouette to extract the gait features.
Model-based methods do not suffer from these problems and are able to represent the temporal motion of body joints.
In this work, we propose an approach based on Graph Convolutional Networks (GCNs) that combines higher-order inputs, and residual networks.
arXiv Detail & Related papers (2022-04-16T18:23:37Z) - Spatial Transformer Network on Skeleton-based Gait Recognition [19.747638780327257]
Gait-TR is a robust skeleton-based gait recognition model based on spatial transformer frameworks and temporal convolutional networks.
Gait-TR achieves substantial improvements over other skeleton-based gait models with higher accuracy and better robustness on the well-known gait dataset CASIA-B.
arXiv Detail & Related papers (2022-04-08T06:53:23Z) - Learning Rich Features for Gait Recognition by Integrating Skeletons and
Silhouettes [20.766540020533803]
This paper proposes a simple yet effective bimodal fusion network, which mines the complementary clues of skeletons and silhouettes to learn rich features for gait identification.
Under the most challenging condition of walking in different clothes on CASIA-B, our method achieves the rank-1 accuracy of 92.1%.
arXiv Detail & Related papers (2021-10-26T04:42:24Z) - Domain Adaptive Robotic Gesture Recognition with Unsupervised
Kinematic-Visual Data Alignment [60.31418655784291]
We propose a novel unsupervised domain adaptation framework which can simultaneously transfer multi-modality knowledge, i.e., both kinematic and visual data, from simulator to real robot.
It remedies the domain gap with enhanced transferable features by using temporal cues in videos, and inherent correlations in multi-modal towards recognizing gesture.
Results show that our approach recovers the performance with great improvement gains, up to 12.91% in ACC and 20.16% in F1score without using any annotations in real robot.
arXiv Detail & Related papers (2021-03-06T09:10:03Z) - ROIAL: Region of Interest Active Learning for Characterizing Exoskeleton
Gait Preference Landscapes [64.87637128500889]
Region of Interest Active Learning (ROIAL) framework actively learns each user's underlying utility function over a region of interest.
ROIAL learns from ordinal and preference feedback, which are more reliable feedback mechanisms than absolute numerical scores.
Results demonstrate the feasibility of recovering gait utility landscapes from limited human trials.
arXiv Detail & Related papers (2020-11-09T22:45:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.