GaitSet: Cross-view Gait Recognition through Utilizing Gait as a Deep
Set
- URL: http://arxiv.org/abs/2102.03247v1
- Date: Fri, 5 Feb 2021 15:49:54 GMT
- Title: GaitSet: Cross-view Gait Recognition through Utilizing Gait as a Deep
Set
- Authors: Hanqing Chao, Kun Wang, Yiwei He, Junping Zhang, Jianfeng Feng
- Abstract summary: Gait is a unique biometric feature that can be recognized at a distance.
Existing gait recognition methods utilize either a gait template or a gait sequence that maintains unnecessary sequential constraints.
We present a novel perspective that utilizes gait as a deep set, which means that a set of gait frames are integrated by a global-local fused deep network.
- Score: 19.017795736485944
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Gait is a unique biometric feature that can be recognized at a distance;
thus, it has broad applications in crime prevention, forensic identification,
and social security. To portray a gait, existing gait recognition methods
utilize either a gait template which makes it difficult to preserve temporal
information, or a gait sequence that maintains unnecessary sequential
constraints and thus loses the flexibility of gait recognition. In this paper,
we present a novel perspective that utilizes gait as a deep set, which means
that a set of gait frames are integrated by a global-local fused deep network
inspired by the way our left- and right-hemisphere processes information to
learn information that can be used in identification. Based on this deep set
perspective, our method is immune to frame permutations, and can naturally
integrate frames from different videos that have been acquired under different
scenarios, such as diverse viewing angles, different clothes, or different
item-carrying conditions. Experiments show that under normal walking
conditions, our single-model method achieves an average rank-1 accuracy of
96.1% on the CASIA-B gait dataset and an accuracy of 87.9% on the OU-MVLP gait
dataset. Under various complex scenarios, our model also exhibits a high level
of robustness. It achieves accuracies of 90.8% and 70.3% on CASIA-B under
bag-carrying and coat-wearing walking conditions respectively, significantly
outperforming the best existing methods. Moreover, the proposed method
maintains a satisfactory accuracy even when only small numbers of frames are
available in the test samples; for example, it achieves 85.0% on CASIA-B even
when using only 7 frames. The source code has been released at
https://github.com/AbnerHqC/GaitSet.
Related papers
- Lazy Layers to Make Fine-Tuned Diffusion Models More Traceable [70.77600345240867]
A novel arbitrary-in-arbitrary-out (AIAO) strategy makes watermarks resilient to fine-tuning-based removal.
Unlike the existing methods of designing a backdoor for the input/output space of diffusion models, in our method, we propose to embed the backdoor into the feature space of sampled subpaths.
Our empirical studies on the MS-COCO, AFHQ, LSUN, CUB-200, and DreamBooth datasets confirm the robustness of AIAO.
arXiv Detail & Related papers (2024-05-01T12:03:39Z) - Whole-body Detection, Recognition and Identification at Altitude and
Range [57.445372305202405]
We propose an end-to-end system evaluated on diverse datasets.
Our approach involves pre-training the detector on common image datasets and fine-tuning it on BRIAR's complex videos and images.
We conduct thorough evaluations under various conditions, such as different ranges and angles in indoor, outdoor, and aerial scenarios.
arXiv Detail & Related papers (2023-11-09T20:20:23Z) - Distillation-guided Representation Learning for Unconstrained Gait Recognition [50.0533243584942]
We propose a framework, termed GAit DEtection and Recognition (GADER), for human authentication in challenging outdoor scenarios.
GADER builds discriminative features through a novel gait recognition method, where only frames containing gait information are used.
We evaluate our method on multiple State-of-The-Arts(SoTA) gait baselines and demonstrate consistent improvements on indoor and outdoor datasets.
arXiv Detail & Related papers (2023-07-27T01:53:57Z) - GaitGS: Temporal Feature Learning in Granularity and Span Dimension for Gait Recognition [34.07501669897291]
GaitGS is a framework that aggregates temporal features simultaneously in both granularity and span dimensions.
Our method demonstrates state-of-the-art performance, achieving Rank-1 accuracy of 98.2%, 96.5%, and 89.7% on two datasets.
arXiv Detail & Related papers (2023-05-31T09:48:25Z) - Learning Gait Representation from Massive Unlabelled Walking Videos: A
Benchmark [11.948554539954673]
This paper proposes a large-scale self-supervised benchmark for gait recognition with contrastive learning.
We collect a large-scale unlabelled gait dataset GaitLU-1M consisting of 1.02M walking sequences.
We evaluate the pre-trained model on four widely-used gait benchmarks, CASIA-B, OU-M, GREW and Gait3D with or without transfer learning.
arXiv Detail & Related papers (2022-06-28T12:33:42Z) - Gait Recognition in the Wild: A Large-scale Benchmark and NAS-based
Baseline [95.88825497452716]
Gait benchmarks empower the research community to train and evaluate high-performance gait recognition systems.
GREW is the first large-scale dataset for gait recognition in the wild.
SPOSGait is the first NAS-based gait recognition model.
arXiv Detail & Related papers (2022-05-05T14:57:39Z) - Spatial Transformer Network on Skeleton-based Gait Recognition [19.747638780327257]
Gait-TR is a robust skeleton-based gait recognition model based on spatial transformer frameworks and temporal convolutional networks.
Gait-TR achieves substantial improvements over other skeleton-based gait models with higher accuracy and better robustness on the well-known gait dataset CASIA-B.
arXiv Detail & Related papers (2022-04-08T06:53:23Z) - WildGait: Learning Gait Representations from Raw Surveillance Streams [1.90365714903665]
Existing methods for gait recognition require cooperative gait scenarios, in which a single person is walking multiple times in a straight line in front of a camera.
We propose a novel weakly supervised learning framework, WildGait, which consists of training a Spatio-Temporal Graph Convolutional Network on a large number of automatically annotated skeleton sequences.
Our results show that, with fine-tuning, we surpass in terms of recognition accuracy the current state-of-the-art pose-based gait recognition solutions.
arXiv Detail & Related papers (2021-05-12T09:11:32Z) - Uncertainty Sets for Image Classifiers using Conformal Prediction [112.54626392838163]
We present an algorithm that modifies any classifier to output a predictive set containing the true label with a user-specified probability, such as 90%.
The algorithm is simple and fast like Platt scaling, but provides a formal finite-sample coverage guarantee for every model and dataset.
Our method modifies an existing conformal prediction algorithm to give more stable predictive sets by regularizing the small scores of unlikely classes after Platt scaling.
arXiv Detail & Related papers (2020-09-29T17:58:04Z) - Uncertainty-Aware Weakly Supervised Action Detection from Untrimmed
Videos [82.02074241700728]
In this paper, we present a prohibitive-level action recognition model that is trained with only video-frame labels.
Our method per person detectors have been trained on large image datasets within Multiple Instance Learning framework.
We show how we can apply our method in cases where the standard Multiple Instance Learning assumption, that each bag contains at least one instance with the specified label, is invalid.
arXiv Detail & Related papers (2020-07-21T10:45:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.