Gait Recognition using Multi-Scale Partial Representation Transformation
with Capsules
- URL: http://arxiv.org/abs/2010.09084v1
- Date: Sun, 18 Oct 2020 19:47:38 GMT
- Title: Gait Recognition using Multi-Scale Partial Representation Transformation
with Capsules
- Authors: Alireza Sepas-Moghaddam, Saeed Ghorbani, Nikolaus F. Troje, Ali Etemad
- Abstract summary: We propose a novel deep network, learning to transfer multi-scale partial gait representations using capsules.
Our network first obtains multi-scale partial representations using a state-of-the-art deep partial feature extractor.
It then recurrently learns the correlations and co-occurrences of the patterns among the partial features in forward and backward directions.
- Score: 22.99694601595627
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Gait recognition, referring to the identification of individuals based on the
manner in which they walk, can be very challenging due to the variations in the
viewpoint of the camera and the appearance of individuals. Current methods for
gait recognition have been dominated by deep learning models, notably those
based on partial feature representations. In this context, we propose a novel
deep network, learning to transfer multi-scale partial gait representations
using capsules to obtain more discriminative gait features. Our network first
obtains multi-scale partial representations using a state-of-the-art deep
partial feature extractor. It then recurrently learns the correlations and
co-occurrences of the patterns among the partial features in forward and
backward directions using Bi-directional Gated Recurrent Units (BGRU). Finally,
a capsule network is adopted to learn deeper part-whole relationships and
assigns more weights to the more relevant features while ignoring the spurious
dimensions. That way, we obtain final features that are more robust to both
viewing and appearance changes. The performance of our method has been
extensively tested on two gait recognition datasets, CASIA-B and OU-MVLP, using
four challenging test protocols. The results of our method have been compared
to the state-of-the-art gait recognition solutions, showing the superiority of
our model, notably when facing challenging viewing and carrying conditions.
Related papers
- DVANet: Disentangling View and Action Features for Multi-View Action
Recognition [56.283944756315066]
We present a novel approach to multi-view action recognition where we guide learned action representations to be separated from view-relevant information in a video.
Our model and method of training significantly outperforms all other uni-modal models on four multi-view action recognition datasets.
arXiv Detail & Related papers (2023-12-10T01:19:48Z) - Distillation-guided Representation Learning for Unconstrained Gait Recognition [50.0533243584942]
We propose a framework, termed GAit DEtection and Recognition (GADER), for human authentication in challenging outdoor scenarios.
GADER builds discriminative features through a novel gait recognition method, where only frames containing gait information are used.
We evaluate our method on multiple State-of-The-Arts(SoTA) gait baselines and demonstrate consistent improvements on indoor and outdoor datasets.
arXiv Detail & Related papers (2023-07-27T01:53:57Z) - Discrepancy-Guided Reconstruction Learning for Image Forgery Detection [10.221066530624373]
We first propose a Discrepancy-Guided (DisGE) to extract forgery-sensitive visual patterns.
We then introduce a Double-Head Reconstruction (DouHR) module to enhance genuine compact visual patterns in different granular spaces.
Under DouHR, we further introduce a Discrepancy-Aggregation Detector (DisAD) to aggregate these genuine compact visual patterns.
arXiv Detail & Related papers (2023-04-26T07:40:43Z) - Semantic Prompt for Few-Shot Image Recognition [76.68959583129335]
We propose a novel Semantic Prompt (SP) approach for few-shot learning.
The proposed approach achieves promising results, improving the 1-shot learning accuracy by 3.67% on average.
arXiv Detail & Related papers (2023-03-24T16:32:19Z) - Exploring Deep Models for Practical Gait Recognition [11.185716724976414]
We present a unified perspective to explore how to construct deep models for state-of-the-art outdoor gait recognition.
Specifically, we challenge the stereotype of shallow gait models and demonstrate the superiority of explicit temporal modeling.
The proposed CNN-based DeepGaitV2 series and Transformer-based SwinGait series exhibit significant performance improvements on Gait3D and GREW.
arXiv Detail & Related papers (2023-03-06T17:19:28Z) - Learning Common Rationale to Improve Self-Supervised Representation for
Fine-Grained Visual Recognition Problems [61.11799513362704]
We propose learning an additional screening mechanism to identify discriminative clues commonly seen across instances and classes.
We show that a common rationale detector can be learned by simply exploiting the GradCAM induced from the SSL objective.
arXiv Detail & Related papers (2023-03-03T02:07:40Z) - Towards a Deeper Understanding of Skeleton-based Gait Recognition [4.812321790984493]
In recent years, most gait recognition methods used the person's silhouette to extract the gait features.
Model-based methods do not suffer from these problems and are able to represent the temporal motion of body joints.
In this work, we propose an approach based on Graph Convolutional Networks (GCNs) that combines higher-order inputs, and residual networks.
arXiv Detail & Related papers (2022-04-16T18:23:37Z) - Distribution Alignment: A Unified Framework for Long-tail Visual
Recognition [52.36728157779307]
We propose a unified distribution alignment strategy for long-tail visual recognition.
We then introduce a generalized re-weight method in the two-stage learning to balance the class prior.
Our approach achieves the state-of-the-art results across all four recognition tasks with a simple and unified framework.
arXiv Detail & Related papers (2021-03-30T14:09:53Z) - Associated Spatio-Temporal Capsule Network for Gait Recognition [36.85667679699001]
State-of-the-art approaches rely on analysis of temporal or spatial characteristics of gait.
ASTCapsNet trained on multi-sensor datasets to analyze multimodal information for gait recognition.
arXiv Detail & Related papers (2021-01-07T09:55:17Z) - Depth Guided Adaptive Meta-Fusion Network for Few-shot Video Recognition [86.31412529187243]
Few-shot video recognition aims at learning new actions with only very few labeled samples.
We propose a depth guided Adaptive Meta-Fusion Network for few-shot video recognition which is termed as AMeFu-Net.
arXiv Detail & Related papers (2020-10-20T03:06:20Z) - View-Invariant Gait Recognition with Attentive Recurrent Learning of
Partial Representations [27.33579145744285]
We propose a network that first learns to extract gait convolutional energy maps (GCEM) from frame-level convolutional features.
It then adopts a bidirectional neural network to learn from split bins of the GCEM, thus exploiting the relations between learned partial recurrent representations.
Our proposed model has been extensively tested on two large-scale CASIA-B and OU-M gait datasets.
arXiv Detail & Related papers (2020-10-18T20:20:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.