Attention-based Shape and Gait Representations Learning for Video-based
Cloth-Changing Person Re-Identification
- URL: http://arxiv.org/abs/2402.03716v1
- Date: Tue, 6 Feb 2024 05:11:46 GMT
- Title: Attention-based Shape and Gait Representations Learning for Video-based
Cloth-Changing Person Re-Identification
- Authors: Vuong D. Nguyen, Samiha Mirza, Pranav Mantini, Shishir K. Shah
- Abstract summary: We deal with the practical problem of Video-based Cloth-Changing Person Re-ID (VCCRe-ID) by proposing "Attention-based Shape and Gait Representations Learning" (ASGL)
Our ASGL framework improves Re-ID performance under clothing variations by learning clothing-invariant gait cues.
Our proposed ST-GAT comprises multi-head attention modules, which are able to enhance the robustness of gait embeddings.
- Score: 1.6385815610837167
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current state-of-the-art Video-based Person Re-Identification (Re-ID)
primarily relies on appearance features extracted by deep learning models.
These methods are not applicable for long-term analysis in real-world scenarios
where persons have changed clothes, making appearance information unreliable.
In this work, we deal with the practical problem of Video-based Cloth-Changing
Person Re-ID (VCCRe-ID) by proposing "Attention-based Shape and Gait
Representations Learning" (ASGL) for VCCRe-ID. Our ASGL framework improves
Re-ID performance under clothing variations by learning clothing-invariant gait
cues using a Spatial-Temporal Graph Attention Network (ST-GAT). Given the
3D-skeleton-based spatial-temporal graph, our proposed ST-GAT comprises
multi-head attention modules, which are able to enhance the robustness of gait
embeddings under viewpoint changes and occlusions. The ST-GAT amplifies the
important motion ranges and reduces the influence of noisy poses. Then, the
multi-head learning module effectively reserves beneficial local temporal
dynamics of movement. We also boost discriminative power of person
representations by learning body shape cues using a GAT. Experiments on two
large-scale VCCRe-ID datasets demonstrate that our proposed framework
outperforms state-of-the-art methods by 12.2% in rank-1 accuracy and 7.0% in
mAP.
Related papers
- Multiple Information Prompt Learning for Cloth-Changing Person Re-Identification [17.948263914620238]
We propose a novel multiple information prompt learning (MIPL) scheme for cloth-changing person ReID.
TheCIS module is designed to decouple clothing information from the original RGB image features.
The Bio-guided attention (BGA) module is proposed to increase the learning intensity of the model for key information.
arXiv Detail & Related papers (2024-11-01T03:08:10Z) - Exploring Stronger Transformer Representation Learning for Occluded Person Re-Identification [2.552131151698595]
We proposed a novel self-supervision and supervision combining transformer-based person re-identification framework, namely SSSC-TransReID.
We designed a self-supervised contrastive learning branch, which can enhance the feature representation for person re-identification without negative samples or additional pre-training.
Our proposed model obtains superior Re-ID performance consistently and outperforms the state-of-the-art ReID methods by large margins on the mean average accuracy (mAP) and Rank-1 accuracy.
arXiv Detail & Related papers (2024-10-21T03:17:25Z) - MIMAFace: Face Animation via Motion-Identity Modulated Appearance Feature Learning [30.61146302275139]
We introduce a Motion-Identity Modulated Appearance Learning Module (MIA) that modulates CLIP features at both motion and identity levels.
We also design an Inter-clip Affinity Learning Module (ICA) to model temporal relationships across clips.
Our method achieves precise facial motion control (i.e., expressions and gaze), faithful identity preservation, and generates animation videos that maintain both intra/inter-clip temporal consistency.
arXiv Detail & Related papers (2024-09-23T16:33:53Z) - Tran-GCN: A Transformer-Enhanced Graph Convolutional Network for Person Re-Identification in Monitoring Videos [8.559235103954341]
Person Re-Identification (Re-ID) has gained popularity in computer vision, enabling cross-camera pedestrian recognition.
We propose a Transformer-enhanced Graph Convolutional Network (Tran-GCN) model to improve Person Re-Identification performance in monitoring videos.
arXiv Detail & Related papers (2024-09-14T09:42:48Z) - Exploring Fine-Grained Representation and Recomposition for Cloth-Changing Person Re-Identification [78.52704557647438]
We propose a novel FIne-grained Representation and Recomposition (FIRe$2$) framework to tackle both limitations without any auxiliary annotation or data.
Experiments demonstrate that FIRe$2$ can achieve state-of-the-art performance on five widely-used cloth-changing person Re-ID benchmarks.
arXiv Detail & Related papers (2023-08-21T12:59:48Z) - Distillation-guided Representation Learning for Unconstrained Gait Recognition [50.0533243584942]
We propose a framework, termed GAit DEtection and Recognition (GADER), for human authentication in challenging outdoor scenarios.
GADER builds discriminative features through a novel gait recognition method, where only frames containing gait information are used.
We evaluate our method on multiple State-of-The-Arts(SoTA) gait baselines and demonstrate consistent improvements on indoor and outdoor datasets.
arXiv Detail & Related papers (2023-07-27T01:53:57Z) - Shape-Erased Feature Learning for Visible-Infrared Person
Re-Identification [90.39454748065558]
Body shape is one of the significant modality-shared cues for VI-ReID.
We propose shape-erased feature learning paradigm that decorrelates modality-shared features in two subspaces.
Experiments on SYSU-MM01, RegDB, and HITSZ-VCM datasets demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2023-04-09T10:22:10Z) - Feature Disentanglement Learning with Switching and Aggregation for
Video-based Person Re-Identification [9.068045610800667]
In video person re-identification (Re-ID), the network must consistently extract features of the target person from successive frames.
Existing methods tend to focus only on how to use temporal information, which often leads to networks being fooled by similar appearances and same backgrounds.
We propose a Disentanglement and Switching and Aggregation Network (DSANet), which segregates the features representing identity and features based on camera characteristics, and pays more attention to ID information.
arXiv Detail & Related papers (2022-12-16T04:27:56Z) - A High-Accuracy Unsupervised Person Re-identification Method Using
Auxiliary Information Mined from Datasets [53.047542904329866]
We make use of auxiliary information mined from datasets for multi-modal feature learning.
This paper proposes three effective training tricks, including Restricted Label Smoothing Cross Entropy Loss (RLSCE), Weight Adaptive Triplet Loss (WATL) and Dynamic Training Iterations (DTI)
arXiv Detail & Related papers (2022-05-06T10:16:18Z) - TCGL: Temporal Contrastive Graph for Self-supervised Video
Representation Learning [79.77010271213695]
We propose a novel video self-supervised learning framework named Temporal Contrastive Graph Learning (TCGL)
Our TCGL integrates the prior knowledge about the frame and snippet orders into graph structures, i.e., the intra-/inter- snippet Temporal Contrastive Graphs (TCG)
To generate supervisory signals for unlabeled videos, we introduce an Adaptive Snippet Order Prediction (ASOP) module.
arXiv Detail & Related papers (2021-12-07T09:27:56Z) - A Self-Supervised Gait Encoding Approach with Locality-Awareness for 3D
Skeleton Based Person Re-Identification [65.18004601366066]
Person re-identification (Re-ID) via gait features within 3D skeleton sequences is a newly-emerging topic with several advantages.
This paper proposes a self-supervised gait encoding approach that can leverage unlabeled skeleton data to learn gait representations for person Re-ID.
arXiv Detail & Related papers (2020-09-05T16:06:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.