Diverse Part Discovery: Occluded Person Re-identification with
Part-Aware Transformer
- URL: http://arxiv.org/abs/2106.04095v1
- Date: Tue, 8 Jun 2021 04:29:07 GMT
- Title: Diverse Part Discovery: Occluded Person Re-identification with
Part-Aware Transformer
- Authors: Yulin Li, Jianfeng He, Tianzhu Zhang, Xiang Liu, Yongdong Zhang, Feng
Wu
- Abstract summary: Occluded person re-identification (Re-ID) is a challenging task as persons are frequently occluded by various obstacles or other persons.
We propose a novel end-to-end Part-Aware Transformer (PAT) for occluded person Re-ID through diverse part discovery.
- Score: 95.02123369512384
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Occluded person re-identification (Re-ID) is a challenging task as persons
are frequently occluded by various obstacles or other persons, especially in
the crowd scenario. To address these issues, we propose a novel end-to-end
Part-Aware Transformer (PAT) for occluded person Re-ID through diverse part
discovery via a transformer encoderdecoder architecture, including a pixel
context based transformer encoder and a part prototype based transformer
decoder. The proposed PAT model enjoys several merits. First, to the best of
our knowledge, this is the first work to exploit the transformer
encoder-decoder architecture for occluded person Re-ID in a unified deep model.
Second, to learn part prototypes well with only identity labels, we design two
effective mechanisms including part diversity and part discriminability.
Consequently, we can achieve diverse part discovery for occluded person Re-ID
in a weakly supervised manner. Extensive experimental results on six
challenging benchmarks for three tasks (occluded, partial and holistic Re-ID)
demonstrate that our proposed PAT performs favorably against stat-of-the-art
methods.
Related papers
- Exploring Stronger Transformer Representation Learning for Occluded Person Re-Identification [2.552131151698595]
We proposed a novel self-supervision and supervision combining transformer-based person re-identification framework, namely SSSC-TransReID.
We designed a self-supervised contrastive learning branch, which can enhance the feature representation for person re-identification without negative samples or additional pre-training.
Our proposed model obtains superior Re-ID performance consistently and outperforms the state-of-the-art ReID methods by large margins on the mean average accuracy (mAP) and Rank-1 accuracy.
arXiv Detail & Related papers (2024-10-21T03:17:25Z) - Transformer for Object Re-Identification: A Survey [69.61542572894263]
Vision Transformers have spurred a growing number of studies delving deeper into Transformer-based Re-ID.
This paper provides a comprehensive review and in-depth analysis of the Transformer-based Re-ID.
Considering the trending unsupervised Re-ID, we propose a new Transformer baseline, UntransReID, achieving state-of-the-art performance.
arXiv Detail & Related papers (2024-01-13T03:17:57Z) - Part Representation Learning with Teacher-Student Decoder for Occluded
Person Re-identification [65.63180725319906]
We propose a Teacher-Student Decoder (TSD) framework for occluded person ReID.
Our proposed TSD consists of a Parsing-aware Teacher Decoder (PTD) and a Standard Student Decoder (SSD)
arXiv Detail & Related papers (2023-12-15T13:54:48Z) - Transformer Based Multi-Grained Features for Unsupervised Person
Re-Identification [9.874360118638918]
We build a dual-branch network architecture based upon a modified Vision Transformer (ViT)
Local tokens output in each branch are reshaped and then uniformly partitioned into multiple stripes to generate part-level features.
Global tokens of two branches are averaged to produce a global feature.
arXiv Detail & Related papers (2022-11-22T13:51:17Z) - Sequential Transformer for End-to-End Person Search [4.920657401819193]
Person search aims to simultaneously localize and recognize a target person from realistic and uncropped gallery images.
In this paper, we propose a novel Sequential Transformer (SeqTR) for end-to-end person search to deal with this challenge.
Our SeqTR contains a detection transformer and a novel re-ID transformer that sequentially addresses detection and re-ID tasks.
arXiv Detail & Related papers (2022-11-06T09:32:30Z) - Integral Migrating Pre-trained Transformer Encoder-decoders for Visual
Object Detection [78.2325219839805]
imTED improves the state-of-the-art of few-shot object detection by up to 7.6% AP.
Experiments on MS COCO dataset demonstrate that imTED consistently outperforms its counterparts by 2.8%.
arXiv Detail & Related papers (2022-05-19T15:11:20Z) - Human-Object Interaction Detection via Disentangled Transformer [63.46358684341105]
We present Disentangled Transformer, where both encoder and decoder are disentangled to facilitate learning of two sub-tasks.
Our method outperforms prior work on two public HOI benchmarks by a sizeable margin.
arXiv Detail & Related papers (2022-04-20T08:15:04Z) - PSTR: End-to-End One-Step Person Search With Transformers [140.32813648935752]
We propose a one-step transformer-based person search framework, PSTR.
PSS module contains a detection encoder-decoder for person detection along with a discriminative re-id decoder for person re-id.
On the challenging PRW benchmark, PSTR achieves a mean average precision (mAP) score of 56.5%.
arXiv Detail & Related papers (2022-04-07T10:22:33Z) - Learning Disentangled Representation Implicitly via Transformer for
Occluded Person Re-Identification [35.40162083252931]
DRL-Net is a representation learning network that handles occluded re-ID without requiring strict person image alignment or any additional supervision.
It measures image similarity by automatically disentangling the representation of undefined semantic components.
The DRL-Net achieves superior re-ID performance consistently and outperforms the state-of-the-art by large margins for Occluded-DukeMTMC.
arXiv Detail & Related papers (2021-07-06T04:24:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.