Dynamic Feature Pruning and Consolidation for Occluded Person
Re-Identification
- URL: http://arxiv.org/abs/2211.14742v2
- Date: Thu, 21 Dec 2023 04:06:43 GMT
- Title: Dynamic Feature Pruning and Consolidation for Occluded Person
Re-Identification
- Authors: YuTeng Ye, Hang Zhou, Jiale Cai, Chenxing Gao, Youjia Zhang, Junle
Wang, Qiang Hu, Junqing Yu, Wei Yang
- Abstract summary: We propose a feature pruning and consolidation (FPC) framework to circumvent explicit human structure parsing.
The framework mainly consists of a sparse encoder, a multi-view feature mathcing module, and a feature consolidation decoder.
Our method outperforms state-of-the-art results by at least 8.6% mAP and 6.0% Rank-1 accuracy on the challenging Occluded-Duke dataset.
- Score: 21.006680330530852
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Occluded person re-identification (ReID) is a challenging problem due to
contamination from occluders. Existing approaches address the issue with prior
knowledge cues, such as human body key points and semantic segmentations, which
easily fail in the presence of heavy occlusion and other humans as occluders.
In this paper, we propose a feature pruning and consolidation (FPC) framework
to circumvent explicit human structure parsing. The framework mainly consists
of a sparse encoder, a multi-view feature mathcing module, and a feature
consolidation decoder. Specifically, the sparse encoder drops less important
image tokens, mostly related to background noise and occluders, solely based on
correlation within the class token attention. Subsequently, the matching stage
relies on the preserved tokens produced by the sparse encoder to identify
k-nearest neighbors in the gallery by measuring the image and patch-level
combined similarity. Finally, we use the feature consolidation module to
compensate pruned features using identified neighbors for recovering essential
information while disregarding disturbance from noise and occlusion.
Experimental results demonstrate the effectiveness of our proposed framework on
occluded, partial, and holistic Re-ID datasets. In particular, our method
outperforms state-of-the-art results by at least 8.6\% mAP and 6.0\% Rank-1
accuracy on the challenging Occluded-Duke dataset.
Related papers
- DDRN:a Data Distribution Reconstruction Network for Occluded Person Re-Identification [5.7703191981015305]
We propose a generative model that leverages data distribution to filter out irrelevant details.
On the Occluded-Duke dataset, we achieved a mAP of 62.4% (+1.1%) and a rank-1 accuracy of 71.3% (+0.6%)
arXiv Detail & Related papers (2024-10-09T06:52:05Z) - SpaRG: Sparsely Reconstructed Graphs for Generalizable fMRI Analysis [8.489318619991534]
Deep learning can help uncover patterns in resting-state functional Magnetic Resonance Imaging (rsfMRI) associated with psychiatric disorders and personal traits.
Yet the problem of interpreting deep learning findings is rarely more evident than in fMRI analyses.
We propose a simple approach to mitigate these challenges grounded on sparsification and self-supervision.
arXiv Detail & Related papers (2024-09-24T18:35:57Z) - UGMAE: A Unified Framework for Graph Masked Autoencoders [67.75493040186859]
We propose UGMAE, a unified framework for graph masked autoencoders.
We first develop an adaptive feature mask generator to account for the unique significance of nodes.
We then design a ranking-based structure reconstruction objective joint with feature reconstruction to capture holistic graph information.
arXiv Detail & Related papers (2024-02-12T19:39:26Z) - Feature Completion Transformer for Occluded Person Re-identification [25.159974510754992]
Occluded person re-identification (Re-ID) is a challenging problem due to the destruction of occluders.
We propose a Feature Completion Transformer (FCFormer) to implicitly complement the semantic information of occluded parts in the feature space.
FCFormer achieves superior performance and outperforms the state-of-the-art methods by significant margins on occluded datasets.
arXiv Detail & Related papers (2023-03-03T01:12:57Z) - Learning Feature Recovery Transformer for Occluded Person
Re-identification [71.18476220969647]
We propose a new approach called Feature Recovery Transformer (FRT) to address the two challenges simultaneously.
To reduce the interference of the noise during feature matching, we mainly focus on visible regions that appear in both images and develop a visibility graph to calculate the similarity.
In terms of the second challenge, based on the developed graph similarity, for each query image, we propose a recovery transformer that exploits the feature sets of its $k$-nearest neighbors in the gallery to recover the complete features.
arXiv Detail & Related papers (2023-01-05T02:36:16Z) - Neighbourhood-guided Feature Reconstruction for Occluded Person
Re-Identification [45.704612531562404]
We propose to reconstruct the feature representation of occluded parts by fully exploiting the information of its neighborhood in a gallery image set.
In the large-scale Occluded-DukeMTMC benchmark, our approach achieves 64.2% mAP and 67.6% rank-1 accuracy.
arXiv Detail & Related papers (2021-05-16T03:53:55Z) - Holistic Guidance for Occluded Person Re-Identification [7.662745552551165]
In real-world video surveillance applications, person re-identification (ReID) suffers from the effects of occlusions and detection errors.
We introduce a novel Holistic Guidance (HG) method that relies only on person identity labels.
Our proposed student-teacher framework is trained to address the problem by matching the distributions of between- and within-class distances (DCDs) of occluded samples with that of holistic (non-occluded) samples.
In addition to this, a joint generative-discriminative backbone is trained with a denoising autoencoder, allowing the system to
arXiv Detail & Related papers (2021-04-13T21:50:29Z) - Generative Partial Visual-Tactile Fused Object Clustering [81.17645983141773]
We propose a Generative Partial Visual-Tactile Fused (i.e., GPVTF) framework for object clustering.
A conditional cross-modal clustering generative adversarial network is then developed to synthesize one modality conditioning on the other modality.
To the end, two pseudo-label based KL-divergence losses are employed to update the corresponding modality-specific encoders.
arXiv Detail & Related papers (2020-12-28T02:37:03Z) - Improving Face Recognition by Clustering Unlabeled Faces in the Wild [77.48677160252198]
We propose a novel identity separation method based on extreme value theory.
It greatly reduces the problems caused by overlapping-identity label noise.
Experiments on both controlled and real settings demonstrate our method's consistent improvements.
arXiv Detail & Related papers (2020-07-14T12:26:50Z) - Peeking into occluded joints: A novel framework for crowd pose
estimation [88.56203133287865]
OPEC-Net is an Image-Guided Progressive GCN module that estimates invisible joints from an inference perspective.
OCPose is the most complex Occluded Pose dataset with respect to average IoU between adjacent instances.
arXiv Detail & Related papers (2020-03-23T19:32:40Z) - High-Order Information Matters: Learning Relation and Topology for
Occluded Person Re-Identification [84.43394420267794]
We propose a novel framework by learning high-order relation and topology information for discriminative features and robust alignment.
Our framework significantly outperforms state-of-the-art by6.5%mAP scores on Occluded-Duke dataset.
arXiv Detail & Related papers (2020-03-18T12:18:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.