Learning Feature Recovery Transformer for Occluded Person
Re-identification
- URL: http://arxiv.org/abs/2301.01879v1
- Date: Thu, 5 Jan 2023 02:36:16 GMT
- Title: Learning Feature Recovery Transformer for Occluded Person
Re-identification
- Authors: Boqiang Xu, Lingxiao He, Jian Liang, Zhenan Sun
- Abstract summary: We propose a new approach called Feature Recovery Transformer (FRT) to address the two challenges simultaneously.
To reduce the interference of the noise during feature matching, we mainly focus on visible regions that appear in both images and develop a visibility graph to calculate the similarity.
In terms of the second challenge, based on the developed graph similarity, for each query image, we propose a recovery transformer that exploits the feature sets of its $k$-nearest neighbors in the gallery to recover the complete features.
- Score: 71.18476220969647
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: One major issue that challenges person re-identification (Re-ID) is the
ubiquitous occlusion over the captured persons. There are two main challenges
for the occluded person Re-ID problem, i.e., the interference of noise during
feature matching and the loss of pedestrian information brought by the
occlusions. In this paper, we propose a new approach called Feature Recovery
Transformer (FRT) to address the two challenges simultaneously, which mainly
consists of visibility graph matching and feature recovery transformer. To
reduce the interference of the noise during feature matching, we mainly focus
on visible regions that appear in both images and develop a visibility graph to
calculate the similarity. In terms of the second challenge, based on the
developed graph similarity, for each query image, we propose a recovery
transformer that exploits the feature sets of its $k$-nearest neighbors in the
gallery to recover the complete features. Extensive experiments across
different person Re-ID datasets, including occluded, partial and holistic
datasets, demonstrate the effectiveness of FRT. Specifically, FRT significantly
outperforms state-of-the-art results by at least 6.2\% Rank-1 accuracy and
7.2\% mAP scores on the challenging Occluded-Duke dataset. The code is
available at https://github.com/xbq1994/Feature-Recovery-Transformer.
Related papers
- DDRN:a Data Distribution Reconstruction Network for Occluded Person Re-Identification [5.7703191981015305]
We propose a generative model that leverages data distribution to filter out irrelevant details.
On the Occluded-Duke dataset, we achieved a mAP of 62.4% (+1.1%) and a rank-1 accuracy of 71.3% (+0.6%)
arXiv Detail & Related papers (2024-10-09T06:52:05Z) - DiffBIR: Towards Blind Image Restoration with Generative Diffusion Prior [70.46245698746874]
We present DiffBIR, a general restoration pipeline that could handle different blind image restoration tasks.
DiffBIR decouples blind image restoration problem into two stages: 1) degradation removal: removing image-independent content; 2) information regeneration: generating the lost image content.
In the first stage, we use restoration modules to remove degradations and obtain high-fidelity restored results.
For the second stage, we propose IRControlNet that leverages the generative ability of latent diffusion models to generate realistic details.
arXiv Detail & Related papers (2023-08-29T07:11:52Z) - Feature Completion Transformer for Occluded Person Re-identification [25.159974510754992]
Occluded person re-identification (Re-ID) is a challenging problem due to the destruction of occluders.
We propose a Feature Completion Transformer (FCFormer) to implicitly complement the semantic information of occluded parts in the feature space.
FCFormer achieves superior performance and outperforms the state-of-the-art methods by significant margins on occluded datasets.
arXiv Detail & Related papers (2023-03-03T01:12:57Z) - Exploring Invariant Representation for Visible-Infrared Person
Re-Identification [77.06940947765406]
Cross-spectral person re-identification, which aims to associate identities to pedestrians across different spectra, faces a main challenge of the modality discrepancy.
In this paper, we address the problem from both image-level and feature-level in an end-to-end hybrid learning framework named robust feature mining network (RFM)
Experiment results on two standard cross-spectral person re-identification datasets, RegDB and SYSU-MM01, have demonstrated state-of-the-art performance.
arXiv Detail & Related papers (2023-02-02T05:24:50Z) - Dynamic Feature Pruning and Consolidation for Occluded Person
Re-Identification [21.006680330530852]
We propose a feature pruning and consolidation (FPC) framework to circumvent explicit human structure parsing.
The framework mainly consists of a sparse encoder, a multi-view feature mathcing module, and a feature consolidation decoder.
Our method outperforms state-of-the-art results by at least 8.6% mAP and 6.0% Rank-1 accuracy on the challenging Occluded-Duke dataset.
arXiv Detail & Related papers (2022-11-27T06:18:40Z) - Counterfactual Intervention Feature Transfer for Visible-Infrared Person
Re-identification [69.45543438974963]
We find graph-based methods in the visible-infrared person re-identification task (VI-ReID) suffer from bad generalization because of two issues.
The well-trained input features weaken the learning of graph topology, making it not generalized enough during the inference process.
We propose a Counterfactual Intervention Feature Transfer (CIFT) method to tackle these problems.
arXiv Detail & Related papers (2022-08-01T16:15:31Z) - On Exploring Pose Estimation as an Auxiliary Learning Task for
Visible-Infrared Person Re-identification [66.58450185833479]
In this paper, we exploit Pose Estimation as an auxiliary learning task to assist the VI-ReID task in an end-to-end framework.
By jointly training these two tasks in a mutually beneficial manner, our model learns higher quality modality-shared and ID-related features.
Experimental results on two benchmark VI-ReID datasets show that the proposed method consistently improves state-of-the-art methods by significant margins.
arXiv Detail & Related papers (2022-01-11T09:44:00Z) - Neighbourhood-guided Feature Reconstruction for Occluded Person
Re-Identification [45.704612531562404]
We propose to reconstruct the feature representation of occluded parts by fully exploiting the information of its neighborhood in a gallery image set.
In the large-scale Occluded-DukeMTMC benchmark, our approach achieves 64.2% mAP and 67.6% rank-1 accuracy.
arXiv Detail & Related papers (2021-05-16T03:53:55Z) - Generative Partial Visual-Tactile Fused Object Clustering [81.17645983141773]
We propose a Generative Partial Visual-Tactile Fused (i.e., GPVTF) framework for object clustering.
A conditional cross-modal clustering generative adversarial network is then developed to synthesize one modality conditioning on the other modality.
To the end, two pseudo-label based KL-divergence losses are employed to update the corresponding modality-specific encoders.
arXiv Detail & Related papers (2020-12-28T02:37:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.