VMRFANet:View-Specific Multi-Receptive Field Attention Network for
Person Re-identification
- URL: http://arxiv.org/abs/2001.07354v1
- Date: Tue, 21 Jan 2020 06:31:18 GMT
- Title: VMRFANet:View-Specific Multi-Receptive Field Attention Network for
Person Re-identification
- Authors: Honglong Cai, Yuedong Fang, Zhiguan Wang, Tingchun Yeh, Jinxing Cheng
- Abstract summary: We propose a novel multi-receptive field attention (MRFA) module that utilizes filters of various sizes to help network focusing on informative pixels.
We present a view-specific mechanism that guides attention module to handle the variation of view conditions.
Our method achieves 95.5% / 88.1% in rank-1 / mAP on Market-1501, 88.9% / 80.0% on DukeMTMC-reID, 81.1% / 78.8% on CUHK03 labeled dataset and 78.9% / 75.3% on CUHK03 detected dataset.
- Score: 3.1498833540989413
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Person re-identification (re-ID) aims to retrieve the same person across
different cameras. In practice, it still remains a challenging task due to
background clutter, variations on body poses and view conditions, inaccurate
bounding box detection, etc. To tackle these issues, in this paper, we propose
a novel multi-receptive field attention (MRFA) module that utilizes filters of
various sizes to help network focusing on informative pixels. Besides, we
present a view-specific mechanism that guides attention module to handle the
variation of view conditions. Moreover, we introduce a Gaussian horizontal
random cropping/padding method which further improves the robustness of our
proposed network. Comprehensive experiments demonstrate the effectiveness of
each component. Our method achieves 95.5% / 88.1% in rank-1 / mAP on
Market-1501, 88.9% / 80.0% on DukeMTMC-reID, 81.1% / 78.8% on CUHK03 labeled
dataset and 78.9% / 75.3% on CUHK03 detected dataset, outperforming current
state-of-the-art methods.
Related papers
- Multiple Information Prompt Learning for Cloth-Changing Person Re-Identification [17.948263914620238]
We propose a novel multiple information prompt learning (MIPL) scheme for cloth-changing person ReID.
TheCIS module is designed to decouple clothing information from the original RGB image features.
The Bio-guided attention (BGA) module is proposed to increase the learning intensity of the model for key information.
arXiv Detail & Related papers (2024-11-01T03:08:10Z) - Domain-Guided Masked Autoencoders for Unique Player Identification [62.87054782745536]
Masked autoencoders (MAEs) have emerged as a superior alternative to conventional feature extractors.
Motivated by human vision, we devise a novel domain-guided masking policy for MAEs termed d-MAE.
We conduct experiments on three large-scale sports datasets.
arXiv Detail & Related papers (2024-03-17T20:14:57Z) - AaP-ReID: Improved Attention-Aware Person Re-identification [2.5761958263376745]
AaP-ReID is a more effective method for person ReID that incorporates channel-wise attention into a ResNet-based architecture.
Our method incorporates the Channel-Wise Attention Bottleneck block and can learn discriminating features by dynamically adjusting the importance ofeach channel in the feature maps.
arXiv Detail & Related papers (2023-09-27T16:54:38Z) - A High-Accuracy Unsupervised Person Re-identification Method Using
Auxiliary Information Mined from Datasets [53.047542904329866]
We make use of auxiliary information mined from datasets for multi-modal feature learning.
This paper proposes three effective training tricks, including Restricted Label Smoothing Cross Entropy Loss (RLSCE), Weight Adaptive Triplet Loss (WATL) and Dynamic Training Iterations (DTI)
arXiv Detail & Related papers (2022-05-06T10:16:18Z) - Learning Instance-level Spatial-Temporal Patterns for Person
Re-identification [80.43222559182072]
We propose a novel Instance-level and Spatial-temporal Disentangled Re-ID method (InSTD) to improve Re-ID accuracy.
In our proposed framework, personalized information such as moving direction is explicitly considered to further narrow down the search space.
The proposed method achieves mAP of 90.8% on Market-1501 and 89.1% on DukeMTMC-reID, improving from the baseline 82.2% and 72.7%, respectively.
arXiv Detail & Related papers (2021-07-31T07:44:47Z) - Neighbourhood-guided Feature Reconstruction for Occluded Person
Re-Identification [45.704612531562404]
We propose to reconstruct the feature representation of occluded parts by fully exploiting the information of its neighborhood in a gallery image set.
In the large-scale Occluded-DukeMTMC benchmark, our approach achieves 64.2% mAP and 67.6% rank-1 accuracy.
arXiv Detail & Related papers (2021-05-16T03:53:55Z) - ES-Net: Erasing Salient Parts to Learn More in Re-Identification [46.624740579314924]
We propose a novel network, Erasing-Salient Net (ES-Net), to learn comprehensive features by erasing the salient areas in an image.
Our ES-Net outperforms state-of-the-art methods on three Person re-ID benchmarks and two Vehicle re-ID benchmarks.
arXiv Detail & Related papers (2021-03-10T08:19:46Z) - Lightweight Multi-Branch Network for Person Re-Identification [6.353193172884524]
This paper proposes a lightweight network that combines global, part-based, and channel features in a unified multi-branch architecture that builds on the resource-efficient OSNet backbone.
Using a well-founded combination of training techniques and design choices, our final model achieves state-of-the-art results on CUHK03 labeled, CUHK03 detected, and Market-1501 with 85.1% mAP / 87.2% rank1, 82.4% mAP / 84.9% rank1, and 91.5% mAP / 96.3% rank1, respectively.
arXiv Detail & Related papers (2021-01-26T13:28:46Z) - Robust Data Hiding Using Inverse Gradient Attention [82.73143630466629]
In the data hiding task, each pixel of cover images should be treated differently since they have divergent tolerabilities.
We propose a novel deep data hiding scheme with Inverse Gradient Attention (IGA), combing the ideas of adversarial learning and attention mechanism.
Empirically, extensive experiments show that the proposed model outperforms the state-of-the-art methods on two prevalent datasets.
arXiv Detail & Related papers (2020-11-21T19:08:23Z) - Towards High Performance Human Keypoint Detection [87.1034745775229]
We find that context information plays an important role in reasoning human body configuration and invisible keypoints.
Inspired by this, we propose a cascaded context mixer ( CCM) which efficiently integrates spatial and channel context information.
To maximize CCM's representation capability, we develop a hard-negative person detection mining strategy and a joint-training strategy.
We present several sub-pixel refinement techniques for postprocessing keypoint predictions to improve detection accuracy.
arXiv Detail & Related papers (2020-02-03T02:24:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.