AaP-ReID: Improved Attention-Aware Person Re-identification
- URL: http://arxiv.org/abs/2309.15780v1
- Date: Wed, 27 Sep 2023 16:54:38 GMT
- Title: AaP-ReID: Improved Attention-Aware Person Re-identification
- Authors: Vipin Gautam, Shitala Prasad and Sharad Sinha
- Abstract summary: AaP-ReID is a more effective method for person ReID that incorporates channel-wise attention into a ResNet-based architecture.
Our method incorporates the Channel-Wise Attention Bottleneck block and can learn discriminating features by dynamically adjusting the importance ofeach channel in the feature maps.
- Score: 2.5761958263376745
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Person re-identification (ReID) is a well-known problem in the field of
computer vision. The primary objective is to identify a specific individual
within a gallery of images. However, this task is challenging due to various
factors, such as pose variations, illumination changes, obstructions, and the
presence ofconfusing backgrounds. Existing ReID methods often fail to capture
discriminative features (e.g., head, shoes, backpacks) and instead capture
irrelevant features when the target is occluded. Motivated by the success of
part-based and attention-based ReID methods, we improve AlignedReID++ and
present AaP-ReID, a more effective method for person ReID that incorporates
channel-wise attention into a ResNet-based architecture. Our method
incorporates the Channel-Wise Attention Bottleneck (CWAbottleneck) block and
can learn discriminating features by dynamically adjusting the importance
ofeach channel in the feature maps. We evaluated Aap-ReID on three benchmark
datasets: Market-1501, DukeMTMC-reID, and CUHK03. When compared with
state-of-the-art person ReID methods, we achieve competitive results with
rank-1 accuracies of 95.6% on Market-1501, 90.6% on DukeMTMC-reID, and 82.4% on
CUHK03.
Related papers
- Multiple Information Prompt Learning for Cloth-Changing Person Re-Identification [17.948263914620238]
We propose a novel multiple information prompt learning (MIPL) scheme for cloth-changing person ReID.
TheCIS module is designed to decouple clothing information from the original RGB image features.
The Bio-guided attention (BGA) module is proposed to increase the learning intensity of the model for key information.
arXiv Detail & Related papers (2024-11-01T03:08:10Z) - ID-Aligner: Enhancing Identity-Preserving Text-to-Image Generation with Reward Feedback Learning [57.91881829308395]
Identity-preserving text-to-image generation (ID-T2I) has received significant attention due to its wide range of application scenarios like AI portrait and advertising.
We present textbfID-Aligner, a general feedback learning framework to enhance ID-T2I performance.
arXiv Detail & Related papers (2024-04-23T18:41:56Z) - A High-Accuracy Unsupervised Person Re-identification Method Using
Auxiliary Information Mined from Datasets [53.047542904329866]
We make use of auxiliary information mined from datasets for multi-modal feature learning.
This paper proposes three effective training tricks, including Restricted Label Smoothing Cross Entropy Loss (RLSCE), Weight Adaptive Triplet Loss (WATL) and Dynamic Training Iterations (DTI)
arXiv Detail & Related papers (2022-05-06T10:16:18Z) - Learning Instance-level Spatial-Temporal Patterns for Person
Re-identification [80.43222559182072]
We propose a novel Instance-level and Spatial-temporal Disentangled Re-ID method (InSTD) to improve Re-ID accuracy.
In our proposed framework, personalized information such as moving direction is explicitly considered to further narrow down the search space.
The proposed method achieves mAP of 90.8% on Market-1501 and 89.1% on DukeMTMC-reID, improving from the baseline 82.2% and 72.7%, respectively.
arXiv Detail & Related papers (2021-07-31T07:44:47Z) - ES-Net: Erasing Salient Parts to Learn More in Re-Identification [46.624740579314924]
We propose a novel network, Erasing-Salient Net (ES-Net), to learn comprehensive features by erasing the salient areas in an image.
Our ES-Net outperforms state-of-the-art methods on three Person re-ID benchmarks and two Vehicle re-ID benchmarks.
arXiv Detail & Related papers (2021-03-10T08:19:46Z) - Lightweight Multi-Branch Network for Person Re-Identification [6.353193172884524]
This paper proposes a lightweight network that combines global, part-based, and channel features in a unified multi-branch architecture that builds on the resource-efficient OSNet backbone.
Using a well-founded combination of training techniques and design choices, our final model achieves state-of-the-art results on CUHK03 labeled, CUHK03 detected, and Market-1501 with 85.1% mAP / 87.2% rank1, 82.4% mAP / 84.9% rank1, and 91.5% mAP / 96.3% rank1, respectively.
arXiv Detail & Related papers (2021-01-26T13:28:46Z) - Attribute-aware Identity-hard Triplet Loss for Video-based Person
Re-identification [51.110453988705395]
Video-based person re-identification (Re-ID) is an important computer vision task.
We introduce a new metric learning method called Attribute-aware Identity-hard Triplet Loss (AITL)
To achieve a complete model of video-based person Re-ID, a multi-task framework with Attribute-driven Spatio-Temporal Attention (ASTA) mechanism is also proposed.
arXiv Detail & Related papers (2020-06-13T09:15:38Z) - Transferable, Controllable, and Inconspicuous Adversarial Attacks on
Person Re-identification With Deep Mis-Ranking [83.48804199140758]
We propose a learning-to-mis-rank formulation to perturb the ranking of the system output.
We also perform a back-box attack by developing a novel multi-stage network architecture.
Our method can control the number of malicious pixels by using differentiable multi-shot sampling.
arXiv Detail & Related papers (2020-04-08T18:48:29Z) - VMRFANet:View-Specific Multi-Receptive Field Attention Network for
Person Re-identification [3.1498833540989413]
We propose a novel multi-receptive field attention (MRFA) module that utilizes filters of various sizes to help network focusing on informative pixels.
We present a view-specific mechanism that guides attention module to handle the variation of view conditions.
Our method achieves 95.5% / 88.1% in rank-1 / mAP on Market-1501, 88.9% / 80.0% on DukeMTMC-reID, 81.1% / 78.8% on CUHK03 labeled dataset and 78.9% / 75.3% on CUHK03 detected dataset.
arXiv Detail & Related papers (2020-01-21T06:31:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.