Batch Coherence-Driven Network for Part-aware Person Re-Identification
- URL: http://arxiv.org/abs/2009.09692v2
- Date: Fri, 7 May 2021 07:21:24 GMT
- Title: Batch Coherence-Driven Network for Part-aware Person Re-Identification
- Authors: Kan Wang, Pengfei Wang, Changxing Ding, and Dacheng Tao
- Abstract summary: Existing part-aware person re-identification methods typically employ two separate steps: namely, body part detection and part-level feature extraction.
We propose NetworkBCDNet that bypasses body part during both the training and testing phases while still semantically aligned features.
- Score: 79.33809815035127
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing part-aware person re-identification methods typically employ two
separate steps: namely, body part detection and part-level feature extraction.
However, part detection introduces an additional computational cost and is
inherently challenging for low-quality images. Accordingly, in this work, we
propose a simple framework named Batch Coherence-Driven Network (BCD-Net) that
bypasses body part detection during both the training and testing phases while
still learning semantically aligned part features. Our key observation is that
the statistics in a batch of images are stable, and therefore that batch-level
constraints are robust. First, we introduce a batch coherence-guided channel
attention (BCCA) module that highlights the relevant channels for each
respective part from the output of a deep backbone model. We investigate
channelpart correspondence using a batch of training images, then impose a
novel batch-level supervision signal that helps BCCA to identify part-relevant
channels. Second, the mean position of a body part is robust and consequently
coherent between batches throughout the training process. Accordingly, we
introduce a pair of regularization terms based on the semantic consistency
between batches. The first term regularizes the high responses of BCD-Net for
each part on one batch in order to constrain it within a predefined area, while
the second encourages the aggregate of BCD-Nets responses for all parts
covering the entire human body. The above constraints guide BCD-Net to learn
diverse, complementary, and semantically aligned part-level features. Extensive
experimental results demonstrate that BCDNet consistently achieves
state-of-the-art performance on four large-scale ReID benchmarks.
Related papers
- Unified Batch Normalization: Identifying and Alleviating the Feature
Condensation in Batch Normalization and a Unified Framework [55.22949690864962]
Batch Normalization (BN) has become an essential technique in contemporary neural network design.
We propose a two-stage unified framework called Unified Batch Normalization (UBN)
UBN significantly enhances performance across different visual backbones and different vision tasks.
arXiv Detail & Related papers (2023-11-27T16:41:31Z) - FDNet: Feature Decoupled Segmentation Network for Tooth CBCT Image [5.455286826028825]
We propose FDNet, a Feature Decoupled Network, to excel in the face of the variable dental conditions encountered in CBCT scans.
The framework's effectiveness is validated through rigorous benchmarks, achieving the top Dice and IoU scores of 85.28% and 75.23%, respectively.
arXiv Detail & Related papers (2023-11-11T12:00:24Z) - GaitStrip: Gait Recognition via Effective Strip-based Feature
Representations and Multi-Level Framework [34.397404430838286]
We present a strip-based multi-level gait recognition network, named GaitStrip, to extract comprehensive gait information at different levels.
To be specific, our high-level branch explores the context of gait sequences and our low-level one focuses on detailed posture changes.
Our GaitStrip achieves state-of-the-art performance in both normal walking and complex conditions.
arXiv Detail & Related papers (2022-03-08T09:49:48Z) - Real-Time Scene Text Detection with Differentiable Binarization and
Adaptive Scale Fusion [62.269219152425556]
segmentation-based scene text detection methods have drawn extensive attention in the scene text detection field.
We propose a Differentiable Binarization (DB) module that integrates the binarization process into a segmentation network.
An efficient Adaptive Scale Fusion (ASF) module is proposed to improve the scale robustness by fusing features of different scales adaptively.
arXiv Detail & Related papers (2022-02-21T15:30:14Z) - Pose-Guided Feature Learning with Knowledge Distillation for Occluded
Person Re-Identification [137.8810152620818]
We propose a network named Pose-Guided Feature Learning with Knowledge Distillation (PGFL-KD)
The PGFL-KD consists of a main branch (MB), and two pose-guided branches, ieno, a foreground-enhanced branch (FEB), and a body part semantics aligned branch (SAB)
Experiments on occluded, partial, and holistic ReID tasks show the effectiveness of our proposed network.
arXiv Detail & Related papers (2021-07-31T03:34:27Z) - BiCnet-TKS: Learning Efficient Spatial-Temporal Representation for Video
Person Re-Identification [86.73532136686438]
We present an efficient spatial-temporal representation for video person re-identification (reID)
We propose a Bilateral Complementary Network (BiCnet) for spatial complementarity modeling.
BiCnet-TKS outperforms state-of-the-arts with about 50% less computations.
arXiv Detail & Related papers (2021-04-30T06:44:34Z) - Toward Real-World BCI: CCSPNet, A Compact Subject-Independent Motor
Imagery Framework [2.0741711594051377]
A conventional brain-computer interface (BCI) requires a complete data gathering, training, and calibration phase for each user before it can be used.
We propose a novel subject-independent BCI framework named CCSPNet that is trained on the motor imagery (MI) paradigm of a large-scale EEG signals database.
The proposed framework applies a wavelet kernel convolutional neural network (WKCNN) and a temporal convolutional neural network (TCNN) in order to represent and extract the diverse spectral features of EEG signals.
arXiv Detail & Related papers (2020-12-25T12:00:47Z) - Gradient-Induced Co-Saliency Detection [81.54194063218216]
Co-saliency detection (Co-SOD) aims to segment the common salient foreground in a group of relevant images.
In this paper, inspired by human behavior, we propose a gradient-induced co-saliency detection method.
arXiv Detail & Related papers (2020-04-28T08:40:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.