Multi-spectral Vehicle Re-identification with Cross-directional
Consistency Network and a High-quality Benchmark
- URL: http://arxiv.org/abs/2208.00632v2
- Date: Mon, 26 Jun 2023 09:50:56 GMT
- Title: Multi-spectral Vehicle Re-identification with Cross-directional
Consistency Network and a High-quality Benchmark
- Authors: Aihua Zheng, Xianpeng Zhu, Zhiqi Ma, Chenglong Li, Jin Tang, Jixin Ma
- Abstract summary: We propose a novel cross-directional consistency network to overcome the discrepancies from both modality and sample aspects.
In particular, we design a new cross-directional center loss to pull the modality centers of each identity close to mitigate cross-modality discrepancy.
Such strategy can generate discriminative multi-spectral feature representations for vehicle Re-ID.
- Score: 32.17985248544149
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: To tackle the challenge of vehicle re-identification (Re-ID) in complex
lighting environments and diverse scenes, multi-spectral sources like visible
and infrared information are taken into consideration due to their excellent
complementary advantages.
However, multi-spectral vehicle Re-ID suffers cross-modality discrepancy
caused by heterogeneous properties of different modalities as well as a big
challenge of the diverse appearance with different views in each identity.
Meanwhile, diverse environmental interference leads to heavy sample
distributional discrepancy in each modality.
In this work, we propose a novel cross-directional consistency network to
simultaneously overcome the discrepancies from both modality and sample
aspects.
In particular, we design a new cross-directional center loss to pull the
modality centers of each identity close to mitigate cross-modality discrepancy,
while the sample centers of each identity close to alleviate the sample
discrepancy. Such strategy can generate discriminative multi-spectral feature
representations for vehicle Re-ID.
In addition, we design an adaptive layer normalization unit to dynamically
adjust individual feature distribution to handle distributional discrepancy of
intra-modality features for robust learning.
To provide a comprehensive evaluation platform, we create a high-quality
RGB-NIR-TIR multi-spectral vehicle Re-ID benchmark (MSVR310), including 310
different vehicles from a broad range of viewpoints, time spans and
environmental complexities.
Comprehensive experiments on both created and public datasets demonstrate the
effectiveness of the proposed approach comparing to the state-of-the-art
methods.
Related papers
- Cross-Modality Perturbation Synergy Attack for Person Re-identification [66.48494594909123]
The main challenge in cross-modality ReID lies in effectively dealing with visual differences between different modalities.
Existing attack methods have primarily focused on the characteristics of the visible image modality.
This study proposes a universal perturbation attack specifically designed for cross-modality ReID.
arXiv Detail & Related papers (2024-01-18T15:56:23Z) - Modality Unifying Network for Visible-Infrared Person Re-Identification [24.186989535051623]
Visible-infrared person re-identification (VI-ReID) is a challenging task due to large cross-modality discrepancies and intra-class variations.
Existing methods mainly focus on learning modality-shared representations by embedding different modalities into the same feature space.
We propose a novel Modality Unifying Network (MUN) to explore a robust auxiliary modality for VI-ReID.
arXiv Detail & Related papers (2023-09-12T14:22:22Z) - DealMVC: Dual Contrastive Calibration for Multi-view Clustering [78.54355167448614]
We propose a novel Dual contrastive calibration network for Multi-View Clustering (DealMVC)
We first design a fusion mechanism to obtain a global cross-view feature. Then, a global contrastive calibration loss is proposed by aligning the view feature similarity graph and the high-confidence pseudo-label graph.
During the training procedure, the interacted cross-view feature is jointly optimized at both local and global levels.
arXiv Detail & Related papers (2023-08-17T14:14:28Z) - Dynamic Enhancement Network for Partial Multi-modality Person
Re-identification [52.70235136651996]
We design a novel dynamic enhancement network (DENet), which allows missing arbitrary modalities while maintaining the representation ability of multiple modalities.
Since the missing state might be changeable, we design a dynamic enhancement module, which dynamically enhances modality features according to the missing state in an adaptive manner.
arXiv Detail & Related papers (2023-05-25T06:22:01Z) - Exploring Invariant Representation for Visible-Infrared Person
Re-Identification [77.06940947765406]
Cross-spectral person re-identification, which aims to associate identities to pedestrians across different spectra, faces a main challenge of the modality discrepancy.
In this paper, we address the problem from both image-level and feature-level in an end-to-end hybrid learning framework named robust feature mining network (RFM)
Experiment results on two standard cross-spectral person re-identification datasets, RegDB and SYSU-MM01, have demonstrated state-of-the-art performance.
arXiv Detail & Related papers (2023-02-02T05:24:50Z) - Discriminative-Region Attention and Orthogonal-View Generation Model for
Vehicle Re-Identification [7.5366501970852955]
Multiple challenges hamper the applications of vision-based vehicle Re-ID methods.
The proposed DRA model can automatically extract the discriminative region features, which can distinguish similar vehicles.
And the OVG model can generate multi-view features based on the input view features to reduce the impact of viewpoint mismatches.
arXiv Detail & Related papers (2022-04-28T07:46:03Z) - Weakly Aligned Feature Fusion for Multimodal Object Detection [52.15436349488198]
multimodal data often suffer from the position shift problem, i.e., the image pair is not strictly aligned.
This problem makes it difficult to fuse multimodal features and puzzles the convolutional neural network (CNN) training.
In this article, we propose a general multimodal detector named aligned region CNN (AR-CNN) to tackle the position shift problem.
arXiv Detail & Related papers (2022-04-21T02:35:23Z) - Multi-Scale Cascading Network with Compact Feature Learning for
RGB-Infrared Person Re-Identification [35.55895776505113]
Multi-Scale Part-Aware Cascading framework (MSPAC) is formulated by aggregating multi-scale fine-grained features from part to global.
Cross-modality correlations can thus be efficiently explored on salient features for distinctive modality-invariant feature learning.
arXiv Detail & Related papers (2020-12-12T15:39:11Z) - DCDLearn: Multi-order Deep Cross-distance Learning for Vehicle
Re-Identification [22.547915009758256]
This paper formulates a multi-order deep cross-distance learning model for vehicle re-identification.
One-view CycleGAN model is developed to alleviate exhaustive and enumerative cross-camera matching problem.
Experiments on three vehicle Re-ID datasets demonstrate that the proposed method achieves significant improvement over the state-of-the-arts.
arXiv Detail & Related papers (2020-03-25T10:46:54Z) - When Relation Networks meet GANs: Relation GANs with Triplet Loss [110.7572918636599]
Training stability is still a lingering concern of generative adversarial networks (GANs)
In this paper, we explore a relation network architecture for the discriminator and design a triplet loss which performs better generalization and stability.
Experiments on benchmark datasets show that the proposed relation discriminator and new loss can provide significant improvement on variable vision tasks.
arXiv Detail & Related papers (2020-02-24T11:35:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.