Parsing-based View-aware Embedding Network for Vehicle Re-Identification
- URL: http://arxiv.org/abs/2004.05021v1
- Date: Fri, 10 Apr 2020 13:06:09 GMT
- Title: Parsing-based View-aware Embedding Network for Vehicle Re-Identification
- Authors: Dechao Meng and Liang Li and Xuejing Liu and Yadong Li and Shijie Yang
and Zhengjun Zha and Xingyu Gao and Shuhui Wang and Qingming Huang
- Abstract summary: We propose a parsing-based view-aware embedding network (PVEN) to achieve the view-aware feature alignment and enhancement for vehicle ReID.
The experiments conducted on three datasets show that our model outperforms state-of-the-art methods by a large margin.
- Score: 138.11983486734576
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vehicle Re-Identification is to find images of the same vehicle from various
views in the cross-camera scenario. The main challenges of this task are the
large intra-instance distance caused by different views and the subtle
inter-instance discrepancy caused by similar vehicles. In this paper, we
propose a parsing-based view-aware embedding network (PVEN) to achieve the
view-aware feature alignment and enhancement for vehicle ReID. First, we
introduce a parsing network to parse a vehicle into four different views, and
then align the features by mask average pooling. Such alignment provides a
fine-grained representation of the vehicle. Second, in order to enhance the
view-aware features, we design a common-visible attention to focus on the
common visible views, which not only shortens the distance among
intra-instances, but also enlarges the discrepancy of inter-instances. The PVEN
helps capture the stable discriminative information of vehicle under different
views. The experiments conducted on three datasets show that our model
outperforms state-of-the-art methods by a large margin.
Related papers
- Cohere3D: Exploiting Temporal Coherence for Unsupervised Representation
Learning of Vision-based Autonomous Driving [73.3702076688159]
We propose a novel contrastive learning algorithm, Cohere3D, to learn coherent instance representations in a long-term input sequence.
We evaluate our algorithm by finetuning the pretrained model on various downstream perception, prediction, and planning tasks.
arXiv Detail & Related papers (2024-02-23T19:43:01Z) - Multi-query Vehicle Re-identification: Viewpoint-conditioned Network,
Unified Dataset and New Metric [30.344288906037345]
We propose a more realistic and easily accessible task, called multi-query vehicle Re-ID.
We design a novel viewpoint-conditioned network (VCNet), which adaptively combines the complementary information from different vehicle viewpoints.
Second, we create a unified benchmark dataset, taken by 6142 cameras from a real-life transportation surveillance system.
Third, we design a new evaluation metric, called mean cross-scene precision (mCSP), which measures the ability of cross-scene recognition.
arXiv Detail & Related papers (2023-05-25T06:22:03Z) - Discriminative-Region Attention and Orthogonal-View Generation Model for
Vehicle Re-Identification [7.5366501970852955]
Multiple challenges hamper the applications of vision-based vehicle Re-ID methods.
The proposed DRA model can automatically extract the discriminative region features, which can distinguish similar vehicles.
And the OVG model can generate multi-view features based on the input view features to reduce the impact of viewpoint mismatches.
arXiv Detail & Related papers (2022-04-28T07:46:03Z) - Pluggable Weakly-Supervised Cross-View Learning for Accurate Vehicle
Re-Identification [53.6218051770131]
Cross-view consistent feature representation is key for accurate vehicle ReID.
Existing approaches resort to supervised cross-view learning using extensive extra viewpoints annotations.
We present a pluggable Weakly-supervised Cross-View Learning (WCVL) module for vehicle ReID.
arXiv Detail & Related papers (2021-03-09T11:51:09Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Viewpoint-Aware Channel-Wise Attentive Network for Vehicle
Re-Identification [30.797140440568455]
Vehicle re-identification (re-ID) matches images of the same vehicle across different cameras.
It is fundamentally challenging because the dramatically different appearance caused by different viewpoints would make the framework fail to match two vehicles of the same identity.
We propose Viewpoint-aware Channel-wise Attention Mechanism (VCAM) by observing the attention mechanism from a different aspect.
arXiv Detail & Related papers (2020-10-12T16:05:41Z) - Orientation-aware Vehicle Re-identification with Semantics-guided Part
Attention Network [33.712450134663236]
We propose a dedicated Semantics-guided Part Attention Network (SPAN) to robustly predict part attention masks for different views of vehicles.
With the help of part attention masks, we can extract discriminative features in each part separately.
Then we introduce Co-occurrence Part-attentive Distance Metric (CPDM) which places greater emphasis on co-occurrence vehicle parts.
arXiv Detail & Related papers (2020-08-26T07:33:09Z) - VehicleNet: Learning Robust Visual Representation for Vehicle
Re-identification [116.1587709521173]
We propose to build a large-scale vehicle dataset (called VehicleNet) by harnessing four public vehicle datasets.
We design a simple yet effective two-stage progressive approach to learning more robust visual representation from VehicleNet.
We achieve the state-of-art accuracy of 86.07% mAP on the private test set of AICity Challenge.
arXiv Detail & Related papers (2020-04-14T05:06:38Z) - The Devil is in the Details: Self-Supervised Attention for Vehicle
Re-Identification [75.3310894042132]
Self-supervised Attention for Vehicle Re-identification (SAVER) is a novel approach to effectively learn vehicle-specific discriminative features.
We show that SAVER improves upon the state-of-the-art on challenging VeRi, VehicleID, Vehicle-1M and VERI-Wild datasets.
arXiv Detail & Related papers (2020-04-14T02:24:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.