Viewpoint-Aware Channel-Wise Attentive Network for Vehicle
Re-Identification
- URL: http://arxiv.org/abs/2010.05810v1
- Date: Mon, 12 Oct 2020 16:05:41 GMT
- Title: Viewpoint-Aware Channel-Wise Attentive Network for Vehicle
Re-Identification
- Authors: Tsai-Shien Chen, Man-Yu Lee, Chih-Ting Liu, Shao-Yi Chien
- Abstract summary: Vehicle re-identification (re-ID) matches images of the same vehicle across different cameras.
It is fundamentally challenging because the dramatically different appearance caused by different viewpoints would make the framework fail to match two vehicles of the same identity.
We propose Viewpoint-aware Channel-wise Attention Mechanism (VCAM) by observing the attention mechanism from a different aspect.
- Score: 30.797140440568455
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vehicle re-identification (re-ID) matches images of the same vehicle across
different cameras. It is fundamentally challenging because the dramatically
different appearance caused by different viewpoints would make the framework
fail to match two vehicles of the same identity. Most existing works solved the
problem by extracting viewpoint-aware feature via spatial attention mechanism,
which, yet, usually suffers from noisy generated attention map or otherwise
requires expensive keypoint labels to improve the quality. In this work, we
propose Viewpoint-aware Channel-wise Attention Mechanism (VCAM) by observing
the attention mechanism from a different aspect. Our VCAM enables the feature
learning framework channel-wisely reweighing the importance of each feature
maps according to the "viewpoint" of input vehicle. Extensive experiments
validate the effectiveness of the proposed method and show that we perform
favorably against state-of-the-arts methods on the public VeRi-776 dataset and
obtain promising results on the 2020 AI City Challenge. We also conduct other
experiments to demonstrate the interpretability of how our VCAM practically
assists the learning framework.
Related papers
- ConMAE: Contour Guided MAE for Unsupervised Vehicle Re-Identification [8.950873153831735]
This work designs a Contour Guided Masked Autoencoder for Unsupervised Vehicle Re-Identification (ConMAE)
Considering that Masked Autoencoder (MAE) has shown excellent performance in self-supervised learning, this work designs a Contour Guided Masked Autoencoder for Unsupervised Vehicle Re-Identification (ConMAE)
arXiv Detail & Related papers (2023-02-11T12:10:25Z) - Complete Solution for Vehicle Re-ID in Surround-view Camera System [10.10765191655754]
Vehicle re-identification (Re-ID) is a critical component of the autonomous driving perception system.
It is difficult to identify the same vehicle in many picture frames due to the unique construction of the fisheye camera.
Our approach combines state-of-the-art accuracy with real-time performance.
arXiv Detail & Related papers (2022-12-08T07:52:55Z) - Online Deep Clustering with Video Track Consistency [85.8868194550978]
We propose an unsupervised clustering-based approach to learn visual features from video object tracks.
We show that exploiting an unsupervised class-agnostic, yet noisy, track generator yields to better accuracy compared to relying on costly and precise track annotations.
arXiv Detail & Related papers (2022-06-07T08:11:00Z) - SurroundDepth: Entangling Surrounding Views for Self-Supervised
Multi-Camera Depth Estimation [101.55622133406446]
We propose a SurroundDepth method to incorporate the information from multiple surrounding views to predict depth maps across cameras.
Specifically, we employ a joint network to process all the surrounding views and propose a cross-view transformer to effectively fuse the information from multiple views.
In experiments, our method achieves the state-of-the-art performance on the challenging multi-camera depth estimation datasets.
arXiv Detail & Related papers (2022-04-07T17:58:47Z) - Multi-Attention-Based Soft Partition Network for Vehicle
Re-Identification [12.319699929810355]
Vehicle re-identification helps in distinguishing between images of the same and other vehicles.
We propose a new vehicle re-identification network based on a multiple soft attention mechanism for capturing various discriminative regions.
Our proposed model achieved a state-of-the-art performance among the attention-based methods without metadata.
arXiv Detail & Related papers (2021-04-21T08:13:17Z) - Pluggable Weakly-Supervised Cross-View Learning for Accurate Vehicle
Re-Identification [53.6218051770131]
Cross-view consistent feature representation is key for accurate vehicle ReID.
Existing approaches resort to supervised cross-view learning using extensive extra viewpoints annotations.
We present a pluggable Weakly-supervised Cross-View Learning (WCVL) module for vehicle ReID.
arXiv Detail & Related papers (2021-03-09T11:51:09Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - V2VNet: Vehicle-to-Vehicle Communication for Joint Perception and
Prediction [74.42961817119283]
We use vehicle-to-vehicle (V2V) communication to improve the perception and motion forecasting performance of self-driving vehicles.
By intelligently aggregating the information received from multiple nearby vehicles, we can observe the same scene from different viewpoints.
arXiv Detail & Related papers (2020-08-17T17:58:26Z) - Vehicle Re-ID for Surround-view Camera System [8.919111930321206]
Vehicle re-identification (ReID) plays a critical role in the perception system of autonomous driving.
There is no existing complete solution for the surround-view system mounted on the vehicle.
We propose a novel quality evaluation mechanism to balance the effect of tracking box's drift and target's consistency.
arXiv Detail & Related papers (2020-06-30T03:25:10Z) - Parsing-based View-aware Embedding Network for Vehicle Re-Identification [138.11983486734576]
We propose a parsing-based view-aware embedding network (PVEN) to achieve the view-aware feature alignment and enhancement for vehicle ReID.
The experiments conducted on three datasets show that our model outperforms state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2020-04-10T13:06:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.