PAMTRI: Pose-Aware Multi-Task Learning for Vehicle Re-Identification
Using Highly Randomized Synthetic Data
- URL: http://arxiv.org/abs/2005.00673v1
- Date: Sat, 2 May 2020 01:29:09 GMT
- Title: PAMTRI: Pose-Aware Multi-Task Learning for Vehicle Re-Identification
Using Highly Randomized Synthetic Data
- Authors: Zheng Tang, Milind Naphade, Stan Birchfield, Jonathan Tremblay,
William Hodge, Ratnesh Kumar, Shuo Wang, Xiaodong Yang
- Abstract summary: Vehicle ReID is challenging due to 1) high intra-class variability (caused by the dependency of shape and appearance on viewpoint), and 2) small inter-class variability (caused by the similarity in shape and appearance between vehicles produced by different manufacturers)
We propose a Pose-Aware Multi-Task Re-Identification (PAMTRI) framework.
It overcomes viewpoint-dependency by explicitly reasoning about vehicle pose and shape via keypoints, heatmaps and segments from pose estimation.
It jointly classifies semantic vehicle attributes (colors and types) while performing ReID, through multi-task learning with the embedded pose representations.
- Score: 34.66187690224724
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In comparison with person re-identification (ReID), which has been widely
studied in the research community, vehicle ReID has received less attention.
Vehicle ReID is challenging due to 1) high intra-class variability (caused by
the dependency of shape and appearance on viewpoint), and 2) small inter-class
variability (caused by the similarity in shape and appearance between vehicles
produced by different manufacturers). To address these challenges, we propose a
Pose-Aware Multi-Task Re-Identification (PAMTRI) framework. This approach
includes two innovations compared with previous methods. First, it overcomes
viewpoint-dependency by explicitly reasoning about vehicle pose and shape via
keypoints, heatmaps and segments from pose estimation. Second, it jointly
classifies semantic vehicle attributes (colors and types) while performing
ReID, through multi-task learning with the embedded pose representations. Since
manually labeling images with detailed pose and attribute information is
prohibitive, we create a large-scale highly randomized synthetic dataset with
automatically annotated vehicle attributes for training. Extensive experiments
validate the effectiveness of each proposed component, showing that PAMTRI
achieves significant improvement over state-of-the-art on two mainstream
vehicle ReID benchmarks: VeRi and CityFlow-ReID. Code and models are available
at https://github.com/NVlabs/PAMTRI.
Related papers
- VehicleGAN: Pair-flexible Pose Guided Image Synthesis for Vehicle Re-identification [27.075761782915496]
This paper proposes to synthesize a large number of vehicle images in the target pose.
Considering the paired data of the same vehicles in different traffic surveillance cameras might be not available in the real world, we propose VehicleGAN.
Because of the feature distribution difference between real and synthetic data, we propose a new Joint Metric Learning (JML) via effective feature-level fusion.
arXiv Detail & Related papers (2023-11-27T19:34:04Z) - Exploring Fine-Grained Representation and Recomposition for Cloth-Changing Person Re-Identification [78.52704557647438]
We propose a novel FIne-grained Representation and Recomposition (FIRe$2$) framework to tackle both limitations without any auxiliary annotation or data.
Experiments demonstrate that FIRe$2$ can achieve state-of-the-art performance on five widely-used cloth-changing person Re-ID benchmarks.
arXiv Detail & Related papers (2023-08-21T12:59:48Z) - ConMAE: Contour Guided MAE for Unsupervised Vehicle Re-Identification [8.950873153831735]
This work designs a Contour Guided Masked Autoencoder for Unsupervised Vehicle Re-Identification (ConMAE)
Considering that Masked Autoencoder (MAE) has shown excellent performance in self-supervised learning, this work designs a Contour Guided Masked Autoencoder for Unsupervised Vehicle Re-Identification (ConMAE)
arXiv Detail & Related papers (2023-02-11T12:10:25Z) - Discriminative-Region Attention and Orthogonal-View Generation Model for
Vehicle Re-Identification [7.5366501970852955]
Multiple challenges hamper the applications of vision-based vehicle Re-ID methods.
The proposed DRA model can automatically extract the discriminative region features, which can distinguish similar vehicles.
And the OVG model can generate multi-view features based on the input view features to reduce the impact of viewpoint mismatches.
arXiv Detail & Related papers (2022-04-28T07:46:03Z) - Pluggable Weakly-Supervised Cross-View Learning for Accurate Vehicle
Re-Identification [53.6218051770131]
Cross-view consistent feature representation is key for accurate vehicle ReID.
Existing approaches resort to supervised cross-view learning using extensive extra viewpoints annotations.
We present a pluggable Weakly-supervised Cross-View Learning (WCVL) module for vehicle ReID.
arXiv Detail & Related papers (2021-03-09T11:51:09Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - VehicleNet: Learning Robust Visual Representation for Vehicle
Re-identification [116.1587709521173]
We propose to build a large-scale vehicle dataset (called VehicleNet) by harnessing four public vehicle datasets.
We design a simple yet effective two-stage progressive approach to learning more robust visual representation from VehicleNet.
We achieve the state-of-art accuracy of 86.07% mAP on the private test set of AICity Challenge.
arXiv Detail & Related papers (2020-04-14T05:06:38Z) - The Devil is in the Details: Self-Supervised Attention for Vehicle
Re-Identification [75.3310894042132]
Self-supervised Attention for Vehicle Re-identification (SAVER) is a novel approach to effectively learn vehicle-specific discriminative features.
We show that SAVER improves upon the state-of-the-art on challenging VeRi, VehicleID, Vehicle-1M and VERI-Wild datasets.
arXiv Detail & Related papers (2020-04-14T02:24:47Z) - Parsing-based View-aware Embedding Network for Vehicle Re-Identification [138.11983486734576]
We propose a parsing-based view-aware embedding network (PVEN) to achieve the view-aware feature alignment and enhancement for vehicle ReID.
The experiments conducted on three datasets show that our model outperforms state-of-the-art methods by a large margin.
arXiv Detail & Related papers (2020-04-10T13:06:09Z) - An Attention-Based Deep Learning Model for Multiple Pedestrian
Attributes Recognition [4.6898263272139795]
This paper provides a novel solution to the problem of automatic characterization of pedestrians in surveillance footage.
We propose a multi-task deep model that uses an element-wise multiplication layer to extract more comprehensive feature representations.
Our experiments were performed on two well-known datasets (RAP and PETA) and point for the superiority of the proposed method with respect to the state-of-the-art.
arXiv Detail & Related papers (2020-04-02T16:21:14Z) - Attribute-guided Feature Learning Network for Vehicle Re-identification [13.75036137728257]
Vehicle re-identification (reID) plays an important role in the automatic analysis of the increasing urban surveillance videos.
This paper proposes a novel Attribute-Guided Network (AGNet), which could learn global representation with the abundant attribute features.
arXiv Detail & Related papers (2020-01-12T06:57:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.