DSAM: A Distance Shrinking with Angular Marginalizing Loss for High
Performance Vehicle Re-identificatio
- URL: http://arxiv.org/abs/2011.06228v3
- Date: Thu, 9 Sep 2021 02:59:59 GMT
- Title: DSAM: A Distance Shrinking with Angular Marginalizing Loss for High
Performance Vehicle Re-identificatio
- Authors: Jiangtao Kong, Yu Cheng, Benjia Zhou, Kai Li, Junliang Xing
- Abstract summary: Vehicle Re-identification (ReID) is an important yet challenging problem in computer vision.
To obtain a high-performance vehicle ReID model, we present a novel Distance Shrinking with Angular Marginalizing (DSAM) loss function.
- Score: 39.45750548358041
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Vehicle Re-identification (ReID) is an important yet challenging problem in
computer vision. Compared to other visual objects like faces and persons,
vehicles simultaneously exhibit much larger intraclass viewpoint variations and
interclass visual similarities, making most exiting loss functions designed for
face recognition and person ReID unsuitable for vehicle ReID. To obtain a
high-performance vehicle ReID model, we present a novel Distance Shrinking with
Angular Marginalizing (DSAM) loss function to perform hybrid learning in both
the Original Feature Space (OFS) and the Feature Angular Space (FAS) using the
local verification and the global identification information. Specifically, it
shrinks the distance between samples of the same class locally in the Original
Feature Space while keeps samples of different classes far away in the Feature
Angular Space. The shrinking and marginalizing operations are performed during
each iteration of the training process and are suitable for different SoftMax
based loss functions. We evaluate the DSAM loss function on three large vehicle
ReID datasets with detailed analyses and extensive comparisons with many
competing vehicle ReID methods. Experimental results show that our DSAM loss
enhances the SoftMax loss by a large margin on the PKU-VD1-Large dataset:
10.41% for mAP, 5.29% for cmc1, and 4.60% for cmc5. Moreover, the mAP is
increased by 9.34% on the PKU-VehicleID dataset and 6.13% on the VeRi-776
dataset. Source code will be released to facilitate further studies in this
research direction.
Related papers
- VehicleGAN: Pair-flexible Pose Guided Image Synthesis for Vehicle Re-identification [27.075761782915496]
This paper proposes to synthesize a large number of vehicle images in the target pose.
Considering the paired data of the same vehicles in different traffic surveillance cameras might be not available in the real world, we propose VehicleGAN.
Because of the feature distribution difference between real and synthetic data, we propose a new Joint Metric Learning (JML) via effective feature-level fusion.
arXiv Detail & Related papers (2023-11-27T19:34:04Z) - V2X-AHD:Vehicle-to-Everything Cooperation Perception via Asymmetric
Heterogenous Distillation Network [13.248981195106069]
We propose a multi-view vehicle-road cooperation perception system, vehicle-to-everything cooperative perception (V2X-AHD)
The V2X-AHD can effectively improve the accuracy of 3D object detection and reduce the number of network parameters, according to this study.
arXiv Detail & Related papers (2023-10-10T13:12:03Z) - LoLep: Single-View View Synthesis with Locally-Learned Planes and
Self-Attention Occlusion Inference [66.45326873274908]
We propose a novel method, LoLep, which regresses Locally-Learned planes from a single RGB image to represent scenes accurately.
Compared to MINE, our approach has an LPIPS reduction of 4.8%-9.0% and an RV reduction of 73.9%-83.5%.
arXiv Detail & Related papers (2023-07-23T03:38:55Z) - Instance-Variant Loss with Gaussian RBF Kernel for 3D Cross-modal
Retriveal [52.41252219453429]
Existing methods treat all instances equally, applying the same penalty strength to instances with varying degrees of difficulty.
This can result in ambiguous convergence or local optima, severely compromising the separability of the feature space.
We propose an Instance-Variant loss to assign different penalty strengths to different instances, improving the space separability.
arXiv Detail & Related papers (2023-05-07T10:12:14Z) - Towards Multimodal Multitask Scene Understanding Models for Indoor
Mobile Agents [49.904531485843464]
In this paper, we discuss the main challenge: insufficient, or even no, labeled data for real-world indoor environments.
We describe MMISM (Multi-modality input Multi-task output Indoor Scene understanding Model) to tackle the above challenges.
MMISM considers RGB images as well as sparse Lidar points as inputs and 3D object detection, depth completion, human pose estimation, and semantic segmentation as output tasks.
We show that MMISM performs on par or even better than single-task models.
arXiv Detail & Related papers (2022-09-27T04:49:19Z) - Discriminative-Region Attention and Orthogonal-View Generation Model for
Vehicle Re-Identification [7.5366501970852955]
Multiple challenges hamper the applications of vision-based vehicle Re-ID methods.
The proposed DRA model can automatically extract the discriminative region features, which can distinguish similar vehicles.
And the OVG model can generate multi-view features based on the input view features to reduce the impact of viewpoint mismatches.
arXiv Detail & Related papers (2022-04-28T07:46:03Z) - A Strong Baseline for Vehicle Re-Identification [1.9573380763700712]
Vehicle Re-ID aims to identify the same vehicle across different cameras.
In this paper, we first analyze the main factors hindering the Vehicle Re-ID performance.
We then present our solutions, specifically targeting the Track 2 of the 5th AI Challenge.
arXiv Detail & Related papers (2021-04-22T03:54:55Z) - Loss Function Discovery for Object Detection via Convergence-Simulation
Driven Search [101.73248560009124]
We propose an effective convergence-simulation driven evolutionary search algorithm, CSE-Autoloss, for speeding up the search progress.
We conduct extensive evaluations of loss function search on popular detectors and validate the good generalization capability of searched losses.
Our experiments show that the best-discovered loss function combinations outperform default combinations by 1.1% and 0.8% in terms of mAP for two-stage and one-stage detectors.
arXiv Detail & Related papers (2021-02-09T08:34:52Z) - Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data
Augmentation [77.60050239225086]
We propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images.
Our approach is fully automatic without any human interaction.
We present a multi-task network for VUS parsing and a multi-stream network for VHI parsing.
arXiv Detail & Related papers (2020-12-15T03:03:38Z) - Discriminative Feature and Dictionary Learning with Part-aware Model for
Vehicle Re-identification [13.556590446194283]
Vehicle re-identification (re-ID) technology is a challenging task since vehicles of the same design or manufacturer show similar appearance.
We propose Triplet Center Loss based Part-aware Model ( TCPM) that leverages the discriminative features in part details of vehicles to refine the accuracy of vehicle re-identification.
Our proposed TCPM has an enormous preference over the existing state-of-the-art methods on benchmark datasets VehicleID and VeRi-776.
arXiv Detail & Related papers (2020-03-16T12:15:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.